Search results for: passive optical networks (PONs)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5051

Search results for: passive optical networks (PONs)

611 Raman Tweezers Spectroscopy Study of Size Dependent Silver Nanoparticles Toxicity on Erythrocytes

Authors: Surekha Barkur, Aseefhali Bankapur, Santhosh Chidangil

Abstract:

Raman Tweezers technique has become prevalent in single cell studies. This technique combines Raman spectroscopy which gives information about molecular vibrations, with optical tweezers which use a tightly focused laser beam for trapping the single cells. Thus Raman Tweezers enabled researchers analyze single cells and explore different applications. The applications of Raman Tweezers include studying blood cells, monitoring blood-related disorders, silver nanoparticle-induced stress, etc. There is increased interest in the toxic effect of nanoparticles with an increase in the various applications of nanoparticles. The interaction of these nanoparticles with the cells may vary with their size. We have studied the effect of silver nanoparticles of sizes 10nm, 40nm, and 100nm on erythrocytes using Raman Tweezers technique. Our aim was to investigate the size dependence of the nanoparticle effect on RBCs. We used 785nm laser (Starbright Diode Laser, Torsana Laser Tech, Denmark) for both trapping and Raman spectroscopic studies. 100 x oil immersion objectives with high numerical aperture (NA 1.3) is used to focus the laser beam into a sample cell. The back-scattered light is collected using the same microscope objective and focused into the spectrometer (Horiba Jobin Vyon iHR320 with 1200grooves/mm grating blazed at 750nm). Liquid nitrogen cooled CCD (Symphony CCD-1024x256-OPEN-1LS) was used for signal detection. Blood was drawn from healthy volunteers in vacutainer tubes and centrifuged to separate the blood components. 1.5 ml of silver nanoparticles was washed twice with distilled water leaving 0.1 ml silver nanoparticles in the bottom of the vial. The concentration of silver nanoparticles is 0.02mg/ml so the 0.03mg of nanoparticles will be present in the 0.1 ml nanoparticles obtained. The 25 ul of RBCs were diluted in 2 ml of PBS solution and then treated with 50 ul (0.015mg) of nanoparticles and incubated in CO2 incubator. Raman spectroscopic measurements were done after 24 hours and 48 hours of incubation. All the spectra were recorded with 10mW laser power (785nm diode laser), 60s of accumulation time and 2 accumulations. Major changes were observed in the peaks 565 cm-1, 1211 cm-1, 1224 cm-1, 1371 cm-1, 1638 cm-1. A decrease in intensity of 565 cm-1, increase in 1211 cm-1 with a reduction in 1224 cm-1, increase in intensity of 1371 cm-1 also peak disappearing at 1635 cm-1 indicates deoxygenation of hemoglobin. Nanoparticles with higher size were showing maximum spectral changes. Lesser changes observed in case of 10nm nanoparticle-treated erythrocyte spectra.

Keywords: erythrocytes, nanoparticle-induced toxicity, Raman tweezers, silver nanoparticles

Procedia PDF Downloads 286
610 The Biopsychosocial Effects of Amputation on Transtibial Amputees in Kwazulu-Natal

Authors: Riyona Chetty, Raisuyah Bhagwan, Nalini Govender

Abstract:

Background: A myriad of physical, psychosocial, and environmental sequelae are associated with limb loss. However, there is a paucity of empirical South African data, which focuses on these sequelae, how they interface with the amputee’s quality of life as well as the challenges they experience following amputation. Objective: This study sought to explore the biopsychosocial effects of amputation and how amputation affected the quality of life of transtibial amputees. Setting: Participants were recruited from a medical facility, under the KwaZulu-Natal Department of Health in South Africa. Methods: A qualitative approach guided this study. Data was collected using one-on-one interviews with 14 unilateral transtibial amputees. Data was analysed thematically. Results: Five broad themes emerged from the inquiry, which captured amputees’ experiences of phantom limb pain, body image disturbances, and their challenges related to adapting to daily activities. Participants also expressed the salience of familial support as well as the importance of psychological interventions to cope. Conclusion: The findings suggested that support networks and professional psychological intervention are imperative in facilitating successful adjustment to the amputation experience. Raising awareness of limb loss in both, rural and urban settings may help reduce the stigma attached to it. Contribution: Quality of life comprises several domains, namely physical, psychological, environmental, and social albeit limited local and international data exists regarding the environmental and social effects. This study brought to the fore the positive and negative effects of amputation in each domain, as well as various strategies that facilitate successful adjustment to amputation.

Keywords: amputation, quality of life, biopsychosocial, phantom limb pain, body image, support

Procedia PDF Downloads 47
609 Business Feasibility of Online Marketing of Food and Beverages Products in India

Authors: Dimpy Shah

Abstract:

The global economy has substantially changed in last three decades. Now almost all markets are transparent and visible for global customers. The corporates are now no more reliant on local markets for trade. The information technology revolution has changed business dynamics and marketing practices of corporate. The markets are divided into two different formats: traditional and virtual. In very short span of time, many e-commerce portals have captured global market. This strategy is well supported by global delivery system of multinational logistic companies. Now the markets are dealing with global supply chain networks, which are more demand driven and customer oriented. The corporate have realized importance of supply chain integration and marketing in this competitive environment. The Indian markets are also significantly affected with all these changes. In terms of population, India is in second place after China. In terms of demography, almost half of the population is of youth. It has been observed that the Indian youth are more inclined towards e-commerce and prefer to buy goods from web portal. Initially, this trend was observed in Indian service sector, textile and electronic goods and now further extended in other product categories. The FMCG companies have also recognized this change and started integration of their supply chain with e-commerce platform. This paper attempts to understand contemporary marketing practices of corporate in e-commerce business in Indian food and beverages segment and also tries to identify innovative marketing practices for proper execution of their strategies. The findings are mainly focused on supply chain re-integration and brand building strategies with proper utilization of social media.

Keywords: FMCG (Fast Moving Consumer Goods), ISCM (Integrated supply chain management), RFID (Radio Frequency Identification), traditional and virtual formats

Procedia PDF Downloads 269
608 Hyper Parameter Optimization of Deep Convolutional Neural Networks for Pavement Distress Classification

Authors: Oumaima Khlifati, Khadija Baba

Abstract:

Pavement distress is the main factor responsible for the deterioration of road structure durability, damage vehicles, and driver comfort. Transportation agencies spend a high proportion of their funds on pavement monitoring and maintenance. The auscultation of pavement distress was based on the manual survey, which was extremely time consuming, labor intensive, and required domain expertise. Therefore, the automatic distress detection is needed to reduce the cost of manual inspection and avoid more serious damage by implementing the appropriate remediation actions at the right time. Inspired by recent deep learning applications, this paper proposes an algorithm for automatic road distress detection and classification using on the Deep Convolutional Neural Network (DCNN). In this study, the types of pavement distress are classified as transverse or longitudinal cracking, alligator, pothole, and intact pavement. The dataset used in this work is composed of public asphalt pavement images. In order to learn the structure of the different type of distress, the DCNN models are trained and tested as a multi-label classification task. In addition, to get the highest accuracy for our model, we adjust the structural optimization hyper parameters such as the number of convolutions and max pooling, filers, size of filters, loss functions, activation functions, and optimizer and fine-tuning hyper parameters that conclude batch size and learning rate. The optimization of the model is executed by checking all feasible combinations and selecting the best performing one. The model, after being optimized, performance metrics is calculated, which describe the training and validation accuracies, precision, recall, and F1 score.

Keywords: distress pavement, hyperparameters, automatic classification, deep learning

Procedia PDF Downloads 82
607 Developing a DNN Model for the Production of Biogas From a Hybrid BO-TPE System in an Anaerobic Wastewater Treatment Plant

Authors: Hadjer Sadoune, Liza Lamini, Scherazade Krim, Amel Djouadi, Rachida Rihani

Abstract:

Deep neural networks are highly regarded for their accuracy in predicting intricate fermentation processes. Their ability to learn from a large amount of datasets through artificial intelligence makes them particularly effective models. The primary obstacle in improving the performance of these models is to carefully choose the suitable hyperparameters, including the neural network architecture (number of hidden layers and hidden units), activation function, optimizer, learning rate, and other relevant factors. This study predicts biogas production from real wastewater treatment plant data using a sophisticated approach: hybrid Bayesian optimization with a tree-structured Parzen estimator (BO-TPE) for an optimised deep neural network (DNN) model. The plant utilizes an Upflow Anaerobic Sludge Blanket (UASB) digester that treats industrial wastewater from soft drinks and breweries. The digester has a working volume of 1574 m3 and a total volume of 1914 m3. Its internal diameter and height were 19 and 7.14 m, respectively. The data preprocessing was conducted with meticulous attention to preserving data quality while avoiding data reduction. Three normalization techniques were applied to the pre-processed data (MinMaxScaler, RobustScaler and StandardScaler) and compared with the Non-Normalized data. The RobustScaler approach has strong predictive ability for estimating the volume of biogas produced. The highest predicted biogas volume was 2236.105 Nm³/d, with coefficient of determination (R2), mean absolute error (MAE), and root mean square error (RMSE) values of 0.712, 164.610, and 223.429, respectively.

Keywords: anaerobic digestion, biogas production, deep neural network, hybrid bo-tpe, hyperparameters tuning

Procedia PDF Downloads 29
606 Experimental Study of Impregnated Diamond Bit Wear During Sharpening

Authors: Rui Huang, Thomas Richard, Masood Mostofi

Abstract:

The lifetime of impregnated diamond bits and their drilling efficiency are in part governed by the bit wear conditions, not only the extent of the diamonds’ wear but also their exposure or protrusion out of the matrix bonding. As much as individual diamonds wear, the bonding matrix does also wear through two-body abrasion (direct matrix-rock contact) and three-body erosion (cuttings trapped in the space between rock and matrix). Although there is some work dedicated to the study of diamond bit wear, there is still a lack of understanding on how matrix erosion and diamond exposure relate to the bit drilling response and drilling efficiency, as well as no literature on the process that governs bit sharpening a procedure commonly implemented by drillers when the extent of diamond polishing yield extremely low rate of penetration. The aim of this research is (i) to derive a correlation between the wear state of the bit and the drilling performance but also (ii) to gain a better understanding of the process associated with tool sharpening. The research effort combines specific drilling experiments and precise mapping of the tool-cutting face (impregnated diamond bits and segments). Bit wear is produced by drilling through a rock sample at a fixed rate of penetration for a given period of time. Before and after each wear test, the bit drilling response and thus efficiency is mapped out using a tailored design experimental protocol. After each drilling test, the bit or segment cutting face is scanned with an optical microscope. The test results show that, under the fixed rate of penetration, diamond exposure increases with drilling distance but at a decreasing rate, up to a threshold exposure that corresponds to the optimum drilling condition for this feed rate. The data further shows that the threshold exposure scale with the rate of penetration up to a point where exposure reaches a maximum beyond which no more matrix can be eroded under normal drilling conditions. The second phase of this research focuses on the wear process referred as bit sharpening. Drillers rely on different approaches (increase feed rate or decrease flow rate) with the aim of tearing worn diamonds away from the bit matrix, wearing out some of the matrix, and thus exposing fresh sharp diamonds and recovering a higher rate of penetration. Although a common procedure, there is no rigorous methodology to sharpen the bit and avoid excessive wear or bit damage. This paper aims to gain some insight into the mechanisms that accompany bit sharpening by carefully tracking diamond fracturing, matrix wear, and erosion and how they relate to drilling parameters recorded while sharpening the tool. The results show that there exist optimal conditions (operating parameters and duration of the procedure) for sharpening that minimize overall bit wear and that the extent of bit sharpening can be monitored in real-time.

Keywords: bit sharpening, diamond exposure, drilling response, impregnated diamond bit, matrix erosion, wear rate

Procedia PDF Downloads 95
605 Loss Function Optimization for CNN-Based Fingerprint Anti-Spoofing

Authors: Yehjune Heo

Abstract:

As biometric systems become widely deployed, the security of identification systems can be easily attacked by various spoof materials. This paper contributes to finding a reliable and practical anti-spoofing method using Convolutional Neural Networks (CNNs) based on the types of loss functions and optimizers. The types of CNNs used in this paper include AlexNet, VGGNet, and ResNet. By using various loss functions including Cross-Entropy, Center Loss, Cosine Proximity, and Hinge Loss, and various loss optimizers which include Adam, SGD, RMSProp, Adadelta, Adagrad, and Nadam, we obtained significant performance changes. We realize that choosing the correct loss function for each model is crucial since different loss functions lead to different errors on the same evaluation. By using a subset of the Livdet 2017 database, we validate our approach to compare the generalization power. It is important to note that we use a subset of LiveDet and the database is the same across all training and testing for each model. This way, we can compare the performance, in terms of generalization, for the unseen data across all different models. The best CNN (AlexNet) with the appropriate loss function and optimizers result in more than 3% of performance gain over the other CNN models with the default loss function and optimizer. In addition to the highest generalization performance, this paper also contains the models with high accuracy associated with parameters and mean average error rates to find the model that consumes the least memory and computation time for training and testing. Although AlexNet has less complexity over other CNN models, it is proven to be very efficient. For practical anti-spoofing systems, the deployed version should use a small amount of memory and should run very fast with high anti-spoofing performance. For our deployed version on smartphones, additional processing steps, such as quantization and pruning algorithms, have been applied in our final model.

Keywords: anti-spoofing, CNN, fingerprint recognition, loss function, optimizer

Procedia PDF Downloads 132
604 An Automated Stock Investment System Using Machine Learning Techniques: An Application in Australia

Authors: Carol Anne Hargreaves

Abstract:

A key issue in stock investment is how to select representative features for stock selection. The objective of this paper is to firstly determine whether an automated stock investment system, using machine learning techniques, may be used to identify a portfolio of growth stocks that are highly likely to provide returns better than the stock market index. The second objective is to identify the technical features that best characterize whether a stock’s price is likely to go up and to identify the most important factors and their contribution to predicting the likelihood of the stock price going up. Unsupervised machine learning techniques, such as cluster analysis, were applied to the stock data to identify a cluster of stocks that was likely to go up in price – portfolio 1. Next, the principal component analysis technique was used to select stocks that were rated high on component one and component two – portfolio 2. Thirdly, a supervised machine learning technique, the logistic regression method, was used to select stocks with a high probability of their price going up – portfolio 3. The predictive models were validated with metrics such as, sensitivity (recall), specificity and overall accuracy for all models. All accuracy measures were above 70%. All portfolios outperformed the market by more than eight times. The top three stocks were selected for each of the three stock portfolios and traded in the market for one month. After one month the return for each stock portfolio was computed and compared with the stock market index returns. The returns for all three stock portfolios was 23.87% for the principal component analysis stock portfolio, 11.65% for the logistic regression portfolio and 8.88% for the K-means cluster portfolio while the stock market performance was 0.38%. This study confirms that an automated stock investment system using machine learning techniques can identify top performing stock portfolios that outperform the stock market.

Keywords: machine learning, stock market trading, logistic regression, cluster analysis, factor analysis, decision trees, neural networks, automated stock investment system

Procedia PDF Downloads 150
603 Using Hierarchical Methodology to Assist the Selection of New Business in Brazilian Companies Incubators

Authors: Izabel Cristina Zattar, Gilberto Passos Lima, Guilherme Schünemann de Oliveira

Abstract:

In Brazil, there are several institutions committed to the development of new businesses based on product innovation. Among them are business incubators, universities and science institutes. Business incubators can be defined as nurseries for new companies, which may be in the technology segment, discussed in this article. Business incubators provide services related to infrastructure, such as physical space and meeting rooms. Besides these services, incubators also offer assistance in the form of information and communication, access to finance, relationship networks and business monitoring and mentoring processes. Business incubators support not all technology companies. One of the business incubators tasks is to assess the nature and feasibility of new business proposals. To assist in this goal, this paper proposes a methodology for evaluating new business using the Analytic Hierarchy Process (AHP). This paper presents the concepts used in the assessing methodology application for new business, concepts that have been tested with positive results in practice. This study counts on three main steps: first, a hierarchy was built, based on new business manuals used by the business incubators. These books and manuals relate business selection requirements, such as the innovation status and other technological aspects. Then, a questionnaire was generated, in order to guide incubator experts in the parity comparisons at all hierarchy levels. The weights of each requirement are calculated from information obtained from the questionnaire responses. Finally, the proposed method was applied to evaluate five new business proposals, which were applying to be part of a company incubator. The main result is the classification of these new businesses, which helped the incubator experts to decide what companies were more eligible to work with. This classification may also be helpful to the decision-making process of business incubators in future selection processes.

Keywords: Analytic Hierarchy Process (AHP), Brazilian companies incubators, technology companies, incubator

Procedia PDF Downloads 393
602 A Multi-Output Network with U-Net Enhanced Class Activation Map and Robust Classification Performance for Medical Imaging Analysis

Authors: Jaiden Xuan Schraut, Leon Liu, Yiqiao Yin

Abstract:

Computer vision in medical diagnosis has achieved a high level of success in diagnosing diseases with high accuracy. However, conventional classifiers that produce an image to-label result provides insufficient information for medical professionals to judge and raise concerns over the trust and reliability of a model with results that cannot be explained. In order to gain local insight into cancerous regions, separate tasks such as imaging segmentation need to be implemented to aid the doctors in treating patients, which doubles the training time and costs which renders the diagnosis system inefficient and difficult to be accepted by the public. To tackle this issue and drive AI-first medical solutions further, this paper proposes a multi-output network that follows a U-Net architecture for image segmentation output and features an additional convolutional neural networks (CNN) module for auxiliary classification output. Class activation maps are a method of providing insight into a convolutional neural network’s feature maps that leads to its classification but in the case of lung diseases, the region of interest is enhanced by U-net-assisted Class Activation Map (CAM) visualization. Therefore, our proposed model combines image segmentation models and classifiers to crop out only the lung region of a chest X-ray’s class activation map to provide a visualization that improves the explainability and is able to generate classification results simultaneously which builds trust for AI-led diagnosis systems. The proposed U-Net model achieves 97.61% accuracy and a dice coefficient of 0.97 on testing data from the COVID-QU-Ex Dataset which includes both diseased and healthy lungs.

Keywords: multi-output network model, U-net, class activation map, image classification, medical imaging analysis

Procedia PDF Downloads 190
601 Teaching English for Specific Purposes to Business Students through Social Media

Authors: Candela Contero Urgal

Abstract:

Using realia to teach English for Specific Purposes (ESP) is a must, as it is thought to be designed to meet the students’ real needs in their professional life. Teachers are then expected to offer authentic materials and set students in authentic contexts where their learning outcomes can be highly meaningful. One way of engaging students is using social networks as a way to bridge the gap between their everyday life and their ESP learning outcomes. It is in ESP, particularly in Business English teaching, that our study focuses, as the ongoing process of digitalization is leading firms to use social media to communicate with potential clients. The present paper is aimed at carrying out a case study in which different digital tools are employed as a way to offer a collection of formats businesses are currently using so as to internationalize and advertise their products and services. A secondary objective of our study will then be to progress on the development of multidisciplinary competencies students are to acquire during their degree. A two-phased study will be presented. The first phase will cover the analysis of course tasks accomplished by undergraduate students at the University of Cadiz (Spain) in their third year of the Degree in Business Management and Administration by comparing the results obtained during the years 2019 to 2021. The second part of our study will present a survey conducted to these students in 2021 and 2022 so as to verify their interest in learning new ways to digitalize as well as internationalize their future businesses. Findings will confirm students’ interest in working with updated realia in their Business English lessons, as a consequence of their strong belief in the necessity to have authentic contexts and didactic resources. Despite the limitations social media can have as a means to teach business English, students will still find it highly beneficial since it will foster their familiarisation with the digital tools they will need to use when they get to the labour market.

Keywords: English for specific purposes, business English, internationalization of higher education, foreign language teaching

Procedia PDF Downloads 104
600 A User Interface for Easiest Way Image Encryption with Chaos

Authors: D. López-Mancilla, J. M. Roblero-Villa

Abstract:

Since 1990, the research on chaotic dynamics has received considerable attention, particularly in light of potential applications of this phenomenon in secure communications. Data encryption using chaotic systems was reported in the 90's as a new approach for signal encoding that differs from the conventional methods that use numerical algorithms as the encryption key. The algorithms for image encryption have received a lot of attention because of the need to find security on image transmission in real time over the internet and wireless networks. Known algorithms for image encryption, like the standard of data encryption (DES), have the drawback of low level of efficiency when the image is large. The encrypting based on chaos proposes a new and efficient way to get a fast and highly secure image encryption. In this work, a user interface for image encryption and a novel and easiest way to encrypt images using chaos are presented. The main idea is to reshape any image into a n-dimensional vector and combine it with vector extracted from a chaotic system, in such a way that the vector image can be hidden within the chaotic vector. Once this is done, an array is formed with the original dimensions of the image and turns again. An analysis of the security of encryption from the images using statistical analysis is made and is used a stage of optimization for image encryption security and, at the same time, the image can be accurately recovered. The user interface uses the algorithms designed for the encryption of images, allowing you to read an image from the hard drive or another external device. The user interface, encrypt the image allowing three modes of encryption. These modes are given by three different chaotic systems that the user can choose. Once encrypted image, is possible to observe the safety analysis and save it on the hard disk. The main results of this study show that this simple method of encryption, using the optimization stage, allows an encryption security, competitive with complicated encryption methods used in other works. In addition, the user interface allows encrypting image with chaos, and to submit it through any public communication channel, including internet.

Keywords: image encryption, chaos, secure communications, user interface

Procedia PDF Downloads 481
599 Working in Multidisciplinary Care Teams: Perspectives from Health Care and Social Service Providers

Authors: Lindy Van Vliet, Saloni Phadke, Anthea Nelson, Ann Gallant

Abstract:

Holistic and patient-centred palliative care and support require an integrated system of care that includes health and social service providers working together to ensure that patients and families have access to the care they need. The objective of this study is to further explore and understand the benefits and challenges of mobilizing multidisciplinary care teams for health care professionals and social service providers. Drawing on an interpretivist, exploratory, qualitative design, our multidisciplinary research team (medicine, nursing and social work) conducted interviews with 15 health care and social service providers in the Ottawa region. Interview data was audio-recorded, transcribed, and analyzed using a reflexive thematic analysis approach. The data deepens our understandings of the facilitators and barriers posed by multidisciplinary care teams. Three main findings emerged: First, the data highlighted the benefits of multidisciplinary care teams for both patient outcomes and quality of life and provider mental health; second, the data showed that the lack of a system-wide integrated communication system reduces the quality of patient care and increases provider stress while working in multidisciplinary care teams; finally, the data demonstrated the existence of implicit hierarchies between disciplines, this coupled with different disciplinary perspectives of palliative care provision can lead to friction and challenges within care teams. These findings will have important implications for the future of palliative care as they will help to facilitate and build stronger person-centred/relationship-centred palliative care practices by naming the challenges faced by multidisciplinary palliative care teams and providing examples of best practices.

Keywords: public health palliative care, palliative care nursing, care networks, integrated health care, palliative care approach, public health, multidisciplinary work, care teams

Procedia PDF Downloads 77
598 System Analysis of Quality Assurance in Online Education

Authors: Keh-Wen Carin Chuang, Kuan-Chou Chen

Abstract:

Our society is in a constant state of change. Technology advancements continue to affect our daily lives. How we work, communicate and entertain ourselves has changed dramatically in the past decades. As our society learns to accept and adapt to the many different technological advances that seem to inundate every part of our lives, the education institutions must migrate from traditional methods of instruction to online education in order to take full advantage of the opportunities provided by these technology advancements. There are many benefits that can be gained for university and society from offering online programs by utilizing advanced technologies. But the programs must not be implemented carelessly. The key to providing a quality online program is the issue of perceived quality, which takes into account the viewpoint of all stakeholders involved. To truly ensure the institutional quality, however, a systemic view of all factors contributing to the quality must be analyzed and linked to one another — allowing education administrators to understand how each factor contributes to the perceived quality of online education. The perceived quality of an online program will be positively reinforced only through an organizational-wide effort that focuses on managed administration, augmenting online program branding, skilled faculty, supportive alumni, student satisfaction, and effective delivery systems — each of which is vital to a quality online program. This study focuses on the concept of quality assurance in the start-up, implementation, and sustainability of online education. A case of online MBA program will be analyzed to explore the quality assurance. The difficulties in promoting online education quality is the fact that universities are complex networks of disciplinary, social, economic, and political fiefdoms, both internal and external factors to the institutions. As such, the system analysis, a systems-thinking approach, on the issue of perceived quality is ideal to investigate the factors and how each factor contributes to the perceived quality in the online education domain.

Keywords: systems thinking, quality assurance, online education, MBA program

Procedia PDF Downloads 231
597 Development of Automated Quality Management System for the Management of Heat Networks

Authors: Nigina Toktasynova, Sholpan Sagyndykova, Zhanat Kenzhebayeva, Maksat Kalimoldayev, Mariya Ishimova, Irbulat Utepbergenov

Abstract:

Any business needs a stable operation and continuous improvement, therefore it is necessary to constantly interact with the environment, to analyze the work of the enterprise in terms of employees, executives and consumers, as well as to correct any inconsistencies of certain types of processes and their aggregate. In the case of heat supply organizations, in addition to suppliers, local legislation must be considered which often is the main regulator of pricing of services. In this case, the process approach used to build a functional organizational structure in these types of businesses in Kazakhstan is a challenge not only in the implementation, but also in ways of analyzing the employee's salary. To solve these problems, we investigated the management system of heating enterprise, including strategic planning based on the balanced scorecard (BSC), quality management in accordance with the standards of the Quality Management System (QMS) ISO 9001 and analysis of the system based on expert judgment using fuzzy inference. To carry out our work we used the theory of fuzzy sets, the QMS in accordance with ISO 9001, BSC according to the method of Kaplan and Norton, method of construction of business processes according to the notation IDEF0, theory of modeling using Matlab software simulation tools and graphical programming LabVIEW. The results of the work are as follows: We determined possibilities of improving the management of heat-supply plant-based on QMS; after the justification and adaptation of software tool it has been used to automate a series of functions for the management and reduction of resources and for the maintenance of the system up to date; an application for the analysis of the QMS based on fuzzy inference has been created with novel organization of communication software with the application enabling the analysis of relevant data of enterprise management system.

Keywords: balanced scorecard, heat supply, quality management system, the theory of fuzzy sets

Procedia PDF Downloads 362
596 Problems and Solutions in the Application of ICP-MS for Analysis of Trace Elements in Various Samples

Authors: Béla Kovács, Éva Bódi, Farzaneh Garousi, Szilvia Várallyay, Áron Soós, Xénia Vágó, Dávid Andrási

Abstract:

In agriculture for analysis of elements in different food and food raw materials, moreover environmental samples generally flame atomic absorption spectrometers (FAAS), graphite furnace atomic absorption spectrometers (GF-AAS), inductively coupled plasma optical emission spectrometers (ICP-OES) and inductively coupled plasma mass spectrometers (ICP-MS) are routinely applied. An inductively coupled plasma mass spectrometer (ICP-MS) is capable for analysis of 70-80 elements in multielemental mode, from 1-5 cm3 volume of a sample, moreover the detection limits of elements are in µg/kg-ng/kg (ppb-ppt) concentration range. All the analytical instruments have different physical and chemical interfering effects analysing the above types of samples. The smaller the concentration of an analyte and the larger the concentration of the matrix the larger the interfering effects. Nowadays there is very important to analyse growingly smaller concentrations of elements. From the above analytical instruments generally the inductively coupled plasma mass spectrometer is capable of analysing the smallest concentration of elements. The applied ICP-MS instrument has Collision Cell Technology (CCT) also. Using CCT mode certain elements have better (smaller) detection limits with 1-3 magnitudes comparing to a normal ICP-MS analytical method. The CCT mode has better detection limits mainly for analysis of selenium, arsenic, germanium, vanadium and chromium. To elaborate an analytical method for trace elements with an inductively coupled plasma mass spectrometer the most important interfering effects (problems) were evaluated: 1) Physical interferences; 2) Spectral interferences (elemental and molecular isobaric); 3) Effect of easily ionisable elements; 4) Memory interferences. Analysing food and food raw materials, moreover environmental samples an other (new) interfering effect emerged in ICP-MS, namely the effect of various matrixes having different evaporation and nebulization effectiveness, moreover having different quantity of carbon content of food and food raw materials, moreover environmental samples. In our research work the effect of different water-soluble compounds furthermore the effect of various quantity of carbon content (as sample matrix) were examined on changes of intensity of the applied elements. So finally we could find “opportunities” to decrease or eliminate the error of the analyses of applied elements (Cr, Co, Ni, Cu, Zn, Ge, As, Se, Mo, Cd, Sn, Sb, Te, Hg, Pb, Bi). To analyse these elements in the above samples, the most appropriate inductively coupled plasma mass spectrometer is a quadrupole instrument applying a collision cell technique (CCT). The extent of interfering effect of carbon content depends on the type of compounds. The carbon content significantly affects the measured concentration (intensities) of the above elements, which can be corrected using different internal standards.

Keywords: elements, environmental and food samples, ICP-MS, interference effects

Procedia PDF Downloads 495
595 AI Peer Review Challenge: Standard Model of Physics vs 4D GEM EOS

Authors: David A. Harness

Abstract:

Natural evolution of ATP cognitive systems is to meet AI peer review standards. ATP process of axiom selection from Mizar to prove a conjecture would be further refined, as in all human and machine learning, by solving the real world problem of the proposed AI peer review challenge: Determine which conjecture forms the higher confidence level constructive proof between Standard Model of Physics SU(n) lattice gauge group operation vs. present non-standard 4D GEM EOS SU(n) lattice gauge group spatially extended operation in which the photon and electron are the first two trace angular momentum invariants of a gravitoelectromagnetic (GEM) energy momentum density tensor wavetrain integration spin-stress pressure-volume equation of state (EOS), initiated via 32 lines of Mathematica code. Resulting gravitoelectromagnetic spectrum ranges from compressive through rarefactive of the central cosmological constant vacuum energy density in units of pascals. Said self-adjoint group operation exclusively operates on the stress energy momentum tensor of the Einstein field equations, introducing quantization directly on the 4D spacetime level, essentially reformulating the Yang-Mills virtual superpositioned particle compounded lattice gauge groups quantization of the vacuum—into a single hyper-complex multi-valued GEM U(1) × SU(1,3) lattice gauge group Planck spacetime mesh quantization of the vacuum. Thus the Mizar corpus already contains all of the axioms required for relevant DeepMath premise selection and unambiguous formal natural language parsing in context deep learning.

Keywords: automated theorem proving, constructive quantum field theory, information theory, neural networks

Procedia PDF Downloads 173
594 Hybridization of Manually Extracted and Convolutional Features for Classification of Chest X-Ray of COVID-19

Authors: M. Bilal Ishfaq, Adnan N. Qureshi

Abstract:

COVID-19 is the most infectious disease these days, it was first reported in Wuhan, the capital city of Hubei in China then it spread rapidly throughout the whole world. Later on 11 March 2020, the World Health Organisation (WHO) declared it a pandemic. Since COVID-19 is highly contagious, it has affected approximately 219M people worldwide and caused 4.55M deaths. It has brought the importance of accurate diagnosis of respiratory diseases such as pneumonia and COVID-19 to the forefront. In this paper, we propose a hybrid approach for the automated detection of COVID-19 using medical imaging. We have presented the hybridization of manually extracted and convolutional features. Our approach combines Haralick texture features and convolutional features extracted from chest X-rays and CT scans. We also employ a minimum redundancy maximum relevance (MRMR) feature selection algorithm to reduce computational complexity and enhance classification performance. The proposed model is evaluated on four publicly available datasets, including Chest X-ray Pneumonia, COVID-19 Pneumonia, COVID-19 CTMaster, and VinBig data. The results demonstrate high accuracy and effectiveness, with 0.9925 on the Chest X-ray pneumonia dataset, 0.9895 on the COVID-19, Pneumonia and Normal Chest X-ray dataset, 0.9806 on the Covid CTMaster dataset, and 0.9398 on the VinBig dataset. We further evaluate the effectiveness of the proposed model using ROC curves, where the AUC for the best-performing model reaches 0.96. Our proposed model provides a promising tool for the early detection and accurate diagnosis of COVID-19, which can assist healthcare professionals in making informed treatment decisions and improving patient outcomes. The results of the proposed model are quite plausible and the system can be deployed in a clinical or research setting to assist in the diagnosis of COVID-19.

Keywords: COVID-19, feature engineering, artificial neural networks, radiology images

Procedia PDF Downloads 72
593 Automatic Lexicon Generation for Domain Specific Dataset for Mining Public Opinion on China Pakistan Economic Corridor

Authors: Tayyaba Azim, Bibi Amina

Abstract:

The increase in the popularity of opinion mining with the rapid growth in the availability of social networks has attracted a lot of opportunities for research in the various domains of Sentiment Analysis and Natural Language Processing (NLP) using Artificial Intelligence approaches. The latest trend allows the public to actively use the internet for analyzing an individual’s opinion and explore the effectiveness of published facts. The main theme of this research is to account the public opinion on the most crucial and extensively discussed development projects, China Pakistan Economic Corridor (CPEC), considered as a game changer due to its promise of bringing economic prosperity to the region. So far, to the best of our knowledge, the theme of CPEC has not been analyzed for sentiment determination through the ML approach. This research aims to demonstrate the use of ML approaches to spontaneously analyze the public sentiment on Twitter tweets particularly about CPEC. Support Vector Machine SVM is used for classification task classifying tweets into positive, negative and neutral classes. Word2vec and TF-IDF features are used with the SVM model, a comparison of the trained model on manually labelled tweets and automatically generated lexicon is performed. The contributions of this work are: Development of a sentiment analysis system for public tweets on CPEC subject, construction of an automatic generation of the lexicon of public tweets on CPEC, different themes are identified among tweets and sentiments are assigned to each theme. It is worth noting that the applications of web mining that empower e-democracy by improving political transparency and public participation in decision making via social media have not been explored and practised in Pakistan region on CPEC yet.

Keywords: machine learning, natural language processing, sentiment analysis, support vector machine, Word2vec

Procedia PDF Downloads 143
592 Linguistic Insights Improve Semantic Technology in Medical Research and Patient Self-Management Contexts

Authors: William Michael Short

Abstract:

Semantic Web’ technologies such as the Unified Medical Language System Metathesaurus, SNOMED-CT, and MeSH have been touted as transformational for the way users access online medical and health information, enabling both the automated analysis of natural-language data and the integration of heterogeneous healthrelated resources distributed across the Internet through the use of standardized terminologies that capture concepts and relationships between concepts that are expressed differently across datasets. However, the approaches that have so far characterized ‘semantic bioinformatics’ have not yet fulfilled the promise of the Semantic Web for medical and health information retrieval applications. This paper argues within the perspective of cognitive linguistics and cognitive anthropology that four features of human meaning-making must be taken into account before the potential of semantic technologies can be realized for this domain. First, many semantic technologies operate exclusively at the level of the word. However, texts convey meanings in ways beyond lexical semantics. For example, transitivity patterns (distributions of active or passive voice) and modality patterns (configurations of modal constituents like may, might, could, would, should) convey experiential and epistemic meanings that are not captured by single words. Language users also naturally associate stretches of text with discrete meanings, so that whole sentences can be ascribed senses similar to the senses of words (so-called ‘discourse topics’). Second, natural language processing systems tend to operate according to the principle of ‘one token, one tag’. For instance, occurrences of the word sound must be disambiguated for part of speech: in context, is sound a noun or a verb or an adjective? In syntactic analysis, deterministic annotation methods may be acceptable. But because natural language utterances are typically characterized by polyvalency and ambiguities of all kinds (including intentional ambiguities), such methods leave the meanings of texts highly impoverished. Third, ontologies tend to be disconnected from everyday language use and so struggle in cases where single concepts are captured through complex lexicalizations that involve profile shifts or other embodied representations. More problematically, concept graphs tend to capture ‘expert’ technical models rather than ‘folk’ models of knowledge and so may not match users’ common-sense intuitions about the organization of concepts in prototypical structures rather than Aristotelian categories. Fourth, and finally, most ontologies do not recognize the pervasively figurative character of human language. However, since the time of Galen the widespread use of metaphor in the linguistic usage of both medical professionals and lay persons has been recognized. In particular, metaphor is a well-documented linguistic tool for communicating experiences of pain. Because semantic medical knowledge-bases are designed to help capture variations within technical vocabularies – rather than the kinds of conventionalized figurative semantics that practitioners as well as patients actually utilize in clinical description and diagnosis – they fail to capture this dimension of linguistic usage. The failure of semantic technologies in these respects degrades the efficiency and efficacy not only of medical research, where information retrieval inefficiencies can lead to direct financial costs to organizations, but also of care provision, especially in contexts of patients’ self-management of complex medical conditions.

Keywords: ambiguity, bioinformatics, language, meaning, metaphor, ontology, semantic web, semantics

Procedia PDF Downloads 126
591 Synergistic Behavior of Polymer Mixtures in Designing Hydrogels for Biomedical Applications

Authors: Maria Bercea, Monica Diana Olteanu

Abstract:

Investigation of polymer systems able to change inside of the body into networks represent an attractive approach, especially when there is a minimally invasive and patient friendly administration. Pharmaceutical formulations based on Pluronic F127 [poly (oxyethylene) (PEO) blocks (70%) and poly(oxypropylene) (PPO) blocks (30%)] present an excellent potential as drug delivery systems. The use of Pluronic F127 alone as gel-forming solution is limited by some characteristics, such as poor mechanical properties, short residence time, high permeability, etc. Investigation of the interactions between the natural and synthetic polymers and surfactants in solution is a subject of great interest from both scientific and practical point of view. As for example, formulations based on Pluronics and chitosan could be used to obtain dual phase transition hydrogels responsive to temperature and pH changes. In this study, different materials were prepared by using poly(vinyl alcohol), chitosan solutions mixed with aqueous solutions of Pluronic F127. The rheological properties of different formulations were investigated in temperature sweep experiments as well as at a constant temperature of 37oC for exploring in-situ gel formation in the human body conditions. In addition, some viscometric investigations were carried out in order to understand the interactions which determine the complex behaviour of these systems. Correlation between the thermodynamic and rheological parameters and phase separation phenomena observed for the investigated systems allowed the dissemination the constitutive response of polymeric materials at different external stimuli, such as temperature and pH. The rheological investigation demonstrated that the viscoelastic moduli of the hydrogels can be tuned depending on concentration of different components as well as pH and temperature conditions and cumulative contributions can be obtained.

Keywords: hydrogel, polymer mixture, stimuli responsive, biomedical applications

Procedia PDF Downloads 343
590 Health and the Politics of Trust: Multi-Drug-Resistant Tuberculosis in Kathmandu

Authors: Mattia Testuzza

Abstract:

Public health is a social endeavour, which involves many different actors: from extremely stratified, structured health systems to unofficial networks of people and knowledge. Health and diseases are an intertwined individual and social experiences. Both patients and health workers navigate this public space through relations of trust. Trust in healthcare goes from the personal trust between a patient and her/his doctor to the trust of both the patient and the health worker in the medical knowledge and the healthcare system. Trust it is not a given, but it is continuously negotiated, given and gained. The key to understand these essential relations of trust in health is to recognise them as a social practice, which therefore implies agency and power. In these terms, health is constantly public and made public, as trust emerges as a meaningfully political phenomenon. Trust as a power relation can be observed at play in the implementation of public health policies such as the WHO’s Directly-Observed Theraphy Short-course (DOTS), and with the increasing concern for drug-resistance that tuberculosis pose, looking at the role of trust in the healthcare delivery system and implementation of public health policies becomes significantly relevant. The ethnographic fieldwork was carried out in four months through observation of the daily practices at the National Tuberculosis Center of Nepal, and semi-structured interviews with MultiDrug-Resistant Tuberculosis (MDR-TB) patients at different stages of the treatment, their relatives, MDR-TB specialised nurses, and doctors. Throughout the research, the role which trust plays in tuberculosis treatment emerged as one fundamental ax that cuts through all the different factors intertwined with drug-resistance development, unfolding a tension between the DOTS policy, which undermines trust, and the day-to-day healthcare relations and practices which cannot function without trust. Trust also stands out as a key component of the solutions to unforeseen issues which develop from the overall uncertainty of the context - for example, political instability and extreme poverty - in which tuberculosis treatment is carried out in Nepal.

Keywords: trust, tuberculosis, drug-resistance, politics of health

Procedia PDF Downloads 245
589 A Methodology Based on Image Processing and Deep Learning for Automatic Characterization of Graphene Oxide

Authors: Rafael do Amaral Teodoro, Leandro Augusto da Silva

Abstract:

Originated from graphite, graphene is a two-dimensional (2D) material that promises to revolutionize technology in many different areas, such as energy, telecommunications, civil construction, aviation, textile, and medicine. This is possible because its structure, formed by carbon bonds, provides desirable optical, thermal, and mechanical characteristics that are interesting to multiple areas of the market. Thus, several research and development centers are studying different manufacturing methods and material applications of graphene, which are often compromised by the scarcity of more agile and accurate methodologies to characterize the material – that is to determine its composition, shape, size, and the number of layers and crystals. To engage in this search, this study proposes a computational methodology that applies deep learning to identify graphene oxide crystals in order to characterize samples by crystal sizes. To achieve this, a fully convolutional neural network called U-net has been trained to segment SEM graphene oxide images. The segmentation generated by the U-net is fine-tuned with a standard deviation technique by classes, which allows crystals to be distinguished with different labels through an object delimitation algorithm. As a next step, the characteristics of the position, area, perimeter, and lateral measures of each detected crystal are extracted from the images. This information generates a database with the dimensions of the crystals that compose the samples. Finally, graphs are automatically created showing the frequency distributions by area size and perimeter of the crystals. This methodological process resulted in a high capacity of segmentation of graphene oxide crystals, presenting accuracy and F-score equal to 95% and 94%, respectively, over the test set. Such performance demonstrates a high generalization capacity of the method in crystal segmentation, since its performance considers significant changes in image extraction quality. The measurement of non-overlapping crystals presented an average error of 6% for the different measurement metrics, thus suggesting that the model provides a high-performance measurement for non-overlapping segmentations. For overlapping crystals, however, a limitation of the model was identified. To overcome this limitation, it is important to ensure that the samples to be analyzed are properly prepared. This will minimize crystal overlap in the SEM image acquisition and guarantee a lower error in the measurements without greater efforts for data handling. All in all, the method developed is a time optimizer with a high measurement value, considering that it is capable of measuring hundreds of graphene oxide crystals in seconds, saving weeks of manual work.

Keywords: characterization, graphene oxide, nanomaterials, U-net, deep learning

Procedia PDF Downloads 157
588 Assessment of Multi-Domain Energy Systems Modelling Methods

Authors: M. Stewart, Ameer Al-Khaykan, J. M. Counsell

Abstract:

Emissions are a consequence of electricity generation. A major option for low carbon generation, local energy systems featuring Combined Heat and Power with solar PV (CHPV) has significant potential to increase energy performance, increase resilience, and offer greater control of local energy prices while complementing the UK’s emissions standards and targets. Recent advances in dynamic modelling and simulation of buildings and clusters of buildings using the IDEAS framework have successfully validated a novel multi-vector (simultaneous control of both heat and electricity) approach to integrating the wide range of primary and secondary plant typical of local energy systems designs including CHP, solar PV, gas boilers, absorption chillers and thermal energy storage, and associated electrical and hot water networks, all operating under a single unified control strategy. Results from this work indicate through simulation that integrated control of thermal storage can have a pivotal role in optimizing system performance well beyond the present expectations. Environmental impact analysis and reporting of all energy systems including CHPV LES presently employ a static annual average carbon emissions intensity for grid supplied electricity. This paper focuses on establishing and validating CHPV environmental performance against conventional emissions values and assessment benchmarks to analyze emissions performance without and with an active thermal store in a notional group of non-domestic buildings. Results of this analysis are presented and discussed in context of performance validation and quantifying the reduced environmental impact of CHPV systems with active energy storage in comparison with conventional LES designs.

Keywords: CHPV, thermal storage, control, dynamic simulation

Procedia PDF Downloads 234
587 Access to Livelihoods for Urban Refugees in Kenya: The Case Study of Somalis Living in Eastleigh

Authors: Nancy Njoka, Manuela Ramos Cacciatore

Abstract:

In Kenya, refugee situations are becoming increasingly protracted, stretching over the years or even decades. As urbanization rates increase, so do the numbers of urban refugees in the country. Refugees living in urban areas face a range of challenges. In their efforts to pursue livelihoods, refugees have identified strategies to confront these challenges. In the same manner, humanitarian actors have come up with different interventions to promote access to livelihoods working through obstacles and barriers created by host governments. This paper seeks to understand the experience of Somali urban refugees living in the urban area of Eastleigh, Nairobi, both by investigating their own actions towards creating avenues to access livelihoods and by understanding their social, economic and policy context in which they forge livelihoods. The empirical data collected through fieldwork in Nairobi in 2020 serves as the basis of this qualitative case study. Drawing upon the themes of urban refugee movement, Somali ethnicity, citizenship discrimination and the livelihoods of refugees, the paper highlights how the actions of the Kenyan government and international non-governmental organization (INGO)s affect access to livelihoods and the consequences of these actions for Somali urban refugees. The results of the paper found that Somali urban refugees are taking active steps to create livelihoods for themselves. This is seen in the growth of Eastleigh as an economic hub in Kenya which is owned and run mostly by Somalis. Indeed, the Somali community is central to the establishment of networks in the neighborhood. Somali urban refugees are marginalized by the Kenyan government, reducing their opportunity to create dignified lives in Eastleigh. Findings also point out the community-based approaches used by INGOs in livelihood interventions. The relevance of this research lies in the interconnection of humanitarian development interventions for protracted refugees and the promotion of livelihoods in an urban and global context.

Keywords: Kenya, livelihoods, Somali, urban refugees

Procedia PDF Downloads 172
586 Tourism Development and Its Role in the Urban Expansion of Al-Khomse City, Libya

Authors: Khaled Klib, Yousri Azzam, Ibrahim Maarouf

Abstract:

Tourism is one of the most important and fastest growing economic activities in the world, which has a prominent role in the growth and development of countries and has become increasingly important as business and trade after the World War II. The tourism development is one of the most important aspects of urban development, which aims to plan and develop tourist attractions and improve the urban environment within cities. Tourism development has become a priority for the urban development policy of cities, particularly those which have many tourist potentials. Complementary services, such as infrastructure, roads’ networks, transportation, and communications are needed for these potentials to function properly. In order to achieve these functionalities, also a new planning for the new areas as an expansion is required, or developing and renovating the existing urban areas according to pre-prepared plans to avoid random expansion of the urban structure of the city. This paper aims to determine the tourist attractions of Al-Khomse city, by reviewing the most important tourist attractions such as the Roman city (Leptis Magna), the geographical location on the Mediterranean coast, the temperate climate and diversity of the natural environment. The paper also examines the reality of the infrastructure and tourist services in the city and its suitability to serve the tourism sector. The paper also includes a proposed for tourism development in the city as one of the city's urban expansion trends, which can guide the development strategy in the future. The paper concludes with a vision for the tourism development areas as one of the trends for urban expansion in the future. The paper also concludes tourism development will have an effective role in the growth and development of urban, economic and social, in addition to preserving the natural environment. The paper recommended the need to emphasize the role of tourism development as one of the pillars and trends for the development policy and expansion of Al-Khomse city, preservation of tourist attractions and natural resources and developing infrastructure and tourist services such as accommodation, entertainment, mobility, and accessibility.

Keywords: tourism, tourist attractions, tourism development, urban expansion

Procedia PDF Downloads 239
585 Alternative Ways to Measure Impacts of Dam Closure to the Structure of Fish Communities of a Neotropical River

Authors: Ana Carolina Lima, Carlos Sérgio Agostinho, Amadeu M. V. M. Soares, Kieran A. Monaghan

Abstract:

Neotropical freshwaters host some of the most biodiverse ecosystems in the world and are among the most threatened by habitat alterations. The high number of species and lack of basic ecological knowledge provides a major obstacle to understanding the effects of environmental change. We assessed the impact of dam closure on the fish communities of a neotropical river by applying simple descriptions of community organizations: Species Abundance Distribution (SAD) and Abundance Biomass Comparison (ABC) curves. Fish data were collected during three distinct time periods (one year before, one year after and five years after closure), at eight sites located downstream of the dam, in the reservoir and reservoir transition zone and upstream of the regulated flow. Dam closure was associated with changes in the structural and functional organization of fish communities at all sites. Species richness tended to increase immediately after dam closure while evenness decreased. Changes in taxonomic structure were accompanied by a change in the distribution of biomass with the proportionate contribution by smaller individuals significantly increased relative to larger individuals. Five years on, richness had fallen to below pre-closure levels at all sites, while the comparative stability of the transformed habitats was reflected by biomass-abundance distribution patterns that approximated pre-disturbance ratios. Despite initial generality, respective sites demonstrated distinct ecological responses that were related to the environmental characteristics of their transformed habitats. This simplistic analysis provides a sensitive and informative assessment of ecological conditions that highlights the impact to ecosystem process and ecological networks and has particular value in regions where detailed ecological knowledge precludes the application of traditional bioassessment methods.

Keywords: ABC curves, SADs, biodiversity, damming, tropical fish

Procedia PDF Downloads 381
584 Digitally Mapping Aboriginal Journey Ways

Authors: Paul Longley Arthur

Abstract:

This paper reports on an Australian Research Council-funded project utilising the Australian digital research infrastructure the ‘Time-Layered Cultural Map of Australia’ (TLCMap) (https://www.tlcmap.org/) [1]. This resource has been developed to help researchers create digital maps from cultural, textual, and historical data, layered with datasets registered on the platform. TLCMap is a set of online tools that allows humanities researchers to compile humanities data using spatio-temporal coordinates – to upload, gather, analyse and visualise data. It is the only purpose-designed, Australian-developed research tool for humanities and social science researchers to identify geographical clusters and parallel journeys by sight. This presentation discusses a series of Aboriginal mapping and visualisation experiments using TLCMap to show how Indigenous knowledge can reconfigure contemporary understandings of space including the urbanised landscape [2, 3]. The research data being generated – investigating the historical movements of Aboriginal people, the distribution of networks, and their relation to land – lends itself to mapping and geo-spatial visualisation and analysis. TLCMap allows researchers to create layers on a 3D map which pinpoint locations with accompanying information, and this has enabled our research team to plot out traditional historical journeys undertaken by Aboriginal people as well as to compile a gazetteer of Aboriginal place names, many of which have largely been undocumented until now [4]. The documented journeys intersect with and overlay many of today’s urban formations including main roads, municipal boundaries, and state borders. The paper questions how such data can be incorporated into a more culturally and ethically responsive understanding of contemporary urban spaces and as well as natural environments [5].

Keywords: spatio-temporal mapping, visualisation, Indigenous knowledge, mobility and migration, research infrastructure

Procedia PDF Downloads 9
583 Application of Artificial Neural Network for Single Horizontal Bare Tube and Bare Tube Bundles (Staggered) of Large Particles: Heat Transfer Prediction

Authors: G. Ravindranath, S. Savitha

Abstract:

This paper presents heat transfer analysis of single horizontal bare tube and heat transfer analysis of staggered arrangement of bare tube bundles bare tube bundles in gas-solid (air-solid) fluidized bed and predictions are done by using Artificial Neural Network (ANN) based on experimental data. Fluidized bed provide nearly isothermal environment with high heat transfer rate to submerged objects i.e. due to through mixing and large contact area between the gas and the particle, a fully fluidized bed has little temperature variation and gas leaves at a temperature which is close to that of the bed. Measurement of average heat transfer coefficient was made by local thermal simulation technique in a cold bubbling air-fluidized bed of size 0.305 m. x 0.305 m. Studies were conducted for single horizontal Bare Tube of length 305mm and 28.6mm outer diameter and for bare tube bundles of staggered arrangement using beds of large (average particle diameter greater than 1 mm) particle (raagi and mustard). Within the range of experimental conditions influence of bed particle diameter ( Dp), Fluidizing Velocity (U) were studied, which are significant parameters affecting heat transfer. Artificial Neural Networks (ANNs) have been receiving an increasing attention for simulating engineering systems due to some interesting characteristics such as learning capability, fault tolerance, and non-linearity. Here, feed-forward architecture and trained by back-propagation technique is adopted to predict heat transfer analysis found from experimental results. The ANN is designed to suit the present system which has 3 inputs and 2 out puts. The network predictions are found to be in very good agreement with the experimental observed values of bare heat transfer coefficient (hb) and nusselt number of bare tube (Nub).

Keywords: fluidized bed, large particles, particle diameter, ANN

Procedia PDF Downloads 360
582 Locating Potential Site for Biomass Power Plant Development in Central Luzon Philippines Using GIS-Based Suitability Analysis

Authors: Bryan M. Baltazar, Marjorie V. Remolador, Klathea H. Sevilla, Imee Saladaga, Loureal Camille Inocencio, Ma. Rosario Concepcion O. Ang

Abstract:

Biomass energy is a traditional source of sustainable energy, which has been widely used in developing countries. The Philippines, specifically Central Luzon, has an abundant source of biomass. Hence, it could supply abundant agricultural residues (rice husks), as feedstock in a biomass power plant. However, locating a potential site for biomass development is a complex process which involves different factors, such as physical, environmental, socio-economic, and risks that are usually diverse and conflicting. Moreover, biomass distribution is highly dispersed geographically. Thus, this study develops an integrated method combining Geographical Information Systems (GIS) and methods for energy planning; Multi-Criteria Decision Analysis (MCDA) and Analytical Hierarchy Process (AHP), for locating suitable site for biomass power plant development in Central Luzon, Philippines by considering different constraints and factors. Using MCDA, a three level hierarchy of factors and constraints was produced, with corresponding weights determined by experts by using AHP. Applying the results, a suitability map for Biomass power plant development in Central Luzon was generated. It showed that the central part of the region has the highest potential for biomass power plant development. It is because of the characteristics of the area such as the abundance of rice fields, with generally flat land surfaces, accessible roads and grid networks, and low risks to flooding and landslide. This study recommends the use of higher accuracy resource maps, and further analysis in selecting the optimum site for biomass power plant development that would account for the cost and transportation of biomass residues.

Keywords: analytic hierarchy process, biomass energy, GIS, multi-criteria decision analysis, site suitability analysis

Procedia PDF Downloads 418