Search results for: artificial insemination
44 Impact of Climate Change on Sea Level Rise along the Coastline of Mumbai City, India
Authors: Chakraborty Sudipta, A. R. Kambekar, Sarma Arnab
Abstract:
Sea-level rise being one of the most important impacts of anthropogenic induced climate change resulting from global warming and melting of icebergs at Arctic and Antarctic, the investigations done by various researchers both on Indian Coast and elsewhere during the last decade has been reviewed in this paper. The paper aims to ascertain the propensity of consistency of different suggested methods to predict the near-accurate future sea level rise along the coast of Mumbai. Case studies at East Coast, Southern Tip and West and South West coast of India have been reviewed. Coastal Vulnerability Index of several important international places has been compared, which matched with Intergovernmental Panel on Climate Change forecasts. The application of Geographic Information System mapping, use of remote sensing technology, both Multi Spectral Scanner and Thematic Mapping data from Landsat classified through Iterative Self-Organizing Data Analysis Technique for arriving at high, moderate and low Coastal Vulnerability Index at various important coastal cities have been observed. Instead of data driven, hindcast based forecast for Significant Wave Height, additional impact of sea level rise has been suggested. Efficacy and limitations of numerical methods vis-à-vis Artificial Neural Network has been assessed, importance of Root Mean Square error on numerical results is mentioned. Comparing between various computerized methods on forecast results obtained from MIKE 21 has been opined to be more reliable than Delft 3D model.
Keywords: Climate change, coastal vulnerability index, global warming, sea level rise.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 156543 The Potential Use of Nanofilters to Supply Potable Water in Persian Gulf and Oman Sea Watershed Basin
Authors: Sara Zamani, Mojtaba Fazeli, Abdollah Rashidi Mehrabadi
Abstract:
In a world worried about water resources with the shadow of drought and famine looming all around, the quality of water is as important as its quantity. The source of all concerns is the constant reduction of per capita quality water for different uses. Iran With an average annual precipitation of 250 mm compared to the 800 mm world average, Iran is considered a water scarce country and the disparity in the rainfall distribution, the limitations of renewable resources and the population concentration in the margins of desert and water scarce areas have intensified the problem. The shortage of per capita renewable freshwater and its poor quality in large areas of the country, which have saline, brackish or hard water resources, and the profusion of natural and artificial pollutant have caused the deterioration of water quality. Among methods of treatment and use of these waters one can refer to the application of membrane technologies, which have come into focus in recent years due to their great advantages. This process is quite efficient in eliminating multi-capacity ions; and due to the possibilities of production at different capacities, application as treatment process in points of use, and the need for less energy in comparison to Reverse Osmosis processes, it can revolutionize the water and wastewater sector in years to come. The article studied the different capacities of water resources in the Persian Gulf and Oman Sea watershed basins, and processes the possibility of using nanofiltration process to treat brackish and non-conventional waters in these basins.Keywords: Membrane processes, saline waters, brackish waters, hard waters, zoning water quality in the Persian Gulf and the Oman Sea Watershed area, nanofiltration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 195342 Fully Automated Methods for the Detection and Segmentation of Mitochondria in Microscopy Images
Authors: Blessing Ojeme, Frederick Quinn, Russell Karls, Shannon Quinn
Abstract:
The detection and segmentation of mitochondria from fluorescence microscopy is crucial for understanding the complex structure of the nervous system. However, the constant fission and fusion of mitochondria and image distortion in the background make the task of detection and segmentation challenging. Although there exists a number of open-source software tools and artificial intelligence (AI) methods designed for analyzing mitochondrial images, the availability of only a few combined expertise in the medical field and AI required to utilize these tools poses a challenge to its full adoption and use in clinical settings. Motivated by the advantages of automated methods in terms of good performance, minimum detection time, ease of implementation, and cross-platform compactibility, this study proposes a fully automated framework for the detection and segmentation of mitochondria using both image shape information and descriptive statistics. Using the low-cost, open-source Python and OpenCV library, the algorithms are implemented in three stages: pre-processing; image binarization; and coarse-to-fine segmentation. The proposed model is validated using the fluorescence mitochondrial dataset. Ground truth labels generated using Labkit were also used to evaluate the performance of our detection and segmentation model using precision, recall and rand index. The study produces good detection and segmentation results and reports the challenges encountered during the image analysis of mitochondrial morphology from the fluorescence mitochondrial dataset. A discussion on the methods and future perspectives of fully automated frameworks concludes the paper.
Keywords: 2D, Binarization, CLAHE, detection, fluorescence microscopy, mitochondria, segmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 46541 Quality Classification and Monitoring Using Adaptive Metric Distance and Neural Networks: Application in Pickling Process
Authors: S. Bouhouche, M. Lahreche, S. Ziani, J. Bast
Abstract:
Modern manufacturing facilities are large scale, highly complex, and operate with large number of variables under closed loop control. Early and accurate fault detection and diagnosis for these plants can minimise down time, increase the safety of plant operations, and reduce manufacturing costs. Fault detection and isolation is more complex particularly in the case of the faulty analog control systems. Analog control systems are not equipped with monitoring function where the process parameters are continually visualised. In this situation, It is very difficult to find the relationship between the fault importance and its consequences on the product failure. We consider in this paper an approach to fault detection and analysis of its effect on the production quality using an adaptive centring and scaling in the pickling process in cold rolling. The fault appeared on one of the power unit driving a rotary machine, this machine can not track a reference speed given by another machine. The length of metal loop is then in continuous oscillation, this affects the product quality. Using a computerised data acquisition system, the main machine parameters have been monitored. The fault has been detected and isolated on basis of analysis of monitored data. Normal and faulty situation have been obtained by an artificial neural network (ANN) model which is implemented to simulate the normal and faulty status of rotary machine. Correlation between the product quality defined by an index and the residual is used to quality classification.Keywords: Modeling, fault detection and diagnosis, parameters estimation, neural networks, Fault Detection and Diagnosis (FDD), pickling process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 157640 Reducing the Imbalance Penalty through Artificial Intelligence Methods Geothermal Production Forecasting: A Case Study for Turkey
Abstract:
In addition to being rich in renewable energy resources, Turkey is one of the countries that promise potential in geothermal energy production with its high installed power, cheapness, and sustainability. Increasing imbalance penalties become an economic burden for organizations, since the geothermal generation plants cannot maintain the balance of supply and demand due to the inadequacy of the production forecasts given in the day-ahead market. A better production forecast reduces the imbalance penalties of market participants and provides a better imbalance in the day ahead market. In this study, using machine learning, deep learning and time series methods, the total generation of the power plants belonging to Zorlu Doğal Electricity Generation, which has a high installed capacity in terms of geothermal, was predicted for the first one-week and first two-weeks of March, then the imbalance penalties were calculated with these estimates and compared with the real values. These modeling operations were carried out on two datasets, the basic dataset and the dataset created by extracting new features from this dataset with the feature engineering method. According to the results, Support Vector Regression from traditional machine learning models outperformed other models and exhibited the best performance. In addition, the estimation results in the feature engineering dataset showed lower error rates than the basic dataset. It has been concluded that the estimated imbalance penalty calculated for the selected organization is lower than the actual imbalance penalty, optimum and profitable accounts.
Keywords: Machine learning, deep learning, time series models, feature engineering, geothermal energy production forecasting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20439 Preparation and Characterization of Pure PVA and PVA/MMT Matrix: Effect of Thermal Treatment
Authors: Albana Hasimi, Edlira Tako, Partizan Malkaj, Elvin Çomo, Blerina Papajani, Mirela Ndrita, Ledjan Malaj
Abstract:
Many endeavors have been exerted during the last years for developing new artificial polymeric membranes, which fulfill the demanded conditions for biomedical uses. One of the most tested polymers is Poly(vinyl alcohol) [PVA]. Our teams are based on the possibility of using PVA for personal protective equipment against COVID-19. In personal protective equipment, we explore the possibility of modifying the properties of the polymer by adding Montmorillonite [MMT]. Heat-treatment above the glass transition temperature is used to improve mechanical properties mainly by increasing the crystallinity of the polymer, which acts as a physical network. Temperature-Modulated Differential Scanning Calorimetry (TMDSC) measurements indicated that the presence of 0.5% MMT in PVA causes a higher Tg value and shaped peak of crystallinity. Decomposition is observed at two of the melting points of the crystals during heating 25-240 oC and overlap of the recrystallization ridges during cooling 240-25 oC. This is indicative of the presence of two types (quality or structure) of polymer crystals. On the other hand, some indication of improvement of the quality of the crystals by heat-treatment is given by the distinct non-reversing contribution to melting. Data on sorption and transport of water in PVA films: PVA pure and PVA/MMT matrix, modified by thermal treatment are presented. The membranes become more rigid as a result of the heat treatment and because of this the water uptake is significantly lower in membranes. That is indicated by analysis of the resulting water uptake kinetics. The presence of 0.5% w/w of MMT has no significant impact on the properties of PVA membranes. Water uptake kinetics deviate from Fick’s law due to slow relaxation of glassy polymer matrix for all types of membranes.
Keywords: Crystallinity, montmorillonite, nanocomposite, poly(vinyl alcohol).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22538 Non-Invasive Data Extraction from Machine Display Units Using Video Analytics
Authors: Ravneet Kaur, Joydeep Acharya, Sudhanshu Gaur
Abstract:
Artificial Intelligence (AI) has the potential to transform manufacturing by improving shop floor processes such as production, maintenance and quality. However, industrial datasets are notoriously difficult to extract in a real-time, streaming fashion thus, negating potential AI benefits. The main example is some specialized industrial controllers that are operated by custom software which complicates the process of connecting them to an Information Technology (IT) based data acquisition network. Security concerns may also limit direct physical access to these controllers for data acquisition. To connect the Operational Technology (OT) data stored in these controllers to an AI application in a secure, reliable and available way, we propose a novel Industrial IoT (IIoT) solution in this paper. In this solution, we demonstrate how video cameras can be installed in a factory shop floor to continuously obtain images of the controller HMIs. We propose image pre-processing to segment the HMI into regions of streaming data and regions of fixed meta-data. We then evaluate the performance of multiple Optical Character Recognition (OCR) technologies such as Tesseract and Google vision to recognize the streaming data and test it for typical factory HMIs and realistic lighting conditions. Finally, we use the meta-data to match the OCR output with the temporal, domain-dependent context of the data to improve the accuracy of the output. Our IIoT solution enables reliable and efficient data extraction which will improve the performance of subsequent AI applications.Keywords: Human machine interface, industrial internet of things, internet of things, optical character recognition, video analytic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 73937 Copper Price Prediction Model for Various Economic Situations
Authors: Haidy S. Ghali, Engy Serag, A. Samer Ezeldin
Abstract:
Copper is an essential raw material used in the construction industry. During 2021 and the first half of 2022, the global market suffered from a significant fluctuation in copper raw material prices due to the aftermath of both the COVID-19 pandemic and the Russia-Ukraine war which exposed its consumers to an unexpected financial risk. Thereto, this paper aims to develop two hybrid price prediction models using artificial neural network and long short-term memory (ANN-LSTM), by Python, that can forecast the average monthly copper prices, traded in the London Metal Exchange; the first model is a multivariate model that forecasts the copper price of the next 1-month and the second is a univariate model that predicts the copper prices of the upcoming three months. Historical data of average monthly London Metal Exchange copper prices are collected from January 2009 till July 2022 and potential external factors are identified and employed in the multivariate model. These factors lie under three main categories: energy prices, and economic indicators of the three major exporting countries of copper depending on the data availability. Before developing the LSTM models, the collected external parameters are analyzed with respect to the copper prices using correlation, and multicollinearity tests in R software; then, the parameters are further screened to select the parameters that influence the copper prices. Then, the two LSTM models are developed, and the dataset is divided into training, validation, and testing sets. The results show that the performance of the 3-month prediction model is better than the 1-month prediction model; but still, both models can act as predicting tools for diverse economic situations.
Keywords: Copper prices, prediction model, neural network, time series forecasting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18736 MAGNI Dynamics: A Vision-Based Kinematic and Dynamic Upper-Limb Model for Intelligent Robotic Rehabilitation
Authors: Alexandros Lioulemes, Michail Theofanidis, Varun Kanal, Konstantinos Tsiakas, Maher Abujelala, Chris Collander, William B. Townsend, Angie Boisselle, Fillia Makedon
Abstract:
This paper presents a home-based robot-rehabilitation instrument, called ”MAGNI Dynamics”, that utilized a vision-based kinematic/dynamic module and an adaptive haptic feedback controller. The system is expected to provide personalized rehabilitation by adjusting its resistive and supportive behavior according to a fuzzy intelligence controller that acts as an inference system, which correlates the user’s performance to different stiffness factors. The vision module uses the Kinect’s skeletal tracking to monitor the user’s effort in an unobtrusive and safe way, by estimating the torque that affects the user’s arm. The system’s torque estimations are justified by capturing electromyographic data from primitive hand motions (Shoulder Abduction and Shoulder Forward Flexion). Moreover, we present and analyze how the Barrett WAM generates a force-field with a haptic controller to support or challenge the users. Experiments show that by shifting the proportional value, that corresponds to different stiffness factors of the haptic path, can potentially help the user to improve his/her motor skills. Finally, potential areas for future research are discussed, that address how a rehabilitation robotic framework may include multisensing data, to improve the user’s recovery process.Keywords: Human-robot interaction, kinect, kinematics, dynamics, haptic control, rehabilitation robotics, artificial intelligence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 131935 Influence of Cyperus rotundus Active Principles Inhibit Viral Multiplication and Stimulate Immune System in Indian White Shrimp Fenneropenaeus indicus against White Spot Syndrome Virus Infection
Authors: T. Citarasu, M. Michaelbabu V. N. Vakharia
Abstract:
The rhizome of Java grass, Cyperus rotundus was extracted different organic polar and non-polar solvents and performed the in vitro antiviral and immunostimulant activities against White Spot Syndrome Virus (WSSV) and Vibrio harveyi respectively. Based on the initial screening the ethyl acetate extract of C. rotundus was strong activities and further it was purified through silica column chromatography and the fractions were screened again for antiviral and immunostimulant activity. Among the different fractions screened against the WSSV and V. harveyi, the fractions, FIII to FV had strong activities. In order to study the in vivo influence of C. rotundus, the fractions (F-III to FV) were pooled and delivered to the F. indicus through artificial feed for 30 days. After the feeding trail the experimental and control diet fed F. indicus were challenged with virulent WSSV and studied the survival, molecular diagnosis, biochemical, haematological, and immunological parameters. Surprisingly, the pooled fractions (F-IV to FVI) incorporated diets helped to significantly (P<0.01) suppressed viral multiplication, showed significant (P<0.01) differences in protein and glucose levels, improved total haemocyte count (THC), coagulase activity, significantly increased (P <= 0.001) prophenol oxidase and intracellular superoxide anion production compared to the control shrimps. Based on the results, C. rotundus extracts effectively suppressed WSSV multiplication and improve the immune system in F. indicus against WSSV infection and this knowledge will helps to develop novel drugs from C. rotundus against WSSV.
Keywords: Antiviral drugs, Cyperus rotundus, Fenneropenaeus indicus, WSSV.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 266034 Investigation of Combined use of MFCC and LPC Features in Speech Recognition Systems
Authors: К. R. Aida–Zade, C. Ardil, S. S. Rustamov
Abstract:
Statement of the automatic speech recognition problem, the assignment of speech recognition and the application fields are shown in the paper. At the same time as Azerbaijan speech, the establishment principles of speech recognition system and the problems arising in the system are investigated. The computing algorithms of speech features, being the main part of speech recognition system, are analyzed. From this point of view, the determination algorithms of Mel Frequency Cepstral Coefficients (MFCC) and Linear Predictive Coding (LPC) coefficients expressing the basic speech features are developed. Combined use of cepstrals of MFCC and LPC in speech recognition system is suggested to improve the reliability of speech recognition system. To this end, the recognition system is divided into MFCC and LPC-based recognition subsystems. The training and recognition processes are realized in both subsystems separately, and recognition system gets the decision being the same results of each subsystems. This results in decrease of error rate during recognition. The training and recognition processes are realized by artificial neural networks in the automatic speech recognition system. The neural networks are trained by the conjugate gradient method. In the paper the problems observed by the number of speech features at training the neural networks of MFCC and LPC-based speech recognition subsystems are investigated. The variety of results of neural networks trained from different initial points in training process is analyzed. Methodology of combined use of neural networks trained from different initial points in speech recognition system is suggested to improve the reliability of recognition system and increase the recognition quality, and obtained practical results are shown.Keywords: Speech recognition, cepstral analysis, Voice activation detection algorithm, Mel Frequency Cepstral Coefficients, features of speech, Cepstral Mean Subtraction, neural networks, Linear Predictive Coding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 91233 Investigation and Identification of a Number of Precious and Semi-Precious Stones Related to Bam Historical Citadel Using Micro Raman Spectroscopy and Scanning Electron Microscopy
Authors: Nazli Darkhal
Abstract:
The use of gems and ornaments has been common in Iran since the beginning of history. The prosperity of the country, the wealth, and the interest of the people of this land in a luxurious and glorious life, combined with beauty, have always attracted the attention of Iranian people to gems and jewelry. Iranians are famous in the world for having a long history of collecting and recognizing precious stones. In this case, we can use the unique treasure of national jewelry. Raman spectroscopy method is one of the oscillating spectroscopy methods that is classified in the group of nondestructive study methods, and like other methods, in addition to several advantages, it also has disadvantages and problems. Micro Raman spectroscopy is one of the different types of Raman spectroscopy in which an optical microscope is combined with a Raman device to provide more capabilities and advantages than its original method. In this way, with the help of Raman spectroscopy and a light microscope, while observing more details from different parts of the historical sample, natural or artificial pigments can be identified in a small part of it. The EDX (Energy Dispersive X ray) electron microscope also functions as the basis for the interaction of the electron beam with the matter. The beams emitted from this interaction can be used to examine samples. In this article, in addition to introducing the micro-Raman spectroscopy method, studies have been conducted on the structure of three samples of existing stones in the historic citadel of Bam. Using this method of study on precious and semi-precious stones, in addition to requiring a short time, can provide us with complete information about the structure and theme of these samples. The results of experiments and gemology of the stones showed that the selected beads are agate and jasper, and they can be placed in the chalcedony group.
Keywords: Bam citadel, precious stones, semi-precious stones, Raman spectroscopy, scanning electron microscope.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 41132 Clustering for Detection of Population Groups at Risk from Anticholinergic Medication
Authors: Amirali Shirazibeheshti, Tarik Radwan, Alireza Ettefaghian, Farbod Khanizadeh, George Wilson, Cristina Luca
Abstract:
Anticholinergic medication has been associated with events such as falls, delirium, and cognitive impairment in older patients. To further assess this, anticholinergic burden scores have been developed to quantify risk. A risk model based on clustering was deployed in a healthcare management system to cluster patients into multiple risk groups according to anticholinergic burden scores of multiple medicines prescribed to patients to facilitate clinical decision-making. To do so, anticholinergic burden scores of drugs were extracted from the literature which categorizes the risk on a scale of 1 to 3. Given the patients’ prescription data on the healthcare database, a weighted anticholinergic risk score was derived per patient based on the prescription of multiple anticholinergic drugs. This study was conducted on 300,000 records of patients currently registered with a major regional UK-based healthcare provider. The weighted risk scores were used as inputs to an unsupervised learning algorithm (mean-shift clustering) that groups patients into clusters that represent different levels of anticholinergic risk. This work evaluates the association between the average risk score and measures of socioeconomic status (index of multiple deprivation) and health (index of health and disability). The clustering identifies a group of 15 patients at the highest risk from multiple anticholinergic medication. Our findings show that this group of patients is located within more deprived areas of London compared to the population of other risk groups. Furthermore, the prescription of anticholinergic medicines is more skewed to female than male patients, suggesting that females are more at risk from this kind of multiple medication. The risk may be monitored and controlled in a healthcare management system that is well-equipped with tools implementing appropriate techniques of artificial intelligence.
Keywords: Anticholinergic medication, socioeconomic status, deprivation, clustering, risk analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 106931 Congolese Wood in the Antwerp Interwar Interior
Authors: M. Jaenen, M. de Bouw, A. Verdonck, M. Leus
Abstract:
During the interwar period artificial materials were often preferred, but many Antwerp architects relied on the application of wood for most of the interior finishing works and furnishings. Archival, literature and on site research of interwar suburban townhouses and the Belgian wood and furniture industry gave a new insight to the application of wood in the interwar interior. Many interwar designers favored the decorative values in all treatments of wood because of its warmth, comfort, good-wearing, and therefore, economic qualities. For the creation of a successful modern interior the texture and surface of the wood becomes as important as the color itself. This aesthetics valuation was the result of the modernization of the wood industry. The development of veneer and plywood gave the possibility to create strong, flat, long and plain wooden surfaces which are capable of retaining their shape. Also the modernization of cutting machines resulted in high quality and diversity in texture of veneer. The flat and plain plywood surfaces were modern decorated with all kinds of veneer-sliced options. In addition, wood species from the former Belgian Colony Congo were imported. Limba (Terminalia superba), kambala (Chlorophora excelsa), mubala (Pentaclethra macrophylla) and sapelli (Entandrophragma cylindricum) were used in the interior of many Antwerp interwar suburban town houses. From the thirties onwards Belgian wood firms established modern manufactures in Congo. There the local wood was dried, cut and prepared for exportation to the harbor of Antwerp. The presence of all kinds of strong and decorative Congolese wood products supported its application in the interwar interior design. The Antwerp architects combined them in their designs for doors, floors, stairs, built-in-furniture, wall paneling and movable furniture.Keywords: Antwerp, Congo, furniture, interwar.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 143330 Game-Tree Simplification by Pattern Matching and Its Acceleration Approach using an FPGA
Authors: Suguru Ochiai, Toru Yabuki, Yoshiki Yamaguchi, Yuetsu Kodama
Abstract:
In this paper, we propose a Connect6 solver which adopts a hybrid approach based on a tree-search algorithm and image processing techniques. The solver must deal with the complicated computation and provide high performance in order to make real-time decisions. The proposed approach enables the solver to be implemented on a single Spartan-6 XC6SLX45 FPGA produced by XILINX without using any external devices. The compact implementation is achieved through image processing techniques to optimize a tree-search algorithm of the Connect6 game. The tree search is widely used in computer games and the optimal search brings the best move in every turn of a computer game. Thus, many tree-search algorithms such as Minimax algorithm and artificial intelligence approaches have been widely proposed in this field. However, there is one fundamental problem in this area; the computation time increases rapidly in response to the growth of the game tree. It means the larger the game tree is, the bigger the circuit size is because of their highly parallel computation characteristics. Here, this paper aims to reduce the size of a Connect6 game tree using image processing techniques and its position symmetric property. The proposed solver is composed of four computational modules: a two-dimensional checkmate strategy checker, a template matching module, a skilful-line predictor, and a next-move selector. These modules work well together in selecting next moves from some candidates and the total amount of their circuits is small. The details of the hardware design for an FPGA implementation are described and the performance of this design is also shown in this paper.Keywords: Connect6, pattern matching, game-tree reduction, hardware direct computation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 197329 A Review of Emerging Technologies in Antennas and Phased Arrays for Avionics Systems
Authors: Muhammad Safi, Abdul Manan
Abstract:
In recent years, research in aircraft avionics systems (i.e., radars and antennas) has grown revolutionary. Aircraft technology is experiencing an increasing inclination from all mechanical to all electrical aircraft, with the introduction of inhabitant air vehicles and drone taxis over the last few years. This develops an overriding need to summarize the history, latest trends, and future development in aircraft avionics research for a better understanding and development of new technologies in the domain of avionics systems. This paper focuses on the future trends in antennas and phased arrays for avionics systems. Along with the general overview of the future avionics trend, this work describes the review of around 50 high-quality research papers on aircraft communication systems. Electric-powered aircrafts have been a hot topic in the modern aircraft world. Electric aircrafts have supremacy over their conventional counterparts. Due to increased drone taxi and urban air mobility, fast and reliable communication is very important, so concepts of Broadband Integrated Digital Avionics Information Exchange Networks (B-IDAIENs) and Modular Avionics are being researched for better communication of future aircraft. A Ku-band phased array antenna based on a modular design can be used in a modular avionics system. Furthermore, integrated avionics is also emerging research in future avionics. The main focus of work in future avionics will be using integrated modular avionics and infra-red phased array antennas, which are discussed in detail in this paper. Other work such as reconfigurable antennas and optical communication, are also discussed in this paper. The future of modern aircraft avionics would be based on integrated modulated avionics and small artificial intelligence-based antennas. Optical and infrared communication will also replace microwave frequencies.
Keywords: AI, avionics systems, communication, electric aircrafts, Infra-red, integrated avionics, modular avionics, phased array, reconfigurable antenna, UAVs.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16728 A Probabilistic Reinforcement-Based Approach to Conceptualization
Authors: Hadi Firouzi, Majid Nili Ahmadabadi, Babak N. Araabi
Abstract:
Conceptualization strengthens intelligent systems in generalization skill, effective knowledge representation, real-time inference, and managing uncertain and indefinite situations in addition to facilitating knowledge communication for learning agents situated in real world. Concept learning introduces a way of abstraction by which the continuous state is formed as entities called concepts which are connected to the action space and thus, they illustrate somehow the complex action space. Of computational concept learning approaches, action-based conceptualization is favored because of its simplicity and mirror neuron foundations in neuroscience. In this paper, a new biologically inspired concept learning approach based on the probabilistic framework is proposed. This approach exploits and extends the mirror neuron-s role in conceptualization for a reinforcement learning agent in nondeterministic environments. In the proposed method, instead of building a huge numerical knowledge, the concepts are learnt gradually from rewards through interaction with the environment. Moreover the probabilistic formation of the concepts is employed to deal with uncertain and dynamic nature of real problems in addition to the ability of generalization. These characteristics as a whole distinguish the proposed learning algorithm from both a pure classification algorithm and typical reinforcement learning. Simulation results show advantages of the proposed framework in terms of convergence speed as well as generalization and asymptotic behavior because of utilizing both success and failures attempts through received rewards. Experimental results, on the other hand, show the applicability and effectiveness of the proposed method in continuous and noisy environments for a real robotic task such as maze as well as the benefits of implementing an incremental learning scenario in artificial agents.
Keywords: Concept learning, probabilistic decision making, reinforcement learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 152627 Radioactivity Assessment of Sediments in Negombo Lagoon Sri Lanka
Authors: H. M. N. L. Handagiripathira
Abstract:
The distributions of naturally occurring and anthropogenic radioactive materials were determined in surface sediments taken at 27 different locations along the bank of Negombo Lagoon in Sri Lanka. Hydrographic parameters of lagoon water and the grain size analyses of the sediment samples were also carried out for this study. The conductivity of the adjacent water was varied from 13.6 mS/cm to 55.4 mS/cm near to the southern end and the northern end of the lagoon, respectively, and equally salinity levels varied from 7.2 psu to 32.1 psu. The average pH in the water was 7.6 and average water temperature was 28.7 °C. The grain size analysis emphasized the mass fractions of the samples as sand (60.9%), fine sand (30.6%) and fine silt+clay (1.3%) in the sampling locations. The surface sediment samples of wet weight, 1 kg each from upper 5-10 cm layer, were oven dried at 105 °C for 24 hours to get a constant weight, homogenized and sieved through a 2 mm sieve (IAEA technical series no. 295). The radioactivity concentrations were determined using gamma spectrometry technique. Ultra Low Background Broad Energy High Purity Ge Detector, BEGe (Model BE5030, Canberra) was used for radioactivity measurement with Canberra Industries' Laboratory Source-less Calibration Software (LabSOCS) mathematical efficiency calibration approach and Geometry composer software. The mean activity concentration was found to be 24 ± 4, 67 ± 9, 181 ± 10, 59 ± 8, 3.5 ± 0.4 and 0.47 ± 0.08 Bq/kg for 238U, 232Th, 40K, 210Pb, 235U and 137Cs respectively. The mean absorbed dose rate in air, radium equivalent activity, external hazard index, annual gonadal dose equivalent and annual effective dose equivalent were 60.8 nGy/h, 137.3 Bq/kg, 0.4, 425.3 mSv/year and 74.6 mSv/year, respectively. The results of this study will provide baseline information on the natural and artificial radioactive isotopes and environmental pollution associated with information on radiological risk.
Keywords: Gamma spectrometry, lagoon, radioactivity, sediments.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 52326 A Methodology for Creating Energy Sustainability in an Enterprise
Authors: John Lamb, Robert Epstein, Vasundhara L. Bhupathi, Sanjeev Kumar Marimekala
Abstract:
As we enter the new era of Artificial Intelligence (AI) and cloud computing, we mostly rely on the machine and natural language processing capabilities of AI, and energy efficient hardware and software devices in almost every industry sector. In these industry sectors, much emphasis is on developing new and innovative methods for producing and conserving energy and to sustain the depletion of natural resources. The core pillars of sustainability are Economic, Environmental, and Social, which are also informally referred to as 3 P's (People, Planet and Profits). The 3 P's play a vital role in creating a core sustainability model in the enterprise. Natural resources are continually being depleted, so there is more focus and growing demand for renewable energy. With this growing demand there is also a growing concern in many industries on how to reduce carbon emission and conserve natural resources while adopting sustainability in the corporate business models and policies. In our paper, we would like to discuss the driving forces such as climate changes, natural disasters, pandemic, disruptive technologies, corporate policies, scaled business models and emerging social media and AI platforms that influence the 3 main pillars of sustainability (3P’s). Through this paper, we would like to bring an overall perspective on enterprise strategies and the primary focus on bringing cultural shifts in adapting energy efficient operational models. Overall, many industries across the globe are incorporating core sustainability principles such as reducing energy costs, reducing greenhouse gas (GHG) emissions, reducing waste and increase recycling, adopting advanced monitoring and metering infrastructure, reducing server footprint and compute resources (shared IT services, cloud computing and application modernization) with the vision for a sustainable environment.
Keywords: AI, cloud computing, machine learning, social media platform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20225 Deorbiting Performance of Electrodynamic Tethers to Mitigate Space Debris
Authors: Giulia Sarego, Lorenzo Olivieri, Andrea Valmorbida, Carlo Bettanini, Giacomo Colombatti, Marco Pertile, Enrico C. Lorenzini
Abstract:
International guidelines recommend removing any artificial body in Low Earth Orbit (LEO) within 25 years from mission completion. Among disposal strategies, electrodynamic tethers appear to be a promising option for LEO, thanks to the limited storage mass and the minimum interface requirements to the host spacecraft. In particular, recent technological advances make it feasible to deorbit large objects with tether lengths of a few kilometers or less. To further investigate such an innovative passive system, the European Union is currently funding the project E.T.PACK – Electrodynamic Tether Technology for Passive Consumable-less Deorbit Kit in the framework of the H2020 Future Emerging Technologies (FET) Open program. The project focuses on the design of an end of life disposal kit for LEO satellites. This kit aims to deploy a taped tether that can be activated at the spacecraft end of life to perform autonomous deorbit within the international guidelines. In this paper, the orbital performance of the E.T.PACK deorbiting kit is compared to other disposal methods. Besides, the orbital decay prediction is parametrized as a function of spacecraft mass and tether system performance. Different values of length, width, and thickness of the tether will be evaluated for various scenarios (i.e., different initial orbital parameters). The results will be compared to other end-of-life disposal methods with similar allocated resources. The analysis of the more innovative system’s performance with the tape coated with a thermionic material, which has a low work-function (LWT), for which no active component for the cathode is required, will also be briefly discussed. The results show that the electrodynamic tether option can be a competitive and performant solution for satellite disposal compared to other deorbit technologies.Keywords: Deorbiting performance, H2020, spacecraft disposal, space electrodynamic tethers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 62624 A Grid-based Neural Network Framework for Multimodal Biometrics
Authors: Sitalakshmi Venkataraman
Abstract:
Recent scientific investigations indicate that multimodal biometrics overcome the technical limitations of unimodal biometrics, making them ideally suited for everyday life applications that require a reliable authentication system. However, for a successful adoption of multimodal biometrics, such systems would require large heterogeneous datasets with complex multimodal fusion and privacy schemes spanning various distributed environments. From experimental investigations of current multimodal systems, this paper reports the various issues related to speed, error-recovery and privacy that impede the diffusion of such systems in real-life. This calls for a robust mechanism that caters to the desired real-time performance, robust fusion schemes, interoperability and adaptable privacy policies. The main objective of this paper is to present a framework that addresses the abovementioned issues by leveraging on the heterogeneous resource sharing capacities of Grid services and the efficient machine learning capabilities of artificial neural networks (ANN). Hence, this paper proposes a Grid-based neural network framework for adopting multimodal biometrics with the view of overcoming the barriers of performance, privacy and risk issues that are associated with shared heterogeneous multimodal data centres. The framework combines the concept of Grid services for reliable brokering and privacy policy management of shared biometric resources along with a momentum back propagation ANN (MBPANN) model of machine learning for efficient multimodal fusion and authentication schemes. Real-life applications would be able to adopt the proposed framework to cater to the varying business requirements and user privacies for a successful diffusion of multimodal biometrics in various day-to-day transactions.Keywords: Back Propagation, Grid Services, MultimodalBiometrics, Neural Networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 191623 The Implications of Technological Advancements on the Constitutional Principles of Contract Law
Authors: Laura Çami (Vorpsi), Xhon Skënderi
Abstract:
In today's rapidly evolving technological landscape, the traditional principles of contract law are facing significant challenges. The emergence of new technologies, such as electronic signatures, smart contracts, and online dispute resolution mechanisms, is transforming the way contracts are formed, interpreted, and enforced. This paper examines the implications of these technological advancements on the constitutional principles of contract law. One of the fundamental principles of contract law is freedom of contract, which ensures that parties have the autonomy to negotiate and enter into contracts as they see fit. However, the use of technology in the contracting process has the potential to disrupt this principle. For example, online platforms and marketplaces often offer standard-form contracts, which may not reflect the specific needs or interests of individual parties. This raises questions about the equality of bargaining power between parties and the extent to which parties are truly free to negotiate the terms of their contracts. Another important principle of contract law is the requirement of consideration, which requires that each party receives something of value in exchange for their promise. The use of digital assets, such as cryptocurrencies, has created new challenges in determining what constitutes valuable consideration in a contract. Due to the ambiguity in this area, disagreements about the legality and enforceability of such contracts may arise. Furthermore, the use of technology in dispute resolution mechanisms, such as online arbitration and mediation, may raise concerns about due process and access to justice. The use of algorithms and artificial intelligence to determine the outcome of disputes may also raise questions about the impartiality and fairness of the process. Finally, it should be noted that there are many different and complex effects of technical improvements on the fundamental constitutional foundations of contract law. As technology continues to evolve, it will be important for policymakers and legal practitioners to consider the potential impacts on contract law and to ensure that the principles of fairness, equality, and access to justice are preserved in the contracting process.
Keywords: Technological advancements, constitutional principles, contract law, smart contracts, online dispute resolution, freedom of contract.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24922 Implementing an Intuitive Reasoner with a Large Weather Database
Authors: Yung-Chien Sun, O. Grant Clark
Abstract:
In this paper, the implementation of a rule-based intuitive reasoner is presented. The implementation included two parts: the rule induction module and the intuitive reasoner. A large weather database was acquired as the data source. Twelve weather variables from those data were chosen as the “target variables" whose values were predicted by the intuitive reasoner. A “complex" situation was simulated by making only subsets of the data available to the rule induction module. As a result, the rules induced were based on incomplete information with variable levels of certainty. The certainty level was modeled by a metric called "Strength of Belief", which was assigned to each rule or datum as ancillary information about the confidence in its accuracy. Two techniques were employed to induce rules from the data subsets: decision tree and multi-polynomial regression, respectively for the discrete and the continuous type of target variables. The intuitive reasoner was tested for its ability to use the induced rules to predict the classes of the discrete target variables and the values of the continuous target variables. The intuitive reasoner implemented two types of reasoning: fast and broad where, by analogy to human thought, the former corresponds to fast decision making and the latter to deeper contemplation. . For reference, a weather data analysis approach which had been applied on similar tasks was adopted to analyze the complete database and create predictive models for the same 12 target variables. The values predicted by the intuitive reasoner and the reference approach were compared with actual data. The intuitive reasoner reached near-100% accuracy for two continuous target variables. For the discrete target variables, the intuitive reasoner predicted at least 70% as accurately as the reference reasoner. Since the intuitive reasoner operated on rules derived from only about 10% of the total data, it demonstrated the potential advantages in dealing with sparse data sets as compared with conventional methods.Keywords: Artificial intelligence, intuition, knowledge acquisition, limited certainty.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 138221 Experimental Determination of Shear Strength Properties of Lightweight Expanded Clay Aggregates Using Direct Shear and Triaxial Tests
Authors: Mahsa Shafaei Bajestani, Mahmoud Yazdani, Aliakbar Golshani
Abstract:
Artificial lightweight aggregates have a wide range of applications in industry and engineering. Nowadays, the usage of this material in geotechnical activities, especially as backfill in retaining walls has been growing due to the specific characteristics which make it a competent alternative to the conventional geotechnical materials. In practice, a material with lower weight but higher shear strength parameters would be ideal as backfill behind retaining walls because of the important roles that these parameters play in decreasing the overall active lateral earth pressure. In this study, two types of Light Expanded Clay Aggregates (LECA) produced in the Leca factory are investigated. LECA is made in a rotary kiln by heating natural clay at different temperatures up to 1200 °C making quasi-spherical aggregates with different sizes ranged from 0 to 25 mm. The loose bulk density of these aggregates is between 300 and 700 kN/m3. The purpose of this research is to determine the stress-strain behavior, shear strength parameters, and the energy absorption of LECA materials. Direct shear tests were conducted at five normal stresses of 25, 50, 75, 100, and 200 kPa. In addition, conventional triaxial compression tests were operated at confining pressures of 50, 100, and 200 kPa to examine stress-strain behavior. The experimental results show a high internal angle of friction and even a considerable amount of nominal cohesion despite the granular structure of LECA. These desirable properties along with the intrinsic low density of these aggregates make LECA as a very proper material in geotechnical applications. Furthermore, the results demonstrate that lightweight aggregates may have high energy absorption that is excellent alternative material in seismic isolations.
Keywords: Expanded clay, direct shear test, triaxial test, shear properties, energy absorption.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 128020 Laboratory Analysis of Stormwater Runoff Hydraulic and Pollutant Removal Performance of Pervious Concrete Based on Seashell By-Products
Authors: Jean-Jacques Randrianarimanana, Nassim Sebaibi, Mohamed Boutouil
Abstract:
In order to solve problems associated with stormwater runoff in urban areas and their effects on natural and artificial water bodies, the integration of new technical solutions to the rainwater drainage becomes even more essential. Permeable pavement systems are one of the most widely used techniques. This paper presents a laboratory analysis of stormwater runoff hydraulic and pollutant removal performance of permeable pavement system using pervious pavements based on seashell products. The laboratory prototype is a square column of 25 cm of side and consists of the surface in pervious concrete, a bedding of 3 cm in height, a geotextile and a subbase layer of 50 cm in height. A series of constant simulated rain events using semi-synthetic runoff which varied in intensity and duration were carried out. The initial vertical saturated hydraulic conductivity of the entire pervious pavement system was 0.25 cm/s (148 L/m2/min). The hydraulic functioning was influenced by both the inlet flow rate value and the test duration. The total water losses including evaporation ranged between 9% to 20% for all hydraulic experiments. The temporal and vertical variability of the pollutant removal efficiency (PRE) of the system were studied for total suspended solids (TSS). The results showed that the PRE along the vertical profile was influenced by the size of the suspended solids, and the pervious paver has the highest capacity to trap pollutant than the other porous layers of the permeable pavement system after the geotextile. The TSS removal efficiency was about 80% for the entire system. The first-flush effect of TSS was observed, but it appeared only at the beginning (2 to 6 min) of the experiments. It has been shown that the PPS can capture first-flush. The project in which this study is integrated aims to contribute to both the valorization of shellfish waste and the sustainable management of rainwater.
Keywords: Hydraulic, pervious concrete, pollutant removal efficiency, seashell by-products, stormwater runoff.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 95819 Hand Gesture Detection via EmguCV Canny Pruning
Authors: N. N. Mosola, S. J. Molete, L. S. Masoebe, M. Letsae
Abstract:
Hand gesture recognition is a technique used to locate, detect, and recognize a hand gesture. Detection and recognition are concepts of Artificial Intelligence (AI). AI concepts are applicable in Human Computer Interaction (HCI), Expert systems (ES), etc. Hand gesture recognition can be used in sign language interpretation. Sign language is a visual communication tool. This tool is used mostly by deaf societies and those with speech disorder. Communication barriers exist when societies with speech disorder interact with others. This research aims to build a hand recognition system for Lesotho’s Sesotho and English language interpretation. The system will help to bridge the communication problems encountered by the mentioned societies. The system has various processing modules. The modules consist of a hand detection engine, image processing engine, feature extraction, and sign recognition. Detection is a process of identifying an object. The proposed system uses Canny pruning Haar and Haarcascade detection algorithms. Canny pruning implements the Canny edge detection. This is an optimal image processing algorithm. It is used to detect edges of an object. The system employs a skin detection algorithm. The skin detection performs background subtraction, computes the convex hull, and the centroid to assist in the detection process. Recognition is a process of gesture classification. Template matching classifies each hand gesture in real-time. The system was tested using various experiments. The results obtained show that time, distance, and light are factors that affect the rate of detection and ultimately recognition. Detection rate is directly proportional to the distance of the hand from the camera. Different lighting conditions were considered. The more the light intensity, the faster the detection rate. Based on the results obtained from this research, the applied methodologies are efficient and provide a plausible solution towards a light-weight, inexpensive system which can be used for sign language interpretation.
Keywords: Canny pruning, hand recognition, machine learning, skin tracking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 130818 Normal and Peaberry Coffee Beans Classification from Green Coffee Bean Images Using Convolutional Neural Networks and Support Vector Machine
Authors: Hira Lal Gope, Hidekazu Fukai
Abstract:
The aim of this study is to develop a system which can identify and sort peaberries automatically at low cost for coffee producers in developing countries. In this paper, the focus is on the classification of peaberries and normal coffee beans using image processing and machine learning techniques. The peaberry is not bad and not a normal bean. The peaberry is born in an only single seed, relatively round seed from a coffee cherry instead of the usual flat-sided pair of beans. It has another value and flavor. To make the taste of the coffee better, it is necessary to separate the peaberry and normal bean before green coffee beans roasting. Otherwise, the taste of total beans will be mixed, and it will be bad. In roaster procedure time, all the beans shape, size, and weight must be unique; otherwise, the larger bean will take more time for roasting inside. The peaberry has a different size and different shape even though they have the same weight as normal beans. The peaberry roasts slower than other normal beans. Therefore, neither technique provides a good option to select the peaberries. Defect beans, e.g., sour, broken, black, and fade bean, are easy to check and pick up manually by hand. On the other hand, the peaberry pick up is very difficult even for trained specialists because the shape and color of the peaberry are similar to normal beans. In this study, we use image processing and machine learning techniques to discriminate the normal and peaberry bean as a part of the sorting system. As the first step, we applied Deep Convolutional Neural Networks (CNN) and Support Vector Machine (SVM) as machine learning techniques to discriminate the peaberry and normal bean. As a result, better performance was obtained with CNN than with SVM for the discrimination of the peaberry. The trained artificial neural network with high performance CPU and GPU in this work will be simply installed into the inexpensive and low in calculation Raspberry Pi system. We assume that this system will be used in under developed countries. The study evaluates and compares the feasibility of the methods in terms of accuracy of classification and processing speed.
Keywords: Convolutional neural networks, coffee bean, peaberry, sorting, support vector machine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 155317 Evaluation of Azo Dye Toxicity Using Some Haematological and Histopathological Alterations in Fish Catla catla
Authors: Barot Jagruti
Abstract:
The textile industry plays a major role in the economy of India and on the other side of the coin it is the major source for water pollution. As azo dyes is the largest dye class they are extensively used in many fields such as textile industry, leather tanning industry, paper production, food, color photography, pharmaceuticals and medicine, cosmetic, hair colorings, wood staining, agricultural, biological and chemical research etc. In addition to these, they can have acute and/or chronic effects on organisms depending on their concentration and length of exposure when they discharged as effluent in the environment. The aim of this study was to assess the genotoxic and histotoxic potentials of environmentally relevant concentrations of C. I. Reactive Red 120 (RR 120) on Catla catla, important edible freshwater fingerlings. For this, healthy Catla catla fingerlings were procured from the Government Fish Farm and acclimatized in 100 L capacity and continuously aerated glass aquarium in laboratory for 15 days. According to APHA some physic-chemical parameters were measured and maintained such as temperature, pH, dissolve oxygen, alkalinity, total hardness. Water along with excreta had been changed every 24 hrs. All fingerlings were fed artificial food palates once a day @ body weight. After 15 days fingerlings were grouped in 5 (10 in each) and exposed to various concentrations of RR 120 (Control, 10, 20, 30 and 40 mg.l-1) and samples (peripheral blood and gills, kidney) were collected and analyzed at 96 hrs. All results were compared with the control. Micronuclei (MN), nuclear buds (NB), fragmented-apoptotic (FA) and bi-nucleated (BN) cells in blood smears and in tissues (gills and kidney cells) were observed. Prominent histopathological alterations were noticed in gills such as aneurism, hyperplasia, degenerated central axis, lifting of gill epithelium, curved secondary gill lamellae etc. Similarly kidney showed some detrimental changes like shrunken glomeruli with increased periglomerular space, degenerated renal tubules etc. Both haematological and histopathological changes clearly reveal the toxic potential of RR 120. This work concludes that water pollution assessment can be done by these two biomarkers which provide baseline to the further chromosomal or molecular work.
Keywords: Catla catla, genotoxicity, histopathlogicalchanges, RR 120azo dye.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 281416 Enhancement of Hardness Related Properties of Grey Cast Iron Powder Reinforced AA7075 Metal Matrix Composites through T6 and T8 Heat Treatments
Authors: S. S. Sharma, P. R. Prabhu, K. Jagannath, Achutha Kini U., Gowri Shankar M. C.
Abstract:
In present global scenario, aluminum alloys are coining the attention of many innovators as competing structural materials for automotive and space applications. Comparing to other challenging alloys, especially, 7xxx series aluminum alloys have been studied seriously because of benefits such as moderate strength; better deforming characteristics and affordable cost. It is expected that substitution of aluminum alloys for steels will result in great improvements in energy economy, durability and recyclability. However, it is necessary to improve the strength and the formability levels at low temperatures in aluminum alloys for still better applications. Aluminum–Zinc–Magnesium with or without other wetting agent denoted as 7XXX series alloys are medium strength heat treatable alloys. In addition to Zn, Mg as major alloying additions, Cu, Mn and Si are the other solute elements which contribute for the improvement in mechanical properties by suitable heat treatment process. Subjecting to suitable treatments like age hardening or cold deformation assisted heat treatments; known as low temperature thermomechanical treatments (LTMT) the challenging properties might be incorporated. T6 is the age hardening or precipitation hardening process with artificial aging cycle whereas T8 comprises of LTMT treatment aged artificially with X% cold deformation. When the cold deformation is provided after solution treatment, there is increase in hardness related properties such as wear resistance, yield and ultimate strength, toughness with the expense of ductility. During precipitation hardening both hardness and strength of the samples are increasing. The hardness value may further improve when room temperature deformation is positively supported with age hardening known as thermomechanical treatment. It is intended to perform heat treatment and evaluate hardness, tensile strength, wear resistance and distribution pattern of reinforcement in the matrix. 2 to 2.5 and 3 to 3.5 times increase in hardness is reported in age hardening and LTMT treatments respectively as compared to as-cast composite. There was better distribution of reinforcements in the matrix, nearly two fold increase in strength levels and up to 5 times increase in wear resistance are also observed in the present study.Keywords: Reinforcement, precipitation, thermomechanical, dislocation, strain hardening.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 220115 Leveraging xAPI in a Corporate e-Learning Environment to Facilitate the Tracking, Modelling, and Predictive Analysis of Learner Behaviour
Authors: Libor Zachoval, Daire O Broin, Oisin Cawley
Abstract:
E-learning platforms, such as Blackboard have two major shortcomings: limited data capture as a result of the limitations of SCORM (Shareable Content Object Reference Model), and lack of incorporation of Artificial Intelligence (AI) and machine learning algorithms which could lead to better course adaptations. With the recent development of Experience Application Programming Interface (xAPI), a large amount of additional types of data can be captured and that opens a window of possibilities from which online education can benefit. In a corporate setting, where companies invest billions on the learning and development of their employees, some learner behaviours can be troublesome for they can hinder the knowledge development of a learner. Behaviours that hinder the knowledge development also raise ambiguity about learner’s knowledge mastery, specifically those related to gaming the system. Furthermore, a company receives little benefit from their investment if employees are passing courses without possessing the required knowledge and potential compliance risks may arise. Using xAPI and rules derived from a state-of-the-art review, we identified three learner behaviours, primarily related to guessing, in a corporate compliance course. The identified behaviours are: trying each option for a question, specifically for multiple-choice questions; selecting a single option for all the questions on the test; and continuously repeating tests upon failing as opposed to going over the learning material. These behaviours were detected on learners who repeated the test at least 4 times before passing the course. These findings suggest that gauging the mastery of a learner from multiple-choice questions test scores alone is a naive approach. Thus, next steps will consider the incorporation of additional data points, knowledge estimation models to model knowledge mastery of a learner more accurately, and analysis of the data for correlations between knowledge development and identified learner behaviours. Additional work could explore how learner behaviours could be utilised to make changes to a course. For example, course content may require modifications (certain sections of learning material may be shown to not be helpful to many learners to master the learning outcomes aimed at) or course design (such as the type and duration of feedback).
Keywords: Compliance Course, Corporate Training, Learner Behaviours, xAPI.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 561