Search results for: hierarchical text classification models
5373 Development of Digital Twin Concept to Detect Abnormal Changes in Structural Behaviour
Authors: Shady Adib, Vladimir Vinogradov, Peter Gosling
Abstract:
Digital Twin (DT) technology is a new technology that appeared in the early 21st century. The DT is defined as the digital representation of living and non-living physical assets. By connecting the physical and virtual assets, data are transmitted smoothly, allowing the virtual asset to fully represent the physical asset. Although there are lots of studies conducted on the DT concept, there is still limited information about the ability of the DT models for monitoring and detecting unexpected changes in structural behaviour in real time. This is due to the large computational efforts required for the analysis and an excessively large amount of data transferred from sensors. This paper aims to develop the DT concept to be able to detect the abnormal changes in structural behaviour in real time using advanced modelling techniques, deep learning algorithms, and data acquisition systems, taking into consideration model uncertainties. finite element (FE) models were first developed offline to be used with a reduced basis (RB) model order reduction technique for the construction of low-dimensional space to speed the analysis during the online stage. The RB model was validated against experimental test results for the establishment of a DT model of a two-dimensional truss. The established DT model and deep learning algorithms were used to identify the location of damage once it has appeared during the online stage. Finally, the RB model was used again to identify the damage severity. It was found that using the RB model, constructed offline, speeds the FE analysis during the online stage. The constructed RB model showed higher accuracy for predicting the damage severity, while deep learning algorithms were found to be useful for estimating the location of damage with small severity.Keywords: data acquisition system, deep learning, digital twin, model uncertainties, reduced basis, reduced order model
Procedia PDF Downloads 995372 Silver Nanoparticles Synthesized in Plant Extract Against Acute Hepatopancreatic Necrosis of Shrimp: Estimated By Multiple Models
Authors: Luz del Carmen Rubí Félix Peña, Jose Adan Felix-Ortiz, Ely Sara Lopez-Alvarez, Wenceslao Valenzuela-Quiñonez
Abstract:
On a global scale, Mexico is the sixth largest producer of farmed white shrimp (Penaeus vannamei). The activity suffered significant economic losses due to acute hepatopancreatic necrosis (AHPND) caused by a strain of Vibrio parahaemolyticus. For control, the first option is the application of antibiotics in food, causing changes in the environment and bacterial communities, which has produced greater virulence and resistance of pathogenic bacteria. An alternative treatment is silver nanoparticles (AgNPs) generated by green synthesis, which have shown an antibacterial capacity by destroying the cell membrane or denaturing the cell. However, the doses at which these are effective are still unknown. The aim is to calculate the minimum inhibitory concentration (MIC) using the Gompertz, Richard, and Logistic model of biosynthesized AgNPs against a strain of V. parahaemolyticus. Through the testing of different formulations of AgNPs synthesized from Euphorbia prostrate (Ep) extracts against V. parahaemolyticus causing AHPND in white shrimp. Aqueous and ethanol extracts were obtained, and the concentration of phenols and flavonoids was quantified. In the antibiograms, AgNPs were formulated in ethanol extracts of Ep (20 and 30%). The inhibition halo at well dilution test were 18±1.7 and 17.67±2.1 mm against V. parahaemolyticus. A broth microdilution was performed with the inhibitory agents (aqueous and ethanolic extracts and AgNPs) and 20 μL of the inoculum of V. parahaemolyticus. The MIC for AgNPs was 6.2-9.3 μg/mL and for ethanol extract of 49-73 mg/mL. The Akaike index (AIC) was used to choose the Gompertz model for ethanol extracts of Ep as the best data descriptor (AIC=204.8, 10%; 45.5, 20%, and 204.8, 30%). The Richards model was at AgNPs ethanol extract with AIC=-9.3 (10%), -17.5 (20 and 30%). The MIC calculated for EP extracts with the modified Gompertz model were 20 mg/mL (10% and 20% extract) and 40 mg/mL at 30%, while Richard was winner for AgNPs-synthesized it was 5 μg/mL (10% and 20%) and 8 μg/mL (30%). The solver tool Excel was used for the calculations of the models and inhibition curves against V.parahaemolyticus.Keywords: green synthesis, euphorbia prostata, phenols, flavonoids, bactericide
Procedia PDF Downloads 1065371 Zero Energy Buildings in Hot-Humid Tropical Climates: Boundaries of the Energy Optimization Grey Zone
Authors: Nakul V. Naphade, Sandra G. L. Persiani, Yew Wah Wong, Pramod S. Kamath, Avinash H. Anantharam, Hui Ling Aw, Yann Grynberg
Abstract:
Achieving zero-energy targets in existing buildings is known to be a difficult task requiring important cuts in the building energy consumption, which in many cases clash with the functional necessities of the building wherever the on-site energy generation is unable to match the overall energy consumption. Between the building’s consumption optimization limit and the energy, target stretches a case-specific optimization grey zone, which requires tailored intervention and enhanced user’s commitment. In the view of the future adoption of more stringent energy-efficiency targets in the context of hot-humid tropical climates, this study aims to define the energy optimization grey zone by assessing the energy-efficiency limit in the state-of-the-art typical mid- and high-rise full AC office buildings, through the integration of currently available technologies. Energy models of two code-compliant generic office-building typologies were developed as a baseline, a 20-storey ‘high-rise’ and a 7-storey ‘mid-rise’. Design iterations carried out on the energy models with advanced market ready technologies in lighting, envelope, plug load management and ACMV systems and controls, lead to a representative energy model of the current maximum technical potential. The simulations showed that ZEB targets could be achieved in fully AC buildings under an average of seven floors only by compromising on energy-intense facilities (as full AC, unlimited power-supply, standard user behaviour, etc.). This paper argues that drastic changes must be made in tropical buildings to span the energy optimization grey zone and achieve zero energy. Fully air-conditioned areas must be rethought, while smart technologies must be integrated with an aggressive involvement and motivation of the users to synchronize with the new system’s energy savings goal.Keywords: energy simulation, office building, tropical climate, zero energy buildings
Procedia PDF Downloads 1845370 Flood Monitoring in the Vietnamese Mekong Delta Using Sentinel-1 SAR with Global Flood Mapper
Authors: Ahmed S. Afifi, Ahmed Magdy
Abstract:
Satellite monitoring is an essential tool to study, understand, and map large-scale environmental changes that affect humans, climate, and biodiversity. The Sentinel-1 Synthetic Aperture Radar (SAR) instrument provides a high collection of data in all-weather, short revisit time, and high spatial resolution that can be used effectively in flood management. Floods occur when an overflow of water submerges dry land that requires to be distinguished from flooded areas. In this study, we use global flood mapper (GFM), a new google earth engine application that allows users to quickly map floods using Sentinel-1 SAR. The GFM enables the users to adjust manually the flood map parameters, e.g., the threshold for Z-value for VV and VH bands and the elevation and slope mask threshold. The composite R:G:B image results by coupling the bands of Sentinel-1 (VH:VV:VH) reduces false classification to a large extent compared to using one separate band (e.g., VH polarization band). The flood mapping algorithm in the GFM and the Otsu thresholding are compared with Sentinel-2 optical data. And the results show that the GFM algorithm can overcome the misclassification of a flooded area in An Giang, Vietnam.Keywords: SAR backscattering, Sentinel-1, flood mapping, disaster
Procedia PDF Downloads 1065369 Primary School Students’ Modeling Processes: Crime Problem
Authors: Neslihan Sahin Celik, Ali Eraslan
Abstract:
As a result of PISA (Program for International Student Assessments) survey that tests how well students can apply the knowledge and skills they have learned at school to real-life challenges, the new and redesigned mathematics education programs in many countries emphasize the necessity for the students to face complex and multifaceted problem situations and gain experience in this sense allowing them to develop new skills and mathematical thinking to prepare them for their future life after school. At this point, mathematical models and modeling approaches can be utilized in the analysis of complex problems which represent real-life situations in which students can actively participate. In particular, model eliciting activities that bring about situations which allow the students to create solutions to problems and which involve mathematical modeling must be used right from primary school years, allowing them to face such complex, real-life situations from early childhood period. A qualitative study was conducted in a university foundation primary school in the city center of a big province in 2013-2014 academic years. The participants were 4th grade students in a primary school. After a four-week preliminary study applied to a fourth-grade classroom, three students included in the focus group were selected using criterion sampling technique. A focus group of three students was videotaped as they worked on the Crime Problem. The conversation of the group was transcribed, examined with students’ written work and then analyzed through the lens of Blum and Ferri’s modeling processing cycle. The results showed that primary fourth-grade students can successfully work with model eliciting problem while they encounter some difficulties in the modeling processes. In particular, they developed new ideas based on different assumptions, identified the patterns among variables and established a variety of models. On the other hand, they had trouble focusing on problems and occasionally had breaks in the process.Keywords: primary school, modeling, mathematical modeling, crime problem
Procedia PDF Downloads 4055368 Classifying Facial Expressions Based on a Motion Local Appearance Approach
Authors: Fabiola M. Villalobos-Castaldi, Nicolás C. Kemper, Esther Rojas-Krugger, Laura G. Ramírez-Sánchez
Abstract:
This paper presents the classification results about exploring the combination of a motion based approach with a local appearance method to describe the facial motion caused by the muscle contractions and expansions that are presented in facial expressions. The proposed feature extraction method take advantage of the knowledge related to which parts of the face reflects the highest deformations, so we selected 4 specific facial regions at which the appearance descriptor were applied. The most common used approaches for feature extraction are the holistic and the local strategies. In this work we present the results of using a local appearance approach estimating the correlation coefficient to the 4 corresponding landmark-localized facial templates of the expression face related to the neutral face. The results let us to probe how the proposed motion estimation scheme based on the local appearance correlation computation can simply and intuitively measure the motion parameters for some of the most relevant facial regions and how these parameters can be used to recognize facial expressions automatically.Keywords: facial expression recognition system, feature extraction, local-appearance method, motion-based approach
Procedia PDF Downloads 4135367 Near Infrared Spectrometry to Determine the Quality of Milk, Experimental Design Setup and Chemometrics: Review
Authors: Meghana Shankara, Priyadarshini Natarajan
Abstract:
Infrared (IR) spectroscopy has revolutionized the way we look at materials around us. Unraveling the pattern in the molecular spectra of materials to analyze the composition and properties of it has been one of the most interesting challenges in modern science. Applications of the IR spectrometry are numerous in the field’s pharmaceuticals, health, food and nutrition, oils, agriculture, construction, polymers, beverage, fabrics and much more limited only by the curiosity of the people. Near Infrared (NIR) spectrometry is applied robustly in analyzing the solids and liquid substances because of its non-destructive analysis method. In this paper, we have reviewed the application of NIR spectrometry in milk quality analysis and have presented the modes of measurement applied in NIRS measurement setup, Design of Experiment (DoE), classification/quantification algorithms used in the case of milk composition prediction like Fat%, Protein%, Lactose%, Solids Not Fat (SNF%) along with different approaches for adulterant identification. We have also discussed the important NIR ranges for the chosen milk parameters. The performance metrics used in the comparison of the various Chemometric approaches include Root Mean Square Error (RMSE), R^2, slope, offset, sensitivity, specificity and accuracyKeywords: chemometrics, design of experiment, milk quality analysis, NIRS measurement modes
Procedia PDF Downloads 2715366 Research on the Spatial Organization and Collaborative Innovation of Innovation Corridors from the Perspective of Ecological Niche: A Case Study of Seven Municipal Districts in Jiangsu Province, China
Authors: Weikang Peng
Abstract:
The innovation corridor is an important spatial carrier to promote regional collaborative innovation, and its development process is the spatial re-organization process of regional innovation resources. This paper takes the Nanjing-Zhenjiang G312 Industrial Innovation Corridor, which involves seven municipal districts in Jiangsu Province, as empirical evidence. Based on multi-source spatial big data in 2010, 2016, and 2022, this paper applies triangulated irregular network (TIN), head/tail breaks, regional innovation ecosystem (RIE) niche fitness evaluation model, and social network analysis to carry out empirical research on the spatial organization and functional structural evolution characteristics of innovation corridors and their correlation with the structural evolution of collaborative innovation network. The results show, first, the development of innovation patches in the corridor has fractal characteristics in time and space and tends to be multi-center and cluster layout along the Nanjing Bypass Highway and National Highway G312. Second, there are large differences in the spatial distribution pattern of niche fitness in the corridor in various dimensions, and the niche fitness of innovation patches along the highway has increased significantly. Third, the scale of the collaborative innovation network in the corridor is expanding fast. The core of the network is shifting from the main urban area to the periphery of the city along the highway, with small-world and hierarchical levels, and the core-edge network structure is highlighted. With the development of the Innovation Corridor, the main collaborative mode in the corridor is changing from collaboration within innovation patches to collaboration between innovation patches, and innovation patches with high ecological suitability tend to be the active areas of collaborative innovation. Overall, polycentric spatial layout, graded functional structure, diversified innovation clusters, and differentiated environmental support play an important role in effectively constructing collaborative innovation linkages and the stable expansion of the scale of collaborative innovation within the innovation corridor.Keywords: innovation corridor development, spatial structure, niche fitness evaluation model, head/tail breaks, innovation network
Procedia PDF Downloads 205365 Quantifying Uncertainties in an Archetype-Based Building Stock Energy Model by Use of Individual Building Models
Authors: Morten Brøgger, Kim Wittchen
Abstract:
Focus on reducing energy consumption in existing buildings at large scale, e.g. in cities or countries, has been increasing in recent years. In order to reduce energy consumption in existing buildings, political incentive schemes are put in place and large scale investments are made by utility companies. Prioritising these investments requires a comprehensive overview of the energy consumption in the existing building stock, as well as potential energy-savings. However, a building stock comprises thousands of buildings with different characteristics making it difficult to model energy consumption accurately. Moreover, the complexity of the building stock makes it difficult to convey model results to policymakers and other stakeholders. In order to manage the complexity of the building stock, building archetypes are often employed in building stock energy models (BSEMs). Building archetypes are formed by segmenting the building stock according to specific characteristics. Segmenting the building stock according to building type and building age is common, among other things because this information is often easily available. This segmentation makes it easy to convey results to non-experts. However, using a single archetypical building to represent all buildings in a segment of the building stock is associated with loss of detail. Thermal characteristics are aggregated while other characteristics, which could affect the energy efficiency of a building, are disregarded. Thus, using a simplified representation of the building stock could come at the expense of the accuracy of the model. The present study evaluates the accuracy of a conventional archetype-based BSEM that segments the building stock according to building type- and age. The accuracy is evaluated in terms of the archetypes’ ability to accurately emulate the average energy demands of the corresponding buildings they were meant to represent. This is done for the buildings’ energy demands as a whole as well as for relevant sub-demands. Both are evaluated in relation to the type- and the age of the building. This should provide researchers, who use archetypes in BSEMs, with an indication of the expected accuracy of the conventional archetype model, as well as the accuracy lost in specific parts of the calculation, due to use of the archetype method.Keywords: building stock energy modelling, energy-savings, archetype
Procedia PDF Downloads 1545364 Assessing Students’ Readiness for an Open and Distance Learning Higher Education Environment
Authors: Upasana G. Singh, Meera Gungea
Abstract:
Learning is no more confined to the traditional classroom, teacher, and student interaction. Many universities offer courses through the Open and Distance Learning (ODL) mode, attracting a diversity of learners in terms of age, gender, and profession to name a few. The ODL mode has surfaced as one of the famous sought-after modes of learning, allowing learners to invest in their educational growth without hampering their personal and professional commitments. This mode of learning, however, requires that those who ultimately choose to adopt it must be prepared to undertake studies through such medium. The purpose of this research is to assess whether students who join universities offering courses through the ODL mode are ready to embark and study within such a framework. This study will be helpful to unveil the challenges students face in such an environment and thus contribute to developing a framework to ease adoption and integration into the ODL environment. Prior to the implementation of e-learning, a readiness assessment is essential for any institution that wants to adopt any form of e-learning. Various e-learning readiness assessment models have been developed over the years. However, this study is based on a conceptual model for e-Learning Readiness Assessment which is a ‘hybrid model’. This hybrid model consists of 4 main parameters: 1) Technological readiness, 2) Culture readiness, 3) Content readiness, and 4) Demographics factors, with 4 sub-areas, namely, technology, innovation, people and self-development. The model also includes the attitudes of users towards the adoption of e-learning as an important aspect of assessing e-learning readiness. For this study, some factors and sub-factors of the hybrid model have been considered and adapted, together with the ‘Attitude’ component. A questionnaire was designed based on the models and students where the target population were students enrolled at the Open University of Mauritius, in undergraduate and postgraduate courses. Preliminary findings indicate that most (68%) learners have an average knowledge about ODL form of learning, despite not many (72%) having previous experience with ODL. Despite learning through ODL 74% of learners preferred hard copy learning material and 48% found difficulty in reading learning material on electronic devices.Keywords: open learning, distance learning, student readiness, a hybrid model
Procedia PDF Downloads 1095363 Identification of Landslide Features Using Back-Propagation Neural Network on LiDAR Digital Elevation Model
Authors: Chia-Hao Chang, Geng-Gui Wang, Jee-Cheng Wu
Abstract:
The prediction of a landslide is a difficult task because it requires a detailed study of past activities using a complete range of investigative methods to determine the changing condition. In this research, first step, LiDAR 1-meter by 1-meter resolution of digital elevation model (DEM) was used to generate six environmental factors of landslide. Then, back-propagation neural networks (BPNN) was adopted to identify scarp, landslide areas and non-landslide areas. The BPNN uses 6 environmental factors in input layer and 1 output layer. Moreover, 6 landslide areas are used as training areas and 4 landslide areas as test areas in the BPNN. The hidden layer is set to be 1 and 2; the hidden layer neurons are set to be 4, 5, 6, 7 and 8; the learning rates are set to be 0.01, 0.1 and 0.5. When using 1 hidden layer with 7 neurons and the learning rate sets to be 0.5, the result of Network training root mean square error is 0.001388. Finally, evaluation of BPNN classification accuracy by the confusion matrix shows that the overall accuracy can reach 94.4%, and the Kappa value is 0.7464.Keywords: digital elevation model, DEM, environmental factors, back-propagation neural network, BPNN, LiDAR
Procedia PDF Downloads 1445362 Subfamilial Relationships within Solanaceae as Inferred from atpB-rbcL Intergenic Spacer
Authors: Syeda Qamarunnisa, Ishrat Jamil, Abid Azhar, Zabta K. Shinwari, Syed Irtifaq Ali
Abstract:
A phylogenetic analysis of family Solanaceae was conducted using sequence data from the chloroplast intergenic atpB-rbcL spacer. Sequence data was generated from 17 species representing 09 out of 14 genera of Solanaceae from Pakistan. Cladogram was constructed using maximum parsimony method and results indicate that Solanaceae is mainly divided into two subfamilies; Solanoideae and Cestroideae. Four major clades within Solanoideae represent tribes; Physaleae, Capsiceae, Datureae and Solaneae are supported by high bootstrap value and the relationships among them are not corroborating with the previous studies. The findings established that subfamily Cestroideae comprised of three genera; Cestrum, Lycium, and Nicotiana with high bootstrap support. Position of Nicotiana inferred with atpB-rbcL sequence is congruent with traditional classification, which placed the taxa in Cestroideae. In the current study Lycium unexpectedly nested with Nicotiana with 100% bootstrap support and identified as a member of tribe Nicotianeae. Expanded sampling of other genera from Pakistan could be valuable towards improving our understanding of intrafamilial relationships within Solanaceae.Keywords: systematics, solanaceae, phylogenetics, intergenic spacer, tribes
Procedia PDF Downloads 4685361 Classification of State Transition by Using a Microwave Doppler Sensor for Wandering Detection
Authors: K. Shiba, T. Kaburagi, Y. Kurihara
Abstract:
With global aging, people who require care, such as people with dementia (PwD), are increasing within many developed countries. And PwDs may wander and unconsciously set foot outdoors, it may lead serious accidents, such as, traffic accidents. Here, round-the-clock monitoring by caregivers is necessary, which can be a burden for the caregivers. Therefore, an automatic wandering detection system is required when an elderly person wanders outdoors, in which case the detection system transmits a ‘moving’ followed by an ‘absence’ state. In this paper, we focus on the transition from the ‘resting’ to the ‘absence’ state, via the ‘moving’ state as one of the wandering transitions. To capture the transition of the three states, our method based on the hidden Markov model (HMM) is built. Using our method, the restraint where the ‘resting’ state and ‘absence’ state cannot be transmitted to each other is applied. To validate our method, we conducted the experiment with 10 subjects. Our results show that the method can classify three states with 0.92 accuracy.Keywords: wander, microwave Doppler sensor, respiratory frequency band, the state transition, hidden Markov model (HMM).
Procedia PDF Downloads 1835360 Machine Learning Strategies for Data Extraction from Unstructured Documents in Financial Services
Authors: Delphine Vendryes, Dushyanth Sekhar, Baojia Tong, Matthew Theisen, Chester Curme
Abstract:
Much of the data that inform the decisions of governments, corporations and individuals are harvested from unstructured documents. Data extraction is defined here as a process that turns non-machine-readable information into a machine-readable format that can be stored, for instance, in a database. In financial services, introducing more automation in data extraction pipelines is a major challenge. Information sought by financial data consumers is often buried within vast bodies of unstructured documents, which have historically required thorough manual extraction. Automated solutions provide faster access to non-machine-readable datasets, in a context where untimely information quickly becomes irrelevant. Data quality standards cannot be compromised, so automation requires high data integrity. This multifaceted task is broken down into smaller steps: ingestion, table parsing (detection and structure recognition), text analysis (entity detection and disambiguation), schema-based record extraction, user feedback incorporation. Selected intermediary steps are phrased as machine learning problems. Solutions leveraging cutting-edge approaches from the fields of computer vision (e.g. table detection) and natural language processing (e.g. entity detection and disambiguation) are proposed.Keywords: computer vision, entity recognition, finance, information retrieval, machine learning, natural language processing
Procedia PDF Downloads 1135359 Factors Influencing the Roles and Responsibilities of Middle Leaders in Saudi and English Primary Schools: A Comparative Critical Study
Authors: Saeed Musaid H. Alzahrani
Abstract:
The role of middle leaders, especially in primary schools, is a multi-faced role that has been subject to changes in nature over recent decades, with claims for more distributed leadership practices. This research examines the way 18 middle leaders in Saudi and English primary schools conceptualise their roles and responsibilities, and different factors influencing those roles and responsibilities. It begins from the premise that both the power of the role and the values of middle leaders are grounded in cultural and political bases, a belief held by the researcher as an 'insider' within the Saudi educational leadership context. The study consisted of a comparative analysis of the role and the responsibilities of middle leaders in Saudi primary schools and their equivalents in English primary schools. A purely qualitative methodological stance was adopted, using in-depth face-to-face semi-structured interviews, observations and document analysis. Middle leaders were asked to reflect deeply on their perceptions and understanding of their roles and explain what they thought influenced their daily practices and responsibilities. The findings suggest that the concept of middle leadership has been influenced by power imposed from above by political authority, via internal and external hierarchical structures, which shapes the nature of the role of the middle leaders and forces them to comply. Middle leaders seem to believe they have the power to make decisions and promote change, but these findings suggest that this is illusory. The power that keeps middle leaders performing is the power of their cultural and religious values. Those values are the resource to which they turn in their search for more energy when they lack support and are short of time taken. Middle leaders in Saudi, just like their equivalents in English schools must comply with the requirements of their role. However, Saudi middle leaders are given no leeway to make decisions or implement change, neither do they have the culture of collegiality that seems to give middle leaders in England more power over their resources and decisions. However, in neither educational setting have middle leaders been given the power to lead, so they remain managers rather than leaders. The findings of this research suggest that there are more similarities between the educational settings of Saudi and England than differences; and in the light of different factors identified in the study, suggest the establishment of a framework for middle leadership, in the hope of enhancing the way the role is practiced.Keywords: middle leader, primary school, power, educational leadership, value, culture, model, Saudi Arabia, England
Procedia PDF Downloads 1925358 Electroencephalography-Based Intention Recognition and Consensus Assessment during Emergency Response
Abstract:
After natural and man-made disasters, robots can bypass the danger, expedite the search, and acquire unprecedented situational awareness to design rescue plans. The hands-free requirement from the first responders excludes the use of tedious manual control and operation. In unknown, unstructured, and obstructed environments, natural-language-based supervision is not amenable for first responders to formulate, and is difficult for robots to understand. Brain-computer interface is a promising option to overcome the limitations. This study aims to test the feasibility of using electroencephalography (EEG) signals to decode human intentions and detect the level of consensus on robot-provided information. EEG signals were classified using machine-learning and deep-learning methods to discriminate search intentions and agreement perceptions. The results show that the average classification accuracy for intention recognition and consensus assessment is 67% and 72%, respectively, proving the potential of incorporating recognizable users’ bioelectrical responses into advanced robot-assisted systems for emergency response.Keywords: consensus assessment, electroencephalogram, emergency response, human-robot collaboration, intention recognition, search and rescue
Procedia PDF Downloads 935357 Study of the Relationship between the Civil Engineering Parameters and the Floating of Buoy Model Which Made from Expanded Polystyrene-Mortar
Authors: Panarat Saengpanya
Abstract:
There were five objectives in this study including the study of housing type with water environment, the physical and mechanical properties of the buoy material, the mechanical properties of the buoy models, the floating of the buoy models and the relationship between the civil engineering parameters and the floating of the buoy. The buoy examples made from Expanded Polystyrene (EPS) covered by 5 mm thickness of mortar with the equal thickness on each side. Specimens are 0.05 m cubes tested at a displacement rate of 0.005 m/min. The existing test method used to assess the parameters relationship is ASTM C 109 to provide comparative results. The results found that the three type of housing with water environment were Stilt Houses, Boat House, and Floating House. EPS is a lightweight material that has been used in engineering applications since at least the 1950s. Its density is about a hundredth of that of mortar, while the mortar strength was found 72 times of EPS. One of the advantage of composite is that two or more materials could be combined to take advantage of the good characteristics of each of the material. The strength of the buoy influenced by mortar while the floating influenced by EPS. Results showed the buoy example compressed under loading. The Stress-Strain curve showed the high secant modulus before reached the peak value. The failure occurred within 10% strain then the strength reduces while the strain was continuing. It was observed that the failure strength reduced by increasing the total volume of examples. For the buoy examples with same area, an increase of the failure strength is found when the high dimension is increased. The results showed the relationship between five parameters including the floating level, the bearing capacity, the volume, the high dimension and the unit weight. The study found increases in high of buoy lead to corresponding decreases in both modulus and compressive strength. The total volume and the unit weight had relationship with the bearing capacity of the buoy.Keywords: floating house, buoy, floating structure, EPS
Procedia PDF Downloads 1465356 A Sustainable Supplier Selection and Order Allocation Based on Manufacturing Processes and Product Tolerances: A Multi-Criteria Decision Making and Multi-Objective Optimization Approach
Authors: Ravi Patel, Krishna K. Krishnan
Abstract:
In global supply chains, appropriate and sustainable suppliers play a vital role in supply chain development and feasibility. In a larger organization with huge number of suppliers, it is necessary to divide suppliers based on their past history of quality and delivery of each product category. Since performance of any organization widely depends on their suppliers, well evaluated selection criteria and decision-making models lead to improved supplier assessment and development. In this paper, SCOR® performance evaluation approach and ISO standards are used to determine selection criteria for better utilization of supplier assessment by using hybrid model of Analytic Hierchchy Problem (AHP) and Fuzzy Techniques for Order Preference by Similarity to Ideal Solution (FTOPSIS). AHP is used to determine the global weightage of criteria which helps TOPSIS to get supplier score by using triangular fuzzy set theory. Both qualitative and quantitative criteria are taken into consideration for the proposed model. In addition, a multi-product and multi-time period model is selected for order allocation. The optimization model integrates multi-objective integer linear programming (MOILP) for order allocation and a hybrid approach for supplier selection. The proposed MOILP model optimizes order allocation based on manufacturing process and product tolerances as per manufacturer’s requirement for quality product. The integrated model and solution approach are tested to find optimized solutions for different scenario. The detailed analysis shows the superiority of proposed model over other solutions which considered individual decision making models.Keywords: AHP, fuzzy set theory, multi-criteria decision making, multi-objective integer linear programming, TOPSIS
Procedia PDF Downloads 1715355 An Innovative Auditory Impulsed EEG and Neural Network Based Biometric Identification System
Authors: Ritesh Kumar, Gitanjali Chhetri, Mandira Bhatia, Mohit Mishra, Abhijith Bailur, Abhinav
Abstract:
The prevalence of the internet and technology in our day to day lives is creating more security issues than ever. The need for protecting and providing a secure access to private and business data has led to the development of many security systems. One of the potential solutions is to employ the bio-metric authentication technique. In this paper we present an innovative biometric authentication method that utilizes a person’s EEG signal, which is acquired in response to an auditory stimulus,and transferred wirelessly to a computer that has the necessary ANN algorithm-Multi layer perceptrol neural network because of is its ability to differentiate between information which is not linearly separable.In order to determine the weights of the hidden layer we use Gaussian random weight initialization. MLP utilizes a supervised learning technique called Back propagation for training the network. The complex algorithm used for EEG classification reduces the chances of intrusion into the protected public or private data.Keywords: EEG signal, auditory evoked potential, biometrics, multilayer perceptron neural network, back propagation rule, Gaussian random weight initialization
Procedia PDF Downloads 4095354 Examples of Parameterization of Stabilizing Controllers with One-Side Coprime Factorization
Authors: Kazuyoshi Mori
Abstract:
Examples of parameterization of stabilizing controllers that require only one of right-/left-coprime factorizations are presented. One parameterization method requires one side coprime factorization. The other requires no coprime factorization. The methods are based on the factorization approach so that a number of models can be applied the method we use in this paper.Keywords: parametrization, coprime factorization, factorization approach, linear systems
Procedia PDF Downloads 3735353 Massively-Parallel Bit-Serial Neural Networks for Fast Epilepsy Diagnosis: A Feasibility Study
Authors: Si Mon Kueh, Tom J. Kazmierski
Abstract:
There are about 1% of the world population suffering from the hidden disability known as epilepsy and major developing countries are not fully equipped to counter this problem. In order to reduce the inconvenience and danger of epilepsy, different methods have been researched by using a artificial neural network (ANN) classification to distinguish epileptic waveforms from normal brain waveforms. This paper outlines the aim of achieving massive ANN parallelization through a dedicated hardware using bit-serial processing. The design of this bit-serial Neural Processing Element (NPE) is presented which implements the functionality of a complete neuron using variable accuracy. The proposed design has been tested taking into consideration non-idealities of a hardware ANN. The NPE consists of a bit-serial multiplier which uses only 16 logic elements on an Altera Cyclone IV FPGA and a bit-serial ALU as well as a look-up table. Arrays of NPEs can be driven by a single controller which executes the neural processing algorithm. In conclusion, the proposed compact NPE design allows the construction of complex hardware ANNs that can be implemented in a portable equipment that suits the needs of a single epileptic patient in his or her daily activities to predict the occurrences of impending tonic conic seizures.Keywords: Artificial Neural Networks (ANN), bit-serial neural processor, FPGA, Neural Processing Element (NPE)
Procedia PDF Downloads 3215352 The Trend of Epidemics in Population and Body Regulation in Iran (1850-1920)
Authors: Seyedfateh Moradi
Abstract:
Medical issues mark the beginning of a new form of epistemology in nineteenth-century Iran. The emergence of epidemic diseases led to the formation of a medical discourse and conflict over the body which displayed itself in the concept of health progress and development. The discourse attributed to this development in the health system defines the general structure of the given period. This discourse manifested itself in the conflict between the traditional and new medicine. The regulation and classification of body and population reveal the nature of this period. The government attempted to adapt itself to the modern and progressive discourse. This paper seeks to reveal part of this rupture and adaptation around epidemics and modern medical discourse. Also, accepting part of the traditional discourse in the new era, or adapting and integrating parts of it indicate a delegation of part of the power of traditional authorities. The delegation of power arose in the context of the discursive hegemony of Western modernism from which there was no escape. This provided the ground for the acceptance of government and emergence of other discourses. Finally, during the reign of Reza Shah (1922-1942), body and population planning changed into the key issues of government, which created serious tensions in society.Keywords: epidemic, population, body, cholera, plague
Procedia PDF Downloads 715351 Predictor Factors in Predictive Model of Soccer Talent Identification among Male Players Aged 14 to 17 Years
Authors: Muhamad Hafiz Ismail, Ahmad H., Nelfianty M. R.
Abstract:
The longitudinal study is conducted to identify predictive factors of soccer talent among male players aged 14 to 17 years. Convenience sampling involving elite respondents (n=20) and sub-elite respondents (n=20) male soccer players. Descriptive statistics were reported as frequencies and percentages. The inferential statistical analysis is used to report the status of reliability, independent samples t-test, paired samples t-test, and multiple regression analysis. Generally, there are differences in mean of height, muscular strength, muscular endurance, cardiovascular endurance, task orientation, cognitive anxiety, self-confidence, juggling skills, short pass skills, long pass skills, dribbling skills, and shooting skills for 20 elite players and sub-elite players. Accordingly, there was a significant difference between pre and post-test for thirteen variables of height, weight, fat percentage, muscle strength, muscle endurance, cardiovascular endurance, flexibility, BMI, task orientation, juggling skills, short pass skills, a long pass skills, and dribbling skills. Based on the first predictive factors (physical), second predictive factors (fitness), third predictive factors (psychological), and fourth predictive factors (skills in playing football) pledged to the soccer talent; four multiple regression models were produced. The first predictive factor (physical) contributed 53.5 percent, supported by height and percentage of fat in soccer talents. The second predictive factor (fitness) contributed 63.2 percent and the third predictive factors (psychology) contributed 66.4 percent of soccer talent. The fourth predictive factors (skills) contributed 59.0 percent of soccer talent. The four multiple regression models could be used as a guide for talent scouting for soccer players of the future.Keywords: soccer talent identification, fitness and physical test, soccer skills test, psychological test
Procedia PDF Downloads 1575350 Application of Hydrological Engineering Centre – River Analysis System (HEC-RAS) to Estuarine Hydraulics
Authors: Julia Zimmerman, Gaurav Savant
Abstract:
This study aims to evaluate the efficacy of the U.S. Army Corp of Engineers’ River Analysis System (HEC-RAS) application to modeling the hydraulics of estuaries. HEC-RAS has been broadly used for a variety of riverine applications. However, it has not been widely applied to the study of circulation in estuaries. This report details the model development and validation of a combined 1D/2D unsteady flow hydraulic model using HEC-RAS for estuaries and they are associated with tidally influenced rivers. Two estuaries, Galveston Bay and Delaware Bay, were used as case studies. Galveston Bay, a bar-built, vertically mixed estuary, was modeled for the 2005 calendar year. Delaware Bay, a drowned river valley estuary, was modeled from October 22, 2019, to November 5, 2019. Water surface elevation was used to validate both models by comparing simulation results to NOAA’s Center for Operational Oceanographic Products and Services (CO-OPS) gauge data. Simulations were run using the Diffusion Wave Equations (DW), the Shallow Water Equations, Eulerian-Lagrangian Method (SWE-ELM), and the Shallow Water Equations Eulerian Method (SWE-EM) and compared for both accuracy and computational resources required. In general, the Diffusion Wave Equations results were found to be comparable to the two Shallow Water equations sets while requiring less computational power. The 1D/2D combined approach was valid for study areas within the 2D flow area, with the 1D flow serving mainly as an inflow boundary condition. Within the Delaware Bay estuary, the HEC-RAS DW model ran in 22 minutes and had an average R² value of 0.94 within the 2-D mesh. The Galveston Bay HEC-RAS DW ran in 6 hours and 47 minutes and had an average R² value of 0.83 within the 2-D mesh. The longer run time and lower R² for Galveston Bay can be attributed to the increased length of the time frame modeled and the greater complexity of the estuarine system. The models did not accurately capture tidal effects within the 1D flow area.Keywords: Delaware bay, estuarine hydraulics, Galveston bay, HEC-RAS, one-dimensional modeling, two-dimensional modeling
Procedia PDF Downloads 1995349 Evaluation of Reliability Flood Control System Based on Uncertainty of Flood Discharge, Case Study Wulan River, Central Java, Indonesia
Authors: Anik Sarminingsih, Krishna V. Pradana
Abstract:
The failure of flood control system can be caused by various factors, such as not considering the uncertainty of designed flood causing the capacity of the flood control system is exceeded. The presence of the uncertainty factor is recognized as a serious issue in hydrological studies. Uncertainty in hydrological analysis is influenced by many factors, starting from reading water elevation data, rainfall data, selection of method of analysis, etc. In hydrological modeling selection of models and parameters corresponding to the watershed conditions should be evaluated by the hydraulic model in the river as a drainage channel. River cross-section capacity is the first defense in knowing the reliability of the flood control system. Reliability of river capacity describes the potential magnitude of flood risk. Case study in this research is Wulan River in Central Java. This river occurring flood almost every year despite some efforts to control floods such as levee, floodway and diversion. The flood-affected areas include several sub-districts, mainly in Kabupaten Kudus and Kabupaten Demak. First step is analyze the frequency of discharge observation from Klambu weir which have time series data from 1951-2013. Frequency analysis is performed using several distribution frequency models such as Gumbel distribution, Normal, Normal Log, Pearson Type III and Log Pearson. The result of the model based on standard deviation overlaps, so the maximum flood discharge from the lower return periods may be worth more than the average discharge for larger return periods. The next step is to perform a hydraulic analysis to evaluate the reliability of river capacity based on the flood discharge resulted from several methods. The selection of the design flood discharge of flood control system is the result of the method closest to bankfull capacity of the river.Keywords: design flood, hydrological model, reliability, uncertainty, Wulan river
Procedia PDF Downloads 2945348 Statistical Variability of Soil Parameters within the Copper Belt Region of the Democratic Republic of the Congo
Authors: Stephan P. Barkhuizen, Deon Greyling, Ryan J. Miller
Abstract:
The accurate determination of the engineering parameters of soil is necessary for the design of geotechnical structures, such as Tailings Storage Facilities. The shear strength and saturated permeability of soil and tailings samples obtained from 14 sites located in the copper belt in the Democratic Republic of the Congo have been tested at six commercial soil laboratories in South Africa. This study compiles a database of the test results proved by the soil laboratories. The samples have been categorised into clay, silt, and sand, based on the Unified Soil Classification System, with tailings kept separate. The effective friction angle (Φ’) and cohesion (c’) were interpreted from the stress paths, in s’:t space, obtained from triaxial tests. The minimum, lower quartile, median, upper quartile, and maximum values for Φ’,c’, and saturated hydraulic conductivity (k) have been determined for the soil sample. The objective is to provide statistics of the measured values of the engineering properties for the TSF borrow material, foundation soils and tailings of this region.Keywords: Democratic Republic of the Congo, laboratory test work, soil engineering parameter variation, tailings storage facilities
Procedia PDF Downloads 645347 The Use of Classifiers in Image Analysis of Oil Wells Profiling Process and the Automatic Identification of Events
Authors: Jaqueline Maria Ribeiro Vieira
Abstract:
Different strategies and tools are available at the oil and gas industry for detecting and analyzing tension and possible fractures in borehole walls. Most of these techniques are based on manual observation of the captured borehole images. While this strategy may be possible and convenient with small images and few data, it may become difficult and suitable to errors when big databases of images must be treated. While the patterns may differ among the image area, depending on many characteristics (drilling strategy, rock components, rock strength, etc.). Previously we developed and proposed a novel strategy capable of detecting patterns at borehole images that may point to regions that have tension and breakout characteristics, based on segmented images. In this work we propose the inclusion of data-mining classification strategies in order to create a knowledge database of the segmented curves. These classifiers allow that, after some time using and manually pointing parts of borehole images that correspond to tension regions and breakout areas, the system will indicate and suggest automatically new candidate regions, with higher accuracy. We suggest the use of different classifiers methods, in order to achieve different knowledge data set configurations.Keywords: image segmentation, oil well visualization, classifiers, data-mining, visual computer
Procedia PDF Downloads 3035346 Implementation of Statistical Parameters to Form an Entropic Mathematical Models
Authors: Gurcharan Singh Buttar
Abstract:
It has been discovered that although these two areas, statistics, and information theory, are independent in their nature, they can be combined to create applications in multidisciplinary mathematics. This is due to the fact that where in the field of statistics, statistical parameters (measures) play an essential role in reference to the population (distribution) under investigation. Information measure is crucial in the study of ambiguity, assortment, and unpredictability present in an array of phenomena. The following communication is a link between the two, and it has been demonstrated that the well-known conventional statistical measures can be used as a measure of information.Keywords: probability distribution, entropy, concavity, symmetry, variance, central tendency
Procedia PDF Downloads 1565345 Digital Architectural Practice as a Challenge for Digital Architectural Technology Elements in the Era of Digital Design
Authors: Ling Liyun
Abstract:
In the field of contemporary architecture, complex forms of architectural works continue to emerge in the world, along with some new terminology emerged: digital architecture, parametric design, algorithm generation, building information modeling, CNC construction and so on. Architects gradually mastered the new skills of mathematical logic in the form of exploration, virtual simulation, and the entire design and coordination in the construction process. Digital construction technology has a greater degree in controlling construction, and ensure its accuracy, creating a series of new construction techniques. As a result, the use of digital technology is an improvement and expansion of the practice of digital architecture design revolution. We worked by reading and analyzing information about the digital architecture development process, a large number of cases, as well as architectural design and construction as a whole process. Thus current developments were introduced and discussed in our paper, such as architectural discourse, design theory, digital design models and techniques, material selecting, as well as artificial intelligence space design. Our paper also pays attention to the representative three cases of digital design and construction experiment at great length in detail to expound high-informatization, high-reliability intelligence, and high-technique in constructing a humane space to cope with the rapid development of urbanization. We concluded that the opportunities and challenges of the shift existed in architectural paradigms, such as the cooperation methods, theories, models, technologies and techniques which were currently employed in digital design research and digital praxis. We also find out that the innovative use of space can gradually change the way people learn, talk, and control information. The past two decades, digital technology radically breaks the technology constraints of industrial technical products, digests the publicity on a particular architectural style (era doctrine). People should not adapt to the machine, but in turn, it’s better to make the machine work for users.Keywords: artificial intelligence, collaboration, digital architecture, digital design theory, material selection, space construction
Procedia PDF Downloads 1365344 An Approach on Intelligent Tolerancing of Car Body Parts Based on Historical Measurement Data
Authors: Kai Warsoenke, Maik Mackiewicz
Abstract:
To achieve a high quality of assembled car body structures, tolerancing is used to ensure a geometric accuracy of the single car body parts. There are two main techniques to determine the required tolerances. The first is tolerance analysis which describes the influence of individually tolerated input values on a required target value. Second is tolerance synthesis to determine the location of individual tolerances to achieve a target value. Both techniques are based on classical statistical methods, which assume certain probability distributions. To ensure competitiveness in both saturated and dynamic markets, production processes in vehicle manufacturing must be flexible and efficient. The dimensional specifications selected for the individual body components and the resulting assemblies have a major influence of the quality of the process. For example, in the manufacturing of forming tools as operating equipment or in the higher level of car body assembly. As part of the metrological process monitoring, manufactured individual parts and assemblies are recorded and the measurement results are stored in databases. They serve as information for the temporary adjustment of the production processes and are interpreted by experts in order to derive suitable adjustments measures. In the production of forming tools, this means that time-consuming and costly changes of the tool surface have to be made, while in the body shop, uncertainties that are difficult to control result in cost-intensive rework. The stored measurement results are not used to intelligently design tolerances in future processes or to support temporary decisions based on real-world geometric data. They offer potential to extend the tolerancing methods through data analysis and machine learning models. The purpose of this paper is to examine real-world measurement data from individual car body components, as well as assemblies, in order to develop an approach for using the data in short-term actions and future projects. For this reason, the measurement data will be analyzed descriptively in the first step in order to characterize their behavior and to determine possible correlations. In the following, a database is created that is suitable for developing machine learning models. The objective is to create an intelligent way to determine the position and number of measurement points as well as the local tolerance range. For this a number of different model types are compared and evaluated. The models with the best result are used to optimize equally distributed measuring points on unknown car body part geometries and to assign tolerance ranges to them. The current results of this investigation are still in progress. However, there are areas of the car body parts which behave more sensitively compared to the overall part and indicate that intelligent tolerancing is useful here in order to design and control preceding and succeeding processes more efficiently.Keywords: automotive production, machine learning, process optimization, smart tolerancing
Procedia PDF Downloads 117