Search results for: Random Field Ising Model
22883 Expert Review on Conceptual Design Model of Assistive Courseware for Low Vision (AC4LV) Learners
Authors: Nurulnadwan Aziz, Ariffin Abdul Mutalib, Siti Mahfuzah Sarif
Abstract:
This paper reports an ongoing project regarding the development of Conceptual Design Model of Assistive Courseware for Low Vision (AC4LV) learners. Having developed the intended model, it has to be validated prior to producing it as guidance for the developers to develop an AC4LV. This study requires two phases of validation process which are through expert review and prototyping method. This paper presents a part of the validation process which is findings from experts review on Conceptual Design Model of AC4LV which has been carried out through a questionnaire. Results from 12 international and local experts from various respectable fields in Human-Computer Interaction (HCI) were discussed and justified. In a nutshell, reviewed Conceptual Design Model of AC4LV was formed. Future works of this study are to validate the reviewed model through prototyping method prior to testing it to the targeted users.Keywords: assistive courseware, conceptual design model, expert review, low vision learners
Procedia PDF Downloads 54622882 OpenFOAM Based Simulation of High Reynolds Number Separated Flows Using Bridging Method of Turbulence
Authors: Sagar Saroha, Sawan S. Sinha, Sunil Lakshmipathy
Abstract:
Reynolds averaged Navier-Stokes (RANS) model is the popular computational tool for prediction of turbulent flows. Being computationally less expensive as compared to direct numerical simulation (DNS), RANS has received wide acceptance in industry and research community as well. However, for high Reynolds number flows, the traditional RANS approach based on the Boussinesq hypothesis is incapacitated to capture all the essential flow characteristics, and thus, its performance is restricted in high Reynolds number flows of practical interest. RANS performance turns out to be inadequate in regimes like flow over curved surfaces, flows with rapid changes in the mean strain rate, duct flows involving secondary streamlines and three-dimensional separated flows. In the recent decade, partially averaged Navier-Stokes (PANS) methodology has gained acceptability among seamless bridging methods of turbulence- placed between DNS and RANS. PANS methodology, being a scale resolving bridging method, is inherently more suitable than RANS for simulating turbulent flows. The superior ability of PANS method has been demonstrated for some cases like swirling flows, high-speed mixing environment, and high Reynolds number turbulent flows. In our work, we intend to evaluate PANS in case of separated turbulent flows past bluff bodies -which is of broad aerodynamic research and industrial application. PANS equations, being derived from base RANS, continue to inherit the inadequacies from the parent RANS model based on linear eddy-viscosity model (LEVM) closure. To enhance PANS’ capabilities for simulating separated flows, the shortcomings of the LEVM closure need to be addressed. Inabilities of the LEVMs have inspired the development of non-linear eddy viscosity models (NLEVM). To explore the potential improvement in PANS performance, in our study we evaluate the PANS behavior in conjugation with NLEVM. Our work can be categorized into three significant steps: (i) Extraction of PANS version of NLEVM from RANS model, (ii) testing the model in the homogeneous turbulence environment and (iii) application and evaluation of the model in the canonical case of separated non-homogeneous flow field (flow past prismatic bodies and bodies of revolution at high Reynolds number). PANS version of NLEVM shall be derived and implemented in OpenFOAM -an open source solver. Homogeneous flows evaluation will comprise the study of the influence of the PANS’ filter-width control parameter on the turbulent stresses; the homogeneous analysis performed over typical velocity fields and asymptotic analysis of Reynolds stress tensor. Non-homogeneous flow case will include the study of mean integrated quantities and various instantaneous flow field features including wake structures. Performance of PANS + NLEVM shall be compared against the LEVM based PANS and LEVM based RANS. This assessment will contribute to significant improvement of the predictive ability of the computational fluid dynamics (CFD) tools in massively separated turbulent flows past bluff bodies.Keywords: bridging methods of turbulence, high Re-CFD, non-linear PANS, separated turbulent flows
Procedia PDF Downloads 14522881 Finite Element Modeling of Aortic Intramural Haematoma Shows Size Matters
Authors: Aihong Zhao, Priya Sastry, Mark L Field, Mohamad Bashir, Arvind Singh, David Richens
Abstract:
Objectives: Intramural haematoma (IMH) is one of the pathologies, along with acute aortic dissection, that present as Acute Aortic Syndrome (AAS). Evidence suggests that unlike aortic dissection, some intramural haematomas may regress with medical management. However, intramural haematomas have been traditionally managed like acute aortic dissections. Given that some of these pathologies may regress with conservative management, it would be useful to be able to identify which of these may not need high risk emergency intervention. A computational aortic model was used in this study to try and identify intramural haematomas with risk of progression to aortic dissection. Methods: We created a computational model of the aorta with luminal blood flow. Reports in the literature have identified 11 mm as the radial clot thickness that is associated with heightened risk of progression of intramural haematoma. Accordingly, haematomas of varying sizes were implanted in the modeled aortic wall to test this hypothesis. The model was exposed to physiological blood flows and the stresses and strains in each layer of the aortic wall were recorded. Results: Size and shape of clot were seen to affect the magnitude of aortic stresses. The greatest stresses and strains were recorded in the intima of the model. When the haematoma exceeded 10 mm in all dimensions, the stress on the intima reached breaking point. Conclusion: Intramural clot size appears to be a contributory factor affecting aortic wall stress. Our computer simulation corroborates clinical evidence in the literature proposing that IMH diameter greater than 11 mm may be predictive of progression. This preliminary report suggests finite element modelling of the aortic wall may be a useful process by which to examine putative variables important in predicting progression or regression of intramural haematoma.Keywords: intramural haematoma, acute aortic syndrome, finite element analysis,
Procedia PDF Downloads 43122880 Pyroelectric Effect on Thermoelectricity of AlInN/GaN Heterostructures
Authors: B. K. Sahoo
Abstract:
Superior thermoelectric (TE) efficiency of AlₓIn₁₋ₓN /GaN heterostructure (HS) requires a minimum value of thermal conductivity (k). A smaller k would lead to even further increase of TE figure of merit (ZT). The built-in polarization (BIP) electric field of AlₓIn₁₋ₓN /GaN HS enhances S, and σ of the HS, however, the effect of BIP field on k of the HS has not been explored. Study of thermal conductivities (k: without BIP and kp: including BIP) vs temperature predicts pyroelectric behavior of HS. Both k and kp show crossover at a temperature Tp. The result shows that below Tp, kp < k due to negative thermal expansion coefficient (TEC). However, above Tp, kp > k. Above Tp, piezoelectric polarization dominates over spontaneous polarization due to positive TEC. This generates more lattice mismatch resulting in the significant contribution of BIP field to thermal conductivity. Thus, Tp can be considered as primary pyroelectric transition temperature of the material as above Tp thermal expansion takes place which is the reason for the secondary pyroelectric effect. It is found that below Tp, kp is decreased; thus enhancing TE efficiency. For x=0.1, 0.2 and 0.3; Tp are close to 200, 210 and 260 K, respectively. Thus, k of the HS can be modified as per requirement by tailoring the Al composition; making it suitable simultaneously for the design of high-temperature pyroelectric sensors and TE module for maximum power production.Keywords: AlₓIn₁₋ₓN/GaN heterostructure, built in polarization, pyroelectric behavior, thermoelectric efficiency
Procedia PDF Downloads 12122879 Application of Model Tree in the Prediction of TBM Rate of Penetration with Synthetic Minority Oversampling Technique
Authors: Ehsan Mehryaar
Abstract:
The rate of penetration is (RoP) one of the vital factors in the cost and time of tunnel boring projects; therefore, predicting it can lead to a substantial increase in the efficiency of the project. RoP is heavily dependent geological properties of the project site and TBM properties. In this study, 151-point data from Queen’s water tunnel is collected, which includes unconfined compression strength, peak slope index, angle with weak planes, and distance between planes of weaknesses. Since the size of the data is small, it was observed that it is imbalanced. To solve that problem synthetic minority oversampling technique is utilized. The model based on the model tree is proposed, where each leaf consists of a support vector machine model. Proposed model performance is then compared to existing empirical equations in the literature.Keywords: Model tree, SMOTE, rate of penetration, TBM(tunnel boring machine), SVM
Procedia PDF Downloads 17422878 Experience of the Formation of Professional Competence of Students of IT-Specialties
Authors: B. I. Zhumagaliyev, L. Sh. Balgabayeva, G. S. Nabiyeva, B. A. Tulegenova, P. Oralkhan, B. S. Kalenova, S. S. Akhmetov
Abstract:
The article describes an approach to build competence in research of Bachelor and Master, which is now an important feature of modern specialist in the field of engineering. Provides an example of methodical teaching methods with the research aspect, is including the formulation of the problem, the method of conducting experiments, analysis of the results. Implementation of methods allows the student to better consolidate their knowledge and skills at the same time to get research. Knowledge on the part of the media requires some training in the subject area and teaching methods.Keywords: professional competence, model of it-specialties, teaching methods, educational technology, decision making
Procedia PDF Downloads 43722877 Managing Inter-Organizational Innovation Project: Systematic Review of Literature
Authors: Lamin B Ceesay, Cecilia Rossignoli
Abstract:
Inter-organizational collaboration is a growing phenomenon in both research and practice. The partnership between organizations enables firms to leverage external resources, experiences, and technology that lie with other firms. This collaborative practice is a source of improved business model performance, technological advancement, and increased competitive advantage for firms. However, the competitive intents, and even diverse institutional logics of firms, make inter-firm innovation-based partnership even more complex, and its governance more challenging. The purpose of this paper is to present a systematic review of research linking the inter-organizational relationship of firms with their innovation practice and specify the different project management issues and gaps addressed in previous research. To do this, we employed a systematic review of the literature on inter-organizational innovation using two complementary scholarly databases - ScienceDirect and Web of Science (WoS). Article scoping relies on the combination of keywords based on similar terms used in the literature:(1) inter-organizational relationship, (2) business network, (3) inter-firm project, and (4) innovation network. These searches were conducted in the title, abstract, and keywords of conceptual and empirical research papers done in English. Our search covers between 2010 to 2019. We applied several exclusion criteria including Papers published outside the years under the review, papers in a language other than English, papers neither listed in WoS nor ScienceDirect and papers that are not sharply related to the inter-organizational innovation-based partnership were removed. After all relevant search criteria were applied, a final list of 84 papers constitutes the data for this review. Our review revealed an increasing evolution of inter-organizational relationship research during the period under the review. The descriptive analysis of papers according to Journal outlets finds that International Journal of Project Management (IJPM), Journal of Industrial Marketing, Journal of Business Research (JBR), etc. are the leading journal outlets for research in the inter-organizational innovation project. The review also finds that Qualitative methods and quantitative approaches respectively are the leading research methods adopted by scholars in the field. However, literature review and conceptual papers constitute the least in the field. During the content analysis of the selected papers, we read the content of each paper and found that the selected papers try to address one of the three phenomena in inter-organizational innovation research: (1) project antecedents; (2) project management and (3) project performance outcomes. We found that these categories are not mutually exclusive, but rather interdependent. This categorization also helped us to organize the fragmented literature in the field. While a significant percentage of the literature discussed project management issues, we found fewer extant literature on project antecedents and performance. As a result of this, we organized the future research agenda addressed in several papers by linking them with the under-researched themes in the field, thus providing great potential to advance future research agenda especially, in the under-researched themes in the field. Finally, our paper reveals that research on inter-organizational innovation project is generally fragmented which hinders a better understanding of the field. Thus, this paper contributes to the understanding of the field by organizing and discussing the extant literature to advance the theory and application of inter-organizational relationship.Keywords: inter-organizational relationship, inter-firm collaboration, innovation projects, project management, systematic review
Procedia PDF Downloads 11422876 Development of In Situ Permeability Test Using Constant Discharge Method for Sandy Soils
Authors: A. Rifa’i, Y. Takeshita, M. Komatsu
Abstract:
The post-rain puddles problem that occurs in the first yard of Prambanan Temple are often disturbing visitor activity. A poodle layer and a drainage system has ever built to avoid such a problem, but puddles still didn’t stop appearing after rain. Permeability parameter needs to be determined by using more simple procedure to find exact method of solution. The instrument modelling were proposed according to the development of field permeability testing instrument. This experiment used proposed Constant Discharge method. Constant Discharge method used a tube poured with constant water flow. The procedure were carried out from unsaturated until saturated soil condition. Volumetric water content (θ) were being monitored by soil moisture measurement device. The results were relationship between k and θ which drawn by numerical approach Van Genutchen model. Parameters θr optimum value obtained from the test was at very dry soil. Coefficient of permeability with a density of 19.8 kN/m3 for unsaturated conditions was in range of 3 x 10-6 cm/sec (Sr= 68 %) until 9.98 x 10-4 cm/sec (Sr= 82 %). The equipment and testing procedure developed in this research was quite effective, simple and easy to be implemented on determining field soil permeability coefficient value of sandy soil. Using constant discharge method in proposed permeability test, value of permeability coefficient under unsaturated condition can be obtained without establish soil water characteristic curve.Keywords: constant discharge method, in situ permeability test, sandy soil, unsaturated conditions
Procedia PDF Downloads 38422875 Analysis of Trends in Environmental Health Research Using Topic Modeling
Authors: Hayoung Cho, Gabi Cho
Abstract:
In response to the continuing increase of demands for living environment safety, the Korean government has established and implemented various environmental health policies and set a high priority to the related R&D. However, the level of related technologies such as environmental risk assessment are still relatively low, and there is a need for detailed investment strategies in the field of environmental health research. As scientific research papers can give valuable implications on the development of a certain field, this study analyzed the global research trends in the field of environmental health over the past 10 years (2005~2015). Research topics were extracted from abstracts of the collected SCI papers using topic modeling to study the changes in research trends and discover emerging technologies. The method of topic modeling can improve the traditional bibliometric approach and provide a more comprehensive review of the global research development. The results of this study are expected to help provide insights for effective policy making and R&D investment direction.Keywords: environmental health, paper analysis, research trends, topic modeling
Procedia PDF Downloads 28822874 An Agent-Based Modeling and Simulation of Human Muscle
Authors: Sina Saadati, Mohammadreza Razzazi
Abstract:
In this article, we have tried to present an agent-based model of human muscle. A suitable model of muscle is necessary for the analysis of mankind's movements. It can be used by clinical researchers who study the influence of motion sicknesses, like Parkinson's disease. It is also useful in the development of a prosthesis that receives the electromyography signals and generates force as a reaction. Since we have focused on computational efficiency in this research, the model can compute the calculations very fast. As far as it concerns prostheses, the model can be known as a charge-efficient method. In this paper, we are about to illustrate an agent-based model. Then, we will use it to simulate the human gait cycle. This method can also be done reversely in the analysis of gait in motion sicknesses.Keywords: agent-based modeling and simulation, human muscle, gait cycle, motion sickness
Procedia PDF Downloads 11422873 Characterization of Complex Electromagnetic Environment Created by Multiple Sources of Electromagnetic Radiation
Authors: Clement Temaneh-Nyah, Josiah Makiche, Josephine Nujoma
Abstract:
This paper considers the characterisation of a complex electromagnetic environment due to multiple sources of electromagnetic radiation as a five-dimensional surface which can be described by a set of several surface sections including: instant EM field intensity distribution maps at a given frequency and altitude, instantaneous spectrum at a given location in space and the time evolution of the electromagnetic field spectrum at a given point in space. This characterization if done over time can enable the exposure levels of Radio Frequency Radiation at every point in the analysis area to be determined and results interpreted based on comparison of the determined RFR exposure level with the safe guidelines for general public exposure given by recognised body such as the International commission on non-ionising radiation protection (ICNIRP), Institute of Electrical and Electronic Engineers (IEEE), the National Radiation Protection Authority (NRPA).Keywords: complex electromagnetic environment, electric field strength, mathematical models, multiple sources
Procedia PDF Downloads 36822872 An Artificial Intelligence Framework to Forecast Air Quality
Authors: Richard Ren
Abstract:
Air pollution is a serious danger to international well-being and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Keywords: air quality prediction, air pollution, artificial intelligence, machine learning algorithms
Procedia PDF Downloads 12722871 Numerical Investigation of Pressure Drop and Erosion Wear by Computational Fluid Dynamics Simulation
Authors: Praveen Kumar, Nitin Kumar, Hemant Kumar
Abstract:
The modernization of computer technology and commercial computational fluid dynamic (CFD) simulation has given better detailed results as compared to experimental investigation techniques. CFD techniques are widely used in different field due to its flexibility and performance. Evaluation of pipeline erosion is complex phenomenon to solve by numerical arithmetic technique, whereas CFD simulation is an easy tool to resolve that type of problem. Erosion wear behaviour due to solid–liquid mixture in the slurry pipeline has been investigated using commercial CFD code in FLUENT. Multi-phase Euler-Lagrange model was adopted to predict the solid particle erosion wear in 22.5° pipe bend for the flow of bottom ash-water suspension. The present study addresses erosion prediction in three dimensional 22.5° pipe bend for two-phase (solid and liquid) flow using finite volume method with standard k-ε turbulence, discrete phase model and evaluation of erosion wear rate with varying velocity 2-4 m/s. The result shows that velocity of solid-liquid mixture found to be highly dominating parameter as compared to solid concentration, density, and particle size. At low velocity, settling takes place in the pipe bend due to low inertia and gravitational effect on solid particulate which leads to high erosion at bottom side of pipeline.Keywords: computational fluid dynamics (CFD), erosion, slurry transportation, k-ε Model
Procedia PDF Downloads 40822870 Preparation of Wool Fiber/Keratin/PVA Film and Study on Their Structure and Properties
Authors: Min Wu, Shuming Shen, Xuhong Yang, Rencheng Tang
Abstract:
Every year, numerous organic wastes from fiber byproducts of the wool textile industry, poor quality raw wools not fit for spinning, horns, nails and feathers from butchery are disposed. These wastes are abundant in keratin which is a renewable material. Wool fiber/keratin/PVA composites with different proportions were prepared in this study, and the influence of the proportions on their structure and properties were studied, aiming to understand the potential application of keratin in the field of biomedicine, degradable wrapper, and cosmetics film, and provide a new way to reuse keratin wastes. The urea / sodium sulfide / sodium dodecyl sulfate (SDS) method was used to dissolve the wool. After filtration and dialysis, the wool keratin solution was achieved. Then the keratin solution and polyvinal (PVA) solution were blended in different proportions, and the wool fibers cut into a certain length were cast into the blended solution. Thereby, various wool fiber/keratin/PVA composite films with different proportions were formed through pouring the solution into a flat box and drying at room temperature. The surface morphology, molecular structure, and mechanical property of the composite films were studied. The results showed that, there are α-helix structure, β-sheet and random coil conformations in the pure keratin film, as well as in the wool fiber. Compared with wool fiber, the crystallinity of keratin decreased. PVA can obviously improve the mechanical property of the blended film. When the blended ratio of keratin and PVA is 20:80, the mechanical property of the blended film is greatly improved. The composite films with 8%-16% of wool fibers have better flexibility than those without wool fibers.Keywords: composite film, keratin, mechanical property, morphological structure, PVA, wool fiber
Procedia PDF Downloads 28622869 Accelerating Quantum Chemistry Calculations: Machine Learning for Efficient Evaluation of Electron-Repulsion Integrals
Authors: Nishant Rodrigues, Nicole Spanedda, Chilukuri K. Mohan, Arindam Chakraborty
Abstract:
A crucial objective in quantum chemistry is the computation of the energy levels of chemical systems. This task requires electron-repulsion integrals as inputs, and the steep computational cost of evaluating these integrals poses a major numerical challenge in efficient implementation of quantum chemical software. This work presents a moment-based machine-learning approach for the efficient evaluation of electron-repulsion integrals. These integrals were approximated using linear combinations of a small number of moments. Machine learning algorithms were applied to estimate the coefficients in the linear combination. A random forest approach was used to identify promising features using a recursive feature elimination approach, which performed best for learning the sign of each coefficient but not the magnitude. A neural network with two hidden layers were then used to learn the coefficient magnitudes along with an iterative feature masking approach to perform input vector compression, identifying a small subset of orbitals whose coefficients are sufficient for the quantum state energy computation. Finally, a small ensemble of neural networks (with a median rule for decision fusion) was shown to improve results when compared to a single network.Keywords: quantum energy calculations, atomic orbitals, electron-repulsion integrals, ensemble machine learning, random forests, neural networks, feature extraction
Procedia PDF Downloads 11422868 The Links between Cardiovascular Risk and Psychological Wellbeing in Elderly
Authors: Laura Sapranaviciute-Zabazlajeva, Abdonas Tamosiunas, Dalia Luksiene, Dalia Virviciute
Abstract:
The cardiovascular diseases (CVD) is the leading cause of death in the EU, especially in the middle aged and elderly population. Psychological wellbeing (PWB) has been linked with better cardiovascular health and survival in the elderly. The aim of the study is to evaluate associations between CVD risk and PWB in middle-aged and elderly population. 10,940 middle aged and older Lithuanians of age 45-74 years, were invited to participate in the study. A study sample was a random and stratified by gender and age. In 2006-2008 7,087 responders participated in the survey, so the response rate was 64.8%. A follow-up study was conducted from 2006 till 2015. New CVD cases and deaths from CVD were evaluated using the Kaunas population-based CVD register and death register of Kaunas. Study results revealed that good PWB predicts longer life in female participants (Log Rank = 13.7, p < 0.001). In the fully adjusted model for socio-demographic, social and CVD risk factors, hazard ratio for CVD mortality risk was lower amongst women with good PWB (HR = 0.28, 95% CI 0.11-0.72), but not significantly for men. Our study concludes, that lower CVD mortality rates is being associated with better PWB in female aged 45-74 years.Keywords: psychological well-being, cardiovascular disease, elderly, survival
Procedia PDF Downloads 36122867 Highly Conductive Polycrystalline Metallic Ring in a Magnetic Field
Authors: Isao Tomita
Abstract:
Electrical conduction in a quasi-one-dimensional polycrystalline metallic ring with a long electron phase coherence length realized at low temperature is investigated. In this situation, the wave nature of electrons is important in the ring, where the electrical current I can be induced by a vector potential that arises from a static magnetic field applied perpendicularly to the ring’s area. It is shown that if the average grain size of the polycrystalline ring becomes large (or comparable to the Fermi wavelength), the electrical current I increases to ~I0, where I0 is a current in a disorder-free ring. The cause of this increasing effect is examined, and this takes place if the electron localization length in the polycrystalline potential increases with increasing grain size, which gives rise to coherent connection of tails of a localized electron wave function in the ring and thus provides highly coherent electrical conduction.Keywords: electrical conduction, electron phase coherence, polycrystalline metal, magnetic field
Procedia PDF Downloads 38822866 Airline Choice Model for Domestic Flights: The Role of Airline Flexibility
Authors: Camila Amin-Puello, Lina Vasco-Diaz, Juan Ramirez-Arias, Claudia Munoz, Carlos Gonzalez-Calderon
Abstract:
Operational flexibility is a fundamental aspect in the field of airlines because although demand is constantly changing, it is the duty of companies to provide a service to users that satisfies their needs in an efficient manner without sacrificing factors such as comfort, safety and other perception variables. The objective of this research is to understand the factors that describe and explain operational flexibility by implementing advanced analytical methods such as exploratory factor analysis and structural equation modeling, examining multiple levels of operational flexibility and understanding how these variable influences users' decision-making when choosing an airline and in turn how it affects the airlines themselves. The use of a hybrid model and latent variables improves the efficiency and accuracy of airline performance prediction in the unpredictable Colombian market. This pioneering study delves into traveler motivations and their impact on domestic flight demand, offering valuable insights to optimize resources and improve the overall traveler experience. Applying the methods, it was identified that low-cost airlines are not useful for flexibility, while users, especially women, found airlines with greater flexibility in terms of ticket costs and flight schedules to be more useful. All of this allows airlines to anticipate and adapt to their customers' needs efficiently: to plan flight capacity appropriately, adjust pricing strategies and improve the overall passenger experience.Keywords: hybrid choice model, airline, business travelers, domestic flights
Procedia PDF Downloads 1322865 Analysis of a IncResU-Net Model for R-Peak Detection in ECG Signals
Authors: Beatriz Lafuente Alcázar, Yash Wani, Amit J. Nimunkar
Abstract:
Cardiovascular Diseases (CVDs) are the leading cause of death globally, and around 80% of sudden cardiac deaths are due to arrhythmias or irregular heartbeats. The majority of these pathologies are revealed by either short-term or long-term alterations in the electrocardiogram (ECG) morphology. The ECG is the main diagnostic tool in cardiology. It is a non-invasive, pain free procedure that measures the heart’s electrical activity and that allows the detecting of abnormal rhythms and underlying conditions. A cardiologist can diagnose a wide range of pathologies based on ECG’s form alterations, but the human interpretation is subjective and it is contingent to error. Moreover, ECG records can be quite prolonged in time, which can further complicate visual diagnosis, and deeply retard disease detection. In this context, deep learning methods have risen as a promising strategy to extract relevant features and eliminate individual subjectivity in ECG analysis. They facilitate the computation of large sets of data and can provide early and precise diagnoses. Therefore, the cardiology field is one of the areas that can most benefit from the implementation of deep learning algorithms. In the present study, a deep learning algorithm is trained following a novel approach, using a combination of different databases as the training set. The goal of the algorithm is to achieve the detection of R-peaks in ECG signals. Its performance is further evaluated in ECG signals with different origins and features to test the model’s ability to generalize its outcomes. Performance of the model for detection of R-peaks for clean and noisy ECGs is presented. The model is able to detect R-peaks in the presence of various types of noise, and when presented with data, it has not been trained. It is expected that this approach will increase the effectiveness and capacity of cardiologists to detect divergences in the normal cardiac activity of their patients.Keywords: arrhythmia, deep learning, electrocardiogram, machine learning, R-peaks
Procedia PDF Downloads 18622864 Analysis of the Engineering Judgement Influence on the Selection of Geotechnical Parameters Characteristic Values
Authors: K. Ivandic, F. Dodigovic, D. Stuhec, S. Strelec
Abstract:
A characteristic value of certain geotechnical parameter results from an engineering assessment. Its selection has to be based on technical principles and standards of engineering practice. It has been shown that the results of engineering assessment of different authors for the same problem and input data are significantly dispersed. A survey was conducted in which participants had to estimate the force that causes a 10 cm displacement at the top of a axially in-situ compressed pile. Fifty experts from all over the world took part in it. The lowest estimated force value was 42% and the highest was 133% of measured force resulting from a mentioned static pile load test. These extreme values result in significantly different technical solutions to the same engineering task. In case of selecting a characteristic value of a geotechnical parameter the importance of the influence of an engineering assessment can be reduced by using statistical methods. An informative annex of Eurocode 1 prescribes the method of selecting the characteristic values of material properties. This is followed by Eurocode 7 with certain specificities linked to selecting characteristic values of geotechnical parameters. The paper shows the procedure of selecting characteristic values of a geotechnical parameter by using a statistical method with different initial conditions. The aim of the paper is to quantify an engineering assessment in the example of determining a characteristic value of a specific geotechnical parameter. It is assumed that this assessment is a random variable and that its statistical features will be determined. For this purpose, a survey research was conducted among relevant experts from the field of geotechnical engineering. Conclusively, the results of the survey and the application of statistical method were compared.Keywords: characteristic values, engineering judgement, Eurocode 7, statistical methods
Procedia PDF Downloads 29622863 A Communication Signal Recognition Algorithm Based on Holder Coefficient Characteristics
Authors: Hui Zhang, Ye Tian, Fang Ye, Ziming Guo
Abstract:
Communication signal modulation recognition technology is one of the key technologies in the field of modern information warfare. At present, communication signal automatic modulation recognition methods are mainly divided into two major categories. One is the maximum likelihood hypothesis testing method based on decision theory, the other is a statistical pattern recognition method based on feature extraction. Now, the most commonly used is a statistical pattern recognition method, which includes feature extraction and classifier design. With the increasingly complex electromagnetic environment of communications, how to effectively extract the features of various signals at low signal-to-noise ratio (SNR) is a hot topic for scholars in various countries. To solve this problem, this paper proposes a feature extraction algorithm for the communication signal based on the improved Holder cloud feature. And the extreme learning machine (ELM) is used which aims at the problem of the real-time in the modern warfare to classify the extracted features. The algorithm extracts the digital features of the improved cloud model without deterministic information in a low SNR environment, and uses the improved cloud model to obtain more stable Holder cloud features and the performance of the algorithm is improved. This algorithm addresses the problem that a simple feature extraction algorithm based on Holder coefficient feature is difficult to recognize at low SNR, and it also has a better recognition accuracy. The results of simulations show that the approach in this paper still has a good classification result at low SNR, even when the SNR is -15dB, the recognition accuracy still reaches 76%.Keywords: communication signal, feature extraction, Holder coefficient, improved cloud model
Procedia PDF Downloads 15622862 Mobile Augmented Reality for Collaboration in Operation
Authors: Chong-Yang Qiao
Abstract:
Mobile augmented reality (MAR) tracking targets from the surroundings and aids operators for interactive data and procedures visualization, potential equipment and system understandably. Operators remotely communicate and coordinate with each other for the continuous tasks, information and data exchange between control room and work-site. In the routine work, distributed control system (DCS) monitoring and work-site manipulation require operators interact in real-time manners. The critical question is the improvement of user experience in cooperative works through applying Augmented Reality in the traditional industrial field. The purpose of this exploratory study is to find the cognitive model for the multiple task performance by MAR. In particular, the focus will be on the comparison between different tasks and environment factors which influence information processing. Three experiments use interface and interaction design, the content of start-up, maintenance and stop embedded in the mobile application. With the evaluation criteria of time demands and human errors, and analysis of the mental process and the behavior action during the multiple tasks, heuristic evaluation was used to find the operators performance with different situation factors, and record the information processing in recognition, interpretation, judgment and reasoning. The research will find the functional properties of MAR and constrain the development of the cognitive model. Conclusions can be drawn that suggest MAR is easy to use and useful for operators in the remote collaborative works.Keywords: mobile augmented reality, remote collaboration, user experience, cognition model
Procedia PDF Downloads 19722861 Ontology for Cross-Site-Scripting (XSS) Attack in Cybersecurity
Authors: Jean Rosemond Dora, Karol Nemoga
Abstract:
In this work, we tackle a frequent problem that frequently occurs in the cybersecurity field which is the exploitation of websites by XSS attacks, which are nowadays considered a complicated attack. These types of attacks aim to execute malicious scripts in a web browser of the client by including code in a legitimate web page. A serious matter is when a website accepts the “user-input” option. Attackers can exploit the web application (if vulnerable), and then steal sensitive data (session cookies, passwords, credit cards, etc.) from the server and/or from the client. However, the difficulty of the exploitation varies from website to website. Our focus is on the usage of ontology in cybersecurity against XSS attacks, on the importance of the ontology, and its core meaning for cybersecurity. We explain how a vulnerable website can be exploited, and how different JavaScript payloads can be used to detect vulnerabilities. We also enumerate some tools to use for an efficient analysis. We present detailed reasoning on what can be done to improve the security of a website in order to resist attacks, and we provide supportive examples. Then, we apply an ontology model against XSS attacks to strengthen the protection of a web application. However, we note that the existence of ontology does not improve the security itself, but it has to be properly used and should require a maximum of security layers to be taken into account.Keywords: cybersecurity, web application vulnerabilities, cyber threats, ontology model
Procedia PDF Downloads 17222860 Multinomial Dirichlet Gaussian Process Model for Classification of Multidimensional Data
Authors: Wanhyun Cho, Soonja Kang, Sanggoon Kim, Soonyoung Park
Abstract:
We present probabilistic multinomial Dirichlet classification model for multidimensional data and Gaussian process priors. Here, we have considered an efficient computational method that can be used to obtain the approximate posteriors for latent variables and parameters needed to define the multiclass Gaussian process classification model. We first investigated the process of inducing a posterior distribution for various parameters and latent function by using the variational Bayesian approximations and important sampling method, and next we derived a predictive distribution of latent function needed to classify new samples. The proposed model is applied to classify the synthetic multivariate dataset in order to verify the performance of our model. Experiment result shows that our model is more accurate than the other approximation methods.Keywords: multinomial dirichlet classification model, Gaussian process priors, variational Bayesian approximation, importance sampling, approximate posterior distribution, marginal likelihood evidence
Procedia PDF Downloads 44422859 The Use of Haar Wavelet Mother Signal Tool for Performance Analysis Response of Distillation Column (Application to Moroccan Case Study)
Authors: Mahacine Amrani
Abstract:
This paper aims at reviewing some Moroccan industrial applications of wavelet especially in the dynamic identification of a process model using Haar wavelet mother response. Two recent Moroccan study cases are described using dynamic data originated by a distillation column and an industrial polyethylene process plant. The purpose of the wavelet scheme is to build on-line dynamic models. In both case studies, a comparison is carried out between the Haar wavelet mother response model and a linear difference equation model. Finally it concludes, on the base of the comparison of the process performances and the best responses, which may be useful to create an estimated on-line internal model control and its application towards model-predictive controllers (MPC). All calculations were implemented using AutoSignal Software.Keywords: process performance, model, wavelets, Haar, Moroccan
Procedia PDF Downloads 31722858 The Effect of Emotional Intelligence on Performance and Motivation of Staff: A Case Study of East Azerbaijan Red Crescent
Authors: Bahram Asghari Aghdam, Ali Mahjoub
Abstract:
The purpose of this study is to evaluate the effect of emotional intelligence on the motivation and performance of East Azarbaijan the Red Crescent staff. In this study, EI is determined as the independent variable component of self awareness, self management, social awareness, and relations management, motivation and performance as dependent variables. The research method is descriptive-survey. In this study, simple random sampling method is used and research sample consists of 130 East Azarbaijan the Red Crescent staff that uses Cochran's formula 100 of them were selected and questionnaires were filled by them. Three types of questionnaires were used in this study for emotional intelligence, consisting of the Bradbury Travis and Jane Greaves standard questionnaire; and for motivation and performance a questionnaire is regulated by the researcher with help of professionals and experts in this field that consists of 33 questions about the motivation and 15 questions about performance and content validity were used to obtain the necessary credit. Reliability by using the Cronbach's alpha coefficient /948 was approved. Also, in this study to test the hypothesis of the Spearman correlation coefficient and linear regressions and determine fitness of variables' of structural equation modeling is used. The results show that emotional intelligence with coefficient /865, motivation and performance of in East Azerbaijan the Red Crescent employees has a positive effect. Based on Friedman Test ranking the most influence in motivation and performance of staff in respondents' opinion is in order of self-awareness, relations management, social awareness and self-management.Keywords: emotional intelligence, self-awareness, self-management, social awareness, relations management, motivation, performance
Procedia PDF Downloads 48522857 Resistive Instability in a Multi Ions Hall Thrusters Plasma
Authors: Sukhmander Singh
Abstract:
Hall thrusters are preferred over chemical thrusters because of its high exhaust velocity (around 10 times higher) and high specific impulse. The propellant Xenon is ionized inside the channel and controlled by the magnetic field. The strength of the magnetic field is such that only electrons get magnetized and ions remain unmagnetized because of larger Larmor radius as compared with the length of the channel of the device. There is quite a possibility of the existence of multi ions in a Hall thruster plasma because of dust contribution or another process which take place in the chamber. In this paper, we have derived the dispersion relation for multi ions resistive instability in a hall plasma. The analytical approach is also used to find out the propagating speed and the growth rate of the instability. In addition, some growing waves are also found to exist in the plasma. The dispersion relation is solved numerically to see the behavior of the instability with the plasma parameters viz, the temperature of plasma species, wave number, drift velocity, collision frequency, magnetic field.Keywords: instability, resisitive, thrusters, waves
Procedia PDF Downloads 31322856 Modeling Core Flooding Experiments for Co₂ Geological Storage Applications
Authors: Avinoam Rabinovich
Abstract:
CO₂ geological storage is a proven technology for reducing anthropogenic carbon emissions, which is paramount for achieving the ambitious net zero emissions goal. Core flooding experiments are an important step in any CO₂ storage project, allowing us to gain information on the flow of CO₂ and brine in the porous rock extracted from the reservoir. This information is important for understanding basic mechanisms related to CO₂ geological storage as well as for reservoir modeling, which is an integral part of a field project. In this work, a different method for constructing accurate models of CO₂-brine core flooding will be presented. Results for synthetic cases and real experiments will be shown and compared with numerical models to exhibit their predictive capabilities. Furthermore, the various mechanisms which impact the CO₂ distribution and trapping in the rock samples will be discussed, and examples from models and experiments will be provided. The new method entails solving an inverse problem to obtain a three-dimensional permeability distribution which, along with the relative permeability and capillary pressure functions, constitutes a model of the flow experiments. The model is more accurate when data from a number of experiments are combined to solve the inverse problem. This model can then be used to test various other injection flow rates and fluid fractions which have not been tested in experiments. The models can also be used to bridge the gap between small-scale capillary heterogeneity effects (sub-core and core scale) and large-scale (reservoir scale) effects, known as the upscaling problem.Keywords: CO₂ geological storage, residual trapping, capillary heterogeneity, core flooding, CO₂-brine flow
Procedia PDF Downloads 7022855 3D Receiver Operator Characteristic Histogram
Authors: Xiaoli Zhang, Xiongfei Li, Yuncong Feng
Abstract:
ROC curves, as a widely used evaluating tool in machine learning field, are the tradeoff of true positive rate and negative rate. However, they are blamed for ignoring some vital information in the evaluation process, such as the amount of information about the target that each instance carries, predicted score given by each classification model to each instance. Hence, in this paper, a new classification performance method is proposed by extending the Receiver Operator Characteristic (ROC) curves to 3D space, which is denoted as 3D ROC Histogram. In the histogram, theKeywords: classification, performance evaluation, receiver operating characteristic histogram, hardness prediction
Procedia PDF Downloads 31422854 Performance Comparison of Different Regression Methods for a Polymerization Process with Adaptive Sampling
Authors: Florin Leon, Silvia Curteanu
Abstract:
Developing complete mechanistic models for polymerization reactors is not easy, because complex reactions occur simultaneously; there is a large number of kinetic parameters involved and sometimes the chemical and physical phenomena for mixtures involving polymers are poorly understood. To overcome these difficulties, empirical models based on sampled data can be used instead, namely regression methods typical of machine learning field. They have the ability to learn the trends of a process without any knowledge about its particular physical and chemical laws. Therefore, they are useful for modeling complex processes, such as the free radical polymerization of methyl methacrylate achieved in a batch bulk process. The goal is to generate accurate predictions of monomer conversion, numerical average molecular weight and gravimetrical average molecular weight. This process is associated with non-linear gel and glass effects. For this purpose, an adaptive sampling technique is presented, which can select more samples around the regions where the values have a higher variation. Several machine learning methods are used for the modeling and their performance is compared: support vector machines, k-nearest neighbor, k-nearest neighbor and random forest, as well as an original algorithm, large margin nearest neighbor regression. The suggested method provides very good results compared to the other well-known regression algorithms.Keywords: batch bulk methyl methacrylate polymerization, adaptive sampling, machine learning, large margin nearest neighbor regression
Procedia PDF Downloads 305