Search results for: NARX (Nonlinear Autoregressive Exogenous Model)
14018 Thermodynamic Properties of Binary Gold-Rare Earth Compounds (Au-RE)
Authors: H. Krarchaa, A. Ferroudj
Abstract:
This work presents the results of thermodynamic properties of intermetallic rare earth-gold compounds at different stoichiometric structures. It mentions the existence of the AuRE AuRE2, Au2RE, Au51RE14, Au6RE, Au3RE and Au4RE phases in the majority of Au-RE phase diagrams. It's observed that equiatomic composition is a common compound for all gold rare earth alloys and it has the highest melting temperature. Enthalpies of the formation of studied compounds are calculated based on a new reformulation of Miedema’s model.Keywords: rare earth element, enthalpy of formation, thermodynamic properties, macroscopic model
Procedia PDF Downloads 2814017 3D CFD Modelling of the Airflow and Heat Transfer in Cold Room Filled with Dates
Authors: Zina Ghiloufi, Tahar Khir
Abstract:
A transient three-dimensional computational fluid dynamics (CFD) model is developed to determine the velocity and temperature distribution in different positions cold room during pre-cooling of dates. The turbulence model used is the k-ω Shear Stress Transport (SST) with the standard wall function, the air. The numerical results obtained show that cooling rate is not uniform inside the room; the product at the medium of room has a slower cooling rate. This cooling heterogeneity has a large effect on the energy consumption during cold storage.Keywords: CFD, cold room, cooling rate, dDates, numerical simulation, k-ω (SST)
Procedia PDF Downloads 23714016 Real Time Data Communication with FlightGear Using Simulink Over a UDP Protocol
Authors: Adil Loya, Ali Haider, Arslan A. Ghaffor, Abubaker Siddique
Abstract:
Simulation and modelling of Unmanned Aero Vehicle (UAV) has gained wide popularity in front of aerospace community. The demand of designing and modelling optimized control system for UAV has increased ten folds since last decade. The reason is next generation warfare is dependent on unmanned technologies. Therefore, this research focuses on the simulation of nonlinear UAV dynamics on Simulink and its integration with Flightgear. There has been lots of research on implementation of optimizing control using Simulink, however, there are fewer known techniques to simulate these dynamics over Flightgear and a tedious technique of acquiring data has been tackled in this research horizon. Sending data to Flightgear is easy but receiving it from Simulink is not that straight forward, i.e. we can only receive control data on the output. However, in this research we have managed to get the data out from the Flightgear by implementation of level 2 s-function block within Simulink. Moreover, the results captured from Flightgear over a Universal Datagram Protocol (UDP) communication are then compared with the attitude signal that were sent previously. This provide useful information regarding the difference in outputs attained from Simulink to Flightgear. It was found that values received on Simulink were in high agreement with that of the Flightgear output. And complete study has been conducted in a discrete way.Keywords: aerospace, flight control, flightgear, communication, Simulink
Procedia PDF Downloads 29014015 Tests for Zero Inflation in Count Data with Measurement Error in Covariates
Authors: Man-Yu Wong, Siyu Zhou, Zhiqiang Cao
Abstract:
In quality of life, health service utilization is an important determinant of medical resource expenditures on Colorectal cancer (CRC) care, a better understanding of the increased utilization of health services is essential for optimizing the allocation of healthcare resources to services and thus for enhancing the service quality, especially for high expenditure on CRC care like Hong Kong region. In assessing the association between the health-related quality of life (HRQOL) and health service utilization in patients with colorectal neoplasm, count data models can be used, which account for over dispersion or extra zero counts. In our data, the HRQOL evaluation is a self-reported measure obtained from a questionnaire completed by the patients, misreports and variations in the data are inevitable. Besides, there are more zero counts from the observed number of clinical consultations (observed frequency of zero counts = 206) than those from a Poisson distribution with mean equal to 1.33 (expected frequency of zero counts = 156). This suggests that excess of zero counts may exist. Therefore, we study tests for detecting zero-inflation in models with measurement error in covariates. Method: Under classical measurement error model, the approximate likelihood function for zero-inflation Poisson regression model can be obtained, then Approximate Maximum Likelihood Estimation(AMLE) can be derived accordingly, which is consistent and asymptotically normally distributed. By calculating score function and Fisher information based on AMLE, a score test is proposed to detect zero-inflation effect in ZIP model with measurement error. The proposed test follows asymptotically standard normal distribution under H0, and it is consistent with the test proposed for zero-inflation effect when there is no measurement error. Results: Simulation results show that empirical power of our proposed test is the highest among existing tests for zero-inflation in ZIP model with measurement error. In real data analysis, with or without considering measurement error in covariates, existing tests, and our proposed test all imply H0 should be rejected with P-value less than 0.001, i.e., zero-inflation effect is very significant, ZIP model is superior to Poisson model for analyzing this data. However, if measurement error in covariates is not considered, only one covariate is significant; if measurement error in covariates is considered, only another covariate is significant. Moreover, the direction of coefficient estimations for these two covariates is different in ZIP regression model with or without considering measurement error. Conclusion: In our study, compared to Poisson model, ZIP model should be chosen when assessing the association between condition-specific HRQOL and health service utilization in patients with colorectal neoplasm. and models taking measurement error into account will result in statistically more reliable and precise information.Keywords: count data, measurement error, score test, zero inflation
Procedia PDF Downloads 28914014 Characteristics of Business Models of Industrial-Internet-of-Things Platforms
Authors: Peter Kress, Alexander Pflaum, Ulrich Loewen
Abstract:
The number of Internet-of-Things (IoT) platforms is steadily increasing across various industries, especially for smart factories, smart homes and smart mobility. Also in the manufacturing industry, the number of Industrial-IoT platforms is growing. Both IT players, start-ups and increasingly also established industry players and small-and-medium-enterprises introduce offerings for the connection of industrial equipment on platforms, enabled by advanced information and communication technology. Beside the offered functionalities, the established ecosystem of partners around a platform is one of the key differentiators to generate a competitive advantage. The key question is how platform operators design the business model around their platform to attract a high number of customers and partners to co-create value for the entire ecosystem. The present research tries to answer this question by determining the key characteristics of business models of successful platforms in the manufacturing industry. To achieve that, the authors selected an explorative qualitative research approach and created an inductive comparative case study. The authors generated valuable descriptive insights of the business model elements (e.g., value proposition, pricing model or partnering model) of various established platforms. Furthermore, patterns across the various cases were identified to derive propositions for the successful design of business models of platforms in the manufacturing industry.Keywords: industrial-internet-of-things, business models, platforms, ecosystems, case study
Procedia PDF Downloads 24614013 A Biophysical Model of CRISPR/Cas9 on- and off-Target Binding for Rational Design of Guide RNAs
Authors: Iman Farasat, Howard M. Salis
Abstract:
The CRISPR/Cas9 system has revolutionized genome engineering by enabling site-directed and high-throughput genome editing, genome insertion, and gene knockdowns in several species, including bacteria, yeast, flies, worms, and human cell lines. This technology has the potential to enable human gene therapy to treat genetic diseases and cancer at the molecular level; however, the current CRISPR/Cas9 system suffers from seemingly sporadic off-target genome mutagenesis that prevents its use in gene therapy. A comprehensive mechanistic model that explains how the CRISPR/Cas9 functions would enable the rational design of the guide-RNAs responsible for target site selection while minimizing unexpected genome mutagenesis. Here, we present the first quantitative model of the CRISPR/Cas9 genome mutagenesis system that predicts how guide-RNA sequences (crRNAs) control target site selection and cleavage activity. We used statistical thermodynamics and law of mass action to develop a five-step biophysical model of cas9 cleavage, and examined it in vivo and in vitro. To predict a crRNA's binding specificities and cleavage rates, we then compiled a nearest neighbor (NN) energy model that accounts for all possible base pairings and mismatches between the crRNA and the possible genomic DNA sites. These calculations correctly predicted crRNA specificity across 5518 sites. Our analysis reveals that cas9 activity and specificity are anti-correlated, and, the trade-off between them is the determining factor in performing an RNA-mediated cleavage with minimal off-targets. To find an optimal solution, we first created a scheme of safe-design criteria for Cas9 target selection by systematic analysis of available high throughput measurements. We then used our biophysical model to determine the optimal Cas9 expression levels and timing that maximizes on-target cleavage and minimizes off-target activity. We successfully applied this approach in bacterial and mammalian cell lines to reduce off-target activity to near background mutagenesis level while maintaining high on-target cleavage rate.Keywords: biophysical model, CRISPR, Cas9, genome editing
Procedia PDF Downloads 40814012 Computing Machinery and Legal Intelligence: Towards a Reflexive Model for Computer Automated Decision Support in Public Administration
Authors: Jacob Livingston Slosser, Naja Holten Moller, Thomas Troels Hildebrandt, Henrik Palmer Olsen
Abstract:
In this paper, we propose a model for human-AI interaction in public administration that involves legal decision-making. Inspired by Alan Turing’s test for machine intelligence, we propose a way of institutionalizing a continuous working relationship between man and machine that aims at ensuring both good legal quality and higher efficiency in decision-making processes in public administration. We also suggest that our model enhances the legitimacy of using AI in public legal decision-making. We suggest that case loads in public administration could be divided between a manual and an automated decision track. The automated decision track will be an algorithmic recommender system trained on former cases. To avoid unwanted feedback loops and biases, part of the case load will be dealt with by both a human case worker and the automated recommender system. In those cases an experienced human case worker will have the role of an evaluator, choosing between the two decisions. This model will ensure that the algorithmic recommender system is not compromising the quality of the legal decision making in the institution. It also enhances the legitimacy of using algorithmic decision support because it provides justification for its use by being seen as superior to human decisions when the algorithmic recommendations are preferred by experienced case workers. The paper outlines in some detail the process through which such a model could be implemented. It also addresses the important issue that legal decision making is subject to legislative and judicial changes and that legal interpretation is context sensitive. Both of these issues requires continuous supervision and adjustments to algorithmic recommender systems when used for legal decision making purposes.Keywords: administrative law, algorithmic decision-making, decision support, public law
Procedia PDF Downloads 21914011 Analysis of Structural Modeling on Digital English Learning Strategy Use
Authors: Gyoomi Kim, Jiyoung Bae
Abstract:
The purpose of this study was to propose a framework that verifies the structural relationships among students’ use of digital English learning strategy (DELS), affective domains, and their individual variables. The study developed a hypothetical model based on previous studies on language learning strategy use as well as digital language learning. The participants were 720 Korean high school students and 430 university students. The instrument was a self-response questionnaire that contained 70 question items based on Oxford’s SILL (Strategy Inventory for Language Learning) as well as the previous studies on language learning strategies in digital learning environment in order to measure DELS and affective domains. The collected data were analyzed through structural equation modeling (SEM). This study used quantitative data analysis procedures: Explanatory factor analysis (EFA) and confirmatory factor analysis (CFA). Firstly, the EFA was conducted in order to verify the hypothetical model; the factor analysis was conducted preferentially to identify the underlying relationships between measured variables of DELS and the affective domain in the EFA process. The hypothetical model was established with six indicators of learning strategies (memory, cognitive, compensation, metacognitive, affective, and social strategies) under the latent variable of the use of DELS. In addition, the model included four indicators (self-confidence, interests, self-regulation, and attitude toward digital learning) under the latent variable of learners’ affective domain. Secondly, the CFA was used to determine the suitability of data and research models, so all data from the present study was used to assess model fits. Lastly, the model also included individual learner factors as covariates and five constructs selected were learners’ gender, the level of English proficiency, the duration of English learning, the period of using digital devices, and previous experience of digital English learning. The results verified from SEM analysis proposed a theoretical model that showed the structural relationships between Korean students’ use of DELS and their affective domains. Therefore, the results of this study help ESL/EFL teachers understand how learners use and develop appropriate learning strategies in digital learning contexts. The pedagogical implication and suggestions for the further study will be also presented.Keywords: Digital English Learning Strategy, DELS, individual variables, learners' affective domains, Structural Equation Modeling, SEM
Procedia PDF Downloads 12614010 The Planner's Pentangle: A Proposal for a 21st-Century Model of Planning for Sustainable Development
Authors: Sonia Hirt
Abstract:
The Planner's Triangle, an oft-cited model that visually defined planning as the search for sustainability to balance the three basic priorities of equity, economy, and environment, has influenced planning theory and practice for a quarter of a century. In this essay, we argue that the triangle requires updating and expansion. Even if planners keep sustainability as their key core aspiration at the center of their imaginary geometry, the triangle's vertices have to be rethought. Planners should move on to a 21st-century concept. We propose a Planner's Pentangle with five basic priorities as vertices of a new conceptual polygon. These five priorities are Wellbeing, Equity, Economy, Environment, and Esthetics (WE⁴). The WE⁴ concept more accurately and fully represents planning’s history. This is especially true in the United States, where public art and public health played pivotal roles in the establishment of the profession in the late 19th and early 20th centuries. It also more accurately represents planning’s future. Both health/wellness and aesthetic concerns are becoming increasingly important in the 21st century. The pentangle can become an effective tool for understanding and visualizing planning's history and present. Planning has a long history of representing urban presents and future as conceptual models in visual form. Such models can play an important role in understanding and shaping practice. For over two decades, one such model, the Planner's Triangle, stood apart as the expression of planning's pursuit for sustainability. But if the model is outdated and insufficiently robust, it can diminish our understanding of planning practice, as well as the appreciation of the profession among non-planners. Thus, we argue for a new conceptual model of what planners do.Keywords: sustainable development, planning for sustainable development, planner's triangle, planner's pentangle, planning and health, planning and art, planning history
Procedia PDF Downloads 14314009 Health Percentage Evaluation for Satellite Electrical Power System Based on Linear Stresses Accumulation Damage Theory
Authors: Lin Wenli, Fu Linchun, Zhang Yi, Wu Ming
Abstract:
To meet the demands of long-life and high-intelligence for satellites, the electrical power system should be provided with self-health condition evaluation capability. Any over-stress events in operations should be recorded. Based on Linear stresses accumulation damage theory, accumulative damage analysis was performed on thermal-mechanical-electrical united stresses for three components including the solar array, the batteries and the power conditioning unit. Then an overall health percentage evaluation model for satellite electrical power system was built. To obtain the accurate quantity for system health percentage, an automatic feedback closed-loop correction method for all coefficients in the evaluation model was present. The evaluation outputs could be referred as taking earlier fault-forecast and interventions for Ground Control Center or Satellites self.Keywords: satellite electrical power system, health percentage, linear stresses accumulation damage, evaluation model
Procedia PDF Downloads 41214008 Spectral Analysis Applied to Variables of Oil Wells Profiling
Authors: Suzana Leitão Russo, Mayara Laysa de Oliveira Silva, José Augusto Andrade Filho, Vitor Hugo Simon
Abstract:
Currently, seismic methods and prospecting methods are commonly applied in the oil industry and, according to the information reported every day; oil is a source of non-renewable energy. It is easier to understand why the ownership of areas of oil extraction is coveted by many nations. It is necessary to think about ways that will enable the maximization of oil production. The technique of spectral analysis can be used to analyze the behavior of the variables already defined in oil well the profile. The main objective is to verify the series dependence of variables, and to model the variables using the frequency domain to observe the model residuals.Keywords: oil, well, spectral analysis, oil extraction
Procedia PDF Downloads 53614007 Study of a Crude Oil Desalting Plant of the National Iranian South Oil Company in Gachsaran by Using Artificial Neural Networks
Authors: H. Kiani, S. Moradi, B. Soltani Soulgani, S. Mousavian
Abstract:
Desalting/dehydration plants (DDP) are often installed in crude oil production units in order to remove water-soluble salts from an oil stream. In order to optimize this process, desalting unit should be modeled. In this research, artificial neural network is used to model efficiency of desalting unit as a function of input parameter. The result of this research shows that the mentioned model has good agreement with experimental data.Keywords: desalting unit, crude oil, neural networks, simulation, recovery, separation
Procedia PDF Downloads 45414006 Forecasting Etching Behavior Silica Sand Using the Design of Experiments Method
Authors: Kefaifi Aissa, Sahraoui Tahar, Kheloufi Abdelkrim, Anas Sabiha, Hannane Farouk
Abstract:
The aim of this study is to show how the Design of Experiments Method (DOE) can be put into use as a practical approach for silica sand etching behavior modeling during its primary step of leaching. In the present work, we have studied etching effect on particle size during a primary step of leaching process on Algerian silica sand with florid acid (HF) at 20% and 30 % during 4 and 8 hours. Therefore, a new purity of the sand is noted depending on the time of leaching. This study was expanded by a numerical approach using a method of experiment design, which shows the influence of each parameter and the interaction between them in the process and approved the obtained experimental results. This model is a predictive approach using hide software. Based on the measured parameters experimentally in the interior of the model, the use of DOE method can make it possible to predict the outside parameters of the model in question and can give us the optimize response without making the experimental measurement.Keywords: acid leaching, design of experiments method(DOE), purity silica, silica etching
Procedia PDF Downloads 28614005 The Main Steamline Break Transient Analysis for Advanced Boiling Water Reactor Using TRACE, PARCS, and SNAP Codes
Authors: H. C. Chang, J. R. Wang, A. L. Ho, S. W. Chen, J. H. Yang, C. Shih, L. C. Wang
Abstract:
To confirm the reactor and containment integrity of the Advanced Boiling Water Reactor (ABWR), we perform the analysis of main steamline break (MSLB) transient by using the TRACE, PARCS, and SNAP codes. The process of the research has four steps. First, the ABWR nuclear power plant (NPP) model is developed by using the above codes. Second, the steady state analysis is performed by using this model. Third, the ABWR model is used to run the analysis of MSLB transient. Fourth, the predictions of TRACE and PARCS are compared with the data of FSAR. The results of TRACE/PARCS and FSAR are similar. According to the TRACE/PARCS results, the reactor and containment integrity of ABWR can be maintained in a safe condition for MSLB.Keywords: advanced boiling water reactor, TRACE, PARCS, SNAP
Procedia PDF Downloads 20914004 Hybrid Direct Numerical Simulation and Large Eddy Simulating Wall Models Approach for the Analysis of Turbulence Entropy
Authors: Samuel Ahamefula
Abstract:
Turbulent motion is a highly nonlinear and complex phenomenon, and its modelling is still very challenging. In this study, we developed a hybrid computational approach to accurately simulate fluid turbulence phenomenon. The focus is coupling and transitioning between Direct Numerical Simulation (DNS) and Large Eddy Simulating Wall Models (LES-WM) regions. In the framework, high-order fidelity fluid dynamical methods are utilized to simulate the unsteady compressible Navier-Stokes equations in the Eulerian format on the unstructured moving grids. The coupling and transitioning of DNS and LES-WM are conducted through the linearly staggered Dirichlet-Neumann coupling scheme. The high-fidelity framework is verified and validated based on namely, DNS ability for capture full range of turbulent scales, giving accurate results and LES-WM efficiency in simulating near-wall turbulent boundary layer by using wall models.Keywords: computational methods, turbulence modelling, turbulence entropy, navier-stokes equations
Procedia PDF Downloads 10314003 A Case Study Demonstrating the Benefits of Low-Carb Eating in an Adult with Latent Autoimmune Diabetes Highlights the Necessity and Effectiveness of These Dietary Therapies
Authors: Jasmeet Kaur, Anup Singh, Shashikant Iyengar, Arun Kumar, Ira Sahay
Abstract:
Latent autoimmune diabetes in adults (LADA) is an irreversible autoimmune disease that affects insulin production. LADA is characterized by the production of Glutamic acid decarboxylase (GAD) antibodies, which is similar to type 1 diabetes. Individuals with LADA may eventually develop overt diabetes and require insulin. In this condition, the pancreas produces little or no insulin, which is a hormone used by the body to allow glucose to enter cells and produce energy. While type 1 diabetes was traditionally associated with children and teenagers, its prevalence has increased in adults as well. LADA is frequently misdiagnosed as type 2 diabetes, especially in adulthood when type 2 diabetes is more common. LADA develops in adulthood, usually after age 30. Managing LADA involves metabolic control with exogenous insulin and prolonging the life of surviving beta cells, thereby slowing the disease's progression. This case study examines the impact of approximately 3 months of low-carbohydrate dietary intervention in a 42-year-old woman with LADA who was initially misdiagnosed as having type 2 diabetes. Her c-peptide was 0.13 and her HbA1c was 9.3% when this trial began. Low-carbohydrate interventions have been shown to improve blood sugar levels, including fasting, post-meal, and random blood sugar levels, as well as haemoglobin levels, blood pressure, energy levels, sleep quality, and satiety levels. The use of low-carbohydrate dietary intervention significantly reduces both hypo- and hyperglycaemia events. During the 3 months of the study, there were 2 to 3 hyperglycaemic events owing to physical stress and a single hypoglycaemic event. Low-carbohydrate dietary therapies lessen insulin dose inaccuracy, which explains why there were fewer hyperglycaemic and hypoglycaemic events. In three months, the glycated haemoglobin (HbA1c) level was reduced from 9.3% to 6.3%. These improvements occur without the need for caloric restriction or physical activity. Stress management was crucial aspect of the treatment plan as stress-induced neuroendocrine hormones can cause immunological dysregulation. Additionally, supplements that support immune system and reduce inflammation were used as part of the treatment during the trial. Long-term studies are needed to track disease development and corroborate the claim that such dietary treatments can prolong the honeymoon phase in LADA. Various factors can contribute to additional autoimmune attacks, so measuring c-peptide is crucial on a regular basis to determine whether insulin levels need to be adjusted.Keywords: autoimmune, diabetes, LADA, low_carb, nutrition
Procedia PDF Downloads 4714002 Improvements in Double Q-Learning for Anomalous Radiation Source Searching
Authors: Bo-Bin Xiaoa, Chia-Yi Liua
Abstract:
In the task of searching for anomalous radiation sources, personnel holding radiation detectors to search for radiation sources may be exposed to unnecessary radiation risk, and automated search using machines becomes a required project. The research uses various sophisticated algorithms, which are double Q learning, dueling network, and NoisyNet, of deep reinforcement learning to search for radiation sources. The simulation environment, which is a 10*10 grid and one shielding wall setting in it, improves the development of the AI model by training 1 million episodes. In each episode of training, the radiation source position, the radiation source intensity, agent position, shielding wall position, and shielding wall length are all set randomly. The three algorithms are applied to run AI model training in four environments where the training shielding wall is a full-shielding wall, a lead wall, a concrete wall, and a lead wall or a concrete wall appearing randomly. The 12 best performance AI models are selected by observing the reward value during the training period and are evaluated by comparing these AI models with the gradient search algorithm. The results show that the performance of the AI model, no matter which one algorithm, is far better than the gradient search algorithm. In addition, the simulation environment becomes more complex, the AI model which applied Double DQN combined Dueling and NosiyNet algorithm performs better.Keywords: double Q learning, dueling network, NoisyNet, source searching
Procedia PDF Downloads 11514001 Subspace Rotation Algorithm for Implementing Restricted Hopfield Network as an Auto-Associative Memory
Authors: Ci Lin, Tet Yeap, Iluju Kiringa
Abstract:
This paper introduces the subspace rotation algorithm (SRA) to train the Restricted Hopfield Network (RHN) as an auto-associative memory. Subspace rotation algorithm is a gradient-free subspace tracking approach based on the singular value decomposition (SVD). In comparison with Backpropagation Through Time (BPTT) on training RHN, it is observed that SRA could always converge to the optimal solution and BPTT could not achieve the same performance when the model becomes complex, and the number of patterns is large. The AUTS case study showed that the RHN model trained by SRA could achieve a better structure of attraction basin with larger radius(in general) than the Hopfield Network(HNN) model trained by Hebbian learning rule. Through learning 10000 patterns from MNIST dataset with RHN models with different number of hidden nodes, it is observed that an several components could be adjusted to achieve a balance between recovery accuracy and noise resistance.Keywords: hopfield neural network, restricted hopfield network, subspace rotation algorithm, hebbian learning rule
Procedia PDF Downloads 12114000 Similar Script Character Recognition on Kannada and Telugu
Authors: Gurukiran Veerapur, Nytik Birudavolu, Seetharam U. N., Chandravva Hebbi, R. Praneeth Reddy
Abstract:
This work presents a robust approach for the recognition of characters in Telugu and Kannada, two South Indian scripts with structural similarities in characters. To recognize the characters exhaustive datasets are required, but there are only a few publicly available datasets. As a result, we decided to create a dataset for one language (source language),train the model with it, and then test it with the target language.Telugu is the target language in this work, whereas Kannada is the source language. The suggested method makes use of Canny edge features to increase character identification accuracy on pictures with noise and different lighting. A dataset of 45,150 images containing printed Kannada characters was created. The Nudi software was used to automatically generate printed Kannada characters with different writing styles and variations. Manual labelling was employed to ensure the accuracy of the character labels. The deep learning models like CNN (Convolutional Neural Network) and Visual Attention neural network (VAN) are used to experiment with the dataset. A Visual Attention neural network (VAN) architecture was adopted, incorporating additional channels for Canny edge features as the results obtained were good with this approach. The model's accuracy on the combined Telugu and Kannada test dataset was an outstanding 97.3%. Performance was better with Canny edge characteristics applied than with a model that solely used the original grayscale images. The accuracy of the model was found to be 80.11% for Telugu characters and 98.01% for Kannada words when it was tested with these languages. This model, which makes use of cutting-edge machine learning techniques, shows excellent accuracy when identifying and categorizing characters from these scripts.Keywords: base characters, modifiers, guninthalu, aksharas, vattakshara, VAN
Procedia PDF Downloads 5513999 Determination of Anchor Lengths by Retaining Walls
Authors: Belabed Lazhar
Abstract:
The dimensioning of the anchored retaining screens passes always by the analysis of their stability. The calculation of anchoring lengths is practically carried out according to the mechanical model suggested by Kranz which is often criticized. The safety is evaluated through the comparison of interior force and external force. The force of anchoring over the length cut behind the failure solid is neglected. The failure surface cuts anchoring in the medium length of sealing. In this article, one proposes a new mechanical model which overcomes these disadvantages (simplifications) and gives interesting results.Keywords: retaining walls, anchoring, stability, mechanical modeling, safety
Procedia PDF Downloads 35213998 Design and Validation of an Aerodynamic Model of the Cessna Citation X Horizontal Stabilizer Using both OpenVSP and Digital Datcom
Authors: Marine Segui, Matthieu Mantilla, Ruxandra Mihaela Botez
Abstract:
This research is the part of a major project at the Research Laboratory in Active Controls, Avionics and Aeroservoelasticity (LARCASE) aiming to improve a Cessna Citation X aircraft cruise performance with an application of the morphing wing technology on its horizontal tail. However, the horizontal stabilizer of the Cessna Citation X turns around its span axis with an angle between -8 and 2 degrees. Within this range, the horizontal stabilizer generates certainly some unwanted drag. To cancel this drag, the LARCASE proposes to trim the aircraft with a horizontal stabilizer equipped by a morphing wing technology. This technology aims to optimize aerodynamic performances by changing the conventional horizontal tail shape during the flight. As a consequence, this technology will be able to generate enough lift on the horizontal tail to balance the aircraft without an unwanted drag generation. To conduct this project, an accurate aerodynamic model of the horizontal tail is firstly required. This aerodynamic model will finally allow precise comparison between a conventional horizontal tail and a morphed horizontal tail results. This paper presents how this aerodynamic model was designed. In this way, it shows how the 2D geometry of the horizontal tail was collected and how the unknown airfoil’s shape of the horizontal tail has been recovered. Finally, the complete horizontal tail airfoil shape was found and a comparison between aerodynamic polar of the real horizontal tail and the horizontal tail found in this paper shows a maximum difference of 0.04 on the lift or the drag coefficient which is very good. Aerodynamic polar data of the aircraft horizontal tail are obtained from the CAE Inc. level D research aircraft flight simulator of the Cessna Citation X.Keywords: aerodynamic, Cessna, citation, coefficient, Datcom, drag, lift, longitudinal, model, OpenVSP
Procedia PDF Downloads 37613997 Transient and Persistent Efficiency Estimation for Electric Grid Utilities Based on Meta-Frontier: Comparative Analysis of China and Japan
Authors: Bai-Chen Xie, Biao Li
Abstract:
With the deepening of international exchanges and investment, the international comparison of power grid firms has become the focus of regulatory authorities. Ignoring the differences in the economic environment, resource endowment, technology, and other aspects of different countries or regions may lead to efficiency bias. Based on the Meta-frontier model, this paper divides China and Japan into two groups by using the data of China and Japan from 2006 to 2020. While preserving the differences between the two countries, it analyzes and compares the efficiency of the transmission and distribution industries of the two countries. Combined with the four-component stochastic frontier model, the efficiency is divided into transient and persistent efficiency. We found that there are obvious differences between the transmission and distribution sectors in China and Japan. On the one hand, the inefficiency of the two countries is mostly caused by long-term and structural problems. The key to improve the efficiency of the two countries is to focus more on solving long-term and structural problems. On the other hand, the long-term and structural problems that cause the inefficiency of the two countries are not the same. Quality factors have different effects on the efficiency of the two countries, and this different effect is captured by the common frontier model but is offset in the overall model. Based on these findings, this paper proposes some targeted policy recommendations.Keywords: transmission and distribution industries, transient efficiency, persistent efficiency, meta-frontier, international comparison
Procedia PDF Downloads 10513996 Complete Ensemble Empirical Mode Decomposition with Adaptive Noise Temporal Convolutional Network for Remaining Useful Life Prediction of Lithium Ion Batteries
Authors: Jing Zhao, Dayong Liu, Shihao Wang, Xinghua Zhu, Delong Li
Abstract:
Uhumanned Underwater Vehicles generally operate in the deep sea, which has its own unique working conditions. Lithium-ion power batteries should have the necessary stability and endurance for use as an underwater vehicle’s power source. Therefore, it is essential to accurately forecast how long lithium-ion batteries will last in order to maintain the system’s reliability and safety. In order to model and forecast lithium battery Remaining Useful Life (RUL), this research suggests a model based on Complete Ensemble Empirical Mode Decomposition with Adaptive noise-Temporal Convolutional Net (CEEMDAN-TCN). In this study, two datasets, NASA and CALCE, which have a specific gap in capacity data fluctuation, are used to verify the model and examine the experimental results in order to demonstrate the generalizability of the concept. The experiments demonstrate the network structure’s strong universality and ability to achieve good fitting outcomes on the test set for various battery dataset types. The evaluation metrics reveal that the CEEMDAN-TCN prediction performance of TCN is 25% to 35% better than that of a single neural network, proving that feature expansion and modal decomposition can both enhance the model’s generalizability and be extremely useful in industrial settings.Keywords: lithium-ion battery, remaining useful life, complete EEMD with adaptive noise, temporal convolutional net
Procedia PDF Downloads 16113995 Green Supply Chain Management and Corporate Performance: The Mediation Mechanism of Information Sharing among Firms
Authors: Seigo Matsuno, Yasuo Uchida, Shozo Tokinaga
Abstract:
This paper proposes and empirically tests a model of the relationships between green supply chain management (GSCM) activities and corporate performance. From the literature review, we identified five constructs, namely, environmental commitment, supplier collaboration, supplier assessment, information sharing among suppliers, and business process improvement. These explanatory variables are used to form a structural model explaining the environmental and economic performance. The model was analyzed using the data from a survey of a sample of manufacturing firms in Japan. The results suggest that the degree of supplier collaboration has an influence on the environmental performance directly. While, the impact of supplier assessment on the environmental performance is mediated by the information sharing and/or business process improvement. And the environmental performance has a positive relationship on the economic performance. Academic and managerial implications of our findings are discussed.Keywords: corporate performance, empirical study, green supply chain management, path modeling
Procedia PDF Downloads 39613994 Optimization of Heat Insulation Structure and Heat Flux Calculation Method of Slug Calorimeter
Authors: Zhu Xinxin, Wang Hui, Yang Kai
Abstract:
Heat flux is one of the most important test parameters in the ground thermal protection test. Slug calorimeter is selected as the main sensor measuring heat flux in arc wind tunnel test due to the convenience and low cost. However, because of excessive lateral heat transfer and the disadvantage of the calculation method, the heat flux measurement error of the slug calorimeter is large. In order to enhance measurement accuracy, the heat insulation structure and heat flux calculation method of slug calorimeter were improved. The heat transfer model of the slug calorimeter was built according to the energy conservation principle. Based on the heat transfer model, the insulating sleeve of the hollow structure was designed, which helped to greatly decrease lateral heat transfer. And the slug with insulating sleeve of hollow structure was encapsulated using a package shell. The improved insulation structure reduced heat loss and ensured that the heat transfer characteristics were almost the same when calibrated and tested. The heat flux calibration test was carried out in arc lamp system for heat flux sensor calibration, and the results show that test accuracy and precision of slug calorimeter are improved greatly. In the meantime, the simulation model of the slug calorimeter was built. The heat flux values in different temperature rise time periods were calculated by the simulation model. The results show that extracting the data of the temperature rise rate as soon as possible can result in a smaller heat flux calculation error. Then the different thermal contact resistance affecting calculation error was analyzed by the simulation model. The contact resistance between the slug and the insulating sleeve was identified as the main influencing factor. The direct comparison calibration correction method was proposed based on only heat flux calibration. The numerical calculation correction method was proposed based on the heat flux calibration and simulation model of slug calorimeter after the simulation model was solved by solving the contact resistance between the slug and the insulating sleeve. The simulation and test results show that two methods can greatly reduce the heat flux measurement error. Finally, the improved slug calorimeter was tested in the arc wind tunnel. And test results show that the repeatability accuracy of improved slug calorimeter is less than 3%. The deviation of measurement value from different slug calorimeters is less than 3% in the same fluid field. The deviation of measurement value between slug calorimeter and Gordon Gage is less than 4% in the same fluid field.Keywords: correction method, heat flux calculation, heat insulation structure, heat transfer model, slug calorimeter
Procedia PDF Downloads 12013993 A Numerical and Experimental Study on Fast Pyrolysis of Single Wood Particle
Authors: Hamid Rezaei, Xiaotao Bi, C. Jim Lim, Anthony Lau, Shahab Sokhansanj
Abstract:
A one-dimensional heat transfer model coupled with the kinetic information has been used to predict the overall pyrolysis mass loss of a single wood particle. The kinetic parameters were determined experimentally and the regime and characteristics of the conversion were evaluated in terms of the particle size and reactor temperature. The order of overall mass loss changed from n=1 at temperatures lower than 350 °C to n=0.5 at temperatures higher that 350 °C. Conversion time analysis showed that particles larger than 0.5 mm were controlled by internal thermal resistances. The valid range of particle size to use the simplified lumped model depends on the fluid temperature around the particles. The critical particle size was 0.6-0.7 mm for the fluid temperature of 500 °C and 0.9-1.0 mm for the fluid temperature of 100 °C. Experimental pyrolysis of moist particles did not show distinct drying and pyrolysis stages. The process was divided into two hypothetical drying and pyrolysis dominated zones and empirical correlations are developed to predict the rate of mass loss in each zone.Keywords: pyrolysis, kinetics, model, single particle
Procedia PDF Downloads 32313992 Effects of Damper Locations and Base Isolators on Seismic Response of a Building Frame
Authors: Azin Shakibabarough, Mojtaba Valinejadshoubi, Ashutosh Bagchi
Abstract:
Structural vibration means repetitive motion that causes fatigue and reduction of the performance of a structure. An earthquake may release high amount of energy that can have adverse effect on all components of a structure. Therefore, decreasing of vibration or maintaining performance of structures such as bridges, dams, roads and buildings is important for life safety and reducing economic loss. When earthquake or any vibration happens, investigation on parts of a structure which sustain the seismic loads is mandatory to provide a safe condition for the occupants. One of the solutions for reducing the earthquake vibration in a structure is using of vibration control devices such as dampers and base isolators. The objective of this study is to investigate the optimal positions of friction dampers and base isolators for better seismic response of 2D frame. For this purpose, a two bay and six story frame with different distribution formats was modeled and some of their responses to earthquake such as inter-story drift, max joint displacement, max axial force and max bending moment were determined and compared using non-linear dynamic analysis.Keywords: fast nonlinear analysis, friction damper, base isolator, seismic vibration control, seismic response
Procedia PDF Downloads 32313991 Leveraging SHAP Values for Effective Feature Selection in Peptide Identification
Authors: Sharon Li, Zhonghang Xia
Abstract:
Post-database search is an essential phase in peptide identification using tandem mass spectrometry (MS/MS) to refine peptide-spectrum matches (PSMs) produced by database search engines. These engines frequently face difficulty differentiating between correct and incorrect peptide assignments. Despite advances in statistical and machine learning methods aimed at improving the accuracy of peptide identification, challenges remain in selecting critical features for these models. In this study, two machine learning models—a random forest tree and a support vector machine—were applied to three datasets to enhance PSMs. SHAP values were utilized to determine the significance of each feature within the models. The experimental results indicate that the random forest model consistently outperformed the SVM across all datasets. Further analysis of SHAP values revealed that the importance of features varies depending on the dataset, indicating that a feature's role in model predictions can differ significantly. This variability in feature selection can lead to substantial differences in model performance, with false discovery rate (FDR) differences exceeding 50% between different feature combinations. Through SHAP value analysis, the most effective feature combinations were identified, significantly enhancing model performance.Keywords: peptide identification, SHAP value, feature selection, random forest tree, support vector machine
Procedia PDF Downloads 3213990 Markov Switching of Conditional Variance
Authors: Josip Arneric, Blanka Skrabic Peric
Abstract:
Forecasting of volatility, i.e. returns fluctuations, has been a topic of interest to portfolio managers, option traders and market makers in order to get higher profits or less risky positions. Based on the fact that volatility is time varying in high frequency data and that periods of high volatility tend to cluster, the most common used models are GARCH type models. As standard GARCH models show high volatility persistence, i.e. integrated behaviour of the conditional variance, it is difficult the predict volatility using standard GARCH models. Due to practical limitations of these models different approaches have been proposed in the literature, based on Markov switching models. In such situations models in which the parameters are allowed to change over time are more appropriate because they allow some part of the model to depend on the state of the economy. The empirical analysis demonstrates that Markov switching GARCH model resolves the problem of excessive persistence and outperforms uni-regime GARCH models in forecasting volatility for selected emerging markets.Keywords: emerging markets, Markov switching, GARCH model, transition probabilities
Procedia PDF Downloads 45613989 Intelligent Agent-Based Model for the 5G mmWave O2I Technology Adoption
Authors: Robert Joseph M. Licup
Abstract:
The deployment of the fifth-generation (5G) mobile system through mmWave frequencies is the new solution in the requirement to provide higher bandwidth readily available for all users. The usage pattern of the mobile users has moved towards either the work from home or online classes set-up because of the pandemic. Previous mobile technologies can no longer meet the high speed, and bandwidth requirement needed, given the drastic shift of transactions to the home. The millimeter-wave (mmWave) underutilized frequency is utilized by the fifth-generation (5G) cellular networks that support multi-gigabit-per-second (Gbps) transmission. However, due to its short wavelengths, high path loss, directivity, blockage sensitivity, and narrow beamwidth are some of the technical challenges that need to be addressed. Different tools, technologies, and scenarios are explored to support network design, accurate channel modeling, implementation, and deployment effectively. However, there is a big challenge on how the consumer will adopt this solution and maximize the benefits offered by the 5G Technology. This research proposes to study the intricacies of technology diffusion, individual attitude, behaviors, and how technology adoption will be attained. The agent based simulation model shaped by the actual applications, technology solution, and related literature was used to arrive at a computational model. The research examines the different attributes, factors, and intricacies that can affect each identified agent towards technology adoption.Keywords: agent-based model, AnyLogic, 5G O21, 5G mmWave solutions, technology adoption
Procedia PDF Downloads 111