Search results for: algebraic code excited linear prediction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6648

Search results for: algebraic code excited linear prediction

6318 Hybrid Approach for Software Defect Prediction Using Machine Learning with Optimization Technique

Authors: C. Manjula, Lilly Florence

Abstract:

Software technology is developing rapidly which leads to the growth of various industries. Now-a-days, software-based applications have been adopted widely for business purposes. For any software industry, development of reliable software is becoming a challenging task because a faulty software module may be harmful for the growth of industry and business. Hence there is a need to develop techniques which can be used for early prediction of software defects. Due to complexities in manual prediction, automated software defect prediction techniques have been introduced. These techniques are based on the pattern learning from the previous software versions and finding the defects in the current version. These techniques have attracted researchers due to their significant impact on industrial growth by identifying the bugs in software. Based on this, several researches have been carried out but achieving desirable defect prediction performance is still a challenging task. To address this issue, here we present a machine learning based hybrid technique for software defect prediction. First of all, Genetic Algorithm (GA) is presented where an improved fitness function is used for better optimization of features in data sets. Later, these features are processed through Decision Tree (DT) classification model. Finally, an experimental study is presented where results from the proposed GA-DT based hybrid approach is compared with those from the DT classification technique. The results show that the proposed hybrid approach achieves better classification accuracy.

Keywords: decision tree, genetic algorithm, machine learning, software defect prediction

Procedia PDF Downloads 303
6317 Machine Learning Techniques to Develop Traffic Accident Frequency Prediction Models

Authors: Rodrigo Aguiar, Adelino Ferreira

Abstract:

Road traffic accidents are the leading cause of unnatural death and injuries worldwide, representing a significant problem of road safety. In this context, the use of artificial intelligence with advanced machine learning techniques has gained prominence as a promising approach to predict traffic accidents. This article investigates the application of machine learning algorithms to develop traffic accident frequency prediction models. Models are evaluated based on performance metrics, making it possible to do a comparative analysis with traditional prediction approaches. The results suggest that machine learning can provide a powerful tool for accident prediction, which will contribute to making more informed decisions regarding road safety.

Keywords: machine learning, artificial intelligence, frequency of accidents, road safety

Procedia PDF Downloads 51
6316 The Development of Large Deformation Stability of Elastomeric Bearings

Authors: Davide Forcellini, James Marshal Kelly

Abstract:

Seismic isolation using multi-layer elastomeric isolators has been used in the United States for more than 20 years. Although isolation bearings normally have a large factor of safety against buckling due to low shear stiffness, this phenomenon has been widely studied. In particular, the linearly elastic theory adopted to study this phenomenon is relatively accurate and adequate for most design purposes. Unfortunately it cannot consider the large deformation response of a bearing when buckling occurs and the unresolved behaviour of the stability of the post-buckled state. The study conducted in this paper may be viewed as a development of the linear theory of multi-layered elastomeric bearing, simply replacing the differential equations by algebraic equations, showing how it is possible to evaluate the post-buckling behaviour and the interactions at large deformations.

Keywords: multi-layer elastomeric isolators, large deformation, compressive load, tensile load, post-buckling behaviour

Procedia PDF Downloads 402
6315 Detecting Earnings Management via Statistical and Neural Networks Techniques

Authors: Mohammad Namazi, Mohammad Sadeghzadeh Maharluie

Abstract:

Predicting earnings management is vital for the capital market participants, financial analysts and managers. The aim of this research is attempting to respond to this query: Is there a significant difference between the regression model and neural networks’ models in predicting earnings management, and which one leads to a superior prediction of it? In approaching this question, a Linear Regression (LR) model was compared with two neural networks including Multi-Layer Perceptron (MLP), and Generalized Regression Neural Network (GRNN). The population of this study includes 94 listed companies in Tehran Stock Exchange (TSE) market from 2003 to 2011. After the results of all models were acquired, ANOVA was exerted to test the hypotheses. In general, the summary of statistical results showed that the precision of GRNN did not exhibit a significant difference in comparison with MLP. In addition, the mean square error of the MLP and GRNN showed a significant difference with the multi variable LR model. These findings support the notion of nonlinear behavior of the earnings management. Therefore, it is more appropriate for capital market participants to analyze earnings management based upon neural networks techniques, and not to adopt linear regression models.

Keywords: earnings management, generalized linear regression, neural networks multi-layer perceptron, Tehran stock exchange

Procedia PDF Downloads 395
6314 Variations of the Modal Characteristics of the Feeding Stage with Different Preloaded Linear Guide

Authors: Jui-Pui Hung, Yong-Run Chen, Wei-Cheng Shih, Chun-Wei Lin

Abstract:

This study was aimed to assess the variations of the modal characteristics of the feeding stage with different linear guide modulus. The dynamic characteristics of the feeding stage were characterized in terms of the modal stiffness, modal frequency and modal damping, which are assessed from the vibration tests. According to the experimental measurements, the actual preload of the linear guide modulus was found to deviate from the rated values as setting in factory. This may be due to the assemblage errors of guide modules. For the stage with linear guides, the dynamic stiffness was affected to change by the preload set on the rolling balls. The variation of the dynamic stiffness at first and second modes is 20.8 and 10.5%, respectively when the linear guide preload is adjusted from medium and high amount. But the modal damping ratio is reduced by 8.97 and 9.65%, respectively. For high-frequency mode, the modal stiffness increases by 171.2% and the damping ratio reduced by 34.4%. Current results demonstrate the importance in the determining the preloaded amount of linear guide modulus in practical application.

Keywords: contact stiffness, feeding stage, linear guides, modal characteristics, pre-load

Procedia PDF Downloads 400
6313 Reliability-Based Codified Design of Concrete Structures

Authors: Naser Alenezi, Ibrahim Alsakkaf, Osama Eid

Abstract:

The main objective of this study is to develop an independent reliability based code for reinforced concrete (R/C) structural components and elements solely for the State of Kuwait and its neighboring countries. The proposed code will take into account the harsh Kuwait’s harsh environment, loading conditions and material strengths. The method for developing such a code is based on structural reliability theory that takes into accounts the specific geographical and the various prescribed societal environment of the Kuwait region. These methods were developed according to the following four components: (1) loads, (2) structural strength, (3) reliability analysis, and (4) achieving target reliability levels (reliability index ’s ). The final product from this study will be a design code for R/C structural elements that include beams and columns, and some other structural members. This reliability-based LRFD design code will provide appropriate, easy, fast, and economical approach for designing R/C structural elements such as, beams and columns, for both houses and bridges, and other concrete structures. In addition, this reliability-based codified design of R/C beams, columns, and, possibly, concrete slabs will improve the design and serviceability of R/C bridge and building systems in Kuwait and neighboring GCC countries. Also, it has the potential to reduce the cost of new concrete structures, as fewer materials are used with more design efficiency.

Keywords: live laod, design, evaluation, structural building

Procedia PDF Downloads 314
6312 Analyzing Students' Writing in an English Code-Mixing Context in Nepali: An Ecological and Systematic Functional Approach

Authors: Binod Duwadi

Abstract:

This article examines the language and literacy practices of English Code-mixing in Nepalese Classroom. Situating the study within an ecological framework, a systematic functional linguistic (SFL) approach was used to analyze students writing in two Neplease schools. Data collection included interviews with teachers, classroom observations, instructional materials, and focal students’ writing samples. Data analyses revealed vastly different language ecologies between the schools owing to sharp socioeconomic stratification, the structural organization of schools, and the pervasiveness of standard language ideology, with stigmatizes English code mixing (ECM) and privileges Standard English in schools. Functional analysis of students’ writing showed that the nature of the writing tasks at the schools created different affordances for exploiting lexicogrammatically choices for meaning making-enhancing them in the case of one school but severely restricting them in the case of another- perpetuating the academic disadvantage for code mixing speakers. Recommendations for structural and attitudinal changes through teacher training and implementation of approaches that engage students’ bidialectal competence for learning are made as important first steps towards addressing educational inequities in Nepalese schools.

Keywords: code-mixing, ecological perspective, systematic functional approach, language and identity

Procedia PDF Downloads 102
6311 Absorbed Dose Measurements for Teletherapy Prediction of Superficial Dose Using Halcyon Linear Accelerator

Authors: Raymond Limen Njinga, Adeneye Samuel Olaolu, Akinyode Ojumoola Ajimo

Abstract:

Introduction: Measurement of entrance dose and dose at different depths is essential to avoid overdose and underdose of patients. The aim of this study is to verify the variation in the absorbed dose using a water-equivalent material. Materials and Methods: The plastic phantom was arranged on the couch of the halcyon linear accelerator by Varian, with the farmer ionization chamber inserted and connected to the electrometer. The image of the setup was taken using the High-Quality Single 1280x1280x16 higher on the service mode to check the alignment with the isocenter. The beam quality TPR₂₀,₁₀ (Tissue phantom ratio) was done to check the beam quality of the machine at a field size of 10 cm x 10 cm. The calibration was done using SAD type set-up at a depth of 5 cm. This process was repeated for ten consecutive weeks, and the values were recorded. Results: The results of the beam output for the teletherapy machine were satisfactory and accepted in comparison with the commissioned measurement of 0.62. The beam quality TPR₂₀,₁₀ (Tissue phantom ratio) was reasonable with respect to the beam quality of the machine at a field size of 10 cm x 10 cm. Conclusion: The results of the beam quality and the absorbed dose rate showed a good consistency over the period of ten weeks with the commissioned measurement value.

Keywords: linear accelerator, absorbed dose rate, isocenter, phantom, ionization chamber

Procedia PDF Downloads 29
6310 The Concentration Analysis of CO2 Using ALOHA Code for Kuosheng Nuclear Power Plant

Authors: W. S. Hsu, Y. Chiang, H. C. Chen, J. R. Wang, S. W. Chen, J. H. Yang, C. Shih

Abstract:

Not only radiation materials, but also the normal chemical material stored in the power plant can cause a risk to the residents. In this research, the ALOHA code was used to perform the concentration analysis under the CO2 storage burst or leakage conditions for Kuosheng nuclear power plant (NPP). The Final Safety Analysis Report (FSAR) and data were used in this study. Additionally, the analysis results of ALOHA code were compared with the R.G. 1.78 failure criteria in order to confirm the control room habitability. The comparison results show that the ALOHA result for burst case was 0.923 g/m3 which was below the criteria. However, the ALOHA results for leakage case was 11.3 g/m3.

Keywords: BWR, ALOHA, habitability, Kuosheng

Procedia PDF Downloads 324
6309 Development of a Context Specific Planning Model for Achieving a Sustainable Urban City

Authors: Jothilakshmy Nagammal

Abstract:

This research paper deals with the different case studies, where the Form-Based Codes are adopted in general and the different implementation methods in particular are discussed to develop a method for formulating a new planning model. The organizing principle of the Form-Based Codes, the transect is used to zone the city into various context specific transects. An approach is adopted to develop the new planning model, city Specific Planning Model (CSPM), as a tool to achieve sustainability for any city in general. A case study comparison method in terms of the planning tools used, the code process adopted and the various control regulations implemented in thirty two different cities are done. The analysis shows that there are a variety of ways to implement form-based zoning concepts: Specific plans, a parallel or optional form-based code, transect-based code /smart code, required form-based standards or design guidelines. The case studies describe the positive and negative results from based zoning, Where it is implemented. From the different case studies on the method of the FBC, it is understood that the scale for formulating the Form-Based Code varies from parts of the city to the whole city. The regulating plan is prepared with the organizing principle as the transect in most of the cases. The various implementation methods adopted in these case studies for the formulation of Form-Based Codes are special districts like the Transit Oriented Development (TOD), traditional Neighbourhood Development (TND), specific plan and Street based. The implementation methods vary from mandatory, integrated and floating. To attain sustainability the research takes the approach of developing a regulating plan, using the transect as the organizing principle for the entire area of the city in general in formulating the Form-Based Codes for the selected Special Districts in the study area in specific, street based. Planning is most powerful when it is embedded in the broader context of systemic change and improvement. Systemic is best thought of as holistic, contextualized and stake holder-owned, While systematic can be thought of more as linear, generalisable, and typically top-down or expert driven. The systemic approach is a process that is based on the system theory and system design principles, which are too often ill understood by the general population and policy makers. The system theory embraces the importance of a global perspective, multiple components, interdependencies and interconnections in any system. In addition, the recognition that a change in one part of a system necessarily alters the rest of the system is a cornerstone of the system theory. The proposed regulating plan taking the transect as an organizing principle and Form-Based Codes to achieve sustainability of the city has to be a hybrid code, which is to be integrated within the existing system - A Systemic Approach with a Systematic Process. This approach of introducing a few form based zones into a conventional code could be effective in the phased replacement of an existing code. It could also be an effective way of responding to the near-term pressure of physical change in “sensitive” areas of the community. With this approach and method the new Context Specific Planning Model is created towards achieving sustainability is explained in detail this research paper.

Keywords: context based planning model, form based code, transect, systemic approach

Procedia PDF Downloads 312
6308 Artificial Neural Network Modeling of a Closed Loop Pulsating Heat Pipe

Authors: Vipul M. Patel, Hemantkumar B. Mehta

Abstract:

Technological innovations in electronic world demand novel, compact, simple in design, less costly and effective heat transfer devices. Closed Loop Pulsating Heat Pipe (CLPHP) is a passive phase change heat transfer device and has potential to transfer heat quickly and efficiently from source to sink. Thermal performance of a CLPHP is governed by various parameters such as number of U-turns, orientations, input heat, working fluids and filling ratio. The present paper is an attempt to predict the thermal performance of a CLPHP using Artificial Neural Network (ANN). Filling ratio and heat input are considered as input parameters while thermal resistance is set as target parameter. Types of neural networks considered in the present paper are radial basis, generalized regression, linear layer, cascade forward back propagation, feed forward back propagation; feed forward distributed time delay, layer recurrent and Elman back propagation. Linear, logistic sigmoid, tangent sigmoid and Radial Basis Gaussian Function are used as transfer functions. Prediction accuracy is measured based on the experimental data reported by the researchers in open literature as a function of Mean Absolute Relative Deviation (MARD). The prediction of a generalized regression ANN model with spread constant of 4.8 is found in agreement with the experimental data for MARD in the range of ±1.81%.

Keywords: ANN models, CLPHP, filling ratio, generalized regression, spread constant

Procedia PDF Downloads 260
6307 Calculation of Detection Efficiency of Horizontal Large Volume Source Using Exvol Code

Authors: M. Y. Kang, Euntaek Yoon, H. D. Choi

Abstract:

To calculate the full energy (FE) absorption peak efficiency for arbitrary volume sample, we developed and verified the EXVol (Efficiency calculator for EXtended Voluminous source) code which is based on effective solid angle method. EXVol is possible to describe the source area as a non-uniform three-dimensional (x, y, z) source. And decompose and set it into several sets of volume units. Users can equally divide (x, y, z) coordinate system to calculate the detection efficiency at a specific position of a cylindrical volume source. By determining the detection efficiency for differential volume units, the total radiative absolute distribution and the correction factor of the detection efficiency can be obtained from the nondestructive measurement of the source. In order to check the performance of the EXVol code, Si ingot of 20 cm in diameter and 50 cm in height were used as a source. The detector was moved at the collimation geometry to calculate the detection efficiency at a specific position and compared with the experimental values. In this study, the performance of the EXVol code was extended to obtain the detection efficiency distribution at a specific position in a large volume source.

Keywords: attenuation, EXVol, detection efficiency, volume source

Procedia PDF Downloads 156
6306 Production Radionuclide Therapy 161-Terbium Using by Talys1.6 and Empire 3.2 Codes in Reactions Cyclotron

Authors: Shohreh Rahimi Lascokalayeh, Hasan Yousefnia, Mojtaba Tajik, Samaneh Zolghadri, Bentehoda Abdolhosseini

Abstract:

In this study, the production of terbium-161 as new therapeutic radionuclide was investigated using TALYS1.6& EMPIRE 3.2 codes. For this purpose, cross section for the reactions reactor to produce 161Tb were extracted by mean of this code In the following step, stopping power of the reactions reactor was calculated by SRIM code. The best reaction in the production of 161Tb is160 Gd(d,n)161Tb Production yield of the 161Tb was obtained by utilization of MATLAB calculation code and based on the charged particle reaction formalism.The results showed that Production yield of the 161Tb was obtained 0.8 (mci/ A*h).

Keywords: terbium161, TALYS1.6, EMPIRE3.2, yield, cross-section

Procedia PDF Downloads 400
6305 Performance Analysis of Bluetooth Low Energy Mesh Routing Algorithm in Case of Disaster Prediction

Authors: Asmir Gogic, Aljo Mujcic, Sandra Ibric, Nermin Suljanovic

Abstract:

Ubiquity of natural disasters during last few decades have risen serious questions towards the prediction of such events and human safety. Every disaster regardless its proportion has a precursor which is manifested as a disruption of some environmental parameter such as temperature, humidity, pressure, vibrations and etc. In order to anticipate and monitor those changes, in this paper we propose an overall system for disaster prediction and monitoring, based on wireless sensor network (WSN). Furthermore, we introduce a modified and simplified WSN routing protocol built on the top of the trickle routing algorithm. Routing algorithm was deployed using the bluetooth low energy protocol in order to achieve low power consumption. Performance of the WSN network was analyzed using a real life system implementation. Estimates of the WSN parameters such as battery life time, network size and packet delay are determined. Based on the performance of the WSN network, proposed system can be utilized for disaster monitoring and prediction due to its low power profile and mesh routing feature.

Keywords: bluetooth low energy, disaster prediction, mesh routing protocols, wireless sensor networks

Procedia PDF Downloads 355
6304 On the Existence of Homotopic Mapping Between Knowledge Graphs and Graph Embeddings

Authors: Jude K. Safo

Abstract:

Knowledge Graphs KG) and their relation to Graph Embeddings (GE) represent a unique data structure in the landscape of machine learning (relative to image, text and acoustic data). Unlike the latter, GEs are the only data structure sufficient for representing hierarchically dense, semantic information needed for use-cases like supply chain data and protein folding where the search space exceeds the limits traditional search methods (e.g. page-rank, Dijkstra, etc.). While GEs are effective for compressing low rank tensor data, at scale, they begin to introduce a new problem of ’data retreival’ which we observe in Large Language Models. Notable attempts by transE, TransR and other prominent industry standards have shown a peak performance just north of 57% on WN18 and FB15K benchmarks, insufficient practical industry applications. They’re also limited, in scope, to next node/link predictions. Traditional linear methods like Tucker, CP, PARAFAC and CANDECOMP quickly hit memory limits on tensors exceeding 6.4 million nodes. This paper outlines a topological framework for linear mapping between concepts in KG space and GE space that preserve cardinality. Most importantly we introduce a traceable framework for composing dense linguistic strcutures. We demonstrate performance on WN18 benchmark this model hits. This model does not rely on Large Langauge Models (LLM) though the applications are certainy relevant here as well.

Keywords: representation theory, large language models, graph embeddings, applied algebraic topology, applied knot theory, combinatorics

Procedia PDF Downloads 43
6303 Code Switching and Language Attitudes of Two 10-11 Years Old Bilingual Child

Authors: Kristiina Teiss

Abstract:

Estonians and children having Estonian as a one of their languages have lately become the fastest growing minority or bilingual group in Finland which underlines the importance of studying this target group. The acquisition of bilingualism by an infant is affected by many different issues like the child’s personal traits, language differences, and different environmental factors such as people´s attitudes towards languages and bilingualism. In the early years the most important factor is the children’s interaction with their parents and siblings. This poster gives an overview to the material and some preliminary findings of ongoing PhD study concerning code-mixing, code-switching and language attitudes of two bilingual 10-11 year old children. Data was collected from two different bilingual families, one of them living in Tampere, Finland and one of them moved during the study to Tallinn, Estonia. The data includes audio recordings of the families’ interactions with their children when they were aged 2-3 years old and then when they were 10-11 years old. The data also includes recorded semi-structured queries of the parents, as well as recorded semi-structured queries of the children when they were in the age of 10-11 years. The features of code-mixing can vary depending on norms or models in the families, or even according to its use by two parents in same family. The practices studied in the ongoing longitudinal case study, based on a framework of ethnography, contain parental conversational strategies and family attitudes as well as CS (code-switching and code-mixing) cases occurring both in children and adult language. The aim of this paper is to find out whether there is a connection between children’s attitudes and their daily language use. It would be also interesting to find some evidence, as to whether living in different countries has different impacts on using two languages. The results of dissertation maid give some directional suggestions on how language maintenance of Estonian-Finnish bilinguals could be supported, although generalizations on the base of case study could not be done.

Keywords: code switching, Estonian, Finnish, language attitudes

Procedia PDF Downloads 341
6302 Quick Response Codes in Physio: A Simple Click to Long-Term Oxygen Therapy Education

Authors: K. W. Lee, C. M. Choi, H. C. Tsang, W. K. Fong, Y. K. Cheng, L. Y. Chan, C. K. Yuen, P. W. Lau, Y. L. To, K. C. Chow

Abstract:

QR (Quick Response) Code is a matrix barcode. It enables users to open websites, photos and other information with mobile devices by just snapping the code. In usual Long Term Oxygen Therapy arrangement, piles of LTOT related information like leaflets from different oxygen service providers are given to patients to choose an appropriate plan according to their needs. If these printed materials are transformed into electronic format (QR Code), it would be more environmentally-friendly. More importantly, electronic materials including LTOT equipment operation and dyspnoea relieving techniques also empower patients in long-term disease management. The objective to this study is to investigate the effect of QR code in patient education on new LTOT users. This study was carried out in medical wards of North District Hospital. Adult patients and relatives who followed commands, were able to use smartphones with internet services and required LTOT arrangement on hospital discharge were recruited. In LTOT arrangement, apart from the usual LTOT education booklets which included patients’ personal information (e.g. oxygen titration and six-minute walk test results etc.), extra leaflets consisted of 1. QR codes of LTOT plans from different oxygen service providers, 2. Education materials of dyspnoea management and 3. Instructions on LTOT equipment operation were given. Upon completion of LTOT arrangement, a questionnaire about the use of QR code on patient education was filled in by patients or relatives. A total of 10 new LTOT users were recruited from November 2017 to January 2018. Initially, 70% of them did not know anything about the QR code, but all of them understood its operation after a simple demonstration. 70% of them agreed that it was convenient to use (20% strongly agree, 40% agree, 10% somewhat agree). 80% of them agreed that QR code could facilitate the retrieval of more LTOT related information (10% strongly agree, 70% agree) while 90% agreed that we should continue delivering QR code leaflets to new LTOT users in the future (30% strongly agree, 40% agree, 20% somewhat agree). It is proven that QR code is a convenient and environmentally-friendly tool to deliver information. It is also relatively easy to be introduced to new users. It has received welcoming feedbacks from current users.

Keywords: long-term oxygen therapy, physiotherapy, patient education, QR code

Procedia PDF Downloads 121
6301 Dido: An Automatic Code Generation and Optimization Framework for Stencil Computations on Distributed Memory Architectures

Authors: Mariem Saied, Jens Gustedt, Gilles Muller

Abstract:

We present Dido, a source-to-source auto-generation and optimization framework for multi-dimensional stencil computations. It enables a large programmer community to easily and safely implement stencil codes on distributed-memory parallel architectures with Ordered Read-Write Locks (ORWL) as an execution and communication back-end. ORWL provides inter-task synchronization for data-oriented parallel and distributed computations. It has been proven to guarantee equity, liveness, and efficiency for a wide range of applications, particularly for iterative computations. Dido consists mainly of an implicitly parallel domain-specific language (DSL) implemented as a source-level transformer. It captures domain semantics at a high level of abstraction and generates parallel stencil code that leverages all ORWL features. The generated code is well-structured and lends itself to different possible optimizations. In this paper, we enhance Dido to handle both Jacobi and Gauss-Seidel grid traversals. We integrate temporal blocking to the Dido code generator in order to reduce the communication overhead and minimize data transfers. To increase data locality and improve intra-node data reuse, we coupled the code generation technique with the polyhedral parallelizer Pluto. The accuracy and portability of the generated code are guaranteed thanks to a parametrized solution. The combination of ORWL features, the code generation pattern and the suggested optimizations, make of Dido a powerful code generation framework for stencil computations in general, and for distributed-memory architectures in particular. We present a wide range of experiments over a number of stencil benchmarks.

Keywords: stencil computations, ordered read-write locks, domain-specific language, polyhedral model, experiments

Procedia PDF Downloads 92
6300 A Semidefinite Model to Quantify Dynamic Forces in the Powertrain of Torque Regulated Bascule Bridge Machineries

Authors: Kodo Sektani, Apostolos Tsouvalas, Andrei Metrikine

Abstract:

The reassessment of existing movable bridges in The Netherlands has created the need for acceptance/rejection criteria to assess whether the machineries are meet certain design demands. However, the existing design code defines a different limit state design, meant for new machineries which is based on a simple linear spring-mass model. Observations show that existing bridges do not confirm the model predictions. In fact, movable bridges are nonlinear systems consisting of mechanical components, such as, gears, electric motors and brakes. Next to that, each movable bridge is characterized by a unique set of parameters. However, in the existing code various variables that describe the physical characteristics of the bridge are neglected or replaced by partial factors. For instance, the damping ratio ζ, which is different for drawbridges compared to bascule bridges, is taken as a constant for all bridge types. In this paper, a model is developed that overcomes some of the limitations of existing modelling approaches to capture the dynamics of the powertrain of a class of bridge machineries First, a semidefinite dynamic model is proposed, which accounts for stiffness, damping, and some additional variables of the physical system, which are neglected by the code, such as nonlinear braking torques. The model gives an upper bound of the peak forces/torques occurring in the powertrain during emergency braking. Second, a discrete nonlinear dynamic model is discussed, with realistic motor torque characteristics during normal operation. This model succeeds to accurately predict the full time history of the occurred stress state of the opening and closing cycle for fatigue purposes.

Keywords: Dynamics of movable bridges, Bridge machinery, Powertrains, Torque measurements

Procedia PDF Downloads 123
6299 Machine Learning Techniques in Seismic Risk Assessment of Structures

Authors: Farid Khosravikia, Patricia Clayton

Abstract:

The main objective of this work is to evaluate the advantages and disadvantages of various machine learning techniques in two key steps of seismic hazard and risk assessment of different types of structures. The first step is the development of ground-motion models, which are used for forecasting ground-motion intensity measures (IM) given source characteristics, source-to-site distance, and local site condition for future events. IMs such as peak ground acceleration and velocity (PGA and PGV, respectively) as well as 5% damped elastic pseudospectral accelerations at different periods (PSA), are indicators of the strength of shaking at the ground surface. Typically, linear regression-based models, with pre-defined equations and coefficients, are used in ground motion prediction. However, due to the restrictions of the linear regression methods, such models may not capture more complex nonlinear behaviors that exist in the data. Thus, this study comparatively investigates potential benefits from employing other machine learning techniques as statistical method in ground motion prediction such as Artificial Neural Network, Random Forest, and Support Vector Machine. The results indicate the algorithms satisfy some physically sound characteristics such as magnitude scaling distance dependency without requiring pre-defined equations or coefficients. Moreover, it is shown that, when sufficient data is available, all the alternative algorithms tend to provide more accurate estimates compared to the conventional linear regression-based method, and particularly, Random Forest outperforms the other algorithms. However, the conventional method is a better tool when limited data is available. Second, it is investigated how machine learning techniques could be beneficial for developing probabilistic seismic demand models (PSDMs), which provide the relationship between the structural demand responses (e.g., component deformations, accelerations, internal forces, etc.) and the ground motion IMs. In the risk framework, such models are used to develop fragility curves estimating exceeding probability of damage for pre-defined limit states, and therefore, control the reliability of the predictions in the risk assessment. In this study, machine learning algorithms like artificial neural network, random forest, and support vector machine are adopted and trained on the demand parameters to derive PSDMs for them. It is observed that such models can provide more accurate estimates of prediction in relatively shorter about of time compared to conventional methods. Moreover, they can be used for sensitivity analysis of fragility curves with respect to many modeling parameters without necessarily requiring more intense numerical response-history analysis.

Keywords: artificial neural network, machine learning, random forest, seismic risk analysis, seismic hazard analysis, support vector machine

Procedia PDF Downloads 80
6298 Intelligent Earthquake Prediction System Based On Neural Network

Authors: Emad Amar, Tawfik Khattab, Fatma Zada

Abstract:

Predicting earthquakes is an important issue in the study of geography. Accurate prediction of earthquakes can help people to take effective measures to minimize the loss of personal and economic damage, such as large casualties, destruction of buildings and broken of traffic, occurred within a few seconds. United States Geological Survey (USGS) science organization provides reliable scientific information of Earthquake Existed throughout history & Preliminary database from the National Center Earthquake Information (NEIC) show some useful factors to predict an earthquake in a seismic area like Aleutian Arc in the U.S. state of Alaska. The main advantage of this prediction method that it does not require any assumption, it makes prediction according to the future evolution of object's time series. The article compares between simulation data result from trained BP and RBF neural network versus actual output result from the system calculations. Therefore, this article focuses on analysis of data relating to real earthquakes. Evaluation results show better accuracy and higher speed by using radial basis functions (RBF) neural network.

Keywords: BP neural network, prediction, RBF neural network, earthquake

Procedia PDF Downloads 464
6297 Hybrid Wavelet-Adaptive Neuro-Fuzzy Inference System Model for a Greenhouse Energy Demand Prediction

Authors: Azzedine Hamza, Chouaib Chakour, Messaoud Ramdani

Abstract:

Energy demand prediction plays a crucial role in achieving next-generation power systems for agricultural greenhouses. As a result, high prediction quality is required for efficient smart grid management and therefore low-cost energy consumption. The aim of this paper is to investigate the effectiveness of a hybrid data-driven model in day-ahead energy demand prediction. The proposed model consists of Discrete Wavelet Transform (DWT), and Adaptive Neuro-Fuzzy Inference System (ANFIS). The DWT is employed to decompose the original signal in a set of subseries and then an ANFIS is used to generate the forecast for each subseries. The proposed hybrid method (DWT-ANFIS) was evaluated using a greenhouse energy demand data for a week and compared with ANFIS. The performances of the different models were evaluated by comparing the corresponding values of Mean Absolute Percentage Error (MAPE). It was demonstrated that discret wavelet transform can improve agricultural greenhouse energy demand modeling.

Keywords: wavelet transform, ANFIS, energy consumption prediction, greenhouse

Procedia PDF Downloads 57
6296 Classifying and Predicting Efficiencies Using Interval DEA Grid Setting

Authors: Yiannis G. Smirlis

Abstract:

The classification and the prediction of efficiencies in Data Envelopment Analysis (DEA) is an important issue, especially in large scale problems or when new units frequently enter the under-assessment set. In this paper, we contribute to the subject by proposing a grid structure based on interval segmentations of the range of values for the inputs and outputs. Such intervals combined, define hyper-rectangles that partition the space of the problem. This structure, exploited by Interval DEA models and a dominance relation, acts as a DEA pre-processor, enabling the classification and prediction of efficiency scores, without applying any DEA models.

Keywords: data envelopment analysis, interval DEA, efficiency classification, efficiency prediction

Procedia PDF Downloads 142
6295 A Method and System for Secure Authentication Using One Time QR Code

Authors: Divyans Mahansaria

Abstract:

User authentication is an important security measure for protecting confidential data and systems. However, the vulnerability while authenticating into a system has significantly increased. Thus, necessary mechanisms must be deployed during the process of authenticating a user to safeguard him/her from the vulnerable attacks. The proposed solution implements a novel authentication mechanism to counter various forms of security breach attacks including phishing, Trojan horse, replay, key logging, Asterisk logging, shoulder surfing, brute force search and others. QR code (Quick Response Code) is a type of matrix barcode or two-dimensional barcode that can be used for storing URLs, text, images and other information. In the proposed solution, during each new authentication request, a QR code is dynamically generated and presented to the user. A piece of generic information is mapped to plurality of elements and stored within the QR code. The mapping of generic information with plurality of elements, randomizes in each new login, and thus the QR code generated for each new authentication request is for one-time use only. In order to authenticate into the system, the user needs to decode the QR code using any QR code decoding software. The QR code decoding software needs to be installed on handheld mobile devices such as smartphones, personal digital assistant (PDA), etc. On decoding the QR code, the user will be presented a mapping between the generic piece of information and plurality of elements using which the user needs to derive cipher secret information corresponding to his/her actual password. Now, in place of the actual password, the user will use this cipher secret information to authenticate into the system. The authentication terminal will receive the cipher secret information and use a validation engine that will decipher the cipher secret information. If the entered secret information is correct, the user will be provided access to the system. Usability study has been carried out on the proposed solution, and the new authentication mechanism was found to be easy to learn and adapt. Mathematical analysis of the time taken to carry out brute force attack on the proposed solution has been carried out. The result of mathematical analysis showed that the solution is almost completely resistant to brute force attack. Today’s standard methods for authentication are subject to a wide variety of software, hardware, and human attacks. The proposed scheme can be very useful in controlling the various types of authentication related attacks especially in a networked computer environment where the use of username and password for authentication is common.

Keywords: authentication, QR code, cipher / decipher text, one time password, secret information

Procedia PDF Downloads 246
6294 Comparison of Different Artificial Intelligence-Based Protein Secondary Structure Prediction Methods

Authors: Jamerson Felipe Pereira Lima, Jeane Cecília Bezerra de Melo

Abstract:

The difficulty and cost related to obtaining of protein tertiary structure information through experimental methods, such as X-ray crystallography or NMR spectroscopy, helped raising the development of computational methods to do so. An approach used in these last is prediction of tridimensional structure based in the residue chain, however, this has been proved an NP-hard problem, due to the complexity of this process, explained by the Levinthal paradox. An alternative solution is the prediction of intermediary structures, such as the secondary structure of the protein. Artificial Intelligence methods, such as Bayesian statistics, artificial neural networks (ANN), support vector machines (SVM), among others, were used to predict protein secondary structure. Due to its good results, artificial neural networks have been used as a standard method to predict protein secondary structure. Recent published methods that use this technique, in general, achieved a Q3 accuracy between 75% and 83%, whereas the theoretical accuracy limit for protein prediction is 88%. Alternatively, to achieve better results, support vector machines prediction methods have been developed. The statistical evaluation of methods that use different AI techniques, such as ANNs and SVMs, for example, is not a trivial problem, since different training sets, validation techniques, as well as other variables can influence the behavior of a prediction method. In this study, we propose a prediction method based on artificial neural networks, which is then compared with a selected SVM method. The chosen SVM protein secondary structure prediction method is the one proposed by Huang in his work Extracting Physico chemical Features to Predict Protein Secondary Structure (2013). The developed ANN method has the same training and testing process that was used by Huang to validate his method, which comprises the use of the CB513 protein data set and three-fold cross-validation, so that the comparative analysis of the results can be made comparing directly the statistical results of each method.

Keywords: artificial neural networks, protein secondary structure, protein structure prediction, support vector machines

Procedia PDF Downloads 583
6293 Electrical Load Estimation Using Estimated Fuzzy Linear Parameters

Authors: Bader Alkandari, Jamal Y. Madouh, Ahmad M. Alkandari, Anwar A. Alnaqi

Abstract:

A new formulation of fuzzy linear estimation problem is presented. It is formulated as a linear programming problem. The objective is to minimize the spread of the data points, taking into consideration the type of the membership function of the fuzzy parameters to satisfy the constraints on each measurement point and to insure that the original membership is included in the estimated membership. Different models are developed for a fuzzy triangular membership. The proposed models are applied to different examples from the area of fuzzy linear regression and finally to different examples for estimating the electrical load on a busbar. It had been found that the proposed technique is more suited for electrical load estimation, since the nature of the load is characterized by the uncertainty and vagueness.

Keywords: fuzzy regression, load estimation, fuzzy linear parameters, electrical load estimation

Procedia PDF Downloads 509
6292 Temperature Rises Characteristics of Distinct Double-Sided Flat Permanent Magnet Linear Generator for Free Piston Engines for Hybrid Vehicles

Authors: Ismail Rahama Adam Hamid

Abstract:

This paper presents the development of a thermal model for a flat, double-sided linear generator designed for use in free-piston engines. The study conducted in this paper examines the influence of temperature on the performance of the permeant magnet linear generator, an integral and pivotal component within the system. This research places particular emphasis on the Neodymium Iron Boron (NdFeB) permanent magnet, which serves as a source of magnetic field for the linear generator. In this study, an internal combustion engine that tends to produce heat is connected to a generator. Considering the temperatures rise from both the combustion process and the thermal contributions of current-carrying conductors and frictional forces. Utilizing Computational Fluid Dynamics (CFD) method, a thermal model of the (NdFeB) magnet within the linear generator is constructed and analyzed. Furthermore, the temperature field is examined to ensure that the linear generator operates under stable conditions without the risk of demagnetization.

Keywords: free piston engine, permanent magnet, linear generator, demagnetization, simulation

Procedia PDF Downloads 13
6291 Mathematical Modeling Pressure Losses of Trapezoidal Labyrinth Channel and Bi-Objective Optimization of the Design Parameters

Authors: Nina Philipova

Abstract:

The influence of the geometric parameters of trapezoidal labyrinth channel on the pressure losses along the labyrinth length is investigated in this work. The impact of the dentate height is studied at fixed values of the dentate angle and the dentate spacing. The objective of the work presented in this paper is to derive a mathematical model of the pressure losses along the labyrinth length depending on the dentate height. The numerical simulations of the water flow movement are performed by using Commercial codes ANSYS GAMBIT and FLUENT. Dripper inlet pressure is set up to be 1 bar. As a result, the mathematical model of the pressure losses is determined as a second-order polynomial by means Commercial code STATISTIKA. Bi-objective optimization is performed by using the mean algebraic function of utility. The optimum value of the dentate height is defined at fixed values of the dentate angle and the dentate spacing. The derived model of the pressure losses and the optimum value of the dentate height are used as a basis for a more successful emitter design.

Keywords: drip irrigation, labyrinth channel hydrodynamics, numerical simulations, Reynolds stress model

Procedia PDF Downloads 128
6290 Solving Extended Linear Complementarity Problems (XLCP) - Wood and Environment

Authors: Liberto Pombal, Christian Dieter Jaekel

Abstract:

The objective of this work is to establish theoretical and numerical conditions for Solving Extended Linear Complementarity Problems (XLCP), with emphasis on the Horizontal Linear Complementarity Problem (HLCP). Two new strategies for solving complementarity problems are presented, using differentiable and penalized functions, which resulted in a natural formalization for the Linear Horizontal case. The computational results of all suggested strategies are also discussed in depth in this paper. The implication in practice allows solving and optimizing, in an innovative way, the (forestry) problems of the value chain of the industrial wood sector in Angola.

Keywords: complementarity, box constrained, optimality conditions, wood and environment

Procedia PDF Downloads 23
6289 Nonlinear Estimation Model for Rail Track Deterioration

Authors: M. Karimpour, L. Hitihamillage, N. Elkhoury, S. Moridpour, R. Hesami

Abstract:

Rail transport authorities around the world have been facing a significant challenge when predicting rail infrastructure maintenance work for a long period of time. Generally, maintenance monitoring and prediction is conducted manually. With the restrictions in economy, the rail transport authorities are in pursuit of improved modern methods, which can provide precise prediction of rail maintenance time and location. The expectation from such a method is to develop models to minimize the human error that is strongly related to manual prediction. Such models will help them in understanding how the track degradation occurs overtime under the change in different conditions (e.g. rail load, rail type, rail profile). They need a well-structured technique to identify the precise time that rail tracks fail in order to minimize the maintenance cost/time and secure the vehicles. The rail track characteristics that have been collected over the years will be used in developing rail track degradation prediction models. Since these data have been collected in large volumes and the data collection is done both electronically and manually, it is possible to have some errors. Sometimes these errors make it impossible to use them in prediction model development. This is one of the major drawbacks in rail track degradation prediction. An accurate model can play a key role in the estimation of the long-term behavior of rail tracks. Accurate models increase the track safety and decrease the cost of maintenance in long term. In this research, a short review of rail track degradation prediction models has been discussed before estimating rail track degradation for the curve sections of Melbourne tram track system using Adaptive Network-based Fuzzy Inference System (ANFIS) model.

Keywords: ANFIS, MGT, prediction modeling, rail track degradation

Procedia PDF Downloads 292