Search results for: linear acceleration method
19985 Pushover Analysis of a Typical Bridge Built in Central Zone of Mexico
Authors: Arturo Galvan, Jatziri Y. Moreno-Martinez, Daniel Arroyo-Montoya, Jose M. Gutierrez-Villalobos
Abstract:
Bridges are one of the most seismically vulnerable structures on highway transportation systems. The general process for assessing the seismic vulnerability of a bridge involves the evaluation of its overall capacity and demand. One of the most common procedures to obtain this capacity is by means of pushover analysis of the structure. Typically, the bridge capacity is assessed using non-linear static methods or non-linear dynamic analyses. The non-linear dynamic approaches use step by step numerical solutions for assessing the capacity with the consuming computer time inconvenience. In this study, a nonlinear static analysis (‘pushover analysis’) was performed to predict the collapse mechanism of a typical bridge built in the central zone of Mexico (Celaya, Guanajuato). The bridge superstructure consists of three simple supported spans with a total length of 76 m: 22 m of the length of extreme spans and 32 m of length of the central span. The deck width is of 14 m and the concrete slab depth is of 18 cm. The bridge is built by means of frames of five piers with hollow box-shaped sections. The dimensions of these piers are 7.05 m height and 1.20 m diameter. The numerical model was created using a commercial software considering linear and non-linear elements. In all cases, the piers were represented by frame type elements with geometrical properties obtained from the structural project and construction drawings of the bridge. The deck was modeled with a mesh of rectangular thin shell (plate bending and stretching) finite elements. The moment-curvature analysis was performed for the sections of the piers of the bridge considering in each pier the effect of confined concrete and its reinforcing steel. In this way, plastic hinges were defined on the base of the piers to carry out the pushover analysis. In addition, time history analyses were performed using 19 accelerograms of real earthquakes that have been registered in Guanajuato. In this way, the displacements produced by the bridge were determined. Finally, pushover analysis was applied through the control of displacements in the piers to obtain the overall capacity of the bridge before the failure occurs. It was concluded that the lateral deformation of the piers due to a critical earthquake occurred in this zone is almost imperceptible due to the geometry and reinforcement demanded by the current design standards and compared to its displacement capacity, they were excessive. According to the analysis, it was found that the frames built with five piers increase the rigidity in the transverse direction of the bridge. Hence it is proposed to reduce these frames of five piers to three piers, maintaining the same geometrical characteristics and the same reinforcement in each pier. Also, the mechanical properties of materials (concrete and reinforcing steel) were maintained. Once a pushover analysis was performed considering this configuration, it was concluded that the bridge would continue having a “correct” seismic behavior, at least for the 19 accelerograms considered in this study. In this way, costs in material, construction, time and labor would be reduced in this study case.Keywords: collapse mechanism, moment-curvature analysis, overall capacity, push-over analysis
Procedia PDF Downloads 15219984 Mapping Method to Solve a Nonlinear Schrodinger Type Equation
Authors: Edamana Vasudevan Krishnan
Abstract:
This paper studies solitons in optical materials with the help of Mapping Method. Two types of nonlinear media have been investigated, namely, the cubic nonlinearity and the quintic nonlinearity. The soliton solutions, shock wave solutions and singular solutions have been derives with certain constraint conditions.Keywords: solitons, integrability, metamaterials, mapping method
Procedia PDF Downloads 49419983 The Investigation of Relationship between Accounting Information and the Value of Companies
Authors: Golamhassan Ghahramani Aghdam, Pedram Bavili Tabrizi
Abstract:
The aim of this research is to investigate the relationship between accounting information and the value of the companies accepted in Tehran Exchange Market. The dependent variable in this research is the value of a company that is measured by price coefficients, and the independent variables are balance sheet information, profit and loss information, cash flow state information, and profit quality characteristics. The profit quality characteristic index is to be related and to be on-time. This research is an application research, and the research population includes all companies that are active in Tehran exchange market. The number of 194 companies was selected by the systematic method as the statistics sample in the period of 2018-2019. The multi-variable linear regression model was used for the hypotheses test. The results show that there is no relationship between accounting information and companies’ value (stock value) that can be due to the lack of efficiency of the investment market and the inability to use the accounting information by investment market activists.Keywords: accounting information, company value, profit quality characteristics, price coefficient
Procedia PDF Downloads 13919982 A Nonlinear Feature Selection Method for Hyperspectral Image Classification
Authors: Pei-Jyun Hsieh, Cheng-Hsuan Li, Bor-Chen Kuo
Abstract:
For hyperspectral image classification, feature reduction is an important pre-processing for avoiding the Hughes phenomena due to the difficulty for collecting training samples. Hence, lots of researches developed feature selection methods such as F-score, HSIC (Hilbert-Schmidt Independence Criterion), and etc., to improve hyperspectral image classification. However, most of them only consider the class separability in the original space, i.e., a linear class separability. In this study, we proposed a nonlinear class separability measure based on kernel trick for selecting an appropriate feature subset. The proposed nonlinear class separability was formed by a generalized RBF kernel with different bandwidths with respect to different features. Moreover, it considered the within-class separability and the between-class separability. A genetic algorithm was applied to tune these bandwidths such that the smallest with-class separability and the largest between-class separability simultaneously. This indicates the corresponding feature space is more suitable for classification. In addition, the corresponding nonlinear classification boundary can separate classes very well. These optimal bandwidths also show the importance of bands for hyperspectral image classification. The reciprocals of these bandwidths can be viewed as weights of bands. The smaller bandwidth, the larger weight of the band, and the more importance for classification. Hence, the descending order of the reciprocals of the bands gives an order for selecting the appropriate feature subsets. In the experiments, three hyperspectral image data sets, the Indian Pine Site data set, the PAVIA data set, and the Salinas A data set, were used to demonstrate the selected feature subsets by the proposed nonlinear feature selection method are more appropriate for hyperspectral image classification. Only ten percent of samples were randomly selected to form the training dataset. All non-background samples were used to form the testing dataset. The support vector machine was applied to classify these testing samples based on selected feature subsets. According to the experiments on the Indian Pine Site data set with 220 bands, the highest accuracies by applying the proposed method, F-score, and HSIC are 0.8795, 0.8795, and 0.87404, respectively. However, the proposed method selects 158 features. F-score and HSIC select 168 features and 217 features, respectively. Moreover, the classification accuracies increase dramatically only using first few features. The classification accuracies with respect to feature subsets of 10 features, 20 features, 50 features, and 110 features are 0.69587, 0.7348, 0.79217, and 0.84164, respectively. Furthermore, only using half selected features (110 features) of the proposed method, the corresponding classification accuracy (0.84168) is approximate to the highest classification accuracy, 0.8795. For other two hyperspectral image data sets, the PAVIA data set and Salinas A data set, we can obtain the similar results. These results illustrate our proposed method can efficiently find feature subsets to improve hyperspectral image classification. One can apply the proposed method to determine the suitable feature subset first according to specific purposes. Then researchers can only use the corresponding sensors to obtain the hyperspectral image and classify the samples. This can not only improve the classification performance but also reduce the cost for obtaining hyperspectral images.Keywords: hyperspectral image classification, nonlinear feature selection, kernel trick, support vector machine
Procedia PDF Downloads 26519981 Weak Instability in Direct Integration Methods for Structural Dynamics
Authors: Shuenn-Yih Chang, Chiu-Li Huang
Abstract:
Three structure-dependent integration methods have been developed for solving equations of motion, which are second-order ordinary differential equations, for structural dynamics and earthquake engineering applications. Although they generally have the same numerical properties, such as explicit formulation, unconditional stability and second-order accuracy, a different performance is found in solving the free vibration response to either linear elastic or nonlinear systems with high frequency modes. The root cause of this different performance in the free vibration responses is analytically explored herein. As a result, it is verified that a weak instability is responsible for the different performance of the integration methods. In general, a weak instability will result in an inaccurate solution or even numerical instability in the free vibration responses of high frequency modes. As a result, a weak instability must be prohibited for time integration methods.Keywords: dynamic analysis, high frequency, integration method, overshoot, weak instability
Procedia PDF Downloads 22419980 Increasing the Frequency of Laser Impulses with Optical Choppers with Rotational Shafts
Authors: Virgil-Florin Duma, Dorin Demian
Abstract:
Optical choppers are among the most common optomechatronic devices, utilized in numerous applications, from radiometry to telescopes and biomedical imaging. The classical configuration has a rotational disk with windows with linear margins. This research points out the laser signals that can be obtained with these classical choppers, as well as with another, novel, patented configuration, of eclipse choppers (i.e., with rotational disks with windows with non-linear margins, oriented outwards or inwards). Approximately triangular laser signals can be obtained with eclipse choppers, in contrast to the approximately sinusoidal – with classical devices. The main topic of this work refers to another, novel device, of choppers with shafts of different shapes and with slits of various profiles (patent pending). A significant improvement which can be obtained (with regard to disk choppers) refers to the chop frequencies of the laser signals. Thus, while 1 kHz is their typical limit for disk choppers, with choppers with shafts, a more than 20 times increase in the chop frequency can be obtained with choppers with shafts. Their transmission functions are also discussed, for different types of laser beams. Acknowledgments: This research is supported by the Romanian National Authority for Scientific Research, through the project PN-III-P2-2.1-BG-2016-0297.Keywords: laser signals, laser systems, optical choppers, optomechatronics, transfer functions, eclipse choppers, choppers with shafts
Procedia PDF Downloads 19119979 Earphone Style Wearable Device for Automatic Guidance Service with Position Sensing
Authors: Dawei Cai
Abstract:
This paper describes a design of earphone style wearable device that may provide an automatic guidance service for visitors. With both position information and orientation information obtained from NFC and terrestrial magnetism sensor, a high level automatic guide service may be realized. To realize the service, a algorithm for position detection using the packet from NFC tags, and developed an algorithm to calculate the device orientation based on the data from acceleration and terrestrial magnetism sensors called as MEMS. If visitors want to know some explanation about an exhibit in front of him, what he has to do is only move to the object and stands for a moment. The identification program will automatically recognize the status based on the information from NFC and MEMS, and start playing explanation content about the exhibit. This service should be useful for improving the understanding of the exhibition items and bring more satisfactory visiting experience without less burden.Keywords: wearable device, MEMS sensor, ubiquitous computing, NFC
Procedia PDF Downloads 23919978 Performance Evaluation of Refinement Method for Wideband Two-Beams Formation
Authors: C. Bunsanit
Abstract:
This paper presents the refinement method for two beams formation of wideband smart antenna. The refinement method for weighting coefficients is based on Fully Spatial Signal Processing by taking Inverse Discrete Fourier Transform (IDFT), and its simulation results are presented using MATLAB. The radiation pattern is created by multiplying the incoming signal with real weights and then summing them together. These real weighting coefficients are computed by IDFT method; however, the range of weight values is relatively wide. Therefore, for reducing this range, the refinement method is used. The radiation pattern concerns with five input parameters to control. These parameters are maximum weighting coefficient, wideband signal, direction of mainbeam, beamwidth, and maximum of minor lobe level. Comparison of the obtained simulation results between using refinement method and taking only IDFT shows that the refinement method works well for wideband two beams formation.Keywords: fully spatial signal processing, beam forming, refinement method, smart antenna, weighting coefficient, wideband
Procedia PDF Downloads 22619977 A Novel Method for Solving Nonlinear Whitham–Broer–Kaup Equation System
Authors: Ayda Nikkar, Roghayye Ahmadiasl
Abstract:
In this letter, a new analytical method called homotopy perturbation method, which does not need small parameter in the equation is implemented for solving the nonlinear Whitham–Broer–Kaup (WBK) partial differential equation. In this method, a homotopy is introduced to be constructed for the equation. The initial approximations can be freely chosen with possible unknown constants which can be determined by imposing the boundary and initial conditions. Comparison of the results with those of exact solution has led us to significant consequences. The results reveal that the HPM is very effective, convenient and quite accurate to systems of nonlinear equations. It is predicted that the HPM can be found widely applicable in engineering.Keywords: homotopy perturbation method, Whitham–Broer–Kaup (WBK) equation, Modified Boussinesq, Approximate Long Wave
Procedia PDF Downloads 31119976 Analysis of Vocal Fold Vibrations from High-Speed Digital Images Based on Dynamic Time Warping
Authors: A. I. A. Rahman, Sh-Hussain Salleh, K. Ahmad, K. Anuar
Abstract:
Analysis of vocal fold vibration is essential for understanding the mechanism of voice production and for improving clinical assessment of voice disorders. This paper presents a Dynamic Time Warping (DTW) based approach to analyze and objectively classify vocal fold vibration patterns. The proposed technique was designed and implemented on a Glottal Area Waveform (GAW) extracted from high-speed laryngeal images by delineating the glottal edges for each image frame. Feature extraction from the GAW was performed using Linear Predictive Coding (LPC). Several types of voice reference templates from simulations of clear, breathy, fry, pressed and hyperfunctional voice productions were used. The patterns of the reference templates were first verified using the analytical signal generated through Hilbert transformation of the GAW. Samples from normal speakers’ voice recordings were then used to evaluate and test the effectiveness of this approach. The classification of the voice patterns using the technique of LPC and DTW gave the accuracy of 81%.Keywords: dynamic time warping, glottal area waveform, linear predictive coding, high-speed laryngeal images, Hilbert transform
Procedia PDF Downloads 23919975 Load Forecasting Using Neural Network Integrated with Economic Dispatch Problem
Authors: Mariyam Arif, Ye Liu, Israr Ul Haq, Ahsan Ashfaq
Abstract:
High cost of fossil fuels and intensifying installations of alternate energy generation sources are intimidating main challenges in power systems. Making accurate load forecasting an important and challenging task for optimal energy planning and management at both distribution and generation side. There are many techniques to forecast load but each technique comes with its own limitation and requires data to accurately predict the forecast load. Artificial Neural Network (ANN) is one such technique to efficiently forecast the load. Comparison between two different ranges of input datasets has been applied to dynamic ANN technique using MATLAB Neural Network Toolbox. It has been observed that selection of input data on training of a network has significant effects on forecasted results. Day-wise input data forecasted the load accurately as compared to year-wise input data. The forecasted load is then distributed among the six generators by using the linear programming to get the optimal point of generation. The algorithm is then verified by comparing the results of each generator with their respective generation limits.Keywords: artificial neural networks, demand-side management, economic dispatch, linear programming, power generation dispatch
Procedia PDF Downloads 18919974 Reduced Differential Transform Methods for Solving the Fractional Diffusion Equations
Authors: Yildiray Keskin, Omer Acan, Murat Akkus
Abstract:
In this paper, the solution of fractional diffusion equations is presented by means of the reduced differential transform method. Fractional partial differential equations have special importance in engineering and sciences. Application of reduced differential transform method to this problem shows the rapid convergence of the sequence constructed by this method to the exact solution. The numerical results show that the approach is easy to implement and accurate when applied to fractional diffusion equations. The method introduces a promising tool for solving many fractional partial differential equations.Keywords: fractional diffusion equations, Caputo fractional derivative, reduced differential transform method, partial
Procedia PDF Downloads 52519973 Electrochemical Detection of the Chemotherapy Agent Methotrexate in vitro from Physiological Fluids Using Functionalized Carbon Nanotube past Electrodes
Authors: Shekher Kummari, V. Sunil Kumar, K. Vengatajalabathy Gobi
Abstract:
A simple, cost-effective, reusable and reagent-free electrochemical biosensor is developed with functionalized multiwall carbon nanotube paste electrode (f-CNTPE) for the sensitive and selective determination of the important chemotherapeutic drug methotrexate (MTX), which is widely used for the treatment of various cancer and autoimmune diseases. The electrochemical response of the fabricated electrode towards the detection of MTX is examined by cyclic voltammetry (CV), differential pulse voltammetry (DPV) and square wave voltammetry (SWV). CV studies have shown that f-CNTPE electrode system exhibited an excellent electrocatalytic activity towards the oxidation of MTX in phosphate buffer (0.2 M) compared with a conventional carbon paste electrode (CPE). The oxidation peak current is enhanced by nearly two times in magnitude. Applying the DPV method under optimized conditions, a linear calibration plot is achieved over a wide range of concentration from 4.0×10⁻⁷ M to 5.5×10⁻⁶ M with the detection limit 1.6×10⁻⁷ M. further, by applying the SWV method a parabolic calibration plot was achieved starting from a very low concentration of 1.0×10⁻⁸ M, and the sensor could detect as low as 2.9×10⁻⁹ M MTX in 10 s and 10 nM were detected in steady state current-time analysis. The f-CNTPE shows very good selectivity towards the specific recognition of MTX in the presence of important biological interference. The electrochemical biosensor detects MTX in-vitro directly from pharmaceutical sample, undiluted urine and human blood serum samples at a concentration range 5.0×10⁻⁷ M with good recovery limits.Keywords: amperometry, electrochemical detection, human blood serum, methotrexate, MWCNT, SWV
Procedia PDF Downloads 30919972 Numerical Solution of Momentum Equations Using Finite Difference Method for Newtonian Flows in Two-Dimensional Cartesian Coordinate System
Authors: Ali Ateş, Ansar B. Mwimbo, Ali H. Abdulkarim
Abstract:
General transport equation has a wide range of application in Fluid Mechanics and Heat Transfer problems. In this equation, generally when φ variable which represents a flow property is used to represent fluid velocity component, general transport equation turns into momentum equations or with its well known name Navier-Stokes equations. In these non-linear differential equations instead of seeking for analytic solutions, preferring numerical solutions is a more frequently used procedure. Finite difference method is a commonly used numerical solution method. In these equations using velocity and pressure gradients instead of stress tensors decreases the number of unknowns. Also, continuity equation, by integrating the system, number of equations is obtained as number of unknowns. In this situation, velocity and pressure components emerge as two important parameters. In the solution of differential equation system, velocities and pressures must be solved together. However, in the considered grid system, when pressure and velocity values are jointly solved for the same nodal points some problems confront us. To overcome this problem, using staggered grid system is a referred solution method. For the computerized solutions of the staggered grid system various algorithms were developed. From these, two most commonly used are SIMPLE and SIMPLER algorithms. In this study Navier-Stokes equations were numerically solved for Newtonian flow, whose mass or gravitational forces were neglected, for incompressible and laminar fluid, as a hydro dynamically fully developed region and in two dimensional cartesian coordinate system. Finite difference method was chosen as the solution method. This is a parametric study in which varying values of velocity components, pressure and Reynolds numbers were used. Differential equations were discritized using central difference and hybrid scheme. The discritized equation system was solved by Gauss-Siedel iteration method. SIMPLE and SIMPLER were used as solution algorithms. The obtained results, were compared for central difference and hybrid as discritization methods. Also, as solution algorithm, SIMPLE algorithm and SIMPLER algorithm were compared to each other. As a result, it was observed that hybrid discritization method gave better results over a larger area. Furthermore, as computer solution algorithm, besides some disadvantages, it can be said that SIMPLER algorithm is more practical and gave result in short time. For this study, a code was developed in DELPHI programming language. The values obtained in a computer program were converted into graphs and discussed. During sketching, the quality of the graph was increased by adding intermediate values to the obtained result values using Lagrange interpolation formula. For the solution of the system, number of grid and node was found as an estimated. At the same time, to indicate that the obtained results are satisfactory enough, by doing independent analysis from the grid (GCI analysis) for coarse, medium and fine grid system solution domain was obtained. It was observed that when graphs and program outputs were compared with similar studies highly satisfactory results were achieved.Keywords: finite difference method, GCI analysis, numerical solution of the Navier-Stokes equations, SIMPLE and SIMPLER algoritms
Procedia PDF Downloads 39119971 A Generalized Family of Estimators for Estimation of Unknown Population Variance in Simple Random Sampling
Authors: Saba Riaz, Syed A. Hussain
Abstract:
This paper is addressing the estimation method of the unknown population variance of the variable of interest. A new generalized class of estimators of the finite population variance has been suggested using the auxiliary information. To improve the precision of the proposed class, known population variance of the auxiliary variable has been used. Mathematical expressions for the biases and the asymptotic variances of the suggested class are derived under large sample approximation. Theoretical and numerical comparisons are made to investigate the performances of the proposed class of estimators. The empirical study reveals that the suggested class of estimators performs better than the usual estimator, classical ratio estimator, classical product estimator and classical linear regression estimator. It has also been found that the suggested class of estimators is also more efficient than some recently published estimators.Keywords: study variable, auxiliary variable, finite population variance, bias, asymptotic variance, percent relative efficiency
Procedia PDF Downloads 22519970 Magnetohydrodynamic Flow of Viscoelastic Nanofluid and Heat Transfer over a Stretching Surface with Non-Uniform Heat Source/Sink and Non-Linear Radiation
Authors: Md. S. Ansari, S. S. Motsa
Abstract:
In this paper, an analysis has been made on the flow of non-Newtonian viscoelastic nanofluid over a linearly stretching sheet under the influence of uniform magnetic field. Heat transfer characteristics is analyzed taking into the effect of nonlinear radiation and non-uniform heat source/sink. Transport equations contain the simultaneous effects of Brownian motion and thermophoretic diffusion of nanoparticles. The relevant partial differential equations are non-dimensionalized and transformed into ordinary differential equations by using appropriate similarity transformations. The transformed, highly nonlinear, ordinary differential equations are solved by spectral local linearisation method. The numerical convergence, error and stability analysis of iteration schemes are presented. The effects of different controlling parameters, namely, radiation, space and temperature-dependent heat source/sink, Brownian motion, thermophoresis, viscoelastic, Lewis number and the magnetic force parameter on the flow field, heat transfer characteristics and nanoparticles concentration are examined. The present investigation has many industrial and engineering applications in the fields of coatings and suspensions, cooling of metallic plates, oils and grease, paper production, coal water or coal–oil slurries, heat exchangers’ technology, and materials’ processing and exploiting.Keywords: magnetic field, nonlinear radiation, non-uniform heat source/sink, similar solution, spectral local linearisation method, Rosseland diffusion approximation
Procedia PDF Downloads 37219969 A Convolution Neural Network PM-10 Prediction System Based on a Dense Measurement Sensor Network in Poland
Authors: Piotr A. Kowalski, Kasper Sapala, Wiktor Warchalowski
Abstract:
PM10 is a suspended dust that primarily has a negative effect on the respiratory system. PM10 is responsible for attacks of coughing and wheezing, asthma or acute, violent bronchitis. Indirectly, PM10 also negatively affects the rest of the body, including increasing the risk of heart attack and stroke. Unfortunately, Poland is a country that cannot boast of good air quality, in particular, due to large PM concentration levels. Therefore, based on the dense network of Airly sensors, it was decided to deal with the problem of prediction of suspended particulate matter concentration. Due to the very complicated nature of this issue, the Machine Learning approach was used. For this purpose, Convolution Neural Network (CNN) neural networks have been adopted, these currently being the leading information processing methods in the field of computational intelligence. The aim of this research is to show the influence of particular CNN network parameters on the quality of the obtained forecast. The forecast itself is made on the basis of parameters measured by Airly sensors and is carried out for the subsequent day, hour after hour. The evaluation of learning process for the investigated models was mostly based upon the mean square error criterion; however, during the model validation, a number of other methods of quantitative evaluation were taken into account. The presented model of pollution prediction has been verified by way of real weather and air pollution data taken from the Airly sensor network. The dense and distributed network of Airly measurement devices enables access to current and archival data on air pollution, temperature, suspended particulate matter PM1.0, PM2.5, and PM10, CAQI levels, as well as atmospheric pressure and air humidity. In this investigation, PM2.5, and PM10, temperature and wind information, as well as external forecasts of temperature and wind for next 24h served as inputted data. Due to the specificity of the CNN type network, this data is transformed into tensors and then processed. This network consists of an input layer, an output layer, and many hidden layers. In the hidden layers, convolutional and pooling operations are performed. The output of this system is a vector containing 24 elements that contain prediction of PM10 concentration for the upcoming 24 hour period. Over 1000 models based on CNN methodology were tested during the study. During the research, several were selected out that give the best results, and then a comparison was made with the other models based on linear regression. The numerical tests carried out fully confirmed the positive properties of the presented method. These were carried out using real ‘big’ data. Models based on the CNN technique allow prediction of PM10 dust concentration with a much smaller mean square error than currently used methods based on linear regression. What's more, the use of neural networks increased Pearson's correlation coefficient (R²) by about 5 percent compared to the linear model. During the simulation, the R² coefficient was 0.92, 0.76, 0.75, 0.73, and 0.73 for 1st, 6th, 12th, 18th, and 24th hour of prediction respectively.Keywords: air pollution prediction (forecasting), machine learning, regression task, convolution neural networks
Procedia PDF Downloads 14919968 Reconfigurable Consensus Achievement of Multi Agent Systems Subject to Actuator Faults in a Leaderless Architecture
Authors: F. Amirarfaei, K. Khorasani
Abstract:
In this paper, reconfigurable consensus achievement of a team of agents with marginally stable linear dynamics and single input channel has been considered. The control algorithm is based on a first order linear protocol. After occurrence of a LOE fault in one of the actuators, using the imperfect information of the effectiveness of the actuators from fault detection and identification module, the control gain is redesigned in a way to still reach consensus. The idea is based on the modeling of change in effectiveness as change of Laplacian matrix. Then as special cases of this class of systems, a team of single integrators as well as double integrators are considered and their behavior subject to a LOE fault is considered. The well-known relative measurements consensus protocol is applied to a leaderless team of single integrator as well as double integrator systems, and Gersgorin disk theorem is employed to determine whether fault occurrence has an effect on system stability and team consensus achievement or not. The analyses show that loss of effectiveness fault in actuator(s) of integrator systems affects neither system stability nor consensus achievement.Keywords: multi-agent system, actuator fault, stability analysis, consensus achievement
Procedia PDF Downloads 33719967 The Construction of Exact Solutions for the Nonlinear Lattice Equation via Coth and Csch Functions Method
Authors: A. Zerarka, W. Djoudi
Abstract:
The method developed in this work uses a generalised coth and csch funtions method to construct new exact travelling solutions to the nonlinear lattice equation. The technique of the homogeneous balance method is used to handle the appropriated solutions.Keywords: coth functions, csch functions, nonlinear partial differential equation, travelling wave solutions
Procedia PDF Downloads 66319966 Using VR as a Training Tool in the Banking Industry
Authors: Bjørn Salskov, Nicolaj Bang, Charlotte Falko
Abstract:
Future labour markets demand employees that can carry out a non-linear task which is still not possible for computers. This means that employees must have well-developed soft-skills to perform at high levels in such a work environment. One of these soft-skills is presenting a message effectively. To be able to present a message effectively, one needs to practice this. To practice effectively, the trainee needs feedback on the current performance. Here VR environments can be used as a practice tool because it gives the trainee a sense of presence and reality. VR environments are becoming a cost-effective training method since it does not demand the presence of an expert to provide this feedback. The research article analysed in this study suggests that VR environment can be used and are able to provide the necessary feedback to the trainee which in turn will help the trainee become better at the task. The research analysed in this review does, however, show that there is a need for a study with larger sample size and a study which runs over a longer period.Keywords: training, presentation, presentation skills, VR training, VR as a training tool, VR and presentation
Procedia PDF Downloads 12219965 Bienzymatic Nanocomposites Biosensors Complexed with Gold Nanoparticles, Polyaniline, Recombinant MN Peroxidase from Corn, and Glucose Oxidase to Measure Glucose
Authors: Anahita Izadyar
Abstract:
Using a recombinant enzyme derived from corn and a simple modification, we are fabricating a facile, fast, and cost-beneficial novel biosensor to measure glucose. We are applying Plant Produced Mn Peroxidase (PPMP), glucose oxidase (GOx), polyaniline (PANI) as conductive polymer and gold nanoparticles (AuNPs) on Au electrode using electrochemical response to detect glucose. We applied the entrapment method of enzyme composition, which is generally used to immobilize conductive polymer and facilitate electron transfer from the enzyme oxidation-reduction center to the sample solution. In this work, the oxidation of glucose on the modified gold electrode was quantified with Linear Sweep Voltammetry(LSV). We expect that the modified biosensor has the potential for monitoring various biofluids.Keywords: plant-produced manganese peroxidase, enzyme-based biosensors, glucose, modified gold nanoparticles electrode, polyaniline
Procedia PDF Downloads 20019964 Geospatial Analysis for Predicting Sinkhole Susceptibility in Greene County, Missouri
Authors: Shishay Kidanu, Abdullah Alhaj
Abstract:
Sinkholes in the karst terrain of Greene County, Missouri, pose significant geohazards, imposing challenges on construction and infrastructure development, with potential threats to lives and property. To address these issues, understanding the influencing factors and modeling sinkhole susceptibility is crucial for effective mitigation through strategic changes in land use planning and practices. This study utilizes geographic information system (GIS) software to collect and process diverse data, including topographic, geologic, hydrogeologic, and anthropogenic information. Nine key sinkhole influencing factors, ranging from slope characteristics to proximity to geological structures, were carefully analyzed. The Frequency Ratio method establishes relationships between attribute classes of these factors and sinkhole events, deriving class weights to indicate their relative importance. Weighted integration of these factors is accomplished using the Analytic Hierarchy Process (AHP) and the Weighted Linear Combination (WLC) method in a GIS environment, resulting in a comprehensive sinkhole susceptibility index (SSI) model for the study area. Employing Jenk's natural break classifier method, the SSI values are categorized into five distinct sinkhole susceptibility zones: very low, low, moderate, high, and very high. Validation of the model, conducted through the Area Under Curve (AUC) and Sinkhole Density Index (SDI) methods, demonstrates a robust correlation with sinkhole inventory data. The prediction rate curve yields an AUC value of 74%, indicating a 74% validation accuracy. The SDI result further supports the success of the sinkhole susceptibility model. This model offers reliable predictions for the future distribution of sinkholes, providing valuable insights for planners and engineers in the formulation of development plans and land-use strategies. Its application extends to enhancing preparedness and minimizing the impact of sinkhole-related geohazards on both infrastructure and the community.Keywords: sinkhole, GIS, analytical hierarchy process, frequency ratio, susceptibility, Missouri
Procedia PDF Downloads 7419963 The Reliability and Shape of the Force-Power-Velocity Relationship of Strength-Trained Males Using an Instrumented Leg Press Machine
Authors: Mark Ashton Newman, Richard Blagrove, Jonathan Folland
Abstract:
The force-velocity profile of an individual has been shown to influence success in ballistic movements, independent of the individuals' maximal power output; therefore, effective and accurate evaluation of an individual’s F-V characteristics and not solely maximal power output is important. The relatively narrow range of loads typically utilised during force-velocity profiling protocols due to the difficulty in obtaining force data at high velocities may bring into question the accuracy of the F-V slope along with predictions pertaining to the maximum force that the system can produce at a velocity of null (F₀) and the theoretical maximum velocity against no load (V₀). As such, the reliability of the slope of the force-velocity profile, as well as V₀, has been shown to be relatively poor in comparison to F₀ and maximal power, and it has been recommended to assess velocity at loads closer to both F₀ and V₀. The aim of the present study was to assess the relative and absolute reliability of an instrumented novel leg press machine which enables the assessment of force and velocity data at loads equivalent to ≤ 10% of one repetition maximum (1RM) through to 1RM during a ballistic leg press movement. The reliability of maximal and mean force, velocity, and power, as well as the respective force-velocity and power-velocity relationships and the linearity of the force-velocity relationship, were evaluated. Sixteen male strength-trained individuals (23.6 ± 4.1 years; 177.1 ± 7.0 cm; 80.0 ± 10.8 kg) attended four sessions; during the initial visit, participants were familiarised with the leg press, modified to include a mounted force plate (Type SP3949, Force Logic, Berkshire, UK) and a Micro-Epsilon WDS-2500-P96 linear positional transducer (LPT) (Micro-Epsilon, Merseyside, UK). Peak isometric force (IsoMax) and a dynamic 1RM, both from a starting position of 81% leg length, were recorded for the dominant leg. Visits two to four saw the participants carry out the leg press movement at loads equivalent to ≤ 10%, 30%, 50%, 70%, and 90% 1RM. IsoMax was recorded during each testing visit prior to dynamic F-V profiling repetitions. The novel leg press machine used in the present study appears to be a reliable tool for measuring F and V-related variables across a range of loads, including velocities closer to V₀ when compared to some of the findings within the published literature. Both linear and polynomial models demonstrated good to excellent levels of reliability for SFV and F₀ respectively, with reliability for V₀ being good using a linear model but poor using a 2nd order polynomial model. As such, a polynomial regression model may be most appropriate when using a similar unilateral leg press setup to predict maximal force production capabilities due to only a 5% difference between F₀ and obtained IsoMax values with a linear model being best suited to predict V₀.Keywords: force-velocity, leg-press, power-velocity, profiling, reliability
Procedia PDF Downloads 5819962 A Study on the Reliability Evaluation of a Timer Card for Air Dryer of the Railway Vehicle
Authors: Chul Su Kim, Jun Ku Lee, Won Jun Lee
Abstract:
The EMU (electric multiple unit) vehicle timer card is a PCB (printed circuit board) for controlling the air-dryer to remove the moisture of the generated air from the air compressor of the braking device. This card is exposed to the lower part of the railway vehicle, so it is greatly affected by the external environment such as temperature and humidity. The main cause of the failure of this timer card is deterioration of soldering area of the PCB surface due to temperature and humidity. Therefore, in the viewpoint of preventive maintenance, it is important to evaluate the reliability of the timer card and predict the replacement cycle to secure the safety of the air braking device is one of the main devices for driving. In this study, the existing and the improved products were evaluated on the reliability through ALT (accelerated life test). In addition, the acceleration factor by the 'Coffin-Manson' equation was obtained, and the remaining lifetime was compared and examined.Keywords: reliability evaluation, timer card, Printed Circuit Board, Accelerated Life Test
Procedia PDF Downloads 27919961 Learning outside the Box by Using Memory Techniques Skill: Case Study in Indonesia Memory Sports Council
Authors: Muhammad Fajar Suardi, Fathimatufzzahra, Dela Isnaini Sendra
Abstract:
Learning is an activity that has been used to do, especially for a student or academics. But a handful of people have not been using and maximizing their brains work and some also do not know a good brain work time in capturing the lessons, so that knowledge is absorbed is also less than the maximum. Indonesia Memory Sports Council (IMSC) is an institution which is engaged in the performance of the brain and the development of effective learning methods by using several techniques that can be used in considering the lessons and knowledge to grasp well, including: loci method, substitution method, and chain method. This study aims to determine the techniques and benefits of using the method given in learning and memorization by applying memory techniques taught by Indonesia Memory Sports Council (IMSC) to students and the difference if not using this method. This research uses quantitative research with survey method addressed to students of Indonesian Memory Sports Council (IMSC). The results of this study indicate that learn, understand and remember the lesson using the techniques of memory which is taught in Indonesia Memory Sport Council is very effective and faster to absorb the lesson than learning without using the techniques of memory, and this affects the academic achievement of students in each educational institution.Keywords: chain method, Indonesia memory sports council, loci method, substitution method
Procedia PDF Downloads 29019960 Intermittent Effect of Coupled Thermal and Acoustic Sources on Combustion: A Spatial Perspective
Authors: Pallavi Gajjar, Vinayak Malhotra
Abstract:
Rockets have been known to have played a predominant role in spacecraft propulsion. The quintessential aspect of combustion-related requirements of a rocket engine is the minimization of the surrounding risks/hazards. Over time, it has become imperative to understand the combustion rate variation in presence of external energy source(s). Rocket propulsion represents a special domain of chemical propulsion assisted by high speed flows in presence of acoustics and thermal source(s). Jet noise leads to a significant loss of resources and every year a huge amount of financial aid is spent to prevent it. External heat source(s) induce high possibility of fire risk/hazards which can sufficiently endanger the operation of a space vehicle. Appreciable work had been done with justifiable simplification and emphasis on the linear variation of external energy source(s), which yields good physical insight but does not cater to accurate predictions. Present work experimentally attempts to understand the correlation between inter-energy conversions with the non-linear placement of external energy source(s). The work is motivated by the need to have better fire safety and enhanced combustion. The specific objectives of the work are a) To interpret the related energy transfer for combustion in presence of alternate external energy source(s) viz., thermal and acoustic, b) To fundamentally understand the role of key controlling parameters viz., separation distance, the number of the source(s), selected configurations and their non-linear variation to resemble real-life cases. An experimental setup was prepared using incense sticks as potential fuel and paraffin wax candles as the external energy source(s). The acoustics was generated using frequency generator, and source(s) were placed at selected locations. Non-equidistant parametric experimentation was carried out, and the effects were noted on regression rate changes. The results are expected to be very helpful in offering a new perspective into futuristic rocket designs and safety.Keywords: combustion, acoustic energy, external energy sources, regression rate
Procedia PDF Downloads 14119959 Physical Activity and Mental Health: A Cross-Sectional Investigation into the Relationship of Specific Physical Activity Domains and Mental Well-Being
Authors: Katja Siefken, Astrid Junge
Abstract:
Background: Research indicates that physical activity (PA) protects us from developing mental disorders. The knowledge regarding optimal domain, intensity, type, context, and amount of PA promotion for the prevention of mental disorders is sparse and incoherent. The objective of this study is to determine the relationship between PA domains and mental well-being, and whether associations vary by domain, amount, context, intensity, and type of PA. Methods: 310 individuals (age: 25 yrs., SD 7; 73% female) completed a questionnaire on personal patterns of their PA behaviour (IPQA) and their mental health (Centre of Epidemiologic Studies Depression Scale (CES-D), Generalized Anxiety Disorder (GAD-7) scale, the subjective physical well-being (FEW-16)). Linear and multiple regression were used for analysis. Findings: Individuals who met the PA recommendation (N=269) reported higher scores on subjective physical well-being than those who did not meet the PA recommendations (N=41). Whilst vigorous intensity PA predicts subjective well-being (β = .122, p = .028), it also correlates with depression. The more vigorously physically active a person is, the higher the depression score (β = .127, p = .026). The strongest impact of PA on mental well-being can be seen in the transport domain. A positive linear correlation on subjective physical well-being (β =.175, p = .002), and a negative linear correlation for anxiety (β =-.142, p = .011) and depression (β = -.164, p = .004) was found. Multiple regression analysis indicates similar results: Time spent in active transport on the bicycle significantly lowers anxiety and depression scores and enhances subjective physical well-being. The more time a participant spends using the bicycle for transport, the lower the depression (β = -.143, p = .013) and anxiety scores (β = -.111,p = .050). Conclusions: Meeting the PA recommendations enhances subjective physical well-being. Active transport has a substantial impact on mental well-being. Findings have implications for policymakers, employers, public health experts and civil society. A stronger focus on the promotion and protection of health through active transport is recommended. Inter-sectoral exchange, outside the health sector, is required. Health systems must engage other sectors in adopting policies that maximize possible health gains.Keywords: active transport, mental well-being, health promotion, psychological disorders
Procedia PDF Downloads 32119958 A Finite Element Method Simulation for Rocket Motor Material Selection
Authors: T. Kritsana, P. Sawitri, P. Teeratas
Abstract:
This article aims to study the effect of pressure on rocket motor case by Finite Element Method simulation to select optimal material in rocket motor manufacturing process. In this study, cylindrical tubes with outside diameter of 122 mm and thickness of 3 mm are used for simulation. Defined rocket motor case materials are AISI4130, AISI1026, AISI1045, AL2024 and AL7075. Internal pressure used for the simulation is 22 MPa. The result from Finite Element Method shows that at a pressure of 22 MPa rocket motor case produced by AISI4130, AISI1045 and AL7075 can be used. A comparison of the result between AISI4130, AISI1045 and AL7075 shows that AISI4130 has minimum principal stress and confirm the results of Finite Element Method by the used of calculation method found that, the results from Finite Element Method has good reliability.Keywords: rocket motor case, finite element method, principal stress, simulation
Procedia PDF Downloads 44919957 A Geometrical Method for the Smoluchowski Equation on the Sphere
Authors: Adriano Valdes-Gomez, Francisco Javier Sevilla
Abstract:
We devise a numerical algorithm to simulate the diffusion of a Brownian particle restricted to the surface of a three-dimensional sphere when the particle is under the effects of an external potential that is coupled linearly. It is obtained using elementary geometry, yet, it converges, in the weak sense, to the solutions to the Smoluchowski equation. Rotations on the sphere, which are the analogs of linear displacements in euclidean spaces, are calculated using algebraic operations and then by a proper scaling, which makes the algorithm efficient and quite simple, especially to what may be the short-time propagator approach. Our findings prove that the global effects of curvature are taken into account in both dynamic and stationary processes, and it is not restricted to work in configuration space, neither restricted to the overdamped limit. We have generalized it successfully to simulate the Kramers or the Ornstein-Uhlenbeck process, where it is necessary to work directly in phase space, and it may be adapted to other two dimensional surfaces with non-constant curvature.Keywords: diffusion on the sphere, Fokker-Planck equation on the sphere, non equilibrium processes on the sphere, numerical methods for diffusion on the sphere
Procedia PDF Downloads 18219956 Customer Experience Management in Food and Beverage Outlet at Indian School of Business: Methodology and Recommendations
Authors: Anupam Purwar
Abstract:
In conventional consumer product industry, stockouts are taken care by carrying buffer stock to check underserving caused by changes in customer demand, incorrect forecast or variability in lead times. But, for food outlets, the alternate of carrying buffer stock is unviable because of indispensable need to serve freshly cooked meals. Besides, the food outlet being the sole provider has no incentives to reduce stockouts, as they have no fear of losing revenue, gross profit, customers and market share. Hence, innovative, easy to implement and practical ways of addressing the twin problem of long queues and poor customer experience needs to be investigated. Current work analyses the demand pattern of 11 different food items across a routine day. Based on this optimum resource allocation for all food items has been carried out by solving a linear programming problem with cost minimization as the objective. Concurrently, recommendations have been devised to address this demand and supply side problem keeping in mind their practicability. Currently, the recommendations are being discussed and implemented at ISB (Indian School of Business) Hyderabad campus.Keywords: F&B industry, resource allocation, demand management, linear programming, LP, queuing analysis
Procedia PDF Downloads 138