Search results for: first order error analysis
36544 Design of Dry Chemical Fire Extinguisher Inspection Equipment in Order to Reduce Ergonomic Risks for Fire Extinguisher Inspectors
Authors: Sitrapee Changmuenwai, Sudaratana Wongweragiat
Abstract:
It is important that a dry chemical fire extinguisher must be inspected for its readiness. For each inspection, the inspectors need to turn the fire extinguisher tank upside down to let the chemical inside the tank move and prevent solidification, which would make the tank not ready for usage when needed. Each tank weighs approximately 16 kg. The inspectors have to turn each tank upside down twice (2 minutes/round). They need to put the tanks over their shoulder close to their ear in order to hear the chemical flow inside the tank or use their hands to feel it. The survey and questionnaire 'The Questionnaire Know Body', which includes neck, left shoulder, upper and lower right arms suggest that all 12 security staffs have the same fatigues. The current dry chemical fire extinguisher inspection affects various ergonomic health problems. Rapid Entire Body Assessment (REBA) is used for evaluation of posture risks so that the working postures may be redesigned or corrected. The dry chemical fire extinguisher inspection equipment has been developed to reduce ergonomic health risks for the inspectors. A REBA analysis has been performed again, and the risk score has been decreased from 13 to 3. In addition, feedbacks from the first trial of the developed equipment show that there are demands to increase the installation in order to reduce the ergonomic health risks.Keywords: dry chemical fire extinguisher inspection equipment, ergonomic, REBA, rapid entire body assessment
Procedia PDF Downloads 12136543 Simplified INS\GPS Integration Algorithm in Land Vehicle Navigation
Authors: Othman Maklouf, Abdunnaser Tresh
Abstract:
Land vehicle navigation is subject of great interest today. Global Positioning System (GPS) is the main navigation system for positioning in such systems. GPS alone is incapable of providing continuous and reliable positioning, because of its inherent dependency on external electromagnetic signals. Inertial Navigation (INS) is the implementation of inertial sensors to determine the position and orientation of a vehicle. The availability of low-cost Micro-Electro-Mechanical-System (MEMS) inertial sensors is now making it feasible to develop INS using an inertial measurement unit (IMU). INS has unbounded error growth since the error accumulates at each step. Usually, GPS and INS are integrated with a loosely coupled scheme. With the development of low-cost, MEMS inertial sensors and GPS technology, integrated INS/GPS systems are beginning to meet the growing demands of lower cost, smaller size, and seamless navigation solutions for land vehicles. Although MEMS inertial sensors are very inexpensive compared to conventional sensors, their cost (especially MEMS gyros) is still not acceptable for many low-end civilian applications (for example, commercial car navigation or personal location systems). An efficient way to reduce the expense of these systems is to reduce the number of gyros and accelerometers, therefore, to use a partial IMU (ParIMU) configuration. For land vehicular use, the most important gyroscope is the vertical gyro that senses the heading of the vehicle and two horizontal accelerometers for determining the velocity of the vehicle. This paper presents a field experiment for a low-cost strap down (ParIMU)\GPS combination, with data post processing for the determination of 2-D components of position (trajectory), velocity and heading. In the present approach, we have neglected earth rotation and gravity variations, because of the poor gyroscope sensitivities of our low-cost IMU (Inertial Measurement Unit) and because of the relatively small area of the trajectory.Keywords: GPS, IMU, Kalman filter, materials engineering
Procedia PDF Downloads 42236542 The Effect of Arbitrary Support Conditions on the Static Behavior of Curved Beams Using the Finite Element Method
Authors: Hossein Mottaghi T., Amir R. Masoodi
Abstract:
This study presents a finite curved element for analyzing the static behavior of curved beams within the elastic range. The objective is to enhance accuracy while reducing the number of elements by incorporating first-order shear deformations of Timoshenko beams. Initially, finite element formulations are developed by considering polynomial initial functions for axial, shear, and rotational deformations for a three-node element. Subsequently, nodal interpolation functions for this element are derived, followed by the construction of the element stiffness matrix. To enable the utilization of the stiffness matrix in the static analysis of curved beams, the constructed matrix in the local coordinates of the element is transformed to the global coordinate system using the rotation matrix. A numerical benchmark example is investigated to assess the accuracy and effectiveness of this method. Moreover, the influence of spring stiffness on the rotation of the endpoint of a clamped beam is examined by substituting each support reaction of the beam with a spring. In the parametric study, the effect of the central angle of the beam on the rotation of the beam's endpoint in a cantilever beam under a concentrated load is examined. This research encompasses various mechanical, geometrical, and boundary configurations to evaluate the static characteristics of curved beams, thus providing valuable insights for their analysis and examination.Keywords: curved beam, finite element method, first-order shear deformation theory, elastic support
Procedia PDF Downloads 7336541 Gnss Aided Photogrammetry for Digital Mapping
Authors: Muhammad Usman Akram
Abstract:
This research work based on GNSS-Aided Photogrammetry for Digital Mapping. It focuses on topographic survey of an area or site which is to be used in future Planning & development (P&D) or can be used for further, examination, exploration, research and inspection. Survey and Mapping in hard-to-access and hazardous areas are very difficult by using traditional techniques and methodologies; as well it is time consuming, labor intensive and has less precision with limited data. In comparison with the advance techniques it is saving with less manpower and provides more precise output with a wide variety of multiple data sets. In this experimentation, Aerial Photogrammetry technique is used where an UAV flies over an area and captures geocoded images and makes a Three-Dimensional Model (3-D Model), UAV operates on a user specified path or area with various parameters; Flight altitude, Ground sampling distance (GSD), Image overlapping, Camera angle etc. For ground controlling, a network of points on the ground would be observed as a Ground Control point (GCP) using Differential Global Positioning System (DGPS) in PPK or RTK mode. Furthermore, that raw data collected by UAV and DGPS will be processed in various Digital image processing programs and Computer Aided Design software. From which as an output we obtain Points Dense Cloud, Digital Elevation Model (DEM) and Ortho-photo. The imagery is converted into geospatial data by digitizing over Ortho-photo, DEM is further converted into Digital Terrain Model (DTM) for contour generation or digital surface. As a result, we get Digital Map of area to be surveyed. In conclusion, we compared processed data with exact measurements taken on site. The error will be accepted if the amount of error is not breached from survey accuracy limits set by concerned institutions.Keywords: photogrammetry, post processing kinematics, real time kinematics, manual data inquiry
Procedia PDF Downloads 3236540 A Single Loop Repetitive Controller for a Four Legs Matrix Converter Unit
Authors: Wesam Rohouma
Abstract:
The aim of this paper is to investigate the use of repetitive controller to regulate the output voltage of three phase four leg matric converter for an Aircraft Ground Power Supply Unit. The proposed controller improve the steady state error and provide good regulation during different loading. Simulation results of 7.5 KW converter are presented to verify the operation of the proposed controller.Keywords: matrix converter, Power electronics, controller, regulation
Procedia PDF Downloads 150636539 Application Reliability Method for the Analysis of the Stability Limit States of Large Concrete Dams
Authors: Mustapha Kamel Mihoubi, Essadik Kerkar, Abdelhamid Hebbouche
Abstract:
According to the randomness of most of the factors affecting the stability of a gravity dam, probability theory is generally used to TESTING the risk of failure and there is a confusing logical transition from the state of stability failed state, so the stability failure process is considered as a probable event. The control of risk of product failures is of capital importance for the control from a cross analysis of the gravity of the consequences and effects of the probability of occurrence of identified major accidents and can incur a significant risk to the concrete dam structures. Probabilistic risk analysis models are used to provide a better understanding the reliability and structural failure of the works, including when calculating stability of large structures to a major risk in the event of an accident or breakdown. This work is interested in the study of the probability of failure of concrete dams through the application of the reliability analysis methods including the methods used in engineering. It is in our case of the use of level II methods via the study limit state. Hence, the probability of product failures is estimated by analytical methods of the type FORM (First Order Reliability Method), SORM (Second Order Reliability Method). By way of comparison, a second level III method was used which generates a full analysis of the problem and involving an integration of the probability density function of, random variables are extended to the field of security by using of the method of Mont-Carlo simulations. Taking into account the change in stress following load combinations: normal, exceptional and extreme the acting on the dam, calculation results obtained have provided acceptable failure probability values which largely corroborate the theory, in fact, the probability of failure tends to increase with increasing load intensities thus causing a significant decrease in strength, especially in the presence of combinations of unique and extreme loads. Shear forces then induce a shift threatens the reliability of the structure by intolerable values of the probability of product failures. Especially, in case THE increase of uplift in a hypothetical default of the drainage system.Keywords: dam, failure, limit state, monte-carlo, reliability, probability, sliding, Taylor
Procedia PDF Downloads 31836538 Detect Critical Thinking Skill in Written Text Analysis. The Use of Artificial Intelligence in Text Analysis vs Chat/Gpt
Authors: Lucilla Crosta, Anthony Edwards
Abstract:
Companies and the market place nowadays struggle to find employees with adequate skills in relation to anticipated growth of their businesses. At least half of workers will need to undertake some form of up-skilling process in the next five years in order to remain aligned with the requests of the market . In order to meet these challenges, there is a clear need to explore the potential uses of AI (artificial Intelligence) based tools in assessing transversal skills (critical thinking, communication and soft skills of different types in general) of workers and adult students while empowering them to develop those same skills in a reliable trustworthy way. Companies seek workers with key transversal skills that can make a difference between workers now and in the future. However, critical thinking seems to be the one of the most imprtant skill, bringing unexplored ideas and company growth in business contexts. What employers have been reporting since years now, is that this skill is lacking in the majority of workers and adult students, and this is particularly visible trough their writing. This paper investigates how critical thinking and communication skills are currently developed in Higher Education environments through use of AI tools at postgraduate levels. It analyses the use of a branch of AI namely Machine Learning and Big Data and of Neural Network Analysis. It also examines the potential effect the acquisition of these skills through AI tools and what kind of effects this has on employability This paper will draw information from researchers and studies both at national (Italy & UK) and international level in Higher Education. The issues associated with the development and use of one specific AI tool Edulai, will be examined in details. Finally comparisons will be also made between these tools and the more recent phenomenon of Chat GPT and forthcomings and drawbacks will be analysed.Keywords: critical thinking, artificial intelligence, higher education, soft skills, chat GPT
Procedia PDF Downloads 11036537 Organizational Commitment and Job Satisfaction of Job Order Personnel in the Overseas Workers Welfare Administration Regional Welfare Office Caraga
Authors: Anne Jane M. Hallasgo
Abstract:
This study assessed the level of job satisfaction and organizational commitment among job order personnel at the Overseas Workers Welfare Administration (OWWA) Regional Welfare Office Caraga. The primary objective of the study was to determine a correlation between the employees’ level of organizational commitment, job satisfaction, and their work performance. A carefully selected sample of twenty-five job orders from the OWWA Regional Welfare Office Caraga participated in the study. These individuals were chosen to represent the organization’s job order workforce. For accuracy and dependability, various types of statistical methods and instruments were employed, including advanced statistical tests like the independent sample T-test, one-way analysis of variance (ANOVA), and Spearman's rank correlation coefficient, as well as descriptive statistics like mean, frequency, and percentage. The study found an acceptable level of job satisfaction regarding work performance. It revealed a significant relationship between affective commitment and job satisfaction concerning leadership and coworkers. A correlation was observed between normative commitment and work performance. The findings suggest that organizations emphasizing positive leadership, fostering supportive coworker relationships, aligning with employee values, and promoting a culture of commitment are likely to enhance both affective and normative commitment, thereby improving overall employee satisfaction. The study recommends designing and implementing a holistic employee well-being program that addresses physical, mental, and emotional health contributing to increased job satisfaction and organizational commitment, creating a healthier and engaged workforce. This research contributes to the understanding of the dynamics of organizational commitment and job satisfaction among job order employees in the public sector.Keywords: affective commitment, continuous commitment, normative commitment, job satisfaction
Procedia PDF Downloads 4836536 Robust Inference with a Skew T Distribution
Authors: M. Qamarul Islam, Ergun Dogan, Mehmet Yazici
Abstract:
There is a growing body of evidence that non-normal data is more prevalent in nature than the normal one. Examples can be quoted from, but not restricted to, the areas of Economics, Finance and Actuarial Science. The non-normality considered here is expressed in terms of fat-tailedness and asymmetry of the relevant distribution. In this study a skew t distribution that can be used to model a data that exhibit inherent non-normal behavior is considered. This distribution has tails fatter than a normal distribution and it also exhibits skewness. Although maximum likelihood estimates can be obtained by solving iteratively the likelihood equations that are non-linear in form, this can be problematic in terms of convergence and in many other respects as well. Therefore, it is preferred to use the method of modified maximum likelihood in which the likelihood estimates are derived by expressing the intractable non-linear likelihood equations in terms of standardized ordered variates and replacing the intractable terms by their linear approximations obtained from the first two terms of a Taylor series expansion about the quantiles of the distribution. These estimates, called modified maximum likelihood estimates, are obtained in closed form. Hence, they are easy to compute and to manipulate analytically. In fact the modified maximum likelihood estimates are equivalent to maximum likelihood estimates, asymptotically. Even in small samples the modified maximum likelihood estimates are found to be approximately the same as maximum likelihood estimates that are obtained iteratively. It is shown in this study that the modified maximum likelihood estimates are not only unbiased but substantially more efficient than the commonly used moment estimates or the least square estimates that are known to be biased and inefficient in such cases. Furthermore, in conventional regression analysis, it is assumed that the error terms are distributed normally and, hence, the well-known least square method is considered to be a suitable and preferred method for making the relevant statistical inferences. However, a number of empirical researches have shown that non-normal errors are more prevalent. Even transforming and/or filtering techniques may not produce normally distributed residuals. Here, a study is done for multiple linear regression models with random error having non-normal pattern. Through an extensive simulation it is shown that the modified maximum likelihood estimates of regression parameters are plausibly robust to the distributional assumptions and to various data anomalies as compared to the widely used least square estimates. Relevant tests of hypothesis are developed and are explored for desirable properties in terms of their size and power. The tests based upon modified maximum likelihood estimates are found to be substantially more powerful than the tests based upon least square estimates. Several examples are provided from the areas of Economics and Finance where such distributions are interpretable in terms of efficient market hypothesis with respect to asset pricing, portfolio selection, risk measurement and capital allocation, etc.Keywords: least square estimates, linear regression, maximum likelihood estimates, modified maximum likelihood method, non-normality, robustness
Procedia PDF Downloads 39736535 Controversies and Contradiction in (IR) Reversibility and the Equilibrium of Reactive Systems
Authors: Joao Teotonio Manzi
Abstract:
Reversibility, irreversibility, equilibrium and steady-state that play a central role in the thermodynamic analysis of processes arising in the context of reactive systems are discussed in this article. Such concepts have generated substantial doubts, even among the most experienced researchers, and engineers, because from the literature, conclusive or definitive statements cannot be extracted. Concepts such as the time-reversibility of irreversible processes seem paradoxical, requiring further analysis. Equilibrium and reversibility, which appear to be of the same nature, have also been re-examined in the light of maximum entropy. The goal of this paper is to revisit and explore these concepts based on classical thermodynamics in order to have a better understanding them due to their impacts on technological advances, as a result, to generate an optimal procedure for designing, monitoring, and engineering optimization. Furthermore, an effective graphic procedure for dimensioning a Plug Flow Reactor has been provided. Thus, to meet the needs of chemical engineering from a simple conceptual analysis but with significant practical effects, a macroscopic approach is taken so as to integrate the different parts of this paper.Keywords: reversibility, equilibrium, steady-state, thermodynamics, reactive system
Procedia PDF Downloads 10636534 Comparison of Two Neural Networks To Model Margarine Age And Predict Shelf-Life Using Matlab
Authors: Phakamani Xaba, Robert Huberts, Bilainu Oboirien
Abstract:
The present study was aimed at developing & comparing two neural-network-based predictive models to predict shelf-life/product age of South African margarine using free fatty acid (FFA), water droplet size (D3.3), water droplet distribution (e-sigma), moisture content, peroxide value (PV), anisidine valve (AnV) and total oxidation (totox) value as input variables to the model. Brick margarine products which had varying ages ranging from fresh i.e. week 0 to week 47 were sourced. The brick margarine products which had been stored at 10 & 25 °C and were characterized. JMP and MATLAB models to predict shelf-life/ margarine age were developed and their performances were compared. The key performance indicators to evaluate the model performances were correlation coefficient (CC), root mean square error (RMSE), and mean absolute percentage error (MAPE) relative to the actual data. The MATLAB-developed model showed a better performance in all three performance indicators. The correlation coefficient of the MATLAB model was 99.86% versus 99.74% for the JMP model, the RMSE was 0.720 compared to 1.005 and the MAPE was 7.4% compared to 8.571%. The MATLAB model was selected to be the most accurate, and then, the number of hidden neurons/ nodes was optimized to develop a single predictive model. The optimized MATLAB with 10 neurons showed a better performance compared to the models with 1 & 5 hidden neurons. The developed models can be used by margarine manufacturers, food research institutions, researchers etc, to predict shelf-life/ margarine product age, optimize addition of antioxidants, extend shelf-life of products and proactively troubleshoot for problems related to changes which have an impact on shelf-life of margarine without conducting expensive trials.Keywords: margarine shelf-life, predictive modelling, neural networks, oil oxidation
Procedia PDF Downloads 19736533 Unveiling the Detailed Turn Off-On Mechanism of Carbon Dots to Different Sized MnO₂ Nanosensor for Selective Detection of Glutathione
Authors: Neeraj Neeraj, Soumen Basu, Banibrata Maity
Abstract:
Glutathione (GSH) is one of the most important biomolecules having small molecular weight, which helps in various cellular functions like regulation of gene, xenobiotic metabolism, preservation of intracellular redox activities, signal transduction, etc. Therefore, the detection of GSH requires huge attention by using extremely selective and sensitive techniques. Herein, a rapid fluorometric nanosensor is designed by combining carbon dots (Cdots) and MnO₂ nanoparticles of different sizes for the detection of GSH. The bottom-up approach, i.e., microwave method, was used for the preparation of the water soluble and greatly fluorescent Cdots by using ascorbic acid as a precursor. MnO₂ nanospheres of different sizes (large, medium, and small) were prepared by varying the ratio of concentration of methionine and KMnO₄ at room temperature, which was confirmed by HRTEM analysis. The successive addition of MnO₂ nanospheres in Cdots results fluorescence quenching. From the fluorescence intensity data, Stern-Volmer quenching constant values (KS-V) were evaluated. From the fluorescence intensity and lifetime analysis, it was found that the degree of fluorescence quenching of Cdots followed the order: large > medium > small. Moreover, fluorescence recovery studies were also performed in the presence of GSH. Fluorescence restoration studies also show the order of turn on follows the same order, i.e., large > medium > small, which was also confirmed by quantum yield and lifetime studies. The limits of detection (LOD) of GSH in presence of Cdots@different sized MnO₂ nanospheres were also evaluated. It was observed thatLOD values were in μM region and lowest in case of large MnO₂ nanospheres. The separation distance (d) between Cdots and the surface of different MnO₂ nanospheres was determined. The d values increase with increase in the size of the MnO₂ nanospheres. In summary, the synthesized Cdots@MnO₂ nanocomposites acted as a rapid, simple, economical as well as environmental-friendly nanosensor for the detection of GSH.Keywords: carbon dots, fluorescence, glutathione, MnO₂ nanospheres, turn off-on
Procedia PDF Downloads 15236532 Effect of Downstream Pressure in Tuning the Flow Control Orifices of Pressure Fed Reaction Control System Thrusters
Authors: Prakash M.N, Mahesh G, Muhammed Rafi K.M, Shiju P. Nair
Abstract:
Introduction: In launch vehicle missions, Reaction Control thrusters are being used for the three-axis stabilization of the vehicle during the coasting phases. A pressure-fed propulsion system is used for the operation of these thrusters due to its less complexity. In liquid stages, these thrusters are designed to draw propellant from the same tank used for the main propulsion system. So in order to regulate the propellant flow rates of these thrusters, flow control orifices are used in feed lines. These orifices are calibrated separately as per the flow rate requirement of individual thrusters for the nominal operating conditions. In some missions, it was observed that the thrusters were operated at higher thrust than nominal. This point was addressed through a series of cold flow and hot tests carried out in-ground and this paper elaborates the details of the same. Discussion: In order to find out the exact reason for this phenomenon, two flight configuration thrusters were identified and hot tested in the ground with calibrated orifices and feed lines. During these tests, the chamber pressure, which is directly proportional to the thrust, is measured. In both cases, chamber pressures higher than the nominal by 0.32bar to 0.7bar were recorded. The increase in chamber pressure is due to an increase in the oxidizer flow rate of both the thrusters. Upon further investigation, it is observed that the calibration of the feed line is done with ambient pressure downstream. But in actual flight conditions, the orifices will be subjected to operate with 10 to 11bar pressure downstream. Due to this higher downstream pressure, the flow through the orifices increases and thereby, the thrusters operate with higher chamber pressure values. Conclusion: As part of further investigatory tests, two numbers of fresh thrusters were realized. Orifice tuning of these thrusters was carried out in three different ways. In the first trial, the orifice tuning was done by simulating 1bar pressure downstream. The second trial was done with the injector assembled downstream. In the third trial, the downstream pressure equal to the flight injection pressure was simulated downstream. Using these calibrated orifices, hot tests were carried out in simulated vacuum conditions. Chamber pressure and flow rate values were exactly matching with the prediction for the second and third trials. But for the first trial, the chamber pressure values obtained in the hot test were more than the prediction. This clearly shows that the flow is detached in the 1st trial and attached for the 2nd & 3rd trials. Hence, the error in tuning the flow control orifices is pinpointed as the reason for this higher chamber pressure observed in flight.Keywords: reaction control thruster, propellent, orifice, chamber pressure
Procedia PDF Downloads 20136531 Developing a DNN Model for the Production of Biogas From a Hybrid BO-TPE System in an Anaerobic Wastewater Treatment Plant
Authors: Hadjer Sadoune, Liza Lamini, Scherazade Krim, Amel Djouadi, Rachida Rihani
Abstract:
Deep neural networks are highly regarded for their accuracy in predicting intricate fermentation processes. Their ability to learn from a large amount of datasets through artificial intelligence makes them particularly effective models. The primary obstacle in improving the performance of these models is to carefully choose the suitable hyperparameters, including the neural network architecture (number of hidden layers and hidden units), activation function, optimizer, learning rate, and other relevant factors. This study predicts biogas production from real wastewater treatment plant data using a sophisticated approach: hybrid Bayesian optimization with a tree-structured Parzen estimator (BO-TPE) for an optimised deep neural network (DNN) model. The plant utilizes an Upflow Anaerobic Sludge Blanket (UASB) digester that treats industrial wastewater from soft drinks and breweries. The digester has a working volume of 1574 m3 and a total volume of 1914 m3. Its internal diameter and height were 19 and 7.14 m, respectively. The data preprocessing was conducted with meticulous attention to preserving data quality while avoiding data reduction. Three normalization techniques were applied to the pre-processed data (MinMaxScaler, RobustScaler and StandardScaler) and compared with the Non-Normalized data. The RobustScaler approach has strong predictive ability for estimating the volume of biogas produced. The highest predicted biogas volume was 2236.105 Nm³/d, with coefficient of determination (R2), mean absolute error (MAE), and root mean square error (RMSE) values of 0.712, 164.610, and 223.429, respectively.Keywords: anaerobic digestion, biogas production, deep neural network, hybrid bo-tpe, hyperparameters tuning
Procedia PDF Downloads 3836530 An Experiment Research on the Effect of Brain-Break in the Classroom on Elementary School Students’ Selective Attention
Authors: Hui Liu, Xiaozan Wang, Jiarong Zhong, Ziming Shao
Abstract:
Introduction: Related research shows that students don’t concentrate on teacher’s speaking in the classroom. The d2 attention test is a time-limited test about selective attention. The d2 attention test can be used to evaluate individual selective attention. Purpose: To use the d2 attention test tool to measure the difference between the attention level of the experimental class and the control class before and after Brain-Break and to explore the effect of Brain-Break in the classroom on students' selective attention. Methods: According to the principle of no difference in pre-test data, two classes in the fourth- grade of Shenzhen Longhua Central Primary School were selected. After 20 minutes of class in the third class in the morning and the third class in the afternoon, about 3-minute Brain-Break intervention was performed in the experimental class for 10 weeks. The normal class in the control class did not intervene. Before and after the experiment, the d2 attention test tool was used to test the attention level of the two-class students. The paired sample t-test and independent sample t-test in SPSS 23.0 was used to test the change in the attention level of the two-class classes around 10 weeks. This article only presents results with significant differences. Results: The independent sample t-test results showed that after ten-week of Brain-Break, the missed errors (E1 t = -2.165 p = 0.042), concentration performance (CP t = 1.866 p = 0.05), and the degree of omissions (Epercent t = -2.375 p = 0.029) in experimental class showed significant differences compared with control class. The students’ error level decreased and the concentration increased. Conclusions: Adding Brain-Break interventions in the classroom can effectively improve the attention level of fourth-grade primary school students to a certain extent, especially can improve the concentration of attention and decrease the error rate in the tasks. The new sport's learning model is worth promotingKeywords: cultural class, micromotor, attention, D2 test
Procedia PDF Downloads 13236529 Countering the Bullwhip Effect by Absorbing It Downstream in the Supply Chain
Authors: Geng Cui, Naoto Imura, Katsuhiro Nishinari, Takahiro Ezaki
Abstract:
The bullwhip effect, which refers to the amplification of demand variance as one moves up the supply chain, has been observed in various industries and extensively studied through analytic approaches. Existing methods to mitigate the bullwhip effect, such as decentralized demand information, vendor-managed inventory, and the Collaborative Planning, Forecasting, and Replenishment System, rely on the willingness and ability of supply chain participants to share their information. However, in practice, information sharing is often difficult to realize due to privacy concerns. The purpose of this study is to explore new ways to mitigate the bullwhip effect without the need for information sharing. This paper proposes a 'bullwhip absorption strategy' (BAS) to alleviate the bullwhip effect by absorbing it downstream in the supply chain. To achieve this, a two-stage supply chain system was employed, consisting of a single retailer and a single manufacturer. In each time period, the retailer receives an order generated according to an autoregressive process. Upon receiving the order, the retailer depletes the ordered amount, forecasts future demand based on past records, and places an order with the manufacturer using the order-up-to replenishment policy. The manufacturer follows a similar process. In essence, the mechanism of the model is similar to that of the beer game. The BAS is implemented at the retailer's level to counteract the bullwhip effect. This strategy requires the retailer to reduce the uncertainty in its orders, thereby absorbing the bullwhip effect downstream in the supply chain. The advantage of the BAS is that upstream participants can benefit from a reduced bullwhip effect. Although the retailer may incur additional costs, if the gain in the upstream segment can compensate for the retailer's loss, the entire supply chain will be better off. Two indicators, order variance and inventory variance, were used to quantify the bullwhip effect in relation to the strength of absorption. It was found that implementing the BAS at the retailer's level results in a reduction in both the retailer's and the manufacturer's order variances. However, when examining the impact on inventory variances, a trade-off relationship was observed. The manufacturer's inventory variance monotonically decreases with an increase in absorption strength, while the retailer's inventory variance does not always decrease as the absorption strength grows. This is especially true when the autoregression coefficient has a high value, causing the retailer's inventory variance to become a monotonically increasing function of the absorption strength. Finally, numerical simulations were conducted for verification, and the results were consistent with our theoretical analysis.Keywords: bullwhip effect, supply chain management, inventory management, demand forecasting, order-to-up policy
Procedia PDF Downloads 7536528 Velocity Logs Error Reduction for In-Service Calibration of Vessel Performance Indicators
Authors: Maria Tsompanoglou, Dimitris Armenis
Abstract:
Vessel behavior in different operational and weather conditions constitutes the main area of interest for the ship operator. Ship speed and fuel consumption are the most decisive parameters in this respect, as their correlation provides information about the economic and environmental efficiency of the vessel, becoming the basis of decision making in terms of maintenance and trading. In the analysis of vessel operational profile for the evaluation of fuel consumption and the equivalent CO2 emissions footprint, the indications of Speed Through Water are widely used. The seasonal and regional variations in seawater characteristics, which are available nowadays, can provide the basis for accurate estimation of the errors in Speed Through Water indications at any time. Accuracy in the speed value on a route basis can enable operator identify the ship fuel and propulsion efficiency and proceed with improvements. This paper discusses case studies, where the actual vessel speed was corrected by a post-processing algorithm. The effects of the vessel correction to standard Key Performance Indicators, as well as operational findings not identified earlier, are also discussed.Keywords: data analytics, MATLAB, vessel performance monitoring, speed through water
Procedia PDF Downloads 30036527 Earth Tremors in Nigeria: A Precursor to Major Disaster?
Authors: Oluseyi Adunola Bamisaiye
Abstract:
The frequency of occurrence of earth tremor in Nigeria has increased tremendously in recent years. Slow earthquakes/ tremor have preceded some large earthquakes in some other regions of the world and the Nigerian case may not be an exception. Timely and careful investigation of these tremors may reveal their relation to large earthquakes and provides important clues to constrain the slip rates on tectonic faults. Thus making it imperative to keep under watch and also study carefully the tectonically active terrains within the country, in order to adequately forecast, prescribe mitigation measures and in order to avoid a major disaster. This report provides new evidence of a slow slip transient in a strongly locked seismogenic zone of the Okemesi fold belt. The aim of this research is to investigate the different methods of earth tremor monitoring using fault slip analysis and mapping of Okemesi hills, which has been the most recent epicenter to most of the recent tremors.Keywords: earth tremor, fault slip, intraplate activities, plate tectonics
Procedia PDF Downloads 15536526 Bibliometric Analysis of Risk Assessment of Inland Maritime Accidents in Bangladesh
Authors: Armana Huq, Wahidur Rahman, Sanwar Kader
Abstract:
Inland waterways in Bangladesh play an important role in providing comfortable and low-cost transportation. However, a maritime accident takes away many lives and creates unwanted hazards every year. This article deals with a comprehensive review of inland waterway accidents in Bangladesh. Additionally, it includes a comparative study between international and local inland research studies based on maritime accidents. Articles from inland waterway areas are analyzed in-depth to make a comprehensive overview of the nature of the academic work, accident and risk management process and different statistical analyses. It is found that empirical analysis based on the available statistical data dominates the research domain. For this study, major maritime accident-related works in the last four decades in Bangladesh (1981-2020) are being analyzed for preparing a bibliometric analysis. A study of maritime accidents of passenger's vessels during (1995-2005) indicates that the predominant causes of accidents in the inland waterways of Bangladesh are collision and adverse weather (77%), out of which collision due to human error alone stands (56%) of all accidents. Another study refers that the major causes of waterway accidents are the collision (60.3%) during 2005-2015. About 92% of this collision occurs due to direct contact with another vessel during this period. Rest 8% of the collision occurs by contact with permanent obstruction on waterway roots. The overall analysis of another study from the last 25 years (1995-2019) shows that one of the main types of accidents is collisions, with about 50.3% of accidents being caused by collisions. The other accident types are cyclone or storm (17%), overload (11.3%), physical failure (10.3%), excessive waves (5.1%), and others (6%). Very few notable works are available in testing or comparing the methods, proposing new methods for risk management, modeling, uncertainty treatment. The purpose of this paper is to provide an overview of the evolution of marine accident-related research domain regarding inland waterway of Bangladesh and attempts to introduce new ideas and methods to abridge the gap between international and national inland maritime-related work domain which can be a catalyst for a safer and sustainable water transportation system in Bangladesh. Another fundamental objective of this paper is to navigate various national maritime authorities and international organizations to implement risk management processes for shipping accident prevention in waterway areas.Keywords: inland waterways, safety, bibliometric analysis, risk management, accidents
Procedia PDF Downloads 18236525 The External Debt in the Context of Economic Growth: The Sample of Turkey
Authors: Ayşen Edirneligil, Mehmet Mucuk
Abstract:
In developing countries, one of the most important restrictions about the economic growth is the lack of national savings which are supposed to finance the investments. In order to overcome this restriction and achieve the higher rate of economic growth by increasing the level of output, countries choose the external borrowing. However, there is a dispute in the literature over the correlation between external debt and economic growth. The aim of this study is to examine the effects of external debt on Turkish economic growth by using VAR analysis with the quarterly data over the period of 2002:01-2014:04. In this respect, Johansen Cointegration Test, Impulse- Response Function and Variance Decomposition Tests will be used for analyses. Empirical findings show that there is no cointegration in the long run.Keywords: external debt, economic growth, Turkish economy, time series analysis
Procedia PDF Downloads 39936524 Numerical Analysis of Jet Grouting Strengthened Pile under Lateral Loading
Authors: Reza Ziaie Moayed, Naeem Gholampoor
Abstract:
Jet grouting strengthened pile (JPP) is one of composite piles used in soft ground improvement. It may improve the vertical and lateral bearing capacity effectively and it has been practically used in a considerable scale. In order to make a further research on load transfer mechanism of single JPP with and without cap under lateral loads, JPP is analyzed by means of FEM analysis. It is resulted that the JPP pile could improve lateral bearing capacity by compared with bored concrete pile which is higher for shorter pile and the biggest bending moment of JPP pile is located in the depth of around 48% of embedded length of the pile. Meanwhile, increase of JPP pile length causes to increase of peak mobilized bending moment. Also, by cap addition, JPP piles will have a much higher lateral bearing capacity and increasing in cohesion of soil layer resulted to increase of lateral bearing capacity of JPP pile. In addition, the numerical results basically coincide with the experimental results presented by other researchers.Keywords: bending moment, FEM analysis, JPP pile, lateral bearing capacity
Procedia PDF Downloads 32636523 Vehicle Gearbox Fault Diagnosis Based on Cepstrum Analysis
Authors: Mohamed El Morsy, Gabriela Achtenová
Abstract:
Research on damage of gears and gear pairs using vibration signals remains very attractive, because vibration signals from a gear pair are complex in nature and not easy to interpret. Predicting gear pair defects by analyzing changes in vibration signal of gears pairs in operation is a very reliable method. Therefore, a suitable vibration signal processing technique is necessary to extract defect information generally obscured by the noise from dynamic factors of other gear pairs. This article presents the value of cepstrum analysis in vehicle gearbox fault diagnosis. Cepstrum represents the overall power content of a whole family of harmonics and sidebands when more than one family of sidebands is present at the same time. The concept for the measurement and analysis involved in using the technique are briefly outlined. Cepstrum analysis is used for detection of an artificial pitting defect in a vehicle gearbox loaded with different speeds and torques. The test stand is equipped with three dynamometers; the input dynamometer serves as the internal combustion engine, the output dynamometers introduce the load on the flanges of the output joint shafts. The pitting defect is manufactured on the tooth side of a gear of the fifth speed on the secondary shaft. Also, a method for fault diagnosis of gear faults is presented based on order cepstrum. The procedure is illustrated with the experimental vibration data of the vehicle gearbox. The results show the effectiveness of cepstrum analysis in detection and diagnosis of the gear condition.Keywords: cepstrum analysis, fault diagnosis, gearbox, vibration signals
Procedia PDF Downloads 37936522 TRACE/FRAPTRAN Analysis of Kuosheng Nuclear Power Plant Dry-Storage System
Authors: J. R. Wang, Y. Chiang, W. Y. Li, H. T. Lin, H. C. Chen, C. Shih, S. W. Chen
Abstract:
The dry-storage systems of nuclear power plants (NPPs) in Taiwan have become one of the major safety concerns. There are two steps considered in this study. The first step is the verification of the TRACE by using VSC-17 experimental data. The results of TRACE were similar to the VSC-17 data. It indicates that TRACE has the respectable accuracy in the simulation and analysis of the dry-storage systems. The next step is the application of TRACE in the dry-storage system of Kuosheng NPP (BWR/6). Kuosheng NPP is the second BWR NPP of Taiwan Power Company. In order to solve the storage of the spent fuels, Taiwan Power Company developed the new dry-storage system for Kuosheng NPP. In this step, the dry-storage system model of Kuosheng NPP was established by TRACE. Then, the steady state simulation of this model was performed and the results of TRACE were compared with the Kuosheng NPP data. Finally, this model was used to perform the safety analysis of Kuosheng NPP dry-storage system. Besides, FRAPTRAN was used tocalculate the transient performance of fuel rods.Keywords: BWR, TRACE, FRAPTRAN, dry-storage
Procedia PDF Downloads 51936521 Speedup Breadth-First Search by Graph Ordering
Abstract:
Breadth-First Search(BFS) is a core graph algorithm that is widely used for graph analysis. As it is frequently used in many graph applications, improve the BFS performance is essential. In this paper, we present a graph ordering method that could reorder the graph nodes to achieve better data locality, thus, improving the BFS performance. Our method is based on an observation that the sibling relationships will dominate the cache access pattern during the BFS traversal. Therefore, we propose a frequency-based model to construct the graph order. First, we optimize the graph order according to the nodes’ visit frequency. Nodes with high visit frequency will be processed in priority. Second, we try to maximize the child nodes overlap layer by layer. As it is proved to be NP-hard, we propose a heuristic method that could greatly reduce the preprocessing overheads. We conduct extensive experiments on 16 real-world datasets. The result shows that our method could achieve comparable performance with the state-of-the-art methods while the graph ordering overheads are only about 1/15.Keywords: breadth-first search, BFS, graph ordering, graph algorithm
Procedia PDF Downloads 13836520 Predicting Customer Purchasing Behaviour in Retail Marketing: A Research for a Supermarket Chain
Authors: Sabri Serkan Güllüoğlu
Abstract:
Analysis can be defined as the process of gathering, recording and researching data related to products and services, in order to learn something. But for marketers, analyses are not only used for learning but also an essential and critical part of the business, because this allows companies to offer products or services which are focused and well targeted. Market analysis also identify market trends, demographics, customer’s buying habits and important information on the competition. Data mining is used instead of traditional research, because it extracts predictive information about customer and sales from large databases. In contrast to traditional research, data mining relies on information that is already available. Simply the goal is to improve the efficiency of supermarkets. In this study, the purpose is to find dependency on products. For instance, which items are bought together, using association rules in data mining. Moreover, this information will be used for improving the profitability of customers such as increasing shopping time and sales of fewer sold items.Keywords: data mining, association rule mining, market basket analysis, purchasing
Procedia PDF Downloads 48336519 Design and Sensitivity Analysis of Photovoltaic/Thermal Solar Collector
Authors: H. M. Farghally, N. M. Ahmed, H. T. El-Madany, D. M. Atia, F. H. Fahmy
Abstract:
Energy is required in almost every aspect of human activities and development of any nation in this world. Increasing fossil fuel price, energy security and climate change have important bearings on sustainable development of any nation. The renewable energy technology is considered one of the drastic approaches taken over the world to reduce the energy problem. The preservation of vegetables by freezing is one of the most important methods of retaining quality in agricultural products over long-term storage periods. Freezing factories show high demand of energy for both heat and electricity; the hybrid Photovoltaic/Thermal (PV/T) systems could be used in order to meet this requirement. This paper presents PV/T system design for freezing factory. Also, the complete mathematical modeling and Matlab Simulink of PV/T collector is introduced. The sensitivity analysis for the manufacturing parameters of PV/T collector is carried out to study their effect on the thermal and electrical efficiency.Keywords: renewable energy, hybrid PV/T system, sensitivity analysis, ecological sciences
Procedia PDF Downloads 29236518 Evaluation of Vehicle Classification Categories: Florida Case Study
Authors: Ren Moses, Jaqueline Masaki
Abstract:
This paper addresses the need for accurate and updated vehicle classification system through a thorough evaluation of vehicle class categories to identify errors arising from the existing system and proposing modifications. The data collected from two permanent traffic monitoring sites in Florida were used to evaluate the performance of the existing vehicle classification table. The vehicle data were collected and classified by the automatic vehicle classifier (AVC), and a video camera was used to obtain ground truth data. The Federal Highway Administration (FHWA) vehicle classification definitions were used to define vehicle classes from the video and compare them to the data generated by AVC in order to identify the sources of misclassification. Six types of errors were identified. Modifications were made in the classification table to improve the classification accuracy. The results of this study include the development of updated vehicle classification table with a reduction in total error by 5.1%, a step by step procedure to use for evaluation of vehicle classification studies and recommendations to improve FHWA 13-category rule set. The recommendations for the FHWA 13-category rule set indicate the need for the vehicle classification definitions in this scheme to be updated to reflect the distribution of current traffic. The presented results will be of interest to States’ transportation departments and consultants, researchers, engineers, designers, and planners who require accurate vehicle classification information for planning, designing and maintenance of transportation infrastructures.Keywords: vehicle classification, traffic monitoring, pavement design, highway traffic
Procedia PDF Downloads 18136517 CFD Simulation for Air-Borne Infection Analysis in AII-Room
Authors: Young Kwon Yang, In Sung Kang, Jung Ha Hwang, Jin Chul Park
Abstract:
The present study is a foundational study for performance improvements on isolation wards to prevent proliferation of secondary infection of infectious diseases such as SARS, H1N1, and MERS inside hospitals. Accordingly, the present study conducted an analysis of the effect of sealing mechanisms and filling of openings on ensuring air tightness performance in isolation wards as well as simulation on air currents in improved isolation wards. The study method is as follows. First, previous studies on aerial infection type and mechanism were reviewed, and the review results were utilized as basic data of analysis on simulation of air current. Second, national and international legislations and regulations in relation to isolation wards as well as case studies on developed nations were investigated in order to identify the problems in isolation wards in Korea and improvement plans. Third, construction and facility plans were compared and analyzed between general and isolation wards focusing on large general hospitals in Korea, thereby conducting comparison and analysis on the performance and effects of air-tightness of general and isolation wards through CFD simulations. The study results showed that isolation wards had better air-tightness performance than that of general wards.Keywords: AII Room, air-borne infection, CFD, computational fluid dynamics
Procedia PDF Downloads 28736516 Estimation of Asphalt Pavement Surfaces Using Image Analysis Technique
Authors: Mohammad A. Khasawneh
Abstract:
Asphalt concrete pavements gradually lose their skid resistance causing safety problems especially under wet conditions and high driving speeds. In order to enact the actual field polishing and wearing process of asphalt pavement surfaces in a laboratory setting, several laboratory-scale accelerated polishing devices were developed by different agencies. To mimic the actual process, friction and texture measuring devices are needed to quantify surface deterioration at different polishing intervals that reflect different stages of the pavement life. The test could still be considered lengthy and to some extent labor-intensive. Therefore, there is a need to come up with another method that can assist in investigating the bituminous pavement surface characteristics in a practical and time-efficient test procedure. The purpose of this paper is to utilize a well-developed image analysis technique to characterize asphalt pavement surfaces without the need to use conventional friction and texture measuring devices in an attempt to shorten and simplify the polishing procedure in the lab. Promising findings showed the possibility of using image analysis in lieu of the labor-sensitive-variable-in-nature friction and texture measurements. It was found that the exposed aggregate surface area of asphalt specimens made from limestone and gravel aggregates produced solid evidence of the validity of this method in describing asphalt pavement surfaces. Image analysis results correlated well with the British Pendulum Numbers (BPN), Polish Values (PV) and Mean Texture Depth (MTD) values.Keywords: friction, image analysis, polishing, statistical analysis, texture
Procedia PDF Downloads 30636515 Numerical Applications of Tikhonov Regularization for the Fourier Multiplier Operators
Authors: Fethi Soltani, Adel Almarashi, Idir Mechai
Abstract:
Tikhonov regularization and reproducing kernels are the most popular approaches to solve ill-posed problems in computational mathematics and applications. And the Fourier multiplier operators are an essential tool to extend some known linear transforms in Euclidean Fourier analysis, as: Weierstrass transform, Poisson integral, Hilbert transform, Riesz transforms, Bochner-Riesz mean operators, partial Fourier integral, Riesz potential, Bessel potential, etc. Using the theory of reproducing kernels, we construct a simple and efficient representations for some class of Fourier multiplier operators Tm on the Paley-Wiener space Hh. In addition, we give an error estimate formula for the approximation and obtain some convergence results as the parameters and the independent variables approaches zero. Furthermore, using numerical quadrature integration rules to compute single and multiple integrals, we give numerical examples and we write explicitly the extremal function and the corresponding Fourier multiplier operators.Keywords: fourier multiplier operators, Gauss-Kronrod method of integration, Paley-Wiener space, Tikhonov regularization
Procedia PDF Downloads 318