Search results for: real time kinematics
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20497

Search results for: real time kinematics

14857 Using Business Simulations and Game-Based Learning for Enterprise Resource Planning Implementation Training

Authors: Carin Chuang, Kuan-Chou Chen

Abstract:

An Enterprise Resource Planning (ERP) system is an integrated information system that supports the seamless integration of all the business processes of a company. Implementing an ERP system can increase efficiencies and decrease the costs while helping improve productivity. Many organizations including large, medium and small-sized companies have already adopted an ERP system for decades. Although ERP system can bring competitive advantages to organizations, the lack of proper training approach in ERP implementation is still a major concern. Organizations understand the importance of ERP training to adequately prepare managers and users. The low return on investment, however, for the ERP training makes the training difficult for knowledgeable workers to transfer what is learned in training to the jobs at workplace. Inadequate and inefficient ERP training limits the value realization and success of an ERP system. That is the need to call for a profound change and innovation for ERP training in both workplace at industry and the Information Systems (IS) education in academia. The innovated ERP training approach can improve the users’ knowledge in business processes and hands-on skills in mastering ERP system. It also can be instructed as educational material for IS students in universities. The purpose of the study is to examine the use of ERP simulation games via the ERPsim system to train the IS students in learning ERP implementation. The ERPsim is the business simulation game developed by ERPsim Lab at HEC Montréal, and the game is a real-life SAP (Systems Applications and Products) ERP system. The training uses the ERPsim system as the tool for the Internet-based simulation games and is designed as online student competitions during the class. The competitions involve student teams with the facilitation of instructor and put the students’ business skills to the test via intensive simulation games on a real-world SAP ERP system. The teams run the full business cycle of a manufacturing company while interacting with suppliers, vendors, and customers through sending and receiving orders, delivering products and completing the entire cash-to-cash cycle. To learn a range of business skills, student needs to adopt individual business role and make business decisions around the products and business processes. Based on the training experiences learned from rounds of business simulations, the findings show that learners have reduced risk in making mistakes that help learners build self-confidence in problem-solving. In addition, the learners’ reflections from their mistakes can speculate the root causes of the problems and further improve the efficiency of the training. ERP instructors teaching with the innovative approach report significant improvements in student evaluation, learner motivation, attendance, engagement as well as increased learner technology competency. The findings of the study can provide ERP instructors with guidelines to create an effective learning environment and can be transferred to a variety of other educational fields in which trainers are migrating towards a more active learning approach.

Keywords: business simulations, ERP implementation training, ERPsim, game-based learning, instructional strategy, training innovation

Procedia PDF Downloads 131
14856 Numerical Implementation and Testing of Fractioning Estimator Method for the Box-Counting Dimension of Fractal Objects

Authors: Abraham Terán Salcedo, Didier Samayoa Ochoa

Abstract:

This work presents a numerical implementation of a method for estimating the box-counting dimension of self-avoiding curves on a planar space, fractal objects captured on digital images; this method is named fractioning estimator. Classical methods of digital image processing, such as noise filtering, contrast manipulation, and thresholding, among others, are used in order to obtain binary images that are suitable for performing the necessary computations of the fractioning estimator. A user interface is developed for performing the image processing operations and testing the fractioning estimator on different captured images of real-life fractal objects. To analyze the results, the estimations obtained through the fractioning estimator are compared to the results obtained through other methods that are already implemented on different available software for computing and estimating the box-counting dimension.

Keywords: box-counting, digital image processing, fractal dimension, numerical method

Procedia PDF Downloads 76
14855 An Automated Bender Element System Used for S-Wave Velocity Tomography during Model Pile Installation

Authors: Yuxin Wu, Yu-Shing Wang, Zitao Zhang

Abstract:

A high-speed and time-lapse S-wave velocity measurement system has been built up for S-wave tomography in sand. This system is based on bender elements and applied to model pile tests in a tailor-made pressurized chamber to monitor the shear wave velocity distribution during pile installation in sand. Tactile pressure sensors are used parallel together with bender elements to monitor the stress changes during the tests. Strain gages are used to monitor the shaft resistance and toe resistance of pile. Since the shear wave velocity (Vs) is determined by the shear modulus of sand and the shaft resistance of pile is also influenced by the shear modulus of sand around the pile, the purposes of this study are to time-lapse monitor the S-wave velocity distribution change at a certain horizontal section during pile installation and to correlate the S-wave velocity distribution and shaft resistance of pile in sand.

Keywords: bender element, pile, shaft resistance, shear wave velocity, tomography

Procedia PDF Downloads 420
14854 Thulium Laser Design and Experimental Verification for NIR and MIR Nonlinear Applications in Specialty Optical Fibers

Authors: Matej Komanec, Tomas Nemecek, Dmytro Suslov, Petr Chvojka, Stanislav Zvanovec

Abstract:

Nonlinear phenomena in the near- and mid-infrared region are attracting scientific attention mainly due to the supercontinuum generation possibilities and subsequent utilizations for ultra-wideband applications like e.g. absorption spectroscopy or optical coherence tomography. Thulium-based fiber lasers provide access to high-power ultrashort pump pulses in the vicinity of 2000 nm, which can be easily exploited for various nonlinear applications. The paper presents a simulation and experimental study of a pulsed thulium laser based for near-infrared (NIR) and mid-infrared (MIR) nonlinear applications in specialty optical fibers. In the first part of the paper the thulium laser is discussed. The thulium laser is based on a gain-switched seed-laser and a series of amplification stages for obtaining output peak powers in the order of kilowatts for pulses shorter than 200 ps in full-width at half-maximum. The pulsed thulium laser is first studied in a simulation software, focusing on seed-laser properties. Afterward, a pre-amplification thulium-based stage is discussed, with the focus of low-noise signal amplification, high signal gain and eliminating pulse distortions during pulse propagation in the gain medium. Following the pre-amplification stage a second gain stage is evaluated with incorporating a thulium-fiber of shorter length with increased rare-earth dopant ratio. Last a power-booster stage is analyzed, where the peak power of kilowatts should be achieved. Examples of analytical study are further validated by the experimental campaign. The simulation model is further corrected based on real components – parameters such as real insertion-losses, cross-talks, polarization dependencies, etc. are included. The second part of the paper evaluates the utilization of nonlinear phenomena, their specific features at the vicinity of 2000 nm, compared to e.g. 1550 nm, and presents supercontinuum modelling, based on the thulium laser pulsed output. Supercontinuum generation simulation is performed and provides reasonably accurate results, once fiber dispersion profile is precisely defined and fiber nonlinearity is known, furthermore input pulse shape and peak power must be known, which is assured thanks to the experimental measurement of the studied thulium pulsed laser. The supercontinuum simulation model is put in relation to designed and characterized specialty optical fibers, which are discussed in the third part of the paper. The focus is placed on silica and mainly on non-silica fibers (fluoride, chalcogenide, lead-silicate) in their conventional, microstructured or tapered variants. Parameters such as dispersion profile and nonlinearity of exploited fibers were characterized either with an accurate model, developed in COMSOL software or by direct experimental measurement to achieve even higher precision. The paper then combines all three studied topics and presents a possible application of such a thulium pulsed laser system working with specialty optical fibers.

Keywords: nonlinear phenomena, specialty optical fibers, supercontinuum generation, thulium laser

Procedia PDF Downloads 313
14853 Statistical Analysis of the Main Causes of Delay Factors of Infrastructure Projects

Authors: Seyed Ali Mohammadiborna, Mehdi Ravanshadnia

Abstract:

Project delays usually detrimentally affect perceptions of project success and can in some instances, result in increased costs and other time-related damages to project stakeholders. One of the realities in the national infrastructure projects is that since the primary stakeholders are state-affiliated, the delay factors of the projects have not been seriously taken into account despite the importance of on-time completion of projects. Project postponement has different economic and social consequences and leads to the technical and economic infeasibility of the infrastructure projects in the form of reduced productivity and exploitation capacity. The present study aimed at investigating delay factors of Iranian national infrastructure projects according to regulatory reports of the Plan and Budget Organization (BPO) of Iran. The present study scrutinized the influence of each of the factors that caused delays in national Iranian infrastructure projects according to the supervision reports of the planning and budget organization in 8 years. For this purpose, the study analyzed the information regarding the impact of 12 key delay factors causing delays in average 4867 projects per year in all provinces. The said factors were classified into the three groups of executive, credit, and financial and environmental-procurement factors.

Keywords: delays, infrastructure, projects, regulatory

Procedia PDF Downloads 133
14852 Differences in the Processing of Sentences with Lexical Ambiguity and Structural Ambiguity: An Experimental Study

Authors: Mariana T. Teixeira, Joana P. Luz

Abstract:

This paper is based on assumptions of psycholinguistics and investigates the processing of ambiguous sentences in Brazilian Portuguese. Specifically, it aims to verify if there is a difference in processing time between sentences with lexical ambiguity and sentences with structural (or syntactic) ambiguity. We hypothesize, based on the Garden Path Theory, that the two types of ambiguity entail different cognitive efforts, since sentences with structural ambiguity require that two structures be processed, whereas ambiguous phrases whose root of ambiguity is in a word require the processing of a single structure, which admits a variation of punctual meaning, within the scope of only one lexical item. In order to test this hypothesis, 25 undergraduate students, whose average age was 27.66 years, native speakers of Brazilian Portuguese, performed a self-monitoring reading task of ambiguous sentences, which had lexical and structural ambiguity. The results suggest that unambiguous sentence processing is faster than ambiguous sentence processing, whether it has lexical or structural ambiguity. In addition, participants presented a mean reading time greater for sentences with syntactic ambiguity than for sentences with lexical ambiguity, evidencing a greater cognitive effort in sentence processing with structural ambiguity.

Keywords: Brazilian portuguese, lexical ambiguity, sentence processing, syntactic ambiguity

Procedia PDF Downloads 221
14851 Semi-Automatic Segmentation of Mitochondria on Transmission Electron Microscopy Images Using Live-Wire and Surface Dragging Methods

Authors: Mahdieh Farzin Asanjan, Erkan Unal Mumcuoglu

Abstract:

Mitochondria are cytoplasmic organelles of the cell, which have a significant role in the variety of cellular metabolic functions. Mitochondria act as the power plants of the cell and are surrounded by two membranes. Significant morphological alterations are often due to changes in mitochondrial functions. A powerful technique in order to study the three-dimensional (3D) structure of mitochondria and its alterations in disease states is Electron microscope tomography. Detection of mitochondria in electron microscopy images due to the presence of various subcellular structures and imaging artifacts is a challenging problem. Another challenge is that each image typically contains more than one mitochondrion. Hand segmentation of mitochondria is tedious and time-consuming and also special knowledge about the mitochondria is needed. Fully automatic segmentation methods lead to over-segmentation and mitochondria are not segmented properly. Therefore, semi-automatic segmentation methods with minimum manual effort are required to edit the results of fully automatic segmentation methods. Here two editing tools were implemented by applying spline surface dragging and interactive live-wire segmentation tools. These editing tools were applied separately to the results of fully automatic segmentation. 3D extension of these tools was also studied and tested. Dice coefficients of 2D and 3D for surface dragging using splines were 0.93 and 0.92. This metric for 2D and 3D for live-wire method were 0.94 and 0.91 respectively. The root mean square symmetric surface distance values of 2D and 3D for surface dragging was measured as 0.69, 0.93. The same metrics for live-wire tool were 0.60 and 2.11. Comparing the results of these editing tools with the results of automatic segmentation method, it shows that these editing tools, led to better results and these results were more similar to ground truth image but the required time was higher than hand-segmentation time

Keywords: medical image segmentation, semi-automatic methods, transmission electron microscopy, surface dragging using splines, live-wire

Procedia PDF Downloads 163
14850 Comparative Analysis of Teachers’ Performance in Private and Public Primary Schools in Oyo State, Nigeria

Authors: Oyetunji John Adenuga

Abstract:

This study on the comparative analysis of the performance of teachers in private and public schools was carried out in Ibadan North West Local Government Area of Oyo State. This study examined the justification for the claim that there is difference in the performance of teachers in private and public primary schools and at the same time identified factors responsible for the difference in the performance of these teachers. A descriptive survey research design was used for the study. Data generated were analysed using t-test and regression analysis. The findings of the study revealed that there is significant difference in the performance of teachers in private and private primary schools in Ibadan North West Local Government Area of Oyo State (t=64.09; df=459; p,.05). The findings also revealed that the method of teaching in private primary schools is significantly different from the method of teaching in public primary schools in Ibadan North West Local Government Area of Oyo State (t=73.08; df=459; p,.05). Findings revealed that school leadership and management have significant contribution on the performance of private and public primary school teachers in Ibadan North West Local Area of Oyo State. Based on the finding, the following recommendations were made: Primary school teachers need to be motivated and rewarded for excellent performance. Primary schools should be properly equipped with teaching-aid facilities, laboratories and libraries. Government should use the findings of this study to improve on teaching materials provided to the primary school teachers in Nigeria. Public primary schools should be designed by education planners, administrators and government. Headmasters, proprietors and teachers of primary schools should look inward and give a performance appraisal and evaluation of themselves form time to time based on subject they taught. Finally, school administrators should be conscious of the way they manage the teachers in schools not only in informal situations but also in formal settings.

Keywords: private education, public education, school leadership, school management, teachers performance

Procedia PDF Downloads 331
14849 Monitoring Saltwater Corrosion on Steel Samples Using Coda Wave Interferometry in MHZ Frequencies

Authors: Maxime Farin, Emmanuel Moulin, Lynda Chehami, Farouk Benmeddour, Pierre Campistron

Abstract:

Assessing corrosion is crucial in the petrochemical and marine industry. Usual ultrasonic methods based on guided waves to detect corrosion can inspect large areas but lack precision. We propose a complementary and sensitive ultrasonic method (~ 10 MHz) based on coda wave interferometry to detect and quantify corrosion at the surface of a steel sample. The method relies on a single piezoelectric transducer, exciting the sample and measuring the scattered coda signals at different instants in time. A laboratory experiment is conducted with a steel sample immersed in salted water for 60~h with parallel coda and temperature measurements to correct coda dependence to temperature variations. Micrometric changes to the sample surface caused by corrosion are detected in the late coda signals, allowing precise corrosion detection. Moreover, a good correlation is found between a parameter quantifying the temperature-corrected stretching of the coda over time with respect to a reference without corrosion and the corrosion surface over the sample recorded with a camera.

Keywords: coda wave interferometry, nondestructive evaluation, corrosion, ultrasonics

Procedia PDF Downloads 222
14848 Effect of Smoking on Tear Break-Up Time and Basal Tear Secretion

Authors: Kalsoom Rani

Abstract:

Tobacco contains nicotine, which causes addiction to many toxic chemicals. In the world, people consume it in the form of smoke, chew, and sniffing, smoke of it is composed of almost 7000 active chemicals, which are very harmful to human health as well as for eye health, inhalation of tobacco smoke and fumes can accelerate and cause many blinding eye diseases. Dry eye and smoking have not been covered extensively in researches; more studies are required to unveil the relationship between smoking and dry eye. This study was conducted to determine the quantity and quality of tears in smokers. 60 subjects participated in the study, which was divided into two groups on the basis of consumption of cigarettes per day with age matched non smokers of 15-50 years. All participants have gone through a study based questioner, eye examination, and diagnostic 'Dry Eye Tests' for evaporative tears evaluation and measurement of basal tear secretion. Subjects were included in the criteria of 10 cigarettes per day with a minimum duration of 1 year; passive smokers for control groups were excluded. The study was carried out in a Medina Teaching Hospital, Faisalabad, Pakistan, ophthalmology department for the duration of 8 months. Mean values for tear break up time (TBUT), was reported 10sec with SD of +3.74 in controlled group, 5sec with SD + 2.32 in smokers and 4sec SD +3.77 heavy smokers in right eye (RE) and left eye (LE) 10.35sec with SD of +3.88 in controlled 5sec with SD + 2.3 in smokers and much reduced TBUT in heavy smokers was 3.85sec SD+2.20. Smoking has a very strong association with TRUT with a significance of P=.00 both eyes. Mean Schirmer-I value of the subjects was reported 12.6mm with SD + 8.37 in RE and 12.59mm with SD + 8.96 LE. The mean Schirmer-II test value was reported in the right, and left eye with a mean value for control was 20.23mm with SD + 8.93, 20.75mm with SD + 8.84 respectively, and in Smokers 9.90mm with SD + 5.74, and 10.07mm with SD + 6.98, and in heavy smokers 7.7mm, SD + 3.22 and 6.9, SD + 3.50 mm, association with smoking showed p=.001 in RE and .003 in LE. Smoking has deteriorated effect on both evaporative tear and aqueous tear secretion and causing symptoms of dry eye burning, itching, redness, and watering with epithelial cell damage.

Keywords: tear break-up time, basal tear secretion, smokers, dry eye

Procedia PDF Downloads 121
14847 Cross-Knowledge Graph Relation Completion for Non-Isomorphic Cross-Lingual Entity Alignment

Authors: Yuhong Zhang, Dan Lu, Chenyang Bu, Peipei Li, Kui Yu, Xindong Wu

Abstract:

The Cross-Lingual Entity Alignment (CLEA) task aims to find the aligned entities that refer to the same identity from two knowledge graphs (KGs) in different languages. It is an effective way to enhance the performance of data mining for KGs with scarce resources. In real-world applications, the neighborhood structures of the same entities in different KGs tend to be non-isomorphic, which makes the representation of entities contain diverse semantic information and then poses a great challenge for CLEA. In this paper, we try to address this challenge from two perspectives. On the one hand, the cross-KG relation completion rules are designed with the alignment constraint of entities and relations to improve the topology isomorphism of two KGs. On the other hand, a representation method combining isomorphic weights is designed to include more isomorphic semantics for counterpart entities, which will benefit the CLEA. Experiments show that our model can improve the isomorphism of two KGs and the alignment performance, especially for two non-isomorphic KGs.

Keywords: knowledge graphs, cross-lingual entity alignment, non-isomorphic, relation completion

Procedia PDF Downloads 115
14846 Static Priority Approach to Under-Frequency Based Load Shedding Scheme in Islanded Industrial Networks: Using the Case Study of Fatima Fertilizer Company Ltd - FFL

Authors: S. H. Kazmi, T. Ahmed, K. Javed, A. Ghani

Abstract:

In this paper static scheme of under-frequency based load shedding is considered for chemical and petrochemical industries with islanded distribution networks relying heavily on the primary commodity to ensure minimum production loss, plant downtime or critical equipment shutdown. A simplistic methodology is proposed for in-house implementation of this scheme using underfrequency relays and a step by step guide is provided including the techniques to calculate maximum percentage overloads, frequency decay rates, time based frequency response and frequency based time response of the system. Case study of FFL electrical system is utilized, presenting the actual system parameters and employed load shedding settings following the similar series of steps. The arbitrary settings are then verified for worst overload conditions (loss of a generation source in this case) and comprehensive system response is then investigated.

Keywords: islanding, under-frequency load shedding, frequency rate of change, static UFLS

Procedia PDF Downloads 479
14845 Exchange Rate Forecasting by Econometric Models

Authors: Zahid Ahmad, Nosheen Imran, Nauman Ali, Farah Amir

Abstract:

The objective of the study is to forecast the US Dollar and Pak Rupee exchange rate by using time series models. For this purpose, daily exchange rates of US and Pakistan for the period of January 01, 2007 - June 2, 2017, are employed. The data set is divided into in sample and out of sample data set where in-sample data are used to estimate as well as forecast the models, whereas out-of-sample data set is exercised to forecast the exchange rate. The ADF test and PP test are used to make the time series stationary. To forecast the exchange rate ARIMA model and GARCH model are applied. Among the different Autoregressive Integrated Moving Average (ARIMA) models best model is selected on the basis of selection criteria. Due to the volatility clustering and ARCH effect the GARCH (1, 1) is also applied. Results of analysis showed that ARIMA (0, 1, 1 ) and GARCH (1, 1) are the most suitable models to forecast the future exchange rate. Further the GARCH (1,1) model provided the volatility with non-constant conditional variance in the exchange rate with good forecasting performance. This study is very useful for researchers, policymakers, and businesses for making decisions through accurate and timely forecasting of the exchange rate and helps them in devising their policies.

Keywords: exchange rate, ARIMA, GARCH, PAK/USD

Procedia PDF Downloads 550
14844 Hydrodynamics of Dual Hybrid Impeller of Stirred Reactor Using Radiotracer

Authors: Noraishah Othman, Siti K. Kamarudin, Norinsan K. Othman, Mohd S. Takriff, Masli I. Rosli, Engku M. Fahmi, Mior A. Khusaini

Abstract:

The present work describes hydrodynamics of mixing characteristics of two dual hybrid impeller consisting of, radial and axial impeller using radiotracer technique. Type A mixer, a Rushton turbine is mounted above a Pitched Blade Turbine (PBT) at common shaft and Type B mixer, a Rushton turbine is mounted below PBT. The objectives of this paper are to investigate the residence time distribution (RTD) of two hybrid mixers and to represent the respective mixers by RTD model. Each type of mixer will experience five radiotracer experiments using Tc99m as source of tracer and scintillation detectors NaI(Tl) are used for tracer detection. The results showed that mixer in parallel model and mixers in series with exchange can represent the flow model in mixer A whereas only mixer in parallel model can represent Type B mixer well than other models. In conclusion, Type A impeller, Rushton impeller above PBT, reduced the presence of dead zone in the mixer significantly rather than Type B.

Keywords: hybrid impeller, residence time distribution (RTD), radiotracer experiments, RTD model

Procedia PDF Downloads 350
14843 Comparing Nonverbal Deception Detection of Police Officers and Human Resources Students in the Czech Republic

Authors: Lenka Mynaříková, Hedvika Boukalová

Abstract:

The study looks at the ability to detect nonverbal deception among police officers and management students in the Czech Republic. Respondents from police departments (n=197) and university students of human resources (n=161) completed a deception detection task and evaluated veracity of the statements of suspects in 21 video clips from real crime investigations. Their evaluations were based on nonverbal behavior. Voices in the video clips were modified so that words were not recognizable, yet paraverbal voice characteristics were preserved. Results suggest that respondents have a tendency to lie bias based on their profession. In the evaluation of video clips, stereotypes also played a significant role. The statements of suspects of a different ethnicity, younger age or specific visual features were considered deceitful more often. Research might be beneficial for training in professions that are in need of deception detection techniques.

Keywords: deception detection, police officers, human resources, forensic psychology, forensic studies, organizational psychology

Procedia PDF Downloads 423
14842 Quantum Technologies, the Practical Challenges to It, and Ideas to Build an Inclusive Quantum Platform, Shoonya Ecosystem (Zero-Point Energy)

Authors: Partha Pratim Kalita

Abstract:

As sound can be converted to light, light can also be deduced to sound. There are technologies to convert light to sound, but there are not many technologies related to the field where sound can be converted to a distinct vibrational sequence of light. Like the laws under which the principles of sound work, there are principles for the light to become quantum in nature. Thus, as we move from sound to the subtler aspects of light, we are moving from 3D to 5D. Either we will be making technologies of 3D in today’s world, or we will be really interested in making technologies of the 5D, depends on our understanding of how quantum 5D works. Right now, the entire world is talking about quantum, which is about the nature and behavior of subatomic particles, which is 5D. In practice, they are using metals and machines based on atomic structures. If we talk of quantum without taking note of the technologies of 5D and beyond, we will only be reinterpreting relative theories in the name of quantum. This paper, therefore, will explore the possibilities of moving towards quantum in its real essence with the Shoonya ecosystem (zero-point energy). In this context, the author shall highlight certain working models developed by him, which are currently in discussion with the Indian government.

Keywords: quantum mechanics, quantum technologies, healthcare, shoonya ecosystem, energy, human consciousness

Procedia PDF Downloads 188
14841 The Mechanisms of Peer-Effects in Education: A Frame-Factor Analysis of Instruction

Authors: Pontus Backstrom

Abstract:

In the educational literature on peer effects, attention has been brought to the fact that the mechanisms creating peer effects are still to a large extent hidden in obscurity. The hypothesis in this study is that the Frame Factor Theory can be used to explain these mechanisms. At heart of the theory is the concept of “time needed” for students to learn a certain curricula unit. The relations between class-aggregated time needed and the actual time available, steers and hinders the actions possible for the teacher. Further, the theory predicts that the timing and pacing of the teachers’ instruction is governed by a “criterion steering group” (CSG), namely the pupils in the 10th-25th percentile of the aptitude distribution in class. The class composition hereby set the possibilities and limitations for instruction, creating peer effects on individual outcomes. To test if the theory can be applied to the issue of peer effects, the study employs multilevel structural equation modelling (M-SEM) on Swedish TIMSS 2015-data (Trends in International Mathematics and Science Study; students N=4090, teachers N=200). Using confirmatory factor analysis (CFA) in the SEM-framework in MPLUS, latent variables are specified according to the theory, such as “limitations of instruction” from TIMSS survey items. The results indicate a good model fit to data of the measurement model. Research is still in progress, but preliminary results from initial M-SEM-models verify a strong relation between the mean level of the CSG and the latent variable of limitations on instruction, a variable which in turn have a great impact on individual students’ test results. Further analysis is required, but so far the analysis indicates a confirmation of the predictions derived from the frame factor theory and reveals that one of the important mechanisms creating peer effects in student outcomes is the effect the class composition has upon the teachers’ instruction in class.

Keywords: compositional effects, frame factor theory, peer effects, structural equation modelling

Procedia PDF Downloads 127
14840 Generalized Up-downlink Transmission using Black-White Hole Entanglement Generated by Two-level System Circuit

Authors: Muhammad Arif Jalil, Xaythavay Luangvilay, Montree Bunruangses, Somchat Sonasang, Preecha Yupapin

Abstract:

Black and white holes form the entangled pair⟨BH│WH⟩, where a white hole occurs when the particle moves at the same speed as light. The entangled black-white hole pair is at the center with the radian between the gap. When the speed of particle motion is slower than light, the black hole is gravitational (positive gravity), where the white hole is smaller than the black hole. On the downstream side, the entangled pair appears to have a black hole outside the gap increases until the white holes disappear, which is the emptiness paradox. On the upstream side, when moving faster than light, white holes form times tunnels, with black holes becoming smaller. It will continue to move faster and further when the black hole disappears and becomes a wormhole (Singularity) that is only a white hole in emptiness (Emptiness). This research studies use of black and white holes generated by a two-level circuit for communication transmission carriers, in which high ability and capacity of data transmission can be obtained. The black and white hole pair can be generated by the two-level system circuit when the speech of a particle on the circuit is equal to the speed of light. The black hole forms when the particle speed has increased from slower to equal to the light speed, while the white hole is established when the particle comes down faster than light. They are bound by the entangled pair, signal and idler, ⟨Signal│Idler⟩, and the virtual ones for the white hole, which has an angular displacement of half of π radian. A two-level system is made from an electronic circuit to create black and white holes bound by the entangled bits that are immune or cloning-free from thieves. Start by creating a wave-particle behavior when its speed is equal to light black hole is in the middle of the entangled pair, which is the two bit gate. The required information can be input into the system and wrapped by the black hole carrier. A timeline (Tunnel) occurs when the wave-particle speed is faster than light, from which the entangle pair is collapsed. The transmitted information is safely in the time tunnel. The required time and space can be modulated via the input for the downlink operation. The downlink is established when the particle speed is given by a frequency(energy) form is down and entered into the entangled gap, where this time the white hole is established. The information with the required destination is wrapped by the white hole and retrieved by the clients at the destination. The black and white holes are disappeared, and the information can be recovered and used.

Keywords: cloning free, time machine, teleportation, two-level system

Procedia PDF Downloads 68
14839 Hydrogen Storage Systems for Enhanced Grid Balancing Services in Wind Energy Conversion Systems

Authors: Nezmin Kayedpour, Arash E. Samani, Siavash Asiaban, Jeroen M. De Kooning, Lieven Vandevelde, Guillaume Crevecoeur

Abstract:

The growing adoption of renewable energy sources, such as wind power, in electricity generation is a significant step towards a sustainable and decarbonized future. However, the inherent intermittency and uncertainty of wind resources pose challenges to the reliable and stable operation of power grids. To address this, hydrogen storage systems have emerged as a promising and versatile technology to support grid balancing services in wind energy conversion systems. In this study, we propose a supplementary control design that enhances the performance of the hydrogen storage system by integrating wind turbine (WT) pitch and torque control systems. These control strategies aim to optimize the hydrogen production process, ensuring efficient utilization of wind energy while complying with grid requirements. The wind turbine pitch control system plays a crucial role in managing the turbine's aerodynamic performance. By adjusting the blade pitch angle, the turbine's rotational speed and power output can be regulated. Our proposed control design dynamically coordinates the pitch angle to match the wind turbine's power output with the optimal hydrogen production rate. This ensures that the electrolyzer receives a steady and optimal power supply, avoiding unnecessary strain on the system during high wind speeds and maximizing hydrogen production during low wind speeds. Moreover, the wind turbine torque control system is incorporated to facilitate efficient operation at varying wind speeds. The torque control system optimizes the energy capture from the wind while limiting mechanical stress on the turbine components. By harmonizing the torque control with hydrogen production requirements, the system maintains stable wind turbine operation, thereby enhancing the overall energy-to-hydrogen conversion efficiency. To enable grid-friendly operation, we introduce a cascaded controller that regulates the electrolyzer's electrical power-current in accordance with grid requirements. This controller ensures that the hydrogen production rate can be dynamically adjusted based on real-time grid demands, supporting grid balancing services effectively. By maintaining a close relationship between the wind turbine's power output and the electrolyzer's current, the hydrogen storage system can respond rapidly to grid fluctuations and contribute to enhanced grid stability. In this paper, we present a comprehensive analysis of the proposed supplementary control design's impact on the overall performance of the hydrogen storage system in wind energy conversion systems. Through detailed simulations and case studies, we assess the system's ability to provide grid balancing services, maximize wind energy utilization, and reduce greenhouse gas emissions.

Keywords: active power control, electrolyzer, grid balancing services, wind energy conversion systems

Procedia PDF Downloads 77
14838 Computer Aided Diagnostic System for Detection and Classification of a Brain Tumor through MRI Using Level Set Based Segmentation Technique and ANN Classifier

Authors: Atanu K Samanta, Asim Ali Khan

Abstract:

Due to the acquisition of huge amounts of brain tumor magnetic resonance images (MRI) in clinics, it is very difficult for radiologists to manually interpret and segment these images within a reasonable span of time. Computer-aided diagnosis (CAD) systems can enhance the diagnostic capabilities of radiologists and reduce the time required for accurate diagnosis. An intelligent computer-aided technique for automatic detection of a brain tumor through MRI is presented in this paper. The technique uses the following computational methods; the Level Set for segmentation of a brain tumor from other brain parts, extraction of features from this segmented tumor portion using gray level co-occurrence Matrix (GLCM), and the Artificial Neural Network (ANN) to classify brain tumor images according to their respective types. The entire work is carried out on 50 images having five types of brain tumor. The overall classification accuracy using this method is found to be 98% which is significantly good.

Keywords: brain tumor, computer-aided diagnostic (CAD) system, gray-level co-occurrence matrix (GLCM), tumor segmentation, level set method

Procedia PDF Downloads 502
14837 Framework for Socio-Technical Issues in Requirements Engineering for Developing Resilient Machine Vision Systems Using Levels of Automation through the Lifecycle

Authors: Ryan Messina, Mehedi Hasan

Abstract:

This research is to examine the impacts of using data to generate performance requirements for automation in visual inspections using machine vision. These situations are intended for design and how projects can smooth the transfer of tacit knowledge to using an algorithm. We have proposed a framework when specifying machine vision systems. This framework utilizes varying levels of automation as contingency planning to reduce data processing complexity. Using data assists in extracting tacit knowledge from those who can perform the manual tasks to assist design the system; this means that real data from the system is always referenced and minimizes errors between participating parties. We propose using three indicators to know if the project has a high risk of failing to meet requirements related to accuracy and reliability. All systems tested achieved a better integration into operations after applying the framework.

Keywords: automation, contingency planning, continuous engineering, control theory, machine vision, system requirements, system thinking

Procedia PDF Downloads 194
14836 Modeling and Optimization of Algae Oil Extraction Using Response Surface Methodology

Authors: I. F. Ejim, F. L. Kamen

Abstract:

Aims: In this experiment, algae oil extraction with a combination of n-hexane and ethanol was investigated. The effects of extraction solvent concentration, extraction time and temperature on the yield and quality of oil were studied using Response Surface Methodology (RSM). Experimental Design: Optimization of algae oil extraction using Box-Behnken design was used to generate 17 experimental runs in a three-factor-three-level design where oil yield, specific gravity, acid value and saponification value were evaluated as the response. Result: In this result, a minimum oil yield of 17% and maximum of 44% was realized. The optimum values for yield, specific gravity, acid value and saponification value from the overlay plot were 40.79%, 0.8788, 0.5056 mg KOH/g and 180.78 mg KOH/g respectively with desirability of 0.801. The maximum point prediction was yield 40.79% at solvent concentration 66.68 n-hexane, temperature of 40.0°C and extraction time of 4 hrs. Analysis of Variance (ANOVA) results showed that the linear and quadratic coefficient were all significant at p<0.05. The experiment was validated and results obtained were with the predicted values. Conclusion: Algae oil extraction was successfully optimized using RSM and its quality indicated it is suitable for many industrial uses.

Keywords: algae oil, response surface methodology, optimization, Box-Bohnken, extraction

Procedia PDF Downloads 329
14835 A Two Level Load Balancing Approach for Cloud Environment

Authors: Anurag Jain, Rajneesh Kumar

Abstract:

Cloud computing is the outcome of rapid growth of internet. Due to elastic nature of cloud computing and unpredictable behavior of user, load balancing is the major issue in cloud computing paradigm. An efficient load balancing technique can improve the performance in terms of efficient resource utilization and higher customer satisfaction. Load balancing can be implemented through task scheduling, resource allocation and task migration. Various parameters to analyze the performance of load balancing approach are response time, cost, data processing time and throughput. This paper demonstrates a two level load balancer approach by combining join idle queue and join shortest queue approach. Authors have used cloud analyst simulator to test proposed two level load balancer approach. The results are analyzed and compared with the existing algorithms and as observed, proposed work is one step ahead of existing techniques.

Keywords: cloud analyst, cloud computing, join idle queue, join shortest queue, load balancing, task scheduling

Procedia PDF Downloads 422
14834 Understanding the Impact of Climate-Induced Rural-Urban Migration on the Technical Efficiency of Maize Production in Malawi

Authors: Innocent Pangapanga-Phiri, Eric Dada Mungatana

Abstract:

This study estimates the effect of climate-induced rural-urban migrants (RUM) on maize productivity. It uses panel data gathered by the National Statistics Office and the World Bank to understand the effect of RUM on the technical efficiency of maize production in rural Malawi. The study runs the two-stage Tobit regression to isolate the real effect of rural-urban migration on the technical efficiency of maize production. The results show that RUM significantly reduces the technical efficiency of maize production. However, the interaction of RUM and climate-smart agriculture has a positive and significant influence on the technical efficiency of maize production, suggesting the need for re-investing migrants’ remittances in agricultural activities.

Keywords: climate-smart agriculture, farm productivity, rural-urban migration, panel stochastic frontier models, two-stage Tobit regression

Procedia PDF Downloads 114
14833 Material Properties Evolution Affecting Demisability for Space Debris Mitigation

Authors: Chetan Mahawar, Sarath Chandran, Sridhar Panigrahi, V. P. Shaji

Abstract:

The ever-growing advancement in space exploration has led to an alarming concern for space debris removal as it restricts further launch operations and adventurous space missions; hence numerous studies have come up with technologies for re-entry predictions and material selection processes for mitigating space debris. The selection of material and operating conditions is determined with the objective of lightweight structure and ability to demise faster subject to spacecraft survivability during its mission. Since the demisability of spacecraft depends on evolving thermal material properties such as emissivity, specific heat capacity, thermal conductivity, radiation intensity, etc. Therefore, this paper presents the analysis of evolving thermal material properties of spacecraft, which affect the demisability process and thus estimate demise time using the demisability model by incorporating evolving thermal properties for sensible heating followed by the complete or partial break-up of spacecraft. The demisability analysis thus concludes the best suitable spacecraft material is based on the least estimated demise time, which fulfills the criteria of design-for-survivability and as well as of design-for-demisability.

Keywords: demisability, emissivity, lightweight, re-entry, survivability

Procedia PDF Downloads 105
14832 Evaluating the Total Costs of a Ransomware-Resilient Architecture for Healthcare Systems

Authors: Sreejith Gopinath, Aspen Olmsted

Abstract:

This paper is based on our previous work that proposed a risk-transference-based architecture for healthcare systems to store sensitive data outside the system boundary, rendering the system unattractive to would-be bad actors. This architecture also allows a compromised system to be abandoned and a new system instance spun up in place to ensure business continuity without paying a ransom or engaging with a bad actor. This paper delves into the details of various attacks we simulated against the prototype system. In the paper, we discuss at length the time and computational costs associated with storing and retrieving data in the prototype system, abandoning a compromised system, and setting up a new instance with existing data. Lastly, we simulate some analytical workloads over the data stored in our specialized data storage system and discuss the time and computational costs associated with running analytics over data in a specialized storage system outside the system boundary. In summary, this paper discusses the total costs of data storage, access, and analytics incurred with the proposed architecture.

Keywords: cybersecurity, healthcare, ransomware, resilience, risk transference

Procedia PDF Downloads 128
14831 An Approach to Tackle Start up Problems Using Applied Games

Authors: Aiswarya Gopal, Kamal Bijlani, Vinoth Rengaraj, R. Jayakrishnan

Abstract:

In the business world, the term “startup” is frequently ringing the bell with the high frequency of young ventures. The main dilemma of startups is the unsuccessful management of the unique risks that have to be confronted in the present world of competition and technology. This research work tried to bring out a game based methodology to improve enough real-world experience among entrepreneurs as well as management students to handle risks and challenges in the field. The game will provide experience to the player to overcome challenges like market problems, running out of cash, poor management, and product problems which can be resolved by a proper strategic approach in the entrepreneurship world. The proposed serious game works on the life cycle of a new software enterprise where the entrepreneur moves from the planning stage to secured financial stage, laying down the basic business structure, and initiates the operations ensuring the increment in confidence level of the player.

Keywords: business model, game based learning, poor management, start up

Procedia PDF Downloads 467
14830 Re-Evaluation of Field X Located in Northern Lake Albert Basin to Refine the Structural Interpretation

Authors: Calorine Twebaze, Jesca Balinga

Abstract:

Field X is located on the Eastern shores of L. Albert, Uganda, on the rift flank where the gross sedimentary fill is typically less than 2,000m. The field was discovered in 2006 and encountered about 20.4m of net pay across three (3) stratigraphic intervals within the discovery well. The field covers an area of 3 km2, with the structural configuration comprising a 3-way dip-closed hanging wall anticline that seals against the basement to the southeast along the bounding fault. Field X had been mapped on reprocessed 3D seismic data, which was originally acquired in 2007 and reprocessed in 2013. The seismic data quality is good across the field, and reprocessing work reduced the uncertainty in the location of the bounding fault and enhanced the lateral continuity of reservoir reflectors. The current study was a re-evaluation of Field X to refine fault interpretation and understand the structural uncertainties associated with the field. The seismic data, and three (3) wells datasets were used during the study. The evaluation followed standard workflows using Petrel software and structural attribute analysis. The process spanned from seismic- -well tie, structural interpretation, and structural uncertainty analysis. Analysis of three (3) well ties generated for the 3 wells provided a geophysical interpretation that was consistent with geological picks. The generated time-depth curves showed a general increase in velocity with burial depth. However, separation in curve trends observed below 1100m was mainly attributed to minimal lateral variation in velocity between the wells. In addition to Attribute analysis, three velocity modeling approaches were evaluated, including the Time-Depth Curve, Vo+ kZ, and Average Velocity Method. The generated models were calibrated at well locations using well tops to obtain the best velocity model for Field X. The Time-depth method resulted in more reliable depth surfaces with good structural coherence between the TWT and depth maps with minimal error at well locations of 2 to 5m. Both the NNE-SSW rift border fault and minor faults in the existing interpretation were reevaluated. However, the new interpretation delineated an E-W trending fault in the northern part of the field that had not been interpreted before. The fault was interpreted at all stratigraphic levels and thus propagates from the basement to the surface and is an active fault today. It was also noted that the entire field is less faulted with more faults in the deeper part of the field. The major structural uncertainties defined included 1) The time horizons due to reduced data quality, especially in the deeper parts of the structure, an error equal to one-third of the reflection time thickness was assumed, 2) Check shot analysis showed varying velocities within the wells thus varying depth values for each well, and 3) Very few average velocity points due to limited wells produced a pessimistic average Velocity model.

Keywords: 3D seismic data interpretation, structural uncertainties, attribute analysis, velocity modelling approaches

Procedia PDF Downloads 46
14829 Association between a Forward Lag of Historical Total Accumulated Gasoline Lead Emissions and Contemporary Autism Prevalence Trends in California, USA

Authors: Mark A. S. Laidlaw, Howard W. Mielke

Abstract:

In California between the late 1920’s and 1986 the lead concentrations in urban soils and dust climbed rapidly following the deposition of greater than 387,000 tonnes of lead emitted from gasoline. Previous research indicates that when children are lead exposed around 90% of the lead is retained in their bones and teeth due to the substitution of lead for calcium. Lead in children’s bones has been shown to accumulate over time and is highest in inner-city urban areas, lower in suburban areas and lowest in rural areas. It is also known that women’s bones demineralize during pregnancy due to the foetus's high demand for calcium. Lead accumulates in women’s bones during childhood and the accumulated lead is subsequently released during pregnancy – a lagged response. This results in calcium plus lead to enter the blood stream and cross the placenta to expose the foetus with lead. In 1970 in the United States, the average age of a first‐time mother was about 21. In 2008, the average age was 25.1. In this study, it is demonstrated that in California there is a forward lagged relationship between the accumulated emissions of lead from vehicle fuel additives and later autism prevalence trends between the 1990’s and current time period. Regression analysis between a 24 year forward lag of accumulated lead emissions and autism prevalence trends in California are associated strongly (R2=0.95, p=0.00000000127). It is hypothesized that autism in genetically susceptible children may stem from vehicle fuel lead emission exposures of their mothers during childhood and that the release of stored lead during subsequent pregnancy resulted in lead exposure of foetuses during a critical developmental period. It is furthermore hypothesized that the 24 years forward lag between lead exposures has occurred because that is time period is the average length for women to enter childbearing age. To test the hypothesis that lead in mothers bones is associated with autism, it is hypothesized that retrospective case-control studies would show an association between the lead in mother’s bones and autism. Furthermore, it is hypothesized that the forward lagged relationship between accumulated historical vehicle fuel lead emissions (or air lead concentrations) and autism prevalence trends will be similar in cities at the national and international scale. If further epidemiological studies indicate a strong relationship between accumulated vehicle fuel lead emissions (or accumulated air lead concentrations) and lead in mother’s bones and autism rates, then urban areas may require extensive soil intervention to prevent the development of autism in children.

Keywords: autism, bones, lead, gasoline, petrol, prevalence

Procedia PDF Downloads 290
14828 The Popular Imagination through the Poem of “Ras B’Nadam”

Authors: Hirreche Baghdad Mohamed

Abstract:

One of the main texts in popular culture in Algeria is a symbolic and imaginary tale, through which the author was able to derive from the world and popular cultural stock and symbolic capital elements that enabled him to create a synthesis between a number of imaginary and real events. Thanks to the level of spirituality that the author was experiencing, he was able to go deep in order to redraw the boundaries of human life in view of its existence and status (life experiences, its end, and its fate). It is a text that is consistent with religious values and has a philosophical depth. This poem can be shared in official and unofficial meetings, during feasts, and during popular celebrations, such as circumcision ceremonies, marriage, and condolences. It has also the ability to draw attention and appeal to the listener and let him travel into the imaginary world. It is the text related to the story of "Ras b’nadem", or "the head of a man", or rather, a "human skull", for which only a few academic studies have been devoted, and there are two copies of it, one attributed to Lakhdar Ibn Khalouf as a matter of suspicion, while the other is attributed to Qadour Ibn Ashour Al-Zarhouni.

Keywords: ras B’Nadam, ras al mahna, lakhdar ibn khalouf, qadour ibn ashour, sufism, melhoun poetry, resistance poetry

Procedia PDF Downloads 182