Search results for: SoC soft error rate
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10419

Search results for: SoC soft error rate

9099 Digital Platform for Psychological Assessment Supported by Sensors and Efficiency Algorithms

Authors: Francisco M. Silva

Abstract:

Technology is evolving, creating an impact on our everyday lives and the telehealth industry. Telehealth encapsulates the provision of healthcare services and information via a technological approach. There are several benefits of using web-based methods to provide healthcare help. Nonetheless, few health and psychological help approaches combine this method with wearable sensors. This paper aims to create an online platform for users to receive self-care help and information using wearable sensors. In addition, researchers developing a similar project obtain a solid foundation as a reference. This study provides descriptions and analyses of the software and hardware architecture. Exhibits and explains a heart rate dynamic and efficient algorithm that continuously calculates the desired sensors' values. Presents diagrams that illustrate the website deployment process and the webserver means of handling the sensors' data. The goal is to create a working project using Arduino compatible hardware. Heart rate sensors send their data values to an online platform. A microcontroller board uses an algorithm to calculate the sensor heart rate values and outputs it to a web server. The platform visualizes the sensor's data, summarizes it in a report, and creates alerts for the user. Results showed a solid project structure and communication from the hardware and software. The web server displays the conveyed heart rate sensor's data on the online platform, presenting observations and evaluations.

Keywords: Arduino, heart rate BPM, microcontroller board, telehealth, wearable sensors, web-based healthcare

Procedia PDF Downloads 126
9098 Interest Rate Prediction with Taylor Rule

Authors: T. Bouchabchoub, A. Bendahmane, A. Haouriqui, N. Attou

Abstract:

This paper presents simulation results of Forex predicting model equations in order to give approximately a prevision of interest rates. First, Hall-Taylor (HT) equations have been used with Taylor rule (TR) to adapt them to European and American Forex Markets. Indeed, initial Taylor Rule equation is conceived for all Forex transactions in every States: It includes only one equation and six parameters. Here, the model has been used with Hall-Taylor equations, initially including twelve equations which have been reduced to only three equations. Analysis has been developed on the following base macroeconomic variables: Real change rate, investment wages, anticipated inflation, realized inflation, real production, interest rates, gap production and potential production. This model has been used to specifically study the impact of an inflation shock on macroeconomic director interest rates.

Keywords: interest rate, Forex, Taylor rule, production, European Central Bank (ECB), Federal Reserve System (FED).

Procedia PDF Downloads 527
9097 Control of Indoor Carbon through Soft Approaches in Himachal Pradesh, India

Authors: Kopal Verma, Umesh C. Kulshrestha

Abstract:

The mountainous regions are very crucial for a country because of their importance for weather, water supply, forests, and various other socio-economic benefits. But the increasing population and its demand for energy and infrastructure have contributed very high loadings of air pollution. Various activities such as cooking, heating, manufacturing, transport, etc. contribute various particulate and gaseous pollutants in the atmosphere. This study was focused upon indoor air pollution and was carried out in four rural households of the Baggi village located in the Hamirpur District of the Himachal Pradesh state. The residents of Baggi village use biomass as fuel for cooking on traditional stove (Chullah). The biomass types include wood (mainly Beul, Grewia Optiva), crop residue and dung cakes. This study aimed to determine the organic carbon (OC), elemental carbon (EC), major cations and anions in the indoor air of each household. During non-cooking hours, it was found that the indoor air contained OC and EC as low as 21µg/m³ and 17µg/m³ respectively. But during cooking hours (with biomass burning), the levels of OC and EC were raised significantly by 91.2% and 85.4% respectively. Then the residents were advised to switch over as per our soft approach options. In the first approach change, they were asked to prepare the meal partially on Chullah using biomass and partially with liquefied petroleum gas (LPG). By doing this change, a considerable reduction in OC (53.1%) and in EC (41.8%) was noticed. The second change of approach included the cooking of entire meal by using LPG. This resulted in the reduction of OC (84.1%) and EC (73.3%) as compared to the values obtained during cooking entirely with biomass. The carbonaceous aerosol levels were higher in the morning hours than in the evening hours because of more biomass burning activity in the morning. According to a general survey done with the residents, the study provided them an awareness about the air pollution and the harmful effects of biomass burning. Some of them correlated their ailments like weakened eyesight, fatigue and respiratory problems with indoor air pollution. This study demonstrated that by replacing biomass with clean fuel such as LPG, the indoor concentrations of EC and OC can be reduced substantially.

Keywords: biomass burning, carbonaceous aerosol, elemental carbon, organic carbon, LPG

Procedia PDF Downloads 118
9096 Leaf Photosynthesis and Water-Use Efficiency of Diverse Legume Species Nodulated by Native Rhizobial Isolates in the Glasshouse

Authors: Lebogang Jane Msiza, Felix Dapare Dakora

Abstract:

Photosynthesis is a process by which plants convert light energy to chemical energy for metabolic processes. Plants are known for converting inorganic CO₂ in the atmosphere to organic C by photosynthesis. A decrease in stomatal conductance causes a decrease in the transpiration rate of leaves, thus increasing the water-use efficiency of plants. Water-use efficiency in plants is conditioned by soil moisture availability and is enhanced under conditions of water deficit. This study evaluated leaf photosynthesis and water-use efficiency in 12 legume species inoculated with 26 rhizobial isolates from soybean, 15 from common bean, 10 from cowpea, 15 from Bambara groundnut, 7 from lessertia and 10 from Kersting bean. Gas-exchange studies were used to measure photosynthesis and water-use efficiency. The results revealed a much higher photosynthetic rate (20.95µmol CO₂ m-2s-1) induced by isolated tutpres to a lower rate (7.06 µmol CO₂ m-2s-1) by isolate mgsa 88. Stomatal conductance ranged from to 0.01 mmol m-2.s-1 by mgsa 88 to 0.12 mmol m-2.s-1 by isolate da-pua 128. Transpiration rate also ranged from 0.09 mmol m-2.s-1 induced by da-pua B2 to 3.28 mmol m-2.s-1 by da-pua 3, while water-use efficiency ranged from 91.32 µmol CO₂ m-1 H₂O elicited by mgsa 106 to 4655.50 µmol CO₂ m-1 H₂O by isolate tutswz 13. The results revealed the highest photosynthetic rate in soybean and the lowest in common bean, and also with higher stomatal conductance and transpiration rates in jack bean and Bambara groundnut. Pigeonpea exhibited much higher water-use efficiency than all the tested legumes. The findings showed significant differences between and among the test legume/rhizobia combinations. Leaf photosynthetic rates are reported to be higher in legumes with high stomatal conductance, which suggests that legume productivity can be improved by manipulating leaf stomatal conductance.

Keywords: legumes, photosynthetic rate, stomatal conductance, water-use efficiency

Procedia PDF Downloads 228
9095 Study on Sintering System of Calcium Barium Sulphoaluminate by XRD Quantitative Analysis

Authors: Xiaopeng Shang, Xin YU, Jun CHANG

Abstract:

Calcium barium sulphoaluminate (CBSA), derived from calcium sulphoaluminate(CSA), has excellent cementitious properties. In this study, the sintering system of CBSA with a theoretical stoichiometric Ca3BaAl6SO16 was investigated. Rietveld refinement was performed using TOPAS 4.2 software to quantitatively calculate the content of CBSA and the actual ionic site occupancy of Ba2+. The results indicate that the contents of Ca4-xBaxAl6SO16 increases with increasing sintering temperature in the 1200℃-1400℃ ranges. When sintered at 1400℃ for 180min, the content of CBSA reaches 88.4%. However, CBSA begins to decompose at 1440℃ and the content of which decreases. The replacement rate of Ba2+ was also enlarged by increasing sintering temperature and prolonged sintering time. Sintering at 1400℃ for 180min is considered as the optimum when replacement rate of Ba2+ and the content of CBSA were taken into account. Ca3.2Ba0.8Al6SO16 with a content of 88.4% was synthesized.

Keywords: calcium barium sulphoaluminate, sintering system, Ba2+ replacement rate, Rietveld refinement

Procedia PDF Downloads 342
9094 The Impact of Natural Resources on Financial Development: The Global Perspective

Authors: Remy Jonkam Oben

Abstract:

Using a time series approach, this study investigates how natural resources impact financial development from a global perspective over the 1980-2019 period. Some important determinants of financial development (economic growth, trade openness, population growth, and investment) have been added to the model as control variables. Unit root tests have revealed that all the variables are integrated into order one. Johansen's cointegration test has shown that the variables are in a long-run equilibrium relationship. The vector error correction model (VECM) has estimated the coefficient of the error correction term (ECT), which suggests that the short-run values of natural resources, economic growth, trade openness, population growth, and investment contribute to financial development converging to its long-run equilibrium level by a 23.63% annual speed of adjustment. The estimated coefficients suggest that global natural resource rent has a statistically-significant negative impact on global financial development in the long-run (thereby validating the financial resource curse) but not in the short-run. Causality test results imply that neither global natural resource rent nor global financial development Granger-causes each other.

Keywords: financial development, natural resources, resource curse hypothesis, time series analysis, Granger causality, global perspective

Procedia PDF Downloads 170
9093 The Impact of Climate Change on Cropland Ecosystem in Tibet Plateau

Authors: Weishou Shen, Chunyan Yang, Zhongliang Li

Abstract:

The crop climate productivity and the distribution of cropland reflect long-term adaption of agriculture to climate. In order to fully understand the impact of climate change on cropland ecosystem in Tibet, the spatiotemporal changes of crop climate productivity and cropland distribution were analyzed with the help of GIS and RS software. Results indicated that the climate change to the direction of wet and warm in Tibet in the recent 30 years, with a rate of 0.79℃/10 yr and 23.28 mm/10yr respectively. Correspondingly, the climate productivity increased gradually, with a rate of 346.3kg/(hm2•10a), of which, the fastest-growing rate of the crop climate productivity is in Southern Tibet Mountain- plain-valley. During the study period, the total cropland area increased from 32.54 million ha to 37.13 million ha, and cropland has expanded to higher altitude area and northward. Overall, increased cropland area and crop climate productivity due to climate change plays a positive role for agriculture in Tibet.

Keywords: climate change, productivity, cropland area, Tibet plateau

Procedia PDF Downloads 378
9092 Design of an Air and Land Multi-Element Expression Pattern of Navigation Electronic Map for Ground Vehicles under United Navigation Mechanism

Authors: Rui Liu, Pengyu Cui, Nan Jiang

Abstract:

At present, there is much research on the application of centralized management and cross-integration application of basic geographic information. However, the idea of information integration and sharing between land, sea, and air navigation targets is not deeply applied into the research of navigation information service, especially in the information expression. Targeting at this problem, the paper carries out works about the expression pattern of navigation electronic map for ground vehicles under air and land united navigation mechanism. At first, with the support from multi-source information fusion of GIS vector data, RS data, GPS data, etc., an air and land united information expression pattern is designed aiming at specific navigation task of emergency rescue in the earthquake. And then, the characteristics and specifications of the united expression of air and land navigation information under the constraints of map load are summarized and transferred into expression rules in the rule bank. At last, the related navigation experiment is implemented to evaluate the effect of the expression pattern. The experiment selects evaluation factors of the navigation task accomplishment time and the navigation error rate as the main index, and make comparisons with the traditional single information expression pattern. To sum up, the research improved the theory of navigation electronic map and laid a certain foundation for the design and realization of united navigation system in the aspect of real-time navigation information delivery.

Keywords: navigation electronic map, united navigation, multi-element expression pattern, multi-source information fusion

Procedia PDF Downloads 199
9091 Using Deep Learning Real-Time Object Detection Convolution Neural Networks for Fast Fruit Recognition in the Tree

Authors: K. Bresilla, L. Manfrini, B. Morandi, A. Boini, G. Perulli, L. C. Grappadelli

Abstract:

Image/video processing for fruit in the tree using hard-coded feature extraction algorithms have shown high accuracy during recent years. While accurate, these approaches even with high-end hardware are computationally intensive and too slow for real-time systems. This paper details the use of deep convolution neural networks (CNNs), specifically an algorithm (YOLO - You Only Look Once) with 24+2 convolution layers. Using deep-learning techniques eliminated the need for hard-code specific features for specific fruit shapes, color and/or other attributes. This CNN is trained on more than 5000 images of apple and pear fruits on 960 cores GPU (Graphical Processing Unit). Testing set showed an accuracy of 90%. After this, trained data were transferred to an embedded device (Raspberry Pi gen.3) with camera for more portability. Based on correlation between number of visible fruits or detected fruits on one frame and the real number of fruits on one tree, a model was created to accommodate this error rate. Speed of processing and detection of the whole platform was higher than 40 frames per second. This speed is fast enough for any grasping/harvesting robotic arm or other real-time applications.

Keywords: artificial intelligence, computer vision, deep learning, fruit recognition, harvesting robot, precision agriculture

Procedia PDF Downloads 420
9090 Experimental Correlation for Erythrocyte Aggregation Rate in Population Balance Modeling

Authors: Erfan Niazi, Marianne Fenech

Abstract:

Red Blood Cells (RBCs) or erythrocytes tend to form chain-like aggregates under low shear rate called rouleaux. This is a reversible process and rouleaux disaggregate in high shear rates. Therefore, RBCs aggregation occurs in the microcirculation where low shear rates are present but does not occur under normal physiological conditions in large arteries. Numerical modeling of RBCs interactions is fundamental in analytical models of a blood flow in microcirculation. Population Balance Modeling (PBM) is particularly useful for studying problems where particles agglomerate and break in a two phase flow systems to find flow characteristics. In this method, the elementary particles lose their individual identity due to continuous destructions and recreations by break-up and agglomeration. The aim of this study is to find RBCs aggregation in a dynamic situation. Simplified PBM was used previously to find the aggregation rate on a static observation of the RBCs aggregation in a drop of blood under the microscope. To find aggregation rate in a dynamic situation we propose an experimental set up testing RBCs sedimentation. In this test, RBCs interact and aggregate to form rouleaux. In this configuration, disaggregation can be neglected due to low shear stress. A high-speed camera is used to acquire video-microscopic pictures of the process. The sizes of the aggregates and velocity of sedimentation are extracted using an image processing techniques. Based on the data collection from 5 healthy human blood samples, the aggregation rate was estimated as 2.7x103(±0.3 x103) 1/s.

Keywords: red blood cell, rouleaux, microfluidics, image processing, population balance modeling

Procedia PDF Downloads 355
9089 3-D Modeling of Particle Size Reduction from Micro to Nano Scale Using Finite Difference Method

Authors: Himanshu Singh, Rishi Kant, Shantanu Bhattacharya

Abstract:

This paper adopts a top-down approach for mathematical modeling to predict the size reduction from micro to nano-scale through persistent etching. The process is simulated using a finite difference approach. Previously, various researchers have simulated the etching process for 1-D and 2-D substrates. It consists of two processes: 1) Convection-Diffusion in the etchant domain; 2) Chemical reaction at the surface of the particle. Since the process requires analysis along moving boundary, partial differential equations involved cannot be solved using conventional methods. In 1-D, this problem is very similar to Stefan's problem of moving ice-water boundary. A fixed grid method using finite volume method is very popular for modelling of etching on a one and two dimensional substrate. Other popular approaches include moving grid method and level set method. In this method, finite difference method was used to discretize the spherical diffusion equation. Due to symmetrical distribution of etchant, the angular terms in the equation can be neglected. Concentration is assumed to be constant at the outer boundary. At the particle boundary, the concentration of the etchant is assumed to be zero since the rate of reaction is much faster than rate of diffusion. The rate of reaction is proportional to the velocity of the moving boundary of the particle. Modelling of the above reaction was carried out using Matlab. The initial particle size was taken to be 50 microns. The density, molecular weight and diffusion coefficient of the substrate were taken as 2.1 gm/cm3, 60 and 10-5 cm2/s respectively. The etch-rate was found to decline initially and it gradually became constant at 0.02µ/s (1.2µ/min). The concentration profile was plotted along with space at different time intervals. Initially, a sudden drop is observed at the particle boundary due to high-etch rate. This change becomes more gradual with time due to declination of etch rate.

Keywords: particle size reduction, micromixer, FDM modelling, wet etching

Procedia PDF Downloads 431
9088 A Mathematical Model for Studying Landing Dynamics of a Typical Lunar Soft Lander

Authors: Johns Paul, Santhosh J. Nalluveettil, P. Purushothaman, M. Premdas

Abstract:

Lunar landing is one of the most critical phases of lunar mission. The lander is provided with a soft landing system to prevent structural damage of lunar module by absorbing the landing shock and also assure stability during landing. Presently available software are not capable to simulate the rigid body dynamics coupled with contact simulation and elastic/plastic deformation analysis. Hence a separate mathematical model has been generated for studying the dynamics of a typical lunar soft lander. Parameters used in the analysis includes lunar surface slope, coefficient of friction, initial touchdown velocity (vertical and horizontal), mass and moment of inertia of lander, crushing force due to energy absorbing material in the legs, number of legs and geometry of lander. The mathematical model is capable to simulate plastic and elastic deformation of honey comb, frictional force between landing leg and lunar soil, surface contact simulation, lunar gravitational force, rigid body dynamics and linkage dynamics of inverted tripod landing gear. The non linear differential equations generated for studying the dynamics of lunar lander is solved by numerical method. Matlab programme has been used as a computer tool for solving the numerical equations. The position of each kinematic joint is defined by mathematical equations for the generation of equation of motion. All hinged locations are defined by position vectors with respect to body fixed coordinate. The vehicle rigid body rotations and motions about body coordinate are only due to the external forces and moments arise from footpad reaction force due to impact, footpad frictional force and weight of vehicle. All these force are mathematically simulated for the generation of equation of motion. The validation of mathematical model is done by two different phases. First phase is the validation of plastic deformation of crushable elements by employing conservation of energy principle. The second phase is the validation of rigid body dynamics of model by simulating a lander model in ADAMS software after replacing the crushable elements to elastic spring element. Simulation of plastic deformation along with rigid body dynamics and contact force cannot be modeled in ADAMS. Hence plastic element of primary strut is replaced with a spring element and analysis is carried out in ADAMS software. The same analysis is also carried out using the mathematical model where the simulation of honeycomb crushing is replaced by elastic spring deformation and compared the results with ADAMS analysis. The rotational motion of linkages and 6 degree of freedom motion of lunar Lander about its CG can be validated by ADAMS software by replacing crushing element to spring element. The model is also validated by the drop test results of 4 leg lunar lander. This paper presents the details of mathematical model generated and its validation.

Keywords: honeycomb, landing leg tripod, lunar lander, primary link, secondary link

Procedia PDF Downloads 351
9087 Movie Genre Preference Prediction Using Machine Learning for Customer-Based Information

Authors: Haifeng Wang, Haili Zhang

Abstract:

Most movie recommendation systems have been developed for customers to find items of interest. This work introduces a predictive model usable by small and medium-sized enterprises (SMEs) who are in need of a data-based and analytical approach to stock proper movies for local audiences and retain more customers. We used classification models to extract features from thousands of customers’ demographic, behavioral and social information to predict their movie genre preference. In the implementation, a Gaussian kernel support vector machine (SVM) classification model and a logistic regression model were established to extract features from sample data and their test error-in-sample were compared. Comparison of error-out-sample was also made under different Vapnik–Chervonenkis (VC) dimensions in the machine learning algorithm to find and prevent overfitting. Gaussian kernel SVM prediction model can correctly predict movie genre preferences in 85% of positive cases. The accuracy of the algorithm increased to 93% with a smaller VC dimension and less overfitting. These findings advance our understanding of how to use machine learning approach to predict customers’ preferences with a small data set and design prediction tools for these enterprises.

Keywords: computational social science, movie preference, machine learning, SVM

Procedia PDF Downloads 260
9086 Analyzing the Effect of Design of Pipe in Shell and Tube Type Heat Exchanger by Measuring Its Heat Transfer Rate by Computation Fluid Dynamics and Thermal Approach

Authors: Dhawal Ladani

Abstract:

Shell and tube type heat exchangers are predominantly used in heat exchange between two fluids and other applications. This paper projects the optimal design of the pipe used in the heat exchanger in such a way to minimize the vibration occurring in the pipe. Paper also consists of the comparison of the different design of the pipe to get the maximize the heat transfer rate by converting laminar flow into the turbulent flow. By the updated design the vibration in the pipe due to the flow is also decreased. Computational Fluid Dynamics and Thermal Heat Transfer analysis are done to justifying the result. Currently, the straight pipe is used in the shell and tube type of heat exchanger where as per the paper the pipe consists of the curvature along with the pipe. Hence, the heat transfer area is also increased and result in the increasing in heat transfer rate. Curvature type design is useful to create turbulence and minimizing the vibration, also. The result will give the output comparison of the effect of laminar flow and the turbulent flow in the heat exchange mechanism, as well as, inverse effect of the boundary layer in heat exchanger is also justified.

Keywords: heat exchanger, heat transfer rate, laminar and turbulent effect, shell and tube

Procedia PDF Downloads 307
9085 Feasibility of Simulating External Vehicle Aerodynamics Using Spalart-Allmaras Turbulence Model with Adjoint Method in OpenFOAM and Fluent

Authors: Arpit Panwar, Arvind Deshpande

Abstract:

The study of external vehicle aerodynamics using Spalart-Allmaras turbulence model with adjoint method was conducted. The accessibility and ease of working with the Fluent module of ANSYS and OpenFOAM were considered. The objective of the study was to understand and analyze the possibility of bringing high-level aerodynamic simulation to the average consumer vehicle. A form-factor of BMW M6 vehicle was designed in Solidworks, which was analyzed in OpenFOAM and Fluent. The turbulence model being a single equation provides much faster convergence rate when clubbed with the adjoint method. Fluent being commercial software still does not allow us to solve Spalart-Allmaras turbulence model using the adjoint method. Hence, the turbulence model was solved using the SIMPLE method in Fluent. OpenFOAM being an open source provide flexibility in simulation but is not user-friendly. It supports solving the defined turbulence model with the adjoint method. The result generated from the simulation gives us acceptable values of drag, when validated with the result of percentage error in drag values for a notch-back vehicle model on an extensive simulation produced at 6th ANSA and μETA conference, Greece. The success of this approach will allow us to bring more aerodynamic vehicle body design to all segments of the automobile and not limiting it to just the high-end sports cars.

Keywords: Spalart-Allmaras turbulence model, OpenFOAM, adjoint method, SIMPLE method, vehicle aerodynamic design

Procedia PDF Downloads 200
9084 Reliability Analysis of Geometric Performance of Onboard Satellite Sensors: A Study on Location Accuracy

Authors: Ch. Sridevi, A. Chalapathi Rao, P. Srinivasulu

Abstract:

The location accuracy of data products is a critical parameter in assessing the geometric performance of satellite sensors. This study focuses on reliability analysis of onboard sensors to evaluate their performance in terms of location accuracy performance over time. The analysis utilizes field failure data and employs the weibull distribution to determine the reliability and in turn to understand the improvements or degradations over a period of time. The analysis begins by scrutinizing the location accuracy error which is the root mean square (RMS) error of differences between ground control point coordinates observed on the product and the map and identifying the failure data with reference to time. A significant challenge in this study is to thoroughly analyze the possibility of an infant mortality phase in the data. To address this, the Weibull distribution is utilized to determine if the data exhibits an infant stage or if it has transitioned into the operational phase. The shape parameter beta plays a crucial role in identifying this stage. Additionally, determining the exact start of the operational phase and the end of the infant stage poses another challenge as it is crucial to eliminate residual infant mortality or wear-out from the model, as it can significantly increase the total failure rate. To address this, an approach utilizing the well-established statistical Laplace test is applied to infer the behavior of sensors and to accurately ascertain the duration of different phases in the lifetime and the time required for stabilization. This approach also helps in understanding if the bathtub curve model, which accounts for the different phases in the lifetime of a product, is appropriate for the data and whether the thresholds for the infant period and wear-out phase are accurately estimated by validating the data in individual phases with Weibull distribution curve fitting analysis. Once the operational phase is determined, reliability is assessed using Weibull analysis. This analysis not only provides insights into the reliability of individual sensors with regards to location accuracy over the required period of time, but also establishes a model that can be applied to automate similar analyses for various sensors and parameters using field failure data. Furthermore, the identification of the best-performing sensor through this analysis serves as a benchmark for future missions and designs, ensuring continuous improvement in sensor performance and reliability. Overall, this study provides a methodology to accurately determine the duration of different phases in the life data of individual sensors. It enables an assessment of the time required for stabilization and provides insights into the reliability during the operational phase and the commencement of the wear-out phase. By employing this methodology, designers can make informed decisions regarding sensor performance with regards to location accuracy, contributing to enhanced accuracy in satellite-based applications.

Keywords: bathtub curve, geometric performance, Laplace test, location accuracy, reliability analysis, Weibull analysis

Procedia PDF Downloads 65
9083 Releasing Two Insect Predators to Control of Aphids Under Open-field Conditions

Authors: Mohamed Ahmed Gesraha, Amany Ramadan Ebeid

Abstract:

Aphids are noxious and serious persistent pests in the open fields worldwide. Many authors studied the possibility of aphid control by applying Ladybirds and Lacewings at different releasing rates under open-field conditions. Results clarify that releasing 3rd instar larvae of Coccinella undecimpunctata at the rate of 1 larva:50 aphid was more effective than 1:100 or 1:200 rates for controlling Aphis gossypii population in Okra field; reflecting more than 90% reduction in the aphid population within 15 days. When Chrysoperla carnea 2nd larval instar were releasing at 1:5, 1:10, and 1:20 (predator: aphid), it was noticed that the former rate was the most effective one, inducing 98.93% reduction in aphid population; while the two other rates reflecting less reduction. Additionally, in the case of double releases, the reduction percentage at the 1:5 rate was 99.63%, emphasize that this rate was the most effective one; the other rates induced 97.05 and 95.64% reduction. Generally, a double release was more effective in all tested rates than the single one because of the cumulative existence of the predators in large numbers at the same period of the experiment. It could be concluded that utilizing insect predators (Coccinella undecimpunctata or Chrysoperla carnea) at an early larval stag were faire enough to reduce the aphids’ populations under open fields conditions.

Keywords: releasing predators, lacewings, ladybird, open fields

Procedia PDF Downloads 173
9082 Increasing a Computer Performance by Overclocking Central Processing Unit (CPU)

Authors: Witthaya Mekhum, Wutthikorn Malikong

Abstract:

The objective of this study is to investigate the increasing desktop computer performance after overclocking central processing unit or CPU by running a computer component at a higher clock rate (more clock cycles per second) than it was designed at the rate of 0.1 GHz for each level or 100 MHz starting at 4000 GHz-4500 GHz. The computer performance is tested for each level with 4 programs, i.e. Hyper PI ver. 0.99b, Cinebench R15, LinX ver.0.6.4 and WinRAR . After the CPU overclock, the computer performance increased. When overclocking CPU at 29% the computer performance tested by Hyper PI ver. 0.99b increased by 10.03% and when tested by Cinebench R15 the performance increased by 20.05% and when tested by LinX Program the performance increased by 16.61%. However, the performance increased only 8.14% when tested with Winrar program. The computer performance did not increase according to the overclock rate because the computer consists of many components such as Random Access Memory or RAM, Hard disk Drive, Motherboard and Display Card, etc.

Keywords: overclock, performance, central processing unit, computer

Procedia PDF Downloads 283
9081 Acclimatation of Bacterial Communities for Biohydrogen Production by Co-Digestion Process in Batch and Continuous Systems

Authors: Gómez Romero Jacob, García Peña Elvia Inés

Abstract:

The co-digestion process of crude cheese whey (CCW) with fruit vegetable waste (FVW) for biohydrogen production was investigated in batch and continuous systems, in stirred 1.8 L bioreactors at 37°C. Five different C/N ratios (7, 17, 21, 31, and 46) were tested in batch systems. While, in continuous system eight conditions were evaluated, hydraulic retention time (from 60 to 10 h) and organic load rate (from 21.96 to 155.87 g COD/L d). Data in batch tests showed a maximum specific biohydrogen production rate of 10.68 mmol H2/Lh and a biohydrogen yield of 449.84 mL H2/g COD at a C/N ratio of 21. In continuous co-digestion system, the optimum hydraulic retention time and organic loading rate were 17.5 h and 80.02 g COD/L d, respectively. Under these conditions, the highest volumetric production hydrogen rate (VPHR) and hydrogen yield were 11.02 mmol H2/L h, 800 mL H2/COD, respectively. A pyrosequencing analysis showed that the main acclimated microbial communities for co-digestion studies consisted of Bifidobacterium, with 85.4% of predominance. Hydrogen producing bacteria such as Klebsiella (9.1%), Lactobacillus (0.97%), Citrobacter (0.21%), Enterobacter (0.27%), and Clostridium (0.18%) were less abundant at this culture period. The microbial population structure was correlated with the lactate, acetate, and butyrate profiles obtained. Results demonstrated that the co-digestion of CCW with FVW improves biohydrogen production due to a better nutrient balance and improvement of the system’s buffering capacity.

Keywords: acclimatation, biohydrogen, co-digestion, microbial community

Procedia PDF Downloads 556
9080 Air Quality Forecast Based on Principal Component Analysis-Genetic Algorithm and Back Propagation Model

Authors: Bin Mu, Site Li, Shijin Yuan

Abstract:

Under the circumstance of environment deterioration, people are increasingly concerned about the quality of the environment, especially air quality. As a result, it is of great value to give accurate and timely forecast of AQI (air quality index). In order to simplify influencing factors of air quality in a city, and forecast the city’s AQI tomorrow, this study used MATLAB software and adopted the method of constructing a mathematic model of PCA-GABP to provide a solution. To be specific, this study firstly made principal component analysis (PCA) of influencing factors of AQI tomorrow including aspects of weather, industry waste gas and IAQI data today. Then, we used the back propagation neural network model (BP), which is optimized by genetic algorithm (GA), to give forecast of AQI tomorrow. In order to verify validity and accuracy of PCA-GABP model’s forecast capability. The study uses two statistical indices to evaluate AQI forecast results (normalized mean square error and fractional bias). Eventually, this study reduces mean square error by optimizing individual gene structure in genetic algorithm and adjusting the parameters of back propagation model. To conclude, the performance of the model to forecast AQI is comparatively convincing and the model is expected to take positive effect in AQI forecast in the future.

Keywords: AQI forecast, principal component analysis, genetic algorithm, back propagation neural network model

Procedia PDF Downloads 228
9079 Avoiding Packet Drop for Improved through Put in the Multi-Hop Wireless N/W

Authors: Manish Kumar Rajak, Sanjay Gupta

Abstract:

Mobile ad hoc networks (MANETs) are infrastructure less and intercommunicate using single-hop and multi-hop paths. Network based congestion avoidance which involves managing the queues in the network devices is an integral part of any network. QoS: A set of service requirements that are met by the network while transferring a packet stream from a source to a destination. Especially in MANETs, packet loss results in increased overheads. This paper presents a new algorithm to avoid congestion using one or more queue on nodes and corresponding flow rate decided in advance for each node. When any node attains an initial value of queue then it sends this status to its downstream nodes which in turn uses the pre-decided flow rate of packet transfer to its upstream nodes. The flow rate on each node is adjusted according to the status received from its upstream nodes. This proposed algorithm uses the existing infrastructure to inform to other nodes about its current queue status.

Keywords: mesh networks, MANET, packet count, threshold, throughput

Procedia PDF Downloads 474
9078 Estimation of Constant Coefficients of Bourgoyne and Young Drilling Rate Model for Drill Bit Wear Prediction

Authors: Ahmed Z. Mazen, Nejat Rahmanian, Iqbal Mujtaba, Ali Hassanpour

Abstract:

In oil and gas well drilling, the drill bit is an important part of the Bottom Hole Assembly (BHA), which is installed and designed to drill and produce a hole by several mechanisms. The efficiency of the bit depends on many drilling parameters such as weight on bit, rotary speed, and mud properties. When the bit is pulled out of the hole, the evaluation of the bit damage must be recorded very carefully to guide engineers in order to select the bits for further planned wells. Having a worn bit for hole drilling may cause severe damage to bit leading to cutter or cone losses in the bottom of hole, where a fishing job will have to take place, and all of these will increase the operating cost. The main factor to reduce the cost of drilling operation is to maximize the rate of penetration by analyzing real-time data to predict the drill bit wear while drilling. There are numerous models in the literature for prediction of the rate of penetration based on drilling parameters, mostly based on empirical approaches. One of the most commonly used approaches is Bourgoyne and Young model, where the rate of penetration can be estimated by the drilling parameters as well as a wear index using an empirical correlation, provided all the constants and coefficients are accurately determined. This paper introduces a new methodology to estimate the eight coefficients for Bourgoyne and Young model using the gPROMS parameters estimation GPE (Version 4.2.0). Real data collected form similar formations (12 ¼’ sections) in two different fields in Libya are used to estimate the coefficients. The estimated coefficients are then used in the equations and applied to nearby wells in the same field to predict the bit wear.

Keywords: Bourgoyne and Young model, bit wear, gPROMS, rate of penetration

Procedia PDF Downloads 154
9077 Landslide Susceptibility Mapping Using Soft Computing in Amhara Saint

Authors: Semachew M. Kassa, Africa M Geremew, Tezera F. Azmatch, Nandyala Darga Kumar

Abstract:

Frequency ratio (FR) and analytical hierarchy process (AHP) methods are developed based on past landslide failure points to identify the landslide susceptibility mapping because landslides can seriously harm both the environment and society. However, it is still difficult to select the most efficient method and correctly identify the main driving factors for particular regions. In this study, we used fourteen landslide conditioning factors (LCFs) and five soft computing algorithms, including Random Forest (RF), Support Vector Machine (SVM), Logistic Regression (LR), Artificial Neural Network (ANN), and Naïve Bayes (NB), to predict the landslide susceptibility at 12.5 m spatial scale. The performance of the RF (F1-score: 0.88, AUC: 0.94), ANN (F1-score: 0.85, AUC: 0.92), and SVM (F1-score: 0.82, AUC: 0.86) methods was significantly better than the LR (F1-score: 0.75, AUC: 0.76) and NB (F1-score: 0.73, AUC: 0.75) method, according to the classification results based on inventory landslide points. The findings also showed that around 35% of the study region was made up of places with high and very high landslide risk (susceptibility greater than 0.5). The very high-risk locations were primarily found in the western and southeastern regions, and all five models showed good agreement and similar geographic distribution patterns in landslide susceptibility. The towns with the highest landslide risk include Amhara Saint Town's western part, the Northern part, and St. Gebreal Church villages, with mean susceptibility values greater than 0.5. However, rainfall, distance to road, and slope were typically among the top leading factors for most villages. The primary contributing factors to landslide vulnerability were slightly varied for the five models. Decision-makers and policy planners can use the information from our study to make informed decisions and establish policies. It also suggests that various places should take different safeguards to reduce or prevent serious damage from landslide events.

Keywords: artificial neural network, logistic regression, landslide susceptibility, naïve Bayes, random forest, support vector machine

Procedia PDF Downloads 82
9076 Experimental Study on Capturing of Magnetic Nanoparticles Transported in an Implant Assisted Cylindrical Tube under Magnetic Field

Authors: Anurag Gaur Nidhi

Abstract:

Targeted drug delivery is a method of delivering medication to a patient in a manner that increases the concentration of the medication in some parts of the body relative to others. Targeted drug delivery seeks to concentrate the medication in the tissues of interest while reducing the relative concentration of the medication in the remaining tissues. This improves efficacy of the while reducing side effects. In the present work, we investigate the effect of magnetic field, flow rate and particle concentration on the capturing of magnetic particles transported in a stent implanted fluidic channel. Iron oxide magnetic nanoparticles (Fe3O4) nanoparticles were synthesized via co-precipitation method. The synthesized Fe3O4 nanoparticles were added in the de-ionized (DI) water to prepare the Fe3O4 magnetic particle suspended fluid. This fluid is transported in a cylindrical tube of diameter 8 mm with help of a peristaltic pump at different flow rate (25-40 ml/min). A ferromagnetic coil of SS 430 has been implanted inside the cylindrical tube to enhance the capturing of magnetic nanoparticles under magnetic field. The capturing of magnetic nanoparticles was observed at different magnetic magnetic field, flow rate and particle concentration. It is observed that capture efficiency increases from 47-67 % at magnetic field 2-5kG, respectively at particle concentration 0.6 mg/ml and at flow rate 30 ml/min. However, the capture efficiency decreases from 65 to 44 % by increasing the flow rate from 25 to 40 ml/min, respectively. Furthermore, it is observed that capture efficiency increases from 51 to 67 % by increasing the particle concentration from 0.3 to 0.6 mg/ml, respectively.

Keywords: capture efficiency, implant assisted-Magnetic drug targeting (IA-MDT), magnetic nanoparticles, In-vitro study

Procedia PDF Downloads 307
9075 Electron-Ion Recombination for Photoionized and Collisionally Ionized Plasmas

Authors: Shahin A. Abdel-Naby, Asad T. Hassan

Abstract:

Astrophysical plasma environments can be classified into collisionally ionized (CP) and photoionizedplasmas (PP). In the PP, ionization is caused by an external radiation field, while it is caused by electron collision in the CP. Accurate and reliable laboratory astrophysical data for electron-ion recombination is needed for plasma modeling for low and high-temperatures. Dielectronic recombination (DR) is the dominant recombination process for the CP for most of the ions. When a free electron is captured by an ion with simultaneous excitation of its core, a doubly-exited intermediate state may be formed. The doubly excited state relaxes either by electron emission (autoionization) or by radiative decay (photon emission). DR process takes place when the relaxation occurs to a bound state by a photon emission. DR calculations at low-temperatures are problematic and challenging since small uncertaintiesin the low-energy DR resonance positions can produce huge uncertainties in DR rate coefficients.DR rate coefficients for N²⁺ and O³⁺ ions are calculated using state-of-the-art multi-configurationBreit-Pauli atomic structure AUTOSTRUCTURE collisional package within the generalized collisional-radiative framework. Level-resolved calculations for RR and DR rate coefficients from the ground and metastable initial states are produced in an intermediate coupling scheme associated withn = 0 and n = 1 core-excitations. DR cross sections for these ions are convoluted with the experimental electron-cooler temperatures to produce DR rate coefficients. Good agreements are foundbetween these rate coefficients and theexperimental measurements performed at CRYRING heavy-ionstorage ring for both ions.

Keywords: atomic data, atomic process, electron-ion collision, plasmas

Procedia PDF Downloads 95
9074 Experimental Study of Impregnated Diamond Bit Wear During Sharpening

Authors: Rui Huang, Thomas Richard, Masood Mostofi

Abstract:

The lifetime of impregnated diamond bits and their drilling efficiency are in part governed by the bit wear conditions, not only the extent of the diamonds’ wear but also their exposure or protrusion out of the matrix bonding. As much as individual diamonds wear, the bonding matrix does also wear through two-body abrasion (direct matrix-rock contact) and three-body erosion (cuttings trapped in the space between rock and matrix). Although there is some work dedicated to the study of diamond bit wear, there is still a lack of understanding on how matrix erosion and diamond exposure relate to the bit drilling response and drilling efficiency, as well as no literature on the process that governs bit sharpening a procedure commonly implemented by drillers when the extent of diamond polishing yield extremely low rate of penetration. The aim of this research is (i) to derive a correlation between the wear state of the bit and the drilling performance but also (ii) to gain a better understanding of the process associated with tool sharpening. The research effort combines specific drilling experiments and precise mapping of the tool-cutting face (impregnated diamond bits and segments). Bit wear is produced by drilling through a rock sample at a fixed rate of penetration for a given period of time. Before and after each wear test, the bit drilling response and thus efficiency is mapped out using a tailored design experimental protocol. After each drilling test, the bit or segment cutting face is scanned with an optical microscope. The test results show that, under the fixed rate of penetration, diamond exposure increases with drilling distance but at a decreasing rate, up to a threshold exposure that corresponds to the optimum drilling condition for this feed rate. The data further shows that the threshold exposure scale with the rate of penetration up to a point where exposure reaches a maximum beyond which no more matrix can be eroded under normal drilling conditions. The second phase of this research focuses on the wear process referred as bit sharpening. Drillers rely on different approaches (increase feed rate or decrease flow rate) with the aim of tearing worn diamonds away from the bit matrix, wearing out some of the matrix, and thus exposing fresh sharp diamonds and recovering a higher rate of penetration. Although a common procedure, there is no rigorous methodology to sharpen the bit and avoid excessive wear or bit damage. This paper aims to gain some insight into the mechanisms that accompany bit sharpening by carefully tracking diamond fracturing, matrix wear, and erosion and how they relate to drilling parameters recorded while sharpening the tool. The results show that there exist optimal conditions (operating parameters and duration of the procedure) for sharpening that minimize overall bit wear and that the extent of bit sharpening can be monitored in real-time.

Keywords: bit sharpening, diamond exposure, drilling response, impregnated diamond bit, matrix erosion, wear rate

Procedia PDF Downloads 99
9073 A Transfer Function Representation of Thermo-Acoustic Dynamics for Combustors

Authors: Myunggon Yoon, Jung-Ho Moon

Abstract:

In this paper, we present a transfer function representation of a general one-dimensional combustor. The input of the transfer function is a heat rate perturbation of a burner and the output is a flow velocity perturbation at the burner. This paper considers a general combustor model composed of multiple cans with different cross sectional areas, along with a non-zero flow rate.

Keywords: combustor, dynamics, thermoacoustics, transfer function

Procedia PDF Downloads 381
9072 Dynamic Fault Diagnosis for Semi-Batch Reactor Under Closed-Loop Control via Independent RBFNN

Authors: Abdelkarim M. Ertiame, D. W. Yu, D. L. Yu, J. B. Gomm

Abstract:

In this paper, a new robust fault detection and isolation (FDI) scheme is developed to monitor a multivariable nonlinear chemical process called the Chylla-Haase polymerization reactor when it is under the cascade PI control. The scheme employs a radial basis function neural network (RBFNN) in an independent mode to model the process dynamics and using the weighted sum-squared prediction error as the residual. The recursive orthogonal Least Squares algorithm (ROLS) is employed to train the model to overcome the training difficulty of the independent mode of the network. Then, another RBFNN is used as a fault classifier to isolate faults from different features involved in the residual vector. The several actuator and sensor faults are simulated in a nonlinear simulation of the reactor in Simulink. The scheme is used to detect and isolate the faults on-line. The simulation results show the effectiveness of the scheme even the process is subjected to disturbances and uncertainties including significant changes in the monomer feed rate, fouling factor, impurity factor, ambient temperature and measurement noise. The simulation results are presented to illustrate the effectiveness and robustness of the proposed method.

Keywords: Robust fault detection, cascade control, independent RBF model, RBF neural networks, Chylla-Haase reactor, FDI under closed-loop control

Procedia PDF Downloads 498
9071 Experimental Investigation of Fluid Dynamic Effects on Crystallisation Scale Growth and Suppression in Agitation Tank

Authors: Prasanjit Das, M. M. K. Khan, M. G. Rasul, Jie Wu, I. Youn

Abstract:

Mineral scale formation is undoubtedly a more serious problem in the mineral industry than other process industries. To better understand scale growth and suppression, an experimental model is proposed in this study for supersaturated crystallised solutions commonly found in mineral process plants. In this experiment, surface crystallisation of potassium nitrate (KNO3) on the wall of the agitation tank and agitation effects on the scale growth and suppression are studied. The new quantitative scale suppression model predicts that at lower agitation speed, the scale growth rate is enhanced and at higher agitation speed, the scale suppression rate increases due to the increased flow erosion effect. A lab-scale agitation tank with and without baffles were used as a benchmark in this study. The fluid dynamic effects on scale growth and suppression in the agitation tank with three different size impellers (diameter 86, 114, 160 mm and model A310 with flow number 0.56) at various ranges of rotational speed (up to 700 rpm) and solution with different concentration (4.5, 4.75 and 5.25 mol/dm3) were investigated. For more elucidation, the effects of the different size of the impeller on wall surface scale growth and suppression rate as well as bottom settled scale accumulation rate are also discussed. Emphasis was placed on applications in the mineral industry, although results are also relevant to other industrial applications.

Keywords: agitation tank, crystallisation, impeller speed, scale

Procedia PDF Downloads 223
9070 Larger Diameter 22 MM-PDC Cutter Greatly Improves Drilling Efficiency of PDC Bit

Authors: Fangyuan Shao, Wei Liu, Deli Gao

Abstract:

With the increasing speed of oil and gas exploration, development and production at home and abroad, the demand for drilling speed up technology is becoming more and more critical to reduce the development cost. Highly efficient and personalized PDC bit is important equipment in the bottom hole assembly (BHA). Therefore, improving the rock-breaking efficiency of PDC bits will help reduce drilling time and drilling cost. Advances in PDC bit technology have resulted in a leapfrogging improvement in the rate of penetration (ROP) of PDC bits over roller cone bits in soft to medium-hard formations. Recently, with the development of PDC technology, the diameter of the PDC tooth can be further expanded. The maximum diameter of the PDC cutter used in this paper is 22 mm. According to the theoretical calculation, under the same depth of cut (DOC), the 22mm-PDC cutter increases the exposure of the cutter, and the increase of PDC cutter diameter helps to increase the cutting area of the PDC cutter. In order to evaluate the cutting performance of the 22 mm-PDC cutter and the existing commonly used cutters, the 16 mm, 19 mm and 22 mm PDC cutter was selected put on a vertical turret lathe (VTL) in the laboratory for cutting tests under different DOCs. The DOCs were 0.5mm, 1.0 mm, 1.5 mm and 2.0 mm, 2.5 mm and 3 mm, respectively. The rock sample used in the experiment was limestone. Results of laboratory tests have shown the new 22 mm-PDC cutter technology greatly improved cutting efficiency. On the one hand, as the DOC increases, the mechanical specific energy (MSE) of all cutters decreases, which means that the cutting efficiency increases. On the other hand, under the same DOC condition, the larger the cutter diameter is, the larger the working area of the cutter is, which leads to higher the cutting efficiency. In view of the high performance of the 22 mm-PDC cutters, which was applied to carry out full-scale bit field experiments. The result shows that the bit with 22mm-PDC cutters achieves a breakthrough improvement of ROP than that with conventional 16mm and 19mm cutters in offset well drilling.

Keywords: polycrystalline diamond compact, 22 mm-PDC cutters, cutting efficiency, mechanical specific energy

Procedia PDF Downloads 204