Search results for: noise measurement
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3654

Search results for: noise measurement

504 Assessment of Five Photoplethysmographic Methods for Estimating Heart Rate Variability

Authors: Akshay B. Pawar, Rohit Y. Parasnis

Abstract:

Heart Rate Variability (HRV) is a widely used indicator of the regulation between the autonomic nervous system (ANS) and the cardiovascular system. Besides being non-invasive, it also has the potential to predict mortality in cases involving critical injuries. The gold standard method for determining HRV is based on the analysis of RR interval time series extracted from ECG signals. However, because it is much more convenient to obtain photoplethysmogramic (PPG) signals as compared to ECG signals (which require the attachment of several electrodes to the body), many researchers have used pulse cycle intervals instead of RR intervals to estimate HRV. They have also compared this method with the gold standard technique. Though most of their observations indicate a strong correlation between the two methods, recent studies show that in healthy subjects, except for a few parameters, the pulse-based method cannot be a surrogate for the standard RR interval- based method. Moreover, the former tends to overestimate short-term variability in heart rate. This calls for improvements in or alternatives to the pulse-cycle interval method. In this study, besides the systolic peak-peak interval method (PP method) that has been studied several times, four recent PPG-based techniques, namely the first derivative peak-peak interval method (P1D method), the second derivative peak-peak interval method (P2D method), the valley-valley interval method (VV method) and the tangent-intersection interval method (TI method) were compared with the gold standard technique. ECG and PPG signals were obtained from 10 young and healthy adults (consisting of both males and females) seated in the armchair position. In order to de-noise these signals and eliminate baseline drift, they were passed through certain digital filters. After filtering, the following HRV parameters were computed from PPG using each of the five methods and also from ECG using the gold standard method: time domain parameters (SDNN, pNN50 and RMSSD), frequency domain parameters (Very low-frequency power (VLF), Low-frequency power (LF), High-frequency power (HF) and Total power or “TP”). Besides, Poincaré plots were also plotted and their SD1/SD2 ratios determined. The resulting sets of parameters were compared with those yielded by the standard method using measures of statistical correlation (correlation coefficient) as well as statistical agreement (Bland-Altman plots). From the viewpoint of correlation, our results show that the best PPG-based methods for the determination of most parameters and Poincaré plots are the P2D method (shows more than 93% correlation with the standard method) and the PP method (mean correlation: 88%) whereas the TI, VV and P1D methods perform poorly (<70% correlation in most cases). However, our evaluation of statistical agreement using Bland-Altman plots shows that none of the five techniques agrees satisfactorily well with the gold standard method as far as time-domain parameters are concerned. In conclusion, excellent statistical correlation implies that certain PPG-based methods provide a good amount of information on the pattern of heart rate variation, whereas poor statistical agreement implies that PPG cannot completely replace ECG in the determination of HRV.

Keywords: photoplethysmography, heart rate variability, correlation coefficient, Bland-Altman plot

Procedia PDF Downloads 310
503 The Effects of Cooling during Baseball Games on Perceived Exertion and Core Temperature

Authors: Chih-Yang Liao

Abstract:

Baseball is usually played outdoors in the warmest months of the year. Therefore, baseball players are susceptible to the influence of the hot environment. It has been shown that hitting performance is increased in games played in warm weather, compared to in cold weather, in Major League Baseball. Intermittent cooling during sporting events can prevent the risk of hyperthermia and increase endurance performance. However, the effects of cooling during baseball games played in a hot environment are unclear. This study adopted a cross-over design. Ten Division I collegiate male baseball players in Taiwan volunteered to participate in this study. Each player played two simulated baseball games, with one day in between. Five of the players received intermittent cooling during the first simulated game, while the other five players received intermittent cooling during the second simulated game. The participants were covered in neck and forehand regions for 6 min with towels that were soaked in icy salt water 3 to 4 times during the games. The participants received the cooling treatment in the dugout when they were not on the field for defense or hitting. During the 2 simulated games, the temperature was 31.1-34.1°C and humidity was 58.2-61.8%, with no difference between the two games. Ratings of perceived exertion, thermal sensation, tympanic and forehead skin temperature immediately after each defensive half-inning and after cooling treatments were recorded. Ratings of perceived exertion were measured using the Borg 10-point scale. The thermal sensation was measured with a 6-point scale. The tympanic and skin temperature was measured with infrared thermometers. The data were analyzed with a two-way analysis of variance with repeated measurement. The results showed that intermitted cooling significantly reduced ratings of perceived exertion and thermal sensation. Forehead skin temperature was also significantly decreased after cooling treatments. However, the tympanic temperature was not significantly different between the two trials. In conclusion, intermittent cooling in the neck and forehead regions was effective in alleviating the perceived exertion and heat sensation. However, this cooling intervention did not affect the core temperature. Whether intermittent cooling has any impact on hitting or pitching performance in baseball players warrants further investigation.

Keywords: baseball, cooling, ratings of perceived exertion, thermal sensation

Procedia PDF Downloads 133
502 The Importance of the Fluctuation in Blood Sugar and Blood Pressure of Insulin-Dependent Diabetic Patients with Chronic Kidney Disease

Authors: Hitoshi Minakuchi, Izumi Takei, Shu Wakino, Koichi Hayashi, Hiroshi Itoh

Abstract:

Objectives: Among type 2 diabetics, patients with CKD(chronic kidney disease), insulin resistance, impaired glyconeogenesis in kidney and reduced degradation of insulin are recognized, and we observed different fluctuational patterns of blood sugar between CKD patients and non-CKD patients. On the other hand, non-dipper type blood pressure change is the risk of organ damage and mortality. We performed cross-sectional study to elucidate the characteristic of the fluctuation of blood glucose and blood pressure at insulin-treated diabetic patients with chronic kidney disease. Methods: From March 2011 to April 2013, at the Ichikawa General Hospital of Tokyo Dental College, we recruited 20 outpatients. All participants are insulin-treated type 2 diabetes with CKD. We collected serum samples, urine samples for several hormone measurements, and performed CGMS(Continuous glucose measurement system), ABPM (ambulatory blood pressure monitoring), brain computed tomography, carotid artery thickness, ankle brachial index, PWV, CVR-R, and analyzed these data statistically. Results: Among all 20 participants, hypoglycemia was decided blood glucose 70mg/dl by CGMS of 9 participants (45.0%). The event of hypoglycemia was recognized lower eGFR (29.8±6.2ml/min:41.3±8.5ml/min, P<0.05), lower HbA1c (6.44±0.57%:7.53±0.49%), higher PWV (1858±97.3cm/s:1665±109.2cm/s), higher serum glucagon (194.2±34.8pg/ml:117.0±37.1pg/ml), higher free cortisol of urine (53.8±12.8μg/day:34.8±7.1μg/day), and higher metanephrin of urine (0.162±0.031mg/day:0.076±0.029mg/day). Non-dipper type blood pressure change in ABPM was detected 8 among 9 participants with hypoglycemia (88.9%), 4 among 11 participants (36.4%) without hypoglycemia. Multiplex logistic-regression analysis revealed that the event of hypoglycemia is the independent factor of non-dipper type blood pressure change. Conclusions: Among insulin-treated type 2 diabetic patients with CKD, the events of hypoglycemia were frequently detected, and can associate with the organ derangements through the medium of non-dipper type blood pressure change.

Keywords: chronic kidney disease, hypoglycemia, non-dipper type blood pressure change, diabetic patients

Procedia PDF Downloads 397
501 Temperature Contour Detection of Salt Ice Using Color Thermal Image Segmentation Method

Authors: Azam Fazelpour, Saeed Reza Dehghani, Vlastimil Masek, Yuri S. Muzychka

Abstract:

The study uses a novel image analysis based on thermal imaging to detect temperature contours created on salt ice surface during transient phenomena. Thermal cameras detect objects by using their emissivities and IR radiance. The ice surface temperature is not uniform during transient processes. The temperature starts to increase from the boundary of ice towards the center of that. Thermal cameras are able to report temperature changes on the ice surface at every individual moment. Various contours, which show different temperature areas, appear on the ice surface picture captured by a thermal camera. Identifying the exact boundary of these contours is valuable to facilitate ice surface temperature analysis. Image processing techniques are used to extract each contour area precisely. In this study, several pictures are recorded while the temperature is increasing throughout the ice surface. Some pictures are selected to be processed by a specific time interval. An image segmentation method is applied to images to determine the contour areas. Color thermal images are used to exploit the main information. Red, green and blue elements of color images are investigated to find the best contour boundaries. The algorithms of image enhancement and noise removal are applied to images to obtain a high contrast and clear image. A novel edge detection algorithm based on differences in the color of the pixels is established to determine contour boundaries. In this method, the edges of the contours are obtained according to properties of red, blue and green image elements. The color image elements are assessed considering their information. Useful elements proceed to process and useless elements are removed from the process to reduce the consuming time. Neighbor pixels with close intensities are assigned in one contour and differences in intensities determine boundaries. The results are then verified by conducting experimental tests. An experimental setup is performed using ice samples and a thermal camera. To observe the created ice contour by the thermal camera, the samples, which are initially at -20° C, are contacted with a warmer surface. Pictures are captured for 20 seconds. The method is applied to five images ,which are captured at the time intervals of 5 seconds. The study shows the green image element carries no useful information; therefore, the boundary detection method is applied on red and blue image elements. In this case study, the results indicate that proposed algorithm shows the boundaries more effective than other edges detection methods such as Sobel and Canny. Comparison between the contour detection in this method and temperature analysis, which states real boundaries, shows a good agreement. This color image edge detection method is applicable to other similar cases according to their image properties.

Keywords: color image processing, edge detection, ice contour boundary, salt ice, thermal image

Procedia PDF Downloads 300
500 Geological Structure Identification in Semilir Formation: An Correlated Geological and Geophysical (Very Low Frequency) Data for Zonation Disaster with Current Density Parameters and Geological Surface Information

Authors: E. M. Rifqi Wilda Pradana, Bagus Bayu Prabowo, Meida Riski Pujiyati, Efraim Maykhel Hagana Ginting, Virgiawan Arya Hangga Reksa

Abstract:

The VLF (Very Low Frequency) method is an electromagnetic method that uses low frequencies between 10-30 KHz which results in a fairly deep penetration. In this study, the VLF method was used for zonation of disaster-prone areas by identifying geological structures in the form of faults. Data acquisition was carried out in Trimulyo Region, Jetis District, Bantul Regency, Special Region of Yogyakarta, Indonesia with 8 measurement paths. This study uses wave transmitters from Japan and Australia to obtain Tilt and Elipt values that can be used to create RAE (Rapat Arus Ekuivalen or Current Density) sections that can be used to identify areas that are easily crossed by electric current. This section will indicate the existence of a geological structure in the form of faults in the study area which is characterized by a high RAE value. In data processing of VLF method, it is obtained Tilt vs Elliptical graph and Moving Average (MA) Tilt vs Moving Average (MA) Elipt graph of each path that shows a fluctuating pattern and does not show any intersection at all. Data processing uses Matlab software and obtained areas with low RAE values that are 0%-6% which shows medium with low conductivity and high resistivity and can be interpreted as sandstone, claystone, and tuff lithology which is part of the Semilir Formation. Whereas a high RAE value of 10% -16% which shows a medium with high conductivity and low resistivity can be interpreted as a fault zone filled with fluid. The existence of the fault zone is strengthened by the discovery of a normal fault on the surface with strike N550W and dip 630E at coordinates X= 433256 and Y= 9127722 so that the activities of residents in the zone such as housing, mining activities and other activities can be avoided to reduce the risk of natural disasters.

Keywords: current density, faults, very low frequency, zonation

Procedia PDF Downloads 156
499 Magnetophotonics 3D MEMS/NEMS System for Quantitative Mitochondrial DNA Defect Profiling

Authors: Dar-Bin Shieh, Gwo-Bin Lee, Chen-Ming Chang, Chen Sheng Yeh, Chih-Chia Huang, Tsung-Ju Li

Abstract:

Mitochondrial defects have a significant impact in many human diseases and aging associated phenotypes. The pathogenic mitochondrial DNA (mtDNA) mutations are diverse and usually present as heteroplasmic. mtDNA 4977bps deletion is one of the common mtDNA defects, and the ratio of mutated versus normal copy is significantly associated with clinical symptoms thus their quantitative detection has become an important unmet needs for advanced disease diagnosis and therapeutic guidelines. This study revealed a Micro-electro-mechanical-system (MEMS) enabled automatic microfluidic chip that only required minimal sample. The system integrated multiple laboratory operation steps into a Lab-on-a-Chip for high-sensitive and prompt measurement. The entire process including magnetic nanoparticle based mtDNA extraction in chip, mutation selective photonic DNA cleavage, and nanoparticle accelerated photonic quantitative polymerase chain reaction (qPCR). All subsystems were packed inside a miniature three-dimensional micro structured system and operated in an automatic manner. Integration of magnetic beads with microfluidic transportation could promptly extract and enrich the specific mtDNA. The near infrared responsive magnetic nanoparticles enabled micro-PCR to be operated by pulse-width-modulation controlled laser pulsing to amplify the desired mtDNA while quantified by fluorescence intensity captured by a complementary metal oxide system array detector. The proportions of pathogenic mtDNA in total DNA were thus obtained. Micro capillary electrophoresis module was used to analyze the amplicone products. In conclusion, this study demonstrated a new magnetophotonic based qPCR MEMS system that successfully detects and quantify specific disease related DNA mutations thus provides a promising future for rapid diagnosis of mitochondria diseases.

Keywords: mitochondrial DNA, micro-electro-mechanical-system, magnetophotonics, PCR

Procedia PDF Downloads 203
498 Improvement of Cardiometabolic after 8 Weeks of Weight Loss Intervention

Authors: Boris Bajer, Andrea Havranova, Miroslav Vlcek, Richard Imrich, Adela Penesova

Abstract:

Lifestyle interventions can prevent the deterioration of impaired glucose tolerance to manifest type 2 diabetes, and also prevent cardiovascular diseases, as it showed many studies (the Finnish Diabetes Prevention Study, Diabetes Prevention Program (DPP), . the China Da Qing Diabetes Prevention Study, etc.) Therefore the aim of our study was to compare the effect of intensified lifestyle intervention on cardiometabolic parameters. Methods: It is an ongoing randomized interventional clinical study (NCT02325804) focused on the reduction of body weight/fat. Intervention: hypocaloric diet (30% restriction of calories) and physical activity 150 minutes/week. Before and after 8 weeks of intervention all patients underwent complete medical examination (measurement of physical fitness, resting metabolic rate (RMR), body composition analysis, oral glucose tolerance test, parameters of lipid metabolism, and other cardiometabolic risk factors. Results: So far 39 patients finished the intervention. The average reduction of body weight was 6,8 + 4,9 kg (0-15 kg; p=0,0006), accompanied with significant reduction of body fat percentage (p ≤ 0,0001), amount of fat mass (p=0,03), waist circumference (p=0.02). Amount of lean mass and RMR remained unchanged. Heart rate (p=0,02), systolic and diastolic blood pressure was reduced (p=0,01 p=0,02 resp.) as well as insulin sensitivity was improved. Lipid parameters also changed - cholesterol, LDL decreased (p=0,05, p=0,04 resp.), while triglycerides showed tendency to decrease (p=0,055). Liver function improved, alanine aminotrasnferase (ALT) were reduced (p=0,01). Physical fitness significantly improved (as measure VO2 max (p=0,02). Conclusion: Results of our study are in line with previous results about the beneficial effect of intensive lifestyle changes on the reduction of cardiometabolic risk factors and improvement of liver function. Supported by grants APVV 15-0228; VEGA 2/0161/16

Keywords: obesity, weight loss, diet lipids, blood pressure, liver enzymes

Procedia PDF Downloads 148
497 The Impact of Protein Content on Athletes’ Body Composition

Authors: G. Vici, L. Cesanelli, L. Belli, R. Ceci, V. Polzonetti

Abstract:

Several factors contribute to success in sport and diet is one of them. Evidence-based sport nutrition guidelines underline the importance of macro- and micro-nutrients’ balance and timing in order to improve athlete’s physical status and performance. Nevertheless, a high content of proteins is commonly found in resistance training athletes’ diet with carbohydrate intake that is not enough or not well planned. The aim of the study was to evaluate the impact of different protein and carbohydrate diet contents on body composition and sport performance on a group of resistance training athletes. Subjects were divided as study group (n=16) and control group (n=14). For a period of 4 months, both groups were subjected to the same resistance training fitness program with study group following a specific diet and control group following an ab libitum diet. Body compositions were evaluated trough anthropometric measurement (weight, height, body circumferences and skinfolds) and Bioimpedence Analysis. Physical strength and training status of individuals were evaluated through the One Repetition Maximum test (RM1). Protein intake in studied group was found to be lower than in control group. There was a statistically significant increase of body weight, free fat mass and body mass cell of studied group respect to the control group. Fat mass remains almost constant. Statistically significant changes were observed in quadriceps and biceps circumferences, with an increase in studied group. The MR1 test showed improvement in study group’s strength but no changes in control group. Usually people consume hyper-proteic diet to achieve muscle mass development. Through this study, it was possible to show that protein intake fixed at 1,7 g/kg/d can meet the individual's needs. In parallel, the increased intake of carbohydrates, focusing on quality and timing of assumption, has enabled the obtainment of desired results with a training protocol supporting a hypertrophic strategy. Therefore, the key point seems related to the planning of a structured program both from a nutritional and training point of view.

Keywords: body composition, diet, exercise, protein

Procedia PDF Downloads 210
496 A Risk Assessment Tool for the Contamination of Aflatoxins on Dried Figs Based on Machine Learning Algorithms

Authors: Kottaridi Klimentia, Demopoulos Vasilis, Sidiropoulos Anastasios, Ihara Diego, Nikolaidis Vasileios, Antonopoulos Dimitrios

Abstract:

Aflatoxins are highly poisonous and carcinogenic compounds produced by species of the genus Aspergillus spp. that can infect a variety of agricultural foods, including dried figs. Biological and environmental factors, such as population, pathogenicity, and aflatoxinogenic capacity of the strains, topography, soil, and climate parameters of the fig orchards, are believed to have a strong effect on aflatoxin levels. Existing methods for aflatoxin detection and measurement, such as high performance liquid chromatography (HPLC), and enzyme-linked immunosorbent assay (ELISA), can provide accurate results, but the procedures are usually time-consuming, sample-destructive, and expensive. Predicting aflatoxin levels prior to crop harvest is useful for minimizing the health and financial impact of a contaminated crop. Consequently, there is interest in developing a tool that predicts aflatoxin levels based on topography and soil analysis data of fig orchards. This paper describes the development of a risk assessment tool for the contamination of aflatoxin on dried figs, based on the location and altitude of the fig orchards, the population of the fungus Aspergillus spp. in the soil, and soil parameters such as pH, saturation percentage (SP), electrical conductivity (EC), organic matter, particle size analysis (sand, silt, clay), the concentration of the exchangeable cations (Ca, Mg, K, Na), extractable P, and trace of elements (B, Fe, Mn, Zn and Cu), by employing machine learning methods. In particular, our proposed method integrates three machine learning techniques, i.e., dimensionality reduction on the original dataset (principal component analysis), metric learning (Mahalanobis metric for clustering), and k-nearest neighbors learning algorithm (KNN), into an enhanced model, with mean performance equal to 85% by terms of the Pearson correlation coefficient (PCC) between observed and predicted values.

Keywords: aflatoxins, Aspergillus spp., dried figs, k-nearest neighbors, machine learning, prediction

Procedia PDF Downloads 164
495 Design and Development of an Autonomous Underwater Vehicle for Irrigation Canal Monitoring

Authors: Mamoon Masud, Suleman Mazhar

Abstract:

Indus river basin’s irrigation system in Pakistan is extremely complex, spanning over 50,000 km. Maintenance and monitoring of this demands enormous resources. This paper describes the development of a streamlined and low-cost autonomous underwater vehicle (AUV) for the monitoring of irrigation canals including water quality monitoring and water theft detection. The vehicle is a hovering-type AUV, designed mainly for monitoring irrigation canals, with fully documented design and open source code. It has a length of 17 inches, and a radius of 3.5 inches with a depth rating of 5m. Multiple sensors are present onboard the AUV for monitoring water quality parameters including pH, turbidity, total dissolved solids (TDS) and dissolved oxygen. A 9-DOF Inertial Measurement Unit (IMU), GY-85, is used, which incorporates an Accelerometer (ADXL345), a Gyroscope (ITG-3200) and a Magnetometer (HMC5883L). The readings from these sensors are fused together using directional cosine matrix (DCM) algorithm, providing the AUV with the heading angle, while a pressure sensor gives the depth of the AUV. 2 sonar-based range sensors are used for obstacle detection, enabling the vehicle to align itself with the irrigation canals edges. 4 thrusters control the vehicle’s surge, heading and heave, providing 3 DOF. The thrusters are controlled using a proportional-integral-derivative (PID) feedback control system, with heading angle and depth being the controller’s input and the thruster motor speed as the output. A flow sensor has been incorporated to monitor canal water level to detect water-theft event in the irrigation system. In addition to water theft detection, the vehicle also provides information on water quality, providing us with the ability to identify the source(s) of water contamination. Detection of such events can provide useful policy inputs for improving irrigation efficiency and reducing water contamination. The AUV being low cost, small sized and suitable for autonomous maneuvering, water level and quality monitoring in the irrigation canals, can be used for irrigation network monitoring at a large scale.

Keywords: the autonomous underwater vehicle, irrigation canal monitoring, water quality monitoring, underwater line tracking

Procedia PDF Downloads 128
494 Application of Acoustic Emissions Related to Drought Can Elicit Antioxidant Responses and Capsaicinoids Content in Chili Pepper Plants

Authors: Laura Helena Caicedo Lopez, Luis Miguel Contreras Medina, Ramon Gerardo Guevara Gonzales, Juan E. Andrade

Abstract:

In this study, we evaluated the effect of three different hydric stress conditions: Low (LHS), medium (MHS), and high (HHS) on capsaicinoid content and enzyme regulation of C. annuum plants. Five main peaks were detected using a 2 Hz resolution vibrometer laser (Polytec-B&K). These peaks or “characteristic frequencies” were used as acoustic emissions (AEs) treatment, transforming these signals into audible sound with the frequency (Hz) content of each hydric stress. Capsaicinoids (CAPs) are the main, secondary metabolites of chili pepper plants and are known to increase during hydric stress conditions or short drought-periods. The AEs treatments were applied in two plant stages: the first one was in the pre-anthesis stage to evaluate the genes that encode the transcription of enzymes responsible for diverse metabolic activities of C. annuum plants. For example, the antioxidant responses such as peroxidase (POD), superoxide dismutase (Mn-SOD). Also, phenyl-alanine ammonia-lyase (PAL) involved in the biosynthesis of the phenylpropanoid compounds. The chalcone synthase (CHS) related to the natural defense mechanisms and species-specific aquaporin (CAPIP-1) that regulate the flow of water into and out of cells. The second stage was at 40 days after flowering (DAF) to evaluate the biochemical effect of AEs related to hydric stress on capsaicinoids production. These two experiments were conducted to identify the molecular responses of C. annuum plants to AE. Moreover, to define AEs could elicit any increase in the capsaicinoids content after a one-week exposition to AEs treatments. The results show that all AEs treatment signals (LHS, MHS, and HHS) were significantly different compared to the non-acoustic emission control (NAE). Also, the AEs induced the up-regulation of POD (~2.8, 2.9, and 3.6, respectively). The gene expression of another antioxidant response was particularly treatment-dependent. The HHS induced and overexpression of Mn-SOD (~0.23) and PAL (~0.33). As well, the MHS only induced an up-regulation of the CHs gene (~0.63). On the other hand, CAPIP-1 gene gas down-regulated by all AEs treatments LHS, MHS, and HHS ~ (-2.4, -0.43 and -6.4, respectively). Likewise, the down-regulation showed particularities depending on the treatment. LHS and MHS induced downregulation of the SOD gene ~ (-1.26 and -1.20 respectively) and PAL (-4.36 and 2.05, respectively). Correspondingly, the LHS and HHS showed the same tendency in the CHs gene, respectively ~ (-1.12 and -1.02, respectively). Regarding the elicitation effect of AE on the capsaicinoids content, additional treatment controls were included. A white noise treatment (WN) to prove the frequency-selectiveness of signals and a hydric stressed group (HS) to compare the CAPs content. Our findings suggest that WN and NAE did not present differences statically. Conversely, HS and all AEs treatments induced a significant increase of capsaicin (Cap) and dihydrocapsaicin (Dcap) after one-week of a treatment. Specifically, the HS plants showed an increase of 8.33 times compared to the NAE and WN treatments and 1.4 times higher than the MHS, which was the AEs treatment with a larger induction of Capsaicinoids among treatments (5.88) and compared to the controls.

Keywords: acoustic emission, capsaicinoids, elicitors, hydric stress, plant signaling

Procedia PDF Downloads 156
493 Coupling Strategy for Multi-Scale Simulations in Micro-Channels

Authors: Dahia Chibouti, Benoit Trouette, Eric Chenier

Abstract:

With the development of micro-electro-mechanical systems (MEMS), understanding fluid flow and heat transfer at the micrometer scale is crucial. In the case where the flow characteristic length scale is narrowed to around ten times the mean free path of gas molecules, the classical fluid mechanics and energy equations are still valid in the bulk flow, but particular attention must be paid to the gas/solid interface boundary conditions. Indeed, in the vicinity of the wall, on a thickness of about the mean free path of the molecules, called the Knudsen layer, the gas molecules are no longer in local thermodynamic equilibrium. Therefore, macroscopic models based on the continuity of velocity, temperature and heat flux jump conditions must be applied at the fluid/solid interface to take this non-equilibrium into account. Although these macroscopic models are widely used, the assumptions on which they depend are not necessarily verified in realistic cases. In order to get rid of these assumptions, simulations at the molecular scale are carried out to study how molecule interaction with walls can change the fluid flow and heat transfers at the vicinity of the walls. The developed approach is based on a kind of heterogeneous multi-scale method: micro-domains overlap the continuous domain, and coupling is carried out through exchanges of information between both the molecular and the continuum approaches. In practice, molecular dynamics describes the fluid flow and heat transfers in micro-domains while the Navier-Stokes and energy equations are used at larger scales. In this framework, two kinds of micro-simulation are performed: i) in bulk, to obtain the thermo-physical properties (viscosity, conductivity, ...) as well as the equation of state of the fluid, ii) close to the walls to identify the relationships between the slip velocity and the shear stress or between the temperature jump and the normal temperature gradient. The coupling strategy relies on an implicit formulation of the quantities extracted from micro-domains. Indeed, using the results of the molecular simulations, a Bayesian regression is performed in order to build continuous laws giving both the behavior of the physical properties, the equation of state and the slip relationships, as well as their uncertainties. These latter allow to set up a learning strategy to optimize the number of micro simulations. In the present contribution, the first results regarding this coupling associated with the learning strategy are illustrated through parametric studies of convergence criteria, choice of basis functions and noise of input data. Anisothermic flows of a Lennard Jones fluid in micro-channels are finally presented.

Keywords: multi-scale, microfluidics, micro-channel, hybrid approach, coupling

Procedia PDF Downloads 153
492 The Analysis of Drill Bit Optimization by the Application of New Electric Impulse Technology in Shallow Water Absheron Peninsula

Authors: Ayshan Gurbanova

Abstract:

Despite based on the fact that drill bit which is the smallest part of bottom hole assembly costs only in between 10% and 15% of the total expenses made, they are the first equipment that is in contact with the formation itself. Hence, it is consequential to choose the appropriate type and dimension of drilling bit, which will prevent majority of problems by not demanding many tripping procedure. However, within the advance in technology, it is now seamless to be beneficial in the terms of many concepts such as subsequent time of operation, energy, expenditure, power and so forth. With the intention of applying the method to Azerbaijan, the field of Shallow Water Absheron Peninsula has been suggested, where the mainland has been located 15 km away from the wildcat wells, named as “NKX01”. It has the water depth of 22 m as indicated. In 2015 and 2016, the seismic survey analysis of 2D and 3D have been conducted in contract area as well as onshore shallow water depth locations. With the aim of indicating clear elucidation, soil stability, possible submersible dangerous scenarios, geohazards and bathymetry surveys have been carried out as well. Within the seismic analysis results, the exact location of exploration wells have been determined and along with this, the correct measurement decisions have been made to divide the land into three productive zones. In the term of the method, Electric Impulse Technology (EIT) is based on discharge energies of electricity within the corrosivity in rock. Take it simply, the highest value of voltages could be created in the less range of nano time, where it is sent to the rock through electrodes’ baring as demonstrated below. These electrodes- higher voltage powered and grounded are placed on the formation which could be obscured in liquid. With the design, it is more seamless to drill horizontal well based on the advantage of loose contact of formation. There is also no chance of worn ability as there are no combustion, mechanical power exist. In the case of energy, the usage of conventional drilling accounts for 1000 𝐽/𝑐𝑚3 , where this value accounts for between 100 and 200 𝐽/𝑐𝑚3 in EIT. Last but not the least, from the test analysis, it has been yielded that it achieves the value of ROP more than 2 𝑚/ℎ𝑟 throughout 15 days. Taking everything into consideration, it is such a fact that with the comparison of data analysis, this method is highly applicable to the fields of Azerbaijan.

Keywords: drilling, drill bit cost, efficiency, cost

Procedia PDF Downloads 51
491 Stable Diffusion, Context-to-Motion Model to Augmenting Dexterity of Prosthetic Limbs

Authors: André Augusto Ceballos Melo

Abstract:

Design to facilitate the recognition of congruent prosthetic movements, context-to-motion translations guided by image, verbal prompt, users nonverbal communication such as facial expressions, gestures, paralinguistics, scene context, and object recognition contributes to this process though it can also be applied to other tasks, such as walking, Prosthetic limbs as assistive technology through gestures, sound codes, signs, facial, body expressions, and scene context The context-to-motion model is a machine learning approach that is designed to improve the control and dexterity of prosthetic limbs. It works by using sensory input from the prosthetic limb to learn about the dynamics of the environment and then using this information to generate smooth, stable movements. This can help to improve the performance of the prosthetic limb and make it easier for the user to perform a wide range of tasks. There are several key benefits to using the context-to-motion model for prosthetic limb control. First, it can help to improve the naturalness and smoothness of prosthetic limb movements, which can make them more comfortable and easier to use for the user. Second, it can help to improve the accuracy and precision of prosthetic limb movements, which can be particularly useful for tasks that require fine motor control. Finally, the context-to-motion model can be trained using a variety of different sensory inputs, which makes it adaptable to a wide range of prosthetic limb designs and environments. Stable diffusion is a machine learning method that can be used to improve the control and stability of movements in robotic and prosthetic systems. It works by using sensory feedback to learn about the dynamics of the environment and then using this information to generate smooth, stable movements. One key aspect of stable diffusion is that it is designed to be robust to noise and uncertainty in the sensory feedback. This means that it can continue to produce stable, smooth movements even when the sensory data is noisy or unreliable. To implement stable diffusion in a robotic or prosthetic system, it is typically necessary to first collect a dataset of examples of the desired movements. This dataset can then be used to train a machine learning model to predict the appropriate control inputs for a given set of sensory observations. Once the model has been trained, it can be used to control the robotic or prosthetic system in real-time. The model receives sensory input from the system and uses it to generate control signals that drive the motors or actuators responsible for moving the system. Overall, the use of the context-to-motion model has the potential to significantly improve the dexterity and performance of prosthetic limbs, making them more useful and effective for a wide range of users Hand Gesture Body Language Influence Communication to social interaction, offering a possibility for users to maximize their quality of life, social interaction, and gesture communication.

Keywords: stable diffusion, neural interface, smart prosthetic, augmenting

Procedia PDF Downloads 87
490 A Literature Review of Precision Agriculture: Applications of Diagnostic Diseases in Corn, Potato, and Rice Based on Artificial Intelligence

Authors: Carolina Zambrana, Grover Zurita

Abstract:

The food loss production that occurs in deficient agricultural production is one of the major problems worldwide. This puts the population's food security and the efficiency of farming investments at risk. It is to be expected that this food security will be achieved with the own and efficient production of each country. It will have an impact on the well-being of its population and, thus, also on food sovereignty. The production losses in quantity and quality occur due to the lack of efficient detection of diseases at an early stage. It is very difficult to solve the agriculture efficiency using traditional methods since it takes a long time to be carried out due to detection imprecision of the main diseases, especially when the production areas are extensive. Therefore, the main objective of this research study is to perform a systematic literature review, of the latest five years, of Precision Agriculture (PA) to be able to understand the state of the art of the set of new technologies, procedures, and optimization processes with Artificial Intelligence (AI). This study will focus on Corns, Potatoes, and Rice diagnostic diseases. The extensive literature review will be performed on Elsevier, Scopus, and IEEE databases. In addition, this research will focus on advanced digital imaging processing and the development of software and hardware for PA. The convolution neural network will be handling special attention due to its outstanding diagnostic results. Moreover, the studied data will be incorporated with artificial intelligence algorithms for the automatic diagnosis of crop quality. Finally, precision agriculture with technology applied to the agricultural sector allows the land to be exploited efficiently. This system requires sensors, drones, data acquisition cards, and global positioning systems. This research seeks to merge different areas of science, control engineering, electronics, digital image processing, and artificial intelligence for the development, in the near future, of a low-cost image measurement system that allows the optimization of crops with AI.

Keywords: precision agriculture, convolutional neural network, deep learning, artificial intelligence

Procedia PDF Downloads 63
489 Perceptions and Expectations by Participants of Monitoring and Evaluation Short Course Training Programmes in Africa

Authors: Mokgophana Ramasobana

Abstract:

Background: At the core of the demand to utilize evidence-based approaches in the policy-making cycle, prioritization of limited financial resources and results driven initiatives is the urgency to develop a cohort of competent Monitoring and Evaluation (M&E) practitioners and public servants. The ongoing strides in the evaluation capacity building (ECB) initiatives are a direct response to produce the highly-sought after M&E skills. Notwithstanding the rapid growth of M&E short courses, participants perceived value and expectation of M&E short courses as a panacea for ECB have not been empirically quantified or measured. The objective of this article is to explicitly illustrate the importance of measuring ECB interventions and understanding what works in ECB and why it works. Objectives: This article illustrates the importance of establishing empirical ECB measurement tools to evaluate ECB interventions in order to ascertain its contribution to the broader evaluation practice. Method: The study was primarily a desktop review of existing literature, juxtaposed by a survey of the participants across the African continent based on the 43 M&E short courses hosted by the Centre for Learning on Evaluation and Results Anglophone Africa (CLEAR-AA) in collaboration with the Department of Planning Monitoring and Evaluation (DPME) Results: The article established that participants perceive short course training as a panacea to improve their M&E practical skill critical to executing their organizational duties. In tandem, participants are likely to demand customized training as opposed to general topics in Evaluation. However, the organizational environments constrain the application of the newly acquired skills. Conclusion: This article aims to contribute to the 'how to' measure ECB interventions discourse and contribute towards the improvement to evaluate ECB interventions. The study finds that participants prefer training courses with longer duration to cover more topics. At the same time, whilst organizations call for customization of programmes, the study found that individual participants demand knowledge of generic and popular evaluation topics.

Keywords: evaluation capacity building, effectiveness and training, monitoring and evaluation (M&E) short course training, perceptions and expectations

Procedia PDF Downloads 117
488 A Measurement and Motor Control System for Free Throw Shots in Basketball Using Gyroscope Sensor

Authors: Niloofar Zebarjad

Abstract:

This research aims at finding a tool to provide basketball players with real-time audio feedback on their shooting form in free throw shots. Free throws played a pivotal role in taking the lead in fierce competitions. The major problem in performing an accurate free throw seems to be improper training. Since the arm movement during the free throw shot is complex, the coach or the athlete might miss the movement details during practice. Hence, there is a necessity to create a system that measures arm movements' critical characteristics and control for improper kinematics. The proposed setup in this study quantifies arm kinematics and provides real-time feedback as an audio signal consisting of a gyroscope sensor. Spatial shoulder angle data are transmitted in a mobile application in real-time and can be saved and processed for statistical and analysis purposes. The proposed system is easy to use, inexpensive, portable, and real-time applicable. Objectives: This research aims to modify and control the free throw using audio feedback and determine if and to what extent the new setup reduces errors in arm formations during throws and finally assesses the successful throw rate. Methods: One group of elite basketball athletes and two novice athletes (control and study group) participated in this study. Each group contains 5 participants being studied in three separate sessions over a week. Results: Empirical results showed enhancements in the free throw shooting style, shot pocket (SP), and locked position (LP). The mean values of shoulder angle were controlled on 25° and 45° for SP and LP, respectively, recommended by valid FIBA references. Conclusion: Throughout the experiments, the system helped correct and control the shoulder angles toward the targeted pattern of shot pocket (SP) and locked position (LP). According to the desired results for arm motion, adding another sensor to measure and control the elbow angle is recommended.

Keywords: audio-feedback, basketball, free-throw, locked-position, motor-control, shot-pocket

Procedia PDF Downloads 267
487 Neural Network Based Control Algorithm for Inhabitable Spaces Applying Emotional Domotics

Authors: Sergio A. Navarro Tuch, Martin Rogelio Bustamante Bello, Leopoldo Julian Lechuga Lopez

Abstract:

In recent years, Mexico’s population has seen a rise of different physiological and mental negative states. Two main consequences of this problematic are deficient work performance and high levels of stress generating and important impact on a person’s physical, mental and emotional health. Several approaches, such as the use of audiovisual stimulus to induce emotions and modify a person’s emotional state, can be applied in an effort to decreases these negative effects. With the use of different non-invasive physiological sensors such as EEG, luminosity and face recognition we gather information of the subject’s current emotional state. In a controlled environment, a subject is shown a series of selected images from the International Affective Picture System (IAPS) in order to induce a specific set of emotions and obtain information from the sensors. The raw data obtained is statistically analyzed in order to filter only the specific groups of information that relate to a subject’s emotions and current values of the physical variables in the controlled environment such as, luminosity, RGB light color, temperature, oxygen level and noise. Finally, a neural network based control algorithm is given the data obtained in order to feedback the system and automate the modification of the environment variables and audiovisual content shown in an effort that these changes can positively alter the subject’s emotional state. During the research, it was found that the light color was directly related to the type of impact generated by the audiovisual content on the subject’s emotional state. Red illumination increased the impact of violent images and green illumination along with relaxing images decreased the subject’s levels of anxiety. Specific differences between men and women were found as to which type of images generated a greater impact in either gender. The population sample was mainly constituted by college students whose data analysis showed a decreased sensibility to violence towards humans. Despite the early stage of the control algorithm, the results obtained from the population sample give us a better insight into the possibilities of emotional domotics and the applications that can be created towards the improvement of performance in people’s lives. The objective of this research is to create a positive impact with the application of technology to everyday activities; nonetheless, an ethical problem arises since this can also be applied to control a person’s emotions and shift their decision making.

Keywords: data analysis, emotional domotics, performance improvement, neural network

Procedia PDF Downloads 121
486 Evolution of Nettlespurge Oil Mud for Drilling Mud System: A Comparative Study of Diesel Oil and Nettlespurge Oil as Oil-Based Drilling Mud

Authors: Harsh Agarwal, Pratikkumar Patel, Maharshi Pathak

Abstract:

Recently the low prices of Crude oil and increase in strict environmental regulations limit limits the use of diesel based muds as these muds are relatively costlier and toxic, as a result disposal of cuttings into the eco-system is a major issue faced by the drilling industries. To overcome these issues faced by the Oil Industry, an attempt has been made to develop oil-in-water emulsion mud system using nettlespurge oil. Nettlespurge oil could be easily available and its cost is around ₹30/litre which is about half the price of diesel in India. Oil-based mud (OBM) was formulated with Nettlespurge oil extracted from Nettlespurge seeds using the Soxhlet extraction method. The formulated nettlespurge oil mud properties were analysed with diesel oil mud properties. The compared properties were rheological properties, yield point and gel strength, and mud density and filtration loss properties, fluid loss and filter cake. The mud density measurement showed that nettlespurge OBM was slightly higher than diesel OBM with mud density values of 9.175 lb/gal and 8.5 lb/gal, respectively, at barite content of 70 g. Thus it has a higher lubricating property. Additionally, the filtration loss test results showed that nettlespurge mud fluid loss volumes, oil was 11 ml, compared to diesel oil mud volume of 15 ml. The filtration loss test indicated that the nettlespurge oil mud with filter cake thickness of 2.2 mm had a cake characteristic of thin and squashy while the diesel oil mud resulted in filter cake thickness of 2.7 mm with cake characteristic of tenacious, rubbery and resilient. The filtration loss test results showed that nettlespurge oil mud fluid loss volumes was much less than the diesel based oil mud. The filtration loss test indicated that the nettlespurge oil mud filter cake thickness less than the diesel oil mud filter cake thickness. So Low formation damage and the emulsion stability effect was analysed with this experiment. The nettlespurge oil-in-water mud system had lower coefficient of friction than the diesel oil based mud system. All the rheological properties have shown better results relative to the diesel based oil mud. Therefore, with all the above mentioned factors and with the data of the conducted experiment we could conclude that the Nettlespurge oil based mud is economically and well as eco-logically much more feasible than the worn out and shabby diesel-based oil mud in the Drilling Industry.

Keywords: economical feasible, ecological feasible, emulsion stability, nettle spurge oil, rheological properties, soxhlet extraction method

Procedia PDF Downloads 187
485 The Use of a Novel Visual Kinetic Demonstration Technique in Student Skill Acquisition of the Sellick Cricoid Force Manoeuvre

Authors: L. Nathaniel-Wurie

Abstract:

The Sellick manoeuvre a.k.a the application of cricoid force (CF), was first described by Brian Sellick in 1961. CF is the application of digital pressure against the cricoid cartilage with the intention of posterior force causing oesophageal compression against the vertebrae. This is designed to prevent passive regurgitation of gastric contents, which is a major cause of morbidity and mortality during emergency airway management inside and outside of the hospital. To the authors knowledge, there is no universally standardised training modality and, therefore, no reliable way to examine if there are appropriate outcomes. If force is not measured during training, how can one surmise that appropriate, accurate, or precise amounts of force are being used routinely. Poor homogeneity in teaching and untested outcomes will correlate with reduced efficacy and increased adverse effects. For this study, the accuracy of force delivery in trained professionals was tested, and outcomes contrasted against a novice control and a novice study group. In this study, 20 operating department practitioners were tested (with a mean experience of 5.3years of performing CF). Subsequent contrast with 40 novice students who were randomised into one of two arms. ‘Arm A’ were explained the procedure, then shown the procedure then asked to perform CF with the corresponding force measurement being taken three times. Arm B had the same process as arm A then before being tested, they had 10, and 30 Newtons applied to their hands to increase intuitive understanding of what the required force equated to, then were asked to apply the equivalent amount of force against a visible force metre and asked to hold that force for 20 seconds which allowed direct visualisation and correction of any over or under estimation. Following this, Arm B were then asked to perform the manoeuvre, and the force generated measured three times. This study shows that there is a wide distribution of force produced by trained professionals and novices performing the procedure for the first time. Our methodology for teaching the manoeuvre shows an improved accuracy, precision, and homogeneity within the group when compared to novices and even outperforms trained practitioners. In conclusion, if this methodology is adopted, it may correlate with higher clinical outcomes, less adverse events, and more successful airway management in critical medical scenarios.

Keywords: airway, cricoid, medical education, sellick

Procedia PDF Downloads 61
484 Characterization of Complex Gold Ores for Preliminary Process Selection: The Case of Kapanda, Ibindi, Mawemeru, and Itumbi in Tanzania

Authors: Sospeter P. Maganga, Alphonce Wikedzi, Mussa D. Budeba, Samwel V. Manyele

Abstract:

This study characterizes complex gold ores (elemental and mineralogical composition, gold distribution, ore grindability, and mineral liberation) for preliminary process selection. About 200 kg of ore samples were collected from each location using systematic sampling by mass interval. Ores were dried, crushed, milled, and split into representative sub-samples (about 1 kg) for elemental and mineralogical composition analyses using X-ray fluorescence (XRF), fire assay finished with Atomic Absorption Spectrometer (AAS), and X-ray Diffraction (XRD) methods, respectively. The gold distribution was studied on size-by-size fractions, while ore grindability was determined using the standard Bond test. The mineral liberation analysis was conducted using ThermoFisher Scientific Mineral Liberation Analyzer (MLA) 650, where unsieved polished grain mounts (80% passing 700 µm) were used as MLA feed. Two MLA measurement modes, X-ray modal analysis (XMOD) and sparse phase liberation-grain X-ray mapping analysis (SPL-GXMAP), were employed. At least two cyanide consumers (Cu, Fe, Pb, and Zn) and kinetics impeders (Mn, S, As, and Bi) were present in all locations investigated. Copper content at Kapanda (0.77% Cu) and Ibindi (7.48% Cu) exceeded the recommended threshold of 0.5% Cu for direct cyanidation. The gold ore at Ibindi indicated a higher rate of grinding compared to other locations. This could be explained by the highest grindability (2.119 g/rev.) and lowest Bond work index (10.213 kWh/t) values. The pyrite-marcasite, chalcopyrite, galena, and siderite were identified as major gold, copper, lead, and iron-bearing minerals, respectively, with potential for economic extraction. However, only gold and copper can be recovered under conventional milling because of grain size issues (galena is exposed by 10%) and process complexity (difficult to concentrate and smelt iron from siderite). Therefore, the preliminary process selection is copper flotation followed by gold cyanidation for Kapanda and Ibindi ores, whereas gold cyanidation with additives such as glycine or ammonia is selected for Mawemeru and Itumbi ores because of low concentrations of Cu, Pb, Fe, and Zn minerals.

Keywords: complex gold ores, mineral liberation, ore characterization, ore grindability

Procedia PDF Downloads 59
483 Cost-Effective Mechatronic Gaming Device for Post-Stroke Hand Rehabilitation

Authors: A. Raj Kumar, S. Bilaloglu

Abstract:

Stroke is a leading cause of adult disability worldwide. We depend on our hands for our activities of daily living(ADL). Although many patients regain the ability to walk, they continue to experience long-term hand motor impairments. As the number of individuals with young stroke is increasing, there is a critical need for effective approaches for rehabilitation of hand function post-stroke. Motor relearning for dexterity requires task-specific kinesthetic, tactile and visual feedback. However, when a stroke results in both sensory and motor impairment, it becomes difficult to ascertain when and what type of sensory substitutions can facilitate motor relearning. In an ideal situation, real-time task-specific data on the ability to learn and data-driven feedback to assist such learning will greatly assist rehabilitation for dexterity. We have found that kinesthetic and tactile information from the unaffected hand can assist patients re-learn the use of optimal fingertip forces during a grasp and lift task. Measurement of fingertip grip force (GF), load forces (LF), their corresponding rates (GFR and LFR), and other metrics can be used to gauge the impairment level and progress during learning. Currently ATI mini force-torque sensors are used in research settings to measure and compute the LF, GF, and their rates while grasping objects of different weights and textures. Use of the ATI sensor is cost prohibitive for deployment in clinical or at-home rehabilitation. A cost effective mechatronic device is developed to quantify GF, LF, and their rates for stroke rehabilitation purposes using off-the-shelf components such as load cells, flexi-force sensors, and an Arduino UNO microcontroller. A salient feature of the device is its integration with an interactive gaming environment to render a highly engaging user experience. This paper elaborates the integration of kinesthetic and tactile sensing through computation of LF, GF and their corresponding rates in real time, information processing, and interactive interfacing through augmented reality for visual feedback.

Keywords: feedback, gaming, kinesthetic, rehabilitation, tactile

Procedia PDF Downloads 230
482 IoT Based Soil Moisture Monitoring System for Indoor Plants

Authors: Gul Rahim Rahimi

Abstract:

The IoT-based soil moisture monitoring system for indoor plants is designed to address the challenges of maintaining optimal moisture levels in soil for plant growth and health. The system utilizes sensor technology to collect real-time data on soil moisture levels, which is then processed and analyzed using machine learning algorithms. This allows for accurate and timely monitoring of soil moisture levels, ensuring plants receive the appropriate amount of water to thrive. The main objectives of the system are twofold: to keep plants fresh and healthy by preventing water deficiency and to provide users with comprehensive insights into the water content of the soil on a daily and hourly basis. By monitoring soil moisture levels, users can identify patterns and trends in water consumption, allowing for more informed decision-making regarding watering schedules and plant care. The scope of the system extends to the agriculture industry, where it can be utilized to minimize the efforts required by farmers to monitor soil moisture levels manually. By automating the process of soil moisture monitoring, farmers can optimize water usage, improve crop yields, and reduce the risk of plant diseases associated with over or under-watering. Key technologies employed in the system include the Capacitive Soil Moisture Sensor V1.2 for accurate soil moisture measurement, the Node MCU ESP8266-12E Board for data transmission and communication, and the Arduino framework for programming and development. Additionally, machine learning algorithms are utilized to analyze the collected data and provide actionable insights. Cloud storage is utilized to store and manage the data collected from multiple sensors, allowing for easy access and retrieval of information. Overall, the IoT-based soil moisture monitoring system offers a scalable and efficient solution for indoor plant care, with potential applications in agriculture and beyond. By harnessing the power of IoT and machine learning, the system empowers users to make informed decisions about plant watering, leading to healthier and more vibrant indoor environments.

Keywords: IoT-based, soil moisture monitoring, indoor plants, water management

Procedia PDF Downloads 29
481 Application of Heritage Clay Roof Tiles in Malaysia’s Government Buildings: Conservation Challenges

Authors: Mohd Sabere Sulaiman, Masyitah Abd Aziz, Norsiah Hassan, Jamilah Halina Abdul Halim, Mohd Saipul Asrafi Haron

Abstract:

The use of clay roof tiles was spread out through Asia and Europe, including Malaysia, since the early 17th Century. Most of the common type of clay roof tiles are used in a flat and rectangular shape, measurement, styles, and characteristics through each tradition and interest, including responsive to the climate. Various types of heritage clay roof tiles were used in Malaysia’s Government Buildings dated 1865, 1919, 1936, and so forth, which mostly were imported from India, France, and Italy. Until now, these heritage clay roof tiles are still found throughout Malaysia, including the ‘Interlocking’ clay roof tile type. This study is to investigate and overview the existence of heritage clay roof tiles used in Malaysia; the ‘interlocking’ type with ‘lip’ and ‘hooks’, through literature reviews as desktop study besides carried out a preliminary observation on various sites and interviews. From the literatures, the last production and used of the local heritage clay roof tiles in Malaysia dated in mid 1900s in Batu Arang, Selangor. The brick factory was abandoned since early 2000s. Although the modern ‘Interlocking’ type were produced to duplicate its form, pattern, and size of the original one, they still facing the problem to blend and merged, which end up dismantling the original version, or replacing one to one condition and even replaced overall with the modern materials. This is quite contradicting with the basic principles of building conservation and had become a challenge. Initial findings from the preliminary observation on site in various state in Malaysia shows some evidence that the heritage clay roof tiles are still intact and been used. Some of them might change to modern roof materials such as metal deck, probably due to easy maintenance and cheaper. Also, some are still struggling to maintain and retain its looks and authenticity of the roof while facing the increasing of material cost. Those improper alteration and changes made is due to lack of knowledge among the owner and end user. Various aspect needs to be considered in order to sustain its usage and its original looks by looking at the proper maintenance aspects of the heritage clay roof tiles to prolong the building life for future generation preferences.

Keywords: challenges, clay, interlocking, maintenance

Procedia PDF Downloads 74
480 A Simple Olfactometer for Odour and Lateralization Thresholds of Chemical Vapours

Authors: Lena Ernstgård, Aishwarya M. Dwivedi, Johan Lundström, Gunnar Johanson

Abstract:

A simple inexpensive olfactometer was constructed to enable valid measures of detection threshold of low concentrations of vapours of chemicals. The delivery system consists of seven syringe pumps, each connected to a Tedlar bag containing a predefined concentration of the test chemical in the air. The seven pumps are connected to a 8-way mixing valve which in turn connects to a birhinal nose piece. Chemical vapor of known concentration is generated by injection of an appropriate amount of the test chemical into a Tedlar bag with a known volume of clean air. Complete vaporization is assured by gentle heating of the bag from the outside with a heat flow. The six test concentrations are obtained by adding different volumes from the starting bag to six new Tedlar bags with known volumes of clean air. One bag contains clean air only. Thus, six different test concentrations and clean air can easily be tested in series by shifting the valve to new positions. Initial in-line measurement with a photoionization detector showed that the delivery system quickly responded to a shift in valve position. Thus 90% of the desired concentration was reached within 15 seconds. The concentrations in the bags are verified daily by gas chromatography. The stability of the system in terms of chemical concentration is monitored in real time by means of a photo-ionization detector. To determine lateralization thresholds, an additional pump supplying clean air is added to the delivery system in a way so that the nostrils can be separately and interchangeably be exposed to clean air and test chemical. Odor and lateralization thresholds were determined for three aldehydes; acrolein, crotonaldehyde, and hexanal in 20 healthy naïve individuals. Aldehydes generally have a strong odour, and the selected aldehydes are also considered to be irritating to mucous membranes. The median odor thresholds of the three aldehydes were 0.017, 0.0008, and 0.097 ppm, respectively. No lateralization threshold could be identified for acrolein, whereas the medians for crotonaldehyde and hexanal were 0.003 and 0.39 ppm, respectively. In conclusion, we constructed a simple, inexpensive olfactometer that allows for stable and easily measurable concentrations of vapors of the test chemical. Our test with aldehydes demonstrates that the system produces valid detection among volunteers in terms of odour and lateralization thresholds.

Keywords: irritation, odour delivery, olfactometer, smell

Procedia PDF Downloads 204
479 Geo-Spatial Distribution of Radio Refractivity and the Influence of Fade Depth on Microwave Propagation Signals over Nigeria

Authors: Olalekan Lawrence Ojo

Abstract:

Designing microwave terrestrial propagation networks requires a thorough evaluation of the severity of multipath fading, especially at frequencies below 10 GHz. In nations like Nigeria, without a large enough databases to support the existing empirical models, the mistakes in the prediction technique intended for the evaluation may be severe. The need for higher bandwidth for various satellite applications makes the investigation of the effects of radio refractivity, fading due to multipath, and Geoclimatic factors on satellite propagation links more important. One of the key elements to take into account for the best functioning of microwave frequencies is the clear air effects. This work has taken into account the geographical distribution of radio refractivity and fades depth over a number of stations in Nigeria. Data from five locations in Nigeria—Akure, Enugu, Jos, Minna, and Sokoto—based on five-year (2017–2021) measurement methods of atmospheric pressure, relative, and humidity temperature—at two levels (ground surface and 100 m heights)—are studied to deduced their effects on signals propagated through a µwave communication links. The assessments included considerations for µwave communication systems as well as the impacts of the dry and wet components of radio refractivity, the effects of the fade depth at various frequencies, and a 20 km link distance. The results demonstrate that the percentage occurrence of the dry terms dominated the radio refractivity constituent at the surface level, contributing a minimum of about 78% and a maximum of about 92%, while at heights of 100 meters, the percentage occurrence of the dry terms dominated the radio refractivity constituent, contributing a minimum of about 79% and a maximum of about 92%. The spatial distribution reveals that, regardless of height, the country's tropical rainforest (TRF) and freshwater swampy mangrove (FWSM) regions reported the greatest values of radio refractivity. The statistical estimate shows that fading values can differ by as much as 1.5 dB, especially near the TRF and FWSM coastlines, even during clear air conditions. The current findings will be helpful for budgeting Earth-space microwave links, particularly for the rollout of Nigeria's 5G and 6G projected microcellular networks.

Keywords: fade depth, geoclimatic factor, refractivity, refractivity gradient

Procedia PDF Downloads 59
478 Real-Time Radiological Monitoring of the Atmosphere Using an Autonomous Aerosol Sampler

Authors: Miroslav Hyza, Petr Rulik, Vojtech Bednar, Jan Sury

Abstract:

An early and reliable detection of an increased radioactivity level in the atmosphere is one of the key aspects of atmospheric radiological monitoring. Although the standard laboratory procedures provide detection limits as low as few µBq/m³, their major drawback is the delayed result reporting: typically a few days. This issue is the main objective of the HAMRAD project, which gave rise to a prototype of an autonomous monitoring device. It is based on the idea of sequential aerosol sampling using a carrousel sample changer combined with a gamma-ray spectrometer. In our hardware configuration, the air is drawn through a filter positioned on the carrousel so that it could be rotated into the measuring position after a preset sampling interval. Filter analysis is performed via a 50% HPGe detector inside an 8.5cm lead shielding. The spectrometer output signal is then analyzed using DSP electronics and Gamwin software with preset nuclide libraries and other analysis parameters. After the counting, the filter is placed into a storage bin with a capacity of 250 filters so that the device can run autonomously for several months depending on the preset sampling frequency. The device is connected to a central server via GPRS/GSM where the user can view monitoring data including raw spectra and technological data describing the state of the device. All operating parameters can be remotely adjusted through a simple GUI. The flow rate is continuously adjustable up to 10 m³/h. The main challenge in spectrum analysis is the natural background subtraction. As detection limits are heavily influenced by the deposited activity of radon decay products and the measurement time is fixed, there must exist an optimal sample decay time (delayed spectrum acquisition). To solve this problem, we adopted a simple procedure based on sequential spectrum acquisition and optimal partial spectral sum with respect to the detection limits for a particular radionuclide. The prototyped device proved to be able to detect atmospheric contamination at the level of mBq/m³ per an 8h sampling.

Keywords: aerosols, atmosphere, atmospheric radioactivity monitoring, autonomous sampler

Procedia PDF Downloads 128
477 Critical Success Factors Influencing Construction Project Performance for Different Objectives: Procurement Phase

Authors: Samart Homthong, Wutthipong Moungnoi

Abstract:

Critical success factors (CSFs) and the criteria to measure project success have received much attention over the decades and are among the most widely researched topics in the context of project management. However, although there have been extensive studies on the subject by different researchers, to date, there has been little agreement on the CSFs. The aim of this study is to identify the CSFs that influence the performance of construction projects, and determine their relative importance for different objectives across five stages in the project life cycle. A considerable literature review was conducted that resulted in the identification of 179 individual factors. These factors were then grouped into nine major categories. A questionnaire survey was used to collect data from three groups of respondents: client representatives, consultants, and contractors. Out of 164 questionnaires distributed, 93 were returned, yielding a response rate of 56.7%. Using the mean score, relative importance index, and weighted average method, the top 10 critical factors for each category were identified. The agreement of survey respondents on those categorised factors were analysed using Spearman’s rank correlation. A one-way analysis of variance was then performed to determine whether the mean scores among the various groups of respondents were statistically significant. The findings indicate the most CSFs in each category in procurement phase are: proper procurement programming of materials (time), stability in the price of materials (cost), and determining quality in the construction (quality). They are then followed by safety equipment acquisition and maintenance (health and safety), budgeting allowed in a contractual arrangement for implementing environmental management activities (environment), completeness of drawing documents (productivity), accurate measurement and pricing of bill of quantities (risk management), adequate communication among the project team (human resource), and adequate cost control measures (client satisfaction). An understanding of CSFs would help all interested parties in the construction industry to improve project performance. Furthermore, the results of this study would help construction professionals and practitioners take proactive measures for effective project management.

Keywords: critical success factors, procurement phase, project life cycle, project performance

Procedia PDF Downloads 169
476 Content Monetization as a Mark of Media Economy Quality

Authors: Bela Lebedeva

Abstract:

Characteristics of the Web as a channel of information dissemination - accessibility and openness, interactivity and multimedia news - become wider and cover the audience quickly, positively affecting the perception of content, but blur out the understanding of the journalistic work. As a result audience and advertisers continue migrating to the Internet. Moreover, online targeting allows monetizing not only the audience (as customarily given to traditional media) but also the content and traffic more accurately. While the users identify themselves with the qualitative characteristics of the new market, its actors are formed. Conflict of interests is laid in the base of the economy of their relations, the problem of traffic tax as an example. Meanwhile, content monetization actualizes fiscal interest of the state too. The balance of supply and demand is often violated due to the political risks, particularly in terms of state capitalism, populism and authoritarian methods of governance such social institutions as the media. A unique example of access to journalistic material, limited by monetization of content is a television channel Dozhd' (Rain) in Russian web space. Its liberal-minded audience has a better possibility for discussion. However, the channel could have been much more successful in terms of unlimited free speech. Avoiding state pressure and censorship its management has decided to save at least online performance and monetizing all of the content for the core audience. The study Methodology was primarily based on the analysis of journalistic content, on the qualitative and quantitative analysis of the audience. Reconstructing main events and relationships of actors on the market for the last six years researcher has reached some conclusions. First, under the condition of content monetization the capitalization of its quality will always strive to quality characteristics of user, thereby identifying him. Vice versa, the user's demand generates high-quality journalism. The second conclusion follows the previous one. The growth of technology, information noise, new political challenges, the economy volatility and the cultural paradigm change – all these factors form the content paying model for an individual user. This model defines him as a beneficiary of specific knowledge and indicates the constant balance of supply and demand other conditions being equal. As a result, a new economic quality of information is created. This feature is an indicator of the market as a self-regulated system. Monetized information quality is less popular than that of the Public Broadcasting Service, but this audience is able to make decisions. These very users keep the niche sectors which have more potential of technology development, including the content monetization ways. The third point of the study allows develop it in the discourse of media space liberalization. This cultural phenomenon may open opportunities for the development of social and economic relations architecture both locally and regionally.

Keywords: content monetization, state capitalism, media liberalization, media economy, information quality

Procedia PDF Downloads 226
475 Vertical Village Buildings as Sustainable Strategy to Re-Attract Mega-Cities in Developing Countries

Authors: M. J. Eichner, Y. S. Sarhan

Abstract:

Overall study purpose has been the evaluation of ‘Vertical Villages’ as a new sustainable building typology, reducing significantly negative impacts of rapid urbanization processes in third world capital cities. Commonly in fast-growing cities, housing and job supply, educational and recreational opportunities, as well as public transportation infrastructure, are not accommodating rapid population growth, exposing people to high noise and emission polluted living environments with low-quality neighborhoods and a lack of recreational areas. Like many others, Egypt’s capital city Cairo, according to the UN facing annual population growth rates of up to 428.000 people, is struggling to address the general deterioration of urban living conditions. New settlements typologies and urban reconstruction approach hardly follow sustainable urbanization principles or socio-ecologic urbanization models with severe effects not only for inhabitants but also for the local environment and global climate. The authors prove that ‘Vertical Village’ buildings can offer a sustainable solution for increasing urban density with at the same time improving the living quality and urban environment significantly. Inserting them within high-density urban fabrics the ecologic and socio-cultural conditions of low-quality neighborhoods can be transformed towards districts, considering all needs of sustainable and social urban life. This study analyzes existing building typologies in Cairo’s «low quality - high density» districts Ard el Lewa, Dokki and Mohandesen according to benchmarks for sustainable residential buildings, identifying major problems and deficits. In 3 case study design projects, the sustainable transformation potential through ‘Vertical Village’ buildings are laid out and comparative studies show the improvement of the urban microclimate, safety, social diversity, sense of community, aesthetics, privacy, efficiency, healthiness and accessibility. The main result of the paper is that the disadvantages of density and overpopulation in developing countries can be converted with ‘Vertical Village’ buildings into advantages, achieving attractive and environmentally friendly living environments with multiple synergies. The paper is documenting based on scientific criteria that mixed-use vertical building structures, designed according to sustainable principles of low rise housing, can serve as an alternative to convert «low quality - high density» districts in megacities, opening a pathway for governments to achieve sustainable urban transformation goals. Neglected informal urban districts, home to millions of the poorer population groups, can be converted into healthier living and working environments.

Keywords: sustainable, architecture, urbanization, urban transformation, vertical village

Procedia PDF Downloads 104