Search results for: radiation processing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5075

Search results for: radiation processing

3125 Automatic Identification of Pectoral Muscle

Authors: Ana L. M. Pavan, Guilherme Giacomini, Allan F. F. Alves, Marcela De Oliveira, Fernando A. B. Neto, Maria E. D. Rosa, Andre P. Trindade, Diana R. De Pina

Abstract:

Mammography is a worldwide image modality used to diagnose breast cancer, even in asymptomatic women. Due to its large availability, mammograms can be used to measure breast density and to predict cancer development. Women with increased mammographic density have a four- to sixfold increase in their risk of developing breast cancer. Therefore, studies have been made to accurately quantify mammographic breast density. In clinical routine, radiologists perform image evaluations through BIRADS (Breast Imaging Reporting and Data System) assessment. However, this method has inter and intraindividual variability. An automatic objective method to measure breast density could relieve radiologist’s workload by providing a first aid opinion. However, pectoral muscle is a high density tissue, with similar characteristics of fibroglandular tissues. It is consequently hard to automatically quantify mammographic breast density. Therefore, a pre-processing is needed to segment the pectoral muscle which may erroneously be quantified as fibroglandular tissue. The aim of this work was to develop an automatic algorithm to segment and extract pectoral muscle in digital mammograms. The database consisted of thirty medio-lateral oblique incidence digital mammography from São Paulo Medical School. This study was developed with ethical approval from the authors’ institutions and national review panels under protocol number 3720-2010. An algorithm was developed, in Matlab® platform, for the pre-processing of images. The algorithm uses image processing tools to automatically segment and extract the pectoral muscle of mammograms. Firstly, it was applied thresholding technique to remove non-biological information from image. Then, the Hough transform is applied, to find the limit of the pectoral muscle, followed by active contour method. Seed of active contour is applied in the limit of pectoral muscle found by Hough transform. An experienced radiologist also manually performed the pectoral muscle segmentation. Both methods, manual and automatic, were compared using the Jaccard index and Bland-Altman statistics. The comparison between manual and the developed automatic method presented a Jaccard similarity coefficient greater than 90% for all analyzed images, showing the efficiency and accuracy of segmentation of the proposed method. The Bland-Altman statistics compared both methods in relation to area (mm²) of segmented pectoral muscle. The statistic showed data within the 95% confidence interval, enhancing the accuracy of segmentation compared to the manual method. Thus, the method proved to be accurate and robust, segmenting rapidly and freely from intra and inter-observer variability. It is concluded that the proposed method may be used reliably to segment pectoral muscle in digital mammography in clinical routine. The segmentation of the pectoral muscle is very important for further quantifications of fibroglandular tissue volume present in the breast.

Keywords: active contour, fibroglandular tissue, hough transform, pectoral muscle

Procedia PDF Downloads 352
3124 Occupational Heat Stress Condition According to Wet Bulb Globe Temperature Index in Textile Processing Unit: A Case Study of Surat, Gujarat, India

Authors: Dharmendra Jariwala, Robin Christian

Abstract:

Thermal exposure is a common problem in every manufacturing industry where heat is used in the manufacturing process. In developing countries like India, a lack of awareness regarding the proper work environmental condition is observed among workers. Improper planning of factory building, arrangement of machineries, ventilation system, etc. play a vital role in the rise of temperature within the manufacturing areas. Due to the uncontrolled thermal stress, workers may be subjected to various heat illnesses from mild disorder to heat stroke. Heat stress is responsible for the health risk and reduction in production. Wet Bulb Globe Temperature (WBGT) index and relative humidity are used to evaluate heat stress conditions. WBGT index is a weighted average of natural wet bulb temperature, globe temperature, dry bulb temperature, which are measured with standard instrument QuestTemp 36 area stress monitor. In this study textile processing units have been selected in the industrial estate in the Surat city. Based on the manufacturing process six locations were identified within the plant at which process was undertaken at 120°C to 180°C. These locations were jet dying machine area, stenter machine area, printing machine, looping machine area, washing area which generate process heat. Office area was also selected for comparision purpose as a sixth location. Present Study was conducted in the winter season and summer season for day and night shift. The results shows that average WBGT index was found above Threshold Limiting Value (TLV) during summer season for day and night shift in all three industries except office area. During summer season highest WBGT index of 32.8°C was found during day shift and 31.5°C was found during night shift at printing machine area. Also during winter season highest WBGT index of 30°C and 29.5°C was found at printing machine area during day shift and night shift respectively.

Keywords: relative humidity, textile industry, thermal stress, WBGT

Procedia PDF Downloads 176
3123 Emotional Processing Difficulties in Recovered Anorexia Nervosa Patients: State or Trait

Authors: Telma Fontao de Castro, Kylee Miller, Maria Xavier Araújo, Isabel Brandao, Sandra Torres

Abstract:

Objective: There is a dearth of research investigating the long-term emotional functioning of individuals recovered from anorexia nervosa (AN). This 15-year longitudinal study aimed to examine whether difficulties in cognitive processing of emotions persisted after long-term AN recovery and its link to anxiety and depression. Method: Twenty-four females, who were tested longitudinally during their acute and recovered AN phases, and 24 healthy control (HC) women, were screened for anxiety, depression, alexithymia, and emotion regulation difficulties (ER; only assessed in recovery phase). Results: Anxiety, depression, and alexithymia levels decreased significantly with AN recovery. However, scores on anxiety and difficulty in identifying feelings (alexithymia factor) remained high when compared to the HC group. Scores on emotion regulation difficulties were also lower in HC group. The abovementioned differences between AN recovered group and HC group in difficulties in identifying and accepting feelings and lack of emotional clarity were no longer present when the effect of anxiety and depression was controlled. Conclusions: Findings suggest that emotional dysfunction tends to decrease in AN recovered phase. However, using an HC group as a reference, we conclude that several emotional difficulties are still increased after long-term AN recovery, in particular, limited access to emotion regulation strategies, and difficulty controlling impulses and engaging in goal-directed behavior, thus suggesting to be a trait vulnerability. In turn, competencies related to emotional clarity and acceptance of emotional responses seem to be state-dependent phenomena linked to anxiety and depression. In sum, managing emotions remains a challenge for individuals recovered from AN. Under this circumstance, maladaptive eating behavior can serve as an affect regulatory function, increasing the risk of relapse. Emotional education and stabilization of depressive and anxious symptomatology after recovery emerge as an important avenue to protect from long-term AN relapse.

Keywords: alexithymia, anorexia nervosa, emotion recognition, emotion regulation

Procedia PDF Downloads 124
3122 Developing Manufacturing Process for the Graphene Sensors

Authors: Abdullah Faqihi, John Hedley

Abstract:

Biosensors play a significant role in the healthcare sectors, scientific and technological progress. Developing electrodes that are easy to manufacture and deliver better electrochemical performance is advantageous for diagnostics and biosensing. They can be implemented extensively in various analytical tasks such as drug discovery, food safety, medical diagnostics, process controls, security and defence, in addition to environmental monitoring. Development of biosensors aims to create high-performance electrochemical electrodes for diagnostics and biosensing. A biosensor is a device that inspects the biological and chemical reactions generated by the biological sample. A biosensor carries out biological detection via a linked transducer and transmits the biological response into an electrical signal; stability, selectivity, and sensitivity are the dynamic and static characteristics that affect and dictate the quality and performance of biosensors. In this research, a developed experimental study for laser scribing technique for graphene oxide inside a vacuum chamber for processing of graphene oxide is presented. The processing of graphene oxide (GO) was achieved using the laser scribing technique. The effect of the laser scribing on the reduction of GO was investigated under two conditions: atmosphere and vacuum. GO solvent was coated onto a LightScribe DVD. The laser scribing technique was applied to reduce GO layers to generate rGO. The micro-details for the morphological structures of rGO and GO were visualised using scanning electron microscopy (SEM) and Raman spectroscopy so that they could be examined. The first electrode was a traditional graphene-based electrode model, made under normal atmospheric conditions, whereas the second model was a developed graphene electrode fabricated under a vacuum state using a vacuum chamber. The purpose was to control the vacuum conditions, such as the air pressure and the temperature during the fabrication process. The parameters to be assessed include the layer thickness and the continuous environment. Results presented show high accuracy and repeatability achieving low cost productivity.

Keywords: laser scribing, lightscribe DVD, graphene oxide, scanning electron microscopy

Procedia PDF Downloads 125
3121 Molecular Approach for the Detection of Lactic Acid Bacteria in the Kenyan Spontaneously Fermented Milk, Mursik

Authors: John Masani Nduko, Joseph Wafula Matofari

Abstract:

Many spontaneously fermented milk products are produced in Kenya, where they are integral to the human diet and play a central role in enhancing food security and income generation via small-scale enterprises. Fermentation enhances product properties such as taste, aroma, shelf-life, safety, texture, and nutritional value. Some of these products have demonstrated therapeutic and probiotic effects although recent reports have linked some to death, biotoxin infections, and esophageal cancer. These products are mostly processed from poor quality raw materials under unhygienic conditions resulting to inconsistent product quality and limited shelf-lives. Though very popular, research on their processing technologies is low, and none of the products has been produced under controlled conditions using starter cultures. To modernize the processing technologies for these products, our study aims at describing the microbiology and biochemistry of a representative Kenyan spontaneously fermented milk product, Mursik using modern biotechnology (DNA sequencing) and their chemical composition. Moreover, co-creation processes reflecting stakeholders’ experiences on traditional fermented milk production technologies and utilization, ideals and senses of value, which will allow the generation of products based on common ground for rapid progress will be discussed. Knowledge of the value of clean starting raw material will be emphasized, the need for the definition of fermentation parameters highlighted, and standard equipment employment to attain controlled fermentation discussed. This presentation will review the available information regarding traditional fermented milk (Mursik) and highlight our current research work on the application of molecular approaches (metagenomics) for the valorization of Mursik production process through starter culture/ probiotic strains isolation and identification, and quality and safety aspects of the product. The importance of the research and future research areas on the same subject will also be highlighted.

Keywords: lactic acid bacteria, high throughput biotechnology, spontaneous fermentation, Mursik

Procedia PDF Downloads 296
3120 Energy Conservation Strategies of Buildings in Hot, Arid Region: Al-Khobar, Saudi Arabia

Authors: M. H. Shwehdi, S. Raja Mohammad

Abstract:

Recently energy savings have become more pronounced as a result of the world financial crises as well the unstable oil prices. Certainly all entities needs to adapt Energy Conservation and Management Strategies due to high monthly consumption of their spread locations and advancements of its telecom systems. These system improvements necessitate the establishment of more exchange centers as well provide energy savings. This paper investigates the impact of HVAC System Characteristics, Operational Strategies, the impact of Envelope Thermal Characteristics, and energy conservation measures. These are classified under three types of measures i.e. Zero-Investment; Low-Investment and High-Investment Energy Conservation Measures. The study shows that the Energy Conservation Measures (ECMs) pertaining to the HVAC system characteristics and operation represent the highest potential for energy reduction, attention should be given to window thermal and solar radiation characteristics when large window areas are used. The type of glazing system needs to be carefully considered in the early design phase of future buildings. Paper will present the thermal optimization of different size centers in the two hot-dry and hot-humid Saudi Arabian city of Al Khobar, East province.

Keywords: energy conservation, optimization, thermal design, intermittent operation, exchange centers, hot-humid climate, Saudi Arabia

Procedia PDF Downloads 454
3119 Geological Structure Identification in Semilir Formation: An Correlated Geological and Geophysical (Very Low Frequency) Data for Zonation Disaster with Current Density Parameters and Geological Surface Information

Authors: E. M. Rifqi Wilda Pradana, Bagus Bayu Prabowo, Meida Riski Pujiyati, Efraim Maykhel Hagana Ginting, Virgiawan Arya Hangga Reksa

Abstract:

The VLF (Very Low Frequency) method is an electromagnetic method that uses low frequencies between 10-30 KHz which results in a fairly deep penetration. In this study, the VLF method was used for zonation of disaster-prone areas by identifying geological structures in the form of faults. Data acquisition was carried out in Trimulyo Region, Jetis District, Bantul Regency, Special Region of Yogyakarta, Indonesia with 8 measurement paths. This study uses wave transmitters from Japan and Australia to obtain Tilt and Elipt values that can be used to create RAE (Rapat Arus Ekuivalen or Current Density) sections that can be used to identify areas that are easily crossed by electric current. This section will indicate the existence of a geological structure in the form of faults in the study area which is characterized by a high RAE value. In data processing of VLF method, it is obtained Tilt vs Elliptical graph and Moving Average (MA) Tilt vs Moving Average (MA) Elipt graph of each path that shows a fluctuating pattern and does not show any intersection at all. Data processing uses Matlab software and obtained areas with low RAE values that are 0%-6% which shows medium with low conductivity and high resistivity and can be interpreted as sandstone, claystone, and tuff lithology which is part of the Semilir Formation. Whereas a high RAE value of 10% -16% which shows a medium with high conductivity and low resistivity can be interpreted as a fault zone filled with fluid. The existence of the fault zone is strengthened by the discovery of a normal fault on the surface with strike N550W and dip 630E at coordinates X= 433256 and Y= 9127722 so that the activities of residents in the zone such as housing, mining activities and other activities can be avoided to reduce the risk of natural disasters.

Keywords: current density, faults, very low frequency, zonation

Procedia PDF Downloads 176
3118 A Single-Channel BSS-Based Method for Structural Health Monitoring of Civil Infrastructure under Environmental Variations

Authors: Yanjie Zhu, André Jesus, Irwanda Laory

Abstract:

Structural Health Monitoring (SHM), involving data acquisition, data interpretation and decision-making system aim to continuously monitor the structural performance of civil infrastructures under various in-service circumstances. The main value and purpose of SHM is identifying damages through data interpretation system. Research on SHM has been expanded in the last decades and a large volume of data is recorded every day owing to the dramatic development in sensor techniques and certain progress in signal processing techniques. However, efficient and reliable data interpretation for damage detection under environmental variations is still a big challenge. Structural damages might be masked because variations in measured data can be the result of environmental variations. This research reports a novel method based on single-channel Blind Signal Separation (BSS), which extracts environmental effects from measured data directly without any prior knowledge of the structure loading and environmental conditions. Despite the successful application in audio processing and bio-medical research fields, BSS has never been used to detect damage under varying environmental conditions. This proposed method optimizes and combines Ensemble Empirical Mode Decomposition (EEMD), Principal Component Analysis (PCA) and Independent Component Analysis (ICA) together to separate structural responses due to different loading conditions respectively from a single channel input signal. The ICA is applying on dimension-reduced output of EEMD. Numerical simulation of a truss bridge, inspired from New Joban Line Arakawa Railway Bridge, is used to validate this method. All results demonstrate that the single-channel BSS-based method can recover temperature effects from mixed structural response recorded by a single sensor with a convincing accuracy. This will be the foundation of further research on direct damage detection under varying environment.

Keywords: damage detection, ensemble empirical mode decomposition (EEMD), environmental variations, independent component analysis (ICA), principal component analysis (PCA), structural health monitoring (SHM)

Procedia PDF Downloads 308
3117 Risk Assessment of Radiation Hazard for a Typical WWER1000: Cancer Risk Analysis during a Hypothetical Accident

Authors: R. Gharari, N. Kojouri, R. Hosseini Aghdam, E. Alibeigi, B. Salmasian

Abstract:

In this research, the WWER1000/V446 (a PWR Russian type reactor) is chosen as the case study. It is assumed that radioactive materials that release into the environment are more than allowable limit due to a complete failure of the ventilation system (reactor stack). In the following, the HOTSPOT and the RASCAL computational codes have been used and coupled with a developed program using MATLAB software to evaluate Total effective dose equivalent (TEDE) and cancer risk according to the BEIR equations for various human organs. In addition, effects of the containment spray system and climate conditions on the TEDE have been investigated. According to the obtained results, there is an inverse correlation between the received dose and the wind speed; the amount of the TEDE for wind speed 2 m/s and is more than wind speed for 14 m/s during the class A of the climate (2.168 and 0.444 mSv, respectively). Also, containment spray system can effect and reduce the amount of the fission products and TEDE. Furthermore, the probability of the cancer risk for women is more than men, and for children is more than adults. In addition, a specific emergency zonal planning is proposed. Results are promising in which the site selection of the WWER1000/V446 were considered safe for the public in this situation.

Keywords: TEDE, total effective dose equivalent, RASCAL and HOTSPOT codes, BEIR equations, cancer risk

Procedia PDF Downloads 166
3116 Automatic Furrow Detection for Precision Agriculture

Authors: Manpreet Kaur, Cheol-Hong Min

Abstract:

The increasing advancement in the robotics equipped with machine vision sensors applied to precision agriculture is a demanding solution for various problems in the agricultural farms. An important issue related with the machine vision system concerns crop row and weed detection. This paper proposes an automatic furrow detection system based on real-time processing for identifying crop rows in maize fields in the presence of weed. This vision system is designed to be installed on the farming vehicles, that is, submitted to gyros, vibration and other undesired movements. The images are captured under image perspective, being affected by above undesired effects. The goal is to identify crop rows for vehicle navigation which includes weed removal, where weeds are identified as plants outside the crop rows. The images quality is affected by different lighting conditions and gaps along the crop rows due to lack of germination and wrong plantation. The proposed image processing method consists of four different processes. First, image segmentation based on HSV (Hue, Saturation, Value) decision tree. The proposed algorithm used HSV color space to discriminate crops, weeds and soil. The region of interest is defined by filtering each of the HSV channels between maximum and minimum threshold values. Then the noises in the images were eliminated by the means of hybrid median filter. Further, mathematical morphological processes, i.e., erosion to remove smaller objects followed by dilation to gradually enlarge the boundaries of regions of foreground pixels was applied. It enhances the image contrast. To accurately detect the position of crop rows, the region of interest is defined by creating a binary mask. The edge detection and Hough transform were applied to detect lines represented in polar coordinates and furrow directions as accumulations on the angle axis in the Hough space. The experimental results show that the method is effective.

Keywords: furrow detection, morphological, HSV, Hough transform

Procedia PDF Downloads 233
3115 Genetic Data of Deceased People: Solving the Gordian Knot

Authors: Inigo de Miguel Beriain

Abstract:

Genetic data of deceased persons are of great interest for both biomedical research and clinical use. This is due to several reasons. On the one hand, many of our diseases have a genetic component; on the other hand, we share genes with a good part of our biological family. Therefore, it would be possible to improve our response considerably to these pathologies if we could use these data. Unfortunately, at the present moment, the status of data on the deceased is far from being satisfactorily resolved by the EU data protection regulation. Indeed, the General Data Protection Regulation has explicitly excluded these data from the category of personal data. This decision has given rise to a fragmented legal framework on this issue. Consequently, each EU member state offers very different solutions. For instance, Denmark considers the data as personal data of the deceased person for a set period of time while some others, such as Spain, do not consider this data as such, but have introduced some specifically focused regulations on this type of data and their access by relatives. This is an extremely dysfunctional scenario from multiple angles, not least of which is scientific cooperation at the EU level. This contribution attempts to outline a solution to this dilemma through an alternative proposal. Its main hypothesis is that, in reality, health data are, in a sense, a rara avis within data in general because they do not refer to one person but to several. Hence, it is possible to think that all of them can be considered data subjects (although not all of them can exercise the corresponding rights in the same way). When the person from whom the data were obtained dies, the data remain as personal data of his or her biological relatives. Hence, the general regime provided for in the GDPR may apply to them. As these are personal data, we could go back to thinking in terms of a general prohibition of data processing, with the exceptions provided for in Article 9.2 and on the legal bases included in Article 6. This may be complicated in practice, given that, since we are dealing with data that refer to several data subjects, it may be complex to refer to some of these bases, such as consent. Furthermore, there are theoretical arguments that may oppose this hypothesis. In this contribution, it is shown, however, that none of these objections is of sufficient substance to delegitimize the argument exposed. Therefore, the conclusion of this contribution is that we can indeed build a general framework on the processing of personal data of deceased persons in the context of the GDPR. This would constitute a considerable improvement over the current regulatory framework, although it is true that some clarifications will be necessary for its practical application.

Keywords: collective data conceptual issues, data from deceased people, genetic data protection issues, GDPR and deceased people

Procedia PDF Downloads 155
3114 Finite Volume Method Simulations of GaN Growth Process in MOVPE Reactor

Authors: J. Skibinski, P. Caban, T. Wejrzanowski, K. J. Kurzydlowski

Abstract:

In the present study, numerical simulations of heat and mass transfer during gallium nitride growth process in Metal Organic Vapor Phase Epitaxy reactor AIX-200/4RF-S is addressed. Existing knowledge about phenomena occurring in the MOVPE process allows to produce high quality nitride based semiconductors. However, process parameters of MOVPE reactors can vary in certain ranges. Main goal of this study is optimization of the process and improvement of the quality of obtained crystal. In order to investigate this subject a series of computer simulations have been performed. Numerical simulations of heat and mass transfer in GaN epitaxial growth process have been performed to determine growth rate for various mass flow rates and pressures of reagents. According to the fact that it’s impossible to determine experimentally the exact distribution of heat and mass transfer inside the reactor during the process, modeling is the only solution to understand the process precisely. Main heat transfer mechanisms during MOVPE process are convection and radiation. Correlation of modeling results with the experiment allows to determine optimal process parameters for obtaining crystals of highest quality.

Keywords: Finite Volume Method, semiconductors, epitaxial growth, metalorganic vapor phase epitaxy, gallium nitride

Procedia PDF Downloads 401
3113 Using Squeezed Vacuum States to Enhance the Sensitivity of Ground Based Gravitational Wave Interferometers beyond the Standard Quantum Limit

Authors: Giacomo Ciani

Abstract:

This paper reviews the impact of quantum noise on modern gravitational wave interferometers and explains how squeezed vacuum states are used to push the noise below the standard quantum limit. With the first detection of gravitational waves from a pair of colliding black holes in September 2015 and subsequent detections including that of gravitational waves from a pair of colliding neutron stars, the ground-based interferometric gravitational wave observatories LIGO and VIRGO have opened the era of gravitational-wave and multi-messenger astronomy. Improving the sensitivity of the detectors is of paramount importance to increase the number and quality of the detections, fully exploiting this new information channel about the universe. Although still in the commissioning phase and not at nominal sensitivity, these interferometers are designed to be ultimately limited by a combination of shot noise and quantum radiation pressure noise, which define an envelope known as the standard quantum limit. Despite the name, this limit can be beaten with the use of advanced quantum measurement techniques, with the use of squeezed vacuum states being currently the most mature and promising. Different strategies for implementation of the technology in the large-scale detectors, in both their frequency-independent and frequency-dependent variations, are presented, together with an analysis of the main technological issues and expected sensitivity gain.

Keywords: gravitational waves, interferometers, squeezed vacuum, standard quantum limit

Procedia PDF Downloads 153
3112 The 6Rs of Radiobiology in Photodynamic Therapy: Review

Authors: Kave Moloudi, Heidi Abrahamse, Blassan P. George

Abstract:

Radiotherapy (RT) and photodynamic therapy (PDT) are both forms of cancer treatment that aim to kill cancer cells while minimizing damage to healthy tissue. The similarity between RT and PDT lies in their mechanism of action. Both treatments use energy to damage cancer cells. RT uses high-energy radiation to damage the DNA of cancer cells, while PDT uses light energy to activate a photosensitizing agent, which produces reactive oxygen species (ROS) that damage the cancer cells. Both treatments require careful planning and monitoring to ensure the correct dose is delivered to the tumor while minimizing damage to surrounding healthy tissue. They are also often used in combination with other treatments, such as surgery or chemotherapy, to improve overall outcomes. However, there are also significant differences between RT and PDT. For example, RT is a non-invasive treatment that can be delivered externally or internally, while PDT requires the injection of a photosensitizing agent and the use of a specialized light source to activate it. Additionally, the side effects and risks associated with each treatment can vary. In this review, we focus on generalizing the 6Rs of radiobiology in PDT, which can open a window for the clinical application of Radio-photodynamic therapy with minimum side effects. Furthermore, this review can open new insight to work on and design new radio-photosensitizer agents in Radio-photodynamic therapy.

Keywords: radiobiology, photodynamic therapy, radiotherapy, 6Rs in radiobiology, ROS, DNA damages, cellular and molecular mechanism, clinical application.

Procedia PDF Downloads 106
3111 The Effect of Gamma rays on Physicochemical Properties of Carboxymethyl Starch

Authors: N. Rajeswara Rao, T. Venkatappa Rao, K. Sowri Babu, N. Srinivas Rao, P. S. V. Shanmukhi

Abstract:

Carboxymethyl Starch (CMS) is a biopolymer derived from starch by the substitution method. CMS is proclaimed to have improved physicochemical properties than native starch. The present work deals with the effect of gamma radiation on the physicochemical properties of CMS. The samples were exposed to gamma irradiation of doses 30, 60 and 90 kGy. The resultant properties were studied with electron spin resonance (ESR), fourier transform infrared spectrometer (FTIR), differential scanning calorimeter (DSC), X-ray diffractometer (XRD) and scanning electron microscopy. Irradiation of CMS by gamma rays initiates cleavage of glucosidic bonds producing different types of radicals. Some of these radicals convert to peroxy radicals by abstracting oxygen. The ESR spectrum of CMS is anisotropic and is thought to be due to the superposition of various component spectra. In order to analyze the ESR spectrum, computer simulations were also employed. ESR spectra are also recorded under different conditions like post-irradiation times, variable temperatures and saturation behavior in order to evaluate the stability of free radicals produced on irradiation. Thermal studies from DSC depict that for CMS the gelatization process was absconded at higher doses. Relative crystallinity was reduced significantly after irradiation from XRD Studies. FTIR studies also confirm the same aspect. From ESR studies, it was concluded that irradiated CMS could be a potential reference material in ESR dosimetry.

Keywords: gamma rays, free radicals, ESR simulations, gelatization

Procedia PDF Downloads 106
3110 Experimental Device for Fluorescence Measurement by Optical Fiber Combined with Dielectrophoretic Sorting in Microfluidic Chips

Authors: Jan Jezek, Zdenek Pilat, Filip Smatlo, Pavel Zemanek

Abstract:

We present a device that combines fluorescence spectroscopy with fiber optics and dielectrophoretic micromanipulation in PDMS (poly-(dimethylsiloxane)) microfluidic chips. The device allows high speed detection (in the order of kHz) of the fluorescence signal, which is coming from the sample by an inserted optical fiber, e.g. from a micro-droplet flow in a microfluidic chip, or even from the liquid flowing in the transparent capillary, etc. The device uses a laser diode at a wavelength suitable for excitation of fluorescence, excitation and emission filters, optics for focusing the laser radiation into the optical fiber, and a highly sensitive fast photodiode for detection of fluorescence. The device is combined with dielectrophoretic sorting on a chip for sorting of micro-droplets according to their fluorescence intensity. The electrodes are created by lift-off technology on a glass substrate, or by using channels filled with a soft metal alloy or an electrolyte. This device found its use in screening of enzymatic reactions and sorting of individual fluorescently labelled microorganisms. The authors acknowledge the support from the Grant Agency of the Czech Republic (GA16-07965S) and Ministry of Education, Youth and Sports of the Czech Republic (LO1212) together with the European Commission (ALISI No. CZ.1.05/2.1.00/01.0017).

Keywords: dielectrophoretic sorting, fiber optics, laser, microfluidic chips, microdroplets, spectroscopy

Procedia PDF Downloads 722
3109 The Effect of Vitamin D Deficiency on Endothelial Function in Atherosclerosis Patients Living in Saudi Arabia

Authors: Wedad Azhar

Abstract:

Vitamin D is an essential fat-soluble vitamin that is required for the maintenance of good health. It is obtained either through exposure to sunlight (ultraviolet B radiation) or through dietary sources. The role of vitamin D is beyond bone health. Indeed, it plays a critical role in the immune system and a broad range of organ functions such as the cardiovascular system. Moreover, vitamin D plays a critical role in the endothelial function, which is one of the main indicators of atherosclerosis. This study is investigating the correlation between vitamin D status and endothelial function in preventing and treating atherosclerosis especially in country that has ample of sunshine but yet, Saudis from suffering from this issue vitamin D deficiency and insufficiency. Ninety participants from both genders and aged 40 to 60will be involved. The participants will be categorised into three groups: the control group will be healthy persons, patients at risk of developing atherosclerosis, patients formally diagnosed atherosclerosis. Half of the participants in each group should already have been taking vitamin D supplementations. Fasting blood samples will be taken from the participants for biochemical assays. Endothelial function will be assist by flow-mediated dilation of the brachial artery. Participants will be asked to complete a questionnaire on their social and economic status, education level, daily exposure to sunlight, smoking status, consumption of supplements and medication, and a food frequency of vitamin D intake. The data will be analysed using SPSS.

Keywords: atherosclerosis, endothelial function, nutrition, vitamin D

Procedia PDF Downloads 296
3108 GPU Based Real-Time Floating Object Detection System

Authors: Jie Yang, Jian-Min Meng

Abstract:

A GPU-based floating object detection scheme is presented in this paper which is designed for floating mine detection tasks. This system uses contrast and motion information to eliminate as many false positives as possible while avoiding false negatives. The GPU computation platform is deployed to allow detecting objects in real-time. From the experimental results, it is shown that with certain configuration, the GPU-based scheme can speed up the computation up to one thousand times compared to the CPU-based scheme.

Keywords: object detection, GPU, motion estimation, parallel processing

Procedia PDF Downloads 477
3107 Influence of 50 Hz, 1m Tesla Electromagnetic Fields on Serum Male Sex Hormones of Male Rats

Authors: Randa M. Mostafa, Y. Moustafa

Abstract:

During our daily life, we are continuously exposed to the extremely low frequency electromagnetic fields (ELF-EMFs) generated by electric appliances. The possible relation between exposure to (ELF-MFs) and adverse health effects has attracted and passed through long debate sessions. Extremely low frequency is a term used to describe radiation frequencies below 300 Hertz (Hz).It is very important for public health because of the widespread use of electrical power at 50-60 Hz in most countries. This study set out to investigate the impact of chronic exposure of male rats to 50- Hz, 1 mTesla (ELF-EMF) of over periods of 1, 2, and 4 weeks on concentration of serum FSH, LH, and testosterone hormones. 60 male albino rats were divided into 6 groups and were continuously exposed to 50-Hz, 1 m Tesla (ELF-EMF) generated by magnetic field chamber for periods of 1, 2, and 4 weeks. For each experimental point, sham treated group was used as a control. Assay of serum testosterone LH, and FSHwere performed. Serum testosterone showed no significant changes. FSH showed significant increase than sham exposed group after 1 week of field exposure. LH showed significant increase than sham exposed group only after 4 weeks of field exposure. A future detailed molecular studies must be carried out to figure out and may be able to explain the possible interactions between ELF-EMF and hypothalamic-pituitary gonadal axis.

Keywords: extremely low frequency electromagnetic fields, testosterone, follicular stimulating hormone, LH

Procedia PDF Downloads 461
3106 A Literature Review of Precision Agriculture: Applications of Diagnostic Diseases in Corn, Potato, and Rice Based on Artificial Intelligence

Authors: Carolina Zambrana, Grover Zurita

Abstract:

The food loss production that occurs in deficient agricultural production is one of the major problems worldwide. This puts the population's food security and the efficiency of farming investments at risk. It is to be expected that this food security will be achieved with the own and efficient production of each country. It will have an impact on the well-being of its population and, thus, also on food sovereignty. The production losses in quantity and quality occur due to the lack of efficient detection of diseases at an early stage. It is very difficult to solve the agriculture efficiency using traditional methods since it takes a long time to be carried out due to detection imprecision of the main diseases, especially when the production areas are extensive. Therefore, the main objective of this research study is to perform a systematic literature review, of the latest five years, of Precision Agriculture (PA) to be able to understand the state of the art of the set of new technologies, procedures, and optimization processes with Artificial Intelligence (AI). This study will focus on Corns, Potatoes, and Rice diagnostic diseases. The extensive literature review will be performed on Elsevier, Scopus, and IEEE databases. In addition, this research will focus on advanced digital imaging processing and the development of software and hardware for PA. The convolution neural network will be handling special attention due to its outstanding diagnostic results. Moreover, the studied data will be incorporated with artificial intelligence algorithms for the automatic diagnosis of crop quality. Finally, precision agriculture with technology applied to the agricultural sector allows the land to be exploited efficiently. This system requires sensors, drones, data acquisition cards, and global positioning systems. This research seeks to merge different areas of science, control engineering, electronics, digital image processing, and artificial intelligence for the development, in the near future, of a low-cost image measurement system that allows the optimization of crops with AI.

Keywords: precision agriculture, convolutional neural network, deep learning, artificial intelligence

Procedia PDF Downloads 81
3105 The Design and Implementation of a Calorimeter for Evaluation of the Thermal Performance of Materials: The Case of Phase Change Materials

Authors: Ebrahim Solgi, Zahra Hamedani, Behrouz Mohammad Kari, Ruwan Fernando, Henry Skates

Abstract:

The use of thermal energy storage (TES) as part of a passive design strategy can reduce a building’s energy demand. TES materials do this by increasing the lag between energy consumption and energy supply by absorbing, storing and releasing energy in a controlled manner. The increase of lightweight construction in the building industry has made it harder to utilize thermal mass. Consequently, Phase Change Materials (PCMs) are a promising alternative as they can be manufactured in thin layers and used with lightweight construction to store latent heat. This research investigates utilizing PCMs, with the first step being measuring their performance under experimental conditions. To do this requires three components. The first is a calorimeter for measuring indoor thermal conditions, the second is a pyranometer for recording the solar conditions: global, diffuse and direct radiation and the third is a data-logger for recording temperature and humidity for the studied period. This paper reports on the design and implementation of an experimental setup used to measure the thermal characteristics of PCMs as part of a wall construction. The experimental model has been simulated with the software EnergyPlus to create a reliable simulation model that warrants further investigation.

Keywords: phase change materials, EnergyPlus, experimental evaluation, night ventilation

Procedia PDF Downloads 261
3104 Achieving Flow at Work: An Experience Sampling Study to Comprehend How Cognitive Task Characteristics and Work Environments Predict Flow Experiences

Authors: Jonas De Kerf, Rein De Cooman, Sara De Gieter

Abstract:

For many decades, scholars have aimed to understand how work can become more meaningful by maximizing both potential and enhancing feelings of satisfaction. One of the largest contributions towards such positive psychology was made with the introduction of the concept of ‘flow,’ which refers to a condition in which people feel intense engagement and effortless action. Since then, valuable research on work-related flow has indicated that this state of mind is related to positive outcomes for both organizations (e.g., social, supportive climates) and workers (e.g., job satisfaction). Yet, scholars still do not fully comprehend how such deep involvement at work is obtained, given the notion that flow is considered a short-term, complex, and dynamic experience. Most research neglects that people who experience flow ought to be optimally challenged so that intense concentration is required. Because attention is at the core of this enjoyable state of mind, this study aims to comprehend how elements that affect workers’ cognitive functioning impact flow at work. Research on cognitive performance suggests that working on mentally demanding tasks (e.g., information processing tasks) requires workers to concentrate deeply, as a result leading to flow experiences. Based on social facilitation theory, working on such tasks in an isolated environment eases concentration. Prior research has indicated that working at home (instead of working at the office) or in a closed office (rather than in an open-plan office) impacts employees’ overall functioning in terms of concentration and productivity. Consequently, we advance such knowledge and propose an interaction by combining cognitive task characteristics and work environments among part-time teleworkers. Hence, we not only aim to shed light on the relation between cognitive tasks and flow but also provide empirical evidence that workers performing such tasks achieve the highest states of flow while working either at home or in closed offices. In July 2022, an experience-sampling study will be conducted that uses a semi-random signal schedule to understand how task and environment predictors together impact part-time teleworkers’ flow. More precisely, about 150 knowledge workers will fill in multiple surveys a day for two consecutive workweeks to report their flow experiences, cognitive tasks, and work environments. Preliminary results from a pilot study indicate that on a between level, tasks high in information processing go along with high self-reported fluent productivity (i.e., making progress). As expected, evidence was found for higher fluency in productivity for workers performing information processing tasks both at home and in a closed office, compared to those performing the same tasks at the office or in open-plan offices. This study expands the current knowledge on work-related flow by looking at a task and environmental predictors that enable workers to obtain such a peak state. While doing so, our findings suggest that practitioners should strive for ideal alignments between tasks and work locations to work with both deep involvement and gratification.

Keywords: cognitive work, office lay-out, work location, work-related flow

Procedia PDF Downloads 104
3103 Effect of Birks Constant and Defocusing Parameter on Triple-to-Double Coincidence Ratio Parameter in Monte Carlo Simulation-GEANT4

Authors: Farmesk Abubaker, Francesco Tortorici, Marco Capogni, Concetta Sutera, Vincenzo Bellini

Abstract:

This project concerns with the detection efficiency of the portable triple-to-double coincidence ratio (TDCR) at the National Institute of Metrology of Ionizing Radiation (INMRI-ENEA) which allows direct activity measurement and radionuclide standardization for pure-beta emitter or pure electron capture radionuclides. The dependency of the simulated detection efficiency of the TDCR, by using Monte Carlo simulation Geant4 code, on the Birks factor (kB) and defocusing parameter has been examined especially for low energy beta-emitter radionuclides such as 3H and 14C, for which this dependency is relevant. The results achieved in this analysis can be used for selecting the best kB factor and the defocusing parameter for computing theoretical TDCR parameter value. The theoretical results were compared with the available ones, measured by the ENEA TDCR portable detector, for some pure-beta emitter radionuclides. This analysis allowed to improve the knowledge of the characteristics of the ENEA TDCR detector that can be used as a traveling instrument for in-situ measurements with particular benefits in many applications in the field of nuclear medicine and in the nuclear energy industry.

Keywords: Birks constant, defocusing parameter, GEANT4 code, TDCR parameter

Procedia PDF Downloads 150
3102 Toward Subtle Change Detection and Quantification in Magnetic Resonance Neuroimaging

Authors: Mohammad Esmaeilpour

Abstract:

One of the important open problems in the field of medical image processing is detection and quantification of small changes. In this poster, we try to investigate that, how the algebraic decomposition techniques can be used for semiautomatically detecting and quantifying subtle changes in Magnetic Resonance (MR) neuroimaging volumes. We mostly focus on the low-rank values of the matrices achieved from decomposing MR image pairs during a period of time. Besides, a skillful neuroradiologist will help the algorithm to distinguish between noises and small changes.

Keywords: magnetic resonance neuroimaging, subtle change detection and quantification, algebraic decomposition, basis functions

Procedia PDF Downloads 477
3101 Acoustic Analysis of Ball Bearings to Identify Localised Race Defect

Authors: M. Solairaju, Nithin J. Thomas, S. Ganesan

Abstract:

Each and every rotating part of a machine element consists of bearings within its structure. In particular, the rolling element bearings such as cylindrical roller bearing and deep groove ball bearings are frequently used. Improper handling, excessive loading, improper lubrication and sealing cause bearing damage. Hence health monitoring of bearings is an important aspect for radiation pattern of bearing vibration is computed using the dipole model. Sound pressure level for defect-free and race defect the prolonged life of machinery and auto motives. This paper presents modeling and analysis of Acoustic response of deep groove ball bearing with localized race defects. Most of the ball bearings, especially in machine tool spindles and high-speed applications are pre-loaded along an axial direction. The present study is carried out with axial preload. Based on the vibration response, the orbit motion of the inner race is studied, and it was found that the oscillation takes place predominantly in the axial direction. Simplified acoustic is estimated. Acoustic response shows a better indication in identifying the defective bearing. The computed sound signal is visualized in diagrammatic representation using Symmetrised Dot Pattern (SDP). SDP gives better visual distinction between the defective and defect-free bearing

Keywords: bearing, dipole, noise, sound

Procedia PDF Downloads 296
3100 Study of the Design and Simulation Work for an Artificial Heart

Authors: Mohammed Eltayeb Salih Elamin

Abstract:

This study discusses the concept of the artificial heart using engineering concepts, of the fluid mechanics and the characteristics of the non-Newtonian fluid. For the purpose to serve heart patients and improve aspects of their lives and since the Statistics review according to world health organization (WHO) says that heart disease and blood vessels are the first cause of death in the world. Statistics shows that 30% of the death cases in the world by the heart disease, so simply we can consider it as the number one leading cause of death in the entire world is heart failure. And since the heart implantation become a very difficult and not always available, the idea of the artificial heart become very essential. So it’s important that we participate in the developing this idea by searching and finding the weakness point in the earlier designs and hoping for improving it for the best of humanity. In this study a pump was designed in order to pump blood to the human body and taking into account all the factors that allows it to replace the human heart, in order to work at the same characteristics and the efficiency of the human heart. The pump was designed on the idea of the diaphragm pump. Three models of blood obtained from the blood real characteristics and all of these models were simulated in order to study the effect of the pumping work on the fluid. After that, we study the properties of this pump by using Ansys15 software to simulate blood flow inside the pump and the amount of stress that it will go under. The 3D geometries modeling was done using SOLID WORKS and the geometries then imported to Ansys design modeler which is used during the pre-processing procedure. The solver used throughout the study is Ansys FLUENT. This is a tool used to analysis the fluid flow troubles and the general well-known term used for this branch of science is known as Computational Fluid Dynamics (CFD). Basically, Design Modeler used during the pre-processing procedure which is a crucial step before the start of the fluid flow problem. Some of the key operations are the geometry creations which specify the domain of the fluid flow problem. Next is mesh generation which means discretization of the domain to solve governing equations at each cell and later, specify the boundary zones to apply boundary conditions for the problem. Finally, the pre–processed work will be saved at the Ansys workbench for future work continuation.

Keywords: Artificial heart, computational fluid dynamic heart chamber, design, pump

Procedia PDF Downloads 461
3099 Numerical Modeling of hybrid Photovoltaic-Thermoelectric Solar Unit by Applying Various Cross-Sections of Cooling Ducts

Authors: Ziba Khalili, Mohsen Sheikholeslami, Ladan Momayez

Abstract:

Combining the photovoltaic/thermal (PVT) systems with a thermoelectric (TE) module can raise energy yields since the TE module boosts the system's energy conversion efficiency. In the current study, a PVT system integrated with a TE module was designed and simulated in ANSYS Fluent 19.2. A copper heat transfer tube (HTT) was employed for cooling the photovoltaic (PV) cells. Four different shapes of HTT cross-section, i.e., circular, square, elliptical, and triangular, with equal cross-section areas were investigated. Also, the influence of Cu-Al2O3/water hybrid nanofluid (0.024% volume concentration), fluid inlet velocity (uᵢ ), and amount of solar radiation (G), on the PV temperature (Tₚᵥ) and system performance were investigated. The ambient temperature (Tₐ), wind speed (u𝓌), and fluid inlet temperature (Tᵢ), were considered to be 25°C, 1 m/s, and 27°C, respectively. According to the obtained data, the triangular case had the greatest impact on reducing the compared to other cases. In the triangular case, examination of the effect of hybrid nanofluid showed that the use of hybrid nanofluid at 800 W/m2 led to a reduction of the TPV by 0.6% compared to water, at 0.19 m/s. Moreover, the thermal efficiency ( ) and the overall electrical efficiency (nₜ) of the system improved by 0.93% and 0.22%, respectively, at 0.19 m/s. In a triangular case where G and were 800 W/m2 and 19 m/s, respectively, the highest amount of, thermal power (Eₜ), and, were obtained as 72.76%, 130.84 W and 12.03%, respectively.

Keywords: electrical performance, photovoltaic/thermal, thermoelectric, hybrid nanofluid, thermal efficiency

Procedia PDF Downloads 79
3098 Problems of Boolean Reasoning Based Biclustering Parallelization

Authors: Marcin Michalak

Abstract:

Biclustering is the way of two-dimensional data analysis. For several years it became possible to express such issue in terms of Boolean reasoning, for processing continuous, discrete and binary data. The mathematical backgrounds of such approach — proved ability of induction of exact and inclusion–maximal biclusters fulfilling assumed criteria — are strong advantages of the method. Unfortunately, the core of the method has quite high computational complexity. In the paper the basics of Boolean reasoning approach for biclustering are presented. In such context the problems of computation parallelization are risen.

Keywords: Boolean reasoning, biclustering, parallelization, prime implicant

Procedia PDF Downloads 126
3097 Solar Calculations of Modified Arch (Semi-Spherical) Type Greenhouse System for Bayburt City

Authors: Uğur Çakir, Erol Şahin, Kemal Çomakli, Ayşegül Çokgez Kuş

Abstract:

Solar energy is thought as main source of all energy sources on the world and it can be used in many applications like agricultural areas, heating cooling or direct electricity production directly or indirectly. Greenhousing is the first one of the agricultural activities that solar energy can be used directly in. Greenhouses offer us suitable conditions which can be controlled easily for the growth of the plant and they are made by using a covering material that allows the sun light entering into the system. Covering material can be glass, fiber glass, plastic or another transparent element. This study investigates the solar energy usability rates and solar energy benefiting rates of a semi-spherical (modified arch) type greenhouse system according to different orientations and positions which exists under climatic conditions of Bayburt. In the concept of this study it is tried to determine the best direction and best sizes of a semi-spherical greenhouse to get best solar benefit from the sun. To achieve this aim a modeling study is made by using MATLAB. However this modeling study is running for some determined shapes and greenhouses it can be used for different shaped greenhouses or buildings. The basic parameters are determined as greenhouse azimuth angle, the rate of size of long edge to short and seasonal solar energy gaining of greenhouse.

Keywords: greenhousing, solar energy, direct radiation, renewable energy

Procedia PDF Downloads 482
3096 Electrodynamic Principles for Generation and Wireless Transfer of Energy

Authors: Steven D. P. Moore

Abstract:

An electrical discharge in the air induces an electromagnetic (EM) wave capable of wireless transfer, reception, and conversion back into electrical discharge at a distant location. Following Norton’s ground wave principles, EM wave radiation (EMR) runs parallel to the Earth’s surface. Energy in an EMR wave can move through the air and be focused to create a spark at a distant location, focused by a receiver to generate a local electrical discharge. This local discharge can be amplified and stored but also has the propensity to initiate another EMR wave. In addition to typical EM waves, lightning is also associated with atmospheric events, trans-ionospheric pulse pairs, the most powerful natural EMR signal on the planet. With each lightning strike, regardless of global position, it generates naturally occurring pulse-pairs that are emitted towards space within a narrow cone. An EMR wave can self-propagate, travel at the speed of light, and, if polarized, contain vector properties. If this reflective pulse could be directed by design through structures that have increased probabilities for lighting strikes, it could theoretically travel near the surface of the Earth at light speed towards a selected receiver for local transformation into electrical energy. Through research, there are several influencing parameters that could be modified to model, test, and increase the potential for adopting this technology towards the goal of developing a global grid that utilizes natural sources of energy.

Keywords: electricity, sparkgap, wireless, electromagnetic

Procedia PDF Downloads 191