Search results for: proportional detector
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 862

Search results for: proportional detector

712 An Experimental Study on the Optimum Installation of Fire Detector for Early Stage Fire Detecting in Rack-Type Warehouses

Authors: Ki Ok Choi, Sung Ho Hong, Dong Suck Kim, Don Mook Choi

Abstract:

Rack type warehouses are different from general buildings in the kinds, amount, and arrangement of stored goods, so the fire risk of rack type warehouses is different from those buildings. The fire pattern of rack type warehouses is different in combustion characteristic and storing condition of stored goods. The initial fire burning rate is different in the surface condition of materials, but the running time of fire is closely related with the kinds of stored materials and stored conditions. The stored goods of the warehouse are consisted of diverse combustibles, combustible liquid, and so on. Fire detection time may be delayed because the residents are less than office and commercial buildings. If fire detectors installed in rack type warehouses are inadaptable, the fire of the warehouse may be the great fire because of delaying of fire detection. In this paper, we studied what kinds of fire detectors are optimized in early detecting of rack type warehouse fire by real-scale fire tests. The fire detectors used in the tests are rate of rise type, fixed type, photo electric type, and aspirating type detectors. We considered optimum fire detecting method in rack type warehouses suggested by the response characteristic and comparative analysis of the fire detectors.

Keywords: fire detector, rack, response characteristic, warehouse

Procedia PDF Downloads 721
711 Sparse Modelling of Cancer Patients’ Survival Based on Genomic Copy Number Alterations

Authors: Khaled M. Alqahtani

Abstract:

Copy number alterations (CNA) are variations in the structure of the genome, where certain regions deviate from the typical two chromosomal copies. These alterations are pivotal in understanding tumor progression and are indicative of patients' survival outcomes. However, effectively modeling patients' survival based on their genomic CNA profiles while identifying relevant genomic regions remains a statistical challenge. Various methods, such as the Cox proportional hazard (PH) model with ridge, lasso, or elastic net penalties, have been proposed but often overlook the inherent dependencies between genomic regions, leading to results that are hard to interpret. In this study, we enhance the elastic net penalty by incorporating an additional penalty that accounts for these dependencies. This approach yields smooth parameter estimates and facilitates variable selection, resulting in a sparse solution. Our findings demonstrate that this method outperforms other models in predicting survival outcomes, as evidenced by our simulation study. Moreover, it allows for a more meaningful interpretation of genomic regions associated with patients' survival. We demonstrate the efficacy of our approach using both real data from a lung cancer cohort and simulated datasets.

Keywords: copy number alterations, cox proportional hazard, lung cancer, regression, sparse solution

Procedia PDF Downloads 21
710 Effect of Birks Constant and Defocusing Parameter on Triple-to-Double Coincidence Ratio Parameter in Monte Carlo Simulation-GEANT4

Authors: Farmesk Abubaker, Francesco Tortorici, Marco Capogni, Concetta Sutera, Vincenzo Bellini

Abstract:

This project concerns with the detection efficiency of the portable triple-to-double coincidence ratio (TDCR) at the National Institute of Metrology of Ionizing Radiation (INMRI-ENEA) which allows direct activity measurement and radionuclide standardization for pure-beta emitter or pure electron capture radionuclides. The dependency of the simulated detection efficiency of the TDCR, by using Monte Carlo simulation Geant4 code, on the Birks factor (kB) and defocusing parameter has been examined especially for low energy beta-emitter radionuclides such as 3H and 14C, for which this dependency is relevant. The results achieved in this analysis can be used for selecting the best kB factor and the defocusing parameter for computing theoretical TDCR parameter value. The theoretical results were compared with the available ones, measured by the ENEA TDCR portable detector, for some pure-beta emitter radionuclides. This analysis allowed to improve the knowledge of the characteristics of the ENEA TDCR detector that can be used as a traveling instrument for in-situ measurements with particular benefits in many applications in the field of nuclear medicine and in the nuclear energy industry.

Keywords: Birks constant, defocusing parameter, GEANT4 code, TDCR parameter

Procedia PDF Downloads 126
709 Experimental Study on Recycled Aggregate Pervious Concrete

Authors: Ji Wenzhan, Zhang Tao, Li Guoyou

Abstract:

Concrete is the most widely used building material in the world. At the same time, the world produces a large amount of construction waste each year. Waste concrete is processed and treated, and the recycled aggregate is used to make pervious concrete, which enables the construction waste to be recycled. Pervious concrete has many advantages such as permeability to water, protection of water resources, and so on. This paper tests the recycled aggregate obtained by crushing high-strength waste concrete (TOU) and low-strength waste concrete (PU), and analyzes the effect of porosity, amount of cement, mineral admixture and recycled aggregate on the strength of permeable concrete. The porosity is inversely proportional to the strength, and the amount of cement used is proportional to the strength. The mineral admixture can effectively improve the workability of the mixture. The quality of recycled aggregates had a significant effect on strength. Compared with concrete using "PU" aggregates, the strength of 7d and 28d concrete using "TOU" aggregates increased by 69.0% and 73.3%, respectively. Therefore, the quality of recycled aggregates should be strictly controlled during production, and the mix ratio should be designed according to different use environments and usage requirements. This test prepared a recycled aggregate permeable concrete with a compressive strength of 35.8 MPa, which can be used for light load roads and provides a reference for engineering applications.

Keywords: recycled aggregate, permeable concrete, compressive strength, permeability

Procedia PDF Downloads 193
708 A Combination of Anisotropic Diffusion and Sobel Operator to Enhance the Performance of the Morphological Component Analysis for Automatic Crack Detection

Authors: Ankur Dixit, Hiroaki Wagatsuma

Abstract:

The crack detection on a concrete bridge is an important and constant task in civil engineering. Chronically, humans are checking the bridge for inspection of cracks to maintain the quality and reliability of bridge. But this process is very long and costly. To overcome such limitations, we have used a drone with a digital camera, which took some images of bridge deck and these images are processed by morphological component analysis (MCA). MCA technique is a very strong application of sparse coding and it explores the possibility of separation of images. In this paper, MCA has been used to decompose the image into coarse and fine components with the effectiveness of two dictionaries namely anisotropic diffusion and wavelet transform. An anisotropic diffusion is an adaptive smoothing process used to adjust diffusion coefficient by finding gray level and gradient as features. These cracks in image are enhanced by subtracting the diffused coarse image into the original image and the results are treated by Sobel edge detector and binary filtering to exhibit the cracks in a fine way. Our results demonstrated that proposed MCA framework using anisotropic diffusion followed by Sobel operator and binary filtering may contribute to an automation of crack detection even in open field sever conditions such as bridge decks.

Keywords: anisotropic diffusion, coarse component, fine component, MCA, Sobel edge detector and wavelet transform

Procedia PDF Downloads 153
707 Study on the DC Linear Stepper Motor to Industrial Applications

Authors: Nolvi Francisco Baggio Filho, Roniele Belusso

Abstract:

Many industrial processes require a precise linear motion. Usually, this movement is achieved with the use of rotary motors combined with electrical control systems and mechanical systems such as gears, pulleys and bearings. Other types of devices are based on linear motors, where the linear motion is obtained directly. The Linear Stepper Motor (MLP) is an excellent solution for industrial applications that require precise positioning and high speed. This study presents an MLP formed by a linear structure and static ferromagnetic material, and a mover structure in which three coils are mounted. Mechanical suspension systems allow a linear movement between static and mover parts, maintaining a constant air gap. The operating principle is based on the tendency of alignment of magnetic flux through the path of least reluctance. The force proportional to the intensity of the electric current and the speed proportional to the frequency of the excitation coils. The study of this device is still based on the use of a numerical and experimental analysis to verify the relationship among electric current applied and planar force developed. In addition, the magnetic field in the air gap region is also monitored.

Keywords: linear stepper motor, planar traction force, reluctance magnetic, industry applications

Procedia PDF Downloads 479
706 Spectroscopic Autoradiography of Alpha Particles on Geologic Samples at the Thin Section Scale Using a Parallel Ionization Multiplier Gaseous Detector

Authors: Hugo Lefeuvre, Jerôme Donnard, Michael Descostes, Sophie Billon, Samuel Duval, Tugdual Oger, Herve Toubon, Paul Sardini

Abstract:

Spectroscopic autoradiography is a method of interest for geological sample analysis. Indeed, researchers may face different issues such as radioelement identification and quantification in the field of environmental studies. Imaging gaseous ionization detectors find their place in geosciences for conducting specific measurements of radioactivity to improve the monitoring of natural processes using naturally-occurring radioactive tracers, but also for the nuclear industry linked to the mining sector. In geological samples, the location and identification of the radioactive-bearing minerals at the thin-section scale remains a major challenge as the detection limit of the usual elementary microprobe techniques is far higher than the concentration of most of the natural radioactive decay products. The spatial distribution of each decay product in the case of uranium in a geomaterial is interesting for relating radionuclides concentration to the mineralogy. The present study aims to provide spectroscopic autoradiography analysis method for measuring the initial energy of alpha particles with a parallel ionization multiplier gaseous detector. The analysis method has been developed thanks to Geant4 modelling of the detector. The track of alpha particles recorded in the gas detector allow the simultaneous measurement of the initial point of emission and the reconstruction of the initial particle energy by a selection based on the linear energy distribution. This spectroscopic autoradiography method was successfully used to reproduce the alpha spectra from a 238U decay chain on a geological sample at the thin-section scale. The characteristics of this measurement are an energy spectrum resolution of 17.2% (FWHM) at 4647 keV and a spatial resolution of at least 50 µm. Even if the efficiency of energy spectrum reconstruction is low (4.4%) compared to the efficiency of a simple autoradiograph (50%), this novel measurement approach offers the opportunity to select areas on an autoradiograph to perform an energy spectrum analysis within that area. This opens up possibilities for the detailed analysis of heterogeneous geological samples containing natural alpha emitters such as uranium-238 and radium-226. This measurement will allow the study of the spatial distribution of uranium and its descendants in geo-materials by coupling scanning electron microscope characterizations. The direct application of this dual modality (energy-position) of analysis will be the subject of future developments. The measurement of the radioactive equilibrium state of heterogeneous geological structures, and the quantitative mapping of 226Ra radioactivity are now being actively studied.

Keywords: alpha spectroscopy, digital autoradiography, mining activities, natural decay products

Procedia PDF Downloads 127
705 Multilevel Modeling of the Progression of HIV/AIDS Disease among Patients under HAART Treatment

Authors: Awol Seid Ebrie

Abstract:

HIV results as an incurable disease, AIDS. After a person is infected with virus, the virus gradually destroys all the infection fighting cells called CD4 cells and makes the individual susceptible to opportunistic infections which cause severe or fatal health problems. Several studies show that the CD4 cells count is the most determinant indicator of the effectiveness of the treatment or progression of the disease. The objective of this paper is to investigate the progression of the disease over time among patient under HAART treatment. Two main approaches of the generalized multilevel ordinal models; namely the proportional odds model and the nonproportional odds model have been applied to the HAART data. Also, the multilevel part of both models includes random intercepts and random coefficients. In general, four models are explored in the analysis and then the models are compared using the deviance information criteria. Of these models, the random coefficients nonproportional odds model is selected as the best model for the HAART data used as it has the smallest DIC value. The selected model shows that the progression of the disease increases as the time under the treatment increases. In addition, it reveals that gender, baseline clinical stage and functional status of the patient have a significant association with the progression of the disease.

Keywords: nonproportional odds model, proportional odds model, random coefficients model, random intercepts model

Procedia PDF Downloads 399
704 Peruvian Diagnostic Reference Levels for Patients Undergoing Different X-Rays Procedures

Authors: Andres Portocarrero Bonifaz, Caterina Sandra Camarena Rodriguez, Ricardo Palma Esparza, Nicolas Antonio Romero Carlos

Abstract:

Reference levels for common X-rays procedures have been set in many protocols. In Peru, during quality control tests, the dose tolerance is set by these international recommendations. Nevertheless, further studies can be made to assess the national reality and relate dose levels with different parameters such as kV, mA/mAs, exposure time, type of processing (digital, digitalized or conventional), etc. In this paper three radiologic procedures were taken into account for study, general X-rays (fixed and mobile), intraoral X-rays (fixed, mobile and portable) and mammography. For this purpose, an Unfors Xi detector was used; the dose was measured at a focus - detector distance which varied depending on the procedure, and was corrected afterward to find the surface entry dose. The data used in this paper was gathered over a period of over 3 years (2015-2018). In addition, each X-ray machine was taken into consideration only once. The results hope to achieve a new standard which reflects the local practice, and address the issues of the ‘Bonn Call for Action’ in Peru. For this purpose, the 75% percentile of the dose of each radiologic procedure was calculated. In future quality control services, those machines with dose values higher than the selected threshold should be informed that they surpass the reference dose levels established in comparison other radiological centers in the country.

Keywords: general X-rays, intraoral X-rays, mammography, reference dose levels

Procedia PDF Downloads 130
703 Method for Tuning Level Control Loops Based on Internal Model Control and Closed Loop Step Test Data

Authors: Arnaud Nougues

Abstract:

This paper describes a two-stage methodology derived from internal model control (IMC) for tuning a proportional-integral-derivative (PID) controller for levels or other integrating processes in an industrial environment. Focus is the ease of use and implementation speed which are critical for an industrial application. Tuning can be done with minimum effort and without the need for time-consuming open-loop step tests on the plant. The first stage of the method applies to levels only: the vessel residence time is calculated from equipment dimensions and used to derive a set of preliminary proportional-integral (PI) settings with IMC. The second stage, re-tuning in closed-loop, applies to levels as well as other integrating processes: a tuning correction mechanism has been developed based on a series of closed-loop simulations with model errors. The tuning correction is done from a simple closed-loop step test and the application of a generic correlation between observed overshoot and integral time correction. A spin-off of the method is that an estimate of the vessel residence time (levels) or open-loop process gain (other integrating process) is obtained from the closed-loop data.

Keywords: closed-loop model identification, IMC-PID tuning method, integrating process control, on-line PID tuning adaptation

Procedia PDF Downloads 195
702 Fusion Neutron Generator Dosimetry and Applications for Medical, Security, and Industry

Authors: Kaouther Bergaui, Nafaa Reguigui, Charles Gary

Abstract:

Characterization and the applications of deuterium-deuterium (DD) neutron generator developed by Adelphie technology and acquired by the National Centre of Nuclear Science and Technology (NCNST) were presented in this work. We study the performance of the neutron generator in terms of neutron yield, production efficiency, and the ionic current as a function of the acceleration voltage at various RF powers. We provide the design and optimization of the PGNAA chamber and thus give insight into the capabilities of the planned PGNAA facility. Additional non-destructive techniques were studied employing the DD neutron generator, such as PGNAA and neutron radiography: The PGNAA is used for determining the concentration of 10B in Si and SiO2 matrices by using a germanium detector HPGe and the results obtained are compared with PGNAA system using a Sodium Iodide detector (NaI (Tl)); Neutron radiography facility was tested and simulated, using a camera device CCD and simulated by the Monte Carlo code; and the explosive detection system (EDS) also simulated using the Monte Carlo code. The study allows us to show that the new models of DD neutron generators are feasible and that superior-quality neutron beams could be produced and used for various applications. The feasibility of Boron neutron capture therapy (BNCT) for cancer treatment using a neutron generator was assessed by optimizing Beam Shaping Assembly (BSA) on a phantom using Monte-Carlo (MCNP6) simulations.

Keywords: neutron generator deuterium-deuterium, Monte Carlo method, radiation, neutron flux, neutron activation analysis, born, neutron radiography, explosive detection, BNCT

Procedia PDF Downloads 168
701 Calibration of Mini TEPC and Measurement of Lineal Energy in a Mixed Radiation Field Produced by Neutrons

Authors: I. C. Cho, W. H. Wen, H. Y. Tsai, T. C. Chao, C. J. Tung

Abstract:

Tissue-equivalent proportional counter (TEPC) is a useful instrument used to measure radiation single-event energy depositions in a subcellular target volume. The quantity of measurements is the microdosimetric lineal energy, which determines the relative biological effectiveness, RBE, for radiation therapy or the radiation-weighting factor, WR, for radiation protection. TEPC is generally used in a mixed radiation field, where each component radiation has its own RBE or WR value. To reduce the pile-up effect during radiotherapy measurements, a miniature TEPC (mini TEPC) with cavity size in the order of 1 mm may be required. In the present work, a homemade mini TEPC with a cylindrical cavity of 1 mm in both the diameter and the height was constructed to measure the lineal energy spectrum of a mixed radiation field with high- and low-LET radiations. Instead of using external radiation beams to penetrate the detector wall, mixed radiation fields were produced by the interactions of neutrons with TEPC walls that contained small plugs of different materials, i.e. Li, B, A150, Cd and N. In all measurements, mini TEPC was placed at the beam port of the Tsing Hua Open-pool Reactor (THOR). Measurements were performed using the propane-based tissue-equivalent gas mixture, i.e. 55% C3H8, 39.6% CO2 and 5.4% N2 by partial pressures. The gas pressure of 422 torr was applied for the simulation of a 1 m diameter biological site. The calibration of mini TEPC was performed using two marking points in the lineal energy spectrum, i.e. proton edge and electron edge. Measured spectra revealed high lineal energy (> 100 keV/m) peaks due to neutron-capture products, medium lineal energy (10 – 100 keV/m) peaks from hydrogen-recoil protons, and low lineal energy (< 10 keV/m) peaks of reactor photons. For cases of Li and B plugs, the high lineal energy peaks were quite prominent. The medium lineal energy peaks were in the decreasing order of Li, Cd, N, A150, and B. The low lineal energy peaks were smaller compared to other peaks. This study demonstrated that internally produced mixed radiations from the interactions of neutrons with different plugs in the TEPC wall provided a useful approach for TEPC measurements of lineal energies.

Keywords: TEPC, lineal energy, microdosimetry, radiation quality

Procedia PDF Downloads 451
700 Coordinated Interference Canceling Algorithm for Uplink Massive Multiple Input Multiple Output Systems

Authors: Messaoud Eljamai, Sami Hidouri

Abstract:

Massive multiple-input multiple-output (MIMO) is an emerging technology for new cellular networks such as 5G systems. Its principle is to use many antennas per cell in order to maximize the network's spectral efficiency. Inter-cellular interference remains a fundamental problem. The use of massive MIMO will not derogate from the rule. It improves performances only when the number of antennas is significantly greater than the number of users. This, considerably, limits the networks spectral efficiency. In this paper, a coordinated detector for an uplink massive MIMO system is proposed in order to mitigate the inter-cellular interference. The proposed scheme combines the coordinated multipoint technique with an interference-cancelling algorithm. It requires the serving cell to send their received symbols, after processing, decision and error detection, to the interfered cells via a backhaul link. Each interfered cell is capable of eliminating intercellular interferences by generating and subtracting the user’s contribution from the received signal. The resulting signal is more reliable than the original received signal. This allows the uplink massive MIMO system to improve their performances dramatically. Simulation results show that the proposed detector improves system spectral efficiency compared to classical linear detectors.

Keywords: massive MIMO, COMP, interference canceling algorithm, spectral efficiency

Procedia PDF Downloads 126
699 Self-Tuning Dead-Beat PD Controller for Pitch Angle Control of a Bench-Top Helicopter

Authors: H. Mansor, S.B. Mohd-Noor, N. I. Othman, N. Tazali, R. I. Boby

Abstract:

This paper presents an improved robust Proportional Derivative controller for a 3-Degree-of-Freedom (3-DOF) bench-top helicopter by using adaptive methodology. Bench-top helicopter is a laboratory scale helicopter used for experimental purposes which is widely used in teaching laboratory and research. Proportional Derivative controller has been developed for a 3-DOF bench-top helicopter by Quanser. Experiments showed that the transient response of designed PD controller has very large steady state error i.e., 50%, which is very serious. The objective of this research is to improve the performance of existing pitch angle control of PD controller on the bench-top helicopter by integration of PD controller with adaptive controller. Usually standard adaptive controller will produce zero steady state error; however response time to reach desired set point is large. Therefore, this paper proposed an adaptive with deadbeat algorithm to overcome the limitations. The output response that is fast, robust and updated online is expected. Performance comparisons have been performed between the proposed self-tuning deadbeat PD controller and standard PD controller. The efficiency of the self-tuning dead beat controller has been proven from the tests results in terms of faster settling time, zero steady state error and capability of the controller to be updated online.

Keywords: adaptive control, deadbeat control, bench-top helicopter, self-tuning control

Procedia PDF Downloads 301
698 The Concentration of Selected Cosmogenic and Anthropogenic Radionuclides in the Ground Layer of the Atmosphere (Polar and Mid-Latitudes Regions)

Authors: A. Burakowska, M. Piotrowski, M. Kubicki, H. Trzaskowska, R. Sosnowiec, B. Myslek-Laurikainen

Abstract:

The most important source of atmospheric radioactivity are radionuclides generated as a result of the impact of primary and secondary cosmic radiation, with the nuclei of nitrogen oxygen and carbon in the upper troposphere and lower stratosphere. This creates about thirty radioisotopes of more than twenty elements. For organisms, the four of them are most important: ³H, ⁷Be, ²²Na, ¹⁴C. The natural radionuclides, which are present in Earth crust, also settle on dust and particles of water vapor. By this means, the derivatives of uranium and thorium, and long-life 40K get into the air. ¹³⁷Cs is the most widespread isotope, that is implemented by humans into the environment. To determine the concentration of radionuclides in the atmosphere, high volume air samplers were used, where the aerosol collection took place on a special filter fabric (Petrianov filter tissue FPP-15-1.5). In 2002 the high volume air sampler AZA-1000 was installed at the Polish Polar Observatory of the Polish Academy of Science in Hornsund, Spitsbergen (77°00’N, 15°33’E), designed to operate in all weather conditions of the cold polar region. Since 1991 (with short breaks) the ASS-500 air sampler has been working, which is located in Swider at the Kalinowski Geophysical Observatory of Geophysics Institute of the Polish Academy of Science (52°07’N, 21°15’E). The following results of radionuclides concentrations were obtained from both stations using gamma spectroscopy analysis: ⁷Be, ¹³⁷Cs, ¹³⁴Cs, ²¹⁰Pb, ⁴⁰K. For gamma spectroscopy analysis HPGe (High Purity Germanium) detector were used. These data were compared with each other. The preliminary results gave evidence that radioactivity measured in aerosols is not proportional to the amount of dust for both studied regions. Furthermore, the results indicate annual variability (seasonal fluctuations) as well as a decrease in the average activity of ⁷Be with increasing latitude. The content of ⁷Be in surface air also indicates the relationship with solar activity cycles.

Keywords: aerosols, air filters, atmospheric beryllium, environmental radionuclides, gamma spectroscopy, mid-latitude regions radionuclides, polar regions radionuclides, solar cycles

Procedia PDF Downloads 115
697 GA3C for Anomalous Radiation Source Detection

Authors: Chia-Yi Liu, Bo-Bin Xiao, Wen-Bin Lin, Hsiang-Ning Wu, Liang-Hsun Huang

Abstract:

In order to reduce the risk of radiation damage that personnel may suffer during operations in the radiation environment, the use of automated guided vehicles to assist or replace on-site personnel in the radiation environment has become a key technology and has become an important trend. In this paper, we demonstrate our proof of concept for autonomous self-learning radiation source searcher in an unknown environment without a map. The research uses GPU version of Asynchronous Advantage Actor-Critic network (GA3C) of deep reinforcement learning to search for radiation sources. The searcher network, based on GA3C architecture, has self-directed learned and improved how search the anomalous radiation source by training 1 million episodes under three simulation environments. In each episode of training, the radiation source position, the radiation source intensity, starting position, are all set randomly in one simulation environment. The input for searcher network is the fused data from a 2D laser scanner and a RGB-D camera as well as the value of the radiation detector. The output actions are the linear and angular velocities. The searcher network is trained in a simulation environment to accelerate the learning process. The well-performance searcher network is deployed to the real unmanned vehicle, Dashgo E2, which mounts LIDAR of YDLIDAR G4, RGB-D camera of Intel D455, and radiation detector made by Institute of Nuclear Energy Research. In the field experiment, the unmanned vehicle is enable to search out the radiation source of the 18.5MBq Na-22 by itself and avoid obstacles simultaneously without human interference.

Keywords: deep reinforcement learning, GA3C, source searching, source detection

Procedia PDF Downloads 91
696 Simulation of the Collimator Plug Design for Prompt-Gamma Activation Analysis in the IEA-R1 Nuclear Reactor

Authors: Carlos G. Santos, Frederico A. Genezini, A. P. Dos Santos, H. Yorivaz, P. T. D. Siqueira

Abstract:

The Prompt-Gamma Activation Analysis (PGAA) is a valuable technique for investigating the elemental composition of various samples. However, the installation of a PGAA system entails specific conditions such as filtering the neutron beam according to the target and providing adequate shielding for both users and detectors. These requirements incur substantial costs, exceeding $100,000, including manpower. Nevertheless, a cost-effective approach involves leveraging an existing neutron beam facility to create a hybrid system integrating PGAA and Neutron Tomography (NT). The IEA-R1 nuclear reactor at IPEN/USP possesses an NT facility with suitable conditions for adapting and implementing a PGAA device. The NT facility offers a thermal flux slightly colder and provides shielding for user protection. The key additional requirement involves designing detector shielding to mitigate high gamma ray background and safeguard the HPGe detector from neutron-induced damage. This study employs Monte Carlo simulations with the MCNP6 code to optimize the collimator plug for PGAA within the IEA-R1 NT facility. Three collimator models are proposed and simulated to assess their effectiveness in shielding gamma and neutron radiation from nucleon fission. The aim is to achieve a focused prompt-gamma signal while shielding ambient gamma radiation. The simulation results indicate that one of the proposed designs is particularly suitable for the PGAA-NT hybrid system.

Keywords: MCNP6.1, neutron, prompt-gamma ray, prompt-gamma activation analysis

Procedia PDF Downloads 44
695 Tuning of Kalman Filter Using Genetic Algorithm

Authors: Hesham Abdin, Mohamed Zakaria, Talaat Abd-Elmonaem, Alaa El-Din Sayed Hafez

Abstract:

Kalman filter algorithm is an estimator known as the workhorse of estimation. It has an important application in missile guidance, especially in lack of accurate data of the target due to noise or uncertainty. In this paper, a Kalman filter is used as a tracking filter in a simulated target-interceptor scenario with noise. It estimates the position, velocity, and acceleration of the target in the presence of noise. These estimations are needed for both proportional navigation and differential geometry guidance laws. A Kalman filter has a good performance at low noise, but a large noise causes considerable errors leads to performance degradation. Therefore, a new technique is required to overcome this defect using tuning factors to tune a Kalman filter to adapt increasing of noise. The values of the tuning factors are between 0.8 and 1.2, they have a specific value for the first half of range and a different value for the second half. they are multiplied by the estimated values. These factors have its optimum values and are altered with the change of the target heading. A genetic algorithm updates these selections to increase the maximum effective range which was previously reduced by noise. The results show that the selected factors have other benefits such as decreasing the minimum effective range that was increased earlier due to noise. In addition to, the selected factors decrease the miss distance for all ranges of this direction of the target, and expand the effective range which leads to increase probability of kill.

Keywords: proportional navigation, differential geometry, Kalman filter, genetic algorithm

Procedia PDF Downloads 487
694 Improvement of the Robust Proportional–Integral–Derivative (PID) Controller Parameters for Controlling the Frequency in the Intelligent Multi-Zone System at the Present of Wind Generation Using the Seeker Optimization Algorithm

Authors: Roya Ahmadi Ahangar, Hamid Madadyari

Abstract:

The seeker optimization algorithm (SOA) is increasingly gaining popularity among the researchers society due to its effectiveness in solving some real-world optimization problems. This paper provides the load-frequency control method based on the SOA for removing oscillations in the power system. A three-zone power system includes a thermal zone, a hydraulic zone and a wind zone equipped with robust proportional-integral-differential (PID) controllers. The result of simulation indicates that load-frequency changes in the wind zone for the multi-zone system are damped in a short period of time. Meanwhile, in the oscillation period, the oscillations amplitude is not significant. The result of simulation emphasizes that the PID controller designed using the seeker optimization algorithm has a robust function and a better performance for oscillations damping compared to the traditional PID controller. The proposed controller’s performance has been compared to the performance of PID controller regulated with Particle Swarm Optimization (PSO) and. Genetic Algorithm (GA) and Artificial Bee Colony (ABC) algorithms in order to show the superior capability of the proposed SOA in regulating the PID controller. The simulation results emphasize the better performance of the optimized PID controller based on SOA compared to the PID controller optimized with PSO, GA and ABC algorithms.

Keywords: load-frequency control, multi zone, robust PID controller, wind generation

Procedia PDF Downloads 283
693 THz Phase Extraction Algorithms for a THz Modulating Interferometric Doppler Radar

Authors: Shaolin Allen Liao, Hual-Te Chien

Abstract:

Various THz phase extraction algorithms have been developed for a novel THz Modulating Interferometric Doppler Radar (THz-MIDR) developed recently by the author. The THz-MIDR differs from the well-known FTIR technique in that it introduces a continuously modulating reference branch, compared to the time-consuming discrete FTIR stepping reference branch. Such change allows real-time tracking of a moving object and capturing of its Doppler signature. The working principle of the THz-MIDR is similar to the FTIR technique: the incoming THz emission from the scene is split by a beam splitter/combiner; one of the beams is continuously modulated by a vibrating mirror or phase modulator and the other split beam is reflected by a reflection mirror; finally both the modulated reference beam and reflected beam are combined by the same beam splitter/combiner and detected by a THz intensity detector (for example, a pyroelectric detector). In order to extract THz phase from the single intensity measurement signal, we have derived rigorous mathematical formulas for 3 Frequency Banded (FB) signals: 1) DC Low-Frequency Banded (LFB) signal; 2) Fundamental Frequency Banded (FFB) signal; and 3) Harmonic Frequency Banded (HFB) signal. The THz phase extraction algorithms are then developed based combinations of 2 or all of these 3 FB signals with efficient algorithms such as Levenberg-Marquardt nonlinear fitting algorithm. Numerical simulation has also been performed in Matlab with simulated THz-MIDR interferometric signal of various Signal to Noise Ratio (SNR) to verify the algorithms.

Keywords: algorithm, modulation, THz phase, THz interferometry doppler radar

Procedia PDF Downloads 308
692 Infant and Child Mortality among the Low Socio-Economic Households in India

Authors: Narendra Kumar

Abstract:

This study uses data from the ‘National Family Health Survey (NFHS-3) 2005-06’ to investigate the predictors of infant and child mortality among low economic households in East and Northeast region. The cross tabulation, life table survival estimates and Cox proportional hazard model techniques have been used to estimate the predictors of infant and child mortality. The life table survival estimates for infant and child mortality shows that infant mortality in female child is lower in comparison to male child but with child mortality, the rates are higher for female in comparison to male child and the Cox proportional hazard model also give highly significant in female in comparison to male child. The infant and child mortality rates among poor households highest in the Central region followed by North and Northeast region and the lowest in South region in comparison to all regions of India. Education of respondent has been found a significant characteristics in both analyzes, further birth interval, respondent occupation, caste/tribe and place of delivery has substantial impact on infant and child mortality among low economic households in East and Northeast region. Finally these findings specified that an increase in parents’ education, improve health care services and improve socioeconomic conditions of low economic households which should in turn raise infant and child survival and should decrease child mortality among low economic households in India.

Keywords: infant, child, mortality, socio-economic, India

Procedia PDF Downloads 288
691 Comparison of Two Maintenance Policies for a Two-Unit Series System Considering General Repair

Authors: Seyedvahid Najafi, Viliam Makis

Abstract:

In recent years, maintenance optimization has attracted special attention due to the growth of industrial systems complexity. Maintenance costs are high for many systems, and preventive maintenance is effective when it increases operations' reliability and safety at a reduced cost. The novelty of this research is to consider general repair in the modeling of multi-unit series systems and solve the maintenance problem for such systems using the semi-Markov decision process (SMDP) framework. We propose an opportunistic maintenance policy for a series system composed of two main units. Unit 1, which is more expensive than unit 2, is subjected to condition monitoring, and its deterioration is modeled using a gamma process. Unit 1 hazard rate is estimated by the proportional hazards model (PHM), and two hazard rate control limits are considered as the thresholds of maintenance interventions for unit 1. Maintenance is performed on unit 2, considering an age control limit. The objective is to find the optimal control limits and minimize the long-run expected average cost per unit time. The proposed algorithm is applied to a numerical example to compare the effectiveness of the proposed policy (policy Ⅰ) with policy Ⅱ, which is similar to policy Ⅰ, but instead of general repair, replacement is performed. Results show that policy Ⅰ leads to lower average cost compared with policy Ⅱ. 

Keywords: condition-based maintenance, proportional hazards model, semi-Markov decision process, two-unit series systems

Procedia PDF Downloads 99
690 Maker-Based Learning in Secondary Mathematics: Investigating Students’ Proportional Reasoning Understanding through Digital Making

Authors: Juan Torralba

Abstract:

Student digital artifacts were investigated, utilizing a qualitative exploratory research design to understand the ways in which students represented their knowledge of seventh-grade proportionality concepts as they participated in maker-based activities that culminated in the creation of digital 3-dimensional models of their dream homes. Representations of the geometric and numeric dimensions of proportionality were analyzed in the written, verbal, and visual data collected from the students. A directed content analysis approach was utilized in the data analysis, as this work aimed to build upon existing research in the field of maker-based STEAM Education. The results from this work show that students can represent their understanding of proportional reasoning through open-ended written responses more accurately than through verbal descriptions or digital artifacts. The geometric and numeric dimensions of proportionality and their respective components of attributes of similarity representation and percents, rates, and ratios representations were the most represented by the students than any other across the data, suggesting a maker-based instructional approach to teaching proportionality in the middle grades may be promising in helping students gain a solid foundation in those components. Recommendations for practice and research are discussed.

Keywords: learning through making, maker-based education, maker education in the middle grades, making in mathematics, the maker movement

Procedia PDF Downloads 50
689 The application of Gel Dosimeters and Comparison with other Dosimeters in Radiotherapy: A Literature Review

Authors: Sujan Mahamud

Abstract:

Purpose: A major challenge in radiotherapy treatment is to deliver precise dose of radiation to the tumor with minimum dose to the healthy normal tissues. Recently, gel dosimetry has emerged as a powerful tool to measure three-dimensional (3D) dose distribution for complex delivery verification and quality assurance. These dosimeters act both as a phantom and detector, thus confirming the versatility of dosimetry technique. The aim of the study is to know the application of Gel Dosimeters in Radiotherapy and find out the comparison with 1D and 2D dimensional dosimeters. Methods and Materials: The study is carried out from Gel Dosimeter literatures. Secondary data and images have been collected from different sources such as different guidelines, books, and internet, etc. Result: Analyzing, verifying, and comparing data from treatment planning system (TPS) is determined that gel dosimeter is a very excellent powerful tool to measure three-dimensional (3D) dose distribution. The TPS calculated data were in very good agreement with the dose distribution measured by the ferrous gel. The overall uncertainty in the ferrous-gel dose determination was considerably reduced using an optimized MRI acquisition protocol and a new MRI scanner. The method developed for comparing measuring gel data with calculated treatment plans, the gel dosimetry method, was proven to be a useful for radiation treatment planning verification. In 1D and 2D Film, the depth dose and lateral for RMSD are 1.8% and 2%, and max (Di-Dj) are 2.5% and 8%. Other side 2D+ ( 3D) Film Gel and Plan Gel for RMSDstruct and RMSDstoch are 2.3% & 3.6% and 1% & 1% and system deviation are -0.6% and 2.5%. The study is investigated that the result fined 2D+ (3D) Film Dosimeter is better than the 1D and 2D Dosimeter. Discussion: Gel Dosimeters is quality control and quality assurance tool which will used the future clinical application.

Keywords: gel dosimeters, phantom, rmsd, QC, detector

Procedia PDF Downloads 130
688 Dose Profiler: A Tracking Device for Online Range Monitoring in Particle Therapy

Authors: G. Battistoni, F. Collamati, E. De Lucia, R. Faccini, C. Mancini-Terracciano, M. Marafini, I. Mattei, S. Muraro, V. Patera, A. Sarti, A. Sciubba, E. Solfaroli Camillocci, M. Toppi, G. Traini, S. M. Valle, C. Voena

Abstract:

Accelerated charged particles, mainly protons and carbon ions, are presently used in Particle Therapy (PT) to treat solid tumors. The precision of PT exploiting the charged particle high localized dose deposition in tissues and biological effectiveness in killing cancer cells demands for an online dose monitoring technique, crucial to improve the quality assurance of treatments: possible patient mis-positionings and biological changes with respect to the CT scan could negatively affect the therapy outcome. In PT the beam range confined in the irradiated target can be monitored thanks to the secondary radiation produced by the interaction of the projectiles with the patient tissue. The Dose Profiler (DP) is a novel device designed to track charged secondary particles and reconstruct their longitudinal emission distribution, correlated to the Bragg peak position. The feasibility of this approach has been demonstrated by dedicated experimental measurements. The DP has been developed in the framework of the INSIDE project, MIUR, INFN and Centro Fermi, Museo Storico della Fisica e Centro Studi e Ricerche 'E. Fermi', Roma, Italy and will be tested at the Proton Therapy center of Trento (Italy) within the end of 2017. The DP combines a tracker, made of six layers of two-view scintillating fibers with square cross section (0.5 x 0.5 mm2) with two layers of two-view scintillating bars (section 12.0 x 0.6 mm2). The electronic readout is performed by silicon photomultipliers. The sensitive area of the tracking planes is 20 x 20 cm2. To optimize the detector layout, a Monte Carlo (MC) simulation based on the FLUKA code has been developed. The complete DP geometry and the track reconstruction code have been fully implemented in the MC. In this contribution, the DP hardware will be described. The expected detector performance computed using a dedicated simulation of a 220 MeV/u carbon ion beam impinging on a PMMA target will be presented, and the result will be discussed in the standard clinical application framework. A possible procedure for real-time beam range monitoring is proposed, following the expectations in actual clinical operation.

Keywords: online range monitoring, particle therapy, quality assurance, tracking detector

Procedia PDF Downloads 226
687 Efficient Estimation for the Cox Proportional Hazards Cure Model

Authors: Khandoker Akib Mohammad

Abstract:

While analyzing time-to-event data, it is possible that a certain fraction of subjects will never experience the event of interest, and they are said to be cured. When this feature of survival models is taken into account, the models are commonly referred to as cure models. In the presence of covariates, the conditional survival function of the population can be modelled by using the cure model, which depends on the probability of being uncured (incidence) and the conditional survival function of the uncured subjects (latency), and a combination of logistic regression and Cox proportional hazards (PH) regression is used to model the incidence and latency respectively. In this paper, we have shown the asymptotic normality of the profile likelihood estimator via asymptotic expansion of the profile likelihood and obtain the explicit form of the variance estimator with an implicit function in the profile likelihood. We have also shown the efficient score function based on projection theory and the profile likelihood score function are equal. Our contribution in this paper is that we have expressed the efficient information matrix as the variance of the profile likelihood score function. A simulation study suggests that the estimated standard errors from bootstrap samples (SMCURE package) and the profile likelihood score function (our approach) are providing similar and comparable results. The numerical result of our proposed method is also shown by using the melanoma data from SMCURE R-package, and we compare the results with the output obtained from the SMCURE package.

Keywords: Cox PH model, cure model, efficient score function, EM algorithm, implicit function, profile likelihood

Procedia PDF Downloads 117
686 The Effect of Mathematical Modeling of Damping on the Seismic Energy Demands

Authors: Selamawit Dires, Solomon Tesfamariam, Thomas Tannert

Abstract:

Modern earthquake engineering and design encompass performance-based design philosophy. The main objective in performance-based design is to achieve a system performing precisely to meet the design objectives so to reduce unintended seismic risks and associated losses. Energy-based earthquake-resistant design is one of the design methodologies that can be implemented in performance-based earthquake engineering. In energy-based design, the seismic demand is usually described as the ratio of the hysteretic to input energy. Once the hysteretic energy is known as a percentage of the input energy, it is distributed among energy-dissipating components of a structure. The hysteretic to input energy ratio is highly dependent on the inherent damping of a structural system. In numerical analysis, damping can be modeled as stiffness-proportional, mass-proportional, or a linear combination of stiffness and mass. In this study, the effect of mathematical modeling of damping on the estimation of seismic energy demands is investigated by considering elastic-perfectly-plastic single-degree-of-freedom systems representing short to long period structures. Furthermore, the seismicity of Vancouver, Canada, is used in the nonlinear time history analysis. According to the preliminary results, the input energy demand is not sensitive to the type of damping models deployed. Hence, consistent results are achieved regardless of the damping models utilized in the numerical analyses. On the other hand, the hysteretic to input energy ratios vary significantly for the different damping models.

Keywords: damping, energy-based seismic design, hysteretic energy, input energy

Procedia PDF Downloads 149
685 Adobe Attenuation Coefficient Determination and Its Comparison with Other Shielding Materials for Energies Found in Common X-Rays Procedures

Authors: Camarena Rodriguez C. S., Portocarrero Bonifaz A., Palma Esparza R., Romero Carlos N. A.

Abstract:

Adobe is a construction material that fulfills the same function as a conventional brick. Widely used since ancient times, it is present in an appreciable percentage of buildings in Latin America. Adobe is a mixture of clay and sand. The interest in the study of the properties of this material arises due to its presence in the infrastructure of hospital´s radiological services, located in places with low economic resources, for the attenuation of radiation. Some materials such as lead and concrete are the most used for shielding and are widely studied in the literature. The present study will determine the mass attenuation coefficient of Adobe. The minimum required thicknesses for the primary and secondary barriers will be estimated for the shielding of radiological facilities where conventional and dental X-rays are performed. For the experimental procedure, an X-ray source emitted direct radiation towards different thicknesses of an Adobe barrier, and a detector was placed on the other side. For this purpose, an UNFORS Xi solid state detector was used, which collected information on the difference of radiation intensity. The initial parameters of the exposure started at 45 kV; and then the tube tension was varied in increments of 5 kV, reaching a maximum of 125 kV. The X-Ray tube was positioned at a distance of 0.5 m from the surface of the Adobe bricks, and the collimation of the radiation beam was set for an area of 0.15 m x 0.15 m. Finally, mathematical methods were applied to determine the mass attenuation coefficient for different energy ranges. In conclusion, the mass attenuation coefficient for Adobe was determined and the approximate thicknesses of the most common Adobe barriers in the hospital buildings were calculated for their later application in the radiological protection.

Keywords: Adobe, attenuation coefficient, radiological protection, shielding, x-rays

Procedia PDF Downloads 138
684 Mobilizing Resources for Social Entrepreneurial Opportunity: A Framework of Engagement Strategy

Authors: Balram Bhushan

Abstract:

The emergence of social entrepreneurship challenges the strict categorization of not-for-profit, for-profit and hybrid organizations. Although the blurring of boundaries helps social entrepreneurial organizations (SEOs) make better use of emerging opportunities, it poses a significant challenge while mobilizing money from different sources. Additionally, for monetary resources, the legal framework of the host country may further complicate the issue by imposing strict accounting standards. Under such circumstances, the resource providers fail to recognize the suitable engagement strategy with the SEO of their choice. Based on the process of value creation and value capture, this paper develops a guiding framework for resource providers to design an appropriate mix of engagement with the identified SEOs. Essentially, social entrepreneurship creates value at the societal level, but value capture is a characteristic of an organization. Additionally, SEOs prefer value creation over value capture. The paper argued that the nature of the relationship between value creation and value capture determines the extent of blurred boundaries of the organization. Accordingly, synergistic, antagonistic and sequential relationships were proposed between value capture and value creation. When value creation is synergistically associated with value creation, the preferred nature of such action falls within the nature of for-profit organizations within the strictest legal framework. Banks offering micro-loans are good examples of this category. Opposite to this, the antagonist relationship between value creation and value capture, where value capture opportunities are sacrificed for value creation, dictates non-profit organizational structure. Examples of this category include non-government organizations and charity organizations. Finally, the sequential relationship between value capture opportunities is followed for value creation opportunities and guides the action closer to the hybrid structure. Examples of this category include organizations where a non-for-profit unit controls for-profit units of the organization either legally or structurally. As an SEO may attempt to utilize multiple entrepreneurial opportunities falling across any of the three relationships between value creation and value capture, the resource providers need to evaluate an appropriate mix of these relationships before designing their engagement strategies. The paper suggests three guiding principles for the engagement strategy. First, the extent of investment should be proportional to the synergistic relationship between value capture and value creation. Second, the subsidized support should be proportional to the sequential relationship. Finally, the funding (charity contribution) should be proportional to the antagonistic relationship. Finally, the resource providers are needed to keep a close watch on the evolving relationship between value creation and value capture for introducing appropriate changes in their engagement strategy.

Keywords: social entrepreneurship, value creation, value capture, entrepreneurial opportunity

Procedia PDF Downloads 105
683 Factors Controlling Durability of Some Egyptian Non-Stylolitic Marbleized Limestone to Salt Weathering

Authors: H. El Shayab, G. M. Kamh, N. G. Abdel Ghafour, M. L. Abdel Latif

Abstract:

Nowadays, marbleized limestone becomes one of the most important sources of the mineral wealth in Egypt as they have beautiful colors (white, grey, rose, yellow and creamy, etc.) make it very suitable for decoration purposes. Non-styolitic marbleized limestone which not contains styolitic surfaces. The current study aims to study different factors controlling durability of non-styolitic marbleized limestone against salt crystallization weathering. The achievement aim of the research was required nine representative samples were collected from the studied areas. Three samples from each of the studied areas. The studied samples was characterized by various instrumental methods before salt weathering, to determine its mineralogical composition, chemical composition and pore physical properties respectively. The obtained results revealed that both of Duwi and Delga studied samples nearly have the same average ∆M% 1.63 and 1.51 respectively and consequently A.I. stage of deformation. On the other hand, average ∆M% of Wata studied samples is 0.29 i.e. lower than two other studied areas. Wata studied samples are more durable against salt crystallization test than Duwi and Delga. The difference in salt crystallization durability may be resulted from one of the following factors: Microscopic textural effect as both of micrite and skeletal percent are in directly proportional to durability of stones to salt weathering. Dolomite mineral present as a secondary are in indirectly proportional to durability of stones to salt weathering. Increase in MgO% also associated with decrease the durability of studied samples against salt crystallization test. Finally, all factors affecting positively against salt crystallization test presents in Wadi Wata studied samples rather than others two areas.

Keywords: marbleized limestone, salt weathering, Wata, salt weathering

Procedia PDF Downloads 294