Search results for: rate of creep method
21680 Investigation of Cavitation in a Centrifugal Pump Using Synchronized Pump Head Measurements, Vibration Measurements and High-Speed Image Recording
Authors: Simon Caba, Raja Abou Ackl, Svend Rasmussen, Nicholas E. Pedersen
Abstract:
It is a challenge to directly monitor cavitation in a pump application during operation because of a lack of visual access to validate the presence of cavitation and its form of appearance. In this work, experimental investigations are carried out in an inline single-stage centrifugal pump with optical access. Hence, it gives the opportunity to enhance the value of CFD tools and standard cavitation measurements. Experiments are conducted using two impellers running in the same volute at 3000 rpm and the same flow rate. One of the impellers used is optimized for lower NPSH₃% by its blade design, whereas the other one is manufactured using a standard casting method. The cavitation is detected by pump performance measurements, vibration measurements and high-speed image recordings. The head drop and the pump casing vibration caused by cavitation are correlated with the visual appearance of the cavitation. The vibration data is recorded in an axial direction of the impeller using accelerometers recording at a sample rate of 131 kHz. The vibration frequency domain data (up to 20 kHz) and the time domain data are analyzed as well as the root mean square values. The high-speed recordings, focusing on the impeller suction side, are taken at 10,240 fps to provide insight into the flow patterns and the cavitation behavior in the rotating impeller. The videos are synchronized with the vibration time signals by a trigger signal. A clear correlation between cloud collapses and abrupt peaks in the vibration signal can be observed. The vibration peaks clearly indicate cavitation, especially at higher NPSHA values where the hydraulic performance is not affected. It is also observed that below a certain NPSHA value, the cavitation started in the inlet bend of the pump. Above this value, cavitation occurs exclusively on the impeller blades. The impeller optimized for NPSH₃% does show a lower NPSH₃% than the standard impeller, but the head drop starts at a higher NPSHA value and is more gradual. Instabilities in the head drop curve of the optimized impeller were observed in addition to a higher vibration level. Furthermore, the cavitation clouds on the suction side appear more unsteady when using the optimized impeller. The shape and location of the cavitation are compared to 3D fluid flow simulations. The simulation results are in good agreement with the experimental investigations. In conclusion, these investigations attempt to give a more holistic view on the appearance of cavitation by comparing the head drop, vibration spectral data, vibration time signals, image recordings and simulation results. Data indicates that a criterion for cavitation detection could be derived from the vibration time-domain measurements, which requires further investigation. Usually, spectral data is used to analyze cavitation, but these investigations indicate that the time domain could be more appropriate for some applications.Keywords: cavitation, centrifugal pump, head drop, high-speed image recordings, pump vibration
Procedia PDF Downloads 18221679 Examining the Extent to Which the Effects of HIV/AIDS Is Addressed in Low Cost Housing Projects in South Africa: The Case of RDP Golf Course Housing Project in Alice Town, Eastern
Authors: Tatenda Manomano
Abstract:
The chronic challenges presented by HIV/AIDS globally have come with extreme negative effects on individuals, families and communities as well as governments. Sub-Saharan Africa remains strongly challenged with South Africa bearing a huge brunt of these. The paper examines the extent to which the effects of HIV/AIDS are addressed in low cost housing projects in South Africa with a case of the RDP Golf Course Housing Project in Alice Town. The study used a triangulation of both qualitative and quantitative methods with the qualitative as the dominant method while the quantitative was less dominant. Findings revealed that infection rate was high; prostitution was high; alcohol abuse was also high; and rape and sexual abuse was also high and there was also lack of hospitals and social workers around the location. These findings prompted this researcher to recommend for proactive policy making that can bolster the challenges faced by these low cost housing projects in accessing health and social services as well as massive campaigns that can promote behavior modification among other things. It is hoped that this paper will be a platform to ring a bell to both government and non-government to augment the campaign against HIV/AIDS in South Africa.Keywords: HIV/AIDS, RDP houses, low cost housing projects, campaigns
Procedia PDF Downloads 40321678 Status of the European Atlas of Natural Radiation
Authors: G. Cinelli, T. Tollefsen, P. Bossew, V. Gruber, R. Braga, M. A. Hernández-Ceballos, M. De Cort
Abstract:
In 2006, the Joint Research Centre (JRC) of the European Commission started the project of the 'European Atlas of Natural Radiation'. The Atlas aims at preparing a collection of maps of Europe displaying the levels of natural radioactivity caused by different sources (indoor and outdoor radon, cosmic radiation, terrestrial radionuclides, terrestrial gamma radiation, etc). The overall goal of the project is to estimate, in geographical resolution, the annual dose that the public may receive from natural radioactivity, combining all the information from the different radiation components. The first map which has been developed is the European map of indoor radon (Rn) since in most cases Rn is the most important contribution to exposure. New versions of the map are realised when new countries join the project or when already participating countries send new data. We show the latest status of this map which currently includes 25 European countries. Second, the JRC has undertaken to map a variable which measures 'what earth delivers' in terms of Rn. The corresponding quantity is called geogenic radon potential (RP). Due to the heterogeneity of data sources across the Europe there is need to develop a harmonized quantity which at the one hand adequately measures or classifies the RP, and on the other hand is suited to accommodate the variety of input data used to estimate this target quantity. Candidates for input quantities which may serve as predictors of the RP, and for which data are available across Europe, to different extent, are Uranium (U) concentration in rocks and soils, soil gas radon and soil permeability, terrestrial gamma dose rate, geological information and indoor data from ground floor. The European Geogenic Radon Map gives the possibility to characterize areas, on European geographical scale, for radon hazard where indoor radon measurements are not available. Parallel to ongoing work on the European Indoor Radon, Geogenic Radon and Cosmic Radiation Maps, we made progress in the development of maps of terrestrial gamma radiation and U, Th and K concentrations in soil and bedrock. We show the first, preliminary map of the terrestrial gamma dose rate, estimated using the data of ambient dose equivalent rate available from the EURDEP system (about 5000 fixed monitoring stations across Europe). Also, the first maps of U, Th, and K concentrations in soil and bedrock are shown in the present work.Keywords: Europe, natural radiation, mapping, indoor radon
Procedia PDF Downloads 29521677 Reliability Qualification Test Plan Derivation Method for Weibull Distributed Products
Authors: Ping Jiang, Yunyan Xing, Dian Zhang, Bo Guo
Abstract:
The reliability qualification test (RQT) is widely used in product development to qualify whether the product meets predetermined reliability requirements, which are mainly described in terms of reliability indices, for example, MTBF (Mean Time Between Failures). It is widely exercised in product development. In engineering practices, RQT plans are mandatorily referred to standards, such as MIL-STD-781 or GJB899A-2009. But these conventional RQT plans in standards are not preferred, as the test plans often require long test times or have high risks for both producer and consumer due to the fact that the methods in the standards only use the test data of the product itself. And the standards usually assume that the product is exponentially distributed, which is not suitable for a complex product other than electronics. So it is desirable to develop an RQT plan derivation method that safely shortens test time while keeping the two risks under control. To meet this end, for the product whose lifetime follows Weibull distribution, an RQT plan derivation method is developed. The merit of the method is that expert judgment is taken into account. This is implemented by applying the Bayesian method, which translates the expert judgment into prior information on product reliability. Then producer’s risk and the consumer’s risk are calculated accordingly. The procedures to derive RQT plans are also proposed in this paper. As extra information and expert judgment are added to the derivation, the derived test plans have the potential to shorten the required test time and have satisfactory low risks for both producer and consumer, compared with conventional test plans. A case study is provided to prove that when using expert judgment in deriving product test plans, the proposed method is capable of finding ideal test plans that not only reduce the two risks but also shorten the required test time as well.Keywords: expert judgment, reliability qualification test, test plan derivation, producer’s risk, consumer’s risk
Procedia PDF Downloads 14321676 Stress and Strain Analysis of Notched Bodies Subject to Non-Proportional Loadings
Authors: Ayhan Ince
Abstract:
In this paper, an analytical simplified method for calculating elasto-plastic stresses strains of notched bodies subject to non-proportional loading paths is discussed. The method was based on the Neuber notch correction, which relates the incremental elastic and elastic-plastic strain energy densities at the notch root and the material constitutive relationship. The validity of the method was presented by comparing computed results of the proposed model against finite element numerical data of notched shaft. The comparison showed that the model estimated notch-root elasto-plastic stresses strains with good accuracy using linear-elastic stresses. The prosed model provides more efficient and simple analysis method preferable to expensive experimental component tests and more complex and time consuming incremental non-linear FE analysis. The model is particularly suitable to perform fatigue life and fatigue damage estimates of notched components subjected to non-proportional loading paths.Keywords: elasto-plastic, stress-strain, notch analysis, nonprortional loadings, cyclic plasticity, fatigue
Procedia PDF Downloads 46721675 A Theoretical Study of Accelerating Neutrons in LINAC Using Magnetic Gradient Method
Authors: Chunduru Amareswara Prasad
Abstract:
The main aim of this proposal it to reveal the secrets of the universe by accelerating neutrons. The proposal idea in its abridged version speaks about the possibility of making neutrons accelerate with help of thermal energy and magnetic energy under controlled conditions. Which is helpful in revealing the hidden secrets of the universe namely dark energy and in finding properties of Higgs boson. The paper mainly speaks about accelerating neutrons to near velocity of light in a LINAC, using magnetic energy by magnetic pressurizers. The center of mass energy of two colliding neutron beams is 94 GeV (~0.5c) can be achieved using this method. The conventional ways to accelerate neutrons has some constraints in accelerating them electromagnetically as they need to be separated from the Tritium or Deuterium nuclei. This magnetic gradient method provides efficient and simple way to accelerate neutrons.Keywords: neutron, acceleration, thermal energy, magnetic energy, Higgs boson
Procedia PDF Downloads 33021674 Analytical Solutions for Tunnel Collapse Mechanisms in Circular Cross-Section Tunnels under Seepage and Seismic Forces
Authors: Zhenyu Yang, Qiunan Chen, Xiaocheng Huang
Abstract:
Reliable prediction of tunnel collapse remains a prominent challenge in the field of civil engineering. In this study, leveraging the nonlinear Hoek-Brown failure criterion and the upper-bound theorem, an analytical solution for the collapse surface of shallowly buried circular tunnels was derived, taking into account the coupled effects of surface loads and pore water pressures. Initially, surface loads and pore water pressures were introduced as external force factors, equating the energy dissipation rate to the external force, yielding our objective function. Subsequently, the variational method was employed for optimization, and the outcomes were juxtaposed with previous research findings. Furthermore, we utilized the deduced equation set to systematically analyze the influence of various rock mass parameters on collapse shape and extent. To validate our analytical solutions, a comparison with prior studies was executed. The corroboration underscored the efficacy of our proposed methodology, offering invaluable insights for collapse risk assessment in practical engineering applications.Keywords: tunnel roof stability, analytical solution, hoek–brown failure criterion, limit analysis
Procedia PDF Downloads 8621673 A TFETI Domain Decompositon Solver for von Mises Elastoplasticity Model with Combination of Linear Isotropic-Kinematic Hardening
Authors: Martin Cermak, Stanislav Sysala
Abstract:
In this paper we present the efficient parallel implementation of elastoplastic problems based on the TFETI (Total Finite Element Tearing and Interconnecting) domain decomposition method. This approach allow us to use parallel solution and compute this nonlinear problem on the supercomputers and decrease the solution time and compute problems with millions of DOFs. In our approach we consider an associated elastoplastic model with the von Mises plastic criterion and the combination of linear isotropic-kinematic hardening law. This model is discretized by the implicit Euler method in time and by the finite element method in space. We consider the system of nonlinear equations with a strongly semismooth and strongly monotone operator. The semismooth Newton method is applied to solve this nonlinear system. Corresponding linearized problems arising in the Newton iterations are solved in parallel by the above mentioned TFETI. The implementation of this problem is realized in our in-house MatSol packages developed in MATLAB.Keywords: isotropic-kinematic hardening, TFETI, domain decomposition, parallel solution
Procedia PDF Downloads 42221672 Dynamic Wind Effects in Tall Buildings: A Comparative Study of Synthetic Wind and Brazilian Wind Standard
Authors: Byl Farney Cunha Junior
Abstract:
In this work the dynamic three-dimensional analysis of a 47-story building located in Goiania city when subjected to wind loads generated using both the Wind Brazilian code, NBR6123 (ABNT, 1988) and the Synthetic-Wind method is realized. To model the frames three different methodologies are used: the shear building model and both bi and three-dimensional finite element models. To start the analysis, a plane frame is initially studied to validate the shear building model and, in order to compare the results of natural frequencies and displacements at the top of the structure the same plane frame was modeled using the finite element method through the SAP2000 V10 software. The same steps were applied to an idealized 20-story spacial frame that helps in the presentation of the stiffness correction process applied to columns. Based on these models the two methods used to generate the Wind loads are presented: a discrete model proposed in the Wind Brazilian code, NBR6123 (ABNT, 1988) and the Synthetic-Wind method. The method uses the Davenport spectrum which is divided into a variety of frequencies to generate the temporal series of loads. Finally, the 47- story building was analyzed using both the three-dimensional finite element method through the SAP2000 V10 software and the shear building model. The models were loaded with Wind load generated by the Wind code NBR6123 (ABNT, 1988) and by the Synthetic-Wind method considering different wind directions. The displacements and internal forces in columns and beams were compared and a comparative study considering a situation of a full elevated reservoir is realized. As can be observed the displacements obtained by the SAP2000 V10 model are greater when loaded with NBR6123 (ABNT, 1988) wind load related to the permanent phase of the structure’s response.Keywords: finite element method, synthetic wind, tall buildings, shear building
Procedia PDF Downloads 27521671 Study of the Efficiency of a Synthetic Wax for Corrosion Protection of Steel in Aggressive Environments
Authors: Laidi Babouri
Abstract:
The remarkable properties of steel, such as hardness and impact resistance, motivate their use in the automotive manufacturing industry. However, due to the very vulnerable environmental conditions of use, the steel that makes up the car body can corrode. This situation is motivating more and more automobile manufacturers to develop research to develop processes minimizing the rate of degradation of the physicomechanical properties of these materials. The present work falls within this perspective; it presents the results of a research study focused on the use of synthetic wax for the protection of steel, type XES (DC04), against corrosion in aggressive environments. The media used in this study are an acid medium with a pH=5.6, a 3% chloride medium, and a dry medium. Evaluation of the protective power of synthetic wax in different environments was carried out using mass loss techniques (immersion), completed by electrochemical techniques (stationary and transient). The results of the immersion of the steel samples, with a surface area of (1.44 cm²), in the various media, for a period of 30 days, using the immersion technique, showed high protective efficiency of synthetic wax in acidic and saline environments, with a lesser degree in a dry environment. Moreover, the study of the protective power, using electrochemical techniques, confirmed the results obtained in static mode (loss of mass), the protective efficiency of synthetic wax, against the corrosion of steel, in different environments, which reaches a maximum rate of 99.87% in a saline environment.Keywords: corrosion, steel, industrial wax, environment, mass loss, electrochemical techniques
Procedia PDF Downloads 7821670 High Performance Liquid Cooling Garment (LCG) Using ThermoCore
Authors: Venkat Kamavaram, Ravi Pare
Abstract:
Modern warfighters experience extreme environmental conditions in many of their operational and training activities. In temperatures exceeding 95°F, the body’s temperature regulation can no longer cool through convection and radiation. In this case, the only cooling mechanism is evaporation. However, evaporative cooling is often compromised by excessive humidity. Natural cooling mechanisms can be further compromised by clothing and protective gear, which trap hot air and moisture close to the body. Creating an efficient heat extraction apparel system that is also lightweight without hindering dexterity or mobility of personnel working in extreme temperatures is a difficult technical challenge and one that needs to be addressed to increase the probability for the future success of the US military. To address this challenge, Oceanit Laboratories, Inc. has developed and patented a Liquid Cooled Garment (LCG) more effective than any on the market today. Oceanit’s LCG is a form-fitting garment with a network of thermally conductive tubes that extracts body heat and can be worn under all authorized and chemical/biological protective clothing. Oceanit specifically designed and developed ThermoCore®, a thermally conductive polymer, for use in this apparel, optimizing the product for thermal conductivity, mechanical properties, manufacturability, and performance temperatures. Thermal Manikin tests were conducted in accordance with the ASTM test method, ASTM F2371, Standard Test Method for Measuring the Heat Removal Rate of Personal Cooling Systems Using a Sweating Heated Manikin, in an environmental chamber using a 20-zone sweating thermal manikin. Manikin test results have shown that Oceanit’s LCG provides significantly higher heat extraction under the same environmental conditions than the currently fielded Environmental Control Vest (ECV) while at the same time reducing the weight. Oceanit’s LCG vests performed nearly 30% better in extracting body heat while weighing 15% less than the ECV. There are NO cooling garments in the market that provide the same thermal extraction performance, form-factor, and reduced weight as Oceanit’s LCG. The two cooling garments that are commercially available and most commonly used are the Environmental Control Vest (ECV) and the Microclimate Cooling Garment (MCG).Keywords: thermally conductive composite, tubing, garment design, form fitting vest, thermocore
Procedia PDF Downloads 11621669 Improving Forecasting Demand for Maintenance Spare Parts: Case Study
Authors: Abdulaziz Afandi
Abstract:
Minimizing the inventory cost, optimizing the inventory quantities, and increasing system operational availability are the main motivations to enhance forecasting demand of spare parts in a major power utility company in Medina. This paper reports in an effort made to optimize the orders quantities of spare parts by improving the method of forecasting the demand. The study focuses on equipment that has frequent spare parts purchase orders with uncertain demand. The pattern of the demand considers a lumpy pattern which makes conventional forecasting methods less effective. A comparison was made by benchmarking various methods of forecasting based on experts’ criteria to select the most suitable method for the case study. Three actual data sets were used to make the forecast in this case study. Two neural networks (NN) approaches were utilized and compared, namely long short-term memory (LSTM) and multilayer perceptron (MLP). The results as expected, showed that the NN models gave better results than traditional forecasting method (judgmental method). In addition, the LSTM model had a higher predictive accuracy than the MLP model.Keywords: neural network, LSTM, MLP, forecasting demand, inventory management
Procedia PDF Downloads 13021668 Circuit Models for Conducted Susceptibility Analyses of Multiconductor Shielded Cables
Authors: Saih Mohamed, Rouijaa Hicham, Ghammaz Abdelilah
Abstract:
This paper presents circuit models to analyze the conducted susceptibility of multiconductor shielded cables in frequency domains using Branin’s method, which is referred to as the method of characteristics. These models, Which can be used directly in the time and frequency domains, take into account the presence of both the transfer impedance and admittance. The conducted susceptibility is studied by using an injection current on the cable shield as the source. Two examples are studied, a coaxial shielded cable and shielded cables with two parallel wires (i.e., twinax cables). This shield has an asymmetry (one slot on the side). Results obtained by these models are in good agreement with those obtained by other methods.Keywords: circuit models, multiconductor shielded cables, Branin’s method, coaxial shielded cable, twinax cables
Procedia PDF Downloads 51921667 Meteorological Risk Assessment for Ships with Fuzzy Logic Designer
Authors: Ismail Karaca, Ridvan Saracoglu, Omer Soner
Abstract:
Fuzzy Logic, an advanced method to support decision-making, is used by various scientists in many disciplines. Fuzzy programming is a product of fuzzy logic, fuzzy rules, and implication. In marine science, fuzzy programming for ships is dramatically increasing together with autonomous ship studies. In this paper, a program to support the decision-making process for ship navigation has been designed. The program is produced in fuzzy logic and rules, by taking the marine accidents and expert opinions into account. After the program was designed, the program was tested by 46 ship accidents reported by the Transportation Safety Investigation Center of Turkey. Wind speed, sea condition, visibility, day/night ratio have been used as input data. They have been converted into a risk factor within the Fuzzy Logic Designer application and fuzzy rules set by marine experts. Finally, the expert's meteorological risk factor for each accident is compared with the program's risk factor, and the error rate was calculated. The main objective of this study is to improve the navigational safety of ships, by using the advance decision support model. According to the study result, fuzzy programming is a robust model that supports safe navigation.Keywords: calculation of risk factor, fuzzy logic, fuzzy programming for ship, safety navigation of ships
Procedia PDF Downloads 19221666 Upgraded Cuckoo Search Algorithm to Solve Optimisation Problems Using Gaussian Selection Operator and Neighbour Strategy Approach
Authors: Mukesh Kumar Shah, Tushar Gupta
Abstract:
An Upgraded Cuckoo Search Algorithm is proposed here to solve optimization problems based on the improvements made in the earlier versions of Cuckoo Search Algorithm. Short comings of the earlier versions like slow convergence, trap in local optima improved in the proposed version by random initialization of solution by suggesting an Improved Lambda Iteration Relaxation method, Random Gaussian Distribution Walk to improve local search and further proposing Greedy Selection to accelerate to optimized solution quickly and by “Study Nearby Strategy” to improve global search performance by avoiding trapping to local optima. It is further proposed to generate better solution by Crossover Operation. The proposed strategy used in algorithm shows superiority in terms of high convergence speed over several classical algorithms. Three standard algorithms were tested on a 6-generator standard test system and the results are presented which clearly demonstrate its superiority over other established algorithms. The algorithm is also capable of handling higher unit systems.Keywords: economic dispatch, gaussian selection operator, prohibited operating zones, ramp rate limits
Procedia PDF Downloads 13221665 A Computational Study of the Electron Transport in HgCdTe Bulk Semiconductor
Abstract:
This paper deals with the use of computational method based on Monte Carlo simulation in order to investigate the transport phenomena of the electron in HgCdTe narrow band gap semiconductor. Via this method we can evaluate the time dependence of the transport parameters: velocity, energy and mobility of electrons through matter (HgCdTe).Keywords: Monte Carlo, transport parameters, HgCdTe, computational mechanics
Procedia PDF Downloads 47721664 Cleaning of Scientific References in Large Patent Databases Using Rule-Based Scoring and Clustering
Authors: Emiel Caron
Abstract:
Patent databases contain patent related data, organized in a relational data model, and are used to produce various patent statistics. These databases store raw data about scientific references cited by patents. For example, Patstat holds references to tens of millions of scientific journal publications and conference proceedings. These references might be used to connect patent databases with bibliographic databases, e.g. to study to the relation between science, technology, and innovation in various domains. Problematic in such studies is the low data quality of the references, i.e. they are often ambiguous, unstructured, and incomplete. Moreover, a complete bibliographic reference is stored in only one attribute. Therefore, a computerized cleaning and disambiguation method for large patent databases is developed in this work. The method uses rule-based scoring and clustering. The rules are based on bibliographic metadata, retrieved from the raw data by regular expressions, and are transparent and adaptable. The rules in combination with string similarity measures are used to detect pairs of records that are potential duplicates. Due to the scoring, different rules can be combined, to join scientific references, i.e. the rules reinforce each other. The scores are based on expert knowledge and initial method evaluation. After the scoring, pairs of scientific references that are above a certain threshold, are clustered by means of single-linkage clustering algorithm to form connected components. The method is designed to disambiguate all the scientific references in the Patstat database. The performance evaluation of the clustering method, on a large golden set with highly cited papers, shows on average a 99% precision and a 95% recall. The method is therefore accurate but careful, i.e. it weighs precision over recall. Consequently, separate clusters of high precision are sometimes formed, when there is not enough evidence for connecting scientific references, e.g. in the case of missing year and journal information for a reference. The clusters produced by the method can be used to directly link the Patstat database with bibliographic databases as the Web of Science or Scopus.Keywords: clustering, data cleaning, data disambiguation, data mining, patent analysis, scientometrics
Procedia PDF Downloads 19521663 Analysis of Hard Turning Process of AISI D3-Thermal Aspects
Authors: B. Varaprasad, C. Srinivasa Rao
Abstract:
In the manufacturing sector, hard turning has emerged as vital machining process for cutting hardened steels. Besides many advantages of hard turning operation, one has to implement to achieve close tolerances in terms of surface finish, high product quality, reduced machining time, low operating cost and environmentally friendly characteristics. In the present study, three-dimensional CAE (Computer Aided Engineering) based simulation of hard turning by using commercial software DEFORM 3D has been compared to experimental results of stresses, temperatures and tool forces in machining of AISI D3 steel using mixed Ceramic inserts (CC6050). In the present analysis, orthogonal cutting models are proposed, considering several processing parameters such as cutting speed, feed, and depth of cut. An exhaustive friction modeling at the tool-work interfaces is carried out. Work material flow around the cutting edge is carefully modeled with adaptive re-meshing simulation capability. In process simulations, feed rate and cutting speed are constant (i.e.,. 0.075 mm/rev and 155 m/min), and analysis is focused on stresses, forces, and temperatures during machining. Close agreement is observed between CAE simulation and experimental values.Keywords: hard turning, computer aided engineering, computational machining, finite element method
Procedia PDF Downloads 45821662 Predicting the Exposure Level of Airborne Contaminants in Occupational Settings via the Well-Mixed Room Model
Authors: Alireza Fallahfard, Ludwig Vinches, Stephane Halle
Abstract:
In the workplace, the exposure level of airborne contaminants should be evaluated due to health and safety issues. It can be done by numerical models or experimental measurements, but the numerical approach can be useful when it is challenging to perform experiments. One of the simplest models is the well-mixed room (WMR) model, which has shown its usefulness to predict inhalation exposure in many situations. However, since the WMR is limited to gases and vapors, it cannot be used to predict exposure to aerosols. The main objective is to modify the WMR model to expand its application to exposure scenarios involving aerosols. To reach this objective, the standard WMR model has been modified to consider the deposition of particles by gravitational settling and Brownian and turbulent deposition. Three deposition models were implemented in the model. The time-dependent concentrations of airborne particles predicted by the model were compared to experimental results conducted in a 0.512 m3 chamber. Polystyrene particles of 1, 2, and 3 µm in aerodynamic diameter were generated with a nebulizer under two air changes per hour (ACH). The well-mixed condition and chamber ACH were determined by the tracer gas decay method. The mean friction velocity on the chamber surfaces as one of the input variables for the deposition models was determined by computational fluid dynamics (CFD) simulation. For the experimental procedure, the particles were generated until reaching the steady-state condition (emission period). Then generation stopped, and concentration measurements continued until reaching the background concentration (decay period). The results of the tracer gas decay tests revealed that the ACHs of the chamber were: 1.4 and 3.0, and the well-mixed condition was achieved. The CFD results showed the average mean friction velocity and their standard deviations for the lowest and highest ACH were (8.87 ± 0.36) ×10-2 m/s and (8.88 ± 0.38) ×10-2 m/s, respectively. The numerical results indicated the difference between the predicted deposition rates by the three deposition models was less than 2%. The experimental and numerical aerosol concentrations were compared in the emission period and decay period. In both periods, the prediction accuracy of the modified model improved in comparison with the classic WMR model. However, there is still a difference between the actual value and the predicted value. In the emission period, the modified WMR results closely follow the experimental data. However, the model significantly overestimates the experimental results during the decay period. This finding is mainly due to an underestimation of the deposition rate in the model and uncertainty related to measurement devices and particle size distribution. Comparing the experimental and numerical deposition rates revealed that the actual particle deposition rate is significant, but the deposition mechanisms considered in the model were ten times lower than the experimental value. Thus, particle deposition was significant and will affect the airborne concentration in occupational settings, and it should be considered in the airborne exposure prediction model. The role of other removal mechanisms should be investigated.Keywords: aerosol, CFD, exposure assessment, occupational settings, well-mixed room model, zonal model
Procedia PDF Downloads 10421661 Medical Image Watermark and Tamper Detection Using Constant Correlation Spread Spectrum Watermarking
Authors: Peter U. Eze, P. Udaya, Robin J. Evans
Abstract:
Data hiding can be achieved by Steganography or invisible digital watermarking. For digital watermarking, both accurate retrieval of the embedded watermark and the integrity of the cover image are important. Medical image security in Teleradiology is one of the applications where the embedded patient record needs to be extracted with accuracy as well as the medical image integrity verified. In this research paper, the Constant Correlation Spread Spectrum digital watermarking for medical image tamper detection and accurate embedded watermark retrieval is introduced. In the proposed method, a watermark bit from a patient record is spread in a medical image sub-block such that the correlation of all watermarked sub-blocks with a spreading code, W, would have a constant value, p. The constant correlation p, spreading code, W and the size of the sub-blocks constitute the secret key. Tamper detection is achieved by flagging any sub-block whose correlation value deviates by more than a small value, ℇ, from p. The major features of our new scheme include: (1) Improving watermark detection accuracy for high-pixel depth medical images by reducing the Bit Error Rate (BER) to Zero and (2) block-level tamper detection in a single computational process with simultaneous watermark detection, thereby increasing utility with the same computational cost.Keywords: Constant Correlation, Medical Image, Spread Spectrum, Tamper Detection, Watermarking
Procedia PDF Downloads 19921660 Finite Element Molecular Modeling: A Structural Method for Large Deformations
Authors: A. Rezaei, M. Huisman, W. Van Paepegem
Abstract:
Atomic interactions in molecular systems are mainly studied by particle mechanics. Nevertheless, researches have also put on considerable effort to simulate them using continuum methods. In early 2000, simple equivalent finite element models have been developed to study the mechanical properties of carbon nanotubes and graphene in composite materials. Afterward, many researchers have employed similar structural simulation approaches to obtain mechanical properties of nanostructured materials, to simplify interface behavior of fiber-reinforced composites, and to simulate defects in carbon nanotubes or graphene sheets, etc. These structural approaches, however, are limited to small deformations due to complicated local rotational coordinates. This article proposes a method for the finite element simulation of molecular mechanics. For ease in addressing the approach, here it is called Structural Finite Element Molecular Modeling (SFEMM). SFEMM method improves the available structural approaches for large deformations, without using any rotational degrees of freedom. Moreover, the method simulates molecular conformation, which is a big advantage over the previous approaches. Technically, this method uses nonlinear multipoint constraints to simulate kinematics of the atomic multibody interactions. Only truss elements are employed, and the bond potentials are implemented through constitutive material models. Because the equilibrium bond- length, bond angles, and bond-torsion potential energies are intrinsic material parameters, the model is independent of initial strains or stresses. In this paper, the SFEMM method has been implemented in ABAQUS finite element software. The constraints and material behaviors are modeled through two Fortran subroutines. The method is verified for the bond-stretch, bond-angle and bond-torsion of carbon atoms. Furthermore, the capability of the method in the conformation simulation of molecular structures is demonstrated via a case study of a graphene sheet. Briefly, SFEMM builds up a framework that offers more flexible features over the conventional molecular finite element models, serving the structural relaxation modeling and large deformations without incorporating local rotational degrees of freedom. Potentially, the method is a big step towards comprehensive molecular modeling with finite element technique, and thereby concurrently coupling an atomistic domain to a solid continuum domain within a single finite element platform.Keywords: finite element, large deformation, molecular mechanics, structural method
Procedia PDF Downloads 15421659 Preparation of Chromium Nanoparticles on Carbon Substrate from Tannery Waste Solution by Chemical Method Compared to Electrokinetic Process
Authors: Mahmoud A. Rabah, Said El Sheikh
Abstract:
This work shows the preparation of chromium nanoparticles from tannery waste solution on glassy carbon by chemical method compared to electrokinetic process. The waste solution contains free and soluble fats, calcium, iron, magnesium and high sodium in addition to the chromium ions. Filtration helps removal of insoluble matters. Diethyl ether successfully extracted soluble fats. The method started by removing calcium as insoluble oxalate salts at hot conditions in a faint acidic medium. The filtrate contains iron, magnesium, chromium ions and sodium chloride in excess. Chromium was separated selectively as insoluble hydroxide sol-gel at pH 6.5, filtered and washed with distilled water. Part of the gel reacted with sulfuric acid to produce chromium sulfate solution having 15-25 g/L concentration. Electrokinetic deposition of chromium nanoparticles on a carbon cathode was carried out using platinum anode under different galvanostatic conditions. The chemical method involved impregnating the carbon specimens with chromium hydroxide gel followed by reduction using hydrazine hydrate or by thermal reduction using hydrogen gas at 1250°C. Chromium grain size was characterized by TEM, FT-IR and SEM. Properties of the Cr grains were correlated to the conditions of the preparation process. Electrodeposition was found to control chromium particles to be more identical in size and shape as compared to the chemical method.Keywords: chromium, electrodeposition, nanoparticles, tannery waste solution
Procedia PDF Downloads 41121658 District Selection for Geotechnical Settlement Suitability Using GIS and Multi Criteria Decision Analysis: A Case Study in Denizli, Turkey
Authors: Erdal Akyol, Mutlu Alkan
Abstract:
Multi criteria decision analysis (MDCA) covers both data and experience. It is very common to solve the problems with many parameters and uncertainties. GIS supported solutions improve and speed up the decision process. Weighted grading as a MDCA method is employed for solving the geotechnical problems. In this study, geotechnical parameters namely soil type; SPT (N) blow number, shear wave velocity (Vs) and depth of underground water level (DUWL) have been engaged in MDCA and GIS. In terms of geotechnical aspects, the settlement suitability of the municipal area was analyzed by the method. MDCA results were compatible with the geotechnical observations and experience. The method can be employed in geotechnical oriented microzoning studies if the criteria are well evaluated.Keywords: GIS, spatial analysis, multi criteria decision analysis, geotechnics
Procedia PDF Downloads 46321657 Real-Time Measurement Approach for Tracking the ΔV10 Estimate Value of DC EAF
Authors: Jin-Lung Guan, Jyh-Cherng Gu, Chun-Wei Huang, Hsin-Hung Chang
Abstract:
This investigation develops a revisable method for estimating the estimate value of equivalent 10 Hz voltage flicker (DV10) of a DC Electric Arc Furnace (EAF). This study also discusses three 161kV DC EAFs by field measurement, with those results indicating that the estimated DV10 value is significantly smaller than the survey value. The key point is that the conventional means of estimating DV10 is inappropriate. There is a main cause as the assumed Qmax is too small. Although DC EAF is regularly operated in a constant MVA mode, the reactive power variation in the Main Transformer (MT) is more significant than that in the Furnace Transformer (FT). A substantial difference exists between estimated maximum reactive power fluctuation (DQmax) and the survey value from actual DC EAF operations. However, this study proposes a revisable method that can obtain a more accurate DV10 estimate than the conventional method.Keywords: voltage flicker, dc EAF, estimate value, DV10
Procedia PDF Downloads 45221656 Investigation of Flexural – Torsion Instability of Struts Using Modified Newmark Method
Authors: Seyed Amin Vakili, Sahar Sadat Vakili, Seyed Ehsan Vakili, Nader Abdoli Yazdi
Abstract:
Differential equations are of fundamental importance in engineering and applied mathematics, since many physical laws and relations appear mathematically in the form of such equations. The equilibrium state of structures consisting of one-dimensional elements can be described by an ordinary differential equation. The response of these kinds of structures under the loading, namely relationship between the displacement field and loading field, can be predicted by the solution of these differential equations and on satisfying the given boundary conditions. When the effect of change of geometry under loading is taken into account in modeling of equilibrium state, then these differential equations are partially integrable in quartered. They also exhibit instability characteristics when the structures are loaded compressively. The purpose of this paper is to represent the ability of the Modified Newmark Method in analyzing flexural-torsional instability of struts for both bifurcation and non-bifurcation structural systems. The results are shown to be very accurate with only a small number of iterations. The method is easily programmed, and has the advantages of simplicity and speeds of convergence and easily is extended to treat material and geometric nonlinearity including no prismatic members and linear and nonlinear spring restraints that would be encountered in frames. In this paper, these abilities of the method will be extended to the system of linear differential equations that govern strut flexural torsional stability.Keywords: instability, torsion, flexural, buckling, modified newmark method stability
Procedia PDF Downloads 36221655 Orthopedic Trauma in Newborn Babies
Authors: Joanna Maj, Awais Hussain, Lyndsey Vu, Catherine Roxas
Abstract:
Background: Bone injuries in babies are common conditions that arise during delivery. Fractures of the clavicle, humerus, femur, and skull are the most common neonatal bone injuries sustained from labor and delivery. During operative deliveries, zealous tractions, ineffective delivery techniques, improper uterine incision, and inadequate relaxation of the uterus can lead to bone fractures in the newborn. Neonatal anatomy is unique. Just as children are not mini-adults, newborns are not mini children. A newborn’s anatomy and physiology are significantly different from a pediatric patient's. In this paper, we describe common orthopedic trauma in newborn babies. We provide a comprehensive overview of the different types of bone injuries in newborns. We hypothesize that the rate of bone fractures sustained at birth is higher in cases of operative deliveries. Methods: Relevant literature was selected by using the PubMed database. Search terms included orthopedic conditions in newborns, neonatal anatomy, and bone fractures in neonates during operative deliveries. Inclusion criteria included age, gender, race, type of bone injury and progression of bone injury. Exclusion criteria were limited in the medical history of cases reviewed and comorbidities. Results: This review finds that a clavicle fracture is the most common type of neonatal orthopedic injury sustained at birth in both operative and non-operative deliveries. We confirm the hypothesis that infants born via operative deliveries have a significantly higher rate of bone fractures than non-cesarean section deliveries. Conclusion: Newborn babies born via operative deliveries have a higher rate of bone fractures of the clavicle, humerus, and femur. A clavicle bone fracture in newborns is most common during emergency operative deliveries in new mothers. We conclude that infants born via an operative delivery sustained more bone injuries than infants born via non-cesarean section deliveries.Keywords: clavicle fracture, humerus fracture, neonates, newborn orthopedics, orthopedic surgery, pediatrics, orthopedic trauma, orthopedic trauma during delivery, cesarean section, obstetrics, neonatal anatomy, neonatal fractures, operative deliveries, labor and delivery, bone injuries in neonates
Procedia PDF Downloads 10521654 High-Resolution ECG Automated Analysis and Diagnosis
Authors: Ayad Dalloo, Sulaf Dalloo
Abstract:
Electrocardiogram (ECG) recording is prone to complications, on analysis by physicians, due to noise and artifacts, thus creating ambiguity leading to possible error of diagnosis. Such drawbacks may be overcome with the advent of high resolution Methods, such as Discrete Wavelet Analysis and Digital Signal Processing (DSP) techniques. This ECG signal analysis is implemented in three stages: ECG preprocessing, features extraction and classification with the aim of realizing high resolution ECG diagnosis and improved detection of abnormal conditions in the heart. The preprocessing stage involves removing spurious artifacts (noise), due to such factors as muscle contraction, motion, respiration, etc. ECG features are extracted by applying DSP and suggested sloping method techniques. These measured features represent peak amplitude values and intervals of P, Q, R, S, R’, and T waves on ECG, and other features such as ST elevation, QRS width, heart rate, electrical axis, QR and QT intervals. The classification is preformed using these extracted features and the criteria for cardiovascular diseases. The ECG diagnostic system is successfully applied to 12-lead ECG recordings for 12 cases. The system is provided with information to enable it diagnoses 15 different diseases. Physician’s and computer’s diagnoses are compared with 90% agreement, with respect to physician diagnosis, and the time taken for diagnosis is 2 seconds. All of these operations are programmed in Matlab environment.Keywords: ECG diagnostic system, QRS detection, ECG baseline removal, cardiovascular diseases
Procedia PDF Downloads 29821653 Keyframe Extraction Using Face Quality Assessment and Convolution Neural Network
Authors: Rahma Abed, Sahbi Bahroun, Ezzeddine Zagrouba
Abstract:
Due to the huge amount of data in videos, extracting the relevant frames became a necessity and an essential step prior to performing face recognition. In this context, we propose a method for extracting keyframes from videos based on face quality and deep learning for a face recognition task. This method has two steps. We start by generating face quality scores for each face image based on the use of three face feature extractors, including Gabor, LBP, and HOG. The second step consists in training a Deep Convolutional Neural Network in a supervised manner in order to select the frames that have the best face quality. The obtained results show the effectiveness of the proposed method compared to the methods of the state of the art.Keywords: keyframe extraction, face quality assessment, face in video recognition, convolution neural network
Procedia PDF Downloads 23821652 Seed Priming Treatments in Common Zinnia (Zinnia elegans) Using Some Plant Extracts
Authors: Atakan Efe Akpınar, Zeynep Demir
Abstract:
Seed priming technologies are frequently used nowadays to increase the germination potential and stress tolerance of seeds. These treatments might be beneficial for native species as well as crops. Different priming treatments can be used depending on the type of plant, the morphology, and the physiology of the seed. Moreover, these may be various physical, chemical, and/or biological treatments. Aiming to improve studies about seed priming, ideas need to be brought into this technological sector related to the agri-seed industry. This study addresses the question of whether seed priming with plant extracts can improve seed vigour and germination performance. By investigating the effects of plant extract priming on various vigour parameters, the research aims to provide insights into the potential benefits of this treatment method. Thus, seed priming was carried out using some plant extracts. Firstly, some plant extracts prepared from plant leaves, roots, or fruit parts were obtained for use in priming treatments. Then, seeds of Common zinnia (Zinnia elegans) were kept in solutions containing plant extracts at 20°C for 48 hours. Seeds without any treatment were evaluated as the control group. At the end of priming applications, seeds are dried superficially at 25°C. Seeds of Common zinnia (Zinnia elegans) were analyzed for vigour (normal germination rate, germination time, germination index etc.). In the future, seed priming applications can expand to multidisciplinary research combining with digital, bioinformatic and molecular tools.Keywords: seed priming, plant extracts, germination, biology
Procedia PDF Downloads 7721651 Effects of Roughness Elements on Heat Transfer During Natural Convection
Abstract:
The present study focused on the investigation of the effects of roughness elements on heat transfer during natural convection in a rectangular cavity using a numerical technique. Roughness elements were introduced on the bottom hot wall with a normalized amplitude (A*/H) of 0.1. Thermal and hydrodynamic behavior was studied using a computational method based on Lattice Boltzmann method (LBM). Numerical studies were performed for a laminar natural convection in the range of Rayleigh number (Ra) from 103 to 106 for a rectangular cavity of aspect ratio (L/H) 2 with a fluid of Prandtl number (Pr) 1.0. The presence of the sinusoidal roughness elements caused a minimum to the maximum decrease in the heat transfer as 7% to 17% respectively compared to the smooth enclosure. The results are presented for mean Nusselt number (Nu), isotherms, and streamlines.Keywords: natural convection, Rayleigh number, surface roughness, Nusselt number, Lattice Boltzmann method
Procedia PDF Downloads 542