Search results for: hazard parameters
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9069

Search results for: hazard parameters

8949 Machine Learning Techniques in Seismic Risk Assessment of Structures

Authors: Farid Khosravikia, Patricia Clayton

Abstract:

The main objective of this work is to evaluate the advantages and disadvantages of various machine learning techniques in two key steps of seismic hazard and risk assessment of different types of structures. The first step is the development of ground-motion models, which are used for forecasting ground-motion intensity measures (IM) given source characteristics, source-to-site distance, and local site condition for future events. IMs such as peak ground acceleration and velocity (PGA and PGV, respectively) as well as 5% damped elastic pseudospectral accelerations at different periods (PSA), are indicators of the strength of shaking at the ground surface. Typically, linear regression-based models, with pre-defined equations and coefficients, are used in ground motion prediction. However, due to the restrictions of the linear regression methods, such models may not capture more complex nonlinear behaviors that exist in the data. Thus, this study comparatively investigates potential benefits from employing other machine learning techniques as statistical method in ground motion prediction such as Artificial Neural Network, Random Forest, and Support Vector Machine. The results indicate the algorithms satisfy some physically sound characteristics such as magnitude scaling distance dependency without requiring pre-defined equations or coefficients. Moreover, it is shown that, when sufficient data is available, all the alternative algorithms tend to provide more accurate estimates compared to the conventional linear regression-based method, and particularly, Random Forest outperforms the other algorithms. However, the conventional method is a better tool when limited data is available. Second, it is investigated how machine learning techniques could be beneficial for developing probabilistic seismic demand models (PSDMs), which provide the relationship between the structural demand responses (e.g., component deformations, accelerations, internal forces, etc.) and the ground motion IMs. In the risk framework, such models are used to develop fragility curves estimating exceeding probability of damage for pre-defined limit states, and therefore, control the reliability of the predictions in the risk assessment. In this study, machine learning algorithms like artificial neural network, random forest, and support vector machine are adopted and trained on the demand parameters to derive PSDMs for them. It is observed that such models can provide more accurate estimates of prediction in relatively shorter about of time compared to conventional methods. Moreover, they can be used for sensitivity analysis of fragility curves with respect to many modeling parameters without necessarily requiring more intense numerical response-history analysis.

Keywords: artificial neural network, machine learning, random forest, seismic risk analysis, seismic hazard analysis, support vector machine

Procedia PDF Downloads 84
8948 Seismic Vulnerability Mitigation of Non-Engineered Buildings

Authors: Muhammad Tariq A. Chaudhary

Abstract:

The tremendous loss of life that resulted in the aftermath of recent earthquakes in developing countries is mostly due to the collapse of non-engineered and semi-engineered building structures. Such structures are used as houses, schools, primary healthcare centres and government offices. These building are classified structurally into two categories viz. non-engineered and semi-engineered. Non-engineered structures include: adobe, Unreinforced Masonry (URM) and wood buildings. Semi-engineered buildings are mostly low-rise (up to 3 story) light concrete frame structures or masonry bearing walls with reinforced concrete slab. This paper presents an overview of the typical damage observed in non-engineered structures and their most likely causes in the past earthquakes with specific emphasis on the performance of such structures in the 2005 Kashmir earthquake. It is demonstrated that seismic performance of these structures can be improved from life-safety viewpoint by adopting simple low-cost modifications to the existing construction practices. Incorporation of some of these practices in the reconstruction efforts after the 2005 Kashmir earthquake are examined in the last section for mitigating seismic risk hazard.

Keywords: Kashmir earthquake, non-engineered buildings, seismic hazard, structural details, structural strengthening

Procedia PDF Downloads 263
8947 How to Reach Net Zero Emissions? On the Permissibility of Negative Emission Technologies and the Danger of Moral Hazards

Authors: Hanna Schübel, Ivo Wallimann-Helmer

Abstract:

In order to reach the goal of the Paris Agreement to not overshoot 1.5°C of warming above pre-industrial levels, various countries including the UK and Switzerland have committed themselves to net zero emissions by 2050. The employment of negative emission technologies (NETs) is very likely going to be necessary for meeting these national objectives as well as other internationally agreed climate targets. NETs are methods of removing carbon from the atmosphere and are thus a means for addressing climate change. They range from afforestation to technological measures such as direct air capture and carbon storage (DACCS), where CO2 is captured from the air and stored underground. As all so-called geoengineering technologies, the development and deployment of NETs are often subject to moral hazard arguments. As these technologies could be perceived as an alternative to mitigation efforts, so the argument goes, they are potentially a dangerous distraction from the main target of mitigating emissions. We think that this is a dangerous argument to make as it may hinder the development of NETs which are an essential element of net zero emission targets. In this paper we argue that the moral hazard argument is only problematic if we do not reflect upon which levels of emissions are at stake in order to meet net zero emissions. In response to the moral hazard argument we develop an account of which levels of emissions in given societies should be mitigated and not be the target of NETs and which levels of emissions can legitimately be a target of NETs. For this purpose, we define four different levels of emissions: the current level of individual emissions, the level individuals emit in order to appear in public without shame, the level of a fair share of individual emissions in the global budget, and finally the baseline of net zero emissions. At each level of emissions there are different subjects to be assigned responsibilities if societies and/or individuals are committed to the target of net zero emissions. We argue that all emissions within one’s fair share do not demand individual mitigation efforts. The same holds with regard to individuals and the baseline level of emissions necessary to appear in public in their societies without shame. Individuals are only under duty to reduce their emissions if they exceed this baseline level. This is different for whole societies. Societies demanding more emissions to appear in public without shame than the individual fair share are under duty to foster emission reductions and are not legitimate to reduce by introducing NETs. NETs are legitimate for reducing emissions only below the level of fair shares and for reaching net zero emissions. Since access to NETs to achieve net zero emissions demands technology not affordable to individuals there are also no full individual responsibilities to achieve net zero emissions. This is mainly a responsibility of societies as a whole.

Keywords: climate change, mitigation, moral hazard, negative emission technologies, responsibility

Procedia PDF Downloads 96
8946 On the Development of a Homogenized Earthquake Catalogue for Northern Algeria

Authors: I. Grigoratos, R. Monteiro

Abstract:

Regions with a significant percentage of non-seismically designed buildings and reduced urban planning are particularly vulnerable to natural hazards. In this context, the project ‘Improved Tools for Disaster Risk Mitigation in Algeria’ (ITERATE) aims at seismic risk mitigation in Algeria. Past earthquakes in North Algeria caused extensive damages, e.g. the El Asnam 1980 moment magnitude (Mw) 7.1 and Boumerdes 2003 Mw 6.8 earthquakes. This paper will address a number of proposed developments and considerations made towards a further improvement of the component of seismic hazard. In specific, an updated earthquake catalog (until year 2018) is compiled, and new conversion equations to moment magnitude are introduced. Furthermore, a network-based method for the estimation of the spatial and temporal distribution of the minimum magnitude of completeness is applied. We found relatively large values for Mc, due to the sparse network, and a nonlinear trend between Mw and body wave (mb) or local magnitude (ML), which are the most common scales reported in the region. Lastly, the resulting b-value of the Gutenberg-Richter distribution is sensitive to the declustering method.

Keywords: conversion equation, magnitude of completeness, seismic events, seismic hazard

Procedia PDF Downloads 145
8945 Analysis of Exponential Nonuniform Transmission Line Parameters

Authors: Mounir Belattar

Abstract:

In this paper the Analysis of voltage waves that propagate along a lossless exponential nonuniform line is presented. For this analysis the parameters of this line are assumed to be varying function of the distance x along the line from the source end. The approach is based on the tow-port networks cascading presentation to derive the ABDC parameters of transmission using Picard-Carson Method which is a powerful method in getting a power series solution for distributed network because it is easy to calculate poles and zeros and solves differential equations such as telegrapher equations by an iterative sequence. So the impedance, admittance voltage and current along the line are expanded as a Taylor series in x/l where l is the total length of the line to obtain at the end, the main transmission line parameters such as voltage response and transmission and reflexion coefficients represented by scattering parameters in frequency domain.

Keywords: ABCD parameters, characteristic impedance exponential nonuniform transmission line, Picard-Carson's method, S parameters, Taylor's series

Procedia PDF Downloads 416
8944 Safety of Ports, Harbours, Marine Terminals: Application of Quantitative Risk Assessment

Authors: Dipak Sonawane, Sudarshan Daga, Somesh Gupta

Abstract:

Quantitative risk assessment (QRA) is a very precise and consistent approach to defining the likelihood, consequence and severity of a major incident/accident. A variety of hazardous cargoes in bulk, such as hydrocarbons and flammable/toxic chemicals, are handled at various ports. It is well known that most of the operations are hazardous, having the potential of damaging property, causing injury/loss of life and, in some cases, the threat of environmental damage. In order to ensure adequate safety towards life, environment and property, the application of scientific methods such as QRA is inevitable. By means of these methods, comprehensive hazard identification, risk assessment and appropriate implementation of Risk Control measures can be carried out. In this paper, the authors, based on their extensive experience in Risk Analysis for ports and harbors, have exhibited how QRA can be used in practice to minimize and contain risk to tolerable levels. A specific case involving the operation for unloading of hydrocarbon at a port is presented. The exercise provides confidence that the method of QRA, as proposed by the authors, can be used appropriately for the identification of hazards and risk assessment of Ports and Terminals.

Keywords: quantitative risk assessment, hazard assessment, consequence analysis, individual risk, societal risk

Procedia PDF Downloads 62
8943 Seismic Response Mitigation of Structures Using Base Isolation System Considering Uncertain Parameters

Authors: Rama Debbarma

Abstract:

The present study deals with the performance of Linear base isolation system to mitigate seismic response of structures characterized by random system parameters. This involves optimization of the tuning ratio and damping properties of the base isolation system considering uncertain system parameters. However, the efficiency of base isolator may reduce if it is not tuned to the vibrating mode it is designed to suppress due to unavoidable presence of system parameters uncertainty. With the aid of matrix perturbation theory and first order Taylor series expansion, the total probability concept is used to evaluate the unconditional response of the primary structures considering random system parameters. For this, the conditional second order information of the response quantities are obtained in random vibration framework using state space formulation. Subsequently, the maximum unconditional root mean square displacement of the primary structures is used as the objective function to obtain optimum damping parameters Numerical study is performed to elucidate the effect of parameters uncertainties on the optimization of parameters of linear base isolator and system performance.

Keywords: linear base isolator, earthquake, optimization, uncertain parameters

Procedia PDF Downloads 401
8942 Experimental Determination of Aluminum 7075-T6 Parameters Using Stabilized Cycle Tests to Predict Thermal Ratcheting

Authors: Armin Rahmatfam, Mohammad Zehsaz, Farid Vakili Tahami, Nasser Ghassembaglou

Abstract:

In this paper the thermal ratcheting, kinematic hardening parameters C, γ, isotropic hardening parameters and also k, b, Q combined isotropic/kinematic hardening parameters have been obtained experimentally from the monotonic, strain controlled cyclic tests at room and elevated temperatures of 20°C, 100°C, and 400°C. These parameters are used in nonlinear combined isotropic/kinematic hardening model to predict better description of the loading and reloading cycles in the cyclic indentation as well as thermal ratcheting. For this purpose, three groups of specimens made of Aluminum 7075-T6 have been investigated. After each test and using stable hysteretic cycles, material parameters have been obtained for using in combined nonlinear isotropic/kinematic hardening models. Also the methodology of obtaining the correct kinematic/isotropic hardening parameters is presented.

Keywords: combined hardening model, kinematic hardening, isotropic hardening, cyclic tests

Procedia PDF Downloads 451
8941 An Experimental Analysis of Squeeze Casting Parameters for 2017 a Wrought Al Alloy

Authors: Mohamed Ben Amar, Najib Souissi, Chedly Bradai

Abstract:

A Taguchi design investigation has been made into the relationship between the ductility and process variables in a squeeze cast 2017A wrought aluminium alloy. The considered process parameters were: squeeze pressure, melt temperature and die preheating temperature. An orthogonal array (OA), main effect, signal-to-noise (S/N) ratio, and the analysis of variance (ANOVA) are employed to analyze the effect of casting parameters. The results have shown that the selected parameters significantly affect the ductility of 2017A wrought Al alloy castings. Optimal squeeze cast process parameters were provided to illustrate the proposed approach and the results were proven to be trustworthy through practical experiments.

Keywords: Taguchi method, squeeze casting, process parameters, ductility, microstructure

Procedia PDF Downloads 370
8940 Seismic Resistant Columns of Buildings against the Differential Settlement of the Foundation

Authors: Romaric Desbrousses, Lan Lin

Abstract:

The objective of this study is to determine how Canadian seismic design provisions affect the column axial load resistance of moment-resisting frame reinforced concrete buildings subjected to the differential settlement of their foundation. To do so, two four-storey buildings are designed in accordance with the seismic design provisions of the Canadian Concrete Design Standards. One building is located in Toronto, which is situated in a moderate seismic hazard zone in Canada, and the other in Vancouver, which is in Canada’s highest seismic hazard zone. A finite element model of each building is developed using SAP 2000. A 100 mm settlement is assigned to the base of the building’s center column. The axial load resistance of the column is represented by the demand capacity ratio. The analysis results show that settlement-induced tensile axial forces have a particularly detrimental effect on the conventional settling columns of the Toronto buildings which fail at a much smaller settlement that those in the Vancouver buildings. The results also demonstrate that particular care should be taken in the design of columns in short-span buildings.

Keywords: Columns, Demand, Foundation differential settlement, Seismic design, Non-linear analysis

Procedia PDF Downloads 110
8939 Number of Necessary Parameters for Parametrization of Stabilizing Controllers for two times two RHinf Systems

Authors: Kazuyoshi Mori

Abstract:

In this paper, we consider the number of parameters for the parametrization of stabilizing controllers for RHinf systems with size 2 × 2. Fortunately, any plant of this model can admit doubly coprime factorization. Thus we can use the Youla parameterization to parametrize the stabilizing contollers . However, Youla parameterization does not give itself the minimal number of parameters. This paper shows that the minimal number of parameters is four. As a result, we show that the Youla parametrization naturally gives the parameterization of stabilizing controllers with minimal numbers.

Keywords: RHinfo, parameterization, number of parameters, multi-input, multi-output systems

Procedia PDF Downloads 390
8938 Intelligent Production Machine

Authors: A. Şahinoğlu, R. Gürbüz, A. Güllü, M. Karhan

Abstract:

This study in production machines, it is aimed that machine will automatically perceive cutting data and alter cutting parameters. The two most important parameters have to be checked in machine control unit are progress feed rate and speeds. These parameters are aimed to be controlled by sounds of machine. Optimum sound’s features introduced to computer. During process, real time data is received and converted by Matlab software. Data is converted into numerical values. According to them progress and speeds decreases/increases at a certain rate and thus optimum sound is acquired. Cutting process is made in respect of optimum cutting parameters. During chip remove progress, features of cutting tools, kind of cut material, cutting parameters and used machine; affects on various parameters. Instead of required parameters need to be measured such as temperature, vibration, and tool wear that emerged during cutting process; detailed analysis of the sound emerged during cutting process will provide detection of various data that included in the cutting process by the much more easy and economic way. The relation between cutting parameters and sound is being identified.

Keywords: cutting process, sound processing, intelligent late, sound analysis

Procedia PDF Downloads 309
8937 Landslide Hazard Zonation Using Satellite Remote Sensing and GIS Technology

Authors: Ankit Tyagi, Reet Kamal Tiwari, Naveen James

Abstract:

Landslide is the major geo-environmental problem of Himalaya because of high ridges, steep slopes, deep valleys, and complex system of streams. They are mainly triggered by rainfall and earthquake and causing severe damage to life and property. In Uttarakhand, the Tehri reservoir rim area, which is situated in the lesser Himalaya of Garhwal hills, was selected for landslide hazard zonation (LHZ). The study utilized different types of data, including geological maps, topographic maps from the survey of India, Landsat 8, and Cartosat DEM data. This paper presents the use of a weighted overlay method in LHZ using fourteen causative factors. The various data layers generated and co-registered were slope, aspect, relative relief, soil cover, intensity of rainfall, seismic ground shaking, seismic amplification at surface level, lithology, land use/land cover (LULC), normalized difference vegetation index (NDVI), topographic wetness index (TWI), stream power index (SPI), drainage buffer and reservoir buffer. Seismic analysis is performed using peak horizontal acceleration (PHA) intensity and amplification factors in the evaluation of the landslide hazard index (LHI). Several digital image processing techniques such as topographic correction, NDVI, and supervised classification were widely used in the process of terrain factor extraction. Lithological features, LULC, drainage pattern, lineaments, and structural features are extracted using digital image processing techniques. Colour, tones, topography, and stream drainage pattern from the imageries are used to analyse geological features. Slope map, aspect map, relative relief are created by using Cartosat DEM data. DEM data is also used for the detailed drainage analysis, which includes TWI, SPI, drainage buffer, and reservoir buffer. In the weighted overlay method, the comparative importance of several causative factors obtained from experience. In this method, after multiplying the influence factor with the corresponding rating of a particular class, it is reclassified, and the LHZ map is prepared. Further, based on the land-use map developed from remote sensing images, a landslide vulnerability study for the study area is carried out and presented in this paper.

Keywords: weighted overlay method, GIS, landslide hazard zonation, remote sensing

Procedia PDF Downloads 107
8936 Sparse Modelling of Cancer Patients’ Survival Based on Genomic Copy Number Alterations

Authors: Khaled M. Alqahtani

Abstract:

Copy number alterations (CNA) are variations in the structure of the genome, where certain regions deviate from the typical two chromosomal copies. These alterations are pivotal in understanding tumor progression and are indicative of patients' survival outcomes. However, effectively modeling patients' survival based on their genomic CNA profiles while identifying relevant genomic regions remains a statistical challenge. Various methods, such as the Cox proportional hazard (PH) model with ridge, lasso, or elastic net penalties, have been proposed but often overlook the inherent dependencies between genomic regions, leading to results that are hard to interpret. In this study, we enhance the elastic net penalty by incorporating an additional penalty that accounts for these dependencies. This approach yields smooth parameter estimates and facilitates variable selection, resulting in a sparse solution. Our findings demonstrate that this method outperforms other models in predicting survival outcomes, as evidenced by our simulation study. Moreover, it allows for a more meaningful interpretation of genomic regions associated with patients' survival. We demonstrate the efficacy of our approach using both real data from a lung cancer cohort and simulated datasets.

Keywords: copy number alterations, cox proportional hazard, lung cancer, regression, sparse solution

Procedia PDF Downloads 21
8935 Contact Address Levels and Human Health Risk of Metals In Milk and Milk Products Bought from Abeokuta, Southwestern Nigeria

Authors: Olukayode Bamgbose, Feyisola Agboola, Adewale M. Taiwo, Olanrewaju Olujimi Oluwole Terebo, Azeez Soyingbe, Akeem Bamgbade

Abstract:

The present study evaluated the contents and health risk assessment of metals determined in milk and milk product samples collected from the Abeokuta market. Forty-five milk and milk product (yoghurt) samples were digested and analysed for selected metals using Atomic Absorption Spectrophotometric method. Health risk assessment was evaluated for hazard quotient (HQ), hazard index (HI), and cancer risk (CR). Data were subjected to descriptive and inferential statistics. The concentrations of Zn, which ranged from 3.24±0.59 to 4.35±0.59 mg/kg, were the highest in the samples. Cr and Cd were measured below the detection limit of the analytical instrument, while the Pb level was higher than the Codex Alimentarius Commission value of 0.02 mg/kg, indicating unsafe for consumption. However, the HQ of Pb and other metals in milk and milk product samples was less than 1.0, thereby establishing no adverse health effects for Pb and other metals. The distribution pattern of metals in milk and milk product samples followed the decreasing order of Zn > Fe > Ni > Co > Cu > Mn > Pb > Cd/Cr. The CR levels of meals were also less than the permissible limit of 1.0 x 10-4, establishing no possible development of cancer. Keywords: adverse effects, cancer, metals, milk, milk product, the permissible limit.

Keywords: adverse effects, cancer, metals, milk, milk product, permissible limit

Procedia PDF Downloads 59
8934 Deterioration Prediction of Pavement Load Bearing Capacity from FWD Data

Authors: Kotaro Sasai, Daijiro Mizutani, Kiyoyuki Kaito

Abstract:

Expressways in Japan have been built in an accelerating manner since the 1960s with the aid of rapid economic growth. About 40 percent in length of expressways in Japan is now 30 years and older and has become superannuated. Time-related deterioration has therefore reached to a degree that administrators, from a standpoint of operation and maintenance, are forced to take prompt measures on a large scale aiming at repairing inner damage deep in pavements. These measures have already been performed for bridge management in Japan and are also expected to be embodied for pavement management. Thus, planning methods for the measures are increasingly demanded. Deterioration of layers around road surface such as surface course and binder course is brought about at the early stages of whole pavement deterioration process, around 10 to 30 years after construction. These layers have been repaired primarily because inner damage usually becomes significant after outer damage, and because surveys for measuring inner damage such as Falling Weight Deflectometer (FWD) survey and open-cut survey are costly and time-consuming process, which has made it difficult for administrators to focus on inner damage as much as they have been supposed to. As expressways today have serious time-related deterioration within them deriving from the long time span since they started to be used, it is obvious the idea of repairing layers deep in pavements such as base course and subgrade must be taken into consideration when planning maintenance on a large scale. This sort of maintenance requires precisely predicting degrees of deterioration as well as grasping the present situations of pavements. Methods for predicting deterioration are determined to be either mechanical or statistical. While few mechanical models have been presented, as far as the authors know of, previous studies have presented statistical methods for predicting deterioration in pavements. One describes deterioration process by estimating Markov deterioration hazard model, while another study illustrates it by estimating Proportional deterioration hazard model. Both of the studies analyze deflection data obtained from FWD surveys and present statistical methods for predicting deterioration process of layers around road surface. However, layers of base course and subgrade remain unanalyzed. In this study, data collected from FWD surveys are analyzed to predict deterioration process of layers deep in pavements in addition to surface layers by a means of estimating a deterioration hazard model using continuous indexes. This model can prevent the loss of information of data when setting rating categories in Markov deterioration hazard model when evaluating degrees of deterioration in roadbeds and subgrades. As a result of portraying continuous indexes, the model can predict deterioration in each layer of pavements and evaluate it quantitatively. Additionally, as the model can also depict probability distribution of the indexes at an arbitrary point and establish a risk control level arbitrarily, it is expected that this study will provide knowledge like life cycle cost and informative content during decision making process referring to where to do maintenance on as well as when.

Keywords: deterioration hazard model, falling weight deflectometer, inner damage, load bearing capacity, pavement

Procedia PDF Downloads 360
8933 Seismic Activity in the Lake Kivu Basin: Implication for Seismic Risk Management

Authors: Didier Birimwiragi Namogo

Abstract:

The Kivu Lake Basin is located in the Western Branch of the East African Rift. In this basin is located a multitude of active faults, on which earthquakes occur regularly. The most recent earthquakes date from 2008, 2015, 2016, 2017 and 2019. The cities of Bukabu and Goma in DR Congo and Giseyi in Rwanda are the most impacted by this intense seismic activity in the region. The magnitude of the strongest earthquakes in the region is 6.1. The 2008 earthquake was particularly destructive, killing several people in DR Congo and Rwanda. This work aims to complete the distribution of seismicity in the region, deduce areas of weakness and establish a hazard map that can assist in seismic risk management. Using the local seismic network of the Goma Volcano Observatory, the earthquakes were relocated, and their focus mechanism was studied. The results show that most of these earthquakes occur on active faults described by Villeneuve in 1938. The alignment of the earthquakes shows a pace that follows directly the directions of the faults described by this author. The study of the focus mechanism of these earthquakes, also shows that these are in particular normal faults whose stresses show an extensive activity. Such study can be used for the establishment of seismic risk management tools.

Keywords: earthquakes, hazard map, faults, focus mechanism

Procedia PDF Downloads 116
8932 A Review on Parametric Optimization of Casting Processes Using Optimization Techniques

Authors: Bhrugesh Radadiya, Jaydeep Shah

Abstract:

In Indian foundry industry, there is a need of defect free casting with minimum production cost in short lead time. Casting defect is a very large issue in foundry shop which increases the rejection rate of casting and wastage of materials. The various parameters influences on casting process such as mold machine related parameters, green sand related parameters, cast metal related parameters, mold related parameters and shake out related parameters. The mold related parameters are most influences on casting defects in sand casting process. This paper review the casting produced by foundry with shrinkage and blow holes as a major defects was analyzed and identified that mold related parameters such as mold temperature, pouring temperature and runner size were not properly set in sand casting process. These parameters were optimized using different optimization techniques such as Taguchi method, Response surface methodology, Genetic algorithm and Teaching-learning based optimization algorithm. Finally, concluded that a Teaching-learning based optimization algorithm give better result than other optimization techniques.

Keywords: casting defects, genetic algorithm, parametric optimization, Taguchi method, TLBO algorithm

Procedia PDF Downloads 705
8931 Effects of Local Ground Conditions on Site Response Analysis Results in Hungary

Authors: Orsolya Kegyes-Brassai, Zsolt Szilvágyi, Ákos Wolf, Richard P. Ray

Abstract:

Local ground conditions have a substantial influence on the seismic response of structures. Their inclusion in seismic hazard assessment and structural design can be realized at different levels of sophistication. However, response results based on more advanced calculation methods e.g. nonlinear or equivalent linear site analysis tend to show significant discrepancies when compared to simpler approaches. This project's main objective was to compare results from several 1-D response programs to Eurocode 8 design spectra. Data from in-situ site investigations were used for assessing local ground conditions at several locations in Hungary. After discussion of the in-situ measurements and calculation methods used, a comprehensive evaluation of all major contributing factors for site response is given. While the Eurocode spectra should account for local ground conditions based on soil classification, there is a wide variation in peak ground acceleration determined from 1-D analyses versus Eurocode. Results show that current Eurocode 8 design spectra may not be conservative enough to account for local ground conditions typical for Hungary.

Keywords: 1-D site response analysis, multichannel analysis of surface waves (MASW), seismic CPT, seismic hazard assessment

Procedia PDF Downloads 230
8930 Chairussyuhur Arman, Totti Tjiptosumirat, Muhammad Gunawan, Mastur, Joko Priyono, Baiq Tri Ratna Erawati

Authors: Maria M. Giannakou, Athanasios K. Ziliaskopoulos

Abstract:

Transmission pipelines carrying natural gas are often routed through populated cities, industrial and environmentally sensitive areas. While the need for these networks is unquestionable, there are serious concerns about the risk these lifeline networks pose to the people, to their habitat and to the critical infrastructures, especially in view of natural disasters such as earthquakes. This work presents an Integrated Pipeline Risk Management methodology (IPRM) for assessing the hazard associated with a natural gas pipeline failure due to natural or manmade disasters. IPRM aims to optimize the allocation of the available resources to countermeasures in order to minimize the impacts of pipeline failure to humans, the environment, the infrastructure and the economic activity. A proposed knapsack mathematical programming formulation is introduced that optimally selects the proper mitigation policies based on the estimated cost – benefit ratios. The proposed model is demonstrated with a small numerical example. The vulnerability analysis of these pipelines and the quantification of consequences from such failures can be useful for natural gas industries on deciding which mitigation measures to implement on the existing pipeline networks with the minimum cost in an acceptable level of hazard.

Keywords: cost benefit analysis, knapsack problem, natural gas distribution network, risk management, risk mitigation

Procedia PDF Downloads 272
8929 Irrigation Water Quality Evaluation in Jiaokou Irrigation District, Guanzhong Basin

Authors: Qiying Zhang, Panpan Xu, Hui Qian

Abstract:

Groundwater is an important water resource in the world, especially in arid and semi-arid regions. In the present study, 141 groundwater samples were collected and analyzed for various physicochemical parameters to assess the irrigation water quality using six indicators (sodium percentage (Na%), sodium adsorption ratio (SAR), magnesium hazard (MH), residual sodium carbonate (RSC), permeability index (PI), and potential salinity (PS)). The results show that the patterns for the average cation and anion concentrations were in decreasing orders of Na > Mg2 > Ca2 > Kand SO42 > HCO3 > Cl > NO3 > CO32 > F, respectively. The values of Na%, MH, and PS show that most of the groundwater samples are not suitable for irrigation. The same conclusion is drawn from the USSL and Wilcox diagrams. PS values indicate that Cland SO42have a great influence on irrigation water in Jiaokou Irrigation District. RSC and PI values indicate that more than half of groundwater samples are suitable for irrigation. The finding is beneficial for the policymakers for future water management schemes to achieve a sustainable development goal.

Keywords: groundwater chemistry, Guanzhong Basin, irrigation water quality evaluation, Jiaokou Irrigation District

Procedia PDF Downloads 178
8928 Optimizing of Machining Parameters of Plastic Material Using Taguchi Method

Authors: Jumazulhisham Abdul Shukor, Mohd. Sazali Said, Roshanizah Harun, Shuib Husin, Ahmad Razlee Ab Kadir

Abstract:

This paper applies Taguchi Optimization Method in determining the best machining parameters for pocket milling process on Polypropylene (PP) using CNC milling machine where the surface roughness is considered and the Carbide inserts cutting tool are used. Three machining parameters; speed, feed rate and depth of cut are investigated along three levels; low, medium and high of each parameter (Taguchi Orthogonal Arrays). The setting of machining parameters were determined by using Taguchi Method and the Signal-to-Noise (S/N) ratio are assessed to define the optimal levels and to predict the effect of surface roughness with assigned parameters based on L9. The final experimental outcomes are presented to prove the optimization parameters recommended by manufacturer are accurate.

Keywords: inserts, milling process, signal-to-noise (S/N) ratio, surface roughness, Taguchi Optimization Method

Procedia PDF Downloads 605
8927 Fire Safety Assessment of At-Risk Groups

Authors: Naser Kazemi Eilaki, Carolyn Ahmer, Ilona Heldal, Bjarne Christian Hagen

Abstract:

Older people and people with disabilities are recognized as at-risk groups when it comes to egress and travel from hazard zone to safe places. One's disability can negatively influence her or his escape time, and this becomes even more important when people from this target group live alone. This research deals with the fire safety of mentioned people's buildings by means of probabilistic methods. For this purpose, fire safety is addressed by modeling the egress of our target group from a hazardous zone to a safe zone. A common type of detached house with a prevalent plan has been chosen for safety analysis, and a limit state function has been developed according to the time-line evacuation model, which is based on a two-zone and smoke development model. An analytical computer model (B-Risk) is used to consider smoke development. Since most of the involved parameters in the fire development model pose uncertainty, an appropriate probability distribution function has been considered for each one of the variables with indeterministic nature. To achieve safety and reliability for the at-risk groups, the fire safety index method has been chosen to define the probability of failure (causalities) and safety index (beta index). An improved harmony search meta-heuristic optimization algorithm has been used to define the beta index. Sensitivity analysis has been done to define the most important and effective parameters for the fire safety of the at-risk group. Results showed an area of openings and intervals to egress exits are more important in buildings, and the safety of people would improve with increasing dimensions of occupant space (building). Fire growth is more critical compared to other parameters in the home without a detector and fire distinguishing system, but in a home equipped with these facilities, it is less important. Type of disabilities has a great effect on the safety level of people who live in the same home layout, and people with visual impairment encounter more risk of capturing compared to visual and movement disabilities.

Keywords: fire safety, at-risk groups, zone model, egress time, uncertainty

Procedia PDF Downloads 80
8926 Effect of Design Parameters on a Two Stage Launch Vehicle Performance

Authors: Assem Sallam, Aly Elzahaby, Ahmed Makled, Mohamed Khalil

Abstract:

Change in design parameters of launch vehicle affects its overall flight path trajectory. In this paper, several design parameters are introduced to study their effect. Selected parameters are the launch vehicle mass, which is presented in the form of payload mass, the maximum allowable angle of attack the launch vehicle can withstand, the flight path angle that is predefined for the launch vehicle second stage, the required inclination and its effect on the launch azimuth and finally by changing the launch pad coordinate. Selected design parameters are studied for their effect on the variation of altitude, ground range, absolute velocity and the flight path angle. The study gives a general mean of adjusting the design parameters to reach the required launch vehicle performance.

Keywords: launch vehicle azimuth, launch vehicle trajectory, launch vehicle payload, launch pad location

Procedia PDF Downloads 291
8925 Symbolic Analysis of Power Spectrum of CMOS Cross Couple Oscillator

Authors: Kittipong Tripetch

Abstract:

This paper proposes for the first time symbolic formula of the power spectrum of cross couple oscillator and its modified circuit. Many principle existed to derived power spectrum in microwave textbook such as impedance, admittance parameters, ABCD, H parameters, etc. It can be compared by graph of power spectrum which methodology is the best from the point of view of practical measurement setup such as condition of impedance parameter which used superposition of current to derived (its current injection of the other port of the circuit is zero, which is impossible in reality). Four Graphs of impedance parameters of cross couple oscillator is proposed. After that four graphs of Scattering parameters of cross couple oscillator will be shown.

Keywords: optimization, power spectrum, impedance parameters, scattering parameter

Procedia PDF Downloads 441
8924 Optimization of End Milling Process Parameters for Minimization of Surface Roughness of AISI D2 Steel

Authors: Pankaj Chandna, Dinesh Kumar

Abstract:

The present work analyses different parameters of end milling to minimize the surface roughness for AISI D2 steel. D2 Steel is generally used for stamping or forming dies, punches, forming rolls, knives, slitters, shear blades, tools, scrap choppers, tyre shredders etc. Surface roughness is one of the main indices that determines the quality of machined products and is influenced by various cutting parameters. In machining operations, achieving desired surface quality by optimization of machining parameters, is a challenging job. In case of mating components the surface roughness become more essential and is influenced by the cutting parameters, because, these quality structures are highly correlated and are expected to be influenced directly or indirectly by the direct effect of process parameters or their interactive effects (i.e. on process environment). In this work, the effects of selected process parameters on surface roughness and subsequent setting of parameters with the levels have been accomplished by Taguchi’s parameter design approach. The experiments have been performed as per the combination of levels of different process parameters suggested by L9 orthogonal array. Experimental investigation of the end milling of AISI D2 steel with carbide tool by varying feed, speed and depth of cut and the surface roughness has been measured using surface roughness tester. Analyses of variance have been performed for mean and signal-to-noise ratio to estimate the contribution of the different process parameters on the process.

Keywords: D2 steel, orthogonal array, optimization, surface roughness, Taguchi methodology

Procedia PDF Downloads 524
8923 Methane Production from Biomedical Waste (Blood)

Authors: Fatima M. Kabbashi, Abdalla M. Abdalla, Hussam K. Hamad, Elias S. Hassan

Abstract:

This study investigates the production of renewable energy (biogas) from biomedical hazard waste (blood) and eco-friendly disposal. Biogas is produced by the bacterial anaerobic digestion of biomaterial (blood). During digestion process bacterial feeding result in breaking down chemical bonds of the biomaterial and changing its features, by the end of the digestion (biogas production) the remains become manure as known. That has led to the economic and eco-friendly disposal of hazard biomedical waste (blood). The samples (Whole blood, Red blood cells 'RBCs', Blood platelet and Fresh Frozen Plasma ‘FFP’) are collected and measured in terms of carbon to nitrogen C/N ratio and total solid, then filled in connected flasks (three flasks) using water displacement method. The results of trails showed that the platelet and FFP failed to produce flammable gas, but via a gas analyzer, it showed the presence of the following gases: CO, HC, CO₂, and NOX. Otherwise, the blood and RBCs produced flammable gases: Methane-nitrous CH₃NO (99.45%), which has a blue color flame and carbon dioxide CO₂ (0.55%), which has red/yellow color flame. Methane-nitrous is sometimes used as fuel for rockets, some aircraft and racing cars.

Keywords: renewable energy, biogas, biomedical waste, blood, anaerobic digestion, eco-friendly disposal

Procedia PDF Downloads 281
8922 On the Survival of Individuals with Type 2 Diabetes Mellitus in the United Kingdom: A Retrospective Case-Control Study

Authors: Njabulo Ncube, Elena Kulinskaya, Nicholas Steel, Dmitry Pshezhetskiy

Abstract:

Life expectancy in the United Kingdom (UK) has been near constant since 2010, particularly for the individuals of 65 years and older. This trend has been also noted in several other countries. This slowdown in the increase of life expectancy was concurrent with the increase in the number of deaths caused by non-communicable diseases. Of particular concern is the world-wide exponential increase in the number of diabetes related deaths. Previous studies have reported increased mortality hazards among diabetics compared to non-diabetics, and on the differing effects of antidiabetic drugs on mortality hazards. This study aimed to estimate the all-cause mortality hazards and related life expectancies among type 2 diabetes (T2DM) patients in the UK using the time-variant Gompertz-Cox model with frailty. The study also aimed to understand the major causes of the change in life expectancy growth in the last decade. A total of 221 182 (30.8% T2DM, 57.6% Males) individuals aged 50 years and above, born between 1930 and 1960, inclusive, and diagnosed between 2000 and 2016, were selected from The Health Improvement Network (THIN) database of the UK primary care data and followed up to 31 December 2016. About 13.4% of participants died during the follow-up period. The overall all-cause mortality hazard ratio of T2DM compared to non-diabetic controls was 1.467 (1.381-1.558) and 1.38 (1.307-1.457) when diagnosed between 50 to 59 years and 60 to 74 years, respectively. The estimated life expectancies among T2DM individuals without further comorbidities diagnosed at the age of 60 years were 2.43 (1930-1939 birth cohort), 2.53 (1940-1949 birth cohort) and 3.28 (1950-1960 birth cohort) years less than those of non-diabetic controls. However, the 1950-1960 birth cohort had a steeper hazard function compared to the 1940-1949 birth cohort for both T2DM and non-diabetic individuals. In conclusion, mortality hazards for people with T2DM continue to be higher than for non-diabetics. The steeper mortality hazard slope for the 1950-1960 birth cohort might indicate the sub-population contributing to a slowdown in the growth of the life expectancy.

Keywords: T2DM, Gompetz-Cox model with frailty, all-cause mortality, life expectancy

Procedia PDF Downloads 104
8921 Causal Estimation for the Left-Truncation Adjusted Time-Varying Covariates under the Semiparametric Transformation Models of a Survival Time

Authors: Yemane Hailu Fissuh, Zhongzhan Zhang

Abstract:

In biomedical researches and randomized clinical trials, the most commonly interested outcomes are time-to-event so-called survival data. The importance of robust models in this context is to compare the effect of randomly controlled experimental groups that have a sense of causality. Causal estimation is the scientific concept of comparing the pragmatic effect of treatments conditional to the given covariates rather than assessing the simple association of response and predictors. Hence, the causal effect based semiparametric transformation model was proposed to estimate the effect of treatment with the presence of possibly time-varying covariates. Due to its high flexibility and robustness, the semiparametric transformation model which shall be applied in this paper has been given much more attention for estimation of a causal effect in modeling left-truncated and right censored survival data. Despite its wide applications and popularity in estimating unknown parameters, the maximum likelihood estimation technique is quite complex and burdensome in estimating unknown parameters and unspecified transformation function in the presence of possibly time-varying covariates. Thus, to ease the complexity we proposed the modified estimating equations. After intuitive estimation procedures, the consistency and asymptotic properties of the estimators were derived and the characteristics of the estimators in the finite sample performance of the proposed model were illustrated via simulation studies and Stanford heart transplant real data example. To sum up the study, the bias of covariates was adjusted via estimating the density function for truncation variable which was also incorporated in the model as a covariate in order to relax the independence assumption of failure time and truncation time. Moreover, the expectation-maximization (EM) algorithm was described for the estimation of iterative unknown parameters and unspecified transformation function. In addition, the causal effect was derived by the ratio of the cumulative hazard function of active and passive experiments after adjusting for bias raised in the model due to the truncation variable.

Keywords: causal estimation, EM algorithm, semiparametric transformation models, time-to-event outcomes, time-varying covariate

Procedia PDF Downloads 103
8920 Human Health Risk Assessment of Mercury-Contaminated Soils in Alebediah Mining Community, Sudan

Authors: Ahmed Elwaleed, Huiho Jeong, Ali H. Abdelbagi, Nguyen Thi Quynh, Koji Arizono, Yasuhiro Ishibashi

Abstract:

Artisanal and small-scale gold mining (ASGM) poses substantial risks to both human health and the environment, particularly through contamination of soil, water, and air. Prolonged exposure to ASGM-contaminated soils can lead to acute or chronic mercury toxicity. This study assesses the human health risks associated with mercury-contaminated soils and tailings in the Alebediah mining community in Sudan. Soil samples were collected from various locations within Alebediah, including ASGM areas, farmlands, and residential areas, along with tailings samples commonly found within ASGM sites. The evaluation of potential health risks to humans included the computation of the estimated daily intake (AvDI), the hazard quotient (HQ), and the hazard index (HI) for both adults and children. The primary exposure route identified as potentially posing a significant health risk was the volatilization of mercury from tailings samples, where mercury concentrations reached up to 25.5 mg/kg. In contrast, other samples within the ASGM area showed elevated mercury levels but did not present significant health risks, with HI values below 1. However, all areas indicated HI values above 1 for the remaining exposure routes. The study observed a decrease in mercury concentration with increasing distance from the ASGM community. Additionally, soil samples revealed elevated mercury levels exceeding background values, prompting an assessment of contamination levels using the enrichment factor (EF). The findings indicated that farmlands and residential areas exhibited depleted EF, while areas surrounding the ASGM community showed none to moderate pollution. In contrast, ASGM areas exhibited significant to extreme pollution. A GIS map was generated to visually depict the extent of mercury pollution, facilitating communication with stakeholders and decision-makers.

Keywords: mercury pollution, artisanal and small-scale gold mining, health risk assessment, hazard index, soil and tailings, enrichment factor

Procedia PDF Downloads 53