Search results for: Edward F. Di Carlo
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 530

Search results for: Edward F. Di Carlo

350 Gas Permeation Behavior of Single and Mixed Gas Components Using an Asymmetric Ceramic Membrane

Authors: Ngozi Claribelle Nwogu, Mohammed Nasir Kajama, Godson Osueke, Edward Gobina

Abstract:

A unique sol–gel dip-coating process to form an asymmetric silica membrane with improved membrane performance and reproducibility has been reported. First, we deposited repeatedly a silica solution on top of a commercial alumina membrane support to improve its structural make up. The coated membrane is further processed under clean room conditions to avoid dust impurity and subsequent drying in an oven for high thermal, chemical and physical stability. The resulting asymmetric membrane exhibits a gradual change in the membrane layer thickness. Compared to a single-layer process using only the membrane support, the dual-layer process improves both flux and selectivity. For the scientifically significant difficulties of natural gas purification, collective CO2, CH4 and H2 gas fluxes and separation factors obtained gave reasonably excellent values. In addition, the membrane selectively separated hydrogen as demonstrated by a high concentration of hydrogen recovery.

Keywords: gas permeation, silica membrane, separation factor, membrane layer thickness

Procedia PDF Downloads 322
349 Reliability Levels of Reinforced Concrete Bridges Obtained by Mixing Approaches

Authors: Adrián D. García-Soto, Alejandro Hernández-Martínez, Jesús G. Valdés-Vázquez, Reyna A. Vizguerra-Alvarez

Abstract:

Reinforced concrete bridges designed by code are intended to achieve target reliability levels adequate for the geographical environment where the code is applicable. Several methods can be used to estimate such reliability levels. Many of them require the establishment of an explicit limit state function (LSF). When such LSF is not available as a close-form expression, the simulation techniques are often employed. The simulation methods are computing intensive and time consuming. Note that if the reliability of real bridges designed by code is of interest, numerical schemes, the finite element method (FEM) or computational mechanics could be required. In these cases, it can be quite difficult (or impossible) to establish a close-form of the LSF, and the simulation techniques may be necessary to compute reliability levels. To overcome the need for a large number of simulations when no explicit LSF is available, the point estimate method (PEM) could be considered as an alternative. It has the advantage that only the probabilistic moments of the random variables are required. However, in the PEM, fitting of the resulting moments of the LSF to a probability density function (PDF) is needed. In the present study, a very simple alternative which allows the assessment of the reliability levels when no explicit LSF is available and without the need of extensive simulations is employed. The alternative includes the use of the PEM, and its applicability is shown by assessing reliability levels of reinforced concrete bridges in Mexico when a numerical scheme is required. Comparisons with results by using the Monte Carlo simulation (MCS) technique are included. To overcome the problem of approximating the probabilistic moments from the PEM to a PDF, a well-known distribution is employed. The approach mixes the PEM and other classic reliability method (first order reliability method, FORM). The results in the present study are in good agreement whit those computed with the MCS. Therefore, the alternative of mixing the reliability methods is a very valuable option to determine reliability levels when no close form of the LSF is available, or if numerical schemes, the FEM or computational mechanics are employed.

Keywords: structural reliability, reinforced concrete bridges, combined approach, point estimate method, monte carlo simulation

Procedia PDF Downloads 318
348 Performance Evaluation of an Inventive Co2 Gas Separation Inorganic Ceramic Membrane System

Authors: Ngozi Claribelle Nwogu, Mohammed Nasir Kajama, Oyoh Kechinyere, Edward Gobina

Abstract:

Atmospheric carbon dioxide emissions are considered as the greatest environmental challenge the world is facing today. The challenges to control the emissions include the recovery of CO2 from flue gas. This concern has been improved due to recent advances in materials process engineering resulting in the development of inorganic gas separation membranes with excellent thermal and mechanical stability required for most gas separations. This paper therefore evaluates the performance of a highly selective inorganic membrane for CO2 recovery applications. Analysis of results obtained is in agreement with experimental literature data. Further results show the prediction performance of the membranes for gas separation and the future direction of research. The materials selection and the membrane preparation techniques are discussed. Method of improving the interface defects in the membrane and its effect on the separation performance has also been reviewed and in addition advances to totally exploit the potential usage of this innovative membrane.

Keywords: carbon dioxide, gas separation, inorganic ceramic membrane, permselectivity

Procedia PDF Downloads 290
347 Local Politics in Taiwan: The Comparison among Magistrates’ Administrative Satisfaction

Authors: Edward Hwang

Abstract:

The efficiency of public policies depends on customer's satisfaction, and the change directions of public policies hinge on customer's assessment; i.e., the performance of public policies in governments is decided by the citizen's administrative satisfaction! The governments in all levels must heave their efficiency and effectiveness of public services to meet the people's substantially multiple needs in order to make citizens trust the governmental operation styles. To pursue the societal equality and justice, governments should treat people equally and provide more services for the disadvantages. The recent Dapu protest event involves Miaoli county Magistrate Liu cheng-hung who arbitrarily demolished houses and destroyed farmlands, and it shifts his popularity into disaffection. Liu case tells us that the political events are lethal to politicians; it cut almost 20% satisfaction degree for Magistrate Liu and hurt KMT support levels nationally. In terms of administrative satisfaction levels, political factors do matter, especially for the derogated events.

Keywords: local politics, administrative satisfaction, Taiwan, customer satisfaction

Procedia PDF Downloads 414
346 The AI Arena: A Framework for Distributed Multi-Agent Reinforcement Learning

Authors: Edward W. Staley, Corban G. Rivera, Ashley J. Llorens

Abstract:

Advances in reinforcement learning (RL) have resulted in recent breakthroughs in the application of artificial intelligence (AI) across many different domains. An emerging landscape of development environments is making powerful RL techniques more accessible for a growing community of researchers. However, most existing frameworks do not directly address the problem of learning in complex operating environments, such as dense urban settings or defense-related scenarios, that incorporate distributed, heterogeneous teams of agents. To help enable AI research for this important class of applications, we introduce the AI Arena: a scalable framework with flexible abstractions for distributed multi-agent reinforcement learning. The AI Arena extends the OpenAI Gym interface to allow greater flexibility in learning control policies across multiple agents with heterogeneous learning strategies and localized views of the environment. To illustrate the utility of our framework, we present experimental results that demonstrate performance gains due to a distributed multi-agent learning approach over commonly-used RL techniques in several different learning environments.

Keywords: reinforcement learning, multi-agent, deep learning, artificial intelligence

Procedia PDF Downloads 129
345 Consideration of Uncertainty in Engineering

Authors: A. Mohammadi, M. Moghimi, S. Mohammadi

Abstract:

Engineers need computational methods which could provide solutions less sensitive to the environmental effects, so the techniques should be used which take the uncertainty to account to control and minimize the risk associated with design and operation. In order to consider uncertainty in engineering problem, the optimization problem should be solved for a suitable range of the each uncertain input variable instead of just one estimated point. Using deterministic optimization problem, a large computational burden is required to consider every possible and probable combination of uncertain input variables. Several methods have been reported in the literature to deal with problems under uncertainty. In this paper, different methods presented and analyzed.

Keywords: uncertainty, Monte Carlo simulated, stochastic programming, scenario method

Procedia PDF Downloads 374
344 Performance Evaluation of a Millimeter-Wave Phased Array Antenna Using Circularly Polarized Elements

Authors: Rawad Asfour, Salam Khamas, Edward A. Ball

Abstract:

This paper is focused on the design of an mm-wave phased array. To date, linear polarization is adapted in the reported designs of phased arrays. However, linear polarization faces several well-known challenges. As such, an advanced design for phased array antennas is required that offers circularly polarized (CP) radiation. A feasible solution for achieving CP phased array antennas is proposed using open-circular loop antennas. To this end, a 3-element circular loop phased array antenna is designed to operate at 28GHz. In addition, the array ability to control the direction of the main lobe is investigated. The results show that the highest achievable field of view (FOV) is 100°, i.e., 50° to the left and 50° to the right-hand side directions. The results are achieved with a CP bandwidth of 15%. Furthermore, the results demonstrate that a high broadside gain of circa 11 dBi can be achieved for the steered beam. Besides, a radiation efficiency of 97 % can also be achieved based on the proposed design.

Keywords: loop antenna, phased array, beam steering, wide bandwidth, circular polarization, CST

Procedia PDF Downloads 258
343 Afghan Refugees as Perpetui Inimici: Revisiting an Early Modern Debate on Enemy Aliens

Authors: Brian Smith

Abstract:

This paper seeks to contrast the contemporary anti-immigration rhetoric since the US pullout of Afghanistan with the debate about enemy aliens in the early modern period. In the seventeenth century, Sir Edward Coke declared that “infidels” should be seen as perpetui inimici (perpetual enemies) since their values were inimical to those of Christian states. As such, they could be perpetually excluded and denied legal standing. Even at that time, these anti-“infidel” arguments clashed with the natural law tradition of hospitality, which assumed that states had a moral responsibility to admit and care for strangers. In particular, this paper looks at the conflict between Hugo Grotius, Samuel Pufendorf, and John Locke. Grotius argues that states have a duty to admit foreigners. He goes out of his way to plan for the admittance of Jewish immigrants. In contrast, Pufendorf claimed that sovereigns had a duty exclude foreigners who would alter the constitutional character of the state. Much like Grotius, Locke argued that non-Christian peoples should be treated as friends and admitted without reservation.

Keywords: enemy aliens, perpetual enemies, hospitality, refugees

Procedia PDF Downloads 58
342 Probabilistic Life Cycle Assessment of the Nano Membrane Toilet

Authors: A. Anastasopoulou, A. Kolios, T. Somorin, A. Sowale, Y. Jiang, B. Fidalgo, A. Parker, L. Williams, M. Collins, E. J. McAdam, S. Tyrrel

Abstract:

Developing countries are nowadays confronted with great challenges related to domestic sanitation services in view of the imminent water scarcity. Contemporary sanitation technologies established in these countries are likely to pose health risks unless waste management standards are followed properly. This paper provides a solution to sustainable sanitation with the development of an innovative toilet system, called Nano Membrane Toilet (NMT), which has been developed by Cranfield University and sponsored by the Bill & Melinda Gates Foundation. The particular technology converts human faeces into energy through gasification and provides treated wastewater from urine through membrane filtration. In order to evaluate the environmental profile of the NMT system, a deterministic life cycle assessment (LCA) has been conducted in SimaPro software employing the Ecoinvent v3.3 database. The particular study has determined the most contributory factors to the environmental footprint of the NMT system. However, as sensitivity analysis has identified certain critical operating parameters for the robustness of the LCA results, adopting a stochastic approach to the Life Cycle Inventory (LCI) will comprehensively capture the input data uncertainty and enhance the credibility of the LCA outcome. For that purpose, Monte Carlo simulations, in combination with an artificial neural network (ANN) model, have been conducted for the input parameters of raw material, produced electricity, NOX emissions, amount of ash and transportation of fertilizer. The given analysis has provided the distribution and the confidence intervals of the selected impact categories and, in turn, more credible conclusions are drawn on the respective LCIA (Life Cycle Impact Assessment) profile of NMT system. Last but not least, the specific study will also yield essential insights into the methodological framework that can be adopted in the environmental impact assessment of other complex engineering systems subject to a high level of input data uncertainty.

Keywords: sanitation systems, nano-membrane toilet, lca, stochastic uncertainty analysis, Monte Carlo simulations, artificial neural network

Procedia PDF Downloads 194
341 Anomaly Detection with ANN and SVM for Telemedicine Networks

Authors: Edward Guillén, Jeisson Sánchez, Carlos Omar Ramos

Abstract:

In recent years, a wide variety of applications are developed with Support Vector Machines -SVM- methods and Artificial Neural Networks -ANN-. In general, these methods depend on intrusion knowledge databases such as KDD99, ISCX, and CAIDA among others. New classes of detectors are generated by machine learning techniques, trained and tested over network databases. Thereafter, detectors are employed to detect anomalies in network communication scenarios according to user’s connections behavior. The first detector based on training dataset is deployed in different real-world networks with mobile and non-mobile devices to analyze the performance and accuracy over static detection. The vulnerabilities are based on previous work in telemedicine apps that were developed on the research group. This paper presents the differences on detections results between some network scenarios by applying traditional detectors deployed with artificial neural networks and support vector machines.

Keywords: anomaly detection, back-propagation neural networks, network intrusion detection systems, support vector machines

Procedia PDF Downloads 312
340 Risk Measure from Investment in Finance by Value at Risk

Authors: Mohammed El-Arbi Khalfallah, Mohamed Lakhdar Hadji

Abstract:

Managing and controlling risk is a topic research in the world of finance. Before a risky situation, the stakeholders need to do comparison according to the positions and actions, and financial institutions must take measures of a particular market risk and credit. In this work, we study a model of risk measure in finance: Value at Risk (VaR), which is a new tool for measuring an entity's exposure risk. We explain the concept of value at risk, your average, tail, and describe the three methods for computing: Parametric method, Historical method, and numerical method of Monte Carlo. Finally, we briefly describe advantages and disadvantages of the three methods for computing value at risk.

Keywords: average value at risk, conditional value at risk, tail value at risk, value at risk

Procedia PDF Downloads 405
339 Impact on Cost of Equity of Accounting and Disclosures

Authors: Abhishek Ranga

Abstract:

The study examined the effect of accounting choice and level of disclosure on the firm’s implied cost of equity in Indian environment. For the study accounting choice was classified as aggressive or conservative depending upon the firm’s choice of accounting methods, accounting policies and accounting estimates. Level of disclosure is the quantum of financial and non-financial information disclosed in firm’s annual report, essentially in note to accounts section, schedules forming part of financial statements and Management Discussion and Analysis report. Regression models were developed with cost of equity as a dependent variable and accounting choice, level of disclosure as an independent variable along with selected control variables. Cost of equity was measured using Edward-Bell-Ohlson (EBO) valuation model, to measure accounting choice Modified-Jones-Model (MJM) was used and level of disclosure was measured using a disclosure index essentially drawn from Botosan study. Results indicated a negative association between the implied cost of equity and conservative accounting choice and also between level of disclosure and cost of equity.

Keywords: aggressive accounting choice, conservative accounting choice, disclosure, implied cost of equity

Procedia PDF Downloads 428
338 Phylogenetic Analysis and a Review of the History of the Accidental Phytoplankter, Phaeodactylum tricornutum Bohlin (Bacillariophyta)

Authors: Jamal S. M. Sabir, Edward C. Theriot, Schonna R. Manning, Abdulrahman L. Al-Malki, Mohammad, Mumdooh J. Sabir, Dwight K. Romanovicz, Nahid H. Hajrah, Robert K. Jansen, Matt P. Ashworth

Abstract:

The diatom Phaeodactylum tricornutum has been used as a model for cell biologists and ecologists for over a century. We have incorporated several new raphid pennates into a three-gene phylogenetic dataset (SSU, rbcL, psbC), and recover Gomphonemopsis sp. as sister to P. tricornutum with 100% BS support. This is the first time a close relative has been identified for P. tricornutum with robust statistical support. We test and reject a succession of hypotheses for other relatives. Our molecular data are statistically significantly incongruent with placement of either or both species among the Cymbellales, an order of diatoms with which both have been associated. We believe that further resolution of the phylogenetic position of P. tricornutum will rely more on increased taxon sampling than increased genetic sampling. Gomphonemopsis is a benthic diatom, and its phylogenetic relationship with P. tricornutum is congruent with the hypothesis that P. tricornutum is a benthic diatom with specific adaptations that lead to active recruitment into the plankton. We hypothesize that other benthic diatoms are likely to have similar adaptations and are not merely passively recruited into the plankton.

Keywords: benthic, diatoms; ecology, Phaeodactylum tricornutum, phylogeny, tychoplankton

Procedia PDF Downloads 204
337 Identifying Factors Contributing to the Spread of Lyme Disease: A Regression Analysis of Virginia’s Data

Authors: Fatemeh Valizadeh Gamchi, Edward L. Boone

Abstract:

This research focuses on Lyme disease, a widespread infectious condition in the United States caused by the bacterium Borrelia burgdorferi sensu stricto. It is critical to identify environmental and economic elements that are contributing to the spread of the disease. This study examined data from Virginia to identify a subset of explanatory variables significant for Lyme disease case numbers. To identify relevant variables and avoid overfitting, linear poisson, and regularization regression methods such as a ridge, lasso, and elastic net penalty were employed. Cross-validation was performed to acquire tuning parameters. The methods proposed can automatically identify relevant disease count covariates. The efficacy of the techniques was assessed using four criteria on three simulated datasets. Finally, using the Virginia Department of Health’s Lyme disease data set, the study successfully identified key factors, and the results were consistent with previous studies.

Keywords: lyme disease, Poisson generalized linear model, ridge regression, lasso regression, elastic net regression

Procedia PDF Downloads 85
336 On Coverage Probability of Confidence Intervals for the Normal Mean with Known Coefficient of Variation

Authors: Suparat Niwitpong, Sa-aat Niwitpong

Abstract:

Statistical inference of normal mean with known coefficient of variation has been investigated recently. This phenomenon occurs normally in environment and agriculture experiments when the scientist knows the coefficient of variation of their experiments. In this paper, we constructed new confidence intervals for the normal population mean with known coefficient of variation. We also derived analytic expressions for the coverage probability of each confidence interval. To confirm our theoretical results, Monte Carlo simulation will be used to assess the performance of these intervals based on their coverage probabilities.

Keywords: confidence interval, coverage probability, expected length, known coefficient of variation

Procedia PDF Downloads 354
335 The Cost of Solar-Centric Renewable Portfolio

Authors: Timothy J. Considine, Edward J. M. Manderson

Abstract:

This paper develops an econometric forecasting system of energy demand coupled with engineering-economic models of energy supply. The framework is used to quantify the impact of state-level renewable portfolio standards (RPSs) achieved predominately with solar generation on electricity rates, electricity consumption, and environmental quality. We perform the analysis using Arizona’s RPS as a case study. We forecast energy demand in Arizona out to 2035, and find by this time the state will require an additional 35 million MWh of electricity generation. If Arizona implements its RPS when supplying this electricity demand, we find there will be a substantial increase in electricity rates (relative to a business-as-usual scenario of reliance on gas-fired generation). Extending the current regime of tax credits can greatly reduce this increase, at the taxpayers’ expense. We find that by 2025 Arizona’s RPS will implicitly abate carbon dioxide emissions at a cost between $101 and $135 per metric ton, and by 2035 abatement costs are between $64 and $112 per metric ton (depending on the future evolution of nature gas prices).

Keywords: electricity demand, renewable portfolio standard, solar, carbon dioxide

Procedia PDF Downloads 446
334 A Risk-Based Approach to Construction Management

Authors: Chloe E. Edwards, Yasaman Shahtaheri

Abstract:

Risk management plays a fundamental role in project planning and delivery. The purpose of incorporating risk management into project management practices is to identify and address uncertainties related to key project-related activities. The uncertainties, known as risk events, can relate to project deliverables that are quantifiable and are often measured by impact to project schedule, cost, or environmental impact. Risk management should be incorporated as an iterative practice throughout the planning, execution, and commissioning phases of a project. This paper specifically examines how risk management contributes to effective project planning and delivery through a case study of a transportation project. This case study focused solely on impacts to project schedule regarding three milestones: readiness for delivery, readiness for testing and commissioning, and completion of the facility. The case study followed the ISO 31000: Risk Management – Guidelines. The key factors that are outlined by these guidelines include understanding the scope and context of the project, conducting a risk assessment including identification, analysis, and evaluation, and lastly, risk treatment through mitigation measures. This process requires continuous consultation with subject matter experts and monitoring to iteratively update the risks accordingly. The risk identification process led to a total of fourteen risks related to design, permitting, construction, and commissioning. The analysis involved running 1,000 Monte Carlo simulations through @RISK 8.0 Industrial software to determine potential milestone completion dates based on the project baseline schedule. These dates include the best case, most likely case, and worst case to provide an estimated delay for each milestone. Evaluation of these results provided insight into which risks were the highest contributors to the projected milestone completion dates. Based on the analysis results, the risk management team was able to provide recommendations for mitigation measures to reduce the likelihood of risks occurring. The risk management team also provided recommendations for managing the identified risks and project activities moving forward to meet the most likely or best-case milestone completion dates.

Keywords: construction management, monte carlo simulation, project delivery, risk assessment, transportation engineering

Procedia PDF Downloads 76
333 A One Dimensional Particle in Cell Model for Excimer Lamps

Authors: W. Benstaali, A. Belasri

Abstract:

In this work we study a planar lamp filled with neon-xenon gas. We use a one-dimensional particle in a cell with Monte Carlo simulation (PIC-MCC) to investigate the effect xenon concentration on the energy deposited on excitation, ionization and ions. A Xe-Ne discharge is studied for a gas pressure of 400 torr. The results show an efficient Xe20-Ne mixture with an applied voltage of 1.2KV; the xenon excitation energy represents 65% form total energy dissipated in the discharge. We have also studied electrical properties and the energy balance a discharge for Xe50-Ne which needs a voltage of 2kv; the xenon energy is than more important.

Keywords: dielectric barrier discharge, efficiency, excitation, lamps

Procedia PDF Downloads 130
332 Finite Sample Inferences for Weak Instrument Models

Authors: Gubhinder Kundhi, Paul Rilstone

Abstract:

It is well established that Instrumental Variable (IV) estimators in the presence of weak instruments can be poorly behaved, in particular, be quite biased in finite samples. Finite sample approximations to the distributions of these estimators are obtained using Edgeworth and Saddlepoint expansions. Departures from normality of the distributions of these estimators are analyzed using higher order analytical corrections in these expansions. In a Monte-Carlo experiment, the performance of these expansions is compared to the first order approximation and other methods commonly used in finite samples such as the bootstrap.

Keywords: bootstrap, Instrumental Variable, Edgeworth expansions, Saddlepoint expansions

Procedia PDF Downloads 280
331 Readiness of Intellectual Capital Measurement: A Review of the Property Development and Investment Industry

Authors: Edward C. W. Chan, Benny C. F. Cheung

Abstract:

In the knowledge economy, the financial indicator is not the unique instrument to gauge the performance of a company. The role of intellectual capital contributing to the company performance is increasing. To measure the company performance due to intellectual capital, the value-added intellectual capital (VAIC) model is adopted to measure the intellectual capital utilisation efficiency of the subject companies. The purpose of this study is to review the readiness of measuring intellectual capital for the Hong Kong listed companies in the property development and property investment industry by using VAIC model. This study covers the financial reports from the representative Hong Kong listed property development companies and property investment companies in the period 2014-2019. The findings from this study indicated the industry is ready for IC measurement employing VAIC framework but not yet ready for using the extended VAIC model.

Keywords: intellectual capital, intellectual capital measurement, property development, property investment, Skandia navigator, VAIC

Procedia PDF Downloads 79
330 Application Reliability Method for the Analysis of the Stability Limit States of Large Concrete Dams

Authors: Mustapha Kamel Mihoubi, Essadik Kerkar, Abdelhamid Hebbouche

Abstract:

According to the randomness of most of the factors affecting the stability of a gravity dam, probability theory is generally used to TESTING the risk of failure and there is a confusing logical transition from the state of stability failed state, so the stability failure process is considered as a probable event. The control of risk of product failures is of capital importance for the control from a cross analysis of the gravity of the consequences and effects of the probability of occurrence of identified major accidents and can incur a significant risk to the concrete dam structures. Probabilistic risk analysis models are used to provide a better understanding the reliability and structural failure of the works, including when calculating stability of large structures to a major risk in the event of an accident or breakdown. This work is interested in the study of the probability of failure of concrete dams through the application of the reliability analysis methods including the methods used in engineering. It is in our case of the use of level II methods via the study limit state. Hence, the probability of product failures is estimated by analytical methods of the type FORM (First Order Reliability Method), SORM (Second Order Reliability Method). By way of comparison, a second level III method was used which generates a full analysis of the problem and involving an integration of the probability density function of, random variables are extended to the field of security by using of the method of Mont-Carlo simulations. Taking into account the change in stress following load combinations: normal, exceptional and extreme the acting on the dam, calculation results obtained have provided acceptable failure probability values which largely corroborate the theory, in fact, the probability of failure tends to increase with increasing load intensities thus causing a significant decrease in strength, especially in the presence of combinations of unique and extreme loads. Shear forces then induce a shift threatens the reliability of the structure by intolerable values of the probability of product failures. Especially, in case THE increase of uplift in a hypothetical default of the drainage system.

Keywords: dam, failure, limit state, monte-carlo, reliability, probability, sliding, Taylor

Procedia PDF Downloads 291
329 A Saturation Attack Simulation on a Navy Warship Based on Discrete-Event Simulation Models

Authors: Yawei Liang

Abstract:

Threat from cruise missiles is among the most dangerous considerations to a warship in the modern era: anti-ship cruise missiles are fast, accurate, and extremely destructive. In this paper, the goal was to use an object-orientated environment to program a simulation to model a scenario in which a lone frigate is attacked by a wave of missiles fired at given intervals. The parameters of the simulation are modified to examine the relationships between different variables in the situation, and an analysis is performed on various aspects of the defending ship’s equipment. Finally, the results are presented, along with a brief discussion.

Keywords: discrete event simulation, Monte Carlo simulation, naval resource management, weapon-target allocation/assignment

Procedia PDF Downloads 55
328 Comparison of Power Consumption of WiFi Inbuilt Internet of Things Device with Bluetooth Low Energy

Authors: Darshana Thomas, Edward Wilkie, James Irvine

Abstract:

The Internet of things (IoT) is currently a highly researched topic, especially within the context of the smart home. These are small sensors that are capable of gathering data and transmitting it to a server. The majority of smart home products use protocols such as ZigBee or Bluetooth Low Energy (BLE). As these small sensors are increasing in number, the need to implement these with much more capable and ubiquitous transmission technology is necessary. The high power consumption is the reason that holds these small sensors back from using other protocols such as the most ubiquitous form of communication, WiFi. Comparing the power consumption of existing transmission technologies to one with WiFi inbuilt, would provide a better understanding for choosing between these technologies. We have developed a small IoT device with WiFi capability and proven that it is much more efficient than the first protocol, 433 MHz. We extend our work in this paper and compare WiFi power consumption with the other most widely used protocol BLE. The experimental results in this paper would conclude whether the developed prototype is capable in terms of power consumption to replace the existing protocol BLE with WiFi.

Keywords: bluetooth, internet of things (IoT), power consumption, WiFi

Procedia PDF Downloads 237
327 The Generalized Pareto Distribution as a Model for Sequential Order Statistics

Authors: Mahdy ‎Esmailian, Mahdi ‎Doostparast, Ahmad ‎Parsian

Abstract:

‎In this article‎, ‎sequential order statistics (SOS) censoring type II samples coming from the generalized Pareto distribution are considered‎. ‎Maximum likelihood (ML) estimators of the unknown parameters are derived on the basis of the available multiple SOS data‎. ‎Necessary conditions for existence and uniqueness of the derived ML estimates are given‎. Due to complexity in the proposed likelihood function‎, ‎a useful re-parametrization is suggested‎. ‎For illustrative purposes‎, ‎a Monte Carlo simulation study is conducted and an illustrative example is analysed‎.

Keywords: bayesian estimation‎, generalized pareto distribution‎, ‎maximum likelihood estimation‎, sequential order statistics

Procedia PDF Downloads 469
326 Design, Construction and Performance Evaluation of a HPGe Detector Shield

Authors: M. Sharifi, M. Mirzaii, F. Bolourinovin, H. Yousefnia, M. Akbari, K. Yousefi-Mojir

Abstract:

A multilayer passive shield composed of low-activity lead (Pb), copper (Cu), tin (Sn) and iron (Fe) was designed and manufactured for a coaxial HPGe detector placed at a surface laboratory for reducing background radiation and radiation dose to the personnel. The performance of the shield was evaluated and efficiency curves of the detector were plotted by using of the various standard sources in different distances. Monte Carlo simulations and a set of TLD chips were used for dose estimation in two distances of 20 and 40 cm. The results show that the shield reduced background spectrum and the personnel dose more than 95%.

Keywords: HPGe shield, background count, personnel dose, efficiency curve

Procedia PDF Downloads 417
325 Experimental Analysis of Advanced Multi-Axial Preforms Conformability to Complex Contours

Authors: Andrew Hardman, Alistair T. McIlhagger, Edward Archer

Abstract:

A degree of research has been undertaken in the determination of 3D textile preforms behaviour to compression with direct comparison to 2D counterparts. Multiscale simulations have been developed to try and accurately analyse the behaviour of varying architectures post-consolidation. However, further understanding is required to experimentally identify the mechanisms and deformations that exist upon conforming to a complex contour. Due to the complexity of 3D textile preforms, determination of yarn behaviour to a complex contour is assessed through consolidation by means of vacuum assisted resin transfer moulding (VARTM), and the resulting mechanisms are investigated by micrograph analysis. Varying architectures; with known areal densities, pic density and thicknesses are assessed for a cohesive study. The resulting performance of each is assessed qualitatively as well as quantitatively from the perspective of material in terms of the change in representative unit cell (RVE) across the curved beam contour, in crimp percentage, tow angle, resin rich areas and binder distortion. A novel textile is developed from the resulting analysis to overcome the observed deformations.

Keywords: comformability, compression, binder architecture, 3D weaving, textile preform

Procedia PDF Downloads 132
324 Towards the Design of Gripper Independent of Substrate Surface Structures

Authors: Annika Schmidt, Ausama Hadi Ahmed, Carlo Menon

Abstract:

End effectors for robotic systems are becoming more and more advanced, resulting in a growing variety of gripping tasks. However, most grippers are application specific. This paper presents a gripper that interacts with an object’s surface rather than being dependent on a defined shape or size. For this purpose, ingressive and astrictive features are combined to achieve the desired gripping capabilities. The developed prototype is tested on a variety of surfaces with different hardness and roughness properties. The results show that the gripping mechanism works on all of the tested surfaces. The influence of the material properties on the amount of the supported load is also studied and the efficiency is discussed.

Keywords: claw, dry adhesion, insects, material properties

Procedia PDF Downloads 324
323 Investigating the Effects of Data Transformations on a Bi-Dimensional Chi-Square Test

Authors: Alexandru George Vaduva, Adriana Vlad, Bogdan Badea

Abstract:

In this research, we conduct a Monte Carlo analysis on a two-dimensional χ2 test, which is used to determine the minimum distance required for independent sampling in the context of chaotic signals. We investigate the impact of transforming initial data sets from any probability distribution to new signals with a uniform distribution using the Spearman rank correlation on the χ2 test. This transformation removes the randomness of the data pairs, and as a result, the observed distribution of χ2 test values differs from the expected distribution. We propose a solution to this problem and evaluate it using another chaotic signal.

Keywords: chaotic signals, logistic map, Pearson’s test, Chi Square test, bivariate distribution, statistical independence

Procedia PDF Downloads 50
322 High Purity Germanium Detector Characterization by Means of Monte Carlo Simulation through Application of Geant4 Toolkit

Authors: Milos Travar, Jovana Nikolov, Andrej Vranicar, Natasa Todorovic

Abstract:

Over the years, High Purity Germanium (HPGe) detectors proved to be an excellent practical tool and, as such, have established their today's wide use in low background γ-spectrometry. One of the advantages of gamma-ray spectrometry is its easy sample preparation as chemical processing and separation of the studied subject are not required. Thus, with a single measurement, one can simultaneously perform both qualitative and quantitative analysis. One of the most prominent features of HPGe detectors, besides their excellent efficiency, is their superior resolution. This feature virtually allows a researcher to perform a thorough analysis by discriminating photons of similar energies in the studied spectra where otherwise they would superimpose within a single-energy peak and, as such, could potentially scathe analysis and produce wrongly assessed results. Naturally, this feature is of great importance when the identification of radionuclides, as well as their activity concentrations, is being practiced where high precision comes as a necessity. In measurements of this nature, in order to be able to reproduce good and trustworthy results, one has to have initially performed an adequate full-energy peak (FEP) efficiency calibration of the used equipment. However, experimental determination of the response, i.e., efficiency curves for a given detector-sample configuration and its geometry, is not always easy and requires a certain set of reference calibration sources in order to account for and cover broader energy ranges of interest. With the goal of overcoming these difficulties, a lot of researches turned towards the application of different software toolkits that implement the Monte Carlo method (e.g., MCNP, FLUKA, PENELOPE, Geant4, etc.), as it has proven time and time again to be a very powerful tool. In the process of creating a reliable model, one has to have well-established and described specifications of the detector. Unfortunately, the documentation that manufacturers provide alongside the equipment is rarely sufficient enough for this purpose. Furthermore, certain parameters tend to evolve and change over time, especially with older equipment. Deterioration of these parameters consequently decreases the active volume of the crystal and can thus affect the efficiencies by a large margin if they are not properly taken into account. In this study, the optimisation method of two HPGe detectors through the implementation of the Geant4 toolkit developed by CERN is described, with the goal of further improving simulation accuracy in calculations of FEP efficiencies by investigating the influence of certain detector variables (e.g., crystal-to-window distance, dead layer thicknesses, inner crystal’s void dimensions, etc.). Detectors on which the optimisation procedures were carried out were a standard traditional co-axial extended range detector (XtRa HPGe, CANBERRA) and a broad energy range planar detector (BEGe, CANBERRA). Optimised models were verified through comparison with experimentally obtained data from measurements of a set of point-like radioactive sources. Acquired results of both detectors displayed good agreement with experimental data that falls under an average statistical uncertainty of ∼ 4.6% for XtRa and ∼ 1.8% for BEGe detector within the energy range of 59.4−1836.1 [keV] and 59.4−1212.9 [keV], respectively.

Keywords: HPGe detector, γ spectrometry, efficiency, Geant4 simulation, Monte Carlo method

Procedia PDF Downloads 82
321 Unconventional Composite Inorganic Membrane Fabrication for Carbon Emissions Mitigation

Authors: Ngozi Nwogu, Godson Osueke, Mamdud Hossain, Edward Gobina

Abstract:

An unconventional composite inorganic ceramic membrane capable in carbon dioxide emission decline was fabricated and tested at laboratory scale to develop in conformism to various environmental guidelines to mitigate the effect of global warming. A review of the existing membrane technologies for carbon capture including the relevant gas transport mechanisms are presented and discussed. Single gas separation experiments using silica modified ceramic membrane with internal diameter 20mm, outside diameter 25mm and length of 368mm deposited on a macro porous supported reactor.was carried out to investigate individual gas permeation behaviours at different pressures and membrane efficiency after a dip coating method. Nitrogen, Carbon dioxide, Argon, Oxygen and Methane pure gases were used to investigate their individual permeation rates at various pressures. Results show that the gas flow rate increases with pressure drop. However at above a pressure of 3bar, CO2 permeability ratio to than the other gases indicated control of a more selective surface adsorptive transport mechanism.

Keywords: carbon dioxide, composite membranes, permeability, transport mechanisms

Procedia PDF Downloads 474