Search results for: mathematical simulation
455 Measuring Systems Interoperability: A Focal Point for Standardized Assessment of Regional Disaster Resilience
Authors: Joel Thomas, Alexa Squirini
Abstract:
The key argument of this research is that every element of systems interoperability is an enabler of regional disaster resilience, and arguably should become a focal point for standardized measurement of communities’ ability to work together. Few resilience research efforts have focused on the development and application of solutions that measurably improve communities’ ability to work together at a regional level, yet a majority of the most devastating and disruptive disasters are those that have had a regional impact. The key findings of the research include a unique theoretical, mathematical, and operational approach to tangibly and defensibly measure and assess systems interoperability required to support crisis information management activities performed by governments, the private sector, and humanitarian organizations. A most effective way for communities to measurably improve regional disaster resilience is through deliberately executed disaster preparedness activities. Developing interoperable crisis information management capabilities is a crosscutting preparedness activity that greatly affects a community’s readiness and ability to work together in times of crisis. Thus, improving communities’ human and technical posture to work together in advance of a crisis, with the ultimate goal of enabling information sharing to support coordination and the careful management of available resources, is a primary means by which communities may improve regional disaster resilience. This model describes how systems interoperability can be qualitatively and quantitatively assessed when characterized as five forms of capital: governance; standard operating procedures; technology; training and exercises; and usage. The unique measurement framework presented defines the relationships between systems interoperability, information sharing and safeguarding, operational coordination, community preparedness and regional disaster resilience, and offers a means by which to implement real-world solutions and measure progress over the course of a multi-year program. The model is being developed and piloted in partnership with the U.S. Department of Homeland Security (DHS) Science and Technology Directorate (S&T) and the North Atlantic Treaty Organization (NATO) Advanced Regional Civil Emergency Coordination Pilot (ARCECP) with twenty-three organizations in Bosnia and Herzegovina, Croatia, Macedonia, and Montenegro. The intended effect of the model implementation is to enable communities to answer two key questions: 'Have we measurably improved crisis information management capabilities as a result of this effort?' and, 'As a result, are we more resilient?'Keywords: disaster, interoperability, measurement, resilience
Procedia PDF Downloads 143454 Signal Processing Techniques for Adaptive Beamforming with Robustness
Authors: Ju-Hong Lee, Ching-Wei Liao
Abstract:
Adaptive beamforming using antenna array of sensors is useful in the process of adaptively detecting and preserving the presence of the desired signal while suppressing the interference and the background noise. For conventional adaptive array beamforming, we require a prior information of either the impinging direction or the waveform of the desired signal to adapt the weights. The adaptive weights of an antenna array beamformer under a steered-beam constraint are calculated by minimizing the output power of the beamformer subject to the constraint that forces the beamformer to make a constant response in the steering direction. Hence, the performance of the beamformer is very sensitive to the accuracy of the steering operation. In the literature, it is well known that the performance of an adaptive beamformer will be deteriorated by any steering angle error encountered in many practical applications, e.g., the wireless communication systems with massive antennas deployed at the base station and user equipment. Hence, developing effective signal processing techniques to deal with the problem due to steering angle error for array beamforming systems has become an important research work. In this paper, we present an effective signal processing technique for constructing an adaptive beamformer against the steering angle error. The proposed array beamformer adaptively estimates the actual direction of the desired signal by using the presumed steering vector and the received array data snapshots. Based on the presumed steering vector and a preset angle range for steering mismatch tolerance, we first create a matrix related to the direction vector of signal sources. Two projection matrices are generated from the matrix. The projection matrix associated with the desired signal information and the received array data are utilized to iteratively estimate the actual direction vector of the desired signal. The estimated direction vector of the desired signal is then used for appropriately finding the quiescent weight vector. The other projection matrix is set to be the signal blocking matrix required for performing adaptive beamforming. Accordingly, the proposed beamformer consists of adaptive quiescent weights and partially adaptive weights. Several computer simulation examples are provided for evaluating and comparing the proposed technique with the existing robust techniques.Keywords: adaptive beamforming, robustness, signal blocking, steering angle error
Procedia PDF Downloads 124453 Optimizing 3D Shape Parameters of Sports Bra Pads in Motion by Finite Element Dynamic Modelling with Inverse Problem Solution
Authors: Jiazhen Chen, Yue Sun, Joanne Yip, Kit-Lun Yick
Abstract:
The design of sports bras poses a considerable challenge due to the difficulty in accurately predicting the wearing result after computer-aided design (CAD). It needs repeated physical try-on or virtual try-on to obtain a comfortable pressure range during motion. Specifically, in the context of running, the exact support area and force exerted on the breasts remain unclear. Consequently, obtaining an effective method to design the sports bra pads shape becomes particularly challenging. This predicament hinders the successful creation and production of sports bras that cater to women's health needs. The purpose of this study is to propose an effective method to obtain the 3D shape of sports bra pads and to understand the relationship between the supporting force and the 3D shape parameters of the pads. Firstly, the static 3D shape of the sports bra pad and human motion data (Running) are obtained by using the 3D scanner and advanced 4D scanning technology. The 3D shape of the sports bra pad is parameterised and simplified by Free-form Deformation (FFD). Then the sub-models of sports bra and human body are constructed by segmenting and meshing them with MSC Apex software. The material coefficient of sports bras is obtained by material testing. The Marc software is then utilised to establish a dynamic contact model between the human breast and the sports bra pad. To realise the reverse design of the sports bra pad, this contact model serves as a forward model for calculating the inverse problem. Based on the forward contact model, the inverse problem of the 3D shape parameters of the sports bra pad with the target bra-wearing pressure range as the boundary condition is solved. Finally, the credibility and accuracy of the simulation are validated by comparing the experimental results with the simulations by the FE model on the pressure distribution. On the one hand, this research allows for a more accurate understanding of the support area and force distribution on the breasts during running. On the other hand, this study can contribute to the customization of sports bra pads for different individuals. It can help to obtain sports bra pads with comfortable dynamic pressure.Keywords: sports bra design, breast motion, running, inverse problem, finite element dynamic model
Procedia PDF Downloads 59452 Optimal Beam for Accelerator Driven Systems
Authors: M. Paraipan, V. M. Javadova, S. I. Tyutyunnikov
Abstract:
The concept of energy amplifier or accelerator driven system (ADS) involves the use of a particle accelerator coupled with a nuclear reactor. The accelerated particle beam generates a supplementary source of neutrons, which allows the subcritical functioning of the reactor, and consequently a safe exploitation. The harder neutron spectrum realized ensures a better incineration of the actinides. The almost generalized opinion is that the optimal beam for ADS is represented by protons with energy around 1 GeV (gigaelectronvolt). In the present work, a systematic analysis of the energy gain for proton beams with energy from 0.5 to 3 GeV and ion beams from deuteron to neon with energies between 0.25 and 2 AGeV is performed. The target is an assembly of metallic U-Pu-Zr fuel rods in a bath of lead-bismuth eutectic coolant. The rods length is 150 cm. A beryllium converter with length 110 cm is used in order to maximize the energy released in the target. The case of a linear accelerator is considered, with a beam intensity of 1.25‧10¹⁶ p/s, and a total accelerator efficiency of 0.18 for proton beam. These values are planned to be achieved in the European Spallation Source project. The energy gain G is calculated as the ratio between the energy released in the target to the energy spent to accelerate the beam. The energy released is obtained through simulation with the code Geant4. The energy spent is calculating by scaling from the data about the accelerator efficiency for the reference particle (proton). The analysis concerns the G values, the net power produce, the accelerator length, and the period between refueling. The optimal energy for proton is 1.5 GeV. At this energy, G reaches a plateau around a value of 8 and a net power production of 120 MW (megawatt). Starting with alpha, ion beams have a higher G than 1.5 GeV protons. A beam of 0.25 AGeV(gigaelectronvolt per nucleon) ⁷Li realizes the same net power production as 1.5 GeV protons, has a G of 15, and needs an accelerator length 2.6 times lower than for protons, representing the best solution for ADS. Beams of ¹⁶O or ²⁰Ne with energy 0.75 AGeV, accelerated in an accelerator with the same length as 1.5 GeV protons produce approximately 900 MW net power, with a gain of 23-25. The study of the evolution of the isotopes composition during irradiation shows that the increase in power production diminishes the period between refueling. For a net power produced of 120 MW, the target can be irradiated approximately 5000 days without refueling, but only 600 days when the net power reaches 1 GW (gigawatt).Keywords: accelerator driven system, ion beam, electrical power, energy gain
Procedia PDF Downloads 140451 The Meaning in Life and the Content of Mental Images of Temporal Mental Simulations in Poles and Americans
Authors: Katarzyna Pasternak
Abstract:
Experiencing the meaning of life is widely recognised as a vital element of well-being and central human motivation. Studies have shown that a higher meaning of life is associated, among other things, with a higher quality of life, higher levels of happiness and better declared health. The subject of the study is the meaning in life measured with The Meaning in Life Questionnaire and the presence of such emotions as nostalgia, awe and hope, and the content of imaginations measured after temporal mental simulations in Americans and Poles. The respondents had to imagine themselves in future, in 40 years and describe two events that would take place at that time. Next, participants assessed the importance of the events described by them, recognised whether during their journey through time they felt awe, hope and nostalgia, and answered the questionnaire examining the meaning in life. 204 (102 from Poland 102 from the USA ) people aged 21 to 60 participated in the study. The study checked whether there were differences in the content of the imaginations of the respondents from Poland and USA, and whether there were statistically significant difference between the declared sense of meaning in life among participants from both countries. The result of the study hane shown that there were no differences in the overall result obtained by the participants in The Meaning in Life Questionnaire , while there were statistically significant differences among the subscales of the questionnaire. It turned out that Americans have a higher presence of meaning in life than Poles, but they obtained lower results in searching of meaning in life. Studies have also shown that there was a statistically significant difference between Poles and Americans in feeling awe after a mental simulation. Poles felt higher level of awe. Images about the future differed between Poles and Americans. Poles judged that the events they described were very important to them. Interestingly, the content of American participants’ imaginations was dominated by topics related to the future of the world, ecology and world peace. There were also ideas about nice moments spent with friends and family. Among Poles, ideas related to professional career and development as well as family events dominated. Research shows that despite the lack of differences in the general meaning in life, Poles are more focused on searching for meaning in life than Americans. The study shows interesting differences between the two cultures.Keywords: meaning in life, mental simulations, imaginations, temporal mental simulations, future, cultural differences
Procedia PDF Downloads 105450 Modeling Standpipe Pressure Using Multivariable Regression Analysis by Combining Drilling Parameters and a Herschel-Bulkley Model
Authors: Seydou Sinde
Abstract:
The aims of this paper are to formulate mathematical expressions that can be used to estimate the standpipe pressure (SPP). The developed formulas take into account the main factors that, directly or indirectly, affect the behavior of SPP values. Fluid rheology and well hydraulics are some of these essential factors. Mud Plastic viscosity, yield point, flow power, consistency index, flow rate, drillstring, and annular geometries are represented by the frictional pressure (Pf), which is one of the input independent parameters and is calculated, in this paper, using Herschel-Bulkley rheological model. Other input independent parameters include the rate of penetration (ROP), applied load or weight on the bit (WOB), bit revolutions per minute (RPM), bit torque (TRQ), and hole inclination and direction coupled in the hole curvature or dogleg (DL). The technique of repeating parameters and Buckingham PI theorem are used to reduce the number of the input independent parameters into the dimensionless revolutions per minute (RPMd), the dimensionless torque (TRQd), and the dogleg, which is already in the dimensionless form of radians. Multivariable linear and polynomial regression technique using PTC Mathcad Prime 4.0 is used to analyze and determine the exact relationships between the dependent parameter, which is SPP, and the remaining three dimensionless groups. Three models proved sufficiently satisfactory to estimate the standpipe pressure: multivariable linear regression model 1 containing three regression coefficients for vertical wells; multivariable linear regression model 2 containing four regression coefficients for deviated wells; and multivariable polynomial quadratic regression model containing six regression coefficients for both vertical and deviated wells. Although that the linear regression model 2 (with four coefficients) is relatively more complex and contains an additional term over the linear regression model 1 (with three coefficients), the former did not really add significant improvements to the later except for some minor values. Thus, the effect of the hole curvature or dogleg is insignificant and can be omitted from the input independent parameters without significant losses of accuracy. The polynomial quadratic regression model is considered the most accurate model due to its relatively higher accuracy for most of the cases. Data of nine wells from the Middle East were used to run the developed models with satisfactory results provided by all of them, even if the multivariable polynomial quadratic regression model gave the best and most accurate results. Development of these models is useful not only to monitor and predict, with accuracy, the values of SPP but also to early control and check for the integrity of the well hydraulics as well as to take the corrective actions should any unexpected problems appear, such as pipe washouts, jet plugging, excessive mud losses, fluid gains, kicks, etc.Keywords: standpipe, pressure, hydraulics, nondimensionalization, parameters, regression
Procedia PDF Downloads 84449 Applicant Perceptions in Admission Process to Higher Education: The Influence of Social Anxiety
Authors: I. Diamant, R. Srouji
Abstract:
Applicant perceptions are attitudes, feelings, and cognitions which individuals have about selection procedures and have been mostly studied in the context of personnel selection. The main aim of the present study is to expand the understanding of applicant perceptions, using the framework of Organizational Justice Theory, in the domain of selection for higher education. The secondary aim is to explore the relationships between individual differences in social anxiety and applicants’ perceptions. The selection process is an accept/reject situation; it was hypothesized that applicants with higher social anxiety would experience negative perceptions and a lower success estimation, especially when subjected to social interaction elements in the process (interview and group simulation). Also, the effects of prior preparation and post-process explanations offered at the end of the selection process were explored. One hundred sixty psychology M.A. program applicants participated in this research, and following the selection process completed questionnaires measuring social anxiety, social exclusion, ratings on several justice dimensions for each of the methods in the selection process, feelings of success, and self-estimation of compatibility. About half of the applicants also received explanations regarding the significance and the aims of the selection process. Results provided support for most of our hypotheses: applicants with higher social anxiety experienced an increased level of social exclusion in the selection process, perceived the selection as less fair and ended with a lower feeling of success relative to those applicants without social anxiety. These relationships were especially salient in the selection procedures which included social interaction. Additionally, preparation for the selection process was positively related to the favorable perception of fairness in the selection process. Finally, contrary to our hypothesis, it was found that explanations did not affect the applicant’s perceptions. The results enhance our understanding of which factors affect applicant perceptions in applicants to higher education studies and contribute uniquely to the understanding of the effect of social anxiety on different aspects of selection experienced by applicants. The findings clearly show that some individuals may be predisposed to react unfavorably to certain selection situations. In an age of increasing awareness towards fairness in evaluation and selection and hiring procedures, these findings may be of relevance and may contribute to the design of future personnel selection methods in general and of higher education selection in particular.Keywords: applicant perceptions, selection and assessment, organizational justice theory, social anxiety
Procedia PDF Downloads 151448 Identification of Fluorinated Methylsiloxanes in Environmental Matrices Near a Manufacturing Plant in Eastern China
Authors: Liqin Zhi, Lin Xu, Wenxia Wei, Yaqi Cai
Abstract:
Recently, replacing some of the methyl groups in polydimethylsiloxanes with other functional groups has been extensively explored to obtain modified polymethylsiloxanes with special properties that enable new industrial applications. Fluorinated polysiloxanes, one type of these modified polysiloxanes, are based on a siloxane backbone with fluorinated groups attached to the side chains of polysiloxanes. As a commercially significant material, poly[methyl(trifluoropropyl)siloxane] (PMTFPS) has sufficient fluorine content to be useful as a fuel-and oil-resistant elastomer, which combines both the chemical and solvent resistance of fluorocarbons and the wide temperature range applicability of organosilicones. PMTFPS products can be used in many applications in which resistance to fuel, oils and hydrocarbon solvents is required, including use as lubricants in bearings, sealants, and elastomers for aerospace and automotive fuel systems. Fluorinated methylsiloxanes, a type of modified methylsiloxane, include tris(trifluoropropyl)trimethylcyclotrisiloxane (D3F) and tetrakis(trifluoropropyl)tetramethylcyclotetrasiloxane (D4F), both of which contain trifluoropropyl groups in the side chains of cyclic methylsiloxanes. D3F, as an important monomer in the manufacture of PMTFPS, is often present as an impurity in PMTFPS. In addition, the synthesis of PMTFPS from D3F could form other fluorinated methylsiloxanes with low molecular weights (such as D4F). The yearly demand and production volumes of D3F increased rapidly all over world. Fluorinated methylsiloxanes might be released into the environment via different pathways during the production and application of PMTFPS. However, there is a lack of data concerning the emission, environmental occurrence and potential environmental impacts of fluorinated methylsiloxanes. Here, we report fluorinated methylsiloxanes (D3F and D4F) in surface water and sediment samples collected near a fluorinated methylsiloxane manufacturing plant in Weihai, China. The concentrations of D3F and D4F in surface water ranged from 3.29 to 291 ng/L and from 7.02 to 168 ng/L, respectively. The concentrations of D3F and D4F in sediment ranged from 11.8 to 5478 ng/g and from 17.2 to 6277 ng/g, respectively. In simulation experiment, the half-lives of D3F and D4F at different pH values (5.2, 6.4, 7.2, 8.3 and 9.2) varied from 80.6 to 154 h and from 267 to 533 h respectively. CF₃(CH₂)₂MeSi(OH)₂ was identified as one of the main hydrolysis products of fluorinated methylsiloxanes. It was also detected in the river samples at concentrations of 72.1-182.9 ng/L. In addition, the slow rearrangement of D3F (spiked concentration = 500 ng/L) to D4F (concentration = 11.0-22.7 ng/L) was also found during 336h hydrolysis experiment.Keywords: fluorinated methylsiloxanes, environmental matrices, hydrolysis, sediment
Procedia PDF Downloads 115447 Understanding Knowledge, Skills and Competency Needs in Digital Health for Current and Future Health Workforce
Authors: Sisira Edirippulige
Abstract:
Background: Digital health education and training (DHET) is imperative for preparing current and future clinicians to work competently in digitally enabled environments. Despite rapid integration of digital health in modern health services, systematic education and training opportunities for health workers is still lacking. Objectives: This study aimed to investigate healthcare professionals’ perspectives and expectations regarding the knowledge, skills and competency needs in digital health for current and future healthcare workforce. Methods: A qualitative study design with semi-structured individual interviews was employed. A purposive sample method was adopted to collect relevant information from the health workers. Inductive thematic analysis was used to analyse data. Interviews were audio-recorded and transcribed verbatim. Consolidated Criteria for Reporting Qualitative Research (COREQ) was followed when we reported this study. Results: Two themes emerged while analysing the data: (1) what to teach in DHET and (2) how to teach DHET. Overall, healthcare professionals agreed that DHET is important for preparing current and future clinicians for working competently in digitally enabled environments. Knowledge relating to what is digital health, types of digital health, use of technology and human factors in digital health were considered as important to be taught in DHET. Skills relating to digital health consultations, clinical information system management and remote monitoring were considered important to be taught. Blended learning which combined e-learning and classroom-based teaching, simulation sessions and clinical rotations were suggested by healthcare professionals as optimal approaches to deliver the above-mentioned content. Conclusions: This study is the first of its kind to investigate health professionals’ perspectives and expectations relating to the knowledge, skills and competency needs in digital health for current and future healthcare workforce. Healthcare workers are keen to acquire relevant knowledge, skills and competencies related to digital health. Different modes of education delivery is of interest to fit in with busy schedule of health workers.Keywords: digital health, telehealth, telemedicine, education, curriculum
Procedia PDF Downloads 149446 Numerical Simulation of Encased Composite Column Bases Subjected to Cyclic Loading
Authors: Eman Ismail, Adnan Masri
Abstract:
Energy dissipation in ductile moment frames occurs mainly through plastic hinge rotations in its members (beams and columns). Generally, plastic hinge locations are pre-determined and limited to the beam ends, where columns are designed to remain elastic in order to avoid premature instability (aka story mechanisms) with the exception of column bases, where a base is 'fixed' in order to provide higher stiffness and stability and to form a plastic hinge. Plastic hinging at steel column bases in ductile moment frames using conventional base connection details is accompanied by several complications (thicker and heavily stiffened connections, larger embedment depths, thicker foundation to accommodate anchor rod embedment, etc.). An encased composite base connection is proposed where a segment of the column beginning at the base up to a certain point along its height is encased in reinforced concrete with headed shear studs welded to the column flanges used to connect the column to the concrete encasement. When the connection is flexurally loaded, stresses are transferred to a reinforced concrete encasement through the headed shear studs, and thereby transferred to the foundation by reinforced concrete mechanics, and axial column forces are transferred through the base-plate assembly. Horizontal base reactions are expected to be transferred by the direct bearing of the outer and inner faces of the flanges; however, investigation of this mechanism is not within the scope of this research. The inelastic and cyclic behavior of the connection will be investigated where it will be subjected to reversed cyclic loading, and rotational ductility will be observed in cases of yielding mechanisms where yielding occurs as flexural yielding in the beam-column, shear yielding in headed studs, and flexural yielding of the reinforced concrete encasement. The findings of this research show that the connection is capable of achieving satisfactory levels of ductility in certain conditions given proper detailing and proportioning of elements.Keywords: seismic design, plastic mechanisms steel structure, moment frame, composite construction
Procedia PDF Downloads 127445 Computational Analysis of Thermal Degradation in Wind Turbine Spars' Equipotential Bonding Subjected to Lightning Strikes
Authors: Antonio A. M. Laudani, Igor O. Golosnoy, Ole T. Thomsen
Abstract:
Rotor blades of large, modern wind turbines are highly susceptible to downward lightning strikes, as well as to triggering upward lightning; consequently, it is necessary to equip them with an effective lightning protection system (LPS) in order to avoid any damage. The performance of existing LPSs is affected by carbon fibre reinforced polymer (CFRP) structures, which lead to lightning-induced damage in the blades, e.g. via electrical sparks. A solution to prevent internal arcing would be to electrically bond the LPS and the composite structures such that to obtain the same electric potential. Nevertheless, elevated temperatures are achieved at the joint interfaces because of high contact resistance, which melts and vaporises some of the epoxy resin matrix around the bonding. The produced high-pressure gasses open up the bonding and can ignite thermal sparks. The objective of this paper is to predict the current density distribution and the temperature field in the adhesive joint cross-section, in order to check whether the resin pyrolysis temperature is achieved and any damage is expected. The finite element method has been employed to solve both the current and heat transfer problems, which are considered weakly coupled. The mathematical model for electric current includes Maxwell-Ampere equation for induced electric field solved together with current conservation, while the thermal field is found from heat diffusion equation. In this way, the current sub-model calculates Joule heat release for a chosen bonding configuration, whereas the thermal analysis allows to determining threshold values of voltage and current density not to be exceeded in order to maintain the temperature across the joint below the pyrolysis temperature, therefore preventing the occurrence of outgassing. In addition, it provides an indication of the minimal number of bonding points. It is worth to mention that the numerical procedures presented in this study can be tailored and applied to any type of joints other than adhesive ones for wind turbine blades. For instance, they can be applied for lightning protection of aerospace bolted joints. Furthermore, they can even be customized to predict the electromagnetic response under lightning strikes of other wind turbine systems, such as nacelle and hub components.Keywords: carbon fibre reinforced polymer, equipotential bonding, finite element method, FEM, lightning protection system, LPS, wind turbine blades
Procedia PDF Downloads 164444 LTE Modelling of a DC Arc Ignition on Cold Electrodes
Authors: O. Ojeda Mena, Y. Cressault, P. Teulet, J. P. Gonnet, D. F. N. Santos, MD. Cunha, M. S. Benilov
Abstract:
The assumption of plasma in local thermal equilibrium (LTE) is commonly used to perform electric arc simulations for industrial applications. This assumption allows to model the arc using a set of magneto-hydromagnetic equations that can be solved with a computational fluid dynamic code. However, the LTE description is only valid in the arc column, whereas in the regions close to the electrodes the plasma deviates from the LTE state. The importance of these near-electrode regions is non-trivial since they define the energy and current transfer between the arc and the electrodes. Therefore, any accurate modelling of the arc must include a good description of the arc-electrode phenomena. Due to the modelling complexity and computational cost of solving the near-electrode layers, a simplified description of the arc-electrode interaction was developed in a previous work to study a steady high-pressure arc discharge, where the near-electrode regions are introduced at the interface between arc and electrode as boundary conditions. The present work proposes a similar approach to simulate the arc ignition in a free-burning arc configuration following an LTE description of the plasma. To obtain the transient evolution of the arc characteristics, appropriate boundary conditions for both the near-cathode and the near-anode regions are used based on recent publications. The arc-cathode interaction is modeled using a non-linear surface heating approach considering the secondary electron emission. On the other hand, the interaction between the arc and the anode is taken into account by means of the heating voltage approach. From the numerical modelling, three main stages can be identified during the arc ignition. Initially, a glow discharge is observed, where the cold non-thermionic cathode is uniformly heated at its surface and the near-cathode voltage drop is in the order of a few hundred volts. Next, a spot with high temperature is formed at the cathode tip followed by a sudden decrease of the near-cathode voltage drop, marking the glow-to-arc discharge transition. During this stage, the LTE plasma also presents an important increase of the temperature in the region adjacent to the hot spot. Finally, the near-cathode voltage drop stabilizes at a few volts and both the electrode and plasma temperatures reach the steady solution. The results after some seconds are similar to those presented for thermionic cathodes.Keywords: arc-electrode interaction, thermal plasmas, electric arc simulation, cold electrodes
Procedia PDF Downloads 122443 Application of Multilinear Regression Analysis for Prediction of Synthetic Shear Wave Velocity Logs in Upper Assam Basin
Authors: Triveni Gogoi, Rima Chatterjee
Abstract:
Shear wave velocity (Vs) estimation is an important approach in the seismic exploration and characterization of a hydrocarbon reservoir. There are varying methods for prediction of S-wave velocity, if recorded S-wave log is not available. But all the available methods for Vs prediction are empirical mathematical models. Shear wave velocity can be estimated using P-wave velocity by applying Castagna’s equation, which is the most common approach. The constants used in Castagna’s equation vary for different lithologies and geological set-ups. In this study, multiple regression analysis has been used for estimation of S-wave velocity. The EMERGE module from Hampson-Russel software has been used here for generation of S-wave log. Both single attribute and multi attributes analysis have been carried out for generation of synthetic S-wave log in Upper Assam basin. Upper Assam basin situated in North Eastern India is one of the most important petroleum provinces of India. The present study was carried out using four wells of the study area. Out of these wells, S-wave velocity was available for three wells. The main objective of the present study is a prediction of shear wave velocities for wells where S-wave velocity information is not available. The three wells having S-wave velocity were first used to test the reliability of the method and the generated S-wave log was compared with actual S-wave log. Single attribute analysis has been carried out for these three wells within the depth range 1700-2100m, which corresponds to Barail group of Oligocene age. The Barail Group is the main target zone in this study, which is the primary producing reservoir of the basin. A system generated list of attributes with varying degrees of correlation appeared and the attribute with the highest correlation was concerned for the single attribute analysis. Crossplot between the attributes shows the variation of points from line of best fit. The final result of the analysis was compared with the available S-wave log, which shows a good visual fit with a correlation of 72%. Next multi-attribute analysis has been carried out for the same data using all the wells within the same analysis window. A high correlation of 85% has been observed between the output log from the analysis and the recorded S-wave. The almost perfect fit between the synthetic S-wave and the recorded S-wave log validates the reliability of the method. For further authentication, the generated S-wave data from the wells have been tied to the seismic and correlated them. Synthetic share wave log has been generated for the well M2 where S-wave is not available and it shows a good correlation with the seismic. Neutron porosity, density, AI and P-wave velocity are proved to be the most significant variables in this statistical method for S-wave generation. Multilinear regression method thus can be considered as a reliable technique for generation of shear wave velocity log in this study.Keywords: Castagna's equation, multi linear regression, multi attribute analysis, shear wave logs
Procedia PDF Downloads 229442 Algorithms Inspired from Human Behavior Applied to Optimization of a Complex Process
Authors: S. Curteanu, F. Leon, M. Gavrilescu, S. A. Floria
Abstract:
Optimization algorithms inspired from human behavior were applied in this approach, associated with neural networks models. The algorithms belong to human behaviors of learning and cooperation and human competitive behavior classes. For the first class, the main strategies include: random learning, individual learning, and social learning, and the selected algorithms are: simplified human learning optimization (SHLO), social learning optimization (SLO), and teaching-learning based optimization (TLBO). For the second class, the concept of learning is associated with competitiveness, and the selected algorithms are sports-inspired algorithms (with Football Game Algorithm, FGA and Volleyball Premier League, VPL) and Imperialist Competitive Algorithm (ICA). A real process, the synthesis of polyacrylamide-based multicomponent hydrogels, where some parameters are difficult to obtain experimentally, is considered as a case study. Reaction yield and swelling degree are predicted as a function of reaction conditions (acrylamide concentration, initiator concentration, crosslinking agent concentration, temperature, reaction time, and amount of inclusion polymer, which could be starch, poly(vinyl alcohol) or gelatin). The experimental results contain 175 data. Artificial neural networks are obtained in optimal form with biologically inspired algorithm; the optimization being perform at two level: structural and parametric. Feedforward neural networks with one or two hidden layers and no more than 25 neurons in intermediate layers were obtained with values of correlation coefficient in the validation phase over 0.90. The best results were obtained with TLBO algorithm, correlation coefficient being 0.94 for an MLP(6:9:20:2) – a feedforward neural network with two hidden layers and 9 and 20, respectively, intermediate neurons. Good results obtained prove the efficiency of the optimization algorithms. More than the good results, what is important in this approach is the simulation methodology, including neural networks and optimization biologically inspired algorithms, which provide satisfactory results. In addition, the methodology developed in this approach is general and has flexibility so that it can be easily adapted to other processes in association with different types of models.Keywords: artificial neural networks, human behaviors of learning and cooperation, human competitive behavior, optimization algorithms
Procedia PDF Downloads 108441 Inversely Designed Chipless Radio Frequency Identification (RFID) Tags Using Deep Learning
Authors: Madhawa Basnayaka, Jouni Paltakari
Abstract:
Fully passive backscattering chipless RFID tags are an emerging wireless technology with low cost, higher reading distance, and fast automatic identification without human interference, unlike already available technologies like optical barcodes. The design optimization of chipless RFID tags is crucial as it requires replacing integrated chips found in conventional RFID tags with printed geometric designs. These designs enable data encoding and decoding through backscattered electromagnetic (EM) signatures. The applications of chipless RFID tags have been limited due to the constraints of data encoding capacity and the ability to design accurate yet efficient configurations. The traditional approach to accomplishing design parameters for a desired EM response involves iterative adjustment of design parameters and simulating until the desired EM spectrum is achieved. However, traditional numerical simulation methods encounter limitations in optimizing design parameters efficiently due to the speed and resource consumption. In this work, a deep learning neural network (DNN) is utilized to establish a correlation between the EM spectrum and the dimensional parameters of nested centric rings, specifically square and octagonal. The proposed bi-directional DNN has two simultaneously running neural networks, namely spectrum prediction and design parameters prediction. First, spectrum prediction DNN was trained to minimize mean square error (MSE). After the training process was completed, the spectrum prediction DNN was able to accurately predict the EM spectrum according to the input design parameters within a few seconds. Then, the trained spectrum prediction DNN was connected to the design parameters prediction DNN and trained two networks simultaneously. For the first time in chipless tag design, design parameters were predicted accurately after training bi-directional DNN for a desired EM spectrum. The model was evaluated using a randomly generated spectrum and the tag was manufactured using the predicted geometrical parameters. The manufactured tags were successfully tested in the laboratory. The amount of iterative computer simulations has been significantly decreased by this approach. Therefore, highly efficient but ultrafast bi-directional DNN models allow rapid and complicated chipless RFID tag designs.Keywords: artificial intelligence, chipless RFID, deep learning, machine learning
Procedia PDF Downloads 50440 Research on the Optimization of Satellite Mission Scheduling
Authors: Pin-Ling Yin, Dung-Ying Lin
Abstract:
Satellites play an important role in our daily lives, from monitoring the Earth's environment and providing real-time disaster imagery to predicting extreme weather events. As technology advances and demands increase, the tasks undertaken by satellites have become increasingly complex, with more stringent resource management requirements. A common challenge in satellite mission scheduling is the limited availability of resources, including onboard memory, ground station accessibility, and satellite power. In this context, efficiently scheduling and managing the increasingly complex satellite missions under constrained resources has become a critical issue that needs to be addressed. The core of Satellite Onboard Activity Planning (SOAP) lies in optimizing the scheduling of the received tasks, arranging them on a timeline to form an executable onboard mission plan. This study aims to develop an optimization model that considers the various constraints involved in satellite mission scheduling, such as the non-overlapping execution periods for certain types of tasks, the requirement that tasks must fall within the contact range of specified types of ground stations during their execution, onboard memory capacity limits, and the collaborative constraints between different types of tasks. Specifically, this research constructs a mixed-integer programming mathematical model and solves it with a commercial optimization package. Simultaneously, as the problem size increases, the problem becomes more difficult to solve. Therefore, in this study, a heuristic algorithm has been developed to address the challenges of using commercial optimization package as the scale increases. The goal is to effectively plan satellite missions, maximizing the total number of executable tasks while considering task priorities and ensuring that tasks can be completed as early as possible without violating feasibility constraints. To verify the feasibility and effectiveness of the algorithm, test instances of various sizes were generated, and the results were validated through feedback from on-site users and compared against solutions obtained from a commercial optimization package. Numerical results show that the algorithm performs well under various scenarios, consistently meeting user requirements. The satellite mission scheduling algorithm proposed in this study can be flexibly extended to different types of satellite mission demands, achieving optimal resource allocation and enhancing the efficiency and effectiveness of satellite mission execution.Keywords: mixed-integer programming, meta-heuristics, optimization, resource management, satellite mission scheduling
Procedia PDF Downloads 25439 Optimized Scheduling of Domestic Load Based on User Defined Constraints in a Real-Time Tariff Scenario
Authors: Madia Safdar, G. Amjad Hussain, Mashhood Ahmad
Abstract:
One of the major challenges of today’s era is peak demand which causes stress on the transmission lines and also raises the cost of energy generation and ultimately higher electricity bills to the end users, and it was used to be managed by the supply side management. However, nowadays this has been withdrawn because of existence of potential in the demand side management (DSM) having its economic and- environmental advantages. DSM in domestic load can play a vital role in reducing the peak load demand on the network provides a significant cost saving. In this paper the potential of demand response (DR) in reducing the peak load demands and electricity bills to the electric users is elaborated. For this purpose the domestic appliances are modeled in MATLAB Simulink and controlled by a module called energy management controller. The devices are categorized into controllable and uncontrollable loads and are operated according to real-time tariff pricing pattern instead of fixed time pricing or variable pricing. Energy management controller decides the switching instants of the controllable appliances based on the results from optimization algorithms. In GAMS software, the MILP (mixed integer linear programming) algorithm is used for optimization. In different cases, different constraints are used for optimization, considering the comforts, needs and priorities of the end users. Results are compared and the savings in electricity bills are discussed in this paper considering real time pricing and fixed tariff pricing, which exhibits the existence of potential to reduce electricity bills and peak loads in demand side management. It is seen that using real time pricing tariff instead of fixed tariff pricing helps to save in the electricity bills. Moreover the simulation results of the proposed energy management system show that the gained power savings lie in high range. It is anticipated that the result of this research will prove to be highly effective to the utility companies as well as in the improvement of domestic DR.Keywords: controllable and uncontrollable domestic loads, demand response, demand side management, optimization, MILP (mixed integer linear programming)
Procedia PDF Downloads 302438 Regeneration of Geological Models Using Support Vector Machine Assisted by Principal Component Analysis
Authors: H. Jung, N. Kim, B. Kang, J. Choe
Abstract:
History matching is a crucial procedure for predicting reservoir performances and making future decisions. However, it is difficult due to uncertainties of initial reservoir models. Therefore, it is important to have reliable initial models for successful history matching of highly heterogeneous reservoirs such as channel reservoirs. In this paper, we proposed a novel scheme for regenerating geological models using support vector machine (SVM) and principal component analysis (PCA). First, we perform PCA for figuring out main geological characteristics of models. Through the procedure, permeability values of each model are transformed to new parameters by principal components, which have eigenvalues of large magnitude. Secondly, the parameters are projected into two-dimensional plane by multi-dimensional scaling (MDS) based on Euclidean distances. Finally, we train an SVM classifier using 20% models which show the most similar or dissimilar well oil production rates (WOPR) with the true values (10% for each). Then, the other 80% models are classified by trained SVM. We select models on side of low WOPR errors. One hundred channel reservoir models are initially generated by single normal equation simulation. By repeating the classification process, we can select models which have similar geological trend with the true reservoir model. The average field of the selected models is utilized as a probability map for regeneration. Newly generated models can preserve correct channel features and exclude wrong geological properties maintaining suitable uncertainty ranges. History matching with the initial models cannot provide trustworthy results. It fails to find out correct geological features of the true model. However, history matching with the regenerated ensemble offers reliable characterization results by figuring out proper channel trend. Furthermore, it gives dependable prediction of future performances with reduced uncertainties. We propose a novel classification scheme which integrates PCA, MDS, and SVM for regenerating reservoir models. The scheme can easily sort out reliable models which have similar channel trend with the reference in lowered dimension space.Keywords: history matching, principal component analysis, reservoir modelling, support vector machine
Procedia PDF Downloads 160437 Nurse-Led Codes: Practical Application in the Emergency Department during a Global Pandemic
Authors: F. DelGaudio, H. Gill
Abstract:
Resuscitation during cardiopulmonary (CPA) arrest is dynamic, high stress, high acuity situation, which can easily lead to communication breakdown, and errors. The care of these high acuity patients has also been shown to increase physiologic stress and task saturation of providers, which can negatively impact the care being provided. These difficulties are further complicated during a global pandemic and pose a significant safety risk to bedside providers. Nurse-led codes are a relatively new concept that may be a potential solution for alleviating some of these difficulties. An experienced nurse who has completed advanced cardiac life support (ACLS), and additional training, assumed the responsibility of directing the mechanics of the appropriate ACLS algorithm. This was done in conjunction with a physician who also acted as a physician leader. The additional nurse-led code training included a multi-disciplinary in situ simulation of a CPA on a suspected COVID-19 patient. During the CPA, the nurse leader’s responsibilities include: ensuring adequate compression depth and rate, minimizing interruptions in chest compressions, the timing of rhythm/pulse checks, and appropriate medication administration. In addition, the nurse leader also functions as a last line safety check for appropriate personal protective equipment and limiting exposure of staff. The use of nurse-led codes for CPA has shown to decrease the cognitive overload and task saturation for the physician, as well as limiting the number of staff being exposed to a potentially infectious patient. The real-world application has allowed physicians to perform and oversee high-risk procedures such as intubation, line placement, and point of care ultrasound, without sacrificing the integrity of the resuscitation. Nurse-led codes have also given the physician the bandwidth to review pertinent medical history, advanced directives, determine reversible causes, and have the end of life conversations with family. While there is a paucity of research on the effectiveness of nurse-led codes, there are many potentially significant benefits. In addition to its value during a pandemic, it may also be beneficial during complex circumstances such as extracorporeal cardiopulmonary resuscitation.Keywords: cardiopulmonary arrest, COVID-19, nurse-led code, task saturation
Procedia PDF Downloads 155436 An Intelligent Text Independent Speaker Identification Using VQ-GMM Model Based Multiple Classifier System
Authors: Ben Soltane Cheima, Ittansa Yonas Kelbesa
Abstract:
Speaker Identification (SI) is the task of establishing identity of an individual based on his/her voice characteristics. The SI task is typically achieved by two-stage signal processing: training and testing. The training process calculates speaker specific feature parameters from the speech and generates speaker models accordingly. In the testing phase, speech samples from unknown speakers are compared with the models and classified. Even though performance of speaker identification systems has improved due to recent advances in speech processing techniques, there is still need of improvement. In this paper, a Closed-Set Tex-Independent Speaker Identification System (CISI) based on a Multiple Classifier System (MCS) is proposed, using Mel Frequency Cepstrum Coefficient (MFCC) as feature extraction and suitable combination of vector quantization (VQ) and Gaussian Mixture Model (GMM) together with Expectation Maximization algorithm (EM) for speaker modeling. The use of Voice Activity Detector (VAD) with a hybrid approach based on Short Time Energy (STE) and Statistical Modeling of Background Noise in the pre-processing step of the feature extraction yields a better and more robust automatic speaker identification system. Also investigation of Linde-Buzo-Gray (LBG) clustering algorithm for initialization of GMM, for estimating the underlying parameters, in the EM step improved the convergence rate and systems performance. It also uses relative index as confidence measures in case of contradiction in identification process by GMM and VQ as well. Simulation results carried out on voxforge.org speech database using MATLAB highlight the efficacy of the proposed method compared to earlier work.Keywords: feature extraction, speaker modeling, feature matching, Mel frequency cepstrum coefficient (MFCC), Gaussian mixture model (GMM), vector quantization (VQ), Linde-Buzo-Gray (LBG), expectation maximization (EM), pre-processing, voice activity detection (VAD), short time energy (STE), background noise statistical modeling, closed-set tex-independent speaker identification system (CISI)
Procedia PDF Downloads 309435 Construction of a Dynamic Model of Cerebral Blood Circulation for Future Integrated Control of Brain State
Authors: Tomohiko Utsuki
Abstract:
Currently, brain resuscitation becomes increasingly important due to revising various clinical guidelines pertinent to emergency care. In brain resuscitation, the control of brain temperature (BT), intracranial pressure (ICP), and cerebral blood flow (CBF) is required for stabilizing physiological state of brain, and is described as the essential treatment points in many guidelines of disorder and/or disease such as brain injury, stroke, and encephalopathy. Thus, an integrated control system of BT, ICP, and CBF will greatly contribute to alleviating the burden on medical staff and improving treatment effect in brain resuscitation. In order to develop such a control system, models related to BT, ICP, and CBF are required for control simulation, because trial and error experiments using patients are not ethically allowed. A static model of cerebral blood circulation from intracranial arteries and vertebral artery to jugular veins has already constructed and verified. However, it is impossible to represent the pooling of blood in blood vessels, which is one cause of cerebral hypertension in this model. And, it is also impossible to represent the pulsing motion of blood vessels caused by blood pressure change which can have an affect on the change of cerebral tissue pressure. Thus, a dynamic model of cerebral blood circulation is constructed in consideration of the elasticity of the blood vessel and the inertia of the blood vessel wall. The constructed dynamic model was numerically analyzed using the normal data, in which each arterial blood flow in cerebral blood circulation, the distribution of blood pressure in the Circle of Willis, and the change of blood pressure along blood flow were calculated for verifying against physiological knowledge. As the result, because each calculated numerical value falling within the generally known normal range, this model has no problem in representing at least the normal physiological state of the brain. It is the next task to verify the accuracy of the present model in the case of disease or disorder. Currently, the construction of a migration model of extracellular fluid and a model of heat transfer in cerebral tissue are in progress for making them parts of an integrated model of brain physiological state, which is necessary for developing an future integrated control system of BT, ICP and CBF. The present model is applicable to constructing the integrated model representing at least the normal condition of brain physiological state by uniting with such models.Keywords: dynamic model, cerebral blood circulation, brain resuscitation, automatic control
Procedia PDF Downloads 153434 The Development of a Precision Irrigation System for Durian
Authors: Chatrabhuti Pipop, Visessri Supattra, Charinpanitkul Tawatchai
Abstract:
Durian is one of the top agricultural products exported by Thailand. There is the massive market potential for the durian industry. While the global demand for Thai durians, especially the demand from China, is very high, Thailand's durian supply is far from satisfying strong demand. Poor agricultural practices result in low yields and poor quality of fruit. Most irrigation systems currently used by the farmers are fixed schedule or fixed rates that ignore actual weather conditions and crop water requirements. In addition, the technologies emerging are too difficult and complex and prices are too high for the farmers to adopt and afford. Many farmers leave the durian trees to grow naturally. With improper irrigation and nutrient management system, durians are vulnerable to a variety of issues, including stunted growth, not flowering, diseases, and death. Technical development or research for durian is much needed to support the wellbeing of the farmers and the economic development of the country. However, there are a limited number of studies or development projects for durian because durian is a perennial crop requiring a long time to obtain the results to report. This study, therefore, aims to address the problem of durian production by developing an autonomous and precision irrigation system. The system is designed and equipped with an industrial programmable controller, a weather station, and a digital flow meter. Daily water requirements are computed based on weather data such as rainfall and evapotranspiration for daily irrigation with variable flow rates. A prediction model is also designed as a part of the system to enhance the irrigation schedule. Before the system was installed in the field, a simulation model was built and tested in a laboratory setting to ensure its accuracy. Water consumption was measured daily before and after the experiment for further analysis. With this system, the crop water requirement is precisely estimated and optimized based on the data from the weather station. Durian will be irrigated at the right amount and at the right time, offering the opportunity for higher yield and higher income to the farmers.Keywords: Durian, precision irrigation, precision agriculture, smart farm
Procedia PDF Downloads 118433 Multi-Omics Integrative Analysis Coupled to Control Theory and Computational Simulation of a Genome-Scale Metabolic Model Reveal Controlling Biological Switches in Human Astrocytes under Palmitic Acid-Induced Lipotoxicity
Authors: Janneth Gonzalez, Andrés Pinzon Velasco, Maria Angarita
Abstract:
Astrocytes play an important role in various processes in the brain, including pathological conditions such as neurodegenerative diseases. Recent studies have shown that the increase in saturated fatty acids such as palmitic acid (PA) triggers pro-inflammatorypathways in the brain. The use of synthetic neurosteroids such as tibolone has demonstrated neuro-protective mechanisms. However, broad studies with a systemic point of view on the neurodegenerative role of PA and the neuro-protective mechanisms of tibolone are lacking. In this study, we performed the integration of multi-omic data (transcriptome and proteome) into a human astrocyte genomic scale metabolic model to study the astrocytic response during palmitate treatment. We evaluated metabolic fluxes in three scenarios (healthy, induced inflammation by PA, and tibolone treatment under PA inflammation). We also applied a control theory approach to identify those reactions that exert more control in the astrocytic system. Our results suggest that PA generates a modulation of central and secondary metabolism, showing a switch in energy source use through inhibition of folate cycle and fatty acid β‐oxidation and upregulation of ketone bodies formation. We found 25 metabolic switches under PA‐mediated cellular regulation, 9 of which were critical only in the inflammatory scenario but not in the protective tibolone one. Within these reactions, inhibitory, total, and directional coupling profiles were key findings, playing a fundamental role in the (de)regulation of metabolic pathways that may increase neurotoxicity and represent potential treatment targets. Finally, the overall framework of our approach facilitates the understanding of complex metabolic regulation, and it can be used for in silico exploration of the mechanisms of astrocytic cell regulation, directing a more complex future experimental work in neurodegenerative diseases.Keywords: astrocytes, data integration, palmitic acid, computational model, multi-omics
Procedia PDF Downloads 97432 A Simulated Evaluation of Model Predictive Control
Authors: Ahmed AlNouss, Salim Ahmed
Abstract:
Process control refers to the techniques to control the variables in a process in order to maintain them at their desired values. Advanced process control (APC) is a broad term within the domain of control where it refers to different kinds of process control and control related tools, for example, model predictive control (MPC), statistical process control (SPC), fault detection and classification (FDC) and performance assessment. APC is often used for solving multivariable control problems and model predictive control (MPC) is one of only a few advanced control methods used successfully in industrial control applications. Advanced control is expected to bring many benefits to the plant operation; however, the extent of the benefits is plant specific and the application needs a large investment. This requires an analysis of the expected benefits before the implementation of the control. In a real plant simulation studies are carried out along with some experimentation to determine the improvement in the performance of the plant due to advanced control. In this research, such an exercise is undertaken to realize the needs of APC application. The main objectives of the paper are as follows: (1) To apply MPC to a number of simulations set up to realize the need of MPC by comparing its performance with that of proportional integral derivatives (PID) controllers. (2) To study the effect of controller parameters on control performance. (3) To develop appropriate performance index (PI) to compare the performance of different controller and develop novel idea to present tuning map of a controller. These objectives were achieved by applying PID controller and a special type of MPC which is dynamic matrix control (DMC) on the multi-tanks process simulated in loop-pro. Then the controller performance has been evaluated by changing the controller parameters. This performance was based on special indices related to the difference between set point and process variable in order to compare the both controllers. The same principle was applied for continuous stirred tank heater (CSTH) and continuous stirred tank reactor (CSTR) processes simulated in Matlab. However, in these processes some developed programs were written to evaluate the performance of the PID and MPC controllers. Finally these performance indices along with their controller parameters were plotted using special program called Sigmaplot. As a result, the improvement in the performance of the control loops was quantified using relevant indices to justify the need and importance of advanced process control. Also, it has been approved that, by using appropriate indices, predictive controller can improve the performance of the control loop significantly.Keywords: advanced process control (APC), control loop, model predictive control (MPC), proportional integral derivatives (PID), performance indices (PI)
Procedia PDF Downloads 407431 Risk and Emotion: Measuring the Effect of Emotion and Other Visceral Factors on Decision Making under Risk
Authors: Michael Mihalicz, Aziz Guergachi
Abstract:
Background: The science of modelling choice preferences has evolved over centuries into an interdisciplinary field contributing to several branches of Microeconomics and Mathematical Psychology. Early theories in Decision Science rested on the logic of rationality, but as it and related fields matured, descriptive theories emerged capable of explaining systematic violations of rationality through cognitive mechanisms underlying the thought processes that guide human behaviour. Cognitive limitations are not, however, solely responsible for systematic deviations from rationality and many are now exploring the effect of visceral factors as the more dominant drivers. The current study builds on the existing literature by exploring sleep deprivation, thermal comfort, stress, hunger, fear, anger and sadness as moderators to three distinct elements that define individual risk preference under Cumulative Prospect Theory. Methodology: This study is designed to compare the risk preference of participants experiencing an elevated affective or visceral state to those in a neutral state using nonparametric elicitation methods across three domains. Two experiments will be conducted simultaneously using different methodologies. The first will determine visceral states and risk preferences randomly over a two-week period by prompting participants to complete an online survey remotely. In each round of questions, participants will be asked to self-assess their current state using Visual Analogue Scales before answering a series of lottery-style elicitation questions. The second experiment will be conducted in a laboratory setting using psychological primes to induce a desired state. In this experiment, emotional states will be recorded using emotion analytics and used a basis for comparison between the two methods. Significance: The expected results include a series of measurable and systematic effects on the subjective interpretations of gamble attributes and evidence supporting the proposition that a portion of the variability in human choice preferences unaccounted for by cognitive limitations can be explained by interacting visceral states. Significant results will promote awareness about the subconscious effect that emotions and other drive states have on the way people process and interpret information, and can guide more effective decision making by informing decision-makers of the sources and consequences of irrational behaviour.Keywords: decision making, emotions, prospect theory, visceral factors
Procedia PDF Downloads 149430 Electromagnetic Energy Harvesting by Using a Rectenna with a Metamaterial Lens
Authors: Ursula D. C. Resende, Fabiano S. Bicalho, Sandro T. M. Gonçalves
Abstract:
The growing demand for cheap and clean energy sources have been motivated by the study and development of distinct technologies and devices able to provide different amounts of energy. In order to supply energy for small loads, the energy from the electromagnetic spectrum can be harvested. This possibility is particularly interesting because this kind of energy is constantly available in the environment and the number of radiofrequency sources is permanently increasing, due to advances in telecommunications services. A rectenna, which is a combination of an antenna and a rectifier circuit, is an equipment that can efficiently perform the electromagnetic energy harvesting. However, since the amount of electromagnetic energy available in the environment is very small, limited values of power can be harvested by the rectenna. Therefore, several technical strategies have been investigated in order to increase this amount of power. In this work, a metamaterial electromagnetic lens is used to improve the electromagnetic energy harvesting. The rectenna investigated was designed and optimized to charge a Li-Ion battery using the electromagnetic energy from an internet Wi-Fi commercial router model TL-WR841HP operating in 2.45 GHz with maximal output power equal to 18 dBm. The rectenna consists of a high directive antenna, a double voltage rectifier circuit and a metamaterial lens. The printed antenna, constituted of two rectangular radiator elements, was projected and optimized by using the Computer Simulation Software (CST) in order to obtain high directivities and values of S11 parameter below -10 dB in 2.45 GHz. The antenna was printed over a double-sided copper fiberglass substrate, FR4, with characterized relative electric permittivity εr = 4.3 and tangent of losses δ = 0.01. The rectifier circuit, which incorporates a circuit for impedance matching and uses the Schottky diode HSMS-2852, was projected and optimized by using Advanced Design Software (ADS) and built over the same FR4 substrate. The metamaterial cell is composed of two Square Split Ring Resonator (S-SRR) and a thin wire in order to operate with negative values of εr and relative magnetic permeability in 2.45 GHz. In order to evaluate the performance of the purposed rectenna two experimental charging tests were performed, one without and other with the metamaterial lens. The result obtained demonstrate that the electromagnetic lens was able to significantly increase the levels of electric current delivered to the battery, approximately 44%.Keywords: electromagnetic energy harvesting, electromagnetic lens, metamaterial, rectenna
Procedia PDF Downloads 143429 The Effects of Native Forests Conservation and Preservation Scenarios on Two Chilean Basins Water Cycle, under Climate Change Conditions
Authors: Hernández Marieta, Aguayo Mauricio, Pedreros María, Llompart Ovidio
Abstract:
The hydrological cycle is influenced by multiple factors, including climate change, land use changes, and anthropogenic activities, all of which threaten water availability and quality worldwide. In recent decades, numerous investigations have used landscape metrics and hydrological modeling to demonstrate the influence of landscape patterns on the hydrological cycle components' natural dynamics. Many of these investigations have determined the repercussions on the quality and availability of water, sedimentation, and erosion regime, mainly in Asian basins. In fact, there is progress in this branch of science, but there are still unanswered questions for our region. This study examines the hydrological response in Chilean basins under various land use change scenarios (LUCC) and the influence of climate change. The components of the water cycle were modeled using a physically distributed type hydrological and hydraulic simulation model based on and oriented to mountain basins TETIS model. Future climate data were derived from Chilean regional simulations using the WRF-MIROC5 model, forced with the RCP 8.5 scenario, at a 25 km resolution for the periods 2030-2060 and 2061-2091. LUCC scenarios were designed based on nature-based solutions, landscape pattern influences, current national and international water conservation legislation, and extreme scenarios of non-preservation and conservation of native forests. The scenarios that demonstrate greater water availability, even under climate change, are those promoting the restoration of native forests in over 30% of the basins, even alongside agricultural activities. Current legislation promoting the restoration of native forests only in riparian zones (30-60 m or 200 m in steeper areas) will not be resilient enough to address future water shortages. Evapotranspiration, direct runoff, and water availability at basin outlets showed the greatest variations due to LUCC. The relationship between hydrological modeling and landscape configuration is an effective tool for establishing future territorial planning that prioritizes water resource protection.Keywords: TETIS, landscape pattern, hydrological process, water availability, Chilean basins
Procedia PDF Downloads 36428 Computer-Aided Drug Repurposing for Mycobacterium Tuberculosis by Targeting Tryptophanyl-tRNA Synthetase
Authors: Neslihan Demirci, Serdar Durdağı
Abstract:
Mycobacterium tuberculosis is still a worldwide disease-causing agent that, according to WHO, led to the death of 1.5 million people from tuberculosis (TB) in 2020. The bacteria reside in macrophages located specifically in the lung. There is a known quadruple drug therapy regimen for TB consisting of isoniazid (INH), rifampin (RIF), pyrazinamide (PZA), and ethambutol (EMB). Over the past 60 years, there have been great contributions to treatment options, such as recently approved delamanid (OPC67683) and bedaquiline (TMC207/R207910), targeting mycolic acid and ATP synthesis, respectively. Also, there are natural compounds that can block the tryptophanyl-tRNA synthetase (TrpRS) enzyme, chuangxinmycin, and indolmycin. Yet, already the drug resistance is reported for those agents. In this study, the newly released TrpRS enzyme structure is investigated for potential inhibitor drugs from already synthesized molecules to help the treatment of resistant cases and to propose an alternative drug for the quadruple drug therapy of tuberculosis. Maestro, Schrodinger is used for docking and molecular dynamic simulations. In-house library containing ~8000 compounds among FDA-approved indole-containing compounds, a total of 57 obtained from the ChemBL were used for both ATP and tryptophan binding pocket docking. Best of indole-containing 57 compounds were subjected to hit expansion and compared later with virtual screening workflow (VSW) results. After docking, VSW was done. Glide-XP docking algorithm was chosen. When compared, VSW alone performed better than the hit expansion module. Best scored compounds were kept for ten ns molecular dynamic simulations by Desmond. Further, 100 ns molecular dynamic simulation was performed for elected molecules according to Z-score. The top three MMGBSA-scored compounds were subjected to steered molecular dynamic (SMD) simulations by Gromacs. While SMD simulations are still being conducted, ponesimod (for multiple sclerosis), vilanterol (β₂ adrenoreceptor agonist), and silodosin (for benign prostatic hyperplasia) were found to have a significant affinity for tuberculosis TrpRS, which is the propulsive force for the urge to expand the research with in vitro studies. Interestingly, top-scored ponesimod has been reported to have a side effect that makes the patient prone to upper respiratory tract infections.Keywords: drug repurposing, molecular dynamics, tryptophanyl-tRNA synthetase, tuberculosis
Procedia PDF Downloads 123427 A Risk-Based Approach to Construction Management
Authors: Chloe E. Edwards, Yasaman Shahtaheri
Abstract:
Risk management plays a fundamental role in project planning and delivery. The purpose of incorporating risk management into project management practices is to identify and address uncertainties related to key project-related activities. The uncertainties, known as risk events, can relate to project deliverables that are quantifiable and are often measured by impact to project schedule, cost, or environmental impact. Risk management should be incorporated as an iterative practice throughout the planning, execution, and commissioning phases of a project. This paper specifically examines how risk management contributes to effective project planning and delivery through a case study of a transportation project. This case study focused solely on impacts to project schedule regarding three milestones: readiness for delivery, readiness for testing and commissioning, and completion of the facility. The case study followed the ISO 31000: Risk Management – Guidelines. The key factors that are outlined by these guidelines include understanding the scope and context of the project, conducting a risk assessment including identification, analysis, and evaluation, and lastly, risk treatment through mitigation measures. This process requires continuous consultation with subject matter experts and monitoring to iteratively update the risks accordingly. The risk identification process led to a total of fourteen risks related to design, permitting, construction, and commissioning. The analysis involved running 1,000 Monte Carlo simulations through @RISK 8.0 Industrial software to determine potential milestone completion dates based on the project baseline schedule. These dates include the best case, most likely case, and worst case to provide an estimated delay for each milestone. Evaluation of these results provided insight into which risks were the highest contributors to the projected milestone completion dates. Based on the analysis results, the risk management team was able to provide recommendations for mitigation measures to reduce the likelihood of risks occurring. The risk management team also provided recommendations for managing the identified risks and project activities moving forward to meet the most likely or best-case milestone completion dates.Keywords: construction management, monte carlo simulation, project delivery, risk assessment, transportation engineering
Procedia PDF Downloads 107426 An Initiative for Improving Pre-Service Teachers’ Pedagogical Content Knowledge in Mathematics
Authors: Taik Kim
Abstract:
Mathematics anxiety has an important consequence for teacher practices that influence students’ attitudes and achievement. Elementary prospective teachers have the highest levels of mathematics anxiety in comparison with other college majors. In his teaching practice, the researcher developed a highly successful teaching model to reduce pre-service teachers’ higher math anxiety and simultaneously to improve their pedagogical math content knowledge. There were eighty one participants from 2015 to 2018 who took the Mathematics for Elementary Teachers I and II. As the analysis data indicated, elementary prospective teachers’ math anxiety was greatly reduced with improving their math pedagogical knowledge. U.S encounters a critical shortage of well qualified educators. To solve the issue, it is essential to engage students in a long-term commitmentto shape better teachers, who will, in turn, produce k-12 school students that are better-prepared for college students. It is imperative that new instructional strategies are implemented to improve student learning and address declining interest, poor preparedness, a lack of diverse representation, and low persistence of students in mathematics. Many four year college students take math courses from the math department in the College of Arts& Science and then take methodology courses from the College of Education. Before taking pedagogy, many students struggle in learning mathematics and lose their confidence. Since the content course focus on college level math, instead of pre service teachers’ teaching area, per se elementary math, they do not have a chance to improve their teaching skills on topics which eventually they teach. The research, a joint appointment of math and math education, has been involved in teaching content and pedagogy. As the result indicated, participants were able to math content at the same time how to teach. In conclusion, the new initiative to use several teaching strategies was able not only to increase elementary prospective teachers’ mathematical skills and knowledge but also to improve their attitude toward mathematics. We need an innovative teaching strategy which implements evidence-based tactics in redesigning a education and math to improve pre service teachers’math skills and which can improve students’ attitude toward math and students’ logical and reasoning skills. Implementation of these best practices in the local school district is particularly important because K-8 teachers are not generally familiar with lab-based instruction. At the same time, local school teachers will learn a new way how to teach math. This study can be a vital teacher education model expanding throughout the State and nationwide. In summary, this study yields invaluable information how to improve teacher education in the elementary level and, eventually, how to enhance K-8 students’ math achievement.Keywords: quality of education and improvement method, teacher education, innovative teaching and learning methodologies, math education
Procedia PDF Downloads 104