Search results for: time varying covariate
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18934

Search results for: time varying covariate

18844 Overview of Environmental and Economic Theories of the Impact of Dams in Different Regions

Authors: Ariadne Katsouras, Andrea Chareunsy

Abstract:

The number of large hydroelectric dams in the world has increased from almost 6,000 in the 1950s to over 45,000 in 2000. Dams are often built to increase the economic development of a country. This can occur in several ways. Large dams take many years to build so the construction process employs many people for a long time and that increased production and income can flow on into other sectors of the economy. Additionally, the provision of electricity can help raise people’s living standards and if the electricity is sold to another country then the money can be used to provide other public goods for the residents of the country that own the dam. Dams are also built to control flooding and provide irrigation water. Most dams are of these types. This paper will give an overview of the environmental and economic theories of the impact of dams in different regions of the world. There is a difference in the degree of environmental and economic impacts due to the varying climates and varying social and political factors of the regions. Production of greenhouse gases from the dam’s reservoir, for instance, tends to be higher in tropical areas as opposed to Nordic environments. However, there are also common impacts due to construction of the dam itself, such as, flooding of land for the creation of the reservoir and displacement of local populations. Economically, the local population tends to benefit least from the construction of the dam. Additionally, if a foreign company owns the dam or the government subsidises the cost of electricity to businesses, then the funds from electricity production do not benefit the residents of the country the dam is built in. So, in the end, the dams can benefit a country economically, but the varying factors related to its construction and how these are dealt with, determine the level of benefit, if any, of the dam. Some of the theories or practices used to evaluate the potential value of a dam include cost-benefit analysis, environmental impacts assessments and regressions. Systems analysis is also a useful method. While these theories have value, there are also possible shortcomings. Cost-benefit analysis converts all the costs and benefits to dollar values, which can be problematic. Environmental impact assessments, likewise, can be incomplete, especially if the assessment does not include feedback effects, that is, they only consider the initial impact. Finally, regression analysis is dependent on the available data and again would not necessarily include feedbacks. Systems analysis is a method that can allow more complex modelling of the environment and the economic system. It would allow a clearer picture to emerge of the impacts and can include a long time frame.

Keywords: comparison, economics, environment, hydroelectric dams

Procedia PDF Downloads 168
18843 Face Recognition Using Eigen Faces Algorithm

Authors: Shweta Pinjarkar, Shrutika Yawale, Mayuri Patil, Reshma Adagale

Abstract:

Face recognition is the technique which can be applied to the wide variety of problems like image and film processing, human computer interaction, criminal identification etc. This has motivated researchers to develop computational models to identify the faces, which are easy and simple to implement. In this, demonstrates the face recognition system in android device using eigenface. The system can be used as the base for the development of the recognition of human identity. Test images and training images are taken directly with the camera in android device.The test results showed that the system produces high accuracy. The goal is to implement model for particular face and distinguish it with large number of stored faces. face recognition system detects the faces in picture taken by web camera or digital camera and these images then checked with training images dataset based on descriptive features. Further this algorithm can be extended to recognize the facial expressions of a person.recognition could be carried out under widely varying conditions like frontal view,scaled frontal view subjects with spectacles. The algorithm models the real time varying lightning conditions. The implemented system is able to perform real-time face detection, face recognition and can give feedback giving a window with the subject's info from database and sending an e-mail notification to interested institutions using android application. Face recognition is the technique which can be applied to the wide variety of problems like image and film processing, human computer interaction, criminal identification etc. This has motivated researchers to develop computational models to identify the faces, which are easy and simple to implement. In this , demonstrates the face recognition system in android device using eigenface. The system can be used as the base for the development of the recognition of human identity. Test images and training images are taken directly with the camera in android device.The test results showed that the system produces high accuracy. The goal is to implement model for particular face and distinguish it with large number of stored faces. face recognition system detects the faces in picture taken by web camera or digital camera and these images then checked with training images dataset based on descriptive features. Further this algorithm can be extended to recognize the facial expressions of a person.recognition could be carried out under widely varying conditions like frontal view,scaled frontal view subjects with spectacles. The algorithm models the real time varying lightning conditions. The implemented system is able to perform real-time face detection, face recognition and can give feedback giving a window with the subject's info from database and sending an e-mail notification to interested institutions using android application.

Keywords: face detection, face recognition, eigen faces, algorithm

Procedia PDF Downloads 335
18842 Contagion and Stock Interdependence in the BRIC+M Block

Authors: Christian Bucio Pacheco, Miriam Magnolia Sosa Castro, María Alejandra Cabello Rosales

Abstract:

This paper aims to analyze the contagion effect among the stock markets of the BRIC+M block (Brazil, Russia, India, China plus Mexico). The contagion effect is proved through increasing on dependence parameters during crisis periods. The dependence parameters are estimated through copula approach in a period of time from July 1997 to December 2015. During this period there are instability and calm episodes, allowing to analyze changes in the relations of dependence. Empirical results show strong evidence of time-varying dependence among the BRIC+M markets and an increasing dependence relation during global financial crisis period.

Keywords: BRIC+M Block, Contagion effect, Copula, dependence

Procedia PDF Downloads 316
18841 Failure Analysis of Fractured Dental Implants

Authors: Rajesh Bansal, Amit Raj Sharma, Vakil Singh

Abstract:

The success and predictability of titanium implants for long durations are well established and there has been a tremendous increase in the popularity of implants among patients as well as clinicians over the last four decades. However, sometimes complications arise, which lead to the loss of the implant as well as the prosthesis. Fracture of dental implants is rare; however, at times, implants or abutment screws fracture and lead to many problems for the clinician and the patient. Possible causes of implant fracture include improper design, overload, fatigue and corrosion. Six retrieved fractured dental implants, with varying diameters and designs, were collected from time to time to examine by scanning electron microscope (SEM) to characterize fracture behavior and assess the mechanism of fracture. In this investigation, it was observed that fracture of the five dental implants occurred due to fatigue crack initiation and propagation from the thread roots.

Keywords: titanium, dental, implant, fracture, failure

Procedia PDF Downloads 45
18840 Effects of Strain-Induced Melt Activation Process on the Structure and Morphology Mg₂Si in Al-15%Mg₂Si Composite

Authors: Reza Eslami-Farsani, Mohammad Alipour

Abstract:

The effect of deformation on the semisolid microstructure and degree of globularity of Al–15%Mg₂Si composite produced by the strain induced melt activation (SIMA) process was studied. Deformation of 25% was used. After deformation, the samples were heated to a temperature above the solidus and below the liquidus point and maintained in the isothermal conditions at three different temperatures (560, 580 and 595 °C) for varying time (5, 10, 20 and 40 min). The microstructural study was carried out on the alloy by the use of optical microscopy. It was observed that strain induced deformation and subsequently melt activation has caused the globular morphology of Mg₂Si particles. The results showed that for the desired microstructures of the alloy during SIMA process, the optimum temperature and time are 595 °C and 40 min respectively.

Keywords: deformation, semisolid, SIMA, Mg₂Si phase, modification

Procedia PDF Downloads 246
18839 Preparation and Quality Control of a Novel Radiolabeled Complex of 166ho for the Treatment of Somatostatin Receptor Expressing Tumours

Authors: H. Yousefnia, A. Golabi Dezfuli, S. Zolghadri, M. Hosntalab

Abstract:

Peptide receptor radionuclide therapy is nowadays used for the treatment of various abnormalities with somatostatin receptors. In this study, 166Ho-DOTATOC was prepared and the best conditions for its radiolabeling was obtained. For this purpose, a certain of DOTATOC was added to a vial containing 166Ho. various experiments by varying ligand concentration, pH, temperature and time were performed to determine the best conditions. Radiochemical purity of the complex was assessed by instant thin layer chromatography method utilizing 0.9% NaCl as the mobile phase. 166Ho-DOTATOC was prepared with radiochemical purity of higher than 95% at the optimized condition (pH=4, temperature: 95° C, time:30 min). In 0.9% NaCl, free Ho cation was developed at Rf of 0.8 while the complex was remained at the front of the paper.

Keywords: Ho-166, neuroendocrine, octreotide, quality control

Procedia PDF Downloads 355
18838 Calculation of Pressure-Varying Langmuir and Brunauer-Emmett-Teller Isotherm Adsorption Parameters

Authors: Trevor C. Brown, David J. Miron

Abstract:

Gas-solid physical adsorption methods are central to the characterization and optimization of the effective surface area, pore size and porosity for applications such as heterogeneous catalysis, and gas separation and storage. Properties such as adsorption uptake, capacity, equilibrium constants and Gibbs free energy are dependent on the composition and structure of both the gas and the adsorbent. However, challenges remain, in accurately calculating these properties from experimental data. Gas adsorption experiments involve measuring the amounts of gas adsorbed over a range of pressures under isothermal conditions. Various constant-parameter models, such as Langmuir and Brunauer-Emmett-Teller (BET) theories are used to provide information on adsorbate and adsorbent properties from the isotherm data. These models typically do not provide accurate interpretations across the full range of pressures and temperatures. The Langmuir adsorption isotherm is a simple approximation for modelling equilibrium adsorption data and has been effective in estimating surface areas and catalytic rate laws, particularly for high surface area solids. The Langmuir isotherm assumes the systematic filling of identical adsorption sites to a monolayer coverage. The BET model is based on the Langmuir isotherm and allows for the formation of multiple layers. These additional layers do not interact with the first layer and the energetics are equal to the adsorbate as a bulk liquid. This BET method is widely used to measure the specific surface area of materials. Both Langmuir and BET models assume that the affinity of the gas for all adsorption sites are identical and so the calculated adsorbent uptake at the monolayer and equilibrium constant are independent of coverage and pressure. Accurate representations of adsorption data have been achieved by extending the Langmuir and BET models to include pressure-varying uptake capacities and equilibrium constants. These parameters are determined using a novel regression technique called flexible least squares for time-varying linear regression. For isothermal adsorption the adsorption parameters are assumed to vary slowly and smoothly with increasing pressure. The flexible least squares for pressure-varying linear regression (FLS-PVLR) approach assumes two distinct types of discrepancy terms, dynamic and measurement for all parameters in the linear equation used to simulate the data. Dynamic terms account for pressure variation in successive parameter vectors, and measurement terms account for differences between observed and theoretically predicted outcomes via linear regression. The resultant pressure-varying parameters are optimized by minimizing both dynamic and measurement residual squared errors. Validation of this methodology has been achieved by simulating adsorption data for n-butane and isobutane on activated carbon at 298 K, 323 K and 348 K and for nitrogen on mesoporous alumina at 77 K with pressure-varying Langmuir and BET adsorption parameters (equilibrium constants and uptake capacities). This modeling provides information on the adsorbent (accessible surface area and micropore volume), adsorbate (molecular areas and volumes) and thermodynamic (Gibbs free energies) variations of the adsorption sites.

Keywords: Langmuir adsorption isotherm, BET adsorption isotherm, pressure-varying adsorption parameters, adsorbate and adsorbent properties and energetics

Procedia PDF Downloads 194
18837 Optimization of Tundish Geometry for Minimizing Dead Volume Using OpenFOAM

Authors: Prateek Singh, Dilshad Ahmad

Abstract:

Growing demand for high-quality steel products has inspired researchers to investigate the unit operations involved in the manufacturing of these products (slabs, rods, sheets, etc.). One such operation is tundish operation, in which a vessel (tundish) acts as a buffer of molten steel for the solidification operation in mold. It is observed that tundish also plays a crucial role in the quality and cleanliness of the steel produced, besides merely acting as a reservoir for the mold. It facilitates removal of dissolved oxygen (inclusions) from the molten steel thus improving its cleanliness. Inclusion removal can be enhanced by increasing the residence time of molten steel in the tundish by incorporation of flow modifiers like dams, weirs, turbo-pad, etc. These flow modifiers also help in reducing the dead or short circuit zones within the tundish which is significant for maintaining thermal and chemical homogeneity of molten steel. Thus, it becomes important to analyze the flow of molten steel in the tundish for different configuration of flow modifiers. In the present work, effect of varying positions and heights/depths of dam and weir on the dead volume in tundish is studied. Steady state thermal and flow profiles of molten steel within the tundish are obtained using OpenFOAM. Subsequently, Residence Time Distribution analysis is performed to obtain the percentage of dead volume in the tundish. Design of Experiment method is then used to configure different tundish geometries for varying positions and heights/depths of dam and weir, and dead volume for each tundish design is obtained. A second-degree polynomial with two-term interactions of independent variables to predict the dead volume in the tundish with positions and heights/depths of dam and weir as variables are computed using Multiple Linear Regression model. This polynomial is then used in an optimization framework to obtain the optimal tundish geometry for minimizing dead volume using Sequential Quadratic Programming optimization.

Keywords: design of experiments, multiple linear regression, OpenFOAM, residence time distribution, sequential quadratic programming optimization, steel, tundish

Procedia PDF Downloads 175
18836 Linear Parameter-Varying Control for Selective Catalytic Reduction Systems

Authors: Jihoon Lim, Patrick Kirchen, Ryozo Nagamune

Abstract:

This paper proposes a linear parameter-varying (LPV) controller capable of reducing nitrogen oxide (NOx) emissions with low ammonia (NH3) slip downstream of selective catalytic reduction (SCR) systems. SCR systems are widely adopted in diesel engines due to high NOx conversion efficiency. However, the nonlinearity of the SCR system and sensor uncertainty result in a challenging control problem. In order to overcome the control challenges, an LPV controller is proposed based on gain-scheduling parameters, that is, exhaust gas temperature and exhaust gas flow rate. Based on experimentally obtained data under the non-road transient driving cycle (NRTC), the simulations firstly show that the proposed controller yields high NOx conversion efficiency with a desired low NH3 slip. The performance of the proposed LPV controller is then compared with other controllers, including a gain-scheduling PID controller and a sliding mode controller. Additionally, the robustness is also demonstrated using the uncertainties ranging from 10 to 30%. The results show that the proposed controller is robustly stable under uncertainties.

Keywords: diesel engine, gain-scheduling control, linear parameter-varying, selective catalytic reduction

Procedia PDF Downloads 123
18835 Automated Evaluation Approach for Time-Dependent Question Answering Pairs on Web Crawler Based Question Answering System

Authors: Shraddha Chaudhary, Raksha Agarwal, Niladri Chatterjee

Abstract:

This work demonstrates a web crawler-based generalized end-to-end open domain Question Answering (QA) system. An efficient QA system requires a significant amount of domain knowledge to answer any question with the aim to find an exact and correct answer in the form of a number, a noun, a short phrase, or a brief piece of text for the user's questions. Analysis of the question, searching the relevant document, and choosing an answer are three important steps in a QA system. This work uses a web scraper (Beautiful Soup) to extract K-documents from the web. The value of K can be calibrated on the basis of a trade-off between time and accuracy. This is followed by a passage ranking process using the MS-Marco dataset trained on 500K queries to extract the most relevant text passage, to shorten the lengthy documents. Further, a QA system is used to extract the answers from the shortened documents based on the query and return the top 3 answers. For evaluation of such systems, accuracy is judged by the exact match between predicted answers and gold answers. But automatic evaluation methods fail due to the linguistic ambiguities inherent in the questions. Moreover, reference answers are often not exhaustive or are out of date. Hence correct answers predicted by the system are often judged incorrect according to the automated metrics. One such scenario arises from the original Google Natural Question (GNQ) dataset which was collected and made available in the year 2016. Use of any such dataset proves to be inefficient with respect to any questions that have time-varying answers. For illustration, if the query is where will be the next Olympics? Gold Answer for the above query as given in the GNQ dataset is “Tokyo”. Since the dataset was collected in the year 2016, and the next Olympics after 2016 were in 2020 that was in Tokyo which is absolutely correct. But if the same question is asked in 2022 then the answer is “Paris, 2024”. Consequently, any evaluation based on the GNQ dataset will be incorrect. Such erroneous predictions are usually given to human evaluators for further validation which is quite expensive and time-consuming. To address this erroneous evaluation, the present work proposes an automated approach for evaluating time-dependent question-answer pairs. In particular, it proposes a metric using the current timestamp along with top-n predicted answers from a given QA system. To test the proposed approach GNQ dataset has been used and the system achieved an accuracy of 78% for a test dataset comprising 100 QA pairs. This test data was automatically extracted using an analysis-based approach from 10K QA pairs of the GNQ dataset. The results obtained are encouraging. The proposed technique appears to have the possibility of developing into a useful scheme for gathering precise, reliable, and specific information in a real-time and efficient manner. Our subsequent experiments will be guided towards establishing the efficacy of the above system for a larger set of time-dependent QA pairs.

Keywords: web-based information retrieval, open domain question answering system, time-varying QA, QA evaluation

Procedia PDF Downloads 73
18834 Execution Time Optimization of Workflow Network with Activity Lead-Time

Authors: Xiaoping Qiu, Binci You, Yue Hu

Abstract:

The executive time of the workflow network has an important effect on the efficiency of the business process. In this paper, the activity executive time is divided into the service time and the waiting time, then the lead time can be extracted from the waiting time. The executive time formulas of the three basic structures in the workflow network are deduced based on the activity lead time. Taken the process of e-commerce logistics as an example, insert appropriate lead time for key activities by using Petri net, and the executive time optimization model is built to minimize the waiting time with the time-cost constraints. Then the solution program-using VC++6.0 is compiled to get the optimal solution, which reduces the waiting time of key activities in the workflow, and verifies the role of lead time in the timeliness of e-commerce logistics.

Keywords: electronic business, execution time, lead time, optimization model, petri net, time workflow network

Procedia PDF Downloads 145
18833 A Survey on Quasi-Likelihood Estimation Approaches for Longitudinal Set-ups

Authors: Naushad Mamode Khan

Abstract:

The Com-Poisson (CMP) model is one of the most popular discrete generalized linear models (GLMS) that handles both equi-, over- and under-dispersed data. In longitudinal context, an integer-valued autoregressive (INAR(1)) process that incorporates covariate specification has been developed to model longitudinal CMP counts. However, the joint likelihood CMP function is difficult to specify and thus restricts the likelihood based estimating methodology. The joint generalized quasilikelihood approach (GQL-I) was instead considered but is rather computationally intensive and may not even estimate the regression effects due to a complex and frequently ill conditioned covariance structure. This paper proposes a new GQL approach for estimating the regression parameters (GQLIII) that are based on a single score vector representation. The performance of GQL-III is compared with GQL-I and separate marginal GQLs (GQL-II) through some simulation experiments and is proved to yield equally efficient estimates as GQL-I and is far more computationally stable.

Keywords: longitudinal, com-Poisson, ill-conditioned, INAR(1), GLMS, GQL

Procedia PDF Downloads 332
18832 Lithium-Ion Battery State of Charge Estimation Using One State Hysteresis Model with Nonlinear Estimation Strategies

Authors: Mohammed Farag, Mina Attari, S. Andrew Gadsden, Saeid R. Habibi

Abstract:

Battery state of charge (SOC) estimation is an important parameter as it measures the total amount of electrical energy stored at a current time. The SOC percentage acts as a fuel gauge if it is compared with a conventional vehicle. Estimating the SOC is, therefore, essential for monitoring the amount of useful life remaining in the battery system. This paper looks at the implementation of three nonlinear estimation strategies for Li-Ion battery SOC estimation. One of the most common behavioral battery models is the one state hysteresis (OSH) model. The extended Kalman filter (EKF), the smooth variable structure filter (SVSF), and the time-varying smoothing boundary layer SVSF are applied on this model, and the results are compared.

Keywords: state of charge estimation, battery modeling, one-state hysteresis, filtering and estimation

Procedia PDF Downloads 409
18831 Change of Flavor Characteristics of Flavor Oil Made Using Sarcodon aspratus (Sarcodon aspratus Berk. S. Ito) According to Extraction Temperature and Extraction Time

Authors: Gyeong-Suk Jo, Soo-Hyun Ji, You-Seok Lee, Jeong-Hwa Kang

Abstract:

To develop an flavor oil using Sarcodon aspratus (Sarcodon aspratus Berk. S. Ito), infiltration extraction method was used to add dried mushroom flavor of Sarcodon aspratus to base olive oil. Edible base oil used during infiltration extraction was pressed olive oil, and infiltration extraction was done while varying extraction temperature to 20, 30, 40 and 50(℃) extraction time to 24 hours, 48 hours and 72 hours. Amount of Sarcodon aspratus added to base oil was 20% compared to 100% of base oil. Production yield of Sarcodon aspratus flavor oil decreased with increasing extraction frequency. Aroma intensity was 2195~2447 (A.U./1㎖), and it increased with increasing extraction temperature and extraction time. Chromaticity of Sarcodon aspratus flavor oil was bright pale yellow with pH of 4.5, sugar content of 71~72 (°Brix), and highest average turbidity of 16.74 (Haze %) shown by the 40℃ group. In the aromatic evaluation, increasing extraction temperature and extraction time resulted in increase of cheese aroma, savory sweet aroma and beef jerky aroma, as well as spicy taste comprised of slight bitter taste, savory taste and slight acrid taste, to make aromatic oil with unique flavor.

Keywords: Flavor Characteristics, Flavor Oil, Infiltration extraction method, mushroom, Sarcodon aspratus (Sarcodon aspratus Berk. S. Ito)

Procedia PDF Downloads 345
18830 Impact Location From Instrumented Mouthguard Kinematic Data In Rugby

Authors: Jazim Sohail, Filipe Teixeira-Dias

Abstract:

Mild traumatic brain injury (mTBI) within non-helmeted contact sports is a growing concern due to the serious risk of potential injury. Extensive research is being conducted looking into head kinematics in non-helmeted contact sports utilizing instrumented mouthguards that allow researchers to record accelerations and velocities of the head during and after an impact. This does not, however, allow the location of the impact on the head, and its magnitude and orientation, to be determined. This research proposes and validates two methods to quantify impact locations from instrumented mouthguard kinematic data, one using rigid body dynamics, the other utilizing machine learning. The rigid body dynamics technique focuses on establishing and matching moments from Euler’s and torque equations in order to find the impact location on the head. The methodology is validated with impact data collected from a lab test with the dummy head fitted with an instrumented mouthguard. Additionally, a Hybrid III Dummy head finite element model was utilized to create synthetic kinematic data sets for impacts from varying locations to validate the impact location algorithm. The algorithm calculates accurate impact locations; however, it will require preprocessing of live data, which is currently being done by cross-referencing data timestamps to video footage. The machine learning technique focuses on eliminating the preprocessing aspect by establishing trends within time-series signals from instrumented mouthguards to determine the impact location on the head. An unsupervised learning technique is used to cluster together impacts within similar regions from an entire time-series signal. The kinematic signals established from mouthguards are converted to the frequency domain before using a clustering algorithm to cluster together similar signals within a time series that may span the length of a game. Impacts are clustered within predetermined location bins. The same Hybrid III Dummy finite element model is used to create impacts that closely replicate on-field impacts in order to create synthetic time-series datasets consisting of impacts in varying locations. These time-series data sets are used to validate the machine learning technique. The rigid body dynamics technique provides a good method to establish accurate impact location of impact signals that have already been labeled as true impacts and filtered out of the entire time series. However, the machine learning technique provides a method that can be implemented with long time series signal data but will provide impact location within predetermined regions on the head. Additionally, the machine learning technique can be used to eliminate false impacts captured by sensors saving additional time for data scientists using instrumented mouthguard kinematic data as validating true impacts with video footage would not be required.

Keywords: head impacts, impact location, instrumented mouthguard, machine learning, mTBI

Procedia PDF Downloads 173
18829 A Comparison of Methods for Estimating Dichotomous Treatment Effects: A Simulation Study

Authors: Jacqueline Y. Thompson, Sam Watson, Lee Middleton, Karla Hemming

Abstract:

Introduction: The odds ratio (estimated via logistic regression) is a well-established and common approach for estimating covariate-adjusted binary treatment effects when comparing a treatment and control group with dichotomous outcomes. Its popularity is primarily because of its stability and robustness to model misspecification. However, the situation is different for the relative risk and risk difference, which are arguably easier to interpret and better suited to specific designs such as non-inferiority studies. So far, there is no equivalent, widely acceptable approach to estimate an adjusted relative risk and risk difference when conducting clinical trials. This is partly due to the lack of a comprehensive evaluation of available candidate methods. Methods/Approach: A simulation study is designed to evaluate the performance of relevant candidate methods to estimate relative risks to represent conditional and marginal estimation approaches. We consider the log-binomial, generalised linear models (GLM) with iteratively weighted least-squares (IWLS) and model-based standard errors (SE); log-binomial GLM with convex optimisation and model-based SEs; log-binomial GLM with convex optimisation and permutation tests; modified-Poisson GLM IWLS and robust SEs; log-binomial generalised estimation equations (GEE) and robust SEs; marginal standardisation and delta method SEs; and marginal standardisation and permutation test SEs. Independent and identically distributed datasets are simulated from a randomised controlled trial to evaluate these candidate methods. Simulations are replicated 10000 times for each scenario across all possible combinations of sample sizes (200, 1000, and 5000), outcomes (10%, 50%, and 80%), and covariates (ranging from -0.05 to 0.7) representing weak, moderate or strong relationships. Treatment effects (ranging from 0, -0.5, 1; on the log-scale) will consider null (H0) and alternative (H1) hypotheses to evaluate coverage and power in realistic scenarios. Performance measures (bias, mean square error (MSE), relative efficiency, and convergence rates) are evaluated across scenarios covering a range of sample sizes, event rates, covariate prognostic strength, and model misspecifications. Potential Results, Relevance & Impact: There are several methods for estimating unadjusted and adjusted relative risks. However, it is unclear which method(s) is the most efficient, preserves type-I error rate, is robust to model misspecification, or is the most powerful when adjusting for non-prognostic and prognostic covariates. GEE estimations may be biased when the outcome distributions are not from marginal binary data. Also, it seems that marginal standardisation and convex optimisation may perform better than GLM IWLS log-binomial.

Keywords: binary outcomes, statistical methods, clinical trials, simulation study

Procedia PDF Downloads 78
18828 Effect of Substrate Temperature on Structure and Properties of Sputtered Transparent Conducting Film of La-Doped BaSnO₃

Authors: Alok Tiwari, Ming Show Wong

Abstract:

Lanthanum (La) doped Barium Tin Oxide (BaSnO₃) film is an excellent alternative for expensive Transparent Conducting Oxides (TCOs) film such as Indium Tin Oxide (ITO). However single crystal film of La-doped BaSnO₃ has been reported with a good amount of conductivity and transparency but in order to improve its reachability, it is important to grow doped BaSO₃ films on an inexpensive substrate. La-doped BaSnO₃ thin films have been grown on quartz substrate by Radio Frequency (RF) sputtering at a different substrate temperature (from 200⁰C to 750⁰C). The thickness of the film measured was varying from 360nm to 380nm with varying substrate temperature. Structure, optical and electrical properties have been studied. The carrier concentration is seen to be decreasing as we enhance the substrate temperature while mobility found to be increased up to 9.3 cm²/V-S. At low substrate temperature resistivity found was lower (< 3x10⁻³ ohm-cm) while sudden enhancement was seen as substrate temperature raises and the trend continues further with increasing substrate temperature. Optical transmittance is getting better with higher substrate temperature from 70% at 200⁰C to > 80% at 750⁰C. Overall, understanding of changes in microstructure, electrical and optical properties of a thin film by varying substrate temperature has been reported successfully.

Keywords: conductivity, perovskite, mobility, TCO film

Procedia PDF Downloads 139
18827 The Use of Boosted Multivariate Trees in Medical Decision-Making for Repeated Measurements

Authors: Ebru Turgal, Beyza Doganay Erdogan

Abstract:

Machine learning aims to model the relationship between the response and features. Medical decision-making researchers would like to make decisions about patients’ course and treatment, by examining the repeated measurements over time. Boosting approach is now being used in machine learning area for these aims as an influential tool. The aim of this study is to show the usage of multivariate tree boosting in this field. The main reason for utilizing this approach in the field of decision-making is the ease solutions of complex relationships. To show how multivariate tree boosting method can be used to identify important features and feature-time interaction, we used the data, which was collected retrospectively from Ankara University Chest Diseases Department records. Dataset includes repeated PF ratio measurements. The follow-up time is planned for 120 hours. A set of different models is tested. In conclusion, main idea of classification with weighed combination of classifiers is a reliable method which was shown with simulations several times. Furthermore, time varying variables will be taken into consideration within this concept and it could be possible to make accurate decisions about regression and survival problems.

Keywords: boosted multivariate trees, longitudinal data, multivariate regression tree, panel data

Procedia PDF Downloads 173
18826 SNP g.1007A>G within the Porcine DNAL4 Gene Affects Sperm Motility Traits

Authors: I. Wiedemann, A. R. Sharifi, A. Mählmeyer, C. Knorr

Abstract:

A requirement for sperm motility is a morphologically intact flagellum with a central axoneme. The flagellar beating is caused by the varying activation and inactivation of dynein molecules which are located in the axoneme. DNAL4 (dynein, axonemal, light chain 4) is regarded as a possible functional candidate gene encoding a small subunit of the dyneins. In the present study, 5814bp of the porcine DNAL4 (GenBank Acc. No. AM284696.1, 6097 bp, 4 exons) were comparatively sequenced using three boars with a high motility (>68%) and three with a low motility (<60%). Primers were self-designed except for those covering exons 1, 2 and 3. Prior to sequencing, the PCR products were purified. Sequencing was performed with an ABI PRISM 3100 Genetic Analyzer using the BigDyeTM Terminator v3.1 Cycle Sequencing Reaction Kit. Finally, 23 SNPs were described and genotyped for 82 AI boars representing the breeds Piétrain, German Large White and German Landrace. The genotypes were used to assess possible associations with standard spermatological parameters (ejaculate volume, density, and sperm motility (undiluted (Motud), 24h (Mot1) and 48h (Mot2) after semen collection) that were regularly recorded on the AI station. The analysis included a total of 8,833 spermatological data sets which ranged from 2 to 295 sets per boar in five years. Only SNP g.1007A>G had a significant effect. Finally, the gene substitution effect using the following statistical model was calculated: Yijk= µ+αi+βj+αβij+b1Sijk+b2Aijk+b3T ijk + b4Vijk+b5(α*A)ijk +b6(β*A)ijk+b7(A*T)ijk+Uijk+eijk where Yijk is the semen characteristics, µ is the general mean, α is the main effect of breed, β is the main effect of season, S is the effect of SNP (g.1007A > G), A is the effect of age at semen collection, V is the effect of diluter, αβ, α*A, β*A, A*T are interactions between the fixed effects, b1-b7 are regression coefficients between y and the respective covariate, U is the random effect of repeated observation on animal and e is the random error. The results from the single marker regression analysis revealed highly significant effects (p < 0.0001) of SNP g.1007A > G on Mot1 resp. on Mot2, resulting in a marked reduction by 11.4% resp. 15.4%. Furthermore a loss of Motud by 4.6% was detected (p < 0.0178). Considering the SNP g.1007A > G as a main factor (dominant-recessive model), significant differences between genotypes AA and AG as well as AA and GG for Mot1 and Mot2 exist. For Motud there was a significant difference between AA and GG.

Keywords: association, DNAL4, porcine, sperm traits

Procedia PDF Downloads 427
18825 Investigation of Corrosion Inhibition Potential of Acalypha chamaedrifolia Leaves Extract towards Mild Steel in Acid Medium

Authors: Stephen Eyije Abechi, Casimir Emmanuel Gimba, Zaharaddeen Nasiru Garba, Sani Shamsudeen, David Ebuka Authur

Abstract:

Corrosion inhibition of mild steel in acid medium using Acalypha chamaedrifolia leaves extract as potential green inhibitor was investigated. Gravimetric (weight loss) technique was used for the corrosion studies. Mild steel coupons of 2cm × 1cm × 0.27 cm dimensions were exposed for varying durations of between 24 to 120 hours, in 1M HCl medium containing a varying concentrations of the leaves extract (0.25g/L, - 1.25g/L). The results show that corrosion rates dropped from a value of 0.49 mgcm-2hr-1 for the uninhibited medium to a value of 0.15 mgcm-2hr-1 for the inhibited medium of 1M HCl in 0.25 g/l of the extract. Values of corrosion inhibition efficiencies of 70.38-85.11% were observed as the concentration of the inhibitor were increased from 0.25g/L, - 1.25g/L. Corrosion Inhibition was found to increase with increase in immersion time and temperature. The magnitude of the Ea indicates that the interaction between the metal surface and the inhibitor was chemisorptions. The Adsorption process fit into the Langmuir isotherm model with a correlation coefficient of 0.97. Evidence from molecular dynamics model shows that Methyl stearate (Line 5) and (3Z, 13Z)-2-methyloctadeca-3,13-dien-1-ol (line 11) were found to have the highest binding energy of -197.69 ± 3.12 and-194.56 ± 10.04 in kcal/mol respectively. The binding energy of these compounds indicates that they would be a very good corrosion inhibitor for mild steel and other Fe related materials.

Keywords: binding energy, corrosion, inhibitor, langmuir isotherm, mild steel

Procedia PDF Downloads 322
18824 Characterization and Comparative Analysis of North Bengal Sand

Authors: Marzia Hoque Tania, Oishy Roy, ASW Kurny, Fahmida Gulshan

Abstract:

This paper presents results of the investigation on the characterization of silica sand of northern region of Bangladesh on the basis of material composition, particle shape, and size, density, transportation, crystallinity, etc. before and after upgradation. The raw sand samples collected from Nilphamari and Lalmonirhat district were studied and compared for the prospect silica as a high valued commodity rather than heavy minerals. The raw sand particles were colorful in appearance with varying particle size distribution. Scanning Electron Microscopy (SEM) showed uniformity in grain size and mineralogical composition. X-ray fluorescence (XRF) analysis indicated the silica content of the as-received sample to be 75%. Thermogravimetric and Differential Thermal Analysis (DTA) did not detect the presence of any organic material. These tests revealed the sample to be alpha-quartz. Samples were washed with organic and inorganic acid with a combination of varying rotation speed, concentration, solid-liquid ratio. Experiments showed the silica content could be enhanced to more than 85% by washing with 15% sulphuric acid in room temperature. Beneficiation can be improved in further work considering the effect of varying temperature or advanced technology.

Keywords: beneficiation, characterization, commercial grade sand, glass sand, silica, upgradation

Procedia PDF Downloads 110
18823 Intelligent Rescheduling Trains for Air Pollution Management

Authors: Kainat Affrin, P. Reshma, G. Narendra Kumar

Abstract:

Optimization of timetable is the need of the day for the rescheduling and routing of trains in real time. Trains are scheduled in parallel with the road transport vehicles to the same destination. As the number of trains is restricted due to single track, customers usually opt for road transport to use frequently. The air pollution increases as the density of vehicles on road transport is increased. Use of an alternate mode of transport like train helps in reducing air-pollution. This paper mainly aims at attracting the passengers to Train transport by proper rescheduling of trains using hybrid of stop-skip algorithm and iterative convex programming algorithm. Rescheduling of train bi-directionally is achieved on a single track with dynamic dual time and varying stops. Introduction of more trains attract customers to use rail transport frequently, thereby decreasing the pollution. The results are simulated using Network Simulator (NS-2).

Keywords: air pollution, AODV, re-scheduling, WSNs

Procedia PDF Downloads 333
18822 Effect of Quenching Medium on the Hardness of Dual Phase Steel Heat Treated at a High Temperature

Authors: Tebogo Mabotsa, Tamba Jamiru, David Ibrahim

Abstract:

Dual phase(DP) steel consists essentially of fine grained equiaxial ferrite and a dispersion of martensite. Martensite is the primary precipitate in DP steels, it is the main resistance to dislocation motion within the material. The objective of this paper is to present a relation between the intercritical annealing holding time and the hardness of a dual phase steel. The initial heat treatment involved heating the specimens to 1000oC and holding the sample at that temperature for 30 minutes. After the initial heat treatment, the samples were heated to 770oC and held for a varying amount of time at constant temperature. The samples were held at 30, 60, and 90 minutes respectively. After heating and holding the samples at the austenite-ferrite phase field, the samples were quenched in water, brine, and oil for each holding time. The experimental results proved that an equation for predicting the hardness of a dual phase steel as a function of the intercritical holding time is possible. The relation between intercritical annealing holding time and hardness of a dual phase steel heat treated at high temperatures is parabolic in nature. Theoretically, the model isdependent on the cooling rate because the model differs for each quenching medium; therefore, a universal hardness equation can be derived where the cooling rate is a variable factor.

Keywords: quenching medium, annealing temperature, dual phase steel, martensite

Procedia PDF Downloads 57
18821 Study of the Effect of the Continuous Electric Field on the Rd Cancer Cell Line by Response Surface Methodology

Authors: Radia Chemlal, Salim Mehenni, Dahbia Leila Anes-boulahbal, Mohamed Kherat, Nabil Mameri

Abstract:

The application of the electric field is considered to be a very promising method in cancer therapy. Indeed, cancer cells are very sensitive to the electric field, although the cellular response is not entirely clear. The tests carried out consisted in subjecting the RD cell line under the effect of the continuous electric field while varying certain parameters (voltage, exposure time, and cell concentration). The response surface methodology (RSM) was used to assess the effect of the chosen parameters, as well as the existence of interactions between them. The results obtained showed that the voltage, the cell concentration as well as the interaction between voltage and exposure time have an influence on the mortality rate of the RD cell line.

Keywords: continuous electric field, RD cancer cell line, RSM, voltage

Procedia PDF Downloads 79
18820 Studies on Mechanisms of Corrosion Inhibition of Acalypha chamaedrifolia Leaves Extract towards Mild Steel in Acid Medium

Authors: Stephen Eyije Abechi, Casimir Emmanuel Gimba, Zaharaddeen Nasiru Garba, Sani Shamsudeen, David Ebuka Authur

Abstract:

The mechanisms of corrosion inhibition of mild steel in acid medium using Acalypha chamaedrifolia leaves extract as potential green inhibitor were investigated. Gravimetric (weight loss) technique was used for the corrosion studies. Mild steel coupons of 2cm × 1cm × 0.27 cm dimensions were exposed for varying durations of between 24 to 120 hours, in 1M HCl medium containing a varying concentrations of the leaves extract (0.25g/L, - 1.25g/L). The results show that corrosion rates dropped from a value of 0.49 mgcm-2hr-1 for the uninhibited medium to a value of 0.15 mgcm-2hr-1 for the inhibited medium of 1M HCl in 0.25 g/l of the extract. Values of corrosion inhibition efficiencies of 70.38-85.11% were observed as the concentration of the inhibitor were increased from 0.25g/L, - 1.25g/L. Corrosion Inhibition was found to increase with increase in immersion time and temperature. The magnitude of the Ea indicates that the interaction between the metal surface and the inhibitor was chemisorptions. The Adsorption process fit into the Langmuir isotherm model with a correlation coefficient of 0.97. Evidence from molecular dynamics model shows that Methyl stearate (Line 5) and (3Z, 13Z)-2-methyloctadeca-3,13-dien-1-ol (line 11) were found to have the highest binding energy of -197.69 ± 3.12 and-194.56 ± 10.04 in kcal/mol respectively. The binding energy of these compounds indicates that they would be a very good corrosion inhibitor for mild steel and other Fe related materials.

Keywords: binding energy, corrosion, inhibitor, Langmuir isotherm, mild steel.

Procedia PDF Downloads 321
18819 On Bianchi Type Cosmological Models in Lyra’s Geometry

Authors: R. K. Dubey

Abstract:

Bianchi type cosmological models have been studied on the basis of Lyra’s geometry. Exact solution has been obtained by considering a time dependent displacement field for constant deceleration parameter and varying cosmological term of the universe. The physical behavior of the different models has been examined for different cases.

Keywords: Bianchi type-I cosmological model, variable gravitational coupling, cosmological constant term, Lyra's model

Procedia PDF Downloads 327
18818 Study of Icons in Enterprise Application Software Context

Authors: Shiva Subhedar, Abhishek Jain, Shivin Mittal

Abstract:

Icons are not merely decorative elements in enterprise applications but very often used because of their many advantages such as compactness, visual appeal, etc. Despite these potential advantages, icons often cause usability problems when they are designed without consideration for their many potential downsides. The aim of the current study was to examine the effect of articulatory distance – the distance between the physical appearance of an interface element and what it actually means. In other words, will the subject find the association of the function and its appearance on the interface natural or is the icon difficult for them to associate with its function. We have calculated response time and quality of identification by varying icon concreteness, the context of usage and subject experience in the enterprise context. The subjects were asked to associate icons (prepared for study purpose) with given function options in context and out of context mode. Response time and their selection were recorded for analysis.

Keywords: HCI, icons, icon concreteness, icon recognition

Procedia PDF Downloads 232
18817 Performance and Voyage Analysis of Marine Gas Turbine Engine, Installed to Power and Propel an Ocean-Going Cruise Ship from Lagos to Jeddah

Authors: Mathias U. Bonet, Pericles Pilidis, Georgios Doulgeris

Abstract:

An aero-derivative marine Gas Turbine engine model is simulated to be installed as the main propulsion prime mover to power a cruise ship which is designed and routed to transport intending Muslim pilgrims for the annual hajj pilgrimage from Nigeria to the Islamic port city of Jeddah in Saudi Arabia. A performance assessment of the Gas Turbine engine has been conducted by examining the effect of varying aerodynamic and hydrodynamic conditions encountered at various geographical locations along the scheduled transit route during the voyage. The investigation focuses on the overall behavior of the Gas Turbine engine employed to power and propel the ship as it operates under ideal and adverse conditions to be encountered during calm and rough weather according to the different seasons of the year under which the voyage may be undertaken. The variation of engine performance under varying operating conditions has been considered as a very important economic issue by determining the time the speed by which the journey is completed as well as the quantity of fuel required for undertaking the voyage. The assessment also focuses on the increased resistance caused by the fouling of the submerged portion of the ship hull surface with its resultant effect on the power output of the engine as well as the overall performance of the propulsion system. Daily ambient temperature levels were obtained by accessing data from the UK Meteorological Office while the varying degree of turbulence along the transit route and according to the Beaufort scale were also obtained as major input variables of the investigation. By assuming the ship to be navigating the Atlantic Ocean and the Mediterranean Sea during winter, spring and summer seasons, the performance modeling and simulation was accomplished through the use of an integrated Gas Turbine performance simulation code known as ‘Turbomach’ along with a Matlab generated code named ‘Poseidon’, all of which have been developed at the Power and Propulsion Department of Cranfield University. As a case study, the results of the various assumptions have further revealed that the marine Gas Turbine is a reliable and available alternative to the conventional marine propulsion prime movers that have dominated the maritime industry before now. The techno-economic and environmental assessment of this type of propulsion prime mover has enabled the determination of the effect of changes in weather and sea conditions on the ship speed as well as trip time and the quantity of fuel required to be burned throughout the voyage.

Keywords: ambient temperature, hull fouling, marine gas turbine, performance, propulsion, voyage

Procedia PDF Downloads 159
18816 Analysing the Interactive Effects of Factors Influencing Sand Production on Drawdown Time in High Viscosity Reservoirs

Authors: Gerald Gwamba, Bo Zhou, Yajun Song, Dong Changyin

Abstract:

The challenges that sand production presents to the oil and gas industry, particularly while working in poorly consolidated reservoirs, cannot be overstated. From restricting production to blocking production tubing, sand production increases the costs associated with production as it elevates the cost of servicing production equipment over time. Production in reservoirs that present with high viscosities, flow rate, cementation, clay content as well as fine sand contents is even more complex and challenging. As opposed to the one-factor at a-time testing, investigating the interactive effects arising from a combination of several factors offers increased reliability of results as well as representation of actual field conditions. It is thus paramount to investigate the conditions leading to the onset of sanding during production to ensure the future sustainability of hydrocarbon production operations under viscous conditions. We adopt the Design of Experiments (DOE) to analyse, using Taguchi factorial designs, the most significant interactive effects of sanding. We propose an optimized regression model to predict the drawdown time at sand production. The results obtained underscore that reservoirs characterized by varying (high and low) levels of viscosity, flow rate, cementation, clay, and fine sand content have a resulting impact on sand production. The only significant interactive effect recorded arises from the interaction between BD (fine sand content and flow rate), while the main effects included fluid viscosity and cementation, with percentage significances recorded as 31.3%, 37.76%, and 30.94%, respectively. The drawdown time model presented could be useful for predicting the time to reach the maximum drawdown pressure under viscous conditions during the onset of sand production.

Keywords: factorial designs, DOE optimization, sand production prediction, drawdown time, regression model

Procedia PDF Downloads 111
18815 The Effect of Improvement Programs in the Mean Time to Repair and in the Mean Time between Failures on Overall Lead Time: A Simulation Using the System Dynamics-Factory Physics Model

Authors: Marcel Heimar Ribeiro Utiyama, Fernanda Caveiro Correia, Dario Henrique Alliprandini

Abstract:

The importance of the correct allocation of improvement programs is of growing interest in recent years. Due to their limited resources, companies must ensure that their financial resources are directed to the correct workstations in order to be the most effective and survive facing the strong competition. However, to our best knowledge, the literature about allocation of improvement programs does not analyze in depth this problem when the flow shop process has two capacity constrained resources. This is a research gap which is deeply studied in this work. The purpose of this work is to identify the best strategy to allocate improvement programs in a flow shop with two capacity constrained resources. Data were collected from a flow shop process with seven workstations in an industrial control and automation company, which process 13.690 units on average per month. The data were used to conduct a simulation with the System Dynamics-Factory Physics model. The main variables considered, due to their importance on lead time reduction, were the mean time between failures and the mean time to repair. The lead time reduction was the output measure of the simulations. Ten different strategies were created: (i) focused time to repair improvement, (ii) focused time between failures improvement, (iii) distributed time to repair improvement, (iv) distributed time between failures improvement, (v) focused time to repair and time between failures improvement, (vi) distributed time to repair and between failures improvement, (vii) hybrid time to repair improvement, (viii) hybrid time between failures improvements, (ix) time to repair improvement strategy towards the two capacity constrained resources, (x) time between failures improvement strategy towards the two capacity constrained resources. The ten strategies tested are variations of the three main strategies for improvement programs named focused, distributed and hybrid. Several comparisons among the effect of the ten strategies in lead time reduction were performed. The results indicated that for the flow shop analyzed, the focused strategies delivered the best results. When it is not possible to perform a large investment on the capacity constrained resources, companies should use hybrid approaches. An important contribution to the academy is the hybrid approach, which proposes a new way to direct the efforts of improvements. In addition, the study in a flow shop with two strong capacity constrained resources (more than 95% of utilization) is an important contribution to the literature. Another important contribution is the problem of allocation with two CCRs and the possibility of having floating capacity constrained resources. The results provided the best improvement strategies considering the different strategies of allocation of improvement programs and different positions of the capacity constrained resources. Finally, it is possible to state that both strategies, hybrid time to repair improvement and hybrid time between failures improvement, delivered best results compared to the respective distributed strategies. The main limitations of this study are mainly regarding the flow shop analyzed. Future work can further investigate different flow shop configurations like a varying number of workstations, different number of products or even different positions of the two capacity constrained resources.

Keywords: allocation of improvement programs, capacity constrained resource, hybrid strategy, lead time, mean time to repair, mean time between failures

Procedia PDF Downloads 92