Search results for: cost prediction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8177

Search results for: cost prediction

7217 Policy Recommendations for Reducing CO2 Emissions in Kenya's Electricity Generation, 2015-2030

Authors: Paul Kipchumba

Abstract:

Kenya is an East African Country lying at the Equator. It had a population of 46 million in 2015 with an annual growth rate of 2.7%, making a population of at least 65 million in 2030. Kenya’s GDP in 2015 was about 63 billion USD with per capita GDP of about 1400 USD. The rural population is 74%, whereas urban population is 26%. Kenya grapples with not only access to energy but also with energy security. There is direct correlation between economic growth, population growth, and energy consumption. Kenya’s energy composition is at least 74.5% from renewable energy with hydro power and geothermal forming the bulk of it; 68% from wood fuel; 22% from petroleum; 9% from electricity; and 1% from coal and other sources. Wood fuel is used by majority of rural and poor urban population. Electricity is mostly used for lighting. As of March 2015 Kenya had installed electricity capacity of 2295 MW, making a per capital electricity consumption of 0.0499 KW. The overall retail cost of electricity in 2015 was 0.009915 USD/ KWh (KES 19.85/ KWh), for installed capacity over 10MW. The actual demand for electricity in 2015 was 3400 MW and the projected demand in 2030 is 18000 MW. Kenya is working on vision 2030 that aims at making it a prosperous middle income economy and targets 23 GW of generated electricity. However, cost and non-cost factors affect generation and consumption of electricity in Kenya. Kenya does not care more about CO2 emissions than on economic growth. Carbon emissions are most likely to be paid by future costs of carbon emissions and penalties imposed on local generating companies by sheer disregard of international law on C02 emissions and climate change. The study methodology was a simulated application of carbon tax on all carbon emitting sources of electricity generation. It should cost only USD 30/tCO2 tax on all emitting sources of electricity generation to have solar as the only source of electricity generation in Kenya. The country has the best evenly distributed global horizontal irradiation. Solar potential after accounting for technology efficiencies such as 14-16% for solar PV and 15-22% for solar thermal is 143.94 GW. Therefore, the paper recommends adoption of solar power for generating all electricity in Kenya in order to attain zero carbon electricity generation in the country.

Keywords: co2 emissions, cost factors, electricity generation, non-cost factors

Procedia PDF Downloads 369
7216 A Comparison between the Results of Hormuz Strait Wave Simulations Using WAVEWATCH-III and MIKE21-SW and Satellite Altimetry Observations

Authors: Fatemeh Sadat Sharifi

Abstract:

In the present study, the capabilities of WAVEWATCH-III and MIKE21-SW for predicting the characteristics of wind waves in Hormuz Strait are evaluated. The GFS wind data (Global Forecast System) were derived. The bathymetry of gride with 2 arc-minute resolution, also were extracted from the ETOPO1. WAVEWATCH-III findings illustrate more valid prediction of wave features comparing to the MIKE-21 SW in deep water. Apparently, in shallow area, the MIKE-21 provides more uniformities with altimetry measurements. This may be due to the merits of the unstructured grid which are used in MIKE-21, leading to better representations of the coastal area. The findings on the direction of waves generated by wind in the modeling area indicate that in some regions, despite the increase in wind speed, significant wave height stays nearly unchanged. This is fundamental because of swift changes in wind track over the Strait of Hormuz. After discussing wind-induced waves in the region, the impact of instability of the surface layer on wave growth has been considered. For this purpose, the average monthly mean air temperature has been used. The results in cold months, when the surface layer is unstable, indicates an acceptable increase in the accuracy of prediction of the indicator wave height.

Keywords: numerical modeling, WAVEWATCH-III, Strait of Hormuz, MIKE21-SW

Procedia PDF Downloads 208
7215 Institutional Design for Managing Irrigation Problems: A Case Study of Farmers'- and Agency-Managed Irrigation Systems of Nepal

Authors: Tirtha Raj Dhakal, Brian Davidson, Bob Farquharson

Abstract:

Institutional design is an important aspect in efficient water resource management. In Nepal, the water supply in both farmers’- and agency-managed irrigation systems has become sub-standard because of the weak institutional framework. This study characterizes both forms of the schemes and links existing institution and governance of the schemes with its performance with reference to cost recovery, maintenance of the schemes and water distribution throughout the schemes. For this, two types of surveys were conducted. A management survey of ten farmers’-managed and five agency-managed schemes of Chitwan valley and its periphery was done. Also, a farm survey comprising 25 farmers from each of head, middle and tail regions of both schemes; Narayani Lift Irrigation Project (agency-managed) and Khageri Irrigation System (farmers’-managed) of Chitwan Valley as a case study was conducted. The results showed that cost recovery of agency-managed schemes in 2015 was less than two percent whereas service fee collection rate in farmers’-managed schemes was nearly 2/3rd that triggered poor maintenance of the schemes and unequal distribution of water throughout the schemes. Also, the institution on practice is unable to create any incentives for farmers for economical use of water as well as willingness to pay for its use. This, thus, compels the need of refined institutional framework which has been suggested in this paper aiming to improve the cost recovery and better water distribution throughout the irrigation schemes.

Keywords: cost recovery, governance, institution, schemes' performance

Procedia PDF Downloads 264
7214 Comparison of Multivariate Adaptive Regression Splines and Random Forest Regression in Predicting Forced Expiratory Volume in One Second

Authors: P. V. Pramila , V. Mahesh

Abstract:

Pulmonary Function Tests are important non-invasive diagnostic tests to assess respiratory impairments and provides quantifiable measures of lung function. Spirometry is the most frequently used measure of lung function and plays an essential role in the diagnosis and management of pulmonary diseases. However, the test requires considerable patient effort and cooperation, markedly related to the age of patients esulting in incomplete data sets. This paper presents, a nonlinear model built using Multivariate adaptive regression splines and Random forest regression model to predict the missing spirometric features. Random forest based feature selection is used to enhance both the generalization capability and the model interpretability. In the present study, flow-volume data are recorded for N= 198 subjects. The ranked order of feature importance index calculated by the random forests model shows that the spirometric features FVC, FEF 25, PEF,FEF 25-75, FEF50, and the demographic parameter height are the important descriptors. A comparison of performance assessment of both models prove that, the prediction ability of MARS with the `top two ranked features namely the FVC and FEF 25 is higher, yielding a model fit of R2= 0.96 and R2= 0.99 for normal and abnormal subjects. The Root Mean Square Error analysis of the RF model and the MARS model also shows that the latter is capable of predicting the missing values of FEV1 with a notably lower error value of 0.0191 (normal subjects) and 0.0106 (abnormal subjects). It is concluded that combining feature selection with a prediction model provides a minimum subset of predominant features to train the model, yielding better prediction performance. This analysis can assist clinicians with a intelligence support system in the medical diagnosis and improvement of clinical care.

Keywords: FEV, multivariate adaptive regression splines pulmonary function test, random forest

Procedia PDF Downloads 314
7213 Influence of Transportation Mode to the Deterioration Rate: Case Study of Food Transport by Ship

Authors: Danijela Tuljak-Suban, Valter Suban

Abstract:

Food as perishable goods represents a specific and sensitive part in the supply chain theory, since changing of its physical or chemical characteristics considerably influences the approach to stock management. The most delicate phase of this process is transportation, where it becomes difficult to ensure stability conditions that limit the deterioration, since the value of the deterioration rate could be easily influenced by the transportation mode. Fuzzy definition of variables allows taking into account these variations. Furthermore an appropriate choice of the defuzzification method permits to adapt results, as much as possible, to real conditions. In the article will be applied the those methods to the relationship between the deterioration rate of perishable goods and transportation by ship, with the aim: (a) to minimize the total costs function, defined as the sum of the ordering cost, holding cost, disposing cost and transportation costs, and (b) to improve supply chain sustainability by reducing the environmental impact and waste disposal costs.

Keywords: perishable goods, fuzzy reasoning, transport by ship, supply chain sustainability

Procedia PDF Downloads 548
7212 A Contemporary Advertising Strategy on Social Networking Sites

Authors: M. S. Aparna, Pushparaj Shetty D.

Abstract:

Nowadays social networking sites have become so popular that the producers or the sellers look for these sites as one of the best options to target the right audience to market their products. There are several tools available to monitor or analyze the social networks. Our task is to identify the right community web pages and find out the behavior analysis of the members by using these tools and formulate an appropriate strategy to market the products or services to achieve the set goals. The advertising becomes more effective when the information of the product/ services come from a known source. The strategy explores great buying influence in the audience on referral marketing. Our methodology proceeds with critical budget analysis and promotes viral influence propagation. In this context, we encompass the vital bits of budget evaluation such as the number of optimal seed nodes or primary influential users activated onset, an estimate coverage spread of nodes and maximum influence propagating distance from an initial seed to an end node. Our proposal for Buyer Prediction mathematical model arises from the urge to perform complex analysis when the probability density estimates of reliable factors are not known or difficult to calculate. Order Statistics and Buyer Prediction mapping function guarantee the selection of optimal influential users at each level. We exercise an efficient tactics of practicing community pages and user behavior to determine the product enthusiasts on social networks. Our approach is promising and should be an elementary choice when there is little or no prior knowledge on the distribution of potential buyers on social networks. In this strategy, product news propagates to influential users on or surrounding networks. By applying the same technique, a user can search friends who are capable to advise better or give referrals, if a product interests him.

Keywords: viral marketing, social network analysis, community web pages, buyer prediction, influence propagation, budget constraints

Procedia PDF Downloads 264
7211 Development, Testing, and Application of a Low-Cost Technology Sulphur Dioxide Monitor as a Tool for use in a Volcanic Emissions Monitoring Network

Authors: Viveka Jackson, Erouscilla Joseph, Denise Beckles, Thomas Christopher

Abstract:

Sulphur Dioxide (SO2) has been defined as a non-flammable, non-explosive, colourless gas, having a pungent, irritating odour, and is one of the main gases emitted from volcanoes. Sulphur dioxide has been recorded in concentrations hazardous to humans (0.25 – 0.5 ppm (~650 – 1300 μg/m3), downwind of many volcanoes and hence warrants constant air-quality monitoring around these sites. It has been linked to an increase in chronic respiratory disease attributed to long-term exposures and alteration in lung and other physiological functions attributed to short-term exposures. Sulphur Springs in Saint Lucia is a highly active geothermal area, located within the Soufrière Volcanic Centre, and is a park widely visited by tourists and locals. It is also a current source of continuous volcanic emissions via its many fumaroles and bubbling pools, warranting concern by residents and visitors to the park regarding the effects of exposure to these gases. In this study, we introduce a novel SO2 measurement system for the monitoring and quantification of ambient levels of airborne volcanic SO2 using low-cost technology. This work involves the extensive production of low-cost SO2 monitors/samplers, as well as field examination in tandem with standard commercial samplers (SO2 diffusion tubes). It also incorporates community involvement in the volcanic monitoring process as non-professional users of the instrument. We intend to present the preliminary monitoring results obtained from the low-cost samplers, to identify the areas in the Park exposed to high concentrations of ambient SO2, and to assess the feasibility of the instrument for non-professional use and application in volcanic settings

Keywords: ambient SO2, community-based monitoring, risk-reduction, sulphur springs, low-cost

Procedia PDF Downloads 470
7210 The Psychosis Prodrome: Biomarkers of the Glutamatergic System and Their Potential Role in Prediction and Treatment

Authors: Peter David Reiss

Abstract:

The concept of the psychosis prodrome has allowed for the identification of adolescent and young adult patients who have a significantly elevated risk of developing schizophrenia spectrum disorders. A number of different interventions have been tested in order to prevent or delay progression of symptoms. To date, there has been no consistent meta-analytical evidence to support efficacy of antipsychotic treatment for patients in the prodromal state, and their use remains therefore inconclusive. Although antipsychotics may manage symptoms transiently, they have not been found to prevent or delay onset of psychotic disorders. Furthermore, pharmacological intervention in high-risk individuals remains controversial, because of the antipsychotic side effect profile in a population in which only about 20 to 35 percent will eventually convert to psychosis over a two-year period, with even after two years conversion rates not exceeding 30 to 40 percent. This general estimate is additionally problematic, in that it ignores the fact that there is significant variation in individual risk among clinical high-risk cases. The current lack of reliable tests for at-risk patients makes it difficult to justify individual treatment decisions. Preventive treatment should ideally be dictated by an individual’s risk while minimizing potentially harmful medication exposure. This requires more accurate predictive assessments by using valid and accessible prognostic markers. The following will compare prediction and risk modification potential of behavioral biomarkers such as disturbances of basic sense of self and emotion awareness, neurocognitive biomarkers such as attention, working and declarative memory, and neurophysiological biomarkers such as glutamatergic abnormalities and NMDA receptor dysfunction. Identification of robust biomarkers could therefore not only provide more reliable means of psychosis prediction, but also help test and develop new clinical interventions targeted at the prodromal state.

Keywords: at-risk mental state, biomarkers, glutamatergic system, NMDA receptor, psychosis prodrome, schizophrenia

Procedia PDF Downloads 199
7209 Pavement Maintenance and Rehabilitation Scheduling Using Genetic Algorithm Based Multi Objective Optimization Technique

Authors: Ashwini Gowda K. S, Archana M. R, Anjaneyappa V

Abstract:

This paper presents pavement maintenance and management system (PMMS) to obtain optimum pavement maintenance and rehabilitation strategies and maintenance scheduling for a network using a multi-objective genetic algorithm (MOGA). Optimal pavement maintenance & rehabilitation strategy is to maximize the pavement condition index of the road section in a network with minimum maintenance and rehabilitation cost during the planning period. In this paper, NSGA-II is applied to perform maintenance optimization; this maintenance approach was expected to preserve and improve the existing condition of the highway network in a cost-effective way. The proposed PMMS is applied to a network that assessed pavement based on the pavement condition index (PCI). The minimum and maximum maintenance cost for a planning period of 20 years obtained from the non-dominated solution was found to be 5.190x10¹⁰ ₹ and 4.81x10¹⁰ ₹, respectively.

Keywords: genetic algorithm, maintenance and rehabilitation, optimization technique, pavement condition index

Procedia PDF Downloads 156
7208 A Tutorial on Model Predictive Control for Spacecraft Maneuvering Problem with Theory, Experimentation and Applications

Authors: O. B. Iskender, K. V. Ling, V. Dubanchet, L. Simonini

Abstract:

This paper discusses the recent advances and future prospects of spacecraft position and attitude control using Model Predictive Control (MPC). First, the challenges of the space missions are summarized, in particular, taking into account the errors, uncertainties, and constraints imposed by the mission, spacecraft and, onboard processing capabilities. The summary of space mission errors and uncertainties provided in categories; initial condition errors, unmodeled disturbances, sensor, and actuator errors. These previous constraints are classified into two categories: physical and geometric constraints. Last, real-time implementation capability is discussed regarding the required computation time and the impact of sensor and actuator errors based on the Hardware-In-The-Loop (HIL) experiments. The rationales behind the scenarios’ are also presented in the scope of space applications as formation flying, attitude control, rendezvous and docking, rover steering, and precision landing. The objectives of these missions are explained, and the generic constrained MPC problem formulations are summarized. Three key design elements used in MPC design: the prediction model, the constraints formulation and the objective cost function are discussed. The prediction models can be linear time invariant or time varying depending on the geometry of the orbit, whether it is circular or elliptic. The constraints can be given as linear inequalities for input or output constraints, which can be written in the same form. Moreover, the recent convexification techniques for the non-convex geometrical constraints (i.e., plume impingement, Field-of-View (FOV)) are presented in detail. Next, different objectives are provided in a mathematical framework and explained accordingly. Thirdly, because MPC implementation relies on finding in real-time the solution to constrained optimization problems, computational aspects are also examined. In particular, high-speed implementation capabilities and HIL challenges are presented towards representative space avionics. This covers an analysis of future space processors as well as the requirements of sensors and actuators on the HIL experiments outputs. The HIL tests are investigated for kinematic and dynamic tests where robotic arms and floating robots are used respectively. Eventually, the proposed algorithms and experimental setups are introduced and compared with the authors' previous work and future plans. The paper concludes with a conjecture that MPC paradigm is a promising framework at the crossroads of space applications while could be further advanced based on the challenges mentioned throughout the paper and the unaddressed gap.

Keywords: convex optimization, model predictive control, rendezvous and docking, spacecraft autonomy

Procedia PDF Downloads 113
7207 Prediction of the Crustal Deformation of Volcán - Nevado Del RUíz in the Year 2020 Using Tropomi Tropospheric Information, Dinsar Technique, and Neural Networks

Authors: Juan Sebastián Hernández

Abstract:

The Nevado del Ruíz volcano, located between the limits of the Departments of Caldas and Tolima in Colombia, presented an unstable behaviour in the course of the year 2020, this volcanic activity led to secondary effects on the crust, which is why the prediction of deformations becomes the task of geoscientists. In the course of this article, the use of tropospheric variables such as evapotranspiration, UV aerosol index, carbon monoxide, nitrogen dioxide, methane, surface temperature, among others, is used to train a set of neural networks that can predict the behaviour of the resulting phase of an unrolled interferogram with the DInSAR technique, whose main objective is to identify and characterise the behaviour of the crust based on the environmental conditions. For this purpose, variables were collected, a generalised linear model was created, and a set of neural networks was created. After the training of the network, validation was carried out with the test data, giving an MSE of 0.17598 and an associated r-squared of approximately 0.88454. The resulting model provided a dataset with good thematic accuracy, reflecting the behaviour of the volcano in 2020, given a set of environmental characteristics.

Keywords: crustal deformation, Tropomi, neural networks (ANN), volcanic activity, DInSAR

Procedia PDF Downloads 107
7206 Strategies in Customer Relationship Management and Customers’ Behavior in Making Decision on Buying Car Insurance of Southeast Insurance Co. Ltd. in Bangkok

Authors: Nattapong Techarattanased, Paweena Sribunrueng

Abstract:

The objective of this study is to investigate strategies in customer relationship management and customers’ behavior in making decision on buying car insurance of Southeast Insurance Co. Ltd. in Bangkok. Subjects in this study included 400 customers with the age over 20 years old to complete questionnaires. The data were analyzed by arithmetic mean and multiple regressions. The results revealed that the customers’ opinions on the strategies in customer relationship management, i.e. customer relationship, customer feedback, customer follow-up, useful service suggestions, customer communication, and service channels were in moderate level but on the customer retention was in high level. Moreover, the strategy in customer relationship management, i.e. customer relationship, and customer feedback had an influence on customers’ buying decision on buying car insurance. The two factors above can be used for the prediction at the rate of 34%. In addition, the strategy in customer relationship management, i.e. customer retention, customer feedback, and useful service suggestions had an influence on the customers’ buying decision on period of being customers. The three factors could be used for the prediction at the rate of 45%.

Keywords: strategies, customer relationship management, behavior in buying decision, car insurance

Procedia PDF Downloads 410
7205 Using Simulation Modeling Approach to Predict USMLE Steps 1 and 2 Performances

Authors: Chau-Kuang Chen, John Hughes, Jr., A. Dexter Samuels

Abstract:

The prediction models for the United States Medical Licensure Examination (USMLE) Steps 1 and 2 performances were constructed by the Monte Carlo simulation modeling approach via linear regression. The purpose of this study was to build robust simulation models to accurately identify the most important predictors and yield the valid range estimations of the Steps 1 and 2 scores. The application of simulation modeling approach was deemed an effective way in predicting student performances on licensure examinations. Also, sensitivity analysis (a/k/a what-if analysis) in the simulation models was used to predict the magnitudes of Steps 1 and 2 affected by changes in the National Board of Medical Examiners (NBME) Basic Science Subject Board scores. In addition, the study results indicated that the Medical College Admission Test (MCAT) Verbal Reasoning score and Step 1 score were significant predictors of the Step 2 performance. Hence, institutions could screen qualified student applicants for interviews and document the effectiveness of basic science education program based on the simulation results.

Keywords: prediction model, sensitivity analysis, simulation method, USMLE

Procedia PDF Downloads 342
7204 Exploring Syntactic and Semantic Features for Text-Based Authorship Attribution

Authors: Haiyan Wu, Ying Liu, Shaoyun Shi

Abstract:

Authorship attribution is to extract features to identify authors of anonymous documents. Many previous works on authorship attribution focus on statistical style features (e.g., sentence/word length), content features (e.g., frequent words, n-grams). Modeling these features by regression or some transparent machine learning methods gives a portrait of the authors' writing style. But these methods do not capture the syntactic (e.g., dependency relationship) or semantic (e.g., topics) information. In recent years, some researchers model syntactic trees or latent semantic information by neural networks. However, few works take them together. Besides, predictions by neural networks are difficult to explain, which is vital in authorship attribution tasks. In this paper, we not only utilize the statistical style and content features but also take advantage of both syntactic and semantic features. Different from an end-to-end neural model, feature selection and prediction are two steps in our method. An attentive n-gram network is utilized to select useful features, and logistic regression is applied to give prediction and understandable representation of writing style. Experiments show that our extracted features can improve the state-of-the-art methods on three benchmark datasets.

Keywords: authorship attribution, attention mechanism, syntactic feature, feature extraction

Procedia PDF Downloads 140
7203 Distangling Biological Noise in Cellular Images with a Focus on Explainability

Authors: Manik Sharma, Ganapathy Krishnamurthi

Abstract:

The cost of some drugs and medical treatments has risen in recent years, that many patients are having to go without. A classification project could make researchers more efficient. One of the more surprising reasons behind the cost is how long it takes to bring new treatments to market. Despite improvements in technology and science, research and development continues to lag. In fact, finding new treatment takes, on average, more than 10 years and costs hundreds of millions of dollars. If successful, we could dramatically improve the industry's ability to model cellular images according to their relevant biology. In turn, greatly decreasing the cost of treatments and ensure these treatments get to patients faster. This work aims at solving a part of this problem by creating a cellular image classification model which can decipher the genetic perturbations in cell (occurring naturally or artificially). Another interesting question addressed is what makes the deep-learning model decide in a particular fashion, which can further help in demystifying the mechanism of action of certain perturbations and paves a way towards the explainability of the deep-learning model.

Keywords: cellular images, genetic perturbations, deep-learning, explainability

Procedia PDF Downloads 116
7202 Predictive Semi-Empirical NOx Model for Diesel Engine

Authors: Saurabh Sharma, Yong Sun, Bruce Vernham

Abstract:

Accurate prediction of NOx emission is a continuous challenge in the field of diesel engine-out emission modeling. Performing experiments for each conditions and scenario cost significant amount of money and man hours, therefore model-based development strategy has been implemented in order to solve that issue. NOx formation is highly dependent on the burn gas temperature and the O2 concentration inside the cylinder. The current empirical models are developed by calibrating the parameters representing the engine operating conditions with respect to the measured NOx. This makes the prediction of purely empirical models limited to the region where it has been calibrated. An alternative solution to that is presented in this paper, which focus on the utilization of in-cylinder combustion parameters to form a predictive semi-empirical NOx model. The result of this work is shown by developing a fast and predictive NOx model by using the physical parameters and empirical correlation. The model is developed based on the steady state data collected at entire operating region of the engine and the predictive combustion model, which is developed in Gamma Technology (GT)-Power by using Direct Injected (DI)-Pulse combustion object. In this approach, temperature in both burned and unburnt zone is considered during the combustion period i.e. from Intake Valve Closing (IVC) to Exhaust Valve Opening (EVO). Also, the oxygen concentration consumed in burnt zone and trapped fuel mass is also considered while developing the reported model.  Several statistical methods are used to construct the model, including individual machine learning methods and ensemble machine learning methods. A detailed validation of the model on multiple diesel engines is reported in this work. Substantial numbers of cases are tested for different engine configurations over a large span of speed and load points. Different sweeps of operating conditions such as Exhaust Gas Recirculation (EGR), injection timing and Variable Valve Timing (VVT) are also considered for the validation. Model shows a very good predictability and robustness at both sea level and altitude condition with different ambient conditions. The various advantages such as high accuracy and robustness at different operating conditions, low computational time and lower number of data points requires for the calibration establishes the platform where the model-based approach can be used for the engine calibration and development process. Moreover, the focus of this work is towards establishing a framework for the future model development for other various targets such as soot, Combustion Noise Level (CNL), NO2/NOx ratio etc.

Keywords: diesel engine, machine learning, NOₓ emission, semi-empirical

Procedia PDF Downloads 116
7201 Conceptual Perimeter Model for Estimating Building Envelope Quantities

Authors: Ka C. Lam, Oluwafunmibi S. Idowu

Abstract:

Building girth is important in building economics and mostly used in quantities take-off of various cost items. Literature suggests that the use of conceptual quantities can improve the accuracy of cost models. Girth or perimeter of a building can be used to estimate conceptual quantities. Hence, the current paper aims to model the perimeter-area function of buildings shapes for use at the conceptual design stage. A detailed literature review on existing building shape indexes was carried out. An empirical approach was used to study the relationship between area and the shortest length of a four-sided orthogonal polygon. Finally, a mathematical approach was used to establish the observed relationships. The empirical results obtained were in agreement with the mathematical model developed. A new equation termed “conceptual perimeter equation” is proposed. The equation can be used to estimate building envelope quantities such as external wall area, external finishing area and scaffolding area before sketch or detailed drawings are prepared.

Keywords: building envelope, building shape index, conceptual quantities, cost modelling, girth

Procedia PDF Downloads 349
7200 Application of Multilinear Regression Analysis for Prediction of Synthetic Shear Wave Velocity Logs in Upper Assam Basin

Authors: Triveni Gogoi, Rima Chatterjee

Abstract:

Shear wave velocity (Vs) estimation is an important approach in the seismic exploration and characterization of a hydrocarbon reservoir. There are varying methods for prediction of S-wave velocity, if recorded S-wave log is not available. But all the available methods for Vs prediction are empirical mathematical models. Shear wave velocity can be estimated using P-wave velocity by applying Castagna’s equation, which is the most common approach. The constants used in Castagna’s equation vary for different lithologies and geological set-ups. In this study, multiple regression analysis has been used for estimation of S-wave velocity. The EMERGE module from Hampson-Russel software has been used here for generation of S-wave log. Both single attribute and multi attributes analysis have been carried out for generation of synthetic S-wave log in Upper Assam basin. Upper Assam basin situated in North Eastern India is one of the most important petroleum provinces of India. The present study was carried out using four wells of the study area. Out of these wells, S-wave velocity was available for three wells. The main objective of the present study is a prediction of shear wave velocities for wells where S-wave velocity information is not available. The three wells having S-wave velocity were first used to test the reliability of the method and the generated S-wave log was compared with actual S-wave log. Single attribute analysis has been carried out for these three wells within the depth range 1700-2100m, which corresponds to Barail group of Oligocene age. The Barail Group is the main target zone in this study, which is the primary producing reservoir of the basin. A system generated list of attributes with varying degrees of correlation appeared and the attribute with the highest correlation was concerned for the single attribute analysis. Crossplot between the attributes shows the variation of points from line of best fit. The final result of the analysis was compared with the available S-wave log, which shows a good visual fit with a correlation of 72%. Next multi-attribute analysis has been carried out for the same data using all the wells within the same analysis window. A high correlation of 85% has been observed between the output log from the analysis and the recorded S-wave. The almost perfect fit between the synthetic S-wave and the recorded S-wave log validates the reliability of the method. For further authentication, the generated S-wave data from the wells have been tied to the seismic and correlated them. Synthetic share wave log has been generated for the well M2 where S-wave is not available and it shows a good correlation with the seismic. Neutron porosity, density, AI and P-wave velocity are proved to be the most significant variables in this statistical method for S-wave generation. Multilinear regression method thus can be considered as a reliable technique for generation of shear wave velocity log in this study.

Keywords: Castagna's equation, multi linear regression, multi attribute analysis, shear wave logs

Procedia PDF Downloads 233
7199 Two-Warehouse Inventory Model for Deteriorating Items with Inventory-Level-Dependent Demand under Two Dispatching Policies

Authors: Lei Zhao, Zhe Yuan, Wenyue Kuang

Abstract:

This paper studies two-warehouse inventory models for a deteriorating item considering that the demand is influenced by inventory levels. The problem mainly focuses on the optimal order policy and the optimal order cycle with inventory-level-dependent demand in two-warehouse system for retailers. It considers the different deterioration rates and the inventory holding costs in owned warehouse (OW) and rented warehouse (RW), and the conditions of transportation cost, allowed shortage and partial backlogging. Two inventory models are formulated: last-in first-out (LIFO) model and first-in-first-out (FIFO) model based on the policy choices of LIFO and FIFO, and a comparative analysis of LIFO model and FIFO model is made. The study finds that the FIFO policy is more in line with realistic operating conditions. Especially when the inventory holding cost of OW is high, and there is no difference or big difference between deterioration rates of OW and RW, the FIFO policy has better applicability. Meanwhile, this paper considers the differences between the effects of warehouse and shelf inventory levels on demand, and then builds retailers’ inventory decision model and studies the factors of the optimal order quantity, the optimal order cycle and the average inventory cost per unit time. To minimize the average total cost, the optimal dispatching policies are provided for retailers’ decisions.

Keywords: FIFO model, inventory-level-dependent, LIFO model, two-warehouse inventory

Procedia PDF Downloads 284
7198 Survival Analysis Based Delivery Time Estimates for Display FAB

Authors: Paul Han, Jun-Geol Baek

Abstract:

In the flat panel display industry, the scheduler and dispatching system to meet production target quantities and the deadline of production are the major production management system which controls each facility production order and distribution of WIP (Work in Process). In dispatching system, delivery time is a key factor for the time when a lot can be supplied to the facility. In this paper, we use survival analysis methods to identify main factors and a forecasting model of delivery time. Of survival analysis techniques to select important explanatory variables, the cox proportional hazard model is used to. To make a prediction model, the Accelerated Failure Time (AFT) model was used. Performance comparisons were conducted with two other models, which are the technical statistics model based on transfer history and the linear regression model using same explanatory variables with AFT model. As a result, the Mean Square Error (MSE) criteria, the AFT model decreased by 33.8% compared to the existing prediction model, decreased by 5.3% compared to the linear regression model. This survival analysis approach is applicable to implementing a delivery time estimator in display manufacturing. And it can contribute to improve the productivity and reliability of production management system.

Keywords: delivery time, survival analysis, Cox PH model, accelerated failure time model

Procedia PDF Downloads 547
7197 DC/DC Boost Converter Applied to Photovoltaic Pumping System Application

Authors: S. Abdourraziq, M. A. Abdourraziq

Abstract:

One of the most famous and important applications of solar energy systems is water pumping. It is often used for irrigation or to supply water in countryside or private firm. However, the cost and the efficiency are still a concern, especially with a continued variation of solar radiation and temperature throughout the day. Then, the improvement of the efficiency of the system components is one of the different solutions to reducing the cost. In this paper, we will present a detailed definition of each element of a PV pumping system, and we will present the different MPPT algorithm used in the literature. Our system consists of a PV panel, a boost converter, a motor-pump set, and a storage tank.

Keywords: PV cell, converter, MPPT, MPP, PV pumping system

Procedia PDF Downloads 163
7196 Crack Width Analysis of Reinforced Concrete Members under Shrinkage Effect by Pseudo-Discrete Crack Model

Authors: F. J. Ma, A. K. H. Kwan

Abstract:

Crack caused by shrinkage movement of concrete is a serious problem especially when restraint is provided. It may cause severe serviceability and durability problems. The existing prediction methods for crack width of concrete due to shrinkage movement are mainly numerical methods under simplified circumstances, which do not agree with each other. To get a more unified prediction method applicable to more sophisticated circumstances, finite element crack width analysis for shrinkage effect should be developed. However, no existing finite element analysis can be carried out to predict the crack width of concrete due to shrinkage movement because of unsolved reasons of conventional finite element analysis. In this paper, crack width analysis implemented by finite element analysis is presented with pseudo-discrete crack model, which combines traditional smeared crack model and newly proposed crack queuing algorithm. The proposed pseudo-discrete crack model is capable of simulating separate and single crack without adopting discrete crack element. And the improved finite element analysis can successfully simulate the stress redistribution when concrete is cracked, which is crucial for predicting crack width, crack spacing and crack number.

Keywords: crack queuing algorithm, crack width analysis, finite element analysis, shrinkage effect

Procedia PDF Downloads 422
7195 Predictive Pathogen Biology: Genome-Based Prediction of Pathogenic Potential and Countermeasures Targets

Authors: Debjit Ray

Abstract:

Horizontal gene transfer (HGT) and recombination leads to the emergence of bacterial antibiotic resistance and pathogenic traits. HGT events can be identified by comparing a large number of fully sequenced genomes across a species or genus, define the phylogenetic range of HGT, and find potential sources of new resistance genes. In-depth comparative phylogenomics can also identify subtle genome or plasmid structural changes or mutations associated with phenotypic changes. Comparative phylogenomics requires that accurately sequenced, complete and properly annotated genomes of the organism. Assembling closed genomes requires additional mate-pair reads or “long read” sequencing data to accompany short-read paired-end data. To bring down the cost and time required of producing assembled genomes and annotating genome features that inform drug resistance and pathogenicity, we are analyzing the performance for genome assembly of data from the Illumina NextSeq, which has faster throughput than the Illumina HiSeq (~1-2 days versus ~1 week), and shorter reads (150bp paired-end versus 300bp paired end) but higher capacity (150-400M reads per run versus ~5-15M) compared to the Illumina MiSeq. Bioinformatics improvements are also needed to make rapid, routine production of complete genomes a reality. Modern assemblers such as SPAdes 3.6.0 running on a standard Linux blade are capable in a few hours of converting mixes of reads from different library preps into high-quality assemblies with only a few gaps. Remaining breaks in scaffolds are generally due to repeats (e.g., rRNA genes) are addressed by our software for gap closure techniques, that avoid custom PCR or targeted sequencing. Our goal is to improve the understanding of emergence of pathogenesis using sequencing, comparative genomics, and machine learning analysis of ~1000 pathogen genomes. Machine learning algorithms will be used to digest the diverse features (change in virulence genes, recombination, horizontal gene transfer, patient diagnostics). Temporal data and evolutionary models can thus determine whether the origin of a particular isolate is likely to have been from the environment (could it have evolved from previous isolates). It can be useful for comparing differences in virulence along or across the tree. More intriguing, it can test whether there is a direction to virulence strength. This would open new avenues in the prediction of uncharacterized clinical bugs and multidrug resistance evolution and pathogen emergence.

Keywords: genomics, pathogens, genome assembly, superbugs

Procedia PDF Downloads 199
7194 Transaction Cost Analysis, Execution Quality, and Best Execution under MiFID II

Authors: Rodrigo Zepeda

Abstract:

Transaction cost analysis (TCA) is a way of analyzing the relative performance of different intermediaries and different trading strategies for trades undertaken in financial instruments. It is a way for an investor to determine the overall quality of execution of a particular trade, and there are many different approaches to undertaking TCA. Under the updated Markets in Financial Instruments Directive (2014/65/EU) (MiFID II), investment firms are required when executing orders, to take all sufficient steps to obtain the best possible result for their clients. This requirement for 'Best Execution' must take into account price, costs, speed, likelihood of execution and settlement, size, nature or any other consideration relevant to the execution of the order. The new regulatory compliance framework under MiFID II will now also apply across a very broad range of financial instruments. This article will provide a comprehensive technical analysis of how TCA and Best Execution will significantly change under MiFID II. It will also explain why harmonization of post-trade reporting requirements under MiFID II could potentially support the development of peer group analysis, which in turn could provide a new and highly advanced framework for TCA that could more effectively support Best Execution requirements under MiFID II. The study is significant because there are no studies that have dealt with TCA and Best Execution under MiFID II in the literature.

Keywords: transaction cost analysis, execution quality, best execution, MiFID II, financial instruments

Procedia PDF Downloads 295
7193 A Machine Learning Approach for Intelligent Transportation System Management on Urban Roads

Authors: Ashish Dhamaniya, Vineet Jain, Rajesh Chouhan

Abstract:

Traffic management is one of the gigantic issue in most of the urban roads in al-most all metropolitan cities in India. Speed is one of the critical traffic parameters for effective Intelligent Transportation System (ITS) implementation as it decides the arrival rate of vehicles on an intersection which are majorly the point of con-gestions. The study aimed to leverage Machine Learning (ML) models to produce precise predictions of speed on urban roadway links. The research objective was to assess how categorized traffic volume and road width, serving as variables, in-fluence speed prediction. Four tree-based regression models namely: Decision Tree (DT), Random Forest (RF), Extra Tree (ET), and Extreme Gradient Boost (XGB)are employed for this purpose. The models' performances were validated using test data, and the results demonstrate that Random Forest surpasses other machine learning techniques and a conventional utility theory-based model in speed prediction. The study is useful for managing the urban roadway network performance under mixed traffic conditions and effective implementation of ITS.

Keywords: stream speed, urban roads, machine learning, traffic flow

Procedia PDF Downloads 75
7192 Reliability Improvement of Power System Networks Using Adaptive Genetic Algorithm

Authors: Alireza Alesaadi

Abstract:

Reliability analysis is a powerful method for determining the weak points of the electrical networks. In designing of electrical network, it is tried to design the most reliable network with minimal system shutting down, but it is usually associated with increasing the cost. In this paper, using adaptive genetic algorithm, a method was presented that provides the most reliable system with a certain economical cost. Finally, the proposed method is applied to a sample network and results will be analyzed.

Keywords: reliability, adaptive genetic algorithm, electrical network, communication engineering

Procedia PDF Downloads 514
7191 Identification of Potential Predictive Biomarkers for Early Diagnosis of Preeclampsia Growth Factors to microRNAs

Authors: Sadia Munir

Abstract:

Preeclampsia is the contributor to the worldwide maternal mortality of approximately 100,000 deaths a year. It complicates about 10% of all pregnancies and is the first cause of maternal admission to intensive care units. Predicting preeclampsia is a major challenge in obstetrics. More importantly, no major progress has been achieved in the treatment of preeclampsia. As placenta is the main cause of the disease, the only way to treat the disease is to extract placental and deliver the baby. In developed countries, the cost of an average case of preeclampsia is estimated at £9000. Interestingly, preeclampsia may have an impact on the health of mother or infant, beyond the pregnancy. We performed a systematic search of PubMed including the combination of terms such as preeclampsia, biomarkers, treatment, hypoxia, inflammation, oxidative stress, vascular endothelial growth factor A, activin A, inhibin A, placental growth factor, transforming growth factor β-1, Nodal, placenta, trophoblast cells, microRNAs. In this review, we have summarized current knowledge on the identification of potential biomarkers for the diagnosis of preeclampsia. Although these studies show promising data in early diagnosis of preeclampsia, the current value of these factors as biomarkers, for the precise prediction of preeclampsia, has its limitation. Therefore, future studies need to be done to support some of the very promising and interesting data to develop affordable and widely available tests for early detection and treatment of preeclampsia.

Keywords: activin, biomarkers, growth factors, miroRNA

Procedia PDF Downloads 446
7190 Shark Detection and Classification with Deep Learning

Authors: Jeremy Jenrette, Z. Y. C. Liu, Pranav Chimote, Edward Fox, Trevor Hastie, Francesco Ferretti

Abstract:

Suitable shark conservation depends on well-informed population assessments. Direct methods such as scientific surveys and fisheries monitoring are adequate for defining population statuses, but species-specific indices of abundance and distribution coming from these sources are rare for most shark species. We can rapidly fill these information gaps by boosting media-based remote monitoring efforts with machine learning and automation. We created a database of shark images by sourcing 24,546 images covering 219 species of sharks from the web application spark pulse and the social network Instagram. We used object detection to extract shark features and inflate this database to 53,345 images. We packaged object-detection and image classification models into a Shark Detector bundle. We developed the Shark Detector to recognize and classify sharks from videos and images using transfer learning and convolutional neural networks (CNNs). We applied these models to common data-generation approaches of sharks: boosting training datasets, processing baited remote camera footage and online videos, and data-mining Instagram. We examined the accuracy of each model and tested genus and species prediction correctness as a result of training data quantity. The Shark Detector located sharks in baited remote footage and YouTube videos with an average accuracy of 89\%, and classified located subjects to the species level with 69\% accuracy (n =\ eight species). The Shark Detector sorted heterogeneous datasets of images sourced from Instagram with 91\% accuracy and classified species with 70\% accuracy (n =\ 17 species). Data-mining Instagram can inflate training datasets and increase the Shark Detector’s accuracy as well as facilitate archiving of historical and novel shark observations. Base accuracy of genus prediction was 68\% across 25 genera. The average base accuracy of species prediction within each genus class was 85\%. The Shark Detector can classify 45 species. All data-generation methods were processed without manual interaction. As media-based remote monitoring strives to dominate methods for observing sharks in nature, we developed an open-source Shark Detector to facilitate common identification applications. Prediction accuracy of the software pipeline increases as more images are added to the training dataset. We provide public access to the software on our GitHub page.

Keywords: classification, data mining, Instagram, remote monitoring, sharks

Procedia PDF Downloads 125
7189 Design Modification in CNC Milling Machine to Reduce the Weight of Structure

Authors: Harshkumar K. Desai, Anuj K. Desai, Jay P. Patel, Snehal V. Trivedi, Yogendrasinh Parmar

Abstract:

The need of continuous improvement in a product or process in this era of global competition leads to apply value engineering for functional and aesthetic improvement in consideration with economic aspect too. Solar industries located at G.I.D.C., Makarpura, Vadodara, Gujarat, India; a manufacturer of variety of CNC Machines had a challenge to analyze the structural design of column, base, carriage and table of CNC Milling Machine in the account of reduction of overall weight of a machine without affecting the rigidity and accuracy at the time of operation. The identified task is the first attempt to validate and optimize the proposed design of ribbed structure statically using advanced modeling and analysis tools in a systematic way. Results of stress and deformation obtained using analysis software are validated with theoretical analysis and found quite satisfactory. Such optimized results offer a weight reduction of the final assembly which is desired by manufacturers in favor of reduction of material cost, processing cost and handling cost finally.

Keywords: CNC milling machine, optimization, finite element analysis (FEA), weight reduction

Procedia PDF Downloads 281
7188 Supply Chain Coordination under Carbon Trading Mechanism in Case of Conflict

Authors: Fuqiang Wang, Jun Liu, Liyan Cai

Abstract:

This paper investigates the coordination of the conflicting two-stage low carbon supply chain consisting of upstream and downstream manufacturers. The conflict means that the upstream manufacturer takes action for carbon emissions reduction under carbon trading mechanism while the downstream manufacturer’s production cost rises. It assumes for the Stackelberg game that the upstream manufacturer plays as a leader and the downstream manufacturer does as a follower. Four kinds of the situation of decentralized decision making, centralized decision-making, the production cost sharing contract and the carbon emissions reduction revenue sharing contract under decentralized decision making are considered. The backward induction approach is adopted to solve the game. The results show that the more intense the conflict is, the lower the efficiency of carbon emissions reduction and the higher the retail price is. The optimal investment of the decentralized supply chain under the two contracts is unchanged and still lower than that of the centralized supply chain. Both the production cost sharing contract and the carbon emissions reduction revenue sharing contract cannot coordinate the supply chain, because that the sharing cost or carbon emissions reduction sharing revenue will transfer through the wholesale price mechanism. As a result, it requires more complicated contract forms to coordinate such a supply chain.

Keywords: cap-and-trade mechanism, carbon emissions reduction, conflict, supply chain coordination

Procedia PDF Downloads 343