Search results for: product modelling
550 Daily Probability Model of Storm Events in Peninsular Malaysia
Authors: Mohd Aftar Abu Bakar, Noratiqah Mohd Ariff, Abdul Aziz Jemain
Abstract:
Storm Event Analysis (SEA) provides a method to define rainfalls events as storms where each storm has its own amount and duration. By modelling daily probability of different types of storms, the onset, offset and cycle of rainfall seasons can be determined and investigated. Furthermore, researchers from the field of meteorology will be able to study the dynamical characteristics of rainfalls and make predictions for future reference. In this study, four categories of storms; short, intermediate, long and very long storms; are introduced based on the length of storm duration. Daily probability models of storms are built for these four categories of storms in Peninsular Malaysia. The models are constructed by using Bernoulli distribution and by applying linear regression on the first Fourier harmonic equation. From the models obtained, it is found that daily probability of storms at the Eastern part of Peninsular Malaysia shows a unimodal pattern with high probability of rain beginning at the end of the year and lasting until early the next year. This is very likely due to the Northeast monsoon season which occurs from November to March every year. Meanwhile, short and intermediate storms at other regions of Peninsular Malaysia experience a bimodal cycle due to the two inter-monsoon seasons. Overall, these models indicate that Peninsular Malaysia can be divided into four distinct regions based on the daily pattern for the probability of various storm events.
Keywords: Daily probability model, monsoon seasons, regions, storm events.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1636549 Implementation of Congestion Management Strategies on Arterial Roads: Case Study of Geelong
Authors: A. Das, L. Hitihamillage, S. Moridpour
Abstract:
Natural disasters are inevitable to the biodiversity. Disasters such as flood, tsunami and tornadoes could be brutal, harsh and devastating. In Australia, flooding is a major issue experienced by different parts of the country. In such crisis, delays in evacuation could decide the life and death of the people living in those regions. Congestion management could become a mammoth task if there are no steps taken before such situations. In the past to manage congestion in such circumstances, many strategies were utilised such as converting the road shoulders to extra lanes or changing the road geometry by adding more lanes. However, expansion of road to resolving congestion problems is not considered a viable option nowadays. The authorities avoid this option due to many reasons, such as lack of financial support and land space. They tend to focus their attention on optimising the current resources they possess and use traffic signals to overcome congestion problems. Traffic Signal Management strategy was considered a viable option, to alleviate congestion problems in the City of Geelong, Victoria. Arterial road with signalised intersections considered in this paper and the traffic data required for modelling collected from VicRoads. Traffic signalling software SIDRA used to model the roads, and the information gathered from VicRoads. In this paper, various signal parameters utilised to assess and improve the corridor performance to achieve the best possible Level of Services (LOS) for the arterial road.
Keywords: Congestion, constraints, management, LOS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 969548 Data-Driven Decision-Making in Digital Entrepreneurship
Authors: Abeba Nigussie Turi, Xiangming Samuel Li
Abstract:
Data-driven business models are more typical for established businesses than early-stage startups that strive to penetrate a market. This paper provided an extensive discussion on the principles of data analytics for early-stage digital entrepreneurial businesses. Here, we developed data-driven decision-making (DDDM) framework that applies to startups prone to multifaceted barriers in the form of poor data access, technical and financial constraints, to state some. The startup DDDM framework proposed in this paper is novel in its form encompassing startup data analytics enablers and metrics aligning with startups' business models ranging from customer-centric product development to servitization which is the future of modern digital entrepreneurship.
Keywords: Startup data analytics, data-driven decision-making, data acquisition, data generation, digital entrepreneurship.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 836547 Empirical Roughness Progression Models of Heavy Duty Rural Pavements
Authors: Nahla H. Alaswadko, Rayya A. Hassan, Bayar N. Mohammed
Abstract:
Empirical deterministic models have been developed to predict roughness progression of heavy duty spray sealed pavements for a dataset representing rural arterial roads. The dataset provides a good representation of the relevant network and covers a wide range of operating and environmental conditions. A sample with a large size of historical time series data for many pavement sections has been collected and prepared for use in multilevel regression analysis. The modelling parameters include road roughness as performance parameter and traffic loading, time, initial pavement strength, reactivity level of subgrade soil, climate condition, and condition of drainage system as predictor parameters. The purpose of this paper is to report the approaches adopted for models development and validation. The study presents multilevel models that can account for the correlation among time series data of the same section and to capture the effect of unobserved variables. Study results show that the models fit the data very well. The contribution and significance of relevant influencing factors in predicting roughness progression are presented and explained. The paper concludes that the analysis approach used for developing the models confirmed their accuracy and reliability by well-fitting to the validation data.
Keywords: Roughness progression, empirical model, pavement performance, heavy duty pavement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 805546 Modeling of Silicon Solar Cell with Anti-Reflecting Coating
Authors: Ankita Gaur, Mouli Karmakar, Shyam
Abstract:
In this study, a silicon solar cell has been modeled and analyzed to enhance its performance by improving the optical properties using an anti-reflecting coating (ARC). The dynamic optical reflectance, transmittance along with the net transmissivity absorptivity product of each layer are assessed as per the diurnal variation of the angle of incidence using MATLAB 2019. The model is tested with various anti-reflective coatings and the performance has also been compared with uncoated cells. ARC improves the optical transmittance of the photon. Higher transmittance of ⁓96.57% with lowest reflectance of ⁓ 1.74% at 12.00 hours was obtained with MgF2 coated silicon cells. The electrical efficiency of the configured solar cell was evaluated for a composite climate of New Delhi, India, for all weather conditions. The annual electricity generation for anti-reflective coated and uncoated crystalline silicon PV Module was observed to be 103.14 KWh and 99.51 KWh, respectively.
Keywords: Anti-reflecting coating, electrical efficiency, reflectance, solar cell, transmittance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 536545 Production and Remanufacturing of Returned Products in Supply Chain using Modified Genetic Algorithm
Authors: Siva Prasad Darla, C. D. Naiju, K. Annamalai, Y. Upendra Sravan
Abstract:
In recent years, environment regulation forcing manufactures to consider recovery activity of end-of- life products and/or return products for refurbishing, recycling, remanufacturing/repair and disposal in supply chain management. In this paper, a mathematical model is formulated for single product production-inventory system considering remanufacturing/reuse of return products and rate of return products follows a demand like function, dependent on purchasing price and acceptance quality level. It is useful in decision making to determine whether to go for remanufacturing or disposal of returned products along with newly produced products to satisfy a stationary demand. In addition, a modified genetic algorithm approach is proposed, inspired by particle swarm optimization method. Numerical analysis of the case study is carried out to validate the model.Keywords: Genetic Algorithm, Particle Swarm Optimization, Production, Remanufacturing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1714544 Verification of Protocol Design using UML - SMV
Authors: Prashanth C.M., K. Chandrashekar Shet
Abstract:
In recent past, the Unified Modeling Language (UML) has become the de facto industry standard for object-oriented modeling of the software systems. The syntax and semantics rich UML has encouraged industry to develop several supporting tools including those capable of generating deployable product (code) from the UML models. As a consequence, ensuring the correctness of the model/design has become challenging and extremely important task. In this paper, we present an approach for automatic verification of protocol model/design. As a case study, Session Initiation Protocol (SIP) design is verified for the property, “the CALLER will not converse with the CALLEE before the connection is established between them ". The SIP is modeled using UML statechart diagrams and the desired properties are expressed in temporal logic. Our prototype verifier “UML-SMV" is used to carry out the verification. We subjected an erroneous SIP model to the UML-SMV, the verifier could successfully detect the error (in 76.26ms) and generate the error trace.
Keywords: Unified Modeling Language, Statechart, Verification, Protocol Design, Model Checking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1857543 Multi-Level Air Quality Classification in China Using Information Gain and Support Vector Machine
Authors: Bingchun Liu, Pei-Chann Chang, Natasha Huang, Dun Li
Abstract:
Machine Learning and Data Mining are the two important tools for extracting useful information and knowledge from large datasets. In machine learning, classification is a wildly used technique to predict qualitative variables and is generally preferred over regression from an operational point of view. Due to the enormous increase in air pollution in various countries especially China, Air Quality Classification has become one of the most important topics in air quality research and modelling. This study aims at introducing a hybrid classification model based on information theory and Support Vector Machine (SVM) using the air quality data of four cities in China namely Beijing, Guangzhou, Shanghai and Tianjin from Jan 1, 2014 to April 30, 2016. China's Ministry of Environmental Protection has classified the daily air quality into 6 levels namely Serious Pollution, Severe Pollution, Moderate Pollution, Light Pollution, Good and Excellent based on their respective Air Quality Index (AQI) values. Using the information theory, information gain (IG) is calculated and feature selection is done for both categorical features and continuous numeric features. Then SVM Machine Learning algorithm is implemented on the selected features with cross-validation. The final evaluation reveals that the IG and SVM hybrid model performs better than SVM (alone), Artificial Neural Network (ANN) and K-Nearest Neighbours (KNN) models in terms of accuracy as well as complexity.
Keywords: Machine learning, air quality classification, air quality index, information gain, support vector machine, cross-validation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 950542 Modelling Phytoremediation Rates of Aquatic Macrophytes in Aquaculture Effluent
Authors: E. A. Kiridi, A. O. Ogunlela
Abstract:
Pollutants from aquacultural practices constitute environmental problems and phytoremediation could offer cheaper environmentally sustainable alternative since equipment using advanced treatment for fish tank effluent is expensive to import, install, operate and maintain, especially in developing countries. The main objective of this research was, therefore, to develop a mathematical model for phytoremediation by aquatic plants in aquaculture wastewater. Other objectives were to evaluate the retention times on phytoremediation rates using the model and to measure the nutrient level of the aquaculture effluent and phytoremediation rates of three aquatic macrophytes, namely; water hyacinth (Eichornia crassippes), water lettuce (Pistial stratoites) and morning glory (Ipomea asarifolia). A completely randomized experimental design was used in the study. Approximately 100 g of each macrophyte were introduced into the hydroponic units and phytoremediation indices monitored at 8 different intervals from the first to the 28th day. The water quality parameters measured were pH and electrical conductivity (EC). Others were concentration of ammonium–nitrogen (NH4+ -N), nitrite- nitrogen (NO2- -N), nitrate- nitrogen (NO3- -N), phosphate –phosphorus (PO43- -P), and biomass value. The biomass produced by water hyacinth was 438.2 g, 600.7 g, 688.2 g and 725.7 g at four 7–day intervals. The corresponding values for water lettuce were 361.2 g, 498.7 g, 561.2 g and 623.7 g and for morning glory were 417.0 g, 567.0 g, 642.0 g and 679.5g. Coefficient of determination was greater than 80% for EC, TDS, NO2- -N, NO3- -N and 70% for NH4+ -N using any of the macrophytes and the predicted values were within the 95% confidence interval of measured values. Therefore, the model is valuable in the design and operation of phytoremediation systems for aquaculture effluent.
Keywords: Phytoremediation, macrophytes, hydroponic unit, aquaculture effluent, mathematical model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1631541 A Novel Cold Asphalt Concrete Mixture for Heavily Trafficked Binder Course
Authors: A. Dulaimi, H. Al Nageim, F. Ruddock, L. Seton
Abstract:
This study aims at developing a novel cold asphalt concrete binder course mixture by using Ordinary Portland Cement (OPC) as a replacement for conventional mineral filler (0%-100%) with new by-product material (LJMU-A2) used as a supplementary cementitious material. With this purpose, cold asphalt concrete binder course mixtures with cationic emulsions were studied by means of stiffness modulus whereas water sensitivity was assessed by measuring the stiffness modulus ratio before and after sample conditioning. The results indicate that a substantial enhancement in the stiffness modulus and a considerable improvement of water sensitivity resistance is achieved by adding LJMU-A2 to the cold asphalt mixtures as a supplementary cementitious material. Moreover, the addition of LJMU-A2 to those mixtures leads to a stiffness modulus after 2-day curing compared to that obtained with Portland cement, which occurs after 7-day curing.Keywords: Binder course, cold mix asphalt, cement, stiffness modulus, water sensitivity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3095540 Effect of Confinement on the Bearing Capacity and Settlement of Spread Foundations
Authors: Tahsin Toma Sabbagh, Ihsan Al-Abboodi, Ali Al-Jazaairry
Abstract:
Allowable-bearing capacity is the competency of soil to safely carries the pressure from the superstructure without experiencing a shear failure with accompanying excessive settlements. Ensuring a safe bearing pressure with respect to failure does not tolerate settlement of the foundation will be within acceptable limits. Therefore, settlement analysis should always be performed since most structures are settlement sensitive. When visualising the movement of a soil wedge in the bearing capacity criterion, both vertically and horizontally, it becomes clear that by confining the soil surrounding the foundation, both the bearing capacity and settlement values improve. In this study, two sizes of spread foundation were considered; (2×4) m and (3×5) m. These represent two real problem case studies of an existing building. The foundations were analysed in terms of dimension as well as position with respect to a confining wall (i.e., sheet piles on both sides). Assuming B is the least foundation dimension, the study comprised the analyses of three distances; (0.1 B), (0.5 B), and (0.75 B) between the sheet piles and foundations alongside three depths of confinement (0.5 B), (1 B), and (1.5 B). Nonlinear three-dimensional finite element analysis (ANSYS) was adopted to perform an analytical investigation on the behaviour of the two foundations contained by the case study. Results showed that confinement of foundations reduced the overall stresses near the foundation by 65% and reduced the vertical displacement by 90%. Moreover, the most effective distance between the confinement wall and the foundation was found to be 0.5 B.
Keywords: Bearing capacity, cohesionless soils, spread footings, soil confinement, soil modelling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 892539 Providing a Practical Model to Reduce Maintenance Costs: A Case Study in Golgohar Company
Authors: Iman Atighi, Jalal Soleimannejad, Ahmad Akbarinasab, Saeid Moradpour
Abstract:
In the past, we could increase profit by increasing product prices. But in the new decade, a competitive market does not let us to increase profit with increase prices. Therefore, the only way to increase profit will be reduce costs. A significant percentage of production costs are the maintenance costs, and analysis of these costs could achieve more profit. Most maintenance strategies such as RCM (Reliability-Center-Maintenance), TPM (Total Productivity Maintenance), PM (Preventive Maintenance) etc., are trying to reduce maintenance costs. In this paper, decreasing the maintenance costs of Concentration Plant of Golgohar Company (GEG) was examined by using of MTBF (Mean Time between Failures) and MTTR (Mean Time to Repair) analyses. These analyses showed that instead of buying new machines and increasing costs in order to promote capacity, the improving of MTBF and MTTR indexes would solve capacity problems in the best way and decrease costs.
Keywords: Golgohar Iron Ore Mining & Industrial Company, maintainability, maintenance costs, reliability-center-maintenance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 660538 The Imaging Methods for Classifying Crispiness of Freeze-Dried Durian using Fuzzy Logic
Authors: Sitthichon Kanitthakun, Pinit Kumhom, Kosin Chamnongthai
Abstract:
In quality control of freeze-dried durian, crispiness is a key quality index of the product. Generally, crispy testing has to be done by a destructive method. A nondestructive testing of the crispiness is required because the samples can be reused for other kinds of testing. This paper proposed a crispiness classification method of freeze-dried durians using fuzzy logic for decision making. The physical changes of a freeze-dried durian include the pores appearing in the images. Three physical features including (1) the diameters of pores, (2) the ratio of the pore area and the remaining area, and (3) the distribution of the pores are considered to contribute to the crispiness. The fuzzy logic is applied for making the decision. The experimental results comparing with food expert opinion showed that the accuracy of the proposed classification method is 83.33 percent.Keywords: Durian, crispiness, freeze drying, pore, fuzzy logic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1973537 A Study of Lean Principles Implementation in the Libyan Healthcare and Industry Sectors
Authors: Nasser M. Amaitik, Ngwan F. Elsagzli
Abstract:
Lean technique is very important in the service and industrial fields. It is defined as an effective tool to eliminate the wastes. In lean the wastes are defined as anything which does not add value to the end product. There are wastes that can be avoided, but some are unavoidable for many reasons.
The present study aims to apply the principles of lean in two different sectors, healthcare and industry. Two case studies have been selected to apply the experimental work. The first case was Al-Jalaa Hospital, while the second case study was the Technical Company of Aluminum Sections in Benghazi, LIBYA. In both case studies the Value Stream Map (VSM) of the current state has been constructed. The proposed plans have been implemented by merging or eliminating procedures or processes.
The results obtained from both case studies showed improvement in Capacity, Idle time and Utilized time.
Keywords: Healthcare service delivery, Idle time, Lean principles, Utilized time, Value stream mapping, Wastes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2330536 Particle Swarm Optimization Based Genetic Algorithm for Two-Stage Transportation Supply Chain
Authors: Siva Prasad Darla, C. D. Naiju, K. Annamalai, S. S. Rajiv Sushanth
Abstract:
Supply chain consists of all stages involved, directly or indirectly, includes all functions involved in fulfilling a customer demand. In two stage transportation supply chain problem, transportation costs are of a significant proportion of final product costs. It is often crucial for successful decisions making approaches in two stage supply chain to explicit account for non-linear transportation costs. In this paper, deterministic demand and finite supply of products was considered. The optimized distribution level and the routing structure from the manufacturing plants to the distribution centres and to the end customers is determined using developed mathematical model and solved by proposed particle swarm optimization based genetic algorithm. Numerical analysis of the case study is carried out to validate the model.Keywords: Genetic Algorithm, Particle Swarm Optimization, Production, Remanufacturing
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1844535 Diffusion of Mobile Entertainment in Malaysia: Drivers and Barriers
Authors: C. C. Wong, P. L. Hiew
Abstract:
This research aims to examine the key success factors for the diffusion of mobile entertainment services in Malaysia. The drivers and barriers observed in this research include perceived benefit; concerns pertaining to pricing, product and technological standardization, privacy and security; as well as influences from peers and community. An analysis of a Malaysian survey of 384 respondents between 18 to 25 years shows that subscribers placed greater importance on perceived benefit of mobile entertainment services compared to other factors. Results of the survey also show that there are strong positive correlations between all the factors, with pricing issue–perceived benefit showing the strongest relationship. This paper aims to provide an extensive study on the drivers and barriers that could be used to derive architecture for entertainment service provision to serve as a guide for telcos to outline suitable approaches in order to encourage mass market adoption of mobile entertainment services in Malaysia.Keywords: Barriers, Correlations, Diffusion, Drivers, Mobile Entertainment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2291534 Design and Analysis of Gauge R&R Studies: Making Decisions Based on ANOVA Method
Authors: Afrooz Moatari Kazerouni
Abstract:
In a competitive production environment, critical decision making are based on data resulted by random sampling of product units. Efficiency of these decisions depends on data quality and also their reliability scale. This point leads to the necessity of a reliable measurement system. Therefore, the conjecture process and analysing the errors contributes to a measurement system known as Measurement System Analysis (MSA). The aim of this research is on determining the necessity and assurance of extensive development in analysing measurement systems, particularly with the use of Repeatability and Reproducibility Gages (GR&R) to improve physical measurements. Nowadays in productive industries, repeatability and reproducibility gages released so well but they are not applicable as well as other measurement system analysis methods. To get familiar with this method and gain a feedback in improving measurement systems, this survey would be on “ANOVA" method as the most widespread way of calculating Repeatability and Reproducibility (R&R).Keywords: Analysis of Variance (ANOVA), MeasurementSystem Analysis (MSA), Part-Operator interaction effect, Repeatability and Reproducibility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4673533 Economic Growth Relations to Domestic and International Air Passenger Transport in Brazil
Authors: Manoela Cabo da Silva, Elton Fernandes, Ricardo Pacheco, Heloisa Pires
Abstract:
This study examined cointegration and causal relationships between economic growth and regular domestic and international passenger air transport in Brazil. Total passengers embarked and disembarked were used as a proxy for air transport activity and gross domestic product (GDP) as a proxy for economic development. The test spanned the period from 2000 to 2015 for domestic passenger traffic and from 1995 to 2015 for international traffic. The results confirm the hypothesis that there is cointegration between passenger traffic series and economic development, showing a bi-directional Granger causal relationship between domestic traffic and economic development and unidirectional influence by economic growth on international passenger air transport demand. Variance decomposition of the series showed that domestic air transport was far more important than international transport to promoting economic development in Brazil.
Keywords: Air passenger transport, cointegration, economic growth, GDP, granger causality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1008532 Reliability Analysis of Underground Pipelines Using Subset Simulation
Authors: Kong Fah Tee, Lutfor Rahman Khan, Hongshuang Li
Abstract:
An advanced Monte Carlo simulation method, called Subset Simulation (SS) for the time-dependent reliability prediction for underground pipelines has been presented in this paper. The SS can provide better resolution for low failure probability level with efficient investigating of rare failure events which are commonly encountered in pipeline engineering applications. In SS method, random samples leading to progressive failure are generated efficiently and used for computing probabilistic performance by statistical variables. SS gains its efficiency as small probability event as a product of a sequence of intermediate events with larger conditional probabilities. The efficiency of SS has been demonstrated by numerical studies and attention in this work is devoted to scrutinise the robustness of the SS application in pipe reliability assessment. It is hoped that the development work can promote the use of SS tools for uncertainty propagation in the decision-making process of underground pipelines network reliability prediction.
Keywords: Underground pipelines, Probability of failure, Reliability and Subset Simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3559531 Temperature Control of Industrial Water Cooler using Hot-gas Bypass
Authors: Jung-in Yoon, Seung-taek Oh, Seung-moon Baek, Jun-hyuk Choi, Jong-yeong Byun, Seok-kwon Jeong, Choon-guen Moon
Abstract:
In this study, we experiment on precise control outlet temperature of water from the water cooler with hot-gas bypass method based on PI control logic for machine tool. Recently, technical trend for machine tools is focused on enhancement of speed and accuracy. High speedy processing causes thermal and structural deformation of objects from the machine tools. Water cooler has to be applied to machine tools to reduce the thermal negative influence with accurate temperature controlling system. The goal of this study is to minimize temperature error in steady state. In addition, control period of an electronic expansion valve were considered to increment of lifetime of the machine tools and quality of product with a water cooler.Keywords: Hot-gas bypass, Water cooler, PI control, Electronic Expansion Valve, Gain tuning
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3161530 Choice Experiment Approach on Evaluation of Non-Market Farming System Outputs: First Results from Lithuanian Case Study
Authors: A. Novikova, L. Rocchi, G. Startiene
Abstract:
Market and non-market outputs are produced jointly in agriculture. Their supply depends on the intensity and type of production. The role of agriculture as an economic activity and its effects are important for the Lithuanian case study, as agricultural land covers more than a half of country. Positive and negative externalities, created in agriculture are not considered in the market. Therefore, specific techniques such as stated preferences methods, in particular choice experiments (CE) are used for evaluation of non-market outputs in agriculture. The main aim of this paper is to present construction of the research path for evaluation of non-market farming system outputs in Lithuania. The conventional and organic farming, covering crops (including both cereal and industrial crops) and livestock (including dairy and cattle) production has been selected. The CE method and nested logit (NL) model were selected as appropriate for evaluation of non-market outputs of different farming systems in Lithuania. A pilot survey was implemented between October–November 2018, in order to test and improve the CE questionnaire. The results of the survey showed that the questionnaire is accepted and well understood by the respondents. The econometric modelling showed that the selected NL model could be used for the main survey. The understanding of the differences between organic and conventional farming by residents was identified. It was revealed that they are more willing to choose organic farming in comparison to conventional farming.
Keywords: Choice experiments, farming system, Lithuania market outputs, non-market outputs.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 605529 Supersonic Flow around a Dihedral Airfoil: Modeling and Experimentation Investigation
Authors: A. Naamane, M. Hasnaoui
Abstract:
Numerical modeling of fluid flows, whether compressible or incompressible, laminar or turbulent presents a considerable contribution in the scientific and industrial fields. However, the development of an approximate model of a supersonic flow requires the introduction of specific and more precise techniques and methods. For this purpose, the object of this paper is modeling a supersonic flow of inviscid fluid around a dihedral airfoil. Based on the thin airfoils theory and the non-dimensional stationary Steichen equation of a two-dimensional supersonic flow in isentropic evolution, we obtained a solution for the downstream velocity potential of the oblique shock at the second order of relative thickness that characterizes a perturbation parameter. This result has been dealt with by the asymptotic analysis and characteristics method. In order to validate our model, the results are discussed in comparison with theoretical and experimental results. Indeed, firstly, the comparison of the results of our model has shown that they are quantitatively acceptable compared to the existing theoretical results. Finally, an experimental study was conducted using the AF300 supersonic wind tunnel. In this experiment, we have considered the incident upstream Mach number over a symmetrical dihedral airfoil wing. The comparison of the different Mach number downstream results of our model with those of the existing theoretical data (relative margin between 0.07% and 4%) and with experimental results (concordance for a deflection angle between 1° and 11°) support the validation of our model with accuracy.
Keywords: Asymptotic modelling, dihedral airfoil, supersonic flow, supersonic wind tunnel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 718528 Analysis of Coal Tar Compositions Produced from Sub-Bituminous Kalimantan Coal Tar
Authors: D. S. Fardhyanti, A. Damayanti
Abstract:
Coal tar is a liquid by-product of coal pyrolysis processes. This liquid oil mixture contains various kinds of useful compounds such as benzoic aromatic compounds and phenolic compounds. These compounds are widely used as raw material for insecticides, dyes, medicines, perfumes, coloring matters, and many others. The coal tar was collected by pyrolysis process of coal obtained from PT Kaltim Prima Coal and Arutmin-Kalimantan. The experiments typically occurred at the atmospheric pressure in a laboratory furnace at temperatures ranging from 300 to 550oC with a heating rate of 10oC/min and a holding time of 1 hour at the pyrolysis temperature. The Gas Chromatography-Mass Spectroscopy (GC-MS) was used to analyze the coal tar components. The obtained coal tar has the viscosity of 3.12 cp, the density of 2.78 g/cm3, the calorific value of 11,048.44 cal/g, and the molecular weight of 222.67. The analysis result showed that the coal tar contained more than 78 chemical compounds such as benzene, cresol, phenol, xylene, naphtalene, etc. The total phenolic compounds contained in coal tar are 33.25% (PT KPC) and 17.58% (Arutmin-Kalimantan). The total naphtalene compounds contained in coal tar is 14.15% (PT KPC) and 17.13% (Arutmin-Kalimantan).Keywords: Coal tar, pyrolysis, gas chromatography-mass spectroscopy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3667527 DRE - A Quality Metric for Component based Software Products
Authors: K. S. Jasmine, R. Vasantha
Abstract:
The overriding goal of software engineering is to provide a high quality system, application or a product. To achieve this goal, software engineers must apply effective methods coupled with modern tools within the context of a mature software process [2]. In addition, it is also must to assure that high quality is realized. Although many quality measures can be collected at the project levels, the important measures are errors and defects. Deriving a quality measure for reusable components has proven to be challenging task now a days. The results obtained from the study are based on the empirical evidence of reuse practices, as emerged from the analysis of industrial projects. Both large and small companies, working in a variety of business domains, and using object-oriented and procedural development approaches contributed towards this study. This paper proposes a quality metric that provides benefit at both project and process level, namely defect removal efficiency (DRE).Keywords: Software Reuse, Defect density, Reuse metrics, Defect Removal efficiency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2811526 Preparation and Properties of Biopolymer from L-Lactide (LL) and ε-Caprolactone (CL)
Authors: A. Buasri, N. Chaiyut, K. Iamma, K. Kongcharoen, K. Cheunsakulpong
Abstract:
Biopolymers have gained much attention as ecofriendly alternatives to petrochemical-based plastics because they are biodegradable and can be produced from renewable feedstocks. One class of biopolyester with many potential environmentally friendly applications is polylactic acid (PLA) and polycaprolactone (PCL). The PLA/PCL biodegradable copolyesters were synthesized by bulk ring-opening copolymerization of successively added Llactide (LL) and ε-caprolactone (CL) in the presence of toluene, using 1-hexanol as initiator and stannous octoate (Sn(Oct)2) as catalyst. Reaction temperature, reaction time and amount of catalyst were evaluated to obtain optimum reaction conditions. The results showed that the %conversion increased with increases in reaction temperature and reaction time, but after a critical amount of catalyst was reached the %conversion decreased. The yield of PLA/PCL biopolymer achieved 98.02% at the reaction temperature 160 °C, amount of catalyst 0.3 mol% and reaction time of 48 h. In addition, the thermal properties of the product were determined by differential scanning calorimetry (DSC) and thermogravimetric analysis (TGA).
Keywords: Biopolymer, Polylactic Acid (PLA), Polycaprolactone (PCL), L-Lactide (LL), ε-Caprolactone (CL)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4502525 Calibration of the Discrete Element Method Using a Large Shear Box
Authors: Corné J. Coetzee, Etienne Horn
Abstract:
One of the main challenges in using the Discrete Element Method (DEM) is to specify the correct input parameter values. In general, the models are sensitive to the input parameter values and accurate results can only be achieved if the correct values are specified. For the linear contact model, micro-parameters such as the particle density, stiffness, coefficient of friction, as well as the particle size and shape distributions are required. There is a need for a procedure to accurately calibrate these parameters before any attempt can be made to accurately model a complete bulk materials handling system. Since DEM is often used to model applications in the mining and quarrying industries, a calibration procedure was developed for materials that consist of relatively large (up to 40 mm in size) particles. A coarse crushed aggregate was used as the test material. Using a specially designed large shear box with a diameter of 590 mm, the confined Young’s modulus (bulk stiffness) and internal friction angle of the material were measured by means of the confined compression test and the direct shear test respectively. DEM models of the experimental setup were developed and the input parameter values were varied iteratively until a close correlation between the experimental and numerical results was achieved. The calibration process was validated by modelling the pull-out of an anchor from a bed of material. The model results compared well with experimental measurement.Keywords: Discrete Element Method (DEM), calibration, shear box, anchor pull-out.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2675524 Analysis of Production Loss on a Linear Walking Worker Line
Authors: Qian Wang, Sylvain Lassalle, Antony R. Mileham, Geraint W. Owen
Abstract:
This paper mathematically analyses the varying magnitude of production loss, which may occur due to idle time (inprocess waiting time and traveling time) on a linear walking worker assembly line. Within this flexible and reconfigurable assembly system, each worker travels down the line carrying out each assembly task at each station; and each worker accomplishes the assembly of a unit from start to finish and then travels back to the first station to start the assembly of a new product. This strategy of system design attempts to combine the flexibility of the U-shaped moving worker assembly cell with the efficiency of the conventional fixed worker assembly line. The paper aims to evaluate the effect of idle time that may offset the labor efficiency of each walking worker providing an insight into the mechanism of such a flexible and reconfigurable assembly system.Keywords: Production lines, manufacturing systems, assemblysystems, walking workers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1902523 Preparation of Vanadium Powder by Hydrogenation and Dehydrogenation
Authors: Weicai Yang, Xianfeng Dong, Dapeng Zeng, Bo Lin, Jun Tang
Abstract:
Low oxygen content vanadium powder was prepared by hydrogenation dehydrogenization (HDH). The effect of purification treatment on hydrogen absorption kinetics of dendritic vanadium was tested, and the effects of milling technique on powder yield and grain size were studied. The crystal phase, oxygen and nitrgen content, and grain size of prepared powder were characterized and analyzed by X-ray diffraction (XRD), oxygen and nitrogen analyzer and grain size analyzer. The results show that the alkaline cleaning can improve the hydrogen absorption of vanadium. The yield of vanadium hydride powder can reach as high as 90% by 4h ball-milling, The resultant product also have an oxygen content less than 600μg/g, and the grain size is smaller than 37μm. Meanwhile, the XRD results show that the phase of hydride vanadium powder is mainly VH0.81. After a hydrogen desorption treatment in vacuum at 700Ôäâ, the phase of the powder converts into V and a little of V2H.Keywords: V-5Cr-5Ti alloy, HDH, microstructures, mechanical properties.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1946522 Influence of Degradative Enzymatic Activities on the Shelf Life of Ready-to-Eat Prickly Pear Fruits
Authors: D. Scalone, R. Palmeri, F. Licciardello, G. Muratore, A. Todaro, G. Spagna
Abstract:
Prickly pear fruit (Opuntia ficus indica L. Miller) belongs to the Cactaceae family. This species is very sensitive to low storage temperatures (< 5°C) which cause damages. The fruits can be peeled, suitably packaged and successfully commercialized as a ready-to-eat product. The main limit to the extension of the shelf life is the production of off-flavors due to different factors, the growth of microorganisms and the action of endogenous enzymes. Lipoxygenase (LOX) and Pectinesterase (PE) are involved in fruit degradation. In particular, LOX pathway is directly responsible for lipid oxidation, and the subsequent production of off-flavours, while PE causes the softening of fruit during maturation. They act on the texture and shelf-life of post-harvest, packaged fruits, as a function of the the grown of microorganisms and packaging technologies used. The aim of this work is to compare the effect of different packaging technologies on the shelf life extension of ready-to-eat prickly pear fruits with regards for the enzymes activities.
Keywords: Enzymes, packaging, prickly pear, shelf life.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1664521 Extraction of Phenol, o-Cresol, and p-Cresol from Coal Tar: Effect of Temperature and Mixing
Authors: Dewi S. Fardhyanti, Panut Mulyono, Wahyudi B. Sediawan, Muslikhin Hidayat
Abstract:
Coal tar is a liquid by-product of the process of coal gasification and carbonation. This liquid oil mixture contains various kinds of useful compounds such as phenol, o-cresol, and p-cresol. These compounds are widely used as raw material for insecticides, dyes, medicines, perfumes, coloring matters, and many others. This research needed to be done that given the optimum conditions for the separation of phenol, o-cresol, and p-cresol from the coal tar by solvent extraction process. The aim of the present work was to study the effect of two kinds of aqueous were used as solvents: methanol and acetone solutions, the effect of temperature (298, 306, and 313K) and mixing (30, 35, and 40rpm) for the separation of phenol, o-cresol, and p-cresol from coal tar by solvent extraction. Results indicated that phenol, o-cresol, and p-cresol in coal tar were selectivity extracted into the solvent phase and these components could be separated by solvent extraction. The aqueous solution of methanol, mass ratio of solvent to feed, Eo/Ro=1, extraction temperature 306K and mixing 35 rpm were the most efficient for extraction of phenol, o-cresol, and p-cresol from coal tar.Keywords: Coal tar, Distribution coefficient, Extraction, Yield.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4534