Search results for: cost prediction
7397 Predicting Global Solar Radiation Using Recurrent Neural Networks and Climatological Parameters
Authors: Rami El-Hajj Mohamad, Mahmoud Skafi, Ali Massoud Haidar
Abstract:
Several meteorological parameters were used for the prediction of monthly average daily global solar radiation on horizontal using recurrent neural networks (RNNs). Climatological data and measures, mainly air temperature, humidity, sunshine duration, and wind speed between 1995 and 2007 were used to design and validate a feed forward and recurrent neural network based prediction systems. In this paper we present our reference system based on a feed-forward multilayer perceptron (MLP) as well as the proposed approach based on an RNN model. The obtained results were promising and comparable to those obtained by other existing empirical and neural models. The experimental results showed the advantage of RNNs over simple MLPs when we deal with time series solar radiation predictions based on daily climatological data.Keywords: recurrent neural networks, global solar radiation, multi-layer perceptron, gradient, root mean square error
Procedia PDF Downloads 4507396 Performance and Economics of Goats Fed Poultry Litter and Rumen Content
Authors: A. Mohammed, A. M. Umar, S. H. Adamu
Abstract:
The study was conducted to evaluate the growth performance and nutrients utilization using 20 entire males of Sahelian goats fed Rumen content (fore-stomach digest) and poultry litter waste (PLW) at various levels of inclusion. The experimental animals were randomly allocated to diet A (Control), B (10% each of FSD and PLW), C (6.67%PLW and 13.33 FSD) and D(13.33% PLW and 6.67% FDS) at the rate of five animals per treatment. After 90 days of feeding trial, It was observed that Diets D had best feed intake and body weight gain which might be due to the good palatability of PLW and less odour of FSD in the diet. Diet C had the least feed cost then followed by diet B and while diet A(control) was more expensive than other treatments. There was the significant difference (P<0.05) between the treatments in the cost of daily feed consumption. Treatment A had the highest value while treatment C recorded the lowest cost of daily feed consumption. There was no significant difference (P > 0.05) between all treatments in terms of Cost of feed kg/ live weight gain, where treatment B had the highest value while the lowest obtained in treatment D. However, it is recommended that more research trial should be carried out to ascertain the true value of incorporating poultry litter waste and fore-stomach digest.Keywords: poultry litter, rumen content, weight gain, economics
Procedia PDF Downloads 6477395 A Study on Performance Prediction in Early Design Stage of Apartment Housing Using Machine Learning
Authors: Seongjun Kim, Sanghoon Shim, Jinwooung Kim, Jaehwan Jung, Sung-Ah Kim
Abstract:
As the development of information and communication technology, the convergence of machine learning of the ICT area and design is attempted. In this way, it is possible to grasp the correlation between various design elements, which was difficult to grasp, and to reflect this in the design result. In architecture, there is an attempt to predict the performance, which is difficult to grasp in the past, by finding the correlation among multiple factors mainly through machine learning. In architectural design area, some attempts to predict the performance affected by various factors have been tried. With machine learning, it is possible to quickly predict performance. The aim of this study is to propose a model that predicts performance according to the block arrangement of apartment housing through machine learning and the design alternative which satisfies the performance such as the daylight hours in the most similar form to the alternative proposed by the designer. Through this study, a designer can proceed with the design considering various design alternatives and accurate performances quickly from the early design stage.Keywords: apartment housing, machine learning, multi-objective optimization, performance prediction
Procedia PDF Downloads 4887394 Prediction of Heavy-Weight Impact Noise and Vibration of Floating Floor Using Modified Impact Spectrum
Authors: Ju-Hyung Kim, Dae-Ho Mun, Hong-Gun Park
Abstract:
When an impact is applied to a floating floor, noise and vibration response of high-frequency range is reduced effectively, while amplifies the response at low-frequency range. This means floating floor can make worse noise condition when heavy-weight impact is applied. The amplified response is the result of interaction between finishing layer (mortar plate) and concrete slab. Because an impact force is not directly delivered to concrete slab, the impact force waveform or spectrum can be changed. In this paper, the changed impact spectrum was derived from several floating floor vibration tests. Based on the measured data, numerical modeling can describe the floating floor response, especially at low-frequency range. As a result, heavy-weight impact noise can be predicted using modified impact spectrum.Keywords: floating floor, heavy-weight impact, prediction, vibration
Procedia PDF Downloads 3737393 The Robot Physician's (Rp - 7) Management and Care in Unstable ICU Oncology Patients
Authors: Alisher Agzamov, Hanan Al Harbi
Abstract:
BACKGROUND: The timely assessment and treatment of ICU Surgical and Medical Oncology patients is important for Oncology surgeons and Medical Oncologists and Intensivists. We hypothesized that the use of Robot Physician’s (RP - 7) ICU management and care in ICU can improve ICU physician rapid response to unstable ICU Oncology patients. METHODS: This is a prospective study using a before-after, cohort-control design to test the effectiveness of RP. We have used RP to make multidisciplinary ICU rounds in the ICU and for Emergency cases. Data concerning several aspects of the RP interaction including the latency of the response, the problem being treated, the intervention that was ordered, and the type of information gathered using the RP were documented. The effect of RP on ICU length of stay and cost was assessed. RESULTS: The use of RP was associated with a reduction in latency of attending physician face-to-face response for routine and urgent pages compared to conventional care (RP: 10.2 +/- 3.3 minutes vs conventional: 220 +/- 80 minutes). The response latencies to Oncology Emergency (8.0 +/- 2.8 vs 150 +/- 55 minutes) and for Respiratory Failure (12 +/- 04 vs 110 +/- 45 minutes) were reduced (P < .001), as was the LOS for patients with AML (5 days) and ARDS (10 day). There was an increase in ICU occupancy by 20 % compared with the prerobot era, and there was an ICU cost savings of KD2.5 million attributable to the use of RP. CONCLUSION: The use of RP enabled rapid face-to-face ICU Intensivist - physician response to unstable ICU Oncology patients and resulted in decreased ICU cost and LOS.Keywords: robot physician, oncology patients, rp - 7 in icu management, cost and icu occupancy
Procedia PDF Downloads 857392 The Optimal Order Policy for the Newsvendor Model under Worker Learning
Authors: Sunantha Teyarachakul
Abstract:
We consider the worker-learning Newsvendor Model, under the case of lost-sales for unmet demand, with the research objective of proposing the cost-minimization order policy and lot size, scheduled to arrive at the beginning of the selling-period. In general, the New Vendor Model is used to find the optimal order quantity for the perishable items such as fashionable products or those with seasonal demand or short-life cycles. Technically, it is used when the product demand is stochastic and available for the single selling-season, and when there is only a one time opportunity for the vendor to purchase, with possibly of long ordering lead-times. Our work differs from the classical Newsvendor Model in that we incorporate the human factor (specifically worker learning) and its influence over the costs of processing units into the model. We describe this by using the well-known Wright’s Learning Curve. Most of the assumptions of the classical New Vendor Model are still maintained in our work, such as the constant per-unit cost of leftover and shortage, the zero initial inventory, as well as the continuous time. Our problem is challenging in the way that the best order quantity in the classical model, which is balancing the over-stocking and under-stocking costs, is no longer optimal. Specifically, when adding the cost-saving from worker learning to such expected total cost, the convexity of the cost function will likely not be maintained. This has called for a new way in determining the optimal order policy. In response to such challenges, we found a number of characteristics related to the expected cost function and its derivatives, which we then used in formulating the optimal ordering policy. Examples of such characteristics are; the optimal order quantity exists and is unique if the demand follows a Uniform Distribution; if the demand follows the Beta Distribution with some specific properties of its parameters, the second derivative of the expected cost function has at most two roots; and there exists the specific level of lot size that satisfies the first order condition. Our research results could be helpful for analysis of supply chain coordination and of the periodic review system for similar problems.Keywords: inventory management, Newsvendor model, order policy, worker learning
Procedia PDF Downloads 4197391 Prediction Fluid Properties of Iranian Oil Field with Using of Radial Based Neural Network
Authors: Abdolreza Memari
Abstract:
In this article in order to estimate the viscosity of crude oil,a numerical method has been used. We use this method to measure the crude oil's viscosity for 3 states: Saturated oil's viscosity, viscosity above the bubble point and viscosity under the saturation pressure. Then the crude oil's viscosity is estimated by using KHAN model and roller ball method. After that using these data that include efficient conditions in measuring viscosity, the estimated viscosity by the presented method, a radial based neural method, is taught. This network is a kind of two layered artificial neural network that its stimulation function of hidden layer is Gaussian function and teaching algorithms are used to teach them. After teaching radial based neural network, results of experimental method and artificial intelligence are compared all together. Teaching this network, we are able to estimate crude oil's viscosity without using KHAN model and experimental conditions and under any other condition with acceptable accuracy. Results show that radial neural network has high capability of estimating crude oil saving in time and cost is another advantage of this investigation.Keywords: viscosity, Iranian crude oil, radial based, neural network, roller ball method, KHAN model
Procedia PDF Downloads 5057390 Analysis of Efficiency Production of Grass Black Jelly (Mesona palustris) in Double Scale
Authors: Irvan Adhin Cholilie, Susinggih Wijana, Yusron Sugiarto
Abstract:
The aim of this research is to compare the results of black grass jelly produced using laboratory scale and double scale. In this research, the production from the laboratory scale is using ingredients of 1 kg black grass jelly added with 5 liters of water, while the double scale is using 5 kg black grass jelly and 75 liters of water. The results of organoleptic tests performed by 30 panelists (general) to the sample gels of grass black powder produced from both of laboratory and double scale are not different significantly in color, odor, flavor, and texture. Proximate test results conducted in both of grass black jelly powder produced in laboratory scale and double scale also have no significant differences in all parameters. Grass black jelly powder from double scale contains water, carbohydrate, crude fiber, and yield in the amount of 12,25 %; 43,7 %; 5,89 %; and 16,28 % respectively. The results of the energy efficiency analysis by boiling, draining, evaporation, drying, and milling processes are 85,11 %; 76,97 %; 99,64 %; 99,99% and 99,39% respectively. The utility needs including water needs for each batch amounted 0.1 m3 and cost Rp 220,5 per batch, the electricity needs for each batch is 20.01 kWh and cost Rp 18569.28 per batch, and LPG needs for each batch is 30 kg costed Rp 234,000.00 so that the total cost spent for the process is Rp 252,789.78 .Keywords: black grass jelly, powder, mass balance, energy balance, cost
Procedia PDF Downloads 3887389 The Low-Cost Design and 3D Printing of Structural Knee Orthotics for Athletic Knee Injury Patients
Authors: Alexander Hendricks, Sean Nevin, Clayton Wikoff, Melissa Dougherty, Jacob Orlita, Rafiqul Noorani
Abstract:
Knee orthotics play an important role in aiding in the recovery of those with knee injuries, especially athletes. However, structural knee orthotics is often very expensive, ranging between $300 and $800. The primary reason for this project was to answer the question: can 3D printed orthotics represent a viable and cost-effective alternative to present structural knee orthotics? The primary objective for this research project was to design a knee orthotic for athletes with knee injuries for a low-cost under $100 and evaluate its effectiveness. The initial design for the orthotic was done in SolidWorks, a computer-aided design (CAD) software available at Loyola Marymount University. After this design was completed, finite element analysis (FEA) was utilized to understand how normal stresses placed upon the knee affected the orthotic. The knee orthotic was then adjusted and redesigned to meet a specified factor-of-safety of 3.25 based on the data gathered during FEA and literature sources. Once the FEA was completed and the orthotic was redesigned based from the data gathered, the next step was to move on to 3D-printing the first design of the knee brace. Subsequently, physical therapy movement trials were used to evaluate physical performance. Using the data from these movement trials, the CAD design of the brace was refined to accommodate the design requirements. The final goal of this research means to explore the possibility of replacing high-cost, outsourced knee orthotics with a readily available low-cost alternative.Keywords: 3D printing, knee orthotics, finite element analysis, design for additive manufacturing
Procedia PDF Downloads 1847388 Predicting and Obtaining New Solvates of Curcumin, Demethoxycurcumin and Bisdemethoxycurcumin Based on the Ccdc Statistical Tools and Hansen Solubility Parameters
Authors: J. Ticona Chambi, E. A. De Almeida, C. A. Andrade Raymundo Gaiotto, A. M. Do Espírito Santo, L. Infantes, S. L. Cuffini
Abstract:
The solubility of active pharmaceutical ingredients (APIs) is challenging for the pharmaceutical industry. The new multicomponent crystalline forms as cocrystal and solvates present an opportunity to improve the solubility of APIs. Commonly, the procedure to obtain multicomponent crystalline forms of a drug starts by screening the drug molecule with the different coformers/solvents. However, it is necessary to develop methods to obtain multicomponent forms in an efficient way and with the least possible environmental impact. The Hansen Solubility Parameters (HSPs) is considered a tool to obtain theoretical knowledge of the solubility of the target compound in the chosen solvent. H-Bond Propensity (HBP), Molecular Complementarity (MC), Coordination Values (CV) are tools used for statistical prediction of cocrystals developed by the Cambridge Crystallographic Data Center (CCDC). The HSPs and the CCDC tools are based on inter- and intra-molecular interactions. The curcumin (Cur), target molecule, is commonly used as an anti‐inflammatory. The demethoxycurcumin (Demcur) and bisdemethoxycurcumin (Bisdcur) are natural analogues of Cur from turmeric. Those target molecules have differences in their solubilities. In this way, the work aimed to analyze and compare different tools for multicomponent forms prediction (solvates) of Cur, Demcur and Biscur. The HSP values were calculated for Cur, Demcur, and Biscur using the chemical group contribution methods and the statistical optimization from experimental data. The HSPmol software was used. From the HSPs of the target molecules and fifty solvents (listed in the HSP books), the relative energy difference (RED) was determined. The probability of the target molecules would be interacting with the solvent molecule was determined using the CCDC tools. A dataset of fifty molecules of different organic solvents was ranked for each prediction method and by a consensus ranking of different combinations: HSP, CV, HBP and MC values. Based on the prediction, 15 solvents were selected as Dimethyl Sulfoxide (DMSO), Tetrahydrofuran (THF), Acetonitrile (ACN), 1,4-Dioxane (DOX) and others. In a starting analysis, the slow evaporation technique from 50°C at room temperature and 4°C was used to obtain solvates. The single crystals were collected by using a Bruker D8 Venture diffractometer, detector Photon100. The data processing and crystal structure determination were performed using APEX3 and Olex2-1.5 software. According to the results, the HSPs (theoretical and optimized) and the Hansen solubility sphere for Cur, Demcur and Biscur were obtained. With respect to prediction analyses, a way to evaluate the predicting method was through the ranking and the consensus ranking position of solvates already reported in the literature. It was observed that the combination of HSP-CV obtained the best results when compared to the other methods. Furthermore, as a result of solvent selected, six new solvates, Cur-DOX, Cur-DMSO, Bicur-DOX, Bircur-THF, Demcur-DOX, Demcur-ACN and a new Biscur hydrate, were obtained. Crystal structures were determined for Cur-DOX, Biscur-DOX, Demcur-DOX and Bicur-Water. Moreover, the unit-cell parameter information for Cur-DMSO, Biscur-THF and Demcur-ACN were obtained. The preliminary results showed that the prediction method is showing a promising strategy to evaluate the possibility of forming multicomponent. It is currently working on obtaining multicomponent single crystals.Keywords: curcumin, HSPs, prediction, solvates, solubility
Procedia PDF Downloads 677387 Prediction of in situ Permeability for Limestone Rock Using Rock Quality Designation Index
Authors: Ahmed T. Farid, Muhammed Rizwan
Abstract:
Geotechnical study for evaluating soil or rock permeability is a highly important parameter. Permeability values for rock formations are more difficult for determination than soil formation as it is an effect of the rock quality and its fracture values. In this research, the prediction of in situ permeability of limestone rock formations was predicted. The limestone rock permeability was evaluated using Lugeon tests (in-situ packer permeability). Different sites which spread all over the Riyadh region of Saudi Arabia were chosen to conduct our study of predicting the in-situ permeability of limestone rock. Correlations were deducted between the values of in-situ permeability of the limestone rock with the value of the rock quality designation (RQD) calculated during the execution of the boreholes of the study areas. The study was performed for different ranges of RQD values measured during drilling of the sites boreholes. The developed correlations are recommended for the onsite determination of the in-situ permeability of limestone rock only. For the other sedimentary formations of rock, more studies are needed for predicting the actual correlations related to each type.Keywords: In situ, packer, permeability, rock, quality
Procedia PDF Downloads 3777386 A Comparison of Image Data Representations for Local Stereo Matching
Authors: André Smith, Amr Abdel-Dayem
Abstract:
The stereo matching problem, while having been present for several decades, continues to be an active area of research. The goal of this research is to find correspondences between elements found in a set of stereoscopic images. With these pairings, it is possible to infer the distance of objects within a scene, relative to the observer. Advancements in this field have led to experimentations with various techniques, from graph-cut energy minimization to artificial neural networks. At the basis of these techniques is a cost function, which is used to evaluate the likelihood of a particular match between points in each image. While at its core, the cost is based on comparing the image pixel data; there is a general lack of consistency as to what image data representation to use. This paper presents an experimental analysis to compare the effectiveness of more common image data representations. The goal is to determine the effectiveness of these data representations to reduce the cost for the correct correspondence relative to other possible matches.Keywords: colour data, local stereo matching, stereo correspondence, disparity map
Procedia PDF Downloads 3757385 Prime Mover Sizing for Base-Loaded Combined Heating and Power Systems
Authors: Djalal Boualili
Abstract:
This article considers the problem of sizing prime movers for combined heating and power (CHP) systems operating at full load to satisfy a fraction of a facility's electric load, i.e. a base load. Prime mover sizing is examined using three criteria: operational cost, carbon dioxide emissions (CDE), and primary energy consumption (PEC). The sizing process leads to consider ratios of conversion factors applied to imported electricity to conversion factors applied to fuel consumed. These ratios are labelled RCost, R CDE, R PEC depending on whether the conversion factors are associated with operational cost, CDE, or PEC, respectively. Analytical results show that in order to achieve savings in operational cost, CDE, or PEC, the ratios must be larger than a unique constant R Min that only depends on the CHP components efficiencies. Savings in operational cost, CDE, or PEC due to CHP operation are explicitly formulated using simple equations. This facilitates the process of comparing the tradeoffs of optimizing the savings of one criterion over the other two – a task that has traditionally been accomplished through computer simulations. A hospital building, located in Chlef, Algeria, was used as an example to apply the methodology presented in this article.Keywords: sizing, heating and power, ratios, energy consumption, carbon dioxide emissions
Procedia PDF Downloads 2347384 The Significance of Seasonality on the Airport Efficiency in Touristic Regions
Authors: Ioanna Pagoni, Annitsa Koumoutsidi
Abstract:
The aim of this paper is to estimate the efficiency of airports that are located in touristic regions. It focuses on the regional airports of Greece, which are located at the mainland and the islands that constitute touristic destinations. Most of these airports share the following characteristics. They operate at levels below capacity with a high level of seasonality to their traffic. In addition, in such airports, the operation of charter and low-cost airlines is significant. The efficiency of the study airports is calculated by using the non-parametric data envelopment analysis during the period of 2010-2016. The selected inputs include several airport infrastructure measures such as passenger terminal size, aircraft parking area, runway length, and the number of check-in counters, while the number of employees in each airport is also used. The number of passengers and aircraft movements are selected as outputs. The effect of seasonality, as well as the operation of charter airlines and low-cost carriers on airport efficiency, is estimated by running proper regression models. Preliminary findings indicate that low-cost and charter airlines contribute to increasing airport efficiency for most of the study airports. The results of this research could be useful for airlines, airport operators, hotel businesses, and other tourism-related operators.Keywords: airport efficiency, data envelopment analysis, low-cost carriers, charter airlines, seasonality
Procedia PDF Downloads 1307383 Replacement Time and Number of Preventive Maintenance Actions for Second-Hand Device
Authors: Wen Liang Chang
Abstract:
In this study, the optimal replacement time and number of preventive maintenance (PM) actions were investigated for a second-hand device. Suppose that a user intends to use a second-hand device for manufacturing products, and that the device is replaced with a new one. Any device failure is rectified through minimal repair, thereby incurring a fixed repair cost to the user. If the new device fails within the FRW period, minimal repair is performed at no cost to the user. After the FRW expires, a failed device is repaired and the cost of repair is incurred by the user. In this study, two profit models were developed, and the optimal replacement time and number of PM actions were determined to maximize profits. Finally, the influence of the optimal replacement time and number of PM actions were elaborated on, using numerical examples.Keywords: second-hand device, preventive maintenance, replacement time, device failure
Procedia PDF Downloads 4727382 Technology Blending as an Innovative Construction Mechanism in the Global South
Authors: Janet Kaningen, Richard N. Kaningen, Jonas Kaningen
Abstract:
This paper aims to discover the best ways to improve production efficiency, cost efficiency, community cohesion, and long-term sustainability in Ghana's housing delivery. Advanced Construction Technologies (ACTs) are set to become the sustainable mainstay of the construction industry due to the demand for innovative housing solutions. Advances in material science, building component production, and assembly technologies are leading to the development of next-generation materials such as polymeric-fiber-based products, light-metal alloys, and eco-materials. Modular housing construction has become more efficient and cost-effective than traditional building methods and is becoming increasingly popular for commercial, industrial, and residential projects. Effective project management and logistics will be imperative in the future speed and cost of modular construction housing.Keywords: technology blending, sustainability, housing, Ghana
Procedia PDF Downloads 957381 Development of Terrorist Threat Prediction Model in Indonesia by Using Bayesian Network
Authors: Hilya Mudrika Arini, Nur Aini Masruroh, Budi Hartono
Abstract:
There are more than 20 terrorist threats from 2002 to 2012 in Indonesia. Despite of this fact, preventive solution through studies in the field of national security in Indonesia has not been conducted comprehensively. This study aims to provide a preventive solution by developing prediction model of the terrorist threat in Indonesia by using Bayesian network. There are eight stages to build the model, started from literature review, build and verify Bayesian belief network to what-if scenario. In order to build the model, four experts from different perspectives are utilized. This study finds several significant findings. First, news and the readiness of terrorist group are the most influent factor. Second, according to several scenarios of the news portion, it can be concluded that the higher positive news proportion, the higher probability of terrorist threat will occur. Therefore, the preventive solution to reduce the terrorist threat in Indonesia based on the model is by keeping the positive news portion to a maximum of 38%.Keywords: Bayesian network, decision analysis, national security system, text mining
Procedia PDF Downloads 3957380 Analysis of Risk Factors Affecting the Motor Insurance Pricing with Generalized Linear Models
Authors: Puttharapong Sakulwaropas, Uraiwan Jaroengeratikun
Abstract:
Casualty insurance business, the optimal premium pricing and adequate cost for an insurance company are important in risk management. Normally, the insurance pure premium can be determined by multiplying the claim frequency with the claim cost. The aim of this research was to study in the application of generalized linear models to select the risk factor for model of claim frequency and claim cost for estimating a pure premium. In this study, the data set was the claim of comprehensive motor insurance, which was provided by one of the insurance company in Thailand. The results of this study found that the risk factors significantly related to pure premium at the 0.05 level consisted of no claim bonus (NCB) and used of the car (Car code).Keywords: generalized linear models, risk factor, pure premium, regression model
Procedia PDF Downloads 4687379 Total Life Cycle Cost and Life Cycle Assessment of Mass Timber Buildings in the US
Authors: Hongmei Gu, Shaobo Liang, Richard Bergman
Abstract:
With current worldwide trend in designs to have net-zero emission buildings to mitigate climate change, widespread use of mass timber products, such as Cross Laminated Timber (CLT), or Nail Laminated Timber (NLT) or Dowel Laminated Timber (DLT) in buildings have been proposed as one approach in reducing Greenhouse Gas (GHG) emissions. Consequentially, mass timber building designs are being adopted more and more by architectures in North America, especially for mid- to high-rise buildings where concrete and steel buildings are currently prevalent, but traditional light-frame wood buildings are not. Wood buildings and their associated wood products have tended to have lower environmental impacts than competing energy-intensive materials. It is common practice to conduct life cycle assessments (LCAs) and life cycle cost analyses on buildings with traditional structural materials like concrete and steel in the building design process. Mass timber buildings with lower environmental impacts, especially GHG emissions, can contribute to the Net Zero-emission goal for the world-building sector. However, the economic impacts from CLT mass timber buildings still vary from the life-cycle cost perspective and environmental trade-offs associated with GHG emissions. This paper quantified the Total Life Cycle Cost and cradle-to-grave GHG emissions of a pre-designed CLT mass timber building and compared it to a functionally-equivalent concrete building. The Total life cycle Eco-cost-efficiency is defined in this study and calculated to discuss the trade-offs for the net-zero emission buildings in a holistic view for both environmental and economic impacts. Mass timber used in buildings for the United States is targeted to the materials from the nation’s sustainable managed forest in order to benefit both national and global environments and economies.Keywords: GHG, economic impact, eco-cost-efficiency, total life-cycle costs
Procedia PDF Downloads 1457378 Development of a Fire Analysis Drone for Smoke Toxicity Measurement for Fire Prediction and Management
Authors: Gabrielle Peck, Ryan Hayes
Abstract:
This research presents the design and creation of a drone gas analyser, aimed at addressing the need for independent data collection and analysis of gas emissions during large-scale fires, particularly wasteland fires. The analyser drone, comprising a lightweight gas analysis system attached to a remote-controlled drone, enables the real-time assessment of smoke toxicity and the monitoring of gases released into the atmosphere during such incidents. The key components of the analyser unit included two gas line inlets connected to glass wool filters, a pump with regulated flow controlled by a mass flow controller, and electrochemical cells for detecting nitrogen oxides, hydrogen cyanide, and oxygen levels. Additionally, a non-dispersive infrared (NDIR) analyser is employed to monitor carbon monoxide (CO), carbon dioxide (CO₂), and hydrocarbon concentrations. Thermocouples can be attached to the analyser to monitor temperature, as well as McCaffrey probes combined with pressure transducers to monitor air velocity and wind direction. These additions allow for monitoring of the large fire and can be used for predictions of fire spread. The innovative system not only provides crucial data for assessing smoke toxicity but also contributes to fire prediction and management. The remote-controlled drone's mobility allows for safe and efficient data collection in proximity to the fire source, reducing the need for human exposure to hazardous conditions. The data obtained from the gas analyser unit facilitates informed decision-making by emergency responders, aiding in the protection of both human health and the environment. This abstract highlights the successful development of a drone gas analyser, illustrating its potential for enhancing smoke toxicity analysis and fire prediction capabilities. The integration of this technology into fire management strategies offers a promising solution for addressing the challenges associated with wildfires and other large-scale fire incidents. The project's methodology and results contribute to the growing body of knowledge in the field of environmental monitoring and safety, emphasizing the practical utility of drones for critical applications.Keywords: fire prediction, drone, smoke toxicity, analyser, fire management
Procedia PDF Downloads 937377 Artificial Neural Network-Based Prediction of Effluent Quality of Wastewater Treatment Plant Employing Data Preprocessing Approaches
Authors: Vahid Nourani, Atefeh Ashrafi
Abstract:
Prediction of treated wastewater quality is a matter of growing importance in water treatment procedure. In this way artificial neural network (ANN), as a robust data-driven approach, has been widely used for forecasting the effluent quality of wastewater treatment. However, developing ANN model based on appropriate input variables is a major concern due to the numerous parameters which are collected from treatment process and the number of them are increasing in the light of electronic sensors development. Various studies have been conducted, using different clustering methods, in order to classify most related and effective input variables. This issue has been overlooked in the selecting dominant input variables among wastewater treatment parameters which could effectively lead to more accurate prediction of water quality. In the presented study two ANN models were developed with the aim of forecasting effluent quality of Tabriz city’s wastewater treatment plant. Biochemical oxygen demand (BOD) was utilized to determine water quality as a target parameter. Model A used Principal Component Analysis (PCA) for input selection as a linear variance-based clustering method. Model B used those variables identified by the mutual information (MI) measure. Therefore, the optimal ANN structure when the result of model B compared with model A showed up to 15% percent increment in Determination Coefficient (DC). Thus, this study highlights the advantage of PCA method in selecting dominant input variables for ANN modeling of wastewater plant efficiency performance.Keywords: Artificial Neural Networks, biochemical oxygen demand, principal component analysis, mutual information, Tabriz wastewater treatment plant, wastewater treatment plant
Procedia PDF Downloads 1347376 Quick Covering Machine for Grain Drying Pavement
Authors: Fatima S. Rodriguez, Victorino T. Taylan, Manolito C. Bulaong, Helen F. Gavino, Vitaliana U. Malamug
Abstract:
In sundrying, the quality of the grains are greatly reduced when paddy grains were caught by the rain unsacked and unstored resulting to reduced profit. The objectives of this study were to design and fabricate a quick covering machine for grain drying pavement to test and evaluate the operating characteristics of the machine according to its deployment speed, recovery speed, deployment time, recovery time, power consumption, aesthetics of laminated sack, conducting partial budget, and cost curve analysis. The machine was able to cover the grains in a 12.8 m x 22.5 m grain drying pavement at an average time of 17.13 s. It consumed 0 .53 W-hr for the deployment and recovery of the cover. The machine entailed an investment cost of $1,344.40 and an annual cost charge of $647.32. Moreover, the savings per year using the quick covering machine was $101.83.Keywords: quick, covering machine, grain, drying pavement
Procedia PDF Downloads 3797375 Risk Reduction of Household Refuse, a Case Study of Shagari Low-Cost, Mubi North (LGA) Adamawa State, Nigeria
Authors: Maryam Tijjani Kolo
Abstract:
Lack of refuse dumping points has made the residents of Shagari low-cost well armed with some health and environmental related hazards. These studies investigate the effect of household refuse on the resident of Shagari low-cost. A well structured questionnaire was administered to elicit views of the respondent in the study area through adopting cluster sampling method. A total of 100 questionnaires were selected and divided into 50, each to both sections of the study area. Interview was conducted to each household head. Data obtained were analyzed using simple parentages to determine the major hazard in the area. Result showed that majority of the household are civil servant and traders, earning reasonable monthly income. 68% of the respondent has experienced the effect of living close to waste dumping areas, which include unpleasant smell and polluted smoke when refuse is burnt, which causes eye and respiratory induction, human injury from broken bottles or sharp objects as well as water, insect and air borne diseases. Hence, the need to urgently address these menace before it overwhelms the capacities of the community becomes paramount. Thus, the community should be given more enlightenment and refuse dumping sites should be created by the local government area.Keywords: household, refuse, refuse dumping points, Shagari low-cost
Procedia PDF Downloads 3237374 Benchmarking Machine Learning Approaches for Forecasting Hotel Revenue
Authors: Rachel Y. Zhang, Christopher K. Anderson
Abstract:
A critical aspect of revenue management is a firm’s ability to predict demand as a function of price. Historically hotels have used simple time series models (regression and/or pick-up based models) owing to the complexities of trying to build casual models of demands. Machine learning approaches are slowly attracting attention owing to their flexibility in modeling relationships. This study provides an overview of approaches to forecasting hospitality demand – focusing on the opportunities created by machine learning approaches, including K-Nearest-Neighbors, Support vector machine, Regression Tree, and Artificial Neural Network algorithms. The out-of-sample performances of above approaches to forecasting hotel demand are illustrated by using a proprietary sample of the market level (24 properties) transactional data for Las Vegas NV. Causal predictive models can be built and evaluated owing to the availability of market level (versus firm level) data. This research also compares and contrast model accuracy of firm-level models (i.e. predictive models for hotel A only using hotel A’s data) to models using market level data (prices, review scores, location, chain scale, etc… for all hotels within the market). The prospected models will be valuable for hotel revenue prediction given the basic characters of a hotel property or can be applied in performance evaluation for an existed hotel. The findings will unveil the features that play key roles in a hotel’s revenue performance, which would have considerable potential usefulness in both revenue prediction and evaluation.Keywords: hotel revenue, k-nearest-neighbors, machine learning, neural network, prediction model, regression tree, support vector machine
Procedia PDF Downloads 1387373 Prediction Modeling of Compression Properties of a Knitted Sportswear Fabric Using Response Surface Method
Authors: Jawairia Umar, Tanveer Hussain, Zulfiqar Ali, Muhammad Maqsood
Abstract:
Different knitted structures and knitted parameters play a vital role in the stretch and recovery management of compression sportswear in addition to the materials use to generate this stretch and recovery behavior of the fabric. The present work was planned to predict the different performance indicators of a compression sportswear fabric with some ground parameters i.e. base yarn stitch length (polyester as base yarn and spandex as plating yarn involve to make a compression fabric) and linear density of the spandex which is a key material of any sportswear fabric. The prediction models were generated by response surface method for performance indicators such as stretch & recovery percentage, compression generated by the garment on body, total elongation on application of high power force and load generated on certain percentage extension in fabric. Certain physical properties of the fabric were also modeled using these two parameters.Keywords: Compression, sportswear, stretch and recovery, statistical model, kikuhime
Procedia PDF Downloads 3827372 Simplified INS\GPS Integration Algorithm in Land Vehicle Navigation
Authors: Othman Maklouf, Abdunnaser Tresh
Abstract:
Land vehicle navigation is subject of great interest today. Global Positioning System (GPS) is the main navigation system for positioning in such systems. GPS alone is incapable of providing continuous and reliable positioning, because of its inherent dependency on external electromagnetic signals. Inertial Navigation (INS) is the implementation of inertial sensors to determine the position and orientation of a vehicle. The availability of low-cost Micro-Electro-Mechanical-System (MEMS) inertial sensors is now making it feasible to develop INS using an inertial measurement unit (IMU). INS has unbounded error growth since the error accumulates at each step. Usually, GPS and INS are integrated with a loosely coupled scheme. With the development of low-cost, MEMS inertial sensors and GPS technology, integrated INS/GPS systems are beginning to meet the growing demands of lower cost, smaller size, and seamless navigation solutions for land vehicles. Although MEMS inertial sensors are very inexpensive compared to conventional sensors, their cost (especially MEMS gyros) is still not acceptable for many low-end civilian applications (for example, commercial car navigation or personal location systems). An efficient way to reduce the expense of these systems is to reduce the number of gyros and accelerometers, therefore, to use a partial IMU (ParIMU) configuration. For land vehicular use, the most important gyroscope is the vertical gyro that senses the heading of the vehicle and two horizontal accelerometers for determining the velocity of the vehicle. This paper presents a field experiment for a low-cost strap down (ParIMU)\GPS combination, with data post processing for the determination of 2-D components of position (trajectory), velocity and heading. In the present approach, we have neglected earth rotation and gravity variations, because of the poor gyroscope sensitivities of our low-cost IMU (Inertial Measurement Unit) and because of the relatively small area of the trajectory.Keywords: GPS, IMU, Kalman filter, materials engineering
Procedia PDF Downloads 4257371 The Prognostic Prediction Value of Positive Lymph Nodes Numbers for the Hypopharyngeal Squamous Cell Carcinoma
Authors: Wendu Pang, Yaxin Luo, Junhong Li, Yu Zhao, Danni Cheng, Yufang Rao, Minzi Mao, Ke Qiu, Yijun Dong, Fei Chen, Jun Liu, Jian Zou, Haiyang Wang, Wei Xu, Jianjun Ren
Abstract:
We aimed to compare the prognostic prediction value of positive lymph node number (PLNN) to the American Joint Committee on Cancer (AJCC) tumor, lymph node, and metastasis (TNM) staging system for patients with hypopharyngeal squamous cell carcinoma (HPSCC). A total of 826 patients with HPSCC from the Surveillance, Epidemiology, and End Results database (2004–2015) were identified and split into two independent cohorts: training (n=461) and validation (n=365). Univariate and multivariate Cox regression analyses were used to evaluate the prognostic effects of PLNN in patients with HPSCC. We further applied six Cox regression models to compare the survival predictive values of the PLNN and AJCC TNM staging system. PLNN showed a significant association with overall survival (OS) and cancer-specific survival (CSS) (P < 0.001) in both univariate and multivariable analyses, and was divided into three groups (PLNN 0, PLNN 1-5, and PLNN>5). In the training cohort, multivariate analysis revealed that the increased PLNN of HPSCC gave rise to significantly poor OS and CSS after adjusting for age, sex, tumor size, and cancer stage; this trend was also verified by the validation cohort. Additionally, the survival model incorporating a composite of PLNN and TNM classification (C-index, 0.705, 0.734) performed better than the PLNN and AJCC TNM models. PLNN can serve as a powerful survival predictor for patients with HPSCC and is a surrogate supplement for cancer staging systems.Keywords: hypopharyngeal squamous cell carcinoma, positive lymph nodes number, prognosis, prediction models, survival predictive values
Procedia PDF Downloads 1587370 An Interpretable Data-Driven Approach for the Stratification of the Cardiorespiratory Fitness
Authors: D.Mendes, J. Henriques, P. Carvalho, T. Rocha, S. Paredes, R. Cabiddu, R. Trimer, R. Mendes, A. Borghi-Silva, L. Kaminsky, E. Ashley, R. Arena, J. Myers
Abstract:
The continued exploration of clinically relevant predictive models continues to be an important pursuit. Cardiorespiratory fitness (CRF) portends clinical vital information and as such its accurate prediction is of high importance. Therefore, the aim of the current study was to develop a data-driven model, based on computational intelligence techniques and, in particular, clustering approaches, to predict CRF. Two prediction models were implemented and compared: 1) the traditional Wasserman/Hansen Equations; and 2) an interpretable clustering approach. Data used for this analysis were from the 'FRIEND - Fitness Registry and the Importance of Exercise: The National Data Base'; in the present study a subset of 10690 apparently healthy individuals were utilized. The accuracy of the models was performed through the computation of sensitivity, specificity, and geometric mean values. The results show the superiority of the clustering approach in the accurate estimation of CRF (i.e., maximal oxygen consumption).Keywords: cardiorespiratory fitness, data-driven models, knowledge extraction, machine learning
Procedia PDF Downloads 2897369 Modelling Conceptual Quantities Using Support Vector Machines
Authors: Ka C. Lam, Oluwafunmibi S. Idowu
Abstract:
Uncertainty in cost is a major factor affecting performance of construction projects. To our knowledge, several conceptual cost models have been developed with varying degrees of accuracy. Incorporating conceptual quantities into conceptual cost models could improve the accuracy of early predesign cost estimates. Hence, the development of quantity models for estimating conceptual quantities of framed reinforced concrete structures using supervised machine learning is the aim of the current research. Using measured quantities of structural elements and design variables such as live loads and soil bearing pressures, response and predictor variables were defined and used for constructing conceptual quantities models. Twenty-four models were developed for comparison using a combination of non-parametric support vector regression, linear regression, and bootstrap resampling techniques. R programming language was used for data analysis and model implementation. Gross soil bearing pressure and gross floor loading were discovered to have a major influence on the quantities of concrete and reinforcement used for foundations. Building footprint and gross floor loading had a similar influence on beams and slabs. Future research could explore the modelling of other conceptual quantities for walls, finishes, and services using machine learning techniques. Estimation of conceptual quantities would assist construction planners in early resource planning and enable detailed performance evaluation of early cost predictions.Keywords: bootstrapping, conceptual quantities, modelling, reinforced concrete, support vector regression
Procedia PDF Downloads 2137368 Model Averaging in a Multiplicative Heteroscedastic Model
Authors: Alan Wan
Abstract:
In recent years, the body of literature on frequentist model averaging in statistics has grown significantly. Most of this work focuses on models with different mean structures but leaves out the variance consideration. In this paper, we consider a regression model with multiplicative heteroscedasticity and develop a model averaging method that combines maximum likelihood estimators of unknown parameters in both the mean and variance functions of the model. Our weight choice criterion is based on a minimisation of a plug-in estimator of the model average estimator's squared prediction risk. We prove that the new estimator possesses an asymptotic optimality property. Our investigation of finite-sample performance by simulations demonstrates that the new estimator frequently exhibits very favourable properties compared to some existing heteroscedasticity-robust model average estimators. The model averaging method hedges against the selection of very bad models and serves as a remedy to variance function misspecification, which often discourages practitioners from modeling heteroscedasticity altogether. The proposed model average estimator is applied to the analysis of two real data sets.Keywords: heteroscedasticity-robust, model averaging, multiplicative heteroscedasticity, plug-in, squared prediction risk
Procedia PDF Downloads 392