Search results for: Semi-Markov Decision Process
3564 Defining a Pathway to Zero Energy Building: A Case Study on Retrofitting an Old Office Building into a Net Zero Energy Building for Hot-Humid Climate
Authors: Kwame B. O. Amoah
Abstract:
This paper focuses on retrofitting an old existing office building to a net-zero energy building (NZEB). An existing small office building in Melbourne, Florida, was chosen as a case study to integrate state-of-the-art design strategies and energy-efficient building systems to improve building performance and reduce energy consumption. The study aimed to explore possible ways to maximize energy savings and renewable energy generation sources to cover the building's remaining energy needs necessary to achieve net-zero energy goals. A series of retrofit options were reviewed and adopted with some significant additional decision considerations. Detailed processes and considerations leading to zero energy are well documented in this study, with lessons learned adequately outlined. Based on building energy simulations, multiple design considerations were investigated, such as emerging state-of-the-art technologies, material selection, improvements to the building envelope, optimization of the HVAC, lighting systems, and occupancy loads analysis, as well as the application of renewable energy sources. The comparative analysis of simulation results was used to determine how specific techniques led to energy saving and cost reductions. The research results indicate that this small office building can meet net-zero energy use after appropriate design manipulations and renewable energy sources.
Keywords: Energy consumption, building energy analysis, energy retrofits, energy-efficiency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3403563 Application of Machine Learning Methods to Online Test Error Detection in Semiconductor Test
Authors: Matthias Kirmse, Uwe Petersohn, Elief Paffrath
Abstract:
As in today's semiconductor industries test costs can make up to 50 percent of the total production costs, an efficient test error detection becomes more and more important. In this paper, we present a new machine learning approach to test error detection that should provide a faster recognition of test system faults as well as an improved test error recall. The key idea is to learn a classifier ensemble, detecting typical test error patterns in wafer test results immediately after finishing these tests. Since test error detection has not yet been discussed in the machine learning community, we define central problem-relevant terms and provide an analysis of important domain properties. Finally, we present comparative studies reflecting the failure detection performance of three individual classifiers and three ensemble methods based upon them. As base classifiers we chose a decision tree learner, a support vector machine and a Bayesian network, while the compared ensemble methods were simple and weighted majority vote as well as stacking. For the evaluation, we used cross validation and a specially designed practical simulation. By implementing our approach in a semiconductor test department for the observation of two products, we proofed its practical applicability.
Keywords: Ensemble methods, fault detection, machine learning, semiconductor test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22743562 Aircraft Automatic Collision Avoidance Using Spiral Geometric Approach
Authors: M. Orefice, V. Di Vito
Abstract:
This paper provides a description of a Collision Avoidance algorithm that has been developed starting from the mathematical modeling of the flight of insects, in terms of spirals and conchospirals geometric paths. It is able to calculate a proper avoidance manoeuver aimed to prevent the infringement of a predefined distance threshold between ownship and the considered intruder, while minimizing the ownship trajectory deviation from the original path and in compliance with the aircraft performance limitations and dynamic constraints. The algorithm is designed in order to be suitable for real-time applications, so that it can be considered for the implementation in the most recent airborne automatic collision avoidance systems using the traffic data received through an ADS-B IN device. The presented approach is able to take into account the rules-of-the-air, due to the possibility to select, through specifically designed decision making logic based on the consideration of the encounter geometry, the direction of the calculated collision avoidance manoeuver that allows complying with the rules-of-the-air, as for instance the fundamental right of way rule. In the paper, the proposed collision avoidance algorithm is presented and its preliminary design and software implementation is described. The applicability of this method has been proved through preliminary simulation tests performed in a 2D environment considering single intruder encounter geometries, as reported and discussed in the paper.
Keywords: collision avoidance, RPAS, spiral geometry, ADS-B based application
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16663561 Removal of Ni(II), Zn(II) and Pb(II) ions from Single Metal Aqueous Solution using Activated Carbon Prepared from Rice Husk
Authors: Mohd F. Taha, Chong F. Kiat, Maizatul S. Shaharun, Anita Ramli
Abstract:
The abundance and availability of rice husk, an agricultural waste, make them as a good source for precursor of activated carbon. In this work, rice husk-based activated carbons were prepared via base treated chemical activation process prior the carbonization process. The effect of carbonization temperatures (400, 600 and 800oC) on their pore structure was evaluated through morphology analysis using scanning electron microscope (SEM). Sample carbonized at 800oC showed better evolution and development of pores as compared to those carbonized at 400 and 600oC. The potential of rice husk-based activated carbon as an alternative adsorbent was investigated for the removal of Ni(II), Zn(II) and Pb(II) from single metal aqueous solution. The adsorption studies using rice husk-based activated carbon as an adsorbent were carried out as a function of contact time at room temperature and the metal ions were analyzed using atomic absorption spectrophotometer (AAS). The ability to remove metal ion from single metal aqueous solution was found to be improved with the increasing of carbonization temperature. Among the three metal ions tested, Pb(II) ion gave the highest adsorption on rice husk-based activated carbon. The results obtained indicate the potential to utilize rice husk as a promising precursor for the preparation of activated carbon for removal of heavy metals.
Keywords: Activated carbon, metal ion adsorption, rice husk, wastewater treatment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27173560 Stackelberg Security Game for Optimizing Security of Federated Internet of Things Platform Instances
Authors: Violeta Damjanovic-Behrendt
Abstract:
This paper presents an approach for optimal cyber security decisions to protect instances of a federated Internet of Things (IoT) platform in the cloud. The presented solution implements the repeated Stackelberg Security Game (SSG) and a model called Stochastic Human behaviour model with AttRactiveness and Probability weighting (SHARP). SHARP employs the Subjective Utility Quantal Response (SUQR) for formulating a subjective utility function, which is based on the evaluations of alternative solutions during decision-making. We augment the repeated SSG (including SHARP and SUQR) with a reinforced learning algorithm called Naïve Q-Learning. Naïve Q-Learning belongs to the category of active and model-free Machine Learning (ML) techniques in which the agent (either the defender or the attacker) attempts to find an optimal security solution. In this way, we combine GT and ML algorithms for discovering optimal cyber security policies. The proposed security optimization components will be validated in a collaborative cloud platform that is based on the Industrial Internet Reference Architecture (IIRA) and its recently published security model.
Keywords: Security, internet of things, cloud computing, Stackelberg security game, machine learning, Naïve Q-learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16453559 Internet Governance based on Multiple-Stakeholders: Opportunities, Issues and Developments
Authors: Martin Hans Knahl
Abstract:
The Internet is the global data communications infrastructure based on the interconnection of both public and private networks using protocols that implement Internetworking on a global scale. Hence the control of protocol and infrastructure development, resource allocation and network operation are crucial and interlinked aspects. Internet Governance is the hotly debated and contentious subject that refers to the global control and operation of key Internet infrastructure such as domain name servers and resources such as domain names. It is impossible to separate technical and political positions as they are interlinked. Furthermore the existence of a global market, transparency and competition impact upon Internet Governance and related topics such as network neutrality and security. Current trends and developments regarding Internet governance with a focus on the policy-making process, security and control have been observed to evaluate current and future implications on the Internet. The multi stakeholder approach to Internet Governance discussed in this paper presents a number of opportunities, issues and developments that will affect the future direction of the Internet. Internet operation, maintenance and advisory organisations such as the Internet Corporation for Assigned Names and Numbers (ICANN) or the Internet Governance Forum (IGF) are currently in the process of formulating policies for future Internet Governance. Given the controversial nature of the issues at stake and the current lack of agreement it is predicted that institutional as well as market governance will remain present for the network access and content.Keywords: Internet Governance, ICANN, Democracy, Security
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18693558 Robust Batch Process Scheduling in Pharmaceutical Industries: A Case Study
Authors: Tommaso Adamo, Gianpaolo Ghiani, Antonio D. Grieco, Emanuela Guerriero
Abstract:
Batch production plants provide a wide range of scheduling problems. In pharmaceutical industries a batch process is usually described by a recipe, consisting of an ordering of tasks to produce the desired product. In this research work we focused on pharmaceutical production processes requiring the culture of a microorganism population (i.e. bacteria, yeasts or antibiotics). Several sources of uncertainty may influence the yield of the culture processes, including (i) low performance and quality of the cultured microorganism population or (ii) microbial contamination. For these reasons, robustness is a valuable property for the considered application context. In particular, a robust schedule will not collapse immediately when a cell of microorganisms has to be thrown away due to a microbial contamination. Indeed, a robust schedule should change locally in small proportions and the overall performance measure (i.e. makespan, lateness) should change a little if at all. In this research work we formulated a constraint programming optimization (COP) model for the robust planning of antibiotics production. We developed a discrete-time model with a multi-criteria objective, ordering the different criteria and performing a lexicographic optimization. A feasible solution of the proposed COP model is a schedule of a given set of tasks onto available resources. The schedule has to satisfy tasks precedence constraints, resource capacity constraints and time constraints. In particular time constraints model tasks duedates and resource availability time windows constraints. To improve the schedule robustness, we modeled the concept of (a, b) super-solutions, where (a, b) are input parameters of the COP model. An (a, b) super-solution is one in which if a variables (i.e. the completion times of a culture tasks) lose their values (i.e. cultures are contaminated), the solution can be repaired by assigning these variables values with a new values (i.e. the completion times of a backup culture tasks) and at most b other variables (i.e. delaying the completion of at most b other tasks). The efficiency and applicability of the proposed model is demonstrated by solving instances taken from a real-life pharmaceutical company. Computational results showed that the determined super-solutions are near-optimal.Keywords: Constraint programming, super-solutions, robust scheduling, batch process, pharmaceutical industries.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19753557 Variability in Near-Surface Ultraviolet Radiation and Its Dependence on Atmospheric Parameters
Authors: Yusuff Idowu Moshood, Sanni Mohammed
Abstract:
Natural radiations such as ultraviolet (UV) radiation sourced from sun are known to be the main causes of skin cancer, sunburn, eye damage, premature aging of skin and other skin related diseases. Its percentage of radiation reaching the earth populace and its impacts are not well known. Its variability in near-surface relating to its impacts on populace depends on some atmospheric parameters. Hence, this work was embarked on to determine the variability in near-surface UV radiation and its dependency on some atmospheric parameters at different time of the day in Offa, Nigeria. The variability was determined using the data obtained from meteorological garden, Science Laboratory Technology Department, Federal Polytechnic Offa, Nigeria. The data obtained were solar UV radiation, solar radiation, temperature, humidity and pressure at 30 minutes interval. Relationships were determined and correlations were derived using SPSS Pearson Correlation tool. The results showed a significant level of correlation with p-value of 0.01 and 0.05 levels. Thus, the results revealed some good relationships between the solar UV radiation and other atmospheric parameters with significance level less than p-value obtained. Inferentially, interdependent relationships were found to exist. Therefore, the nature of relationship obtained could be a yardstick for decision making in short term environmental planning on solar UV radiation depending of some atmospheric parameters within Offa locality.
Keywords: Correlation, inferential, radiation, yardstick.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7823556 Load Balancing in Heterogeneous P2P Systems using Mobile Agents
Authors: Neeraj Nehra, R. B. Patel, V. K. Bhat
Abstract:
Use of the Internet and the World-Wide-Web (WWW) has become widespread in recent years and mobile agent technology has proliferated at an equally rapid rate. In this scenario load balancing becomes important for P2P systems. Beside P2P systems can be highly heterogeneous, i.e., they may consists of peers that range from old desktops to powerful servers connected to internet through high-bandwidth lines. There are various loads balancing policies came into picture. Primitive one is Message Passing Interface (MPI). Its wide availability and portability make it an attractive choice; however the communication requirements are sometimes inefficient when implementing the primitives provided by MPI. In this scenario we use the concept of mobile agent because Mobile agent (MA) based approach have the merits of high flexibility, efficiency, low network traffic, less communication latency as well as highly asynchronous. In this study we present decentralized load balancing scheme using mobile agent technology in which when a node is overloaded, task migrates to less utilized nodes so as to share the workload. However, the decision of which nodes receive migrating task is made in real-time by defining certain load balancing policies. These policies are executed on PMADE (A Platform for Mobile Agent Distribution and Execution) in decentralized manner using JuxtaNet and various load balancing metrics are discussed.Keywords: Mobile Agents, Agent host, Agent Submitter, PMADE.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17433555 Deposition Rate and Energy Enhancements of TiN Thin-Film in a Magnetized Sheet Plasma Source
Authors: Hamdi Muhyuddin D. Barra, Henry J. Ramos
Abstract:
Titanium nitride (TiN) has been synthesized using the sheet plasma negative ion source (SPNIS). The parameters used for its effective synthesis has been determined from previous experiments and studies. In this study, further enhancement of the deposition rate of TiN synthesis and advancement of the SPNIS operation is presented. This is primarily achieved by the addition of Sm-Co permanent magnets and a modification of the configuration in the TiN deposition process. The magnetic enhancement is aimed at optimizing the sputtering rate and the sputtering yield of the process. The Sm-Co permanent magnets are placed below the Ti target for better sputtering by argon. The Ti target is biased from –250V to – 350V and is sputtered by Ar plasma produced at discharge current of 2.5–4A and discharge potential of 60–90V. Steel substrates of dimensions 20x20x0.5mm3 were prepared with N2:Ar volumetric ratios of 1:3, 1:5 and 1:10. Ocular inspection of samples exhibit bright gold color associated with TiN. XRD characterization confirmed the effective TiN synthesis as all samples exhibit the (200) and (311) peaks of TiN and the non-stoichiometric Ti2N (220) facet. Cross-sectional SEM results showed increase in the TiN deposition rate of up to 0.35μm/min. This doubles what was previously obtained [1]. Scanning electron micrograph results give a comparative morphological picture of the samples. Vickers hardness results gave the largest hardness value of 21.094GPa.Keywords: Chemical vapor deposition, Magnetized sheetplasma, Thin-film synthesis, Titanium nitride.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16653554 An Experimental Study on the Effect of Operating Parameters during the Micro-Electro-Discharge Machining of Ni Based Alloy
Authors: Asma Perveen, M. P. Jahan
Abstract:
Ni alloys have managed to cover wide range of applications such as automotive industries, oil gas industries, and aerospace industries. However, these alloys impose challenges while using conventional machining technologies. On the other hand, Micro-Electro-Discharge machining (micro-EDM) is a non-conventional machining method that uses controlled sparks energy to remove material irrespective of the materials hardness. There has been always a huge interest from the industries for developing optimum methodology and parameters in order to enhance the productivity of micro-EDM in terms of reducing machining time and tool wear for different alloys. Therefore, the aims of this study are to investigate the effects of the micro-EDM process parameters, in order to find their optimal values. The input process parameters include voltage, capacitance, and electrode rotational speed, whereas the output parameters considered are machining time, entrance diameter of hole, overcut, tool wear, and crater size. The surface morphology and element characterization are also investigated with the use of SEM and EDX analysis. The experimental result indicates the reduction of machining time with the increment of discharge energy. Discharge energy also contributes to the enlargement of entrance diameter as well as overcut. In addition, tool wears show reduction with the increase of discharge energy. Moreover, crater size is found to be increased in size along with the increment of discharge energy.Keywords: Micro EDM, Ni alloy, discharge energy, micro-holes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13353553 The Analysis of Defects Prediction in Injection Molding
Authors: Mehdi Moayyedian, Kazem Abhary, Romeo Marian
Abstract:
This paper presents an evaluation of a plastic defect in injection molding before it occurs in the process; it is known as the short shot defect. The evaluation of different parameters which affect the possibility of short shot defect is the aim of this paper. The analysis of short shot possibility is conducted via SolidWorks Plastics and Taguchi method to determine the most significant parameters. Finite Element Method (FEM) is employed to analyze two circular flat polypropylene plates of 1 mm thickness. Filling time, part cooling time, pressure holding time, melt temperature and gate type are chosen as process and geometric parameters, respectively. A methodology is presented herein to predict the possibility of the short-shot occurrence. The analysis determined melt temperature is the most influential parameter affecting the possibility of short shot defect with a contribution of 74.25%, and filling time with a contribution of 22%, followed by gate type with a contribution of 3.69%. It was also determined the optimum level of each parameter leading to a reduction in the possibility of short shot are gate type at level 1, filling time at level 3 and melt temperature at level 3. Finally, the most significant parameters affecting the possibility of short shot were determined to be melt temperature, filling time, and gate type.Keywords: Injection molding, plastic defects, short shot, Taguchi method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15323552 Effect of Crude Oil Particle Elasticity on the Separation Efficiency of a Hydrocyclone
Authors: M. H. Narasingha, K. Pana-Suppamassadu, P. Narataruksa
Abstract:
The separation efficiency of a hydrocyclone has extensively been considered on the rigid particle assumption. A collection of experimental studies have demonstrated their discrepancies from the modeling and simulation results. These discrepancies caused by the actual particle elasticity have generally led to a larger amount of energy consumption in the separation process. In this paper, the influence of particle elasticity on the separation efficiency of a hydrocyclone system was investigated through the Finite Element (FE) simulations using crude oil droplets as the elastic particles. A Reitema-s design hydrocyclone with a diameter of 8 mm was employed to investigate the separation mechanism of the crude oil droplets from water. The cut-size diameter eter of the crude oil was 10 - Ðçm in order to fit with the operating range of the adopted hydrocylone model. Typical parameters influencing the performance of hydrocyclone were varied with the feed pressure in the range of 0.3 - 0.6 MPa and feed concentration between 0.05 – 0.1 w%. In the simulation, the Finite Element scheme was applied to investigate the particle-flow interaction occurred in the crude oil system during the process. The interaction of a single oil droplet at the size of 10 - Ðçm to the flow field was observed. The feed concentration fell in the dilute flow regime so the particle-particle interaction was ignored in the study. The results exhibited the higher power requirement for the separation of the elastic particulate system when compared with the rigid particulate system.Keywords: Hydrocyclone, separation efficiency, strain energy density, strain rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18043551 Optimization of the Dental Direct Digital Imaging by Applying the Self-Recognition Technology
Authors: Mina Dabirinezhad, Mohsen Bayat Pour, Amin Dabirinejad
Abstract:
This paper is intended to introduce the technology to solve some of the deficiencies of the direct digital radiology. Nowadays, digital radiology is the latest progression in dental imaging, which has become an essential part of dentistry. There are two main parts of the direct digital radiology comprised of an intraoral X-ray machine and a sensor (digital image receptor). The dentists and the dental nurses experience afflictions during the taking image process by the direct digital X-ray machine. For instance, sometimes they need to readjust the sensor in the mouth of the patient to take the X-ray image again due to the low quality of that. Another problem is, the position of the sensor may move in the mouth of the patient and it triggers off an inappropriate image for the dentists. It means that it is a time-consuming process for dentists or dental nurses. On the other hand, taking several the X-ray images brings some problems for the patient such as being harmful to their health and feeling pain in their mouth due to the pressure of the sensor to the jaw. The author provides a technology to solve the above-mentioned issues that is called “Self-Recognition Direct Digital Radiology” (SDDR). This technology is based on the principle that the intraoral X-ray machine is capable to diagnose the location of the sensor in the mouth of the patient automatically. In addition, to solve the aforementioned problems, SDDR technology brings out fewer environmental impacts in comparison to the previous version.
Keywords: Dental direct digital imaging, digital image receptor, digital x-ray machine, and environmental impacts.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5973550 JEWEL: A Cosmological Model Due to the Geometrical Displacement of Galactic Object Like Black, White and Worm Holes
Authors: Francesco Pia
Abstract:
Stellar objects such as black, white and worm holes can be the subject of speculative reasoning if represented in a simplified and geometric form in order to be able to move them; and the cosmological model is one of the most important contents in relation to speculations that can then open the way to other aspects that are not strictly speculative but practical, precisely in the Universe represented by us. In this work, thanks to the hypothesis of a very large number of black, white and worm holes present in our Universe, we imagine that they can be moved; it was therefore thought to align them on a plane and following a redistribution, and the boundaries of this plane were ideally joined, giving rise to a sphere that has the stellar objects examined radially distributed. Thanks to geometrical displacements of these stellar objects that do not make each one of them lose their functionality in the region in which they are located, at the end of the speculative process it is possible to highlight a spherical layer that allows a flow from the outside and inside this spherical shell allowing to relate to other external and internal spherical layers; this aspect that seems useful to describe the universe we live in, for example inside one of the spherical shells just described. The name "Jewel" was chosen because, imagining the speculative process present in this work at the end of steps, the cosmological model tends to be "luminous". This cosmological model includes, for each internal part of a generic layer, different and numerous moments of our universe thanks to an eternal flow inward. There are many aspects to explore, one of these is the connection between the outermost and the inside of the spherical layers.
Keywords: Black hole, cosmological model, cosmology, white hole.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5713549 A Software Tool Design for Cerebral Infarction of MR Images
Authors: Kyoung-Jong Park, Woong-Gi Jeon, Hee-Cheol Kim, Dong-Eog Kim, Heung-Kook Choi
Abstract:
The brain MR imaging-based clinical research and analysis system were specifically built and the development for a large-scale data was targeted. We used the general clinical data available for building large-scale data. Registration period for the selection of the lesion ROI and the region growing algorithm was used and the Mesh-warp algorithm for matching was implemented. The accuracy of the matching errors was modified individually. Also, the large ROI research data can accumulate by our developed compression method. In this way, the correctly decision criteria to the research result was suggested. The experimental groups were age, sex, MR type, patient ID and smoking which can easily be queries. The result data was visualized of the overlapped images by a color table. Its data was calculated by the statistical package. The evaluation for the utilization of this system in the chronic ischemic damage in the area has done from patients with the acute cerebral infarction. This is the cause of neurologic disability index location in the center portion of the lateral ventricle facing. The corona radiate was found in the position. Finally, the system reliability was measured both inter-user and intra-user registering correlation.
Keywords: Software tool design, Cerebral infarction, Brain MR image, Registration
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16633548 Removal of Volatile Organic Compounds from Contaminated Surfactant Solution using Co-Curren Vacuum Stripping
Authors: Pornchai Suriya-Amrit, Suratsawadee Kungsanant, Boonyarach Kitiyanan
Abstract:
There has been a growing interest in utilizing surfactants in remediation processes to separate the hydrophobic volatile organic compounds (HVOCs) from aqueous solution. One attractive process is cloud point extraction (CPE), which utilizes nonionic surfactants as a separating agent. Since the surfactant cost is a key determination of the economic viability of the process, it is important that the surfactants are recycled and reused. This work aims to study the performance of the co-current vacuum stripping using a packed column for HVOCs removal from contaminated surfactant solution. Six types HVOCs are selected as contaminants. The studied surfactant is the branched secondary alcohol ethoxylates (AEs), Tergitol TMN-6 (C14H30O2). The volatility and the solubility of HVOCs in surfactant system are determined in terms of an apparent Henry’s law constant and a solubilization constant, respectively. Moreover, the HVOCs removal efficiency of vacuum stripping column is assessed in terms of percentage of HVOCs removal and the overall liquid phase volumetric mass transfer coefficient. The apparent Henry’s law constant of benzenz , toluene, and ethyl benzene were 7.00×10-5, 5.38×10-5, 3.35× 10-5 respectively. The solubilization constant of benzene, toluene, and ethyl benzene were 1.71, 2.68, 7.54 respectively. The HVOCs removal for all solute were around 90 percent.
Keywords: Apparent Henry’s law constant, Branched secondary alcohol ethoxylates, Vacuum Stripping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16893547 Vulnerability of Indian Agriculture to Climate Change: A Study of the Himalayan Region State
Authors: Rajendra Kumar Isaac, Monisha Isaac
Abstract:
Climate variability and changes are the emerging challenges for Indian agriculture with the growing population to ensure national food security. A study was conducted to assess the Climatic Change effects in medium to low altitude areas of the Himalayan region causing changes in land use and cereal crop productivity with the various climatic parameters. The rainfall and temperature changes from 1951 to 2013 were studied at four locations of varying altitudes, namely Hardwar, Rudra Prayag, Uttar Kashi and Tehri Garwal. It was observed that there is noticeable increment in temperature on all the four locations. It was surprisingly observed that the mean rainfall intensity of 30 minutes duration has increased at the rate of 0.1 mm/hours since 2000. The study shows that the combined effect of increasing temperature, rainfall, runoff and urbanization at the mid-Himalayan region is causing an increase in various climatic disasters and changes in agriculture patterns. A noticeable change in cropping patterns, crop productivity and land use change was observed. Appropriate adaptation and mitigation strategies are necessary to ensure that sustainable and climate-resilient agriculture. Appropriate information is necessary for farmers, as well as planners and decision makers for developing, disseminating and adopting climate-smart technologies.Keywords: Climate variability, agriculture, land use, mitigation strategies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21433546 Adjustment of a PET Scanner for PEPT
Authors: Alireza Sadrmomtaz
Abstract:
Positron emission particle tracking (PEPT) is a technique in which a single radioactive tracer particle can be accurately tracked as it moves. A limitation of PET is that in order to reconstruct a tomographic image it is necessary to acquire a large volume of data (millions of events), so it is difficult to study rapidly changing systems. By considering this fact, PEPT is a very fast process compared with PET. In PEPT detecting both photons defines a line and the annihilation is assumed to have occurred somewhere along this line. The location of the tracer can be determined to within a few mm from coincident detection of a small number of pairs of back-to-back gamma rays and using triangulation. This can be achieved many times per second and the track of a moving particle can be reliably followed. This technique was invented at the University of Birmingham [1]. The attempt in PEPT is not to form an image of the tracer particle but simply to determine its location with time. If this tracer is followed for a long enough period within a closed, circulating system it explores all possible types of motion. The application of PEPT to industrial process systems carried out at the University of Birmingham is categorized in two subjects: the behaviour of granular materials and viscous fluids. Granular materials are processed in industry for example in the manufacture of pharmaceuticals, ceramics, food, polymers and PEPT has been used in a number of ways to study the behaviour of these systems [2]. PEPT allows the possibility of tracking a single particle within the bed [3]. Also PEPT has been used for studying systems such as: fluid flow, viscous fluids in mixers [4], using a neutrally-buoyant tracer particle [5].Keywords: PET, BGO, Particle Tracking, ECAT 931, List mode, PEPT.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14033545 Artificial Neural Network Modeling and Genetic Algorithm Based Optimization of Hydraulic Design Related to Seepage under Concrete Gravity Dams on Permeable Soils
Authors: Muqdad Al-Juboori, Bithin Datta
Abstract:
Hydraulic structures such as gravity dams are classified as essential structures, and have the vital role in providing strong and safe water resource management. Three major aspects must be considered to achieve an effective design of such a structure: 1) The building cost, 2) safety, and 3) accurate analysis of seepage characteristics. Due to the complexity and non-linearity relationships of the seepage process, many approximation theories have been developed; however, the application of these theories results in noticeable errors. The analytical solution, which includes the difficult conformal mapping procedure, could be applied for a simple and symmetrical problem only. Therefore, the objectives of this paper are to: 1) develop a surrogate model based on numerical simulated data using SEEPW software to approximately simulate seepage process related to a hydraulic structure, 2) develop and solve a linked simulation-optimization model based on the developed surrogate model to describe the seepage occurring under a concrete gravity dam, in order to obtain optimum and safe design at minimum cost. The result shows that the linked simulation-optimization model provides an efficient and optimum design of concrete gravity dams.Keywords: Artificial neural network, concrete gravity dam, genetic algorithm, seepage analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13773544 Visualization and Indexing of Spectral Databases
Authors: Tibor Kulcsar, Gabor Sarossy, Gabor Bereznai, Robert Auer, Janos Abonyi
Abstract:
On-line (near infrared) spectroscopy is widely used to support the operation of complex process systems. Information extracted from spectral database can be used to estimate unmeasured product properties and monitor the operation of the process. These techniques are based on looking for similar spectra by nearest neighborhood algorithms and distance based searching methods. Search for nearest neighbors in the spectral space is an NP-hard problem, the computational complexity increases by the number of points in the discrete spectrum and the number of samples in the database. To reduce the calculation time some kind of indexing could be used. The main idea presented in this paper is to combine indexing and visualization techniques to reduce the computational requirement of estimation algorithms by providing a two dimensional indexing that can also be used to visualize the structure of the spectral database. This 2D visualization of spectral database does not only support application of distance and similarity based techniques but enables the utilization of advanced clustering and prediction algorithms based on the Delaunay tessellation of the mapped spectral space. This means the prediction has not to use the high dimension space but can be based on the mapped space too. The results illustrate that the proposed method is able to segment (cluster) spectral databases and detect outliers that are not suitable for instance based learning algorithms.
Keywords: indexing high dimensional databases, dimensional reduction, clustering, similarity, k-nn algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17693543 Finding Pareto Optimal Front for the Multi-Mode Time, Cost Quality Trade-off in Project Scheduling
Authors: H. Iranmanesh, M. R. Skandari, M. Allahverdiloo
Abstract:
Project managers are the ultimate responsible for the overall characteristics of a project, i.e. they should deliver the project on time with minimum cost and with maximum quality. It is vital for any manager to decide a trade-off between these conflicting objectives and they will be benefited of any scientific decision support tool. Our work will try to determine optimal solutions (rather than a single optimal solution) from which the project manager will select his desirable choice to run the project. In this paper, the problem in project scheduling notated as (1,T|cpm,disc,mu|curve:quality,time,cost) will be studied. The problem is multi-objective and the purpose is finding the Pareto optimal front of time, cost and quality of a project (curve:quality,time,cost), whose activities belong to a start to finish activity relationship network (cpm) and they can be done in different possible modes (mu) which are non-continuous or discrete (disc), and each mode has a different cost, time and quality . The project is constrained to a non-renewable resource i.e. money (1,T). Because the problem is NP-Hard, to solve the problem, a meta-heuristic is developed based on a version of genetic algorithm specially adapted to solve multi-objective problems namely FastPGA. A sample project with 30 activities is generated and then solved by the proposed method.Keywords: FastPGA, Multi-Execution Activity Mode, ParetoOptimality, Project Scheduling, Time-Cost-Quality Trade-Off.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16843542 The Modulation of Self-interest Instruction on the Fair-Proposing Behavior in Ultimatum Game
Authors: N. S. Yen, T. H. Yang, W. H. Huang, Y. F. Fang, H. W. Cho
Abstract:
Ultimatum game is an experimental paradigm to study human decision making. There are two players, a proposer and a responder, to split a fixed amount of money. According to the traditional economic theory on ultimatum game, proposer should propose the selfish offers to responder as much as possible to maximize proposer’s own outcomes. However, most evidences had showed that people chose more fair offers, hence two hypotheses – fairness favoring and strategic concern were proposed. In current study, we induced the motivation in participants to be either selfish or altruistic, and manipulated the task variables, the stake sizes (NT$100, 1000, 10000) and the share sizes (the 40%, 30%, 20%, 10% of the sum as selfish offers, and the 60%, 70%, 80%, 90% of the sum as altruistic offers), to examine the two hypotheses. The results showed that most proposers chose more fair offers with longer reaction times (RTs) no matter in choosing between the fair and selfish offers, or between the fair and altruistic offers. However, the proposers received explicit self-interest instruction chose more selfish offers accompanied with longer RTs in choosing between the fair and selfish offers. Therefore, the results supported the strategic concern hypothesis that previous proposers choosing the fair offers might be resulted from the fear of rejection by responders. Proposers would become more self-interest if the fear of being rejected is eliminated.Keywords: Ultimatum game, self-interest, altruistic, fear of rejection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8653541 Interpreting Chopin’s Music Today: Mythologization of Art: Kitsch
Authors: Ilona Bala
Abstract:
The subject of this abstract is related to the notion of 'popular music', a notion that should be treated with extreme care, particularly when applied to Frederic Chopin, one of the greatest composers of Romanticism. By ‘popular music’, we mean a category of everyday music, set against the more intellectual kind, referred to as ‘classical’. We only need to look back to the culture of the nineteenth century to realize that this ‘popular music’ refers to the ‘music of the low’. It can be studied from a sociological viewpoint, or as sociological aesthetics. However, we cannot ignore the fact that, very quickly, this music spread to the wealthiest strata of the European society of the nineteenth century, while likewise the lowest classes often listen to the intellectual classical music, so pleasant to listen to. Further, we can observe that a sort of ‘sacralisation of kitsch’ occurs at the intersection between the classical and popular music. This process is the topic of this contribution. We will start by investigating the notion of kitsch through the study of Chopin’s popular compositions. However, before considering the popularisation of this music in today’s culture, we will have to focus on the use of the word kitsch in Chopin’s times, through his own musical aesthetics. Finally, the objective here will be to negate the theory that art is simply the intellectual definition of aesthetics. A kitsch can, obviously, only work on the emotivity of the masses, as it represents one of the features of culture-language (the words which the masses identify with). All art is transformed, becoming something outdated or even outmoded. Here, we are truly within a process of mythologization of art, through the study of the aesthetic reception of the musical work.Keywords: F. Chopin, musical work, popular music, romantic music, mythologization of art, kitsch.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12723540 Design and Implementation of a Software Platform Based on Artificial Intelligence for Product Recommendation
Authors: G. Settanni, A. Panarese, R. Vaira, A. Galiano
Abstract:
Nowadays, artificial intelligence is used successfully in the field of e-commerce for its ability to learn from a large amount of data. In this research study, a prototype software platform was designed and implemented in order to suggest to users the most suitable products for their needs. The platform includes a recommender system based on artificial intelligence algorithms that provide suggestions and decision support to the customer. Specifically, support vector machine algorithms have been implemented combined with natural language processing techniques that allow the user to interact with the system, express their requests and receive suggestions. The interested user can access the web platform on the internet using a computer, tablet or mobile phone, register, provide the necessary information and view the products that the system deems them the most appropriate. The platform also integrates a dashboard that allows the use of the various functions, which the platform is equipped with, in an intuitive and simple way. Also, Long Short-Term Memory algorithms have been implemented and trained on historical data in order to predict customer scores of the different items. Items with the highest scores are recommended to customers.
Keywords: Deep Learning, Long Short-Term Memory, Machine Learning, Recommender Systems, Support Vector Machine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3263539 Land Layout and Urban Design of New Cities in Underdeveloped Areas of China: A Case Study of Xixian New Area
Authors: Libin Ouyang
Abstract:
China has experienced a very fast urbanization process in the past two decades. Due to the uncoordinated characteristics of regional development in China, a large number of people from rural areas or small towns have flooded into regional central cities, which are building new cities around them due to the shortage of construction land or the need for urban development. However, the construction of some new cities has not achieved the expected effect, the absorption capacity of industry and population is limited, and the phenomenon of capital and land waste is obvious. This paper takes Xixian New Area in Shaanxi Province, an inland area in Northwest China, as an example, and tries to analyze the reasons for the lack of vitality in Xixian New Area from the perspectives of land use layout and urban design. This paper will also select the Energy-Finance-Trade Start-up Area in Xixian New Area as an important research site, and study how to optimize the land use layout and urban design to ease the population of big cities, effectively solve the problems of big cities, improve the vitality and attractiveness of the new city, and promote the sustainable development of the new city. The study can provide reference for urban planning practitioners and policy makers, provide theoretical help for the construction of new cities in other underdeveloped regions of China, and provide certain case references for the construction of cities in other developing countries in the process of rapid urbanization.
Keywords: New city, land use layout, urban design, urban planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2843538 Sensitivity of the SHARC Model to Variations of Manning Coefficient and Effect of “n“ on the Sediment Materials Entry into the Eastern Water intake- A Case in the Dez Diversion Weir in Iran
Authors: M.R.Mansoujian, A.Rohani, N.Hedayat , M.Qamari, M. Osroosh
Abstract:
Permanent rivers are the main sources of renewable water supply for the croplands under the irrigation and drainage schemes. They are also the major source of sediment loads transport into the storage reservoirs of the hydro-electrical dams, diversion weirs and regulating dams. Sedimentation process results from soil erosion which is related to poor watershed management and human intervention ion in the hydraulic regime of the rivers. These could change the hydraulic behavior and as such, leads to riverbed and river bank scouring, the consequences of which would be sediment load transport into the dams and therefore reducing the flow discharge in water intakes. The present paper investigate sedimentation process by varying the Manning coefficient "n" by using the SHARC software along the watercourse in the Dez River. Results indicated that the optimum "n" within that river range is 0.0315 at which quantity minimum sediment loads are transported into the Eastern intake. Comparison of the model results with those obtained by those from the SSIIM software within the same river reach showed a very close proximity between them. This suggests a relative accuracy with which the model can simulate the hydraulic flow characteristics and therefore its suitability as a powerful analytical tool for project feasibility studies and project implementation.Keywords: Sediment transport, Manning coefficient, Eastern Intake, SHARC, Dez River.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16813537 Inferring User Preference Using Distance Dependent Chinese Restaurant Process and Weighted Distribution for a Content Based Recommender System
Authors: Bagher Rahimpour Cami, Hamid Hassanpour, Hoda Mashayekhi
Abstract:
Nowadays websites provide a vast number of resources for users. Recommender systems have been developed as an essential element of these websites to provide a personalized environment for users. They help users to retrieve interested resources from large sets of available resources. Due to the dynamic feature of user preference, constructing an appropriate model to estimate the user preference is the major task of recommender systems. Profile matching and latent factors are two main approaches to identify user preference. In this paper, we employed the latent factor and profile matching to cluster the user profile and identify user preference, respectively. The method uses the Distance Dependent Chines Restaurant Process as a Bayesian nonparametric framework to extract the latent factors from the user profile. These latent factors are mapped to user interests and a weighted distribution is used to identify user preferences. We evaluate the proposed method using a real-world data-set that contains news tweets of a news agency (BBC). The experimental results and comparisons show the superior recommendation accuracy of the proposed approach related to existing methods, and its ability to effectively evolve over time.Keywords: Content-based recommender systems, dynamic user modeling, extracting user interests, predicting user preference.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8153536 Logic Programming and Artificial Neural Networks in Pharmacological Screening of Schinus Essential Oils
Authors: José Neves, M. Rosário Martins, Fátima Candeias, Diana Ferreira, Sílvia Arantes, Júlio Cruz-Morais, Guida Gomes, Joaquim Macedo, António Abelha, Henrique Vicente
Abstract:
Some plants of genus Schinus have been used in the folk medicine as topical antiseptic, digestive, purgative, diuretic, analgesic or antidepressant, and also for respiratory and urinary infections. Chemical composition of essential oils of S. molle and S. terebinthifolius had been evaluated and presented high variability according with the part of the plant studied and with the geographic and climatic regions. The pharmacological properties, namely antimicrobial, anti-tumoural and anti-inflammatory activities are conditioned by chemical composition of essential oils. Taking into account the difficulty to infer the pharmacological properties of Schinus essential oils without hard experimental approach, this work will focus on the development of a decision support system, in terms of its knowledge representation and reasoning procedures, under a formal framework based on Logic Programming, complemented with an approach to computing centered on Artificial Neural Networks and the respective Degree-of-Confidence that one has on such an occurrence.Keywords: Artificial neuronal networks, essential oils, knowledge representation and reasoning, logic programming, Schinus molle L, Schinus terebinthifolius raddi.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24223535 Noise Reduction in Web Data: A Learning Approach Based on Dynamic User Interests
Authors: Julius Onyancha, Valentina Plekhanova
Abstract:
One of the significant issues facing web users is the amount of noise in web data which hinders the process of finding useful information in relation to their dynamic interests. Current research works consider noise as any data that does not form part of the main web page and propose noise web data reduction tools which mainly focus on eliminating noise in relation to the content and layout of web data. This paper argues that not all data that form part of the main web page is of a user interest and not all noise data is actually noise to a given user. Therefore, learning of noise web data allocated to the user requests ensures not only reduction of noisiness level in a web user profile, but also a decrease in the loss of useful information hence improves the quality of a web user profile. Noise Web Data Learning (NWDL) tool/algorithm capable of learning noise web data in web user profile is proposed. The proposed work considers elimination of noise data in relation to dynamic user interest. In order to validate the performance of the proposed work, an experimental design setup is presented. The results obtained are compared with the current algorithms applied in noise web data reduction process. The experimental results show that the proposed work considers the dynamic change of user interest prior to elimination of noise data. The proposed work contributes towards improving the quality of a web user profile by reducing the amount of useful information eliminated as noise.Keywords: Web log data, web user profile, user interest, noise web data learning, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1734