Search results for: Pareto Optimal Solution
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3902

Search results for: Pareto Optimal Solution

2642 A Novel Approach to Handle Uncertainty in Health System Variables for Hospital Admissions

Authors: Manisha Rathi, Thierry Chaussalet

Abstract:

Hospital staff and managers are under pressure and concerned for effective use and management of scarce resources. The hospital admissions require many decisions that have complex and uncertain consequences for hospital resource utilization and patient flow. It is challenging to predict risk of admissions and length of stay of a patient due to their vague nature. There is no method to capture the vague definition of admission of a patient. Also, current methods and tools used to predict patients at risk of admission fail to deal with uncertainty in unplanned admission, LOS, patients- characteristics. The main objective of this paper is to deal with uncertainty in health system variables, and handles uncertain relationship among variables. An introduction of machine learning techniques along with statistical methods like Regression methods can be a proposed solution approach to handle uncertainty in health system variables. A model that adapts fuzzy methods to handle uncertain data and uncertain relationships can be an efficient solution to capture the vague definition of admission of a patient.

Keywords: Admission, Fuzzy, Regression, Uncertainty

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1420
2641 Multimodal Reasoning in a Knowledge Engineering Framework for Product Support

Authors: Rossitza M. Setchi, Nikolaos Lagos

Abstract:

Problem solving has traditionally been one of the principal research areas for artificial intelligence. Yet, although artificial intelligence reasoning techniques have been employed in several product support systems, the benefit of integrating product support, knowledge engineering, and problem solving, is still unclear. This paper studies the synergy of these areas and proposes a knowledge engineering framework that integrates product support systems and artificial intelligence techniques. The framework includes four spaces; the data, problem, hypothesis, and solution ones. The data space incorporates the knowledge needed for structured reasoning to take place, the problem space contains representations of problems, and the hypothesis space utilizes a multimodal reasoning approach to produce appropriate solutions in the form of virtual documents. The solution space is used as the gateway between the system and the user. The proposed framework enables the development of product support systems in terms of smaller, more manageable steps while the combination of different reasoning techniques provides a way to overcome the lack of documentation resources.

Keywords: Knowledge engineering framework, product support, case-based reasoning, model-based reasoning, multimodal reasoning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1796
2640 Evaluation of a Hybrid Knowledge-Based System Using Fuzzy Approach

Authors: Kamalendu Pal

Abstract:

This paper describes the main features of a knowledge-based system evaluation method. System evaluation is placed in the context of a hybrid legal decision-support system, Advisory Support for Home Settlement in Divorce (ASHSD). Legal knowledge for ASHSD is represented in two forms, as rules and previously decided cases. Besides distinguishing the two different forms of knowledge representation, the paper outlines the actual use of these forms in a computational framework that is designed to generate a plausible solution for a given case, by using rule-based reasoning (RBR) and case-based reasoning (CBR) in an integrated environment. The nature of suitability assessment of a solution has been considered as a multiple criteria decision-making process in ASHAD evaluation. The evaluation was performed by a combination of discussions and questionnaires with different user groups. The answers to questionnaires used in this evaluations method have been measured as a fuzzy linguistic term. The finding suggests that fuzzy linguistic evaluation is practical and meaningful in knowledge-based system development purpose. 

Keywords: Case-based reasoning, decision-support system, fuzzy linguistic term, rule-based reasoning, system evaluation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1632
2639 Integrated Subset Split for Balancing Network Utilization and Quality of Routing

Authors: S. V. Kasmir Raja, P. Herbert Raj

Abstract:

The overlay approach has been widely used by many service providers for Traffic Engineering (TE) in large Internet backbones. In the overlay approach, logical connections are set up between edge nodes to form a full mesh virtual network on top of the physical topology. IP routing is then run over the virtual network. Traffic engineering objectives are achieved through carefully routing logical connections over the physical links. Although the overlay approach has been implemented in many operational networks, it has a number of well-known scaling issues. This paper proposes a new approach to achieve traffic engineering without full-mesh overlaying with the help of integrated approach and equal subset split method. Traffic engineering needs to determine the optimal routing of traffic over the existing network infrastructure by efficiently allocating resource in order to optimize traffic performance on an IP network. Even though constraint-based routing [1] of Multi-Protocol Label Switching (MPLS) is developed to address this need, since it is not widely tested or debugged, Internet Service Providers (ISPs) resort to TE methods under Open Shortest Path First (OSPF), which is the most commonly used intra-domain routing protocol. Determining OSPF link weights for optimal network performance is an NP-hard problem. As it is not possible to solve this problem, we present a subset split method to improve the efficiency and performance by minimizing the maximum link utilization in the network via a small number of link weight modifications. The results of this method are compared against results of MPLS architecture [9] and other heuristic methods.

Keywords: Constraint based routing, Link Utilization, Subsetsplit method and Traffic Engineering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1397
2638 Optimum Time Coordination of Overcurrent Relays using Two Phase Simplex Method

Authors: Prashant P. Bedekar, Sudhir R. Bhide, Vijay S. Kale

Abstract:

Overcurrent (OC) relays are the major protection devices in a distribution system. The operating time of the OC relays are to be coordinated properly to avoid the mal-operation of the backup relays. The OC relay time coordination in ring fed distribution networks is a highly constrained optimization problem which can be stated as a linear programming problem (LPP). The purpose is to find an optimum relay setting to minimize the time of operation of relays and at the same time, to keep the relays properly coordinated to avoid the mal-operation of relays. This paper presents two phase simplex method for optimum time coordination of OC relays. The method is based on the simplex algorithm which is used to find optimum solution of LPP. The method introduces artificial variables to get an initial basic feasible solution (IBFS). Artificial variables are removed using iterative process of first phase which minimizes the auxiliary objective function. The second phase minimizes the original objective function and gives the optimum time coordination of OC relays.

Keywords: Constrained optimization, LPP, Overcurrent relaycoordination, Two-phase simplex method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3007
2637 Comprehensive Characteristics of the Municipal Solid Waste Generated in the Faculty of Engineering, UKM

Authors: A. Salsabili, M.Aghajani Mir, S.Saheri, Noor Ezlin Ahmad Basri

Abstract:

The main aims in this research are to study the solid waste generation in the Faculty of Engineering and Built Environment in the UKM and at the same time to determine composition and some of the waste characteristics likewise: moisture content, density, pH and C/N ratio. For this purpose multiple campaigns were conducted to collect the wastes produced in all hostels, faculties, offices and so on, during 24th of February till 2nd of March 2009, measure and investigate them with regard to both physical and chemical characteristics leading to highlight the necessary management policies. Research locations are Faculty of Engineering and the Canteen nearby that. From the result gained, the most suitable solid waste management solution will be proposed to UKM. The average solid waste generation rate in UKM is 203.38 kg/day. The composition of solid waste generated are glass, plastic, metal, aluminum, organic and inorganic waste and others waste. From the laboratory result, the average moisture content, density, pH and C/N ratio values from the solid waste generated are 49.74%, 165.1 kg/m3, 5.3, and 7:1 respectively. Since, the food waste (organic waste) were the most dominant component, around 62% from the total waste generated hence, the most suitable solid waste management solution is composting.

Keywords: Solid Waste, Waste Management, Characterizationand Composition

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3596
2636 Sulphur-Mediated Precipitation of Pt/Fe/Co/CrIons in Liquid-Liquid and Gas-Liquid Chloride Systems

Authors: J. Siame, H. Kasaini

Abstract:

The proof of concept experiments were conducted to determine the feasibility of using small amounts of Dissolved Sulphur (DS) from the gaseous phase to precipitate platinum ions in chloride media. Two sets of precipitation experiments were performed in which the source of sulphur atoms was either a thiosulphate solution (Na2S2O3) or a sulphur dioxide gas (SO2). In liquid-liquid (L-L) system, complete precipitation of Pt was achieved at small dosages of Na2S2O3 (0.01 – 1.0 M) in a time interval of 3-5 minutes. On the basis of this result, gas absorption tests were carried out mainly to achieve sulphur solubility equivalent to 0.018 M. The idea that huge amounts of precious metals could be recovered selectively from their dilute solutions by utilizing the waste SO2 streams at low pressure seemed attractive from the economic and environmental point of views. Therefore, mass transfer characteristics of SO2 gas associated with reactive absorption across the gas-liquid (G-L) interface were evaluated under different conditions of pressure (0.5 – 2 bar), solution temperature ranges from 20 – 50 oC and acid strength (1 – 4 M, HCl). This paper concludes with information about selective precipitation of Pt in the presence of cations (Fe2+, Co2+, and Cr3+) in a CSTR and recommendation to scale up laboratory data to industrial pilot scale operations.

Keywords: CSTR, diffusivity, platinum, selective precipitation, sulphur dioxide, thiosulphate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2157
2635 Developing a Sustainable Educational Portal for the D-Grid Community

Authors: Viktor Achter, Sebastian Breuers, Marc Seifert, Ulrich Lang, Joachim Götze, Bernd Reuther, Paul Müller

Abstract:

Within the last years, several technologies have been developed to help building e-learning portals. Most of them follow approaches that deliver a vast amount of functionalities, suitable for class-like learning. The SuGI project, as part of the D-Grid (funded by the BMBF), targets on delivering a highly scalable and sustainable learning solution to provide materials (e.g. learning modules, training systems, webcasts, tutorials, etc.) containing knowledge about Grid computing to the D-Grid community. In this article, the process of the development of an e-learning portal focused on the requirements of this special user group is described. Furthermore, it deals with the conceptual and technical design of an e-learning portal, addressing the special needs of heterogeneous target groups. The main focus lies on the quality management of the software development process, Web templates for uploading new contents, the rich search and filter functionalities which will be described from a conceptual as well as a technical point of view. Specifically, it points out best practices as well as concepts to provide a sustainable solution to a relatively unknown and highly heterogeneous community.

Keywords: D-Grid, e-learning, e-science, Grid computing, SuGI.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1346
2634 Integrated Simulation and Optimization for Carbon Capture and Storage System

Authors: Taekyoon Park, Seok Goo Lee, Sung Ho Kim, Ung Lee, Jong Min Lee, Chonghun Han

Abstract:

CO2 capture and storage/sequestration (CCS) is a key technology for addressing the global warming issue. This paper proposes an integrated model for the whole chain of CCS, from a power plant to a reservoir. The integrated model is further utilized to determine optimal operating conditions and study responses to various changes in input variables.

Keywords: CCS, Caron Dioxide, Carbon Capture and Storage, Simulation, Optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2686
2633 Thermo-Sensitive Hydrogel: Control of Hydrophilic-Hydrophobic Transition

Authors: Wanwipa Siriwatwechakul, Nutte Teraphongphom, Vatcharani Ngaotheppitak, Sureeporn Kunataned

Abstract:

The study investigated the hydrophilic to hydrophobic transition of modified polyacrylamide hydrogel with the inclusion of N-isopropylacrylamide (NIAM). The modification was done by mimicking micellar polymerization, which resulted in better arrangement of NIAM chains in the polyacrylamide network. The degree of NIAM arrangement is described by NH number. The hydrophilic to hydrophobic transition was measured through the partition coefficient, K, of Orange II and Methylene Blue in hydrogel and in water. These dyes were chosen as a model for solutes with different degree of hydrophobicity. The study showed that the hydrogel with higher NH values resulted in better solubility of both dyes. Moreover, in temperature above the lower critical solution temperature (LCST) of Poly(N-isopropylacrylamide) (PNIAM)also caused the collapse of NIPAM chains which results in a more hydrophobic environment that increases the solubility of Methylene Blue and decreases the solubility of Orange II in the hydrogels with NIPAM present.

Keywords: Thermo-sensitive hydrogel, partition coefficient, the lower critical solution temperature (LCST), micellar polymerization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2856
2632 Accurate And Efficient Global Approximation using Adaptive Polynomial RSM for Complex Mechanical and Vehicular Performance Models

Authors: Y. Z. Wu, Z. Dong, S. K. You

Abstract:

Global approximation using metamodel for complex mathematical function or computer model over a large variable domain is often needed in sensibility analysis, computer simulation, optimal control, and global design optimization of complex, multiphysics systems. To overcome the limitations of the existing response surface (RS), surrogate or metamodel modeling methods for complex models over large variable domain, a new adaptive and regressive RS modeling method using quadratic functions and local area model improvement schemes is introduced. The method applies an iterative and Latin hypercube sampling based RS update process, divides the entire domain of design variables into multiple cells, identifies rougher cells with large modeling error, and further divides these cells along the roughest dimension direction. A small number of additional sampling points from the original, expensive model are added over the small and isolated rough cells to improve the RS model locally until the model accuracy criteria are satisfied. The method then combines local RS cells to regenerate the global RS model with satisfactory accuracy. An effective RS cells sorting algorithm is also introduced to improve the efficiency of model evaluation. Benchmark tests are presented and use of the new metamodeling method to replace complex hybrid electrical vehicle powertrain performance model in vehicle design optimization and optimal control are discussed.

Keywords: Global approximation, polynomial response surface, domain decomposition, domain combination, multiphysics modeling, hybrid powertrain optimization

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1908
2631 A Study on Removal Characteristics of (Mn2+) from Aqueous Solution by CNT

Authors: Nassereldeen A. Kabashi, Suleyman A. Muyibi. Mohammed E. Saeed., Farhana I. Yahya

Abstract:

It is important to remove manganese from water because of its effects on human and the environment. Human activities are one of the biggest contributors for excessive manganese concentration in the environment. The proposed method to remove manganese in aqueous solution by using adsorption as in carbon nanotubes (CNT) at different parameters: The parameters are CNT dosage, pH, agitation speed and contact time. Different pHs are pH 6.0, pH 6.5, pH 7.0, pH 7.5 and pH 8.0, CNT dosages are 5mg, 6.25mg, 7.5mg, 8.75mg or 10mg, contact time are 10 min, 32.5 min, 55 min, 87.5 min and 120 min while the agitation speeds are 100rpm, 150rpm, 200rpm, 250rpm and 300rpm. The parameters chosen for experiments are based on experimental design done by using Central Composite Design, Design Expert 6.0 with 4 parameters, 5 levels and 2 replications. Based on the results, condition set at pH 7.0, agitation speed of 300 rpm, 7.5mg and contact time 55 minutes gives the highest removal with 75.5%. From ANOVA analysis in Design Expert 6.0, the residual concentration will be very much affected by pH and CNT dosage. Initial manganese concentration is 1.2mg/L while the lowest residual concentration achieved is 0.294mg/L, which almost satisfy DOE Malaysia Standard B requirement. Therefore, further experiments must be done to remove manganese from model water to the required standard (0.2 mg/L) with the initial concentration set to 0.294 mg/L.

Keywords: Adsorption, CNT, DOE, Manganese, Parameters.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1471
2630 Orthogonal Array Application and Response Surface Method Approach for Optimal Product Values: An Application for Oil Blending Process

Authors: Christopher C. Ihueze, Constance C. Obiuto, Christian E. Okafor, Charles C. Okpala

Abstract:

This paper presents a methodical approach for designing and optimizing process parameters in oil blending industries. Twenty seven replicated experiments were conducted for production of A-Z crown super oil (SAE20W/50) employing L9 orthogonal array to establish process response parameters. Power law model was fitted to experimental data and the obtained model was optimized applying the central composite design (CCD) of response surface methodology (RSM). Quadratic model was found to be significant for production of A-Z crown supper oil. The study recognized and specified four new lubricant formulations that conform to ISO oil standard in the course of analyzing the batch productions of A-Z crown supper oil as: L1: KV = 21.8293Cst, BS200 = 9430.00Litres, Ad102=11024.00Litres, PVI = 2520 Litres, L2: KV = 22.513Cst, BS200 = 12430.00 Litres, Ad102 = 11024.00 Litres, PVI = 2520 Litres, L3: KV = 22.1671Cst, BS200 = 9430.00 Litres, Ad102 = 10481.00 Litres, PVI= 2520 Litres, L4: KV = 22.8605Cst, BS200 = 12430.00 Litres, Ad102 = 10481.00 Litres, PVI = 2520 Litres. The analysis of variance showed that quadratic model is significant for kinematic viscosity production while the R-sq value statistic of 0.99936 showed that the variation of kinematic viscosity is due to its relationship with the control factors. This study therefore resulted to appropriate blending proportions of lubricants base oil and additives and recommends the optimal kinematic viscosity of A-Z crown super oil (SAE20W/50) to be 22.86Cst.

Keywords: Additives, control factors, kinematic viscosity, lubricant, orthogonal array, process parameter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1945
2629 Fractal Analysis of 16S rRNA Gene Sequences in Archaea Thermophiles

Authors: T. Holden, G. Tremberger, Jr, E. Cheung, R. Subramaniam, R. Sullivan, N. Gadura, P. Schneider, P. Marchese, A. Flamholz, T. Cheung, D. Lieberman

Abstract:

A nucleotide sequence can be expressed as a numerical sequence when each nucleotide is assigned its proton number. A resulting gene numerical sequence can be investigated for its fractal dimension in terms of evolution and chemical properties for comparative studies. We have investigated such nucleotide fluctuation in the 16S rRNA gene of archaea thermophiles. The studied archaea thermophiles were archaeoglobus fulgidus, methanothermobacter thermautotrophicus, methanocaldococcus jannaschii, pyrococcus horikoshii, and thermoplasma acidophilum. The studied five archaea-euryarchaeota thermophiles have fractal dimension values ranging from 1.93 to 1.97. Computer simulation shows that random sequences would have an average of about 2 with a standard deviation about 0.015. The fractal dimension was found to correlate (negative correlation) with the thermophile-s optimal growth temperature with R2 value of 0.90 (N =5). The inclusion of two aracheae-crenarchaeota thermophiles reduces the R2 value to 0.66 (N = 7). Further inclusion of two bacterial thermophiles reduces the R2 value to 0.50 (N =9). The fractal dimension is correlated (positive) to the sequence GC content with an R2 value of 0.89 for the five archaea-euryarchaeota thermophiles (and 0.74 for the entire set of N = 9), although computer simulation shows little correlation. The highest correlation (positive) was found to be between the fractal dimension and di-nucleotide Shannon entropy. However Shannon entropy and sequence GC content were observed to correlate with optimal growth temperature having an R2 of 0.8 (negative), and 0.88 (positive), respectively, for the entire set of 9 thermophiles; thus the correlation lacks species specificity. Together with another correlation study of bacterial radiation dosage with RecA repair gene sequence fractal dimension, it is postulated that fractal dimension analysis is a sensitive tool for studying the relationship between genotype and phenotype among closely related sequences.

Keywords: Fractal dimension, archaea thermophiles, Shannon entropy, GC content

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1779
2628 MFCA: An Environmental Management Accounting Technique for Optimal Resource Efficiency in Production Processes

Authors: Omolola A. Tajelawi, Hari L. Garbharran

Abstract:

Revenue leakages are one of the major challenges manufacturers face in production processes, as most of the input materials that should emanate as products from the lines are lost as waste. Rather than generating income from material input which is meant to end-up as products, losses are further incurred as costs in order to manage waste generated. In addition, due to the lack of a clear view of the flow of resources on the lines from input to output stage, acquiring information on the true cost of waste generated have become a challenge. This has therefore given birth to the conceptualization and implementation of waste minimization strategies by several manufacturing industries. This paper reviews the principles and applications of three environmental management accounting tools namely Activity-based Costing (ABC), Life-Cycle Assessment (LCA) and Material Flow Cost Accounting (MFCA) in the manufacturing industry and their effectiveness in curbing revenue leakages. The paper unveils the strengths and limitations of each of the tools; beaming a searchlight on the tool that could allow for optimal resource utilization, transparency in production process as well as improved cost efficiency. Findings from this review reveal that MFCA may offer superior advantages with regards to the provision of more detailed information (both in physical and monetary terms) on the flow of material inputs throughout the production process compared to the other environmental accounting tools. This paper therefore makes a case for the adoption of MFCA as a viable technique for the identification and reduction of waste in production processes, and also for effective decision making by production managers, financial advisors and other relevant stakeholders.

Keywords: MFCA, environmental management accounting, resource efficiency, waste reduction, revenue losses.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4407
2627 Side Effects of Dental Tooth Whitening: Data from Literature

Authors: Saimir Heta, Ilma Robo, Emela Dalloshi, Nevila Alliu, Vera Ostreni

Abstract:

The dental whitening process, beyond the fact that it is a mini-invasive dental treatment, has effects on the dental structure or on the pulp of the tooth where it is applied. The electronic search was performed using keywords to find articles published within the last 10 years about side effects, assessed as such, of the minimally invasive dental bleaching treatment. The aim of the study was to evaluate the side effects of bleaching based on the percentage and type of solution used, where the latter was evaluated on the basic solution used for bleaching. The side effects of bleaching are evaluated in selected articles depending on the method of bleaching application, which means it is carried out with recommended solutions, or with mixtures of alternative solutions or substances based on Internet information. The dental bleaching process has side effects which have not yet been definitively evaluated, experimentally in large samples of individuals or animals (mice or cattle) to arrive at accurate numerical conclusions. The trend of publications about this topic is increasing in recent years, as long as the trend for aesthetic facial treatments, including dental ones, is increasing.

Keywords: Teeth whitening, side effects, permanent teeth, formed dental apex.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25
2626 A Robust Extrapolation Method for Curtailed Aperture Reconstruction in Acoustic Imaging

Authors: R. Bremananth

Abstract:

Acoustic Imaging based sound localization using microphone array is a challenging task in digital-signal processing. Discrete Fourier transform (DFT) based near-field acoustical holography (NAH) is an important acoustical technique for sound source localization and provide an efficient solution to the ill-posed problem. However, in practice, due to the usage of small curtailed aperture and its consequence of significant spectral leakage, the DFT could not reconstruct the active-region-of-sound (AROS) effectively, especially near the edges of aperture. In this paper, we emphasize the fundamental problems of DFT-based NAH, provide a solution to spectral leakage effect by the extrapolation based on linear predictive coding and 2D Tukey windowing. This approach has been tested to localize the single and multi-point sound sources. We observe that incorporating extrapolation technique increases the spatial resolution, localization accuracy and reduces spectral leakage when small curtail aperture with a lower number of sensors accounts.

Keywords: Acoustic Imaging, Discrete Fourier Transform (DFT), k-space wavenumber, Near-Field Acoustical Holography (NAH), Source Localization, Spectral Leakage.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1693
2625 Data Envelopment Analysis with Partially Perfect Objects

Authors: Alexander Y. Vaninsky

Abstract:

This paper presents a simplified version of Data Envelopment Analysis (DEA) - a conventional approach to evaluating the performance and ranking of competitive objects characterized by two groups of factors acting in opposite directions: inputs and outputs. DEA with a Perfect Object (DEA PO) augments the group of actual objects with a virtual Perfect Object - the one having greatest outputs and smallest inputs. It allows for obtaining an explicit analytical solution and making a step to an absolute efficiency. This paper develops this approach further and introduces a DEA model with Partially Perfect Objects. DEA PPO consecutively eliminates the smallest relative inputs or greatest relative outputs, and applies DEA PO to the reduced collections of indicators. The partial efficiency scores are combined to get the weighted efficiency score. The computational scheme remains simple, like that of DEA PO, but the advantage of the DEA PPO is taking into account all of the inputs and outputs for each actual object. Firm evaluation is considered as an example.

Keywords: Data Envelopment Analysis, Perfect object, Partially perfect object, Partial efficiency, Explicit solution, Simplified algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1697
2624 New Adaptive Linear Discriminante Analysis for Face Recognition with SVM

Authors: Mehdi Ghayoumi

Abstract:

We have applied new accelerated algorithm for linear discriminate analysis (LDA) in face recognition with support vector machine. The new algorithm has the advantage of optimal selection of the step size. The gradient descent method and new algorithm has been implemented in software and evaluated on the Yale face database B. The eigenfaces of these approaches have been used to training a KNN. Recognition rate with new algorithm is compared with gradient.

Keywords: lda, adaptive, svm, face recognition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1422
2623 Effect of Atmospheric Pressure on the Flow at the Outlet of a Propellant Nozzle

Authors: R. Haoui

Abstract:

The purpose of this work is to simulate the flow at the exit of Vulcan 1 engine of European launcher Ariane 5. The geometry of the propellant nozzle is already determined using the characteristics method. The pressure in the outlet section of the nozzle is less than atmospheric pressure on the ground, causing the existence of oblique and normal shock waves at the exit. During the rise of the launcher, the atmospheric pressure decreases and the shock wave disappears. The code allows the capture of shock wave at exit of nozzle. The numerical technique uses the Flux Vector Splitting method of Van Leer to ensure convergence and avoid the calculation instabilities. The Courant, Friedrichs and Lewy coefficient (CFL) and mesh size level are selected to ensure the numerical convergence. The nonlinear partial derivative equations system which governs this flow is solved by an explicit unsteady numerical scheme by the finite volume method. The accuracy of the solution depends on the size of the mesh and also the step of time used in the discretized equations. We have chosen in this study the mesh that gives us a stationary solution with good accuracy.

Keywords: Launchers, supersonic flow, finite volume, nozzles, shock wave.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 877
2622 Cost Benefit Analysis: Evaluation among the Millimetre Wavebands and SHF Bands of Small Cell 5G Networks

Authors: Emanuel Teixeira, Anderson Ramos, Marisa Lourenço, Fernando J. Velez, Jon M. Peha

Abstract:

This article discusses the benefit cost analysis aspects of millimetre wavebands (mmWaves) and Super High Frequency (SHF). The devaluation along the distance of the carrier-to-noise-plus-interference ratio with the coverage distance is assessed by considering two different path loss models, the two-slope urban micro Line-of-Sight (UMiLoS) for the SHF band and the modified Friis propagation model, for frequencies above 24 GHz. The equivalent supported throughput is estimated at the 5.62, 28, 38, 60 and 73 GHz frequency bands and the influence of carrier-to-noise-plus-interference ratio in the radio and network optimization process is explored. Mostly owing to the lessening caused by the behaviour of the two-slope propagation model for SHF band, the supported throughput at this band is higher than at the millimetre wavebands only for the longest cell lengths. The benefit cost analysis of these pico-cellular networks was analysed for regular cellular topologies, by considering the unlicensed spectrum. For shortest distances, we can distinguish an optimal of the revenue in percentage terms for values of the cell length, R ≈ 10 m for the millimeter wavebands and for longest distances an optimal of the revenue can be observed at R ≈ 550 m for the 5.62 GHz. It is possible to observe that, for the 5.62 GHz band, the profit is slightly inferior than for millimetre wavebands, for the shortest Rs, and starts to increase for cell lengths approximately equal to the ratio between the break-point distance and the co-channel reuse factor, achieving a maximum for values of R approximately equal to 550 m.

Keywords: 5G, millimetre wavebands, super high-frequency band, SINR, signal-to-interference-plus-noise ratio, cost benefit analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 722
2621 GPS and SMS-Based Child Tracking System Using Smart Phone

Authors: A. Al-Mazloum, E. Omer, M. F. A. Abdullah

Abstract:

Recently many cases of missing children between ages 14 and 17 years are reported. Parents always worry about the possibility of kidnapping of their children. This paper proposes an Android based solution to aid parents to track their children in real time. Nowadays, most mobile phones are equipped with location services capabilities allowing us to get the device’s geographic position in real time. The proposed solution takes the advantage of the location services provided by mobile phone since most of kids carry mobile phones. The mobile application use the GPS and SMS services found in Android mobile phones. It allows the parent to get their child’s location on a real time map. The system consists of two sides, child side and parent side. A parent’s device main duty is to send a request location SMS to the child’s device to get the location of the child. On the other hand, the child’s device main responsibility is to reply the GPS position to the parent’s device upon request.

Keywords: Child Tracking System, Global Positioning System (GPS), SMS-based Mobile Application.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15516
2620 Mathematical Modeling of the AMCs Cross-Contamination Removal in the FOUPs: Finite Element Formulation and Application in FOUP’s Decontamination

Authors: N. Santatriniaina, J. Deseure, T.Q. Nguyen, H. Fontaine, C. Beitia, L. Rakotomanana

Abstract:

Nowadays, with the increasing of the wafer's size and the decreasing of critical size of integrated circuit manufacturing in modern high-tech, microelectronics industry needs a maximum attention to challenge the contamination control. The move to 300 [mm] is accompanied by the use of Front Opening Unified Pods for wafer and his storage. In these pods an airborne cross contamination may occur between wafers and the pods. A predictive approach using modeling and computational methods is very powerful method to understand and qualify the AMCs cross contamination processes. This work investigates the required numerical tools which are employed in order to study the AMCs cross-contamination transfer phenomena between wafers and FOUPs. Numerical optimization and finite element formulation in transient analysis were established. Analytical solution of one dimensional problem was developed and the calibration process of physical constants was performed. The least square distance between the model (analytical 1D solution) and the experimental data are minimized. The behavior of the AMCs intransient analysis was determined. The model framework preserves the classical forms of the diffusion and convection-diffusion equations and yields to consistent form of the Fick's law. The adsorption process and the surface roughness effect were also traduced as a boundary condition using the switch condition Dirichlet to Neumann and the interface condition. The methodology is applied, first using the optimization methods with analytical solution to define physical constants, and second using finite element method including adsorption kinetic and the switch of Dirichlet to Neumann condition.

Keywords: AMCs, FOUP, cross-contamination, adsorption, diffusion, numerical analysis, wafers, Dirichlet to Neumann, finite elements methods, Fick’s law, optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3176
2619 On Solution of Interval Valued Intuitionistic Fuzzy Assignment Problem Using Similarity Measure and Score Function

Authors: Gaurav Kumar, Rakesh Kumar Bajaj

Abstract:

The primary objective of the paper is to propose a new method for solving assignment problem under uncertain situation. In the classical assignment problem (AP), zpqdenotes the cost for assigning the qth job to the pth person which is deterministic in nature. Here in some uncertain situation, we have assigned a cost in the form of composite relative degree Fpq instead of  and this replaced cost is in the maximization form. In this paper, it has been solved and validated by the two proposed algorithms, a new mathematical formulation of IVIF assignment problem has been presented where the cost has been considered to be an IVIFN and the membership of elements in the set can be explained by positive and negative evidences. To determine the composite relative degree of similarity of IVIFS the concept of similarity measure and the score function is used for validating the solution which is obtained by Composite relative similarity degree method. Further, hypothetical numeric illusion is conducted to clarify the method’s effectiveness and feasibility developed in the study. Finally, conclusion and suggestion for future work are also proposed.

Keywords: Assignment problem, Interval-valued Intuitionistic Fuzzy Sets, Similarity Measures, score function.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3013
2618 Cadmium Filter Cake of a Hydrometallurgical Zinc Smelter as a New Source for the Biological Synthesis of CdS Quantum Dots

Authors: Mehran Bakhshi, Mohammad Raouf Hosseini, Mohammadhosein Rahimi

Abstract:

The cadmium sulfide nanoparticles were synthesized from the nickel-cadmium cake of a hydrometallurgical zinc producing plant and sodium sulfide as Cd2+ and S-2 sources, respectively. Also, the synthesis process was performed by using the secretions of Bacillus licheniformis as bio-surfactant. Initially, in order to obtain a cadmium rich solution, two following steps were carried out: 1) Alkaline leaching for the removal of zinc oxide from the cake, and 2) acidic leaching to dissolve cadmium from the remained solid residue. Afterward, the obtained CdSO4 solution was used for the nanoparticle biosynthesis. Nanoparticles were characterized by the energy dispersive spectroscopy (EDS) and X-ray diffraction (XRD) to confirm the formation of CdS crystals with cubic structure. Also, transmission electron microscopy (TEM) was applied to determine the particle sizes which were in 2-10 nm range. Moreover, the presence of the protein containing bio-surfactants was approved by using infrared analysis (FTIR). In addition, the absorbance below 400 nm confirms quantum particles’ size. Finally, it was shown that valuable CdS quantum dots could be obtained from the industrial waste products via environment-friendly biological approaches.

Keywords: Biosynthesis, cadmium cake, cadmium sulfide, nanoparticle, zinc smelter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1535
2617 Concept of a Pseudo-Lower Bound Solution for Reinforced Concrete Slabs

Authors: M. De Filippo, J. S. Kuang

Abstract:

In construction industry, reinforced concrete (RC) slabs represent fundamental elements of buildings and bridges. Different methods are available for analysing the structural behaviour of slabs. In the early ages of last century, the yield-line method has been proposed to attempt to solve such problem. Simple geometry problems could easily be solved by using traditional hand analyses which include plasticity theories. Nowadays, advanced finite element (FE) analyses have mainly found their way into applications of many engineering fields due to the wide range of geometries to which they can be applied. In such cases, the application of an elastic or a plastic constitutive model would completely change the approach of the analysis itself. Elastic methods are popular due to their easy applicability to automated computations. However, elastic analyses are limited since they do not consider any aspect of the material behaviour beyond its yield limit, which turns to be an essential aspect of RC structural performance. Furthermore, their applicability to non-linear analysis for modeling plastic behaviour gives very reliable results. Per contra, this type of analysis is computationally quite expensive, i.e. not well suited for solving daily engineering problems. In the past years, many researchers have worked on filling this gap between easy-to-implement elastic methods and computationally complex plastic analyses. This paper aims at proposing a numerical procedure, through which a pseudo-lower bound solution, not violating the yield criterion, is achieved. The advantages of moment distribution are taken into account, hence the increase in strength provided by plastic behaviour is considered. The lower bound solution is improved by detecting over-yielded moments, which are used to artificially rule the moment distribution among the rest of the non-yielded elements. The proposed technique obeys Nielsen’s yield criterion. The outcome of this analysis provides a simple, yet accurate, and non-time-consuming tool of predicting the lower-bound solution of the collapse load of RC slabs. By using this method, structural engineers can find the fracture patterns and ultimate load bearing capacity. The collapse triggering mechanism is found by detecting yield-lines. An application to the simple case of a square clamped slab is shown, and a good match was found with the exact values of collapse load.

Keywords: Computational mechanics, lower bound method, reinforced concrete slabs, yield-line.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1096
2616 High Sensitivity Crack Detection and Locating with Optimized Spatial Wavelet Analysis

Authors: A. Ghanbari Mardasi, N. Wu, C. Wu

Abstract:

In this study, a spatial wavelet-based crack localization technique for a thick beam is presented. Wavelet scale in spatial wavelet transformation is optimized to enhance crack detection sensitivity. A windowing function is also employed to erase the edge effect of the wavelet transformation, which enables the method to detect and localize cracks near the beam/measurement boundaries. Theoretical model and vibration analysis considering the crack effect are first proposed and performed in MATLAB based on the Timoshenko beam model. Gabor wavelet family is applied to the beam vibration mode shapes derived from the theoretical beam model to magnify the crack effect so as to locate the crack. Relative wavelet coefficient is obtained for sensitivity analysis by comparing the coefficient values at different positions of the beam with the lowest value in the intact area of the beam. Afterward, the optimal wavelet scale corresponding to the highest relative wavelet coefficient at the crack position is obtained for each vibration mode, through numerical simulations. The same procedure is performed for cracks with different sizes and positions in order to find the optimal scale range for the Gabor wavelet family. Finally, Hanning window is applied to different vibration mode shapes in order to overcome the edge effect problem of wavelet transformation and its effect on the localization of crack close to the measurement boundaries. Comparison of the wavelet coefficients distribution of windowed and initial mode shapes demonstrates that window function eases the identification of the cracks close to the boundaries.

Keywords: Edge effect, scale optimization, small crack locating, spatial wavelet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 951
2615 Aircraft Selection Process Using Preference Analysis for Reference Ideal Solution (PARIS)

Authors: C. Ardil

Abstract:

Multiple criteria decision making analysis (MCDMA) methods are applied to many real - life problems in different fields of engineering science and technology. The "preference analysis for reference ideal solution (PARIS)" method is proposed for an efficient MCDMA evaluation of decision problems. The multiple criteria aircraft evaluation approach is based on the integrated the mean weight, entropy weight, PARIS, and TOPSIS method, which eliminates the subjective importance weight assignment process. The evaluation criteria were identified from an extensive literature review of aircraft selection process. The aim of this study is to propose an efficient methodology for handling the aircraft selection process in which the proposed method solves effectively the MCDMA problem. A numerical example is presented to demonstrate the applicability and validity of the proposed MCDMA approach. 

Keywords: aircraft selection, aircraft, multiple criteria decision making, multiple criteria decision making analysis, mean weight, entropy weight, MCDMA, PARIS, TOPSIS, VIKOR, ELECTRE, PROMETHEE

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 536
2614 Studying the Effect of Ethanol and Operating Temperature on Purification of Lactulose Syrup Containing Lactose

Authors: N. Zanganeh, M. Zabet

Abstract:

Lactulose is a synthetic disaccharide which has remarkable applications in food and pharmaceutical fields. Lactulose is not found in nature and it is produced by isomerization reaction of lactose in an alkaline environment. It should be noted that this reaction has a very low yield since significant amount of lactose stays un-reacted in the system. Basically, purification of lactulose is difficult and costly. Previous studies have revealed that solubility of lactose and lactulose are significantly different in ethanol. Considering the fact that solubility is also affected by temperature itself, we investigated the effect of ethanol and temperature on separation process of lactose from the syrup containing lactose and lactulose. For this purpose, a saturated solution containing lactulose and lactose was made at three different temperatures; 25⁰C (room temperature), 31⁰C, and 37⁰C first.  Five samples containing 2g saturated solution was taken and then 2g, 3g, 4g, 5g, and 6g ethanol separately was added to the sampling tubes. Sampling tubes were kept at respective temperatures afterward. The concentration of lactose and lactulose after separation process measured and analyzed by High Performance Liquid Chromatography (HPLC). Results showed that ethanol has such a greater impact than operating temperature on purification process. Also, it was observed that the maximum rate of separation occurred at initial amount of added ethanol.

Keywords: Ethanol, lactose, lactulose syrup, purification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1121
2613 Treatment of Wastewater from Wet Scrubbers in Secondary Lead Smelters for Recycling and Lead Recovery

Authors: Mahmoud A. Rabah

Abstract:

The present study shows a method to recover lead metal from wastewater of wet scrubber in secondary lead smelter. The wastewater is loaded with 42,000 ppm of insoluble lead compounds (TSP) submicron in diameter. The technical background benefits the use of cationic polyfloc solution to flocculate these colloidal solids before press filtration. The polymer solution is injected in the wastewater stream in a countercurrent flow design. The study demonstrates the effect of polymer dose, temperature, pH, flow velocity of the wastewater and different filtration media on the filtration extent. Results indicated that filtration rate (¦r), quality of purified water, purifying efficiency (¦e) and floc diameter decrease regularly with increase in mass flow rate and velocity up to turbulence of 0.5 m.sec-1. Laminar flow is in favor of flocculation. Polyfloc concentration of 0.75 – 1.25 g/m3 wastewater is convenient. Increasing temperature of the wastewater and pneumatic pressure of filtration enhances ¦r. High pH value deforms floc formation and assists degradation of the filtration fabric. The overall efficiency of the method amounts to 93.2 %. Lead metal was recovered from the filtrate cake using carbon as a reducing agent at 900°C.

Keywords: Wastewater, wet scrubbers, filtration, secondary lead.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3443