Search results for: validation process.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5755

Search results for: validation process.

4255 Application Quality Function Deployment (QFD) Tool in Design of Aero Pumps Based on System Engineering

Authors: Z. Soleymani, M. Amirzadeh

Abstract:

Quality Function Deployment (QFD) was developed in 1960 in Japan and introduced in 1983 in America and Europe. The paper presents a real application of this technique in a way that the method of applying QFD in design and production aero fuel pumps has been considered. While designing a product and in order to apply system engineering process, the first step is identification customer needs then its transition to engineering parameters. Since each change in deign after production process leads to extra human costs and also increase in products quality risk, QFD can make benefits in sale by meeting customer expectations. Since the needs identified as well, the use of QFD tool can lead to increase in communications and less deviation in design and production phases, finally it leads to produce the products with defined technical attributes.

Keywords: Customer voice, engineering parameters, QFD, gear pump.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1695
4254 Using Single Decision Tree to Assess the Impact of Cutting Conditions on Vibration

Authors: S. Ghorbani, N. I. Polushin

Abstract:

Vibration during machining process is crucial since it affects cutting tool, machine, and workpiece leading to a tool wear, tool breakage, and an unacceptable surface roughness. This paper applies a nonparametric statistical method, single decision tree (SDT), to identify factors affecting on vibration in machining process. Workpiece material (AISI 1045 Steel, AA2024 Aluminum alloy, A48-class30 Gray Cast Iron), cutting tool (conventional, cutting tool with holes in toolholder, cutting tool filled up with epoxy-granite), tool overhang (41-65 mm), spindle speed (630-1000 rpm), feed rate (0.05-0.075 mm/rev) and depth of cut (0.05-0.15 mm) were used as input variables, while vibration was the output parameter. It is concluded that workpiece material is the most important parameters for natural frequency followed by cutting tool and overhang.

Keywords: Cutting condition, vibration, natural frequency, decision tree, CART algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1406
4253 Design Optimization of Ferrocement-Laminated Plate Using Genetic Algorithm

Authors: M. Rokonuzzaman, Z. Gürdal

Abstract:

This paper describes the design optimization of ferrocement-laminated plate made up of reinforcing steel wire mesh(es) and cement mortar. For the improvement of the designing process, the plate is modeled as a multi-layer medium, dividing the ferrocement plate into layers of mortar and ferrocement. The mortar layers are assumed to be isotropic in nature and the ferrocement layers are assumed to be orthotropic. The ferrocement layers are little stiffer, but much more costlier, than the mortar layers due the presence of steel wire mesh. The optimization is performed for minimum weight design of the laminate using a genetic algorithm. The optimum designs are discussed for different plate configurations and loadings, and it is compared with the worst designs obtained at the final generation. The paper provides a procedure for the designers in decision-making process.

Keywords: Buckling, Ferrocement-Laminated Plate, Genetic Algorithm, Plate Theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2155
4252 Experimental Analysis of Diesel Hydrotreating Reactor to Development a Simplified Tool for Process Real- time Optimization

Authors: S.Shokri, S.Zahedi, M.Ahmadi Marvast, B. Baloochi, H.Ganji

Abstract:

In this research, a systematic investigation was carried out to determine the optimum conditions of HDS reactor. Moreover, a suitable model was developed for a rigorous RTO (real time optimization) loop of HDS (Hydro desulfurization) process. A systematic experimental series was designed based on CCD (Central Composite design) and carried out in the related pilot plant to tune the develop model. The designed variables in the experiments were Temperature, LHSV and pressure. However, the hydrogen over fresh feed ratio was remained constant. The ranges of these variables were respectively equal to 320-380ºC, 1- 21/hr and 50-55 bar. a power law kinetic model was also developed for our further research in the future .The rate order and activation energy , power of reactant concentration and frequency factor of this model was respectively equal to 1.4, 92.66 kJ/mol and k0=2.7*109 .

Keywords: Statistical model, Multiphase Reactors, Gas oil, Hydrodesulfurization, Optimization, Kinetics

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2654
4251 Statistically Significant Differences of Carbon Dioxide and Carbon Monoxide Emission in Photocopying Process

Authors: Kiurski S. Jelena, Kecić S. Vesna, Oros B. Ivana

Abstract:

Experimental results confirmed the temporal variation of carbon dioxide and carbon monoxide concentration during the working shift of the photocopying process in a small photocopying shop in Novi Sad, Serbia. The statistically significant differences of target gases were examined with two-way analysis of variance without replication followed by Scheffe's post hoc test. The existence of statistically significant differences was obtained for carbon monoxide emission which is pointed out with F-values (12.37 and 31.88) greater than Fcrit (6.94) in contrary to carbon dioxide emission (F-values of 1.23 and 3.12 were less than Fcrit).  Scheffe's post hoc test indicated that sampling point A (near the photocopier machine) and second time interval contribute the most on carbon monoxide emission.

Keywords: Analysis of variance, carbon dioxide, carbon monoxide, photocopying indoor, Scheffe's test

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1578
4250 Computing Entropy for Ortholog Detection

Authors: Hsing-Kuo Pao, John Case

Abstract:

Biological sequences from different species are called or-thologs if they evolved from a sequence of a common ancestor species and they have the same biological function. Approximations of Kolmogorov complexity or entropy of biological sequences are already well known to be useful in extracting similarity information between such sequences -in the interest, for example, of ortholog detection. As is well known, the exact Kolmogorov complexity is not algorithmically computable. In prac-tice one can approximate it by computable compression methods. How-ever, such compression methods do not provide a good approximation to Kolmogorov complexity for short sequences. Herein is suggested a new ap-proach to overcome the problem that compression approximations may notwork well on short sequences. This approach is inspired by new, conditional computations of Kolmogorov entropy. A main contribution of the empir-ical work described shows the new set of entropy-based machine learning attributes provides good separation between positive (ortholog) and nega-tive (non-ortholog) data - better than with good, previously known alter-natives (which do not employ some means to handle short sequences well).Also empirically compared are the new entropy based attribute set and a number of other, more standard similarity attributes sets commonly used in genomic analysis. The various similarity attributes are evaluated by cross validation, through boosted decision tree induction C5.0, and by Receiver Operating Characteristic (ROC) analysis. The results point to the conclu-sion: the new, entropy based attribute set by itself is not the one giving the best prediction; however, it is the best attribute set for use in improving the other, standard attribute sets when conjoined with them.

Keywords: compression, decision tree, entropy, ortholog, ROC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1807
4249 The Impact of 21st Century Technology in Higher Education: The Role of Artificial Intelligence

Authors: Josefina Bengoechea, Alex Bell

Abstract:

Higher education, with its brick-and-mortar facilities and credits-based on hours of study, was developed to serve the needs of a national, industrial, analogue economy. However, the ongoing process of globalization on the one hand, and the emergence of ever-changing needs of employers on the other hand, make this type of process-based education obsolete, and exclusive to students who can afford to pay a full-time tuition and dedicate 4 years of their lives exclusively to study. The creative destruction brought about by new technologies in the 21st century will not only reconfigure the labour market, as millions of jobs will be lost to Artificial Intelligence. The purpose of this paper is to consider if the implementation of technology is the solution to the problems faced in higher education. The paper builds upon a constructivist approach, combining a literature review and research on key publications.

Keywords: Artificial intelligence, employability, labour market, new technology in higher education.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 675
4248 Improvement of Ground Truth Data for Eye Location on Infrared Driver Recordings

Authors: Sorin Valcan, Mihail Găianu

Abstract:

Labeling is a very costly and time consuming process which aims to generate datasets for training neural networks in several functionalities and projects. For driver monitoring system projects, the need of labeled images has a significant impact on the budget and distribution of effort. This paper presents the modifications done to a ground truth data generation algorithm for 2D eyes location on infrared images with drivers in order to improve the quality of the data and performance of the trained neural networks. The algorithm restrictions become tougher which makes it more accurate but also less constant. The resulting dataset becomes smaller and shall not be altered by any kind of manual labels adjustment before being used in the neural networks training process. These changes resulted in a much better performance of the trained neural networks.

Keywords: Labeling automation, infrared camera, driver monitoring, eye detection, Convolutional Neural Networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 369
4247 Neighborhood Sustainability Assessment Tools: A Conceptual Framework for Their Use in Building Adaptive Capacity to Climate Change

Authors: Sally Naji, Julie Gwilliam

Abstract:

Climate change remains a challenging matter for the human and the built environment in the 21st century, where the need to consider adaptation to climate change in the development process is paramount. However, there remains a lack of information regarding how we should prepare responses to this issue, such as through developing organized and sophisticated tools enabling the adaptation process. This study aims to build a systematic framework approach to investigate the potentials that Neighborhood Sustainability Assessment tools (NSA) might offer in enabling both the analysis of the emerging adaptive capacity to climate change. The analysis of the framework presented in this paper aims to discuss this issue in three main phases. The first part attempts to link sustainability and climate change, in the context of adaptive capacity. It is argued that in deciding to promote sustainability in the context of climate change, both the resilience and vulnerability processes become central. However, there is still a gap in the current literature regarding how the sustainable development process can respond to climate change. As well as how the resilience of practical strategies might be evaluated. It is suggested that the integration of the sustainability assessment processes with both the resilience thinking process, and vulnerability might provide important components for addressing the adaptive capacity to climate change. A critical review of existing literature is presented illustrating the current lack of work in this field, integrating these three concepts in the context of addressing the adaptive capacity to climate change. The second part aims to identify the most appropriate scale at which to address the built environment for the climate change adaptation. It is suggested that the neighborhood scale can be considered as more suitable than either the building or urban scales. It then presents the example of NSAs, and discusses the need to explore their potential role in promoting the adaptive capacity to climate change. The third part of the framework presents a comparison among three example NSAs, BREEAM Communities, LEED-ND, and CASBEE-UD. These three tools have been selected as the most developed and comprehensive assessment tools that are currently available for the neighborhood scale. This study concludes that NSAs are likely to present the basis for an organized framework to address the practical process for analyzing and yet promoting Adaptive Capacity to Climate Change. It is further argued that vulnerability (exposure & sensitivity) and resilience (Interdependence & Recovery) form essential aspects to be addressed in the future assessment of NSA’s capability to adapt to both short and long term climate change impacts. Finally, it is acknowledged that further work is now required to understand impact assessment in terms of the range of physical sectors (Water, Energy, Transportation, Building, Land Use and Ecosystems), Actor and stakeholder engagement as well as a detailed evaluation of the NSA indicators, together with a barriers diagnosis process.

Keywords: Adaptive capacity, climate change, NSA tools, resilience, vulnerability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2140
4246 Identification of Social Responsibility Factors within Mega Construction Projects

Authors: Ali Alotaibi, Francis Edum-Fotwe, Andrew Price /

Abstract:

Mega construction projects create buildings and major infrastructure to respond to work and life requirements while playing a vital role in promoting any nation’s economy. However, the industry is often criticised for not balancing economic, environmental and social dimensions of their projects, with emphasis typically on one aspect to the detriment of the others. This has resulted in many negative impacts including environmental pollution, waste throughout the project lifecycle, low productivity, and avoidable accidents. The identification of comprehensive Social Responsibility (SR) indicators, which combine social, environmental and economic aspects, is urgently needed. This is particularly the case in the context of the Kingdom of Saudi Arabia (KSA), which often has mega public construction projects. The aim of this paper is to develop a set of wide-ranging SR indicators which encompass social, economic and environmental aspects unique to the KSA. A qualitative approach was applied to explore relevant indicators through a review of the existing literature, international standards and reports. A list of appropriate indicators was developed, and its comprehensiveness was corroborated by interviews with experts on mega construction projects working with SR concepts in the KSA. The findings present 39 indicators and their metrics, covering 10 economic, 12 environmental and 17 social aspects of SR mapped against their references. These indicators are a valuable reference for decision-makers and academics in the KSA to understand factors related to SR in mega construction projects. The indicators are related to mega construction projects within the KSA and require validation in a real case scenario or within a different industry to demonstrate their generalisability.

Keywords: Social responsibility, construction projects, economic, social, environmental, indicators.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1823
4245 Supersonic Flow around a Dihedral Airfoil: Modeling and Experimentation Investigation

Authors: A. Naamane, M. Hasnaoui

Abstract:

Numerical modeling of fluid flows, whether compressible or incompressible, laminar or turbulent presents a considerable contribution in the scientific and industrial fields. However, the development of an approximate model of a supersonic flow requires the introduction of specific and more precise techniques and methods. For this purpose, the object of this paper is modeling a supersonic flow of inviscid fluid around a dihedral airfoil. Based on the thin airfoils theory and the non-dimensional stationary Steichen equation of a two-dimensional supersonic flow in isentropic evolution, we obtained a solution for the downstream velocity potential of the oblique shock at the second order of relative thickness that characterizes a perturbation parameter. This result has been dealt with by the asymptotic analysis and characteristics method. In order to validate our model, the results are discussed in comparison with theoretical and experimental results. Indeed, firstly, the comparison of the results of our model has shown that they are quantitatively acceptable compared to the existing theoretical results. Finally, an experimental study was conducted using the AF300 supersonic wind tunnel. In this experiment, we have considered the incident upstream Mach number over a symmetrical dihedral airfoil wing. The comparison of the different Mach number downstream results of our model with those of the existing theoretical data (relative margin between 0.07% and 4%) and with experimental results (concordance for a deflection angle between 1° and 11°) support the validation of our model with accuracy.

Keywords: Asymptotic modelling, dihedral airfoil, supersonic flow, supersonic wind tunnel.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 688
4244 The Influence of Pad Thermal Diffusivity over Heat Transfer into the PCBs Structure

Authors: Mihai Brânzei, Ioan Plotog, Ion Pencea

Abstract:

The Pads have unique values of thermophysical properties (THP) having important contribution over heat transfer into the PCB structure. Materials with high thermal diffusivity (TD) rapidly adjust their temperature to that of their surroundings, because the HT is quick in compare to their volumetric heat capacity (VHC). In the paper is presenting the diffusivity tests (ASTM E1461 flash method) for PCBs with different core materials. In the experiments, the multilayer structure of PCBA was taken into consideration, an equivalent property referring to each of experimental structure be practically measured. Concerning to entire structure, the THP emphasize the major contribution of substrate in establishing of reflow soldering process (RSP) heat transfer necessities. This conclusion offer practical solution for heat transfer time constant calculation as function of thickness and substrate material diffusivity with an acceptable error estimation.

Keywords: heat transfer time constant, packaging, reflowsoldering process, thermal diffusivity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2313
4243 Aircraft Selection Using Preference Optimization Programming (POP)

Authors: C. Ardil

Abstract:

A multiple-criteria decision support system is proposed for the best aircraft selection decision. Various strategic, economic, environmental, and risk-related factors can directly or indirectly influence this choice, and they should be taken into account in the decision-making process. The paper suggests a multiple-criteria analysis to aid in the airline management's decision-making process when choosing an appropriate aircraft. In terms of the suggested approach, an integrated entropic preference optimization programming (POP) for fleet modeling risk analysis is applied. The findings of the study of multiple criteria analysis indicate that the A321(neo) aircraft type is the best alternative in this particular optimization instance. The proposed methodology can be applied to other complex engineering problems involving multiple criteria analysis.

Keywords: Aircraft selection, decision making, multiple criteria decision making, preference optimization programming, POP, entropic weight method, TOPSIS, WSM, WPM

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 577
4242 Optimizing Approach for Sifting Process to Solve a Common Type of Empirical Mode Decomposition Mode Mixing

Authors: Saad Al-Baddai, Karema Al-Subari, Elmar Lang, Bernd Ludwig

Abstract:

Empirical mode decomposition (EMD), a new data-driven of time-series decomposition, has the advantage of supposing that a time series is non-linear or non-stationary, as is implicitly achieved in Fourier decomposition. However, the EMD suffers of mode mixing problem in some cases. The aim of this paper is to present a solution for a common type of signals causing of EMD mode mixing problem, in case a signal suffers of an intermittency. By an artificial example, the solution shows superior performance in terms of cope EMD mode mixing problem comparing with the conventional EMD and Ensemble Empirical Mode decomposition (EEMD). Furthermore, the over-sifting problem is also completely avoided; and computation load is reduced roughly six times compared with EEMD, an ensemble number of 50.

Keywords: Empirical mode decomposition, mode mixing, sifting process, over-sifting.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 966
4241 Faults Forecasting System

Authors: Hanaa E.Sayed, Hossam A. Gabbar, Shigeji Miyazaki

Abstract:

This paper presents Faults Forecasting System (FFS) that utilizes statistical forecasting techniques in analyzing process variables data in order to forecast faults occurrences. FFS is proposing new idea in detecting faults. Current techniques used in faults detection are based on analyzing the current status of the system variables in order to check if the current status is fault or not. FFS is using forecasting techniques to predict future timing for faults before it happens. Proposed model is applying subset modeling strategy and Bayesian approach in order to decrease dimensionality of the process variables and improve faults forecasting accuracy. A practical experiment, designed and implemented in Okayama University, Japan, is implemented, and the comparison shows that our proposed model is showing high forecasting accuracy and BEFORE-TIME.

Keywords: Bayesian Techniques, Faults Detection, Forecasting techniques, Multivariate Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1530
4240 Teachers' Conceptions as a Basis for the Design of an Educational Application: Case Perioperative Nursing

Authors: Antti Pirhonen, Minna Silvennoinen

Abstract:

The only relevant basis for the design of an educational application are objectives of learning for the content area. This study analyses the process in which the real – not only the formal – objectives could work as the starting point for the construction of an educational game. The application context is the education of perioperative nursing. The process is based on the panel discussions of nursing teachers. In the panels, the teachers elaborated the objectives. The transcribed discussions were analysed in terms of the conceptions of learning and teaching of perioperative nursing. The outcome of the study is first the elaborated objectives, which will be used in the implementation of an educational game for the needs of pre-, intra and post-operative nursing skills learning. Second, the study shows that different views of learning are necessary to be understood in order to design an appropriate educational application.

Keywords: Perioperative nursing, conceptions of learning, educational applications.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1585
4239 Collaborative Research between Malaysian and Australian Universities on Learning Analytics: Challenges and Strategies

Authors: Z. Tasir, S. N. Kew, D. West, Z. Abdullah, D. Toohey

Abstract:

Research on Learning Analytics is progressively developing in the higher education field by concentrating on the process of students' learning. Therefore, a research project between Malaysian and Australian Universities was initiated in 2015 to look at the use of Learning Analytics to support the development of teaching practice. The focal point of this article is to discuss and share the experiences of Malaysian and Australian universities in the process of developing the collaborative research on Learning Analytics. Three aspects of this will be discussed: 1) Establishing an international research project and team members, 2) cross-cultural understandings, and 3) ways of working in relation to the practicalities of the project. This article is intended to benefit other researchers by highlighting the challenges as well as the strategies used in this project to ensure such collaborative research succeeds.

Keywords: Academic research project, collaborative research, cross-cultural understanding, international research project.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1464
4238 Managing User Expectations in Information Systems Development

Authors: Linda, Sau-ling Lai

Abstract:

This paper provides new ways to explore the old problem of failure of information systems development in an organisation. Based on the theory of cognitive dissonance, information systems (IS) failure is defined as a gap between what the users expect from an information system and how well these expectations are met by the perceived performance of the delivered system. Bridging the expectation-perception gap requires that IS professionals make a radical change from being the proprietor of information systems and products to being service providers. In order to deliver systems and services that IS users perceive as valuable, IS people must become expert in determining and assessing users- expectations and perceptions. It is also suggested that the IS community, in general, has given relatively little attention to the front-end process of requirements specification for IS development. There is a simplistic belief that requirements are obtainable from users, they are then translatable into a formal specification. The process of information needs analysis is problematic and worthy of investigation.

Keywords: Information Systems Development, Cognitive Dissonance, Expectation-Perception Gap, Requirements Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3627
4237 Construction Technology of Modified Vacuum Pre-Loading Method for Slurry Dredged Soil

Authors: Ali H. Mahfouz, Gao Ming-Jun, Mohamad Sharif

Abstract:

Slurry dredged soil at coastal area has a high water content, poor permeability, and low surface intensity. Hence, it is infeasible to use vacuum preloading method to treat this type of soil foundation. For the special case of super soft ground, a floating bridge is first constructed on muddy soil and used as a service road and platform for implementing the modified vacuum preloading method. The modified technique of vacuum preloading and its construction process for the super soft soil foundation improvement is then studied. Application of modified vacuum preloading method shows that the technology and its construction process are highly suitable for improving the super soft soil foundation in coastal areas.

Keywords: Super soft foundation, dredger fill, vacuum preloading, foundation treatment, construction technology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1893
4236 Application of Recycled Tungsten Carbide Powder for Fabrication of Iron Based Powder Metallurgy Alloy

Authors: Yukinori Taniguchi, Kazuyoshi Kurita, Kohei Mizuta, Keigo Nishitani, Ryuichi Fukuda

Abstract:

Tungsten carbide is widely used as a tool material in metal manufacturing process. Since tungsten is typical rare metal, establishment of recycle process of tungsten carbide tools and restore into cemented carbide material bring great impact to metal manufacturing industry. Recently, recycle process of tungsten carbide has been developed and established gradually. However, the demands for quality of cemented carbide tool are quite severe because hardness, toughness, anti-wear ability, heat resistance, fatigue strength and so on should be guaranteed for precision machining and tool life. Currently, it is hard to restore the recycled tungsten carbide powder entirely as raw material for new processed cemented carbide tool. In this study, to suggest positive use of recycled tungsten carbide powder, we have tried to fabricate a carbon based sintered steel which shows reinforced mechanical properties with recycled tungsten carbide powder. We have made set of newly designed sintered steels. Compression test of sintered specimen in density ratio of 0.85 (which means 15% porosity inside) has been conducted. As results, at least 1.7 times higher in nominal strength in the amount of 7.0 wt.% was shown in recycled WC powder. The strength reached to over 600 MPa for the Fe-WC-Co-Cu sintered alloy. Wear test has been conducted by using ball-on-disk type friction tester using 5 mm diameter ball with normal force of 2 N in the dry conditions. Wear amount after 1,000 m running distance shows that about 1.5 times longer life was shown in designed sintered alloy. Since results of tensile test showed that same tendency in previous testing, it is concluded that designed sintered alloy can be used for several mechanical parts with special strength and anti-wear ability in relatively low cost due to recycled tungsten carbide powder.

Keywords: Tungsten carbide, recycle process, compression test, powder metallurgy, anti-wear ability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1444
4235 Refining Waste Spent Hydroprocessing Catalyst and Their Metal Recovery

Authors: Meena Marafi, Mohan S. Rana

Abstract:

Catalysts play an important role in producing valuable fuel products in petroleum refining; but, due to feedstock’s impurities catalyst gets deactivated with carbon and metal deposition. The disposal of spent catalyst falls under the category of hazardous industrial waste that requires strict agreement with environmental regulations. The spent hydroprocessing catalyst contains Mo, V and Ni at high concentrations that have been found to be economically significant for recovery. Metal recovery process includes deoiling, decoking, grinding, dissolving and treatment with complexing leaching agent such as ethylene diamine tetra acetic acid (EDTA). The process conditions have been optimized as a function of time, temperature and EDTA concentration in presence of ultrasonic agitation. The results indicated that optimum condition established through this approach could recover 97%, 94% and 95% of the extracted Mo, V and Ni, respectively, while 95% EDTA was recovered after acid treatment.

Keywords: Spent catalyst, deactivation, hydrotreating, spent catalyst.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1289
4234 Multi-objective Optimisation of Composite Laminates under Heat and Moisture Effects using a Hybrid Neuro-GA Algorithm

Authors: M. R. Ghasemi, A. Ehsani

Abstract:

In this paper, the optimum weight and cost of a laminated composite plate is seeked, while it undergoes the heaviest load prior to a complete failure. Various failure criteria are defined for such structures in the literature. In this work, the Tsai-Hill theory is used as the failure criterion. The theory of analysis was based on the Classical Lamination Theory (CLT). A newly type of Genetic Algorithm (GA) as an optimization technique with a direct use of real variables was employed. Yet, since the optimization via GAs is a long process, and the major time is consumed through the analysis, Radial Basis Function Neural Networks (RBFNN) was employed in predicting the output from the analysis. Thus, the process of optimization will be carried out through a hybrid neuro-GA environment, and the procedure will be carried out until a predicted optimum solution is achieved.

Keywords: Composite Laminates, GA, Multi-objectiveOptimisation, Neural Networks, RBFNN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1610
4233 An Energy Integration Approach on UHDE Ammonia Process

Authors: Alnouss M. Ahmed, Al-Nuaimi A. Ibrahim

Abstract:

In this paper, the energy performance of a selected UHDE Ammonia plant is optimized by conducting heat integration through waste heat recovery and the synthesis of a heat exchange network (HEN). Minimum hot and cold utility requirements were estimated through IChemE spreadsheet. Supporting simulation was carried out using HYSYS software. The results showed that there is no need for heating utility while the required cold utility was found to be around 268,714 kW. Hence a threshold pinch case was faced. Then, the hot and cold streams were matched appropriately. Also, waste heat recovered resulted with savings in HP and LP steams of approximately 51.0% and 99.6%, respectively. An economic analysis on proposed HEN showed very attractive overall payback period not exceeding 3 years. In general, a net saving approaching 35% was achieved in implementing heat optimization of current studied UHDE Ammonia process.

Keywords: Ammonia, Energy Optimization, Heat Exchange Network and Techno-Economic Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4527
4232 An Automated Method to Segment and Classify Masses in Mammograms

Authors: Viet Dzung Nguyen, Duc Thuan Nguyen, Tien Dzung Nguyen, Van Thanh Pham

Abstract:

Mammography is the most effective procedure for an early diagnosis of the breast cancer. Nowadays, people are trying to find a way or method to support as much as possible to the radiologists in diagnosis process. The most popular way is now being developed is using Computer-Aided Detection (CAD) system to process the digital mammograms and prompt the suspicious region to radiologist. In this paper, an automated CAD system for detection and classification of massive lesions in mammographic images is presented. The system consists of three processing steps: Regions-Of- Interest detection, feature extraction and classification. Our CAD system was evaluated on Mini-MIAS database consisting 322 digitalized mammograms. The CAD system-s performance is evaluated using Receiver Operating Characteristics (ROC) and Freeresponse ROC (FROC) curves. The archived results are 3.47 false positives per image (FPpI) and sensitivity of 85%.

Keywords: classification, computer-aided detection, featureextraction, mass detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1633
4231 A Goal-Driven Crime Scripting Framework

Authors: Hashem Dehghanniri

Abstract:

Crime scripting is a simple and effective crime modeling technique that aims to improve understanding of security analysts about security and crime incidents. Low-quality scripts provide a wrong, incomplete, or sophisticated understanding of the crime commission process, which oppose the purpose of their application, e.g., identifying effective and cost-efficient situational crime prevention (SCP) measures. One important and overlooked factor in generating quality scripts is the crime scripting method. This study investigates the problems within the existing crime scripting practices and proposes a crime scripting approach that contributes to generating quality crime scripts. It was validated by experienced crime scripters. This framework helps analysts develop better crime scripts and contributes to their effective application, e.g., SCP measures identification or policy-making.

Keywords: Attack modeling, crime commission process, crime script, situational crime prevention.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 652
4230 Degradation of Amitriptyline Hydrochloride, Methyl Salicylate and 2-Phenoxyethanol in Water Systems by the Combination UV/Cl2

Authors: F. Javier Benitez, Francisco J. Real, Juan Luis Acero, Francisco Casas

Abstract:

Three emerging contaminants (amitriptyline hydrochloride, methyl salicylate and 2-phenoxyethanol) frequently found in waste-waters were selected to be individually degraded in ultra-pure water by the combined advanced oxidation process constituted by UV radiation and chlorine. The influence of pH, initial chlorine concentration and nature of the contaminants was firstly explored. The trend for the reactivity of the selected compounds was deduced: amitriptyline hydrochloride > methyl salicylate > 2-phenoxyethanol. A later kinetic study was carried out and focused on the specific evaluation of the first-order rate constants and the determination of the partial contribution to the global reaction of the direct photochemical pathway and the radical pathway. A comparison between the rate constant values among photochemical experiments without and with the presence of Cl2 reveals a clear increase in the oxidation efficiency of the combined process with respect to the photochemical reaction alone. In a second stage, the simultaneous oxidation of mixtures of the selected contaminants in several types of water (ultrapure water, surface water from a reservoir, and two secondary effluents) was also performed by the same combination UV/Cl2 under more realistic operating conditions. The efficiency of this combined system UV/Cl2 was compared to other oxidants such as the UV/S2O82- and UV/H2O2 AOPs. Results confirmed that the UV/Cl2 system provides higher elimination efficiencies among the AOPs tested.

Keywords: Emerging contaminants, amitriptyline, methyl salicylate, 2-phenoxyethanol, chlorination, photolysis, rate constants, UV/chlorine advanced oxidation process.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1519
4229 Land Surface Temperature and Biophysical Factors in Urban Planning

Authors: Illyani Ibrahim, Azizan Abu Samah, Rosmadi Fauzi

Abstract:

Land surface temperature (LST) is an important parameter to study in urban climate. The understanding of the influence of biophysical factors could improve the establishment of modeling urban thermal landscape. It is well established that climate hold a great influence on the urban landscape. However, it has been recognize that climate has a low priority in urban planning process, due to the complex nature of its influence. This study will focus on the relatively cloud free Landsat Thematic Mapper image of the study area, acquired on the 2nd March 2006. Correlation analyses were conducted to identify the relationship of LST to the biophysical factors; vegetation indices, impervious surface, and albedo to investigate the variation of LST. We suggest that the results can be considered by the stackholders during decision-making process to create a cooler and comfortable environment in the urban landscape for city dwellers.

Keywords: Biophysical factors, land surface temperature, urban planning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2066
4228 Statistical Process Optimization Through Multi-Response Surface Methodology

Authors: S. Raissi, R- Eslami Farsani

Abstract:

In recent years, response surface methodology (RSM) has brought many attentions of many quality engineers in different industries. Most of the published literature on robust design methodology is basically concerned with optimization of a single response or quality characteristic which is often most critical to consumers. For most products, however, quality is multidimensional, so it is common to observe multiple responses in an experimental situation. Through this paper interested person will be familiarize with this methodology via surveying of the most cited technical papers. It is believed that the proposed procedure in this study can resolve a complex parameter design problem with more than two responses. It can be applied to those areas where there are large data sets and a number of responses are to be optimized simultaneously. In addition, the proposed procedure is relatively simple and can be implemented easily by using ready-made standard statistical packages.

Keywords: Multi-Response Surface Methodology (MRSM), Design of Experiments (DOE), Process modeling, Quality improvement; Robust Design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4431
4227 Comparing Machine Learning Estimation of Fuel Consumption of Heavy-Duty Vehicles

Authors: Victor Bodell, Lukas Ekstrom, Somayeh Aghanavesi

Abstract:

Fuel consumption (FC) is one of the key factors in determining expenses of operating a heavy-duty vehicle. A customer may therefore request an estimate of the FC of a desired vehicle. The modular design of heavy-duty vehicles allows their construction by specifying the building blocks, such as gear box, engine and chassis type. If the combination of building blocks is unprecedented, it is unfeasible to measure the FC, since this would first r equire the construction of the vehicle. This paper proposes a machine learning approach to predict FC. This study uses around 40,000 vehicles specific and o perational e nvironmental c onditions i nformation, such as road slopes and driver profiles. A ll v ehicles h ave d iesel engines and a mileage of more than 20,000 km. The data is used to investigate the accuracy of machine learning algorithms Linear regression (LR), K-nearest neighbor (KNN) and Artificial n eural n etworks (ANN) in predicting fuel consumption for heavy-duty vehicles. Performance of the algorithms is evaluated by reporting the prediction error on both simulated data and operational measurements. The performance of the algorithms is compared using nested cross-validation and statistical hypothesis testing. The statistical evaluation procedure finds that ANNs have the lowest prediction error compared to LR and KNN in estimating fuel consumption on both simulated and operational data. The models have a mean relative prediction error of 0.3% on simulated data, and 4.2% on operational data.

Keywords: Artificial neural networks, fuel consumption, machine learning, regression, statistical tests.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 792
4226 The Fake News Impact on the Public Policy Cycle: A Systemic Analysis through Documentary Survey

Authors: Aron Miranda Burgos, Ergon Cugler de Moraes Silva

Abstract:

In the present article, it is observed that the constant advancement of issues related to misinformation impacts the guarantee of the public policy cycle. Thus, it is found that the dissemination of false information has a direct influence on each of the component stages of this cycle. Therefore, in order to maintain scientific and theoretical credibility in the qualitative analysis process, it was necessary to logically interpose the concepts of firehosing of falsehood, fake news, public policy cycle, as well as using the epistemological and pragmatic mechanism at the intersection of such academic concepts, such as the scientific method. It was found, through the analysis of official documents and public notes, how the multiple theoretical perspectives evidence the commitment of the provision and elaboration of public policies, verifying the way in which the fake news impact each part of the process in this atmosphere.

Keywords: Firehosing of falsehood, governance, misinformation, post-truth.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 831