Search results for: reduce order aeroelastic model (ROAM)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 30462

Search results for: reduce order aeroelastic model (ROAM)

29832 Temperature Distribution in Friction Stir Welding Using Finite Element Method

Authors: Armansyah, I. P. Almanar, M. Saiful Bahari Shaari, M. Shamil Jaffarullah, Nur’amirah Busu, M. Arif Fadzleen Zainal Abidin, M. Amlie A. Kasim

Abstract:

Temperature distribution in Friction Stir Welding (FSW) of 6061-T6 Aluminum Alloy is modeled using the Finite Element Method (FEM). In order to obtain temperature distribution in the welded aluminum plates during welding operation, transient thermal finite element analyses are performed. Heat input from tool shoulder and tool pin are considered in the model. A moving heat source with a heat distribution simulating the heat generated by frictions between tool shoulder and workpiece is used in the analysis. Three-dimensional model for simulated process is carried out by using Altair HyperWork, a commercially available software. Transient thermal finite element analyses are performed in order to obtain the temperature distribution in the welded Aluminum plates during welding operation. The developed model was then used to show the effect of various input parameters such as total rate of welding speed and rotational speed on temperature distribution in the workpiece.

Keywords: frictions stir welding, temperature distribution, finite element method, altair hyperwork

Procedia PDF Downloads 528
29831 Improving Automotive Efficiency through Lean Management Tools: A Case Study

Authors: Raed El-Khalil, Hussein Zeaiter

Abstract:

Managing and improving efficiency in the current highly competitive global automotive industry demands that companies adopt leaner and more flexible systems. During the past 20 years the domestic automotive industry in North America has been focusing on establishing new management strategies in order to meet market demands. 98The lean management process also known as Toyota Manufacturing Process (TPS) or lean manufacturing encompasses tools and techniques that were established in order to provide the best quality product with the fastest lead time at the lowest cost. The following paper presents a study that focused on improving labor efficiency at one of the Big Three (Ford, GM, Chrysler LLC) domestic automotive facility in North America. The objective of the study was to utilize several lean management tools in order to optimize the efficiency and utilization levels at the “Pre-Marriage” chassis area in a truck manufacturing and assembly facility. Utilizing three different lean tools (i.e. Standardization of work, 7 Wastes, and 5S) this research was able to improve efficiency by 51%, utilization by 246%, and reduce operations by 14%. The return on investment calculated based on the improvements made was 284%.

Keywords: lean manufacturing, standardized work, operation efficiency, utilization

Procedia PDF Downloads 499
29830 Disintegration of Deuterons by Photons Reaction Model for GEANT4 with Dibaryon Formalism

Authors: Jae Won Shin, Chang Ho Hyun

Abstract:

A disintegration of deuterons by photons (dγ → np) reaction model for GEANT4 is developed in this work. An effective field theory with dibaryon fields Introducing a dibaryon field, we can take into account the effective range contribution to the propagator up to infinite order, and it consequently makes the convergence of the theory better than the pionless effective field theory without dibaryon fields. We develop a hadronic model for GEANT4 which is specialized for the disintegration of the deuteron by photons, dγ → np. For the description of two-nucleon interactions, we employ an effective field theory so called pionless theory with dibaryon fields (dEFT). In spite of its simplicity, the theory has proven very effective and useful in the applications to various two-nucleon systems and processes at low energies. We apply the new model of GEANT4 (G4dEFT) to the calculation of total and differential cross sections in dγ → np, and obtain good agreements to experimental data for a wide range of incoming photon energies.

Keywords: dγ → np, dibaryon fields, effective field theory, GEANT4

Procedia PDF Downloads 362
29829 Mitigation of High Voltage Equipment Design Deficiencies for Improved Operation and Maintenance

Authors: Riyad Awad, Abdulmohsen Alghadeer, Meshari Otaibi

Abstract:

Proper operation and maintenance (O&M) activities of high voltage equipment can lead to an increased asset lifecycle and maintain its integrity and reliability. Such a vital process is important to be proactively considered during equipment design and manufacturing phases by removing and eliminating any obstacles in the equipment which adversely affect the (O&M) activities. This paper presents a gap analysis pertaining to difficulties in performing operations and maintenance (O&M) high voltage electrical equipment, includes power transformers, switch gears, motor control center, disconnect switches and circuit breakers. The difficulties are gathered from field personnel, equipment design review comments, quality management system, and lessons learned database. The purpose of the gap analysis is to mitigate and prevent the (O&M) difficulties as early as possible in the design stage of the equipment lifecycle. The paper concludes with several recommendations and corrective actions for all identified gaps in order to reduce the cost (O&M) difficulties and improve the equipment lifecycle.

Keywords: operation and maintenance, high voltage equipment, equipment lifecycle, reduce the cost of maintenance

Procedia PDF Downloads 151
29828 Investigating the Dynamics of Knowledge Acquisition in Learning Using Differential Equations

Authors: Gilbert Makanda, Roelf Sypkens

Abstract:

A mathematical model for knowledge acquisition in teaching and learning is proposed. In this study we adopt the mathematical model that is normally used for disease modelling into teaching and learning. We derive mathematical conditions which facilitate knowledge acquisition. This study compares the effects of dropping out of the course at early stages with later stages of learning. The study also investigates effect of individual interaction and learning from other sources to facilitate learning. The study fits actual data to a general mathematical model using Matlab ODE45 and lsqnonlin to obtain a unique mathematical model that can be used to predict knowledge acquisition. The data used in this study was obtained from the tutorial test results for mathematics 2 students from the Central University of Technology, Free State, South Africa in the department of Mathematical and Physical Sciences. The study confirms already known results that increasing dropout rates and forgetting taught concepts reduce the population of knowledgeable students. Increasing teaching contacts and access to other learning materials facilitate knowledge acquisition. The effect of increasing dropout rates is more enhanced in the later stages of learning than earlier stages. The study opens up a new direction in further investigations in teaching and learning using differential equations.

Keywords: differential equations, knowledge acquisition, least squares nonlinear, dynamical systems

Procedia PDF Downloads 355
29827 Numerical Study on the Performance of Upgraded Victorian Brown Coal in an Ironmaking Blast Furnace

Authors: Junhai Liao, Yansong Shen, Aibing Yu

Abstract:

A 3D numerical model is developed to simulate the complicated in-furnace combustion phenomena in the lower part of an ironmaking blast furnace (BF) while using pulverized coal injection (PCI) technology to reduce the consumption of relatively expensive coke. The computational domain covers blowpipe-tuyere-raceway-coke bed in the BF. The model is validated against experimental data in terms of gaseous compositions and coal burnout. Parameters, such as coal properties and some key operational variables, play an important role on the performance of coal combustion. Their diverse effects on different combustion characteristics are examined in the domain, in terms of gas compositions, temperature, and burnout. The heat generated by the combustion of upgraded Victorian brown coal is able to meet the heating requirement of a BF, hence making upgraded brown coal injected into BF possible. It is evidenced that the model is suitable to investigate the mechanism of the PCI operation in a BF. Prediction results provide scientific insights to optimize and control of the PCI operation. This model cuts the cost to investigate and understand the comprehensive combustion phenomena of upgraded Victorian brown coal in a full-scale BF.

Keywords: blast furnace, numerical study, pulverized coal injection, Victorian brown coal

Procedia PDF Downloads 232
29826 Automated Adaptions of Semantic User- and Service Profile Representations by Learning the User Context

Authors: Nicole Merkle, Stefan Zander

Abstract:

Ambient Assisted Living (AAL) describes a technological and methodological stack of (e.g. formal model-theoretic semantics, rule-based reasoning and machine learning), different aspects regarding the behavior, activities and characteristics of humans. Hence, a semantic representation of the user environment and its relevant elements are required in order to allow assistive agents to recognize situations and deduce appropriate actions. Furthermore, the user and his/her characteristics (e.g. physical, cognitive, preferences) need to be represented with a high degree of expressiveness in order to allow software agents a precise evaluation of the users’ context models. The correct interpretation of these context models highly depends on temporal, spatial circumstances as well as individual user preferences. In most AAL approaches, model representations of real world situations represent the current state of a universe of discourse at a given point in time by neglecting transitions between a set of states. However, the AAL domain currently lacks sufficient approaches that contemplate on the dynamic adaptions of context-related representations. Semantic representations of relevant real-world excerpts (e.g. user activities) help cognitive, rule-based agents to reason and make decisions in order to help users in appropriate tasks and situations. Furthermore, rules and reasoning on semantic models are not sufficient for handling uncertainty and fuzzy situations. A certain situation can require different (re-)actions in order to achieve the best results with respect to the user and his/her needs. But what is the best result? To answer this question, we need to consider that every smart agent requires to achieve an objective, but this objective is mostly defined by domain experts who can also fail in their estimation of what is desired by the user and what not. Hence, a smart agent has to be able to learn from context history data and estimate or predict what is most likely in certain contexts. Furthermore, different agents with contrary objectives can cause collisions as their actions influence the user’s context and constituting conditions in unintended or uncontrolled ways. We present an approach for dynamically updating a semantic model with respect to the current user context that allows flexibility of the software agents and enhances their conformance in order to improve the user experience. The presented approach adapts rules by learning sensor evidence and user actions using probabilistic reasoning approaches, based on given expert knowledge. The semantic domain model consists basically of device-, service- and user profile representations. In this paper, we present how this semantic domain model can be used in order to compute the probability of matching rules and actions. We apply this probability estimation to compare the current domain model representation with the computed one in order to adapt the formal semantic representation. Our approach aims at minimizing the likelihood of unintended interferences in order to eliminate conflicts and unpredictable side-effects by updating pre-defined expert knowledge according to the most probable context representation. This enables agents to adapt to dynamic changes in the environment which enhances the provision of adequate assistance and affects positively the user satisfaction.

Keywords: ambient intelligence, machine learning, semantic web, software agents

Procedia PDF Downloads 268
29825 Meta Model for Optimum Design Objective Function of Steel Frames Subjected to Seismic Loads

Authors: Salah R. Al Zaidee, Ali S. Mahdi

Abstract:

Except for simple problems of statically determinate structures, optimum design problems in structural engineering have implicit objective functions where structural analysis and design are essential within each searching loop. With these implicit functions, the structural engineer is usually enforced to write his/her own computer code for analysis, design, and searching for optimum design among many feasible candidates and cannot take advantage of available software for structural analysis, design, and searching for the optimum solution. The meta-model is a regression model used to transform an implicit objective function into objective one and leads in turn to decouple the structural analysis and design processes from the optimum searching process. With the meta-model, well-known software for structural analysis and design can be used in sequence with optimum searching software. In this paper, the meta-model has been used to develop an explicit objective function for plane steel frames subjected to dead, live, and seismic forces. Frame topology is assumed as predefined based on architectural and functional requirements. Columns and beams sections and different connections details are the main design variables in this study. Columns and beams are grouped to reduce the number of design variables and to make the problem similar to that adopted in engineering practice. Data for the implicit objective function have been generated based on analysis and assessment for many design proposals with CSI SAP software. These data have been used later in SPSS software to develop a pure quadratic nonlinear regression model for the explicit objective function. Good correlations with a coefficient, R2, in the range from 0.88 to 0.99 have been noted between the original implicit functions and the corresponding explicit functions generated with meta-model.

Keywords: meta-modal, objective function, steel frames, seismic analysis, design

Procedia PDF Downloads 230
29824 Enhancement Dynamic Cars Detection Based on Optimized HOG Descriptor

Authors: Mansouri Nabila, Ben Jemaa Yousra, Motamed Cina, Watelain Eric

Abstract:

Research and development efforts in intelligent Advanced Driver Assistance Systems (ADAS) seek to save lives and reduce the number of on-road fatalities. For traffic and emergency monitoring, the essential but challenging task is vehicle detection and tracking in reasonably short time. This purpose needs first of all a powerful dynamic car detector model. In fact, this paper presents an optimized HOG process based on shape and motion parameters fusion. Our proposed approach mains to compute HOG by bloc feature from foreground blobs using configurable research window and pathway in order to overcome the shortcoming in term of computing time of HOG descriptor and improve their dynamic application performance. Indeed we prove in this paper that HOG by bloc descriptor combined with motion parameters is a very suitable car detector which reaches in record time a satisfactory recognition rate in dynamic outside area and bypasses several popular works without using sophisticated and expensive architectures such as GPU and FPGA.

Keywords: car-detector, HOG, motion, computing time

Procedia PDF Downloads 315
29823 How to Perform Proper Indexing?

Authors: Watheq Mansour, Waleed Bin Owais, Mohammad Basheer Kotit, Khaled Khan

Abstract:

Efficient query processing is one of the utmost requisites in any business environment to satisfy consumer needs. This paper investigates the various types of indexing models, viz. primary, secondary, and multi-level. The investigation is done under the ambit of various types of queries to which each indexing model performs with efficacy. This study also discusses the inherent advantages and disadvantages of each indexing model and how indexing models can be chosen based on a particular environment. This paper also draws parallels between various indexing models and provides recommendations that would help a Database administrator to zero-in on a particular indexing model attributed to the needs and requirements of the production environment. In addition, to satisfy industry and consumer needs attributed to the colossal data generation nowadays, this study has proposed two novel indexing techniques that can be used to index highly unstructured and structured Big Data with efficacy. The study also briefly discusses some best practices that the industry should follow in order to choose an indexing model that is apposite to their prerequisites and requirements.

Keywords: indexing, hashing, latent semantic indexing, B-tree

Procedia PDF Downloads 148
29822 Mathematical Model That Using Scrambling and Message Integrity Methods in Audio Steganography

Authors: Mohammed Salem Atoum

Abstract:

The success of audio steganography is to ensure imperceptibility of the embedded message in stego file and withstand any form of intentional or un-intentional degradation of message (robustness). Audio steganographic that utilized LSB of audio stream to embed message gain a lot of popularity over the years in meeting the perceptual transparency, robustness and capacity. This research proposes an XLSB technique in order to circumvent the weakness observed in LSB technique. Scrambling technique is introduce in two steps; partitioning the message into blocks followed by permutation each blocks in order to confuse the contents of the message. The message is embedded in the MP3 audio sample. After extracting the message, the permutation codebook is used to re-order it into its original form. Md5sum and SHA-256 are used to verify whether the message is altered or not during transmission. Experimental result shows that the XLSB performs better than LSB.

Keywords: XLSB, scrambling, audio steganography, security

Procedia PDF Downloads 358
29821 Document Analysis for Modelling iTV Advertising towards Impulse Purchase

Authors: Azizah Che Omar

Abstract:

The study provides a systematic literature review which analyzed the literature for the purpose of looking for concepts, theories, approaches and guidelines in order to propose a conceptual design model of interactive television advertising toward impulse purchase (iTVAdIP). An extensive review of literature was purposely carried out to understand the concepts of interactive television (iTV). Therefore, some elements; iTV guidelines, advertising theories, persuasive approaches, and the impulse purchase elements were analyzed to reach the scope of this work. The extensive review was also a necessity to achieve the objective of this study, which was to determine the concept of iTVAdIP design model. Through systematic review analysis, this study discovered that all the previous models did not emphasize the conceptual design model of interactive television advertising. As a result, the finding showed that the concept of the proposed model should contain the iTV guidelines, advertising theory, persuasive approach and impulse purchase elements. In addition, a summary diagram for the development of the proposed model is depicted to provide clearer understanding towards the concepts of conceptual design model of iTVAdIP.

Keywords: impulse purchase, interactive television advertising, human computer interaction, advertising theories

Procedia PDF Downloads 359
29820 A Tuning Method for Microwave Filter via Complex Neural Network and Improved Space Mapping

Authors: Shengbiao Wu, Weihua Cao, Min Wu, Can Liu

Abstract:

This paper presents an intelligent tuning method of microwave filter based on complex neural network and improved space mapping. The tuning process consists of two stages: the initial tuning and the fine tuning. At the beginning of the tuning, the return loss of the filter is transferred to the passband via the error of phase. During the fine tuning, the phase shift caused by the transmission line and the higher order mode is removed by the curve fitting. Then, an Cauchy method based on the admittance parameter (Y-parameter) is used to extract the coupling matrix. The influence of the resonant cavity loss is eliminated during the parameter extraction process. By using processed data pairs (the amount of screw variation and the variation of the coupling matrix), a tuning model is established by the complex neural network. In view of the improved space mapping algorithm, the mapping relationship between the actual model and the ideal model is established, and the amplitude and direction of the tuning is constantly updated. Finally, the tuning experiment of the eight order coaxial cavity filter shows that the proposed method has a good effect in tuning time and tuning precision.

Keywords: microwave filter, scattering parameter, coupling matrix, intelligent tuning

Procedia PDF Downloads 290
29819 Application of Principal Component Analysis and Ordered Logit Model in Diabetic Kidney Disease Progression in People with Type 2 Diabetes

Authors: Mequanent Wale Mekonen, Edoardo Otranto, Angela Alibrandi

Abstract:

Diabetic kidney disease is one of the main microvascular complications caused by diabetes. Several clinical and biochemical variables are reported to be associated with diabetic kidney disease in people with type 2 diabetes. However, their interrelations could distort the effect estimation of these variables for the disease's progression. The objective of the study is to determine how the biochemical and clinical variables in people with type 2 diabetes are interrelated with each other and their effects on kidney disease progression through advanced statistical methods. First, principal component analysis was used to explore how the biochemical and clinical variables intercorrelate with each other, which helped us reduce a set of correlated biochemical variables to a smaller number of uncorrelated variables. Then, ordered logit regression models (cumulative, stage, and adjacent) were employed to assess the effect of biochemical and clinical variables on the order-level response variable (progression of kidney function) by considering the proportionality assumption for more robust effect estimation. This retrospective cross-sectional study retrieved data from a type 2 diabetic cohort in a polyclinic hospital at the University of Messina, Italy. The principal component analysis yielded three uncorrelated components. These are principal component 1, with negative loading of glycosylated haemoglobin, glycemia, and creatinine; principal component 2, with negative loading of total cholesterol and low-density lipoprotein; and principal component 3, with negative loading of high-density lipoprotein and a positive load of triglycerides. The ordered logit models (cumulative, stage, and adjacent) showed that the first component (glycosylated haemoglobin, glycemia, and creatinine) had a significant effect on the progression of kidney disease. For instance, the cumulative odds model indicated that the first principal component (linear combination of glycosylated haemoglobin, glycemia, and creatinine) had a strong and significant effect on the progression of kidney disease, with an effect or odds ratio of 0.423 (P value = 0.000). However, this effect was inconsistent across levels of kidney disease because the first principal component did not meet the proportionality assumption. To address the proportionality problem and provide robust effect estimates, alternative ordered logit models, such as the partial cumulative odds model, the partial adjacent category model, and the partial continuation ratio model, were used. These models suggested that clinical variables such as age, sex, body mass index, medication (metformin), and biochemical variables such as glycosylated haemoglobin, glycemia, and creatinine have a significant effect on the progression of kidney disease.

Keywords: diabetic kidney disease, ordered logit model, principal component analysis, type 2 diabetes

Procedia PDF Downloads 23
29818 Improving the Run Times of Existing and Historical Demand Models Using Simple Python Scripting

Authors: Abhijeet Ostawal, Parmjit Lall

Abstract:

The run times for a large strategic model that we were managing had become too long leading to delays in project delivery, increased costs and loss in productivity. Software developers are continuously working towards developing more efficient tools by changing their algorithms and processes. The issue faced by our team was how do you apply the latest technologies on validated existing models which are based on much older versions of software that do not have the latest software capabilities. The multi-model transport model that we had could only be run in sequential assignment order. Recent upgrades to the software now allowed the assignment to be run in parallel, a concept called parallelization. Parallelization is a Python script working only within the latest version of the software. A full model transfer to the latest version was not possible due to time, budget and the potential changes in trip assignment. This article is to show the method to adapt and update the Python script in such a way that it can be used in older software versions by calling the latest version and then recalling the old version for assignment model without affecting the results. Through a process of trial-and-error run time savings of up to 30-40% have been achieved. Assignment results were maintained within the older version and through this learning process we’ve applied this methodology to other even older versions of the software resulting in huge time savings, more productivity and efficiency for both client and consultant.

Keywords: model run time, demand model, parallelisation, python scripting

Procedia PDF Downloads 104
29817 Impact of VARK Learning Model at Tertiary Level Education

Authors: Munazza A. Mirza, Khawar Khurshid

Abstract:

Individuals are generally associated with different learning styles, which have been explored extensively in recent past. The learning styles refer to the potential of an individual by which s/he can easily comprehend and retain information. Among various learning style models, VARK is the most accepted model which categorizes the learners with respect to their sensory characteristics. Based on the number of preferred learning modes, the learners can be categorized as uni-modal, bi-modal, tri-modal, or quad/multi-modal. Although there is a prevalent belief in the learning styles, however, the model is not being frequently and effectively utilized in the higher education. This research describes the identification model to validate teacher’s didactic practice and student’s performance linkage with the learning styles. The identification model is recommended to check the effective application and evaluation of the various learning styles. The proposed model is a guideline to effectively implement learning styles inventory in order to ensure that it will validate performance linkage with learning styles. If performance is linked with learning styles, this may help eradicate the distrust on learning style theory. For this purpose, a comprehensive study was conducted to compare and understand how VARK inventory model is being used to identify learning preferences and their correlation with learner’s performance. A comparative analysis of the findings of these studies is presented to understand the learning styles of tertiary students in various disciplines. It is concluded with confidence that the learning styles of students cannot be associated with any specific discipline. Furthermore, there is not enough empirical proof to link performance with learning styles.

Keywords: learning style, VARK, sensory preferences, identification model, didactic practices

Procedia PDF Downloads 263
29816 High-Fidelity 1D Dynamic Model of a Hydraulic Servo Valve Using 3D Computational Fluid Dynamics and Electromagnetic Finite Element Analysis

Authors: D. Henninger, A. Zopey, T. Ihde, C. Mehring

Abstract:

The dynamic performance of a 4-way solenoid operated hydraulic spool valve has been analyzed by means of a one-dimensional modeling approach capturing flow, magnetic and fluid forces, valve inertia forces, fluid compressibility, and damping. Increased model accuracy was achieved by analyzing the detailed three-dimensional electromagnetic behavior of the solenoids and flow behavior through the spool valve body for a set of relevant operating conditions, thereby allowing the accurate mapping of flow and magnetic forces on the moving valve body, in lieu of representing the respective forces by lower-order models or by means of simplistic textbook correlations. The resulting high-fidelity one-dimensional model provided the basis for specific and timely design modification eliminating experimentally observed valve oscillations.

Keywords: dynamic performance model, high-fidelity model, 1D-3D decoupled analysis, solenoid-operated hydraulic servo valve, CFD and electromagnetic FEA

Procedia PDF Downloads 162
29815 Design Optimization of a Compact Quadrupole Electromagnet for CLS 2.0

Authors: Md. Armin Islam, Les Dallin, Mark Boland, W. J. Zhang

Abstract:

This paper reports a study on the optimal magnetic design of a compact quadrupole electromagnet for the Canadian Light Source (CLS 2.0). The nature of the design is to determine a quadrupole with low relative higher order harmonics and better field quality. The design problem was formulated as an optimization model, in which the objective function is the higher order harmonics (multipole errors) and the variable to be optimized is the material distribution on the pole. The higher order harmonics arose in the quadrupole due to truncating the ideal hyperbola at a certain point to make the pole. In this project, the arisen harmonics have been optimized both transversely and longitudinally by adjusting material on the poles in a controlled way. For optimization, finite element analysis (FEA) has been conducted. A better higher order harmonics amplitudes and field quality have been achieved through the optimization. On the basis of the optimized magnetic design, electrical and cooling calculation has been performed for the magnet.

Keywords: drift, electrical, and cooling calculation, integrated field, magnetic field gradient, multipole errors, quadrupole

Procedia PDF Downloads 135
29814 Synthesis and Characterization of Thiourea-Formaldehyde Coated Fe3O4 (TUF@Fe3O4) and Its Application for Adsorption of Methylene Blue

Authors: Saad M. Alshehri, Tansir Ahamad

Abstract:

Thiourea-Formaldehyde Pre-Polymer (TUF) was prepared by the reaction thiourea and formaldehyde in basic medium and used as a coating materials for magnetite Fe3O4. The synthesized polymer coated microspheres (TUF@Fe3O4) was characterized using FTIR, TGA SEM and TEM. Its BET surface area was up to 1680 m2 g_1. The adsorption capacity of this ACF product was evaluated in its adsorption of Methylene Blue (MB) in water under different pH values and different temperature. We found that the adsorption process was well described both by the Langmuir and Freundlich isotherm model. The kinetic processes of MB adsorption onto TUF@Fe3O4 were described in order to provide a more clear interpretation of the adsorption rate and uptake mechanism. The overall kinetic data was acceptably explained by a pseudo second-order rate model. Evaluated ∆Go and ∆Ho specify the spontaneous and exothermic nature of the reaction. The adsorption takes place with a decrease in entropy (∆So is negative). The monolayer capacity for MB was up to 450 mg g_1 and was one of the highest among similar polymeric products. It was due to its large BET surface area.

Keywords: TGA, FTIR, magentite, thiourea formaldehyde resin, methylene blue, adsorption

Procedia PDF Downloads 326
29813 A Study on the Performance of 2-PC-D Classification Model

Authors: Nurul Aini Abdul Wahab, Nor Syamim Halidin, Sayidatina Aisah Masnan, Nur Izzati Romli

Abstract:

There are many applications of principle component method for reducing the large set of variables in various fields. Fisher’s Discriminant function is also a popular tool for classification. In this research, the researcher focuses on studying the performance of Principle Component-Fisher’s Discriminant function in helping to classify rice kernels to their defined classes. The data were collected on the smells or odour of the rice kernel using odour-detection sensor, Cyranose. 32 variables were captured by this electronic nose (e-nose). The objective of this research is to measure how well a combination model, between principle component and linear discriminant, to be as a classification model. Principle component method was used to reduce all 32 variables to a smaller and manageable set of components. Then, the reduced components were used to develop the Fisher’s Discriminant function. In this research, there are 4 defined classes of rice kernel which are Aromatic, Brown, Ordinary and Others. Based on the output from principle component method, the 32 variables were reduced to only 2 components. Based on the output of classification table from the discriminant analysis, 40.76% from the total observations were correctly classified into their classes by the PC-Discriminant function. Indirectly, it gives an idea that the classification model developed has committed to more than 50% of misclassifying the observations. As a conclusion, the Fisher’s Discriminant function that was built on a 2-component from PCA (2-PC-D) is not satisfying to classify the rice kernels into its defined classes.

Keywords: classification model, discriminant function, principle component analysis, variable reduction

Procedia PDF Downloads 325
29812 A Summary-Based Text Classification Model for Graph Attention Networks

Authors: Shuo Liu

Abstract:

In Chinese text classification tasks, redundant words and phrases can interfere with the formation of extracted and analyzed text information, leading to a decrease in the accuracy of the classification model. To reduce irrelevant elements, extract and utilize text content information more efficiently and improve the accuracy of text classification models. In this paper, the text in the corpus is first extracted using the TextRank algorithm for abstraction, the words in the abstract are used as nodes to construct a text graph, and then the graph attention network (GAT) is used to complete the task of classifying the text. Testing on a Chinese dataset from the network, the classification accuracy was improved over the direct method of generating graph structures using text.

Keywords: Chinese natural language processing, text classification, abstract extraction, graph attention network

Procedia PDF Downloads 86
29811 Applying Lean Six Sigma in an Emergency Department, of a Private Hospital

Authors: Sarah Al-Lumai, Fatima Al-Attar, Nour Jamal, Badria Al-Dabbous, Manal Abdulla

Abstract:

Today, many commonly used Industrial Engineering tools and techniques are being used in hospitals around the world for the goal of producing a more efficient and effective healthcare system. A common quality improvement methodology known as Lean Six-Sigma has been successful in manufacturing industries and recently in healthcare. The objective of our project is to use the Lean Six-Sigma methodology to reduce waiting time in the Emergency Department (ED), in a local private hospital. Furthermore, a comprehensive literature review was conducted to evaluate the success of Lean Six-Sigma in the ED. According to the study conducted by Ibn Sina Hospital, in Morocco, the most common problem that patients complain about is waiting time. To ensure patient satisfaction many hospitals such as North Shore University Hospital were able to reduce waiting time up to 37% by using Lean Six-Sigma. Other hospitals, such as John Hopkins’s medical center used Lean Six-Sigma successfully to enhance the overall patient flow that ultimately decreased waiting time. Furthermore, it was found that capacity constraints, such as staff shortages and lack of beds were one of the main reasons behind long waiting time. With the use of Lean Six-Sigma and bed management, hospitals like Memorial Hermann Southwest Hospital were able to reduce patient delays. Moreover, in order to successfully implement Lean Six-Sigma in our project, two common methodologies were considered, DMAIC and DMADV. After the assessment of both methodologies, it was found that DMAIC was a more suitable approach to our project because it is more concerned with improving an already existing process. With many of its successes, Lean Six-Sigma has its limitation especially in healthcare; but limitations can be minimized if properly approached.

Keywords: lean six sigma, DMAIC, hospital, methodology

Procedia PDF Downloads 483
29810 A Mechanism of Reusable, Portable, and Reliable Script Generator on Android

Authors: Kuei-Chun Liu, Yu-Yu Lai, Ching-Hong Wu

Abstract:

A good automated testing tool could reduce as much as possible the manual work done by testers. Traditional record-replay testing tool provides an automated testing solution by recording mouse coordinates as test scripts, but it will be easily broken if any change of resolutions. Therefore, more and more testers design multiple test scripts to automate the testing process for different devices. In order to improve the traditional record-replay approach and reduce the effort that the testers spending on writing test scripts, we propose an approach for generating the Android application test scripts based on accessibility service without connecting to a computer. This approach simulates user input actions and replays them correctly even at the different conditions such as the internet connection is unstable when the device under test, the different resolutions on Android devices. In this paper, we describe how to generate test scripts automatically and make a comparison with existing tools for Android such as Robotium, Appium, UIAutomator, and MonkeyTalk.

Keywords: accessibility service, Appium, automated testing, MonkeyTalk, Robotium, testing, UIAutomator

Procedia PDF Downloads 364
29809 Second Order MIMO Sliding Mode Controller for Nonlinear Modeled Wind Turbine

Authors: Alireza Toloei, Ahmad R. Saffary, Reza Ghasemi

Abstract:

Due to the growing need for energy and limited fossil resources, the use of renewable energy, particularly wind is strongly favored. We all wind energy can’t be saved. Betz law, 59% of the total kinetic energy of the wind turbine is extracting. Therefore turbine control to achieve maximum performance and maintain stable conditions seem necessary. In this article, we plan for a horizontal axis wind turbine variable-speed variable-pitch nonlinear controller to obtain maximum output power. The model presented in this article, including a wide range of wind turbines are horizontal axis. However, the parameters used in this model is from Vestas V29 225 kW wind turbine. We designed second order sliding mode controller, which was robust in the face of changes in wind speed and it eliminated chattering by using of super twisting algorithm. Finally, using MATLAB software to simulate the results we considered the accuracy of the simulation results.

Keywords: non linear controller, robust, sliding mode, kinetic energy

Procedia PDF Downloads 484
29808 Improvement of the Aerodynamic Behaviour of a Land Rover Discovery 4 in Turbulent Flow Using Computational Fluid Dynamics (CFD)

Authors: Ahmed Al-Saadi, Ali Hassanpour, Tariq Mahmud

Abstract:

The main objective of this study is to investigate ways to reduce the aerodynamic drag coefficient and to increase the stability of the full-size Sport Utility Vehicle using three-dimensional Computational Fluid Dynamics (CFD) simulation. The baseline model in the simulation was the Land Rover Discovery 4. Many aerodynamic devices and external design modifications were used in this study. These reduction aerodynamic techniques were tested individually or in combination to get the best design. All new models have the same capacity and comfort of the baseline model. Uniform freestream velocity of the air at inlet ranging from 28 m/s to 40 m/s was used. ANSYS Fluent software (version 16.0) was used to simulate all models. The drag coefficient obtained from the ANSYS Fluent for the baseline model was validated with experimental data. It is found that the use of modern aerodynamic add-on devices and modifications has a significant effect in reducing the aerodynamic drag coefficient.

Keywords: aerodynamics, RANS, sport utility vehicle, turbulent flow

Procedia PDF Downloads 305
29807 Formation of an Empire in the 21st Century: Theoretical Approach in International Relations and a Worldview of the New World Order

Authors: Rami Georg Johann

Abstract:

Against the background of the current geopolitical constellations, the author looks at various empire models, which are discussed and compared with each other with regard to their stability and functioning. The focus is on the fifth concept as a possible new world order in the 21st century. These will be discussed and compared to one another according to their stability and functioning. All empires to be designed will be conceptualised based on one, two, three, four, and five worlds. All worlds are made up of a different constellation of states and relating coalitions. All systems will be discussed in detail. The one-world-system, the“Western Empire,” will be presented as a possible solution to a new world order in the 21st century (fifth concept). The term “Western” in “Western Empire” describes the Western concept after World War II. This Western concept was the result of two horrible world wars in the 20th century.” With this in mind, the fifth concept forms a stable empire system, the “Western Empire,” by political measures tied to two issues. Thus, this world order provides a significantly higher long-term stability in contrast to all other empire models (comprising five, four, three, or two worlds). Confrontations and threats of war are reduced to a minimum. The two issues mentioned are “merger” and “competition.” These are the main differences in forming an empire compared to all empires and realms in the history of mankind. The fifth concept of this theory, the “Western Empire,” acts explicitly as a counter model. The Western Empire (fifth concept) is formed by the merger of world powers without war. Thus, a world order without competition is created. This merged entity secures long-term peace, stability, democratic values, freedom, human rights, equality, and justice in the new world order.

Keywords: empire formation, theory of international relations, Western Empire, world order

Procedia PDF Downloads 135
29806 Crashworthiness Optimization of an Automotive Front Bumper in Composite Material

Authors: S. Boria

Abstract:

In the last years, the crashworthiness of an automotive body structure can be improved, since the beginning of the design stage, thanks to the development of specific optimization tools. It is well known how the finite element codes can help the designer to investigate the crashing performance of structures under dynamic impact. Therefore, by coupling nonlinear mathematical programming procedure and statistical techniques with FE simulations, it is possible to optimize the design with reduced number of analytical evaluations. In engineering applications, many optimization methods which are based on statistical techniques and utilize estimated models, called meta-models, are quickly spreading. A meta-model is an approximation of a detailed simulation model based on a dataset of input, identified by the design of experiments (DOE); the number of simulations needed to build it depends on the number of variables. Among the various types of meta-modeling techniques, Kriging method seems to be excellent in accuracy, robustness and efficiency compared to other ones when applied to crashworthiness optimization. Therefore the application of such meta-model was used in this work, in order to improve the structural optimization of a bumper for a racing car in composite material subjected to frontal impact. The specific energy absorption represents the objective function to maximize and the geometrical parameters subjected to some design constraints are the design variables. LS-DYNA codes were interfaced with LS-OPT tool in order to find the optimized solution, through the use of a domain reduction strategy. With the use of the Kriging meta-model the crashworthiness characteristic of the composite bumper was improved.

Keywords: composite material, crashworthiness, finite element analysis, optimization

Procedia PDF Downloads 248
29805 The SBO/LOCA Analysis of TRACE/SNAP for Kuosheng Nuclear Power Plant

Authors: J. R. Wang, H. T. Lin, Y. Chiang, H. C. Chen, C. Shih

Abstract:

Kuosheng Nuclear Power Plant (NPP) is located on the northern coast of Taiwan. Its nuclear steam supply system is a type of BWR/6 designed and built by General Electric on a twin unit concept. First, the methodology of Kuosheng NPP SPU (Stretch Power Uprate) safety analysis TRACE/SNAP model was developed in this research. Then, in order to estimate the safety of Kuosheng NPP under the more severe condition, the SBO (Station Blackout) + LOCA (Loss-of-Coolant Accident) transient analysis of Kuosheng NPP SPU TRACE/SNAP model was performed. Besides, the animation model of Kuosheng NPP was presented using the animation function of SNAP with TRACE/SNAP analysis results.

Keywords: TRACE, safety analysis, BWR/6, severe accident

Procedia PDF Downloads 698
29804 Intelligent Process Data Mining for Monitoring for Fault-Free Operation of Industrial Processes

Authors: Hyun-Woo Cho

Abstract:

The real-time fault monitoring and diagnosis of large scale production processes is helpful and necessary in order to operate industrial process safely and efficiently producing good final product quality. Unusual and abnormal events of the process may have a serious impact on the process such as malfunctions or breakdowns. This work try to utilize process measurement data obtained in an on-line basis for the safe and some fault-free operation of industrial processes. To this end, this work evaluated the proposed intelligent process data monitoring framework based on a simulation process. The monitoring scheme extracts the fault pattern in the reduced space for the reliable data representation. Moreover, this work shows the results of using linear and nonlinear techniques for the monitoring purpose. It has shown that the nonlinear technique produced more reliable monitoring results and outperforms linear methods. The adoption of the qualitative monitoring model helps to reduce the sensitivity of the fault pattern to noise.

Keywords: process data, data mining, process operation, real-time monitoring

Procedia PDF Downloads 626
29803 Optimization of the Jatropha curcas Supply Chain as a Criteria for the Implementation of Future Collection Points in Rural Areas of Manabi-Ecuador

Authors: Boris G. German, Edward Jiménez, Sebastián Espinoza, Andrés G. Chico, Ricardo A. Narváez

Abstract:

The unique flora and fauna of The Galapagos Islands has leveraged a tourism-driven growth in the islands. Nonetheless, such development is energy-intensive and requires thousands of gallons of diesel each year for thermoelectric electricity generation. The needed transport of fossil fuels from the continent has generated oil spillages and affectations to the fragile ecosystem of the islands. The Zero Fossil Fuels initiative for The Galapagos proposed by the Ecuadorian government as an alternative to reduce the use of fossil fuels in the islands, considers the replacement of diesel in thermoelectric generators, by Jatropha curcas vegetable oil. However, the Jatropha oil supply cannot entirely cover yet the demand for electricity generation in Galapagos. Within this context, the present work aims to provide an optimization model that can be used as a selection criterion for approving new Jatropha Curcas collection points in rural areas of Manabi-Ecuador. For this purpose, existing Jatropha collection points in Manabi were grouped under three regions: north (7 collection points), center (4 collection points) and south (9 collection points). Field work was carried out in every region in order to characterize the collection points, to establish local Jatropha supply and to determine transportation costs. Data collection was complemented using GIS software and an objective function was defined in order to determine the profit associated to Jatropha oil production. The market price of both Jatropha oil and residual cake, were considered for the total revenue; whereas Jatropha price, transportation and oil extraction costs were considered for the total cost. The tonnes of Jatropha fruit and seed, transported from collection points to the extraction plant, were considered as variables. The maximum and minimum amount of the collected Jatropha from each region constrained the optimization problem. The supply chain was optimized using linear programming in order to maximize the profits. Finally, a sensitivity analysis was performed in order to find a profit-based criterion for the acceptance of future collection points in Manabi. The maximum profit reached a value of $ 4,616.93 per year, which represented a total Jatropha collection of 62.3 tonnes Jatropha per year. The northern region of Manabi had the biggest collection share (69%), followed by the southern region (17%). The criteria for accepting new Jatropha collection points in the rural areas of Manabi can be defined by the current maximum profit of the zone and by the variation in the profit when collection points are removed one at a time. The definition of new feasible collection points plays a key role in the supply chain associated to Jatropha oil production. Therefore, a mathematical model that assists decision makers in establishing new collection points while assuring profitability, contributes to guarantee a continued Jatropha oil supply for Galapagos and a sustained economic growth in the rural areas of Ecuador.

Keywords: collection points, Jatropha curcas, linear programming, supply chain

Procedia PDF Downloads 418