Search results for: Control Approach.
6061 A Data-Driven Approach for Studying the Washout Effects of Rain on Air Pollution
Abstract:
Air pollution is a serious environmental threat on a global scale and can cause harm to human health, morbidity and premature mortality. Reliable monitoring and control systems are therefore necessary to develop coping skills against the hazards associated with this phenomenon. However, existing environmental monitoring means often do not provide a sufficient response due to practical and technical limitations. Commercial microwave links that form the infrastructure for transmitting data between cell phone towers can be harnessed to map rain at high tempo-spatial resolution. Rainfall causes a decrease in the signal strength received by these wireless communication links allowing it to be used as a built-in sensor network to map the phenomenon. In this study, we point to the potential that lies in this system to indirectly monitor areas where air pollution is reduced. The relationship between pollutant wash-off and rainfall provides an opportunity to acquire important spatial information about air quality using existing cell-phone tower signals. Since the density of microwave communication networks is high relative to any dedicated sensor arrays, it could be possible to rely on this available observation tool for studying precipitation scavenging on air pollutants, for model needs and more.
Keywords: Air pollution, commercial microwave links, rainfall.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9046060 Applying the Crystal Model Approach on Light Nuclei for Calculating Radii and Density Distribution
Authors: A. Amar
Abstract:
A new model namely, the crystal model, has been modified to calculate radius and density distribution of light nuclei up to 8Be. The crystal model has been modified according to solid state physics which uses the analogy between nucleon distribution and atoms distribution in the crystal. The model has analytical analysis to calculate the radius where the density distribution of light nuclei has been obtained from the analogy of crystal lattice. The distribution of nucleons over crystal has been discussed in general form. The equation used to calculate binding energy was taken from the solid-state model of repulsive and attractive force. The numbers of the protons were taken to control repulsive force where the atomic number was responsible for the attractive force. The parameter has been calculated from the crystal model was found to be proportional to the radius of the nucleus. The density distribution of light nuclei was taken as a summation of two clusters distribution as in 6Li=alpha+deuteron configuration. A test has been done on the data obtained for radius and density distribution using double folding for d+6,7Li with M3Y nucleon-nucleon interaction. Good agreement has been obtained for both radius and density distribution of light nuclei. The model failed to calculate the radius of 9Be, so modifications should be done to overcome discrepancy.
Keywords: nuclear lattice, crystal model, light nuclei, nuclear density distributions
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4306059 Reverse Logistics Information Management Using Ontological Approach
Authors: F. Lhafiane, A. Elbyed, M. Bouchoum
Abstract:
Reverse Logistics (RL) Network is considered as complex and dynamic network that involves many stakeholders such as: suppliers, manufactures, warehouse, retails and costumers, this complexity is inherent in such process due to lack of perfect knowledge or conflicting information. Ontologies on the other hand can be considered as an approach to overcome the problem of sharing knowledge and communication among the various reverse logistics partners. In this paper we propose a semantic representation based on hybrid architecture for building the Ontologies in ascendant way, this method facilitates the semantic reconciliation between the heterogeneous information systems that support reverse logistics processes and product data.
Keywords: Reverse Logistics, information management, heterogeneity, Ontologies, semantic web.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29666058 Statistical Optimization of Medium Components for Biomass Production of Chlorella pyrenoidosa under Autotrophic Conditions and Evaluation of Its Biochemical Composition under Stress Conditions
Authors: N. P. Dhull, K. Gupta, R. Soni, D. K. Rahi, S. K. Soni
Abstract:
The aim of the present work was to statistically design an autotrophic medium for maximum biomass production by Chlorella pyrenoidosa using response surface methodology. After evaluating one factor at a time approach, K2HPO4, KNO3, MgSO4.7H2O and NaHCO3 were preferred over the other components of the fog’s medium as most critical autotrophic medium components. The study showed that the maximum biomass yield was achieved while the concentrations of MgSO4.7H2O, K2HPO4, KNO3 and NaHCO3 were 0.409 g/L, 0.24 g/L, 1.033 g/L, and 3.265 g/L, respectively. The study reported that the biomass productivity of C. pyrenoidosa improved from 0.14 g/L in defined fog’s medium to 1.40 g/L in modified fog’s medium resulting 10 fold increase. The biochemical composition biosynthesis of C. pyrenoidosa was altered using nitrogen limiting stress bringing about 5.23 fold increase in lipid content than control (cell without stress), as analyzed by FTIR integration method.
Keywords: Autotrophic condition, Chlorella pyrenoidosa, FTIR, Response Surface Methodology, Optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24406057 Simultaneous Tuning of Static Var Compensator and Power System Stabilizer Employing Real- Coded Genetic Algorithm
Authors: S. Panda, N. P. Patidar, R. Singh
Abstract:
Power system stability enhancement by simultaneous tuning of a Power System Stabilizer (PSS) and a Static Var Compensator (SVC)-based controller is thoroughly investigated in this paper. The coordination among the proposed damping stabilizers and the SVC internal voltage regulators has also been taken into consideration. The design problem is formulated as an optimization problem with a time-domain simulation-based objective function and Real-Coded Genetic Algorithm (RCGA) is employed to search for optimal controller parameters. The proposed stabilizers are tested on a weakly connected power system with different disturbances and loading conditions. The nonlinear simulation results are presented to show the effectiveness and robustness of the proposed control schemes over a wide range of loading conditions and disturbances. Further, the proposed design approach is found to be robust and improves stability effectively even under small disturbance and unbalanced fault conditions.
Keywords: Real-Coded Genetic Algorithm (RCGA), Static Var Compensator (SVC), Power System Stabilizer (PSS), Low Frequency Oscillations, Power System Stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22556056 Design of an Intelligent Tutor using a Multiagent Approach
Authors: Kamel Khoualdi, Radia Benghezal
Abstract:
Research in distributed artificial intelligence and multiagent systems consider how a set of distributed entities can interact and coordinate their actions in order to solve a given problem. In this paper an overview of this concept and its evolution is presented particularly its application in the design of intelligent tutoring systems. An intelligent tutor based on the concept of agent and centered specifically on the design of a pedagogue agent is illustrated. Our work has two goals: the first one concerns the architecture aspect and the design of a tutor using multiagent approach. The second one deals particularly with the design of a part of a tutor system: the pedagogue agent.Keywords: Intelligent tutoring systems, Multiagent systems, Pedagogue agent, Planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14906055 Feature Selection for Breast Cancer Diagnosis: A Case-Based Wrapper Approach
Authors: Mohammad Darzi, Ali AsgharLiaei, Mahdi Hosseini, HabibollahAsghari
Abstract:
This article addresses feature selection for breast cancer diagnosis. The present process contains a wrapper approach based on Genetic Algorithm (GA) and case-based reasoning (CBR). GA is used for searching the problem space to find all of the possible subsets of features and CBR is employed to estimate the evaluation result of each subset. The results of experiment show that the proposed model is comparable to the other models on Wisconsin breast cancer (WDBC) dataset.Keywords: Case-based reasoning; Breast cancer diagnosis; Genetic algorithm; Wrapper feature selection
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28756054 Simulation Data Management Approach for Developing Adaptronic Systems – The W-Model Methodology
Authors: Roland S. Nattermann, Reiner Anderl
Abstract:
Existing proceeding-models for the development of mechatronic systems provide a largely parallel action in the detailed development. This parallel approach is to take place also largely independent of one another in the various disciplines involved. An approach for a new proceeding-model provides a further development of existing models to use for the development of Adaptronic Systems. This approach is based on an intermediate integration and an abstract modeling of the adaptronic system. Based on this system-model a simulation of the global system behavior, due to external and internal factors or Forces is developed. For the intermediate integration a special data management system is used. According to the presented approach this data management system has a number of functions that are not part of the "normal" PDM functionality. Therefore a concept for a new data management system for the development of Adaptive system is presented in this paper. This concept divides the functions into six layers. In the first layer a system model is created, which divides the adaptronic system based on its components and the various technical disciplines. Moreover, the parameters and properties of the system are modeled and linked together with the requirements and the system model. The modeled parameters and properties result in a network which is analyzed in the second layer. From this analysis necessary adjustments to individual components for specific manipulation of the system behavior can be determined. The third layer contains an automatic abstract simulation of the system behavior. This simulation is a precursor for network analysis and serves as a filter. By the network analysis and simulation changes to system components are examined and necessary adjustments to other components are calculated. The other layers of the concept treat the automatic calculation of system reliability, the "normal" PDM-functionality and the integration of discipline-specific data into the system model. A prototypical implementation of an appropriate data management with the addition of an automatic system development is being implemented using the data management system ENOVIA SmarTeam V5 and the simulation system MATLAB.
Keywords: Adaptronic, Data-Management, LOEWE-CentreAdRIA
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23686053 A Fuzzy Decision Making Approach for Supplier Selection in Healthcare Industry
Authors: Zeynep Sener, Mehtap Dursun
Abstract:
Supplier evaluation and selection is one of the most important components of an effective supply chain management system. Due to the expanding competition in healthcare, selecting the right medical device suppliers offers great potential for increasing quality while decreasing costs. This paper proposes a fuzzy decision making approach for medical supplier selection. A real-world medical device supplier selection problem is presented to illustrate the application of the proposed decision methodology.
Keywords: Fuzzy decision making, fuzzy multiple objective programming, medical supply chain, supplier selection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26726052 Concurrent Approach to Data Parallel Model using Java
Authors: Bala Dhandayuthapani Veerasamy
Abstract:
Parallel programming models exist as an abstraction of hardware and memory architectures. There are several parallel programming models in commonly use; they are shared memory model, thread model, message passing model, data parallel model, hybrid model, Flynn-s models, embarrassingly parallel computations model, pipelined computations model. These models are not specific to a particular type of machine or memory architecture. This paper expresses the model program for concurrent approach to data parallel model through java programming.Keywords: Concurrent, Data Parallel, JDK, Parallel, Thread
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20976051 A Relationship between Two Stabilizing Controllers and Its Application to Two-Stage Compensator Design without Coprime Factorizability – Single-Input Single-Output Case –
Authors: Kazuyoshi Mori
Abstract:
In this paper, we first show a relationship between two stabilizing controllers, which presents an extended feedback system using two stabilizing controllers. Then, we apply this relationship to the two-stage compensator design. In this paper, we consider singleinput single-output plants. On the other hand, we do not assume the coprime factorizability of the model. Thus, the results of this paper are based on the factorization approach only, so that they can be applied to numerous linear systems.Keywords: Relationship among Compensators, Two-Stage Compensator Design, Parametrization of Stabilizing Controllers, Factorization Approach
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12106050 User-Perceived Quality Factors for Certification Model of Web-Based System
Authors: Jamaiah H. Yahaya, Aziz Deraman, Abdul Razak Hamdan, Yusmadi Yah Jusoh
Abstract:
One of the most essential issues in software products is to maintain it relevancy to the dynamics of the user’s requirements and expectation. Many studies have been carried out in quality aspect of software products to overcome these problems. Previous software quality assessment models and metrics have been introduced with strengths and limitations. In order to enhance the assurance and buoyancy of the software products, certification models have been introduced and developed. From our previous experiences in certification exercises and case studies collaborating with several agencies in Malaysia, the requirements for user based software certification approach is identified and demanded. The emergence of social network applications, the new development approach such as agile method and other varieties of software in the market have led to the domination of users over the software. As software become more accessible to the public through internet applications, users are becoming more critical in the quality of the services provided by the software. There are several categories of users in web-based systems with different interests and perspectives. The classifications and metrics are identified through brain storming approach with includes researchers, users and experts in this area. The new paradigm in software quality assessment is the main focus in our research. This paper discusses the classifications of users in web-based software system assessment and their associated factors and metrics for quality measurement. The quality model is derived based on IEEE structure and FCM model. The developments are beneficial and valuable to overcome the constraints and improve the application of software certification model in future.
Keywords: Software certification model, user centric approach, software quality factors, metrics and measurements, web-based system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21486049 Conceptual Model for Knowledge Sharing Model in Creating Idea for Mobile Application
Authors: Hanafizan Hussain
Abstract:
This study shows that several projects will be conducted at the workshop in which using the conceptual model for knowledge sharing approach to create an idea for mobile application. The sharing idea has been done through the collaborative activity in which a group of different field sought to define the mobile application which will lead to new media approach of using social media platform. The collaborative activity will be provided and implemented in the form of one day workshop to determine the approach towards the theme given. The activity later will be continued for four weeks for the participant to prepare for the pitch day workshop. This paper shows the pitch of idea including the interface and prototype for the said products. The collaboration between the members with different field of study shows that social media influenced the knowledge sharing model and its creation or innovations. One of the projects supported a collaborative activity in which a group of young designers sought to define the knowledge sharing model of their ability in creating idea for mobile applications.Keywords: Mobile application, collaborative activity, conceptual knowledge sharing model, social media platform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7256048 Thresholding Approach for Automatic Detection of Pseudomonas aeruginosa Biofilms from Fluorescence in situ Hybridization Images
Authors: Zonglin Yang, Tatsuya Akiyama, Kerry S. Williamson, Michael J. Franklin, Thiruvarangan Ramaraj
Abstract:
Pseudomonas aeruginosa is an opportunistic pathogen that forms surface-associated microbial communities (biofilms) on artificial implant devices and on human tissue. Biofilm infections are difficult to treat with antibiotics, in part, because the bacteria in biofilms are physiologically heterogeneous. One measure of biological heterogeneity in a population of cells is to quantify the cellular concentrations of ribosomes, which can be probed with fluorescently labeled nucleic acids. The fluorescent signal intensity following fluorescence in situ hybridization (FISH) analysis correlates to the cellular level of ribosomes. The goals here are to provide computationally and statistically robust approaches to automatically quantify cellular heterogeneity in biofilms from a large library of epifluorescent microscopy FISH images. In this work, the initial steps were developed toward these goals by developing an automated biofilm detection approach for use with FISH images. The approach allows rapid identification of biofilm regions from FISH images that are counterstained with fluorescent dyes. This methodology provides advances over other computational methods, allowing subtraction of spurious signals and non-biological fluorescent substrata. This method will be a robust and user-friendly approach which will enable users to semi-automatically detect biofilm boundaries and extract intensity values from fluorescent images for quantitative analysis of biofilm heterogeneity.
Keywords: Image informatics, Pseudomonas aeruginosa, biofilm, FISH, computer vision, data visualization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11816047 Prediction of Binding Free Energies for Dyes Removal Using Computational Chemistry
Authors: R. Chanajaree, D. Luanwiset, K. Pongpratea
Abstract:
Dye removal is an environmental concern because the textile industries have been increasing by world population and industrialization. Adsorption is the technique to find adsorbents to remove dyes from wastewater. This method is low-cost and effective for dye removal. This work tries to develop effective adsorbents using the computational approach because it will be able to predict the possibility of the adsorbents for specific dyes in terms of binding free energies. The computational approach is faster and cheaper than the experimental approach in case of finding the best adsorbents. All starting structures of dyes and adsorbents are optimized by quantum calculation. The complexes between dyes and adsorbents are generated by the docking method. The obtained binding free energies from docking are compared to binding free energies from the experimental data. The calculated energies can be ranked as same as the experimental results. In addition, this work also shows the possible orientation of the complexes. This work used two experimental groups of the complexes of the dyes and adsorbents. In the first group, there are chitosan (adsorbent) and two dyes (reactive red (RR) and direct sun yellow (DY)). In the second group, there are poly(1,2-epoxy-3-phenoxy) propane (PEPP), which is the adsorbent, and 2 dyes of bromocresol green (BCG) and alizarin yellow (AY).
Keywords: Dye removal, binding free energies, quantum calculation, docking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7216046 A Metric-Set and Model Suggestion for Better Software Project Cost Estimation
Authors: Murat Ayyıldız, Oya Kalıpsız, Sırma Yavuz
Abstract:
Software project effort estimation is frequently seen as complex and expensive for individual software engineers. Software production is in a crisis. It suffers from excessive costs. Software production is often out of control. It has been suggested that software production is out of control because we do not measure. You cannot control what you cannot measure. During last decade, a number of researches on cost estimation have been conducted. The metric-set selection has a vital role in software cost estimation studies; its importance has been ignored especially in neural network based studies. In this study we have explored the reasons of those disappointing results and implemented different neural network models using augmented new metrics. The results obtained are compared with previous studies using traditional metrics. To be able to make comparisons, two types of data have been used. The first part of the data is taken from the Constructive Cost Model (COCOMO'81) which is commonly used in previous studies and the second part is collected according to new metrics in a leading international company in Turkey. The accuracy of the selected metrics and the data samples are verified using statistical techniques. The model presented here is based on Multi-Layer Perceptron (MLP). Another difficulty associated with the cost estimation studies is the fact that the data collection requires time and care. To make a more thorough use of the samples collected, k-fold, cross validation method is also implemented. It is concluded that, as long as an accurate and quantifiable set of metrics are defined and measured correctly, neural networks can be applied in software cost estimation studies with successKeywords: Software Metrics, Software Cost Estimation, Neural Network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19576045 Sensitivity Analysis during the Optimization Process Using Genetic Algorithms
Authors: M. A. Rubio, A. Urquia
Abstract:
Genetic algorithms (GA) are applied to the solution of high-dimensional optimization problems. Additionally, sensitivity analysis (SA) is usually carried out to determine the effect on optimal solutions of changes in parameter values of the objective function. These two analyses (i.e., optimization and sensitivity analysis) are computationally intensive when applied to high-dimensional functions. The approach presented in this paper consists in performing the SA during the GA execution, by statistically analyzing the data obtained of running the GA. The advantage is that in this case SA does not involve making additional evaluations of the objective function and, consequently, this proposed approach requires less computational effort than conducting optimization and SA in two consecutive steps.Keywords: Optimization, sensitivity, genetic algorithms, model calibration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14746044 An Inverse Heat Transfer Algorithm for Predicting the Thermal Properties of Tumors during Cryosurgery
Authors: Mohamed Hafid, Marcel Lacroix
Abstract:
This study aimed at developing an inverse heat transfer approach for predicting the time-varying freezing front and the temperature distribution of tumors during cryosurgery. Using a temperature probe pressed against the layer of tumor, the inverse approach is able to predict simultaneously the metabolic heat generation and the blood perfusion rate of the tumor. Once these parameters are predicted, the temperature-field and time-varying freezing fronts are determined with the direct model. The direct model rests on one-dimensional Pennes bioheat equation. The phase change problem is handled with the enthalpy method. The Levenberg-Marquardt Method (LMM) combined to the Broyden Method (BM) is used to solve the inverse model. The effect (a) of the thermal properties of the diseased tissues; (b) of the initial guesses for the unknown thermal properties; (c) of the data capture frequency; and (d) of the noise on the recorded temperatures is examined. It is shown that the proposed inverse approach remains accurate for all the cases investigated.
Keywords: Cryosurgery, inverse heat transfer, Levenberg-Marquardt method, thermal properties, Pennes model, enthalpy method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14996043 Optimal Capacitor Allocation for loss reduction in Distribution System Using Fuzzy and Plant Growth Simulation Algorithm
Authors: R. Srinivasa Rao
Abstract:
This paper presents a new and efficient approach for capacitor placement in radial distribution systems that determine the optimal locations and size of capacitor with an objective of improving the voltage profile and reduction of power loss. The solution methodology has two parts: in part one the loss sensitivity factors are used to select the candidate locations for the capacitor placement and in part two a new algorithm that employs Plant growth Simulation Algorithm (PGSA) is used to estimate the optimal size of capacitors at the optimal buses determined in part one. The main advantage of the proposed method is that it does not require any external control parameters. The other advantage is that it handles the objective function and the constraints separately, avoiding the trouble to determine the barrier factors. The proposed method is applied to 9 and 34 bus radial distribution systems. The solutions obtained by the proposed method are compared with other methods. The proposed method has outperformed the other methods in terms of the quality of solution.Keywords: Distribution systems, Capacitor allocation, Loss reduction, Fuzzy, PGSA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22796042 Information Fusion for Identity Verification
Authors: Girija Chetty, Monica Singh
Abstract:
In this paper we propose a novel approach for ascertaining human identity based on fusion of profile face and gait biometric cues The identification approach based on feature learning in PCA-LDA subspace, and classification using multivariate Bayesian classifiers allows significant improvement in recognition accuracy for low resolution surveillance video scenarios. The experimental evaluation of the proposed identification scheme on a publicly available database [2] showed that the fusion of face and gait cues in joint PCA-LDA space turns out to be a powerful method for capturing the inherent multimodality in walking gait patterns, and at the same time discriminating the person identity..
Keywords: Biometrics, gait recognition, PCA, LDA, Eigenface, Fisherface, Multivariate Gaussian Classifier
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17796041 Color Image Segmentation Using Kekre-s Algorithm for Vector Quantization
Authors: H. B. Kekre, Tanuja K. Sarode, Bhakti Raul
Abstract:
In this paper we propose segmentation approach based on Vector Quantization technique. Here we have used Kekre-s fast codebook generation algorithm for segmenting low-altitude aerial image. This is used as a preprocessing step to form segmented homogeneous regions. Further to merge adjacent regions color similarity and volume difference criteria is used. Experiments performed with real aerial images of varied nature demonstrate that this approach does not result in over segmentation or under segmentation. The vector quantization seems to give far better results as compared to conventional on-the-fly watershed algorithm.Keywords: Image Segmentation, , Codebook, Codevector, data compression, Encoding
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21956040 Peculiarities of Comprehending the Subjective Well- Being by Student with High and Low Level of Emotional Intelligence
Authors: Veronika Pivkina, Alla Kim, Khon Nataliya
Abstract:
In this paper, the actuality of the study, and the role of subjective well-being problem in modern psychology and the comprehending of subjective well-being by current students is defined. The purpose of this research is to educe peculiarities of comprehending of subjective well-being by students with various levels of emotional intelligence. Methods of research are adapted Russian-Language questionnaire of K. Riff 'The scales of psychological well-being'; emotional intelligence questionnaire of D. V. Lusin. The research involved 72 students from different universities and disciplines aged between 18 and 24. Analyzing the results of the studies, it can be concluded that the understanding of happiness in different groups of students with high and low levels of overall emotional intelligence is different, as well as differentiated by gender. Students with a higher level of happiness possess more capacity and higher need to control their emotions, to cause and maintain the desired emotions and control something undesirable.
Keywords: Subjective well-being, emotional intelligence, psychology of comprehending, students.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16826039 Analyzing Defects with Failure Assessment Diagrams of Gas Pipelines
Authors: Alfred Hasanaj, Ardit Gjeta, Miranda Kullolli
Abstract:
The approach in analyzing defects on different pipe lines is conducted through Failure Assessment Diagram (FAD). These methods of analyses have further extended in recent years. This approach is used to identify and stress out a solution for the defects which randomly occur with gas pipes such are corrosion defects, gauge defects, and combination of defects where gauge and dents are included. Few of the defects are to be analyzed in this paper where our main focus will be the fracture of cast Iron pipes, elastic-plastic failure and plastic collapse of X52 steel pipes for gas transport. We need to conduct a calculation of probability of the defects in order to predict and avoid such costly defects.
Keywords: Defects, Failure Assessment Diagrams, Safety Factor Steel Pipes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29476038 Energy Efficiency Analysis of Crossover Technologies in Industrial Applications
Authors: W. Schellong
Abstract:
Industry accounts for one-third of global final energy demand. Crossover technologies (e.g. motors, pumps, process heat, and air conditioning) play an important role in improving energy efficiency. These technologies are used in many applications independent of the production branch. Especially electrical power is used by drives, pumps, compressors, and lightning. The paper demonstrates the algorithm of the energy analysis by some selected case studies for typical industrial processes. The energy analysis represents an essential part of energy management systems (EMS). Generally, process control system (PCS) can support EMS. They provide information about the production process, and they organize the maintenance actions. Combining these tools into an integrated process allows the development of an energy critical equipment strategy. Thus, asset and energy management can use the same common data to improve the energy efficiency.
Keywords: Crossover technologies, data management, energy analysis, energy efficiency, process control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9616037 Optimal Capacitor Placement in a Radial Distribution System using Plant Growth Simulation Algorithm
Authors: R. Srinivasa Rao, S. V. L. Narasimham
Abstract:
This paper presents a new and efficient approach for capacitor placement in radial distribution systems that determine the optimal locations and size of capacitor with an objective of improving the voltage profile and reduction of power loss. The solution methodology has two parts: in part one the loss sensitivity factors are used to select the candidate locations for the capacitor placement and in part two a new algorithm that employs Plant growth Simulation Algorithm (PGSA) is used to estimate the optimal size of capacitors at the optimal buses determined in part one. The main advantage of the proposed method is that it does not require any external control parameters. The other advantage is that it handles the objective function and the constraints separately, avoiding the trouble to determine the barrier factors. The proposed method is applied to 9, 34, and 85-bus radial distribution systems. The solutions obtained by the proposed method are compared with other methods. The proposed method has outperformed the other methods in terms of the quality of solution.
Keywords: Distribution systems, Capacitor placement, loss reduction, Loss sensitivity factors, PGSA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 52826036 A Monte Carlo Method to Data Stream Analysis
Authors: Kittisak Kerdprasop, Nittaya Kerdprasop, Pairote Sattayatham
Abstract:
Data stream analysis is the process of computing various summaries and derived values from large amounts of data which are continuously generated at a rapid rate. The nature of a stream does not allow a revisit on each data element. Furthermore, data processing must be fast to produce timely analysis results. These requirements impose constraints on the design of the algorithms to balance correctness against timely responses. Several techniques have been proposed over the past few years to address these challenges. These techniques can be categorized as either dataoriented or task-oriented. The data-oriented approach analyzes a subset of data or a smaller transformed representation, whereas taskoriented scheme solves the problem directly via approximation techniques. We propose a hybrid approach to tackle the data stream analysis problem. The data stream has been both statistically transformed to a smaller size and computationally approximated its characteristics. We adopt a Monte Carlo method in the approximation step. The data reduction has been performed horizontally and vertically through our EMR sampling method. The proposed method is analyzed by a series of experiments. We apply our algorithm on clustering and classification tasks to evaluate the utility of our approach.Keywords: Data Stream, Monte Carlo, Sampling, DensityEstimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14176035 Changes in Subjective and Objective Measures of Performance in Ramadan
Authors: H. Alabed, K. Abuzayan, J. Waterhouse
Abstract:
The Muslim faith requires individuals to fast between the hours of sunrise and sunset during the month of Ramadan. Our recent work has concentrated on some of the changes that take place during the daytime when fasting. A questionnaire was developed to assess subjective estimates of physical, mental and social activities, and fatigue. Four days were studied: in the weeks before and after Ramadan (control days) and during the first and last weeks of Ramadan (experimental days). On each of these four days, this questionnaire was given several times during the daytime and once after the fast had been broken and just before individuals retired at night. During Ramadan, daytime mental, physical and social activities all decreased below control values but then increased to abovecontrol values in the evening. The desires to perform physical and mental activities showed very similar patterns. That is, individuals tried to conserve energy during the daytime in preparation for the evenings when they ate and drank, often with friends. During Ramadan also, individuals were more fatigued in the daytime and napped more often than on control days. This extra fatigue probably reflected decreased sleep, individuals often having risen earlier (before sunrise, to prepare for fasting) and retired later (to enable recovery from the fast). Some physiological measures and objective measures of performance (including the response to a bout of exercise) have also been investigated. Urine osmolality fell during the daytime on control days as subjects drank, but rose in Ramadan to reach values at sunset indicative of dehydration. Exercise performance was also compromised, particularly late in the afternoon when the fast had lasted several hours. Self-chosen exercise work-rates fell and a set amount of exercise felt more arduous. There were also changes in heart rate and lactate accumulation in the blood, indicative of greater cardiovascular and metabolic stress caused by the exercise in subjects who had been fasting. Daytime fasting in Ramadan produces widespread effects which probably reflect combined effects of sleep loss and restrictions to intakes of water and food.Keywords: Drinking, Eating, Mental Performance, Physical Performance, Social Activity, Sleepiness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17476034 Fast Short-Term Electrical Load Forecasting under High Meteorological Variability with a Multiple Equation Time Series Approach
Authors: Charline David, Alexandre Blondin Massé, Arnaud Zinflou
Abstract:
We present a multiple equation time series approach for the short-term load forecasting applied to the electrical power load consumption for the whole Quebec province, in Canada. More precisely, we take into account three meteorological variables — temperature, cloudiness and wind speed —, and we use meteorological measurements taken at different locations on the territory. Our final model shows an average MAPE score of 1.79% over an 8-years dataset.
Keywords: Short-term load forecasting, special days, time series, multiple equations, parallelization, clustering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2896033 Brand Identity Creation for Thai Halal Brands
Authors: Pibool Waijittragum
Abstract:
The purpose of this paper is to synthesize the research result of brand Identities of Thai Halal brands which related to the way of life for Thai Muslims. The results will be transforming to Thai Halal Brands packaging and label design. The expected benefit is an alternative of marketing strategy for brand building process for Halal products in Thailand. Four elements of marketing strategies which necessary for the brand identity creation is the research framework: consists of Attributes, Benefits, Values and Personality. The research methodology was applied using qualitative and quantitative; 19 marketing experts with dynamic roles in Thai consumer products were interviewed. In addition, a field survey of 122 Thai Muslims selected from 175 Muslim communities in Bangkok was studied. Data analysis will be according to 5 categories of Thai Halal product: 1) Meat 2) Vegetable and Fruits 3) Instant foods and Garnishing ingredient 4) Beverages, Desserts and Snacks 5) Hygienic daily products.
The results will explain some suitable approach for brand Identities of Thai Halal brands as are: 1) Benefit approach as the characteristics of the product with its benefit. The brand identity created transform to the packaging design should be clear and display a fresh product 2) Value approach as the value of products that affect to consumers’ perception. The brand identity created transform to the packaging design should be simply look and using a trustful image 3) Personality approach as the reflection of consumers thought. The brand identity created transform to the packaging design should be sincere, enjoyable, merry, flamboyant look and using a humoristic image.
Keywords: Marketing strategies, Brand identity, Packaging and Label Design, Thai Halal products.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22216032 Tabu Search Approach to Solve Routing Issues in Communication Networks
Authors: Anant Oonsivilai, Wichai Srisuruk, Boonruang Marungsri, Thanatchai Kulworawanichpong
Abstract:
Optimal routing in communication networks is a major issue to be solved. In this paper, the application of Tabu Search (TS) in the optimum routing problem where the aim is to minimize the computational time and improvement of quality of the solution in the communication have been addressed. The goal is to minimize the average delays in the communication. The effectiveness of Tabu Search method is shown by the results of simulation to solve the shortest path problem. Through this approach computational cost can be reduced.Keywords: Communication networks, optimum routing network, tabu search algorithm, shortest path.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2096