Search results for: hot strip process adjustment
4732 Performance Comparison of Different Regression Methods for a Polymerization Process with Adaptive Sampling
Authors: Florin Leon, Silvia Curteanu
Abstract:
Developing complete mechanistic models for polymerization reactors is not easy, because complex reactions occur simultaneously; there is a large number of kinetic parameters involved and sometimes the chemical and physical phenomena for mixtures involving polymers are poorly understood. To overcome these difficulties, empirical models based on sampled data can be used instead, namely regression methods typical of machine learning field. They have the ability to learn the trends of a process without any knowledge about its particular physical and chemical laws. Therefore, they are useful for modeling complex processes, such as the free radical polymerization of methyl methacrylate achieved in a batch bulk process. The goal is to generate accurate predictions of monomer conversion, numerical average molecular weight and gravimetrical average molecular weight. This process is associated with non-linear gel and glass effects. For this purpose, an adaptive sampling technique is presented, which can select more samples around the regions where the values have a higher variation. Several machine learning methods are used for the modeling and their performance is compared: support vector machines, k-nearest neighbor, k-nearest neighbor and random forest, as well as an original algorithm, large margin nearest neighbor regression. The suggested method provides very good results compared to the other well-known regression algorithms.Keywords: Adaptive sampling, batch bulk methyl methacrylate polymerization, large margin nearest neighbor regression, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14024731 The Use of Performance Indicators for Evaluating Models of Drying Jackfruit (Artocarpus heterophyllus L.): Page, Midilli, and Lewis
Authors: D. S. C. Soares, D. G. Costa, J. T. S., A. K. S. Abud, T. P. Nunes, A. M. Oliveira Júnior
Abstract:
Mathematical models of drying are used for the purpose of understanding the drying process in order to determine important parameters for design and operation of the dryer. The jackfruit is a fruit with high consumption in the Northeast and perishability. It is necessary to apply techniques to improve their conservation for longer in order to diffuse it by regions with low consumption. This study aimed to analyze several mathematical models (Page, Lewis, and Midilli) to indicate one that best fits the conditions of convective drying process using performance indicators associated with each model: accuracy (Af) and noise factors (Bf), mean square error (RMSE) and standard error of prediction (% SEP). Jackfruit drying was carried out in convective type tray dryer at a temperature of 50°C for 9 hours. It is observed that the model Midili was more accurate with Af: 1.39, Bf: 1.33, RMSE: 0.01%, and SEP: 5.34. However, the use of the Model Midilli is not appropriate for purposes of control process due to need four tuning parameters. With the performance indicators used in this paper, the Page model showed similar results with only two parameters. It is concluded that the best correlation between the experimental and estimated data is given by the Page’s model.
Keywords: Drying, models, jackfruit.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24234730 Handwriting Velocity Modeling by Artificial Neural Networks
Authors: Mohamed Aymen Slim, Afef Abdelkrim, Mohamed Benrejeb
Abstract:
The handwriting is a physical demonstration of a complex cognitive process learnt by man since his childhood. People with disabilities or suffering from various neurological diseases are facing so many difficulties resulting from problems located at the muscle stimuli (EMG) or signals from the brain (EEG) and which arise at the stage of writing. The handwriting velocity of the same writer or different writers varies according to different criteria: age, attitude, mood, writing surface, etc. Therefore, it is interesting to reconstruct an experimental basis records taking, as primary reference, the writing speed for different writers which would allow studying the global system during handwriting process. This paper deals with a new approach of the handwriting system modeling based on the velocity criterion through the concepts of artificial neural networks, precisely the Radial Basis Functions (RBF) neural networks. The obtained simulation results show a satisfactory agreement between responses of the developed neural model and the experimental data for various letters and forms then the efficiency of the proposed approaches.
Keywords: ElectroMyoGraphic (EMG) signals, Experimental approach, Handwriting process, Radial Basis Functions (RBF) neural networks, Velocity Modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23174729 On the Parameter Optimization of Fuzzy Inference Systems
Authors: Erika Martinez Ramirez, Rene V. Mayorga
Abstract:
Nowadays, more engineering systems are using some kind of Artificial Intelligence (AI) for the development of their processes. Some well-known AI techniques include artificial neural nets, fuzzy inference systems, and neuro-fuzzy inference systems among others. Furthermore, many decision-making applications base their intelligent processes on Fuzzy Logic; due to the Fuzzy Inference Systems (FIS) capability to deal with problems that are based on user knowledge and experience. Also, knowing that users have a wide variety of distinctiveness, and generally, provide uncertain data, this information can be used and properly processed by a FIS. To properly consider uncertainty and inexact system input values, FIS normally use Membership Functions (MF) that represent a degree of user satisfaction on certain conditions and/or constraints. In order to define the parameters of the MFs, the knowledge from experts in the field is very important. This knowledge defines the MF shape to process the user inputs and through fuzzy reasoning and inference mechanisms, the FIS can provide an “appropriate" output. However an important issue immediately arises: How can it be assured that the obtained output is the optimum solution? How can it be guaranteed that each MF has an optimum shape? A viable solution to these questions is through the MFs parameter optimization. In this Paper a novel parameter optimization process is presented. The process for FIS parameter optimization consists of the five simple steps that can be easily realized off-line. Here the proposed process of FIS parameter optimization it is demonstrated by its implementation on an Intelligent Interface section dealing with the on-line customization / personalization of internet portals applied to E-commerce.Keywords: Artificial Intelligence, Fuzzy Logic, Fuzzy InferenceSystems, Nonlinear Optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19844728 An Experimental Investigation on the Effect of Deep cold Rolling Parameters on Surface Roughness and Hardness of AISI 4140 Steel
Authors: P. R. Prabhu, S. M. Kulkarni, S. S. Sharma
Abstract:
Deep cold rolling (DCR) is a cold working process, which easily produces a smooth and work-hardened surface by plastic deformation of surface irregularities. In the present study, the influence of main deep cold rolling process parameters on the surface roughness and the hardness of AISI 4140 steel were studied by using fractional factorial design of experiments. The assessment of the surface integrity aspects on work material was done, in terms of identifying the predominant factor amongst the selected parameters, their order of significance and setting the levels of the factors for minimizing surface roughness and/or maximizing surface hardness. It was found that the ball diameter, rolling force, initial surface roughness and number of tool passes are the most pronounced parameters, which have great effects on the work piece-s surface during the deep cold rolling process. A simple, inexpensive and newly developed DCR tool, with interchangeable collet for using different ball diameters, was used throughout the experimental work presented in this paper.
Keywords: Deep cold rolling, design of experiments, surface hardness, surface roughness
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21574727 Improvement of Reaction Technology of Decalin Halogenation
Authors: Dmitriy Yu. Korulkin, Ravshan M. Nuraliev, Raissa A. Muzychkina
Abstract:
In this research paper were investigated the main regularities of a radical bromination reaction of decalin. There had been studied the temperature effect, durations of reaction, frequency rate of process, a ratio of initial components, type and number of the initiator on decalin bromination degree. There were specified optimum conditions of synthesis of a perbromodecalin by the method of a decalin bromination. There are developed the technological flowchart of receiving a perbromodecalin and the mass balance of process on the first and the subsequent loadings of components. The results of research of antibacterial and antifungal activity of synthesized bromoderivatives have been represented.
Keywords: Decalin, optimum technology, perbromodecalin, radical bromination.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22184726 Mathematical Modeling to Predict Surface Roughness in CNC Milling
Authors: Ab. Rashid M.F.F., Gan S.Y., Muhammad N.Y.
Abstract:
Surface roughness (Ra) is one of the most important requirements in machining process. In order to obtain better surface roughness, the proper setting of cutting parameters is crucial before the process take place. This research presents the development of mathematical model for surface roughness prediction before milling process in order to evaluate the fitness of machining parameters; spindle speed, feed rate and depth of cut. 84 samples were run in this study by using FANUC CNC Milling α-Τ14ιE. Those samples were randomly divided into two data sets- the training sets (m=60) and testing sets(m=24). ANOVA analysis showed that at least one of the population regression coefficients was not zero. Multiple Regression Method was used to determine the correlation between a criterion variable and a combination of predictor variables. It was established that the surface roughness is most influenced by the feed rate. By using Multiple Regression Method equation, the average percentage deviation of the testing set was 9.8% and 9.7% for training data set. This showed that the statistical model could predict the surface roughness with about 90.2% accuracy of the testing data set and 90.3% accuracy of the training data set.
Keywords: Surface roughness, regression analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21334725 A Study of the Variables in the Optimisation of a Platinum Precipitation Process
Authors: Tebogo Phetla, Edison Muzenda, M Belaid
Abstract:
This study investigated possible ways to improve the efficiency of the platinum precipitation process using ammonium chloride by reducing the platinum content reporting to the effluent. The ore treated consist of five platinum group metals namely, ruthenium, rhodium, iridium, platinum, palladium and a precious metal gold. Gold, ruthenium, rhodium and iridium were extracted prior the platinum precipitation process. Temperature, reducing agent, flow rate and potential difference were the variables controlled to determine the operation conditions for optimum platinum precipitation efficiency. Hydrogen peroxide was added as the oxidizing agent at the temperature of 85-90oC and potential difference of 700-850mV was the variable used to check the oxidizing state of platinum. The platinum was further purified at temperature between 60-65oC, potential difference above 700 mV, ammonium chloride of 200 l, and at these conditions the platinum content reporting to the effluent was reduced to less than 300ppm, resulting in optimum platinum precipitation efficiency and purity of 99.9%.Keywords: Platinum Group Metals (PGM), Potential difference, Precipitation, Redox reactions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 47914724 Process Development of Safe and Ready-to-eat Raw Oyster Meat by Irradiation Technology
Authors: Pattama Ratana-Arporn, Pongtep Wilaipun
Abstract:
White scar oyster (Crassostrea belcheri) is often eaten raw and being the leading vehicle for foodborne disease, especially Salmonella Weltevreden which exposed the prominent and most resistant to radiation. Gamma irradiation at a low dose of 1 kGy was enough to eliminate S. Weltevreden contaminated in oyster meat at a level up to 5 log CFU/g while it still retain the raw characteristics and equivalent sensory quality as the non-irradiated one. Process development of ready-to-eat chilled oyster meat was conducted by shucking the meat, individually packed in plastic bags, subjected to 1 kGy gamma radiation at chilled condition and then stored in 4oC refrigerated temperature. Microbiological determination showed the absence of S. Weltevreden (5 log CFU/g initial inoculated) along the whole storage time of 30 days. Sensory evaluation indicated the decreasing in sensory scores along storage time which determining the product shelf life to be 18 days compared to 15 days of nonirradiated one. The most advantage of developed process was to provide the safe raw oyster to consumers and in addition sensory quality retained and 3-day extension shelf life also exist.Keywords: decontamination, food safety, irradiation, oyster, Salmonella Weltevreden
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16844723 Decision Support System for Flood Crisis Management using Artificial Neural Network
Authors: Muhammad Aqil, Ichiro Kita, Akira Yano, Nishiyama Soichi
Abstract:
This paper presents an alternate approach that uses artificial neural network to simulate the flood level dynamics in a river basin. The algorithm was developed in a decision support system environment in order to enable users to process the data. The decision support system is found to be useful due to its interactive nature, flexibility in approach and evolving graphical feature and can be adopted for any similar situation to predict the flood level. The main data processing includes the gauging station selection, input generation, lead-time selection/generation, and length of prediction. This program enables users to process the flood level data, to train/test the model using various inputs and to visualize results. The program code consists of a set of files, which can as well be modified to match other purposes. This program may also serve as a tool for real-time flood monitoring and process control. The running results indicate that the decision support system applied to the flood level seems to have reached encouraging results for the river basin under examination. The comparison of the model predictions with the observed data was satisfactory, where the model is able to forecast the flood level up to 5 hours in advance with reasonable prediction accuracy. Finally, this program may also serve as a tool for real-time flood monitoring and process control.Keywords: Decision Support System, Neural Network, Flood Level
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16284722 Simultaneous Saccharification and Fermentation(SSF) of Sugarcane Bagasse - Kinetics and Modeling
Authors: E.Sasikumar, T.Viruthagiri
Abstract:
Simultaneous Saccharification and Fermentation (SSF) of sugarcane bagasse by cellulase and Pachysolen tannophilus MTCC *1077 were investigated in the present study. Important process variables for ethanol production form pretreated bagasse were optimized using Response Surface Methodology (RSM) based on central composite design (CCD) experiments. A 23 five level CCD experiments with central and axial points was used to develop a statistical model for the optimization of process variables such as incubation temperature (25–45°) X1, pH (5.0–7.0) X2 and fermentation time (24–120 h) X3. Data obtained from RSM on ethanol production were subjected to the analysis of variance (ANOVA) and analyzed using a second order polynomial equation and contour plots were used to study the interactions among three relevant variables of the fermentation process. The fermentation experiments were carried out using an online monitored modular fermenter 2L capacity. The processing parameters setup for reaching a maximum response for ethanol production was obtained when applying the optimum values for temperature (32°C), pH (5.6) and fermentation time (110 h). Maximum ethanol concentration (3.36 g/l) was obtained from 50 g/l pretreated sugarcane bagasse at the optimized process conditions in aerobic batch fermentation. Kinetic models such as Monod, Modified Logistic model, Modified Logistic incorporated Leudeking – Piret model and Modified Logistic incorporated Modified Leudeking – Piret model have been evaluated and the constants were predicted.
Keywords: Sugarcane bagasse, ethanol, optimization, Pachysolen tannophilus.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23074721 Pin type Clamping Attachment for Remote Setup of Machining Process
Authors: Afzeri, R. Muhida, Darmawan, A. N. Berahim
Abstract:
Sharing the manufacturing facility through remote operation and monitoring of a machining process is challenge for effective use the production facility. Several automation tools in term of hardware and software are necessary for successfully remote operation of a machine. This paper presents a prototype of workpiece holding attachment for remote operation of milling process by self configuration the workpiece setup. The prototype is designed with mechanism to reorient the work surface into machining spindle direction with high positioning accuracy. Variety of parts geometry is hold by attachment to perform single setup machining. Pin type with array pattern additionally clamps the workpiece surface from two opposite directions for increasing the machining rigidity. Optimum pins configuration for conforming the workpiece geometry with minimum deformation is determined through hybrid algorithms, Genetic Algorithms (GA) and Particle Swarm Optimization (PSO). Prototype with intelligent optimization technique enables to hold several variety of workpiece geometry which is suitable for machining low of repetitive production in remote operation.Keywords: Optimization, Remote machining, GeneticAlgorithms, Machining Fixture.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26414720 Destination of the Solid Waste Generated at the Agricultural Products Wholesale Market in Brazil
Authors: C de Almeida, I. M. Dal Fabbro
Abstract:
The Brazilian Agricultural Products Wholesale Market fits well as example of residues generating system, reaching 750 metric tons per month of total residues, from which 600 metric tons are organic material and 150 metric tons are recyclable materials. Organic material is basically composed of fruit, vegetables and flowers leftovers from the products commercialization. The recyclable compounds are generate from packing material employed in the commercialization process. This research work devoted efforts in carrying quantitative analysis of the residues generated in the agricultural enterprise at its final destination. Data survey followed the directions implemented by the Residues Management Program issued by the agricultural enterprise. It was noticed from that analysis the necessity of changing the logistics applied to the recyclable material collecting process. However, composting process was elected as the organic compounds destination which is considered adequate for a material composed of significant percentage of organic matter far higher than wood, cardboard and plastics contents.
Keywords: Composting, environment, recycling, solid waste.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20204719 Specialized Reduced Models of Dynamic Flows in 2-Stroke Engines
Authors: S. Cagin, X. Fischer, E. Delacourt, N. Bourabaa, C. Morin, D. Coutellier, B. Carré, S. Loumé
Abstract:
The complexity of scavenging by ports and its impact on engine efficiency create the need to understand and to model it as realistically as possible. However, there are few empirical scavenging models and these are highly specialized. In a design optimization process, they appear very restricted and their field of use is limited. This paper presents a comparison of two methods to establish and reduce a model of the scavenging process in 2-stroke diesel engines. To solve the lack of scavenging models, a CFD model has been developed and is used as the referent case. However, its large size requires a reduction. Two techniques have been tested depending on their fields of application: The NTF method and neural networks. They both appear highly appropriate drastically reducing the model’s size (over 90% reduction) with a low relative error rate (under 10%). Furthermore, each method produces a reduced model which can be used in distinct specialized fields of application: the distribution of a quantity (mass fraction for example) in the cylinder at each time step (pseudo-dynamic model) or the qualification of scavenging at the end of the process (pseudo-static model).
Keywords: Diesel engine, Design optimization, Model reduction, Neural network, NTF algorithm, Scavenging.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13314718 The Synthetic T2 Quality Control Chart and its Multi-Objective Optimization
Authors: Francisco Aparisi, Marco A. de Luna
Abstract:
In some real applications of Statistical Process Control it is necessary to design a control chart to not detect small process shifts, but keeping a good performance to detect moderate and large shifts in the quality. In this work we develop a new quality control chart, the synthetic T2 control chart, that can be designed to cope with this objective. A multi-objective optimization is carried out employing Genetic Algorithms, finding the Pareto-optimal front of non-dominated solutions for this optimization problem.Keywords: Multi-objective optimization, Quality Control, SPC, Synthetic T2 control chart.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15604717 The Sign in the Communication Process
Authors: S. Pesina, T. Solonchak
Abstract:
In the process of information transmission (concept verbalization) we deal mostly with the substance (contents), and then pay attention to the form. Recalling events from the remote past, often we cannot exactly reproduce specific heard or pronounced words, as well as the syntactic structures. We remember events, feelings, images; we recall the general contents of the discourse. The thought gets a specific language form only during the concept verbalization phase. With minimum time for pondering, depending on the language competence level, the grammar and syntactic shaping often occurs automatically with the use of famous models and stereotypes. This means that the language form adapts itself to the consciousness, and not vice versa.
Keywords: Lexical eidos, phenomenology, noema, polysemantic word, semantic core.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19424716 A Contribution to 3D Modeling of Manufacturing Tolerance Optimization
Authors: F. Sebaa, A. Cheikh, M. Rahou
Abstract:
The study of the generated defects on manufactured parts shows the difficulty to maintain parts in their positions during the machining process and to estimate them during the pre-process plan. This work presents a contribution to the development of 3D models for the optimization of the manufacturing tolerances. An experimental study allows the measurement of the defects of part positioning for the determination of ε and the choice of an optimal setup of the part. An approach of 3D tolerance based on the small displacements method permits the determination of the manufacturing errors upstream. A developed tool, allows an automatic generation of the tolerance intervals along the three axes.Keywords: Manufacturing tolerances, 3D modeling, optimization, errors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15754715 Placement of Implants in Palatum of a Teenager without Maxillary Incisor Teeth
Authors: Luan Mavriqi, Ilma Robo, Emin Kuzimi, Egresa Baca
Abstract:
The process of skeletal growth in an adolescent significantly affects the displacement of implants placed in the palatine suture. The problems caused by this process have an impact on the dental function and aesthetics of the affected data. If fixed prostheses are placed based on implants, the whole structure would impede maxillary growth. This is the significant difference between the maxilla and the mandible, as the lower jaw has no growth process that affects the movement of the implants or the latter to inhibit the growth of the jaw. In a teenager patient an accident occurred accompanied by loss of maxillary central incisors. The main complaint of patients is aesthetics and phonetics. Dental history of patients refers to the presence of a Maryland bridge that was accompanied by dissatisfaction on the part of the patient. Implant placement is not indicated as jaw augmentation may lead to displacement of the implant. The treatment plan includes the placement of implants in the palatum where this bone thickness allows as a procedure.in this article only the first stage of treatment is presented. Implant treatment is ongoing, will be followed by the second phase of treatment when the patient has reached the age of 18 years.
Keywords: Implants, palatum, adolescent, primary incisor teeth.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3984714 Logistical Optimization of Nuclear Waste Flows during Decommissioning
Authors: G. Dottavio, M. F. Andrade, F. Renard, V. Cheutet, A.-L. L. S. Vercraene, P. Hoang, S. Briet, R. Dachicourt, Y. Baizet
Abstract:
An important number of technological equipment and high-skilled workers over long periods of time have to be mobilized during nuclear decommissioning processes. The related operations generate complex flows of waste and high inventory levels, associated to information flows of heterogeneous types. Taking into account that more than 10 decommissioning operations are on-going in France and about 50 are expected toward 2025: A big challenge is addressed today. The management of decommissioning and dismantling of nuclear installations represents an important part of the nuclear-based energy lifecycle, since it has an environmental impact as well as an important influence on the electricity cost and therefore the price for end-users. Bringing new technologies and new solutions into decommissioning methodologies is thus mandatory to improve the quality, cost and delay efficiency of these operations. The purpose of our project is to improve decommissioning management efficiency by developing a decision-support framework dedicated to plan nuclear facility decommissioning operations and to optimize waste evacuation by means of a logistic approach. The target is to create an easy-to-handle tool capable of i) predicting waste flows and proposing the best decommissioning logistics scenario and ii) managing information during all the steps of the process and following the progress: planning, resources, delays, authorizations, saturation zones, waste volume, etc. In this article we present our results from waste nuclear flows simulation during decommissioning process, including discrete-event simulation supported by FLEXSIM 3-D software. This approach was successfully tested and our works confirms its ability to improve this type of industrial process by identifying the critical points of the chain and optimizing it by identifying improvement actions. This type of simulation, executed before the start of the process operations on the basis of a first conception, allow ‘what-if’ process evaluation and help to ensure quality of the process in an uncertain context. The simulation of nuclear waste flows before evacuation from the site will help reducing the cost and duration of the decommissioning process by optimizing the planning and the use of resources, transitional storage and expensive radioactive waste containers. Additional benefits are expected for the governance system of the waste evacuation since it will enable a shared responsibility of the waste flows.
Keywords: Nuclear decommissioning, logistical optimization, decision-support framework, waste management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15564713 An Approach for Coagulant Dosage Optimization Using Soft Jar Test: A Case Study of Bangkhen Water Treatment Plant
Authors: Ninlawat Phuangchoke, Waraporn Viyanon, Setta Sasananan
Abstract:
The most important process of the water treatment plant process is coagulation, which uses alum and poly aluminum chloride (PACL). Therefore, determining the dosage of alum and PACL is the most important factor to be prescribed. This research applies an artificial neural network (ANN), which uses the Levenberg–Marquardt algorithm to create a mathematical model (Soft Jar Test) for chemical dose prediction, as used for coagulation, such as alum and PACL, with input data consisting of turbidity, pH, alkalinity, conductivity, and, oxygen consumption (OC) of the Bangkhen Water Treatment Plant (BKWTP), under the authority of the Metropolitan Waterworks Authority of Thailand. The data were collected from 1 January 2019 to 31 December 2019 in order to cover the changing seasons of Thailand. The input data of ANN are divided into three groups: training set, test set, and validation set. The coefficient of determination and the mean absolute errors of the alum model are 0.73, 3.18 and the PACL model are 0.59, 3.21, respectively.
Keywords: Soft jar test, jar test, water treatment plant process, artificial neural network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6674712 Slow, Wet and Catalytic Pyrolysis of Fowl Manure
Authors: Renzo Carta, Mario Cruccu, Francesco Desogus
Abstract:
This work presents the experimental results obtained at a pilot plant which works with a slow, wet and catalytic pyrolysis process of dry fowl manure. This kind of process mainly consists in the cracking of the organic matrix and in the following reaction of carbon with water, which is either already contained in the organic feed or added, to produce carbon monoxide and hydrogen. Reactions are conducted in a rotating reactor maintained at a temperature of 500°C; the required amount of water is about 30% of the dry organic feed. This operation yields a gas containing about 59% (on a volume basis) of hydrogen, 17% of carbon monoxide and other products such as light hydrocarbons (methane, ethane, propane) and carbon monoxide in lesser amounts. The gas coming from the reactor can be used to produce not only electricity, through internal combustion engines, but also heat, through direct combustion in industrial boilers. Furthermore, as the produced gas is devoid of both solid particles and pollutant species (such as dioxins and furans), the process (in this case applied to fowl manure) can be considered as an optimal way for the disposal and the contemporary energetic valorization of organic materials, in such a way that is not damaging to the environment.Keywords: Brushwood, fowl manure, kenaf, pilot plant, pyrolysis, pyrolysis gas.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23694711 Interaction of Elevated Carbon Dioxide and Temperature on Strawberry (Fragaria × ananassa) Growth and Fruit Yield
Authors: Himali N. Balasooriya, Kithsiri B. Dassanayake, Saman Seneweera, Said Ajlouni
Abstract:
Increase in atmospheric CO2 concentration [CO2] and ambient temperature associated with changing climatic conditions will have significant impacts on agriculture crop productivity and quality. Independent effects of the above two environmental variables on the growth, yield and quality of strawberry were well documented. Higher temperatures over the optimum range (20-25ºC) lead to crop failures, while elevated [CO2] stimulated plant growth and yield but compromised the physical quality of fruits. However, there is very limited understanding of the interaction between these variables on the plant growth, yield and quality. Therefore, this study was designed to investigate the interactive effect of high temperature and elevated [CO2] on growth, yield and quality of strawberries. Strawberry cultivars ‘Albion’ and ‘San Andreas’ were grown under six different combinations of two temperatures (25 and 30ºC) and three [CO2] (400, 650 and 950 µmol mol-1) in controlled-environmental growth chambers. Plant growth measurements such as plant height, canopy area, number of flowers, and fruit yield were measured during phonological development. Photosynthesis and transpiration, the ratio of intercellular to atmospheric [CO2] (Ci/Ca) were measured to estimate the physiological adjustment to climate stress. The impact of temperature and [CO2] interaction on growth and yield of strawberry was significant (p < 0.05). Across both cultivars, highest fruit yields were observed at 650 µmol mol-1 [CO2], which was particularly clear at 25°C. The fruit yield gradually decreased at 30°C under all the treatment combinations. However, photosynthesis rates were highest at 650 µmol mol-1 [CO2] but no increment was found at 900 µmol mol-1 [CO2]. Interestingly, Ci/Ca ratio increased with increasing atmospheric [CO2] which was predominant at high temperature. Similarly, fruit yield was substantially reduced at high [CO2] under high temperature. Our findings suggest that increased Ci/Ca ratio at high temperature is likely reduces the photosynthesis and thus yield response to elevated [CO2].
Keywords: Atmospheric [CO2], fruit yield, strawberry, temperature.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9864710 Integration Methods and Processes of Product Design and Flexible Production for Direct Production within the iCIM 3000 System
Authors: Roman Ružarovský, Radovan Holubek, Daynier Rolando Delgado Sobrino
Abstract:
Currently is characterized production engineering together with the integration of industrial automation and robotics such very quick view of to manufacture the products. The production range is continuously changing, expanding and producers have to be flexible in this regard. It means that need to offer production possibilities, which can respond to the quick change. Engineering product development is focused on supporting CAD software, such systems are mainly used for product design. That manufacturers are competitive, it should be kept procured machines made available capable of responding to output flexibility. In response to that problem is the development of flexible manufacturing systems, consisting of various automated systems. The integration of flexible manufacturing systems and subunits together with product design and of engineering is a possible solution for this issue. Integration is possible through the implementation of CIM systems. Such a solution and finding a hyphen between CAD and procurement system ICIM 3000 from Festo Co. is engaged in the research project and this contribution. This can be designed the products in CAD systems and watch the manufacturing process from order to shipping by the development of methods and processes of integration, This can be modeled in CAD systems products and watch the manufacturing process from order to shipping to develop methods and processes of integration, which will improve support for product design parameters by monitoring of the production process, by creating of programs for production using the CAD and therefore accelerates the a total of process from design to implementation.
Keywords: CAD- Computer Aided Design, CAM- Computer Aided Manufacturing, CIM- Computer integrated manufacturing, iCIM 3000, integration, direct production from CAD.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23014709 Retrieving Extended High Dynamic Range from Digital Negative Image - An Experiment on Architectural Photo Imaging
Authors: See Zi Siang, Khairul Hazrin Hashim, Harold Thwaites, Lee Xia Sheng, Ooi Wooi Har
Abstract:
The paper explores the development of an optimization of method and apparatus for retrieving extended high dynamic range from digital negative image. Architectural photo imaging can benefit from high dynamic range imaging (HDRI) technique for preserving and presenting sufficient luminance in the shadow and highlight clipping image areas. The HDRI technique that requires multiple exposure images as the source of HDRI rendering may not be effective in terms of time efficiency during the acquisition process and post-processing stage, considering it has numerous potential imaging variables and technical limitations during the multiple exposure process. This paper explores an experimental method and apparatus that aims to expand the dynamic range from digital negative image in HDRI environment. The method and apparatus explored is based on a single source of RAW image acquisition for the use of HDRI post-processing. It will cater the optimization in order to avoid and minimize the conventional HDRI photographic errors caused by different physical conditions during the photographing process and the misalignment of multiple exposed image sequences. The study observes the characteristics and capabilities of RAW image format as digital negative used for the retrieval of extended high dynamic range process in HDRI environment.
Keywords: High Dynamic Range Image, Photography Workflow Optimization, Digital Negative Image, Architectural Image
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16184708 An Automatic Model Transformation Methodology Based on Semantic and Syntactic Comparisons and the Granularity Issue Involved
Authors: Tiexin Wang, Sebastien Truptil, Frederick Benaben
Abstract:
Model transformation, as a pivotal aspect of Modeldriven engineering, attracts more and more attentions both from researchers and practitioners. Many domains (enterprise engineering, software engineering, knowledge engineering, etc.) use model transformation principles and practices to serve to their domain specific problems; furthermore, model transformation could also be used to fulfill the gap between different domains: by sharing and exchanging knowledge. Since model transformation has been widely used, there comes new requirement on it: effectively and efficiently define the transformation process and reduce manual effort that involved in. This paper presents an automatic model transformation methodology based on semantic and syntactic comparisons, and focuses particularly on granularity issue that existed in transformation process. Comparing to the traditional model transformation methodologies, this methodology serves to a general purpose: crossdomain methodology. Semantic and syntactic checking measurements are combined into a refined transformation process, which solves the granularity issue. Moreover, semantic and syntactic comparisons are supported by software tool; manual effort is replaced in this way.Keywords: Automatic model transformation, granularity issue, model-driven engineering, semantic and syntactic comparisons.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21914707 Study of Compost Maturity during Humification Process using UV-Spectroscopy
Authors: N. Sanmanee, K. Panishkan, K. Obsuwan, S. Dharmvanij
Abstract:
The increments of aromatic structures are widely used to monitor the degree of humification. Compost derived from mix manures mixed with agricultural wastes was studied. The compost collected at day 0, 7, 14, 21, 28, 35, 49, 77, 91, 105, and 119 was divided into 3 stages, initial stage at day 0, thermophilic stage during day 1-48, and mature stage during day 49-119. The change of highest absorptions at wavelength range between 210-235 nm during day 0- 49 implied that small molecules such as nitrates and carboxylic occurred faster than the aromatic molecules that were found at wavelength around 280 nm. The ratio of electron-transfer band at wavelength 253 nm by the benzonoid band at wavelength 230 nm (E253/E230) also gradually increased during the fermenting period indicating the presence of O-containing functional groups. This was in agreement with the shift change from aliphatic to aromatic structures as shown by the relationship with C/N and H/C ratios (r = - 0.631 and -0.717, p< 0.05) since both were decreasing. Although the amounts of humic acid (HA) were not different much during the humification process, the UV spectral deconvolution showed better qualitative characteristics to help in determining the compost quality. From this study, the compost should be used at day 49 and should not be kept longer than 3 months otherwise the quality of HA would decline regardless of the amounts of HA that might be rising. This implied that other processes, such as mineralization had an influence on the humification process changing HA-s structure and its qualities.
Keywords: Compost maturity, UV spectroscopy, humification, humic acid
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16614706 Reduce of Fermentation Time in Composting Process by Using a Special Microbial Consortium
Authors: S.H. Mirdamadian, S.M. Khayam-Nekoui, H. Ghanavati
Abstract:
Composting is the process in which municipal solid waste (MSW) and other organic waste materials such as biosolids and manures are decomposed through the action of bacteria and other microorganisms into a stable granular material which, applied to land, as soil conditioner. Microorganisms, especially those that are able to degrade polymeric organic material have a key role in speed up this process. The aim of this study has been established to isolation of microorganisms with high ability to production extracellular enzymes for degradation of natural polymers that are exists in MSW for decreasing time of degradation phase. Our experimental study for isolation designed in two phases: in first phase we isolated degrading microorganism with selected media that consist a special natural polymer such as cellulose, starch, lipids and etc as sole source of carbon. In second phase we selected microorganism that had high degrading enzyme production with enzymatic assay for seed production. However, our findings in pilot scale have indicated that usage of this microbial consortium had high efficiency for decreasing degradation phase.Keywords: Biodegradation, Compost, Municipal Solid Waste, Waste Management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21744705 An Exploration of the Dimensions of Place-Making: A South African Case Study
Authors: W. J. Strydom, K. Puren
Abstract:
Place-making is viewed here as an empowering process in which people represent, improve and maintain their spatial (natural or built) environment. With the above-mentioned in mind, place-making is multi-dimensional and include a spatial dimension (including visual properties or the end product/plan), a procedural dimension during which (negotiation/discussion of ideas with all relevant stakeholders in terms of end product/plan) and a psychological dimension (inclusion of intrinsic values and meanings related to a place in the end product/plan). These three represent dimensions of place-making. The purpose of this paper is to explore these dimensions of place-making in a case study of a local community in Ikageng, Potchefstroom, North-West Province, South Africa. This case study represents an inclusive process that strives to empower a local community (forcefully relocated due to Apartheid legislation in South Africa). This case study focussed on the inclusion of participants in the decision-making process regarding their daily environment. By means of focus group discussions and a collaborative design workshop, data is generated and ultimately creates a linkage with the theoretical dimensions of place-making. This paper contributes to the field of spatial planning due to the exploration of the dimensions of place-making and the relevancy of this process on spatial planning (especially in a South African setting).
Keywords: Case study, place-making, spatial planning, spatial dimension, procedural dimension, psychological dimension.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17084704 Development of a Kinetic Model for the Photodegradation of 4-Chlorophenol using a XeBr Excilamp
Authors: M. Gomez, M. D. Murcia, E. Gomez, J. L. Gomez, N. Christofi
Abstract:
Excilamps are new UV sources with great potential for application in wastewater treatment. In the present work, a XeBr excilamp emitting radiation at 283 nm has been used for the photodegradation of 4-chlorophenol within a range of concentrations from 50 to 500 mg L-1. Total removal of 4-chlorophenol was achieved for all concentrations assayed. The two main photoproduct intermediates formed along the photodegradation process, benzoquinone and hydroquinone, although not being completely removed, remain at very low residual concentrations. Such concentrations are insignificant compared to the 4-chlorophenol initial ones and non-toxic. In order to simulate the process and scaleup, a kinetic model has been developed and validated from the experimental data.Keywords: 4-chlorophenol, excilamps, kinetic model, photodegradation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13864703 Daemon- Based Distributed Deadlock Detection and Resolution
Authors: Z. RahimAlipour, A. T. Haghighat
Abstract:
detecting the deadlock is one of the important problems in distributed systems and different solutions have been proposed for it. Among the many deadlock detection algorithms, Edge-chasing has been the most widely used. In Edge-chasing algorithm, a special message called probe is made and sent along dependency edges. When the initiator of a probe receives the probe back the existence of a deadlock is revealed. But these algorithms are not problem-free. One of the problems associated with them is that they cannot detect some deadlocks and they even identify false deadlocks. A key point not mentioned in the literature is that when the process is waiting to obtain the required resources and its execution has been blocked, how it can actually respond to probe messages in the system. Also the question of 'which process should be victimized in order to achieve a better performance when multiple cycles exist within one single process in the system' has received little attention. In this paper, one of the basic concepts of the operating system - daemon - will be used to solve the problems mentioned. The proposed Algorithm becomes engaged in sending probe messages to the mandatory daemons and collects enough information to effectively identify and resolve multi-cycle deadlocks in distributed systems.Keywords: Distributed system, distributed deadlock detectionand resolution, daemon, false deadlock.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1937