Search results for: Markov Decision Process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6479

Search results for: Markov Decision Process

4859 Mathematical Modeling to Predict Surface Roughness in CNC Milling

Authors: Ab. Rashid M.F.F., Gan S.Y., Muhammad N.Y.

Abstract:

Surface roughness (Ra) is one of the most important requirements in machining process. In order to obtain better surface roughness, the proper setting of cutting parameters is crucial before the process take place. This research presents the development of mathematical model for surface roughness prediction before milling process in order to evaluate the fitness of machining parameters; spindle speed, feed rate and depth of cut. 84 samples were run in this study by using FANUC CNC Milling α-Τ14ιE. Those samples were randomly divided into two data sets- the training sets (m=60) and testing sets(m=24). ANOVA analysis showed that at least one of the population regression coefficients was not zero. Multiple Regression Method was used to determine the correlation between a criterion variable and a combination of predictor variables. It was established that the surface roughness is most influenced by the feed rate. By using Multiple Regression Method equation, the average percentage deviation of the testing set was 9.8% and 9.7% for training data set. This showed that the statistical model could predict the surface roughness with about 90.2% accuracy of the testing data set and 90.3% accuracy of the training data set.

Keywords: Surface roughness, regression analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2125
4858 A Study of the Variables in the Optimisation of a Platinum Precipitation Process

Authors: Tebogo Phetla, Edison Muzenda, M Belaid

Abstract:

This study investigated possible ways to improve the efficiency of the platinum precipitation process using ammonium chloride by reducing the platinum content reporting to the effluent. The ore treated consist of five platinum group metals namely, ruthenium, rhodium, iridium, platinum, palladium and a precious metal gold. Gold, ruthenium, rhodium and iridium were extracted prior the platinum precipitation process. Temperature, reducing agent, flow rate and potential difference were the variables controlled to determine the operation conditions for optimum platinum precipitation efficiency. Hydrogen peroxide was added as the oxidizing agent at the temperature of 85-90oC and potential difference of 700-850mV was the variable used to check the oxidizing state of platinum. The platinum was further purified at temperature between 60-65oC, potential difference above 700 mV, ammonium chloride of 200 l, and at these conditions the platinum content reporting to the effluent was reduced to less than 300ppm, resulting in optimum platinum precipitation efficiency and purity of 99.9%.

Keywords: Platinum Group Metals (PGM), Potential difference, Precipitation, Redox reactions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4784
4857 Process Development of Safe and Ready-to-eat Raw Oyster Meat by Irradiation Technology

Authors: Pattama Ratana-Arporn, Pongtep Wilaipun

Abstract:

White scar oyster (Crassostrea belcheri) is often eaten raw and being the leading vehicle for foodborne disease, especially Salmonella Weltevreden which exposed the prominent and most resistant to radiation. Gamma irradiation at a low dose of 1 kGy was enough to eliminate S. Weltevreden contaminated in oyster meat at a level up to 5 log CFU/g while it still retain the raw characteristics and equivalent sensory quality as the non-irradiated one. Process development of ready-to-eat chilled oyster meat was conducted by shucking the meat, individually packed in plastic bags, subjected to 1 kGy gamma radiation at chilled condition and then stored in 4oC refrigerated temperature. Microbiological determination showed the absence of S. Weltevreden (5 log CFU/g initial inoculated) along the whole storage time of 30 days. Sensory evaluation indicated the decreasing in sensory scores along storage time which determining the product shelf life to be 18 days compared to 15 days of nonirradiated one. The most advantage of developed process was to provide the safe raw oyster to consumers and in addition sensory quality retained and 3-day extension shelf life also exist.

Keywords: decontamination, food safety, irradiation, oyster, Salmonella Weltevreden

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1678
4856 Simultaneous Saccharification and Fermentation(SSF) of Sugarcane Bagasse - Kinetics and Modeling

Authors: E.Sasikumar, T.Viruthagiri

Abstract:

Simultaneous Saccharification and Fermentation (SSF) of sugarcane bagasse by cellulase and Pachysolen tannophilus MTCC *1077 were investigated in the present study. Important process variables for ethanol production form pretreated bagasse were optimized using Response Surface Methodology (RSM) based on central composite design (CCD) experiments. A 23 five level CCD experiments with central and axial points was used to develop a statistical model for the optimization of process variables such as incubation temperature (25–45°) X1, pH (5.0–7.0) X2 and fermentation time (24–120 h) X3. Data obtained from RSM on ethanol production were subjected to the analysis of variance (ANOVA) and analyzed using a second order polynomial equation and contour plots were used to study the interactions among three relevant variables of the fermentation process. The fermentation experiments were carried out using an online monitored modular fermenter 2L capacity. The processing parameters setup for reaching a maximum response for ethanol production was obtained when applying the optimum values for temperature (32°C), pH (5.6) and fermentation time (110 h). Maximum ethanol concentration (3.36 g/l) was obtained from 50 g/l pretreated sugarcane bagasse at the optimized process conditions in aerobic batch fermentation. Kinetic models such as Monod, Modified Logistic model, Modified Logistic incorporated Leudeking – Piret model and Modified Logistic incorporated Modified Leudeking – Piret model have been evaluated and the constants were predicted.

Keywords: Sugarcane bagasse, ethanol, optimization, Pachysolen tannophilus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2295
4855 Pin type Clamping Attachment for Remote Setup of Machining Process

Authors: Afzeri, R. Muhida, Darmawan, A. N. Berahim

Abstract:

Sharing the manufacturing facility through remote operation and monitoring of a machining process is challenge for effective use the production facility. Several automation tools in term of hardware and software are necessary for successfully remote operation of a machine. This paper presents a prototype of workpiece holding attachment for remote operation of milling process by self configuration the workpiece setup. The prototype is designed with mechanism to reorient the work surface into machining spindle direction with high positioning accuracy. Variety of parts geometry is hold by attachment to perform single setup machining. Pin type with array pattern additionally clamps the workpiece surface from two opposite directions for increasing the machining rigidity. Optimum pins configuration for conforming the workpiece geometry with minimum deformation is determined through hybrid algorithms, Genetic Algorithms (GA) and Particle Swarm Optimization (PSO). Prototype with intelligent optimization technique enables to hold several variety of workpiece geometry which is suitable for machining low of repetitive production in remote operation.

Keywords: Optimization, Remote machining, GeneticAlgorithms, Machining Fixture.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2632
4854 Destination of the Solid Waste Generated at the Agricultural Products Wholesale Market in Brazil

Authors: C de Almeida, I. M. Dal Fabbro

Abstract:

The Brazilian Agricultural Products Wholesale Market fits well as example of residues generating system, reaching 750 metric tons per month of total residues, from which 600 metric tons are organic material and 150 metric tons are recyclable materials. Organic material is basically composed of fruit, vegetables and flowers leftovers from the products commercialization. The recyclable compounds are generate from packing material employed in the commercialization process. This research work devoted efforts in carrying quantitative analysis of the residues generated in the agricultural enterprise at its final destination. Data survey followed the directions implemented by the Residues Management Program issued by the agricultural enterprise. It was noticed from that analysis the necessity of changing the logistics applied to the recyclable material collecting process. However, composting process was elected as the organic compounds destination which is considered adequate for a material composed of significant percentage of organic matter far higher than wood, cardboard and plastics contents.

Keywords: Composting, environment, recycling, solid waste.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2013
4853 Specialized Reduced Models of Dynamic Flows in 2-Stroke Engines

Authors: S. Cagin, X. Fischer, E. Delacourt, N. Bourabaa, C. Morin, D. Coutellier, B. Carré, S. Loumé

Abstract:

The complexity of scavenging by ports and its impact on engine efficiency create the need to understand and to model it as realistically as possible. However, there are few empirical scavenging models and these are highly specialized. In a design optimization process, they appear very restricted and their field of use is limited. This paper presents a comparison of two methods to establish and reduce a model of the scavenging process in 2-stroke diesel engines. To solve the lack of scavenging models, a CFD model has been developed and is used as the referent case. However, its large size requires a reduction. Two techniques have been tested depending on their fields of application: The NTF method and neural networks. They both appear highly appropriate drastically reducing the model’s size (over 90% reduction) with a low relative error rate (under 10%). Furthermore, each method produces a reduced model which can be used in distinct specialized fields of application: the distribution of a quantity (mass fraction for example) in the cylinder at each time step (pseudo-dynamic model) or the qualification of scavenging at the end of the process (pseudo-static model).

Keywords: Diesel engine, Design optimization, Model reduction, Neural network, NTF algorithm, Scavenging.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1323
4852 The Synthetic T2 Quality Control Chart and its Multi-Objective Optimization

Authors: Francisco Aparisi, Marco A. de Luna

Abstract:

In some real applications of Statistical Process Control it is necessary to design a control chart to not detect small process shifts, but keeping a good performance to detect moderate and large shifts in the quality. In this work we develop a new quality control chart, the synthetic T2 control chart, that can be designed to cope with this objective. A multi-objective optimization is carried out employing Genetic Algorithms, finding the Pareto-optimal front of non-dominated solutions for this optimization problem.

Keywords: Multi-objective optimization, Quality Control, SPC, Synthetic T2 control chart.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1548
4851 The Sign in the Communication Process

Authors: S. Pesina, T. Solonchak

Abstract:

In the process of information transmission (concept verbalization) we deal mostly with the substance (contents), and then pay attention to the form. Recalling events from the remote past, often we cannot exactly reproduce specific heard or pronounced words, as well as the syntactic structures. We remember events, feelings, images; we recall the general contents of the discourse. The thought gets a specific language form only during the concept verbalization phase. With minimum time for pondering, depending on the language competence level, the grammar and syntactic shaping often occurs automatically with the use of famous models and stereotypes. This means that the language form adapts itself to the consciousness, and not vice versa.

Keywords: Lexical eidos, phenomenology, noema, polysemantic word, semantic core.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1935
4850 A Contribution to 3D Modeling of Manufacturing Tolerance Optimization

Authors: F. Sebaa, A. Cheikh, M. Rahou

Abstract:

The study of the generated defects on manufactured parts shows the difficulty to maintain parts in their positions during the machining process and to estimate them during the pre-process plan. This work presents a contribution to the development of 3D models for the optimization of the manufacturing tolerances. An experimental study allows the measurement of the defects of part positioning for the determination of ε and the choice of an optimal setup of the part. An approach of 3D tolerance based on the small displacements method permits the determination of the manufacturing errors upstream. A developed tool, allows an automatic generation of the tolerance intervals along the three axes.

Keywords: Manufacturing tolerances, 3D modeling, optimization, errors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1567
4849 Genetic Programming: Principles, Applications and Opportunities for Hydrological Modelling

Authors: Oluwaseun K. Oyebode, Josiah A. Adeyemo

Abstract:

Hydrological modelling plays a crucial role in the planning and management of water resources, most especially in water stressed regions where the need to effectively manage the available water resources is of critical importance. However, due to the complex, nonlinear and dynamic behaviour of hydro-climatic interactions, achieving reliable modelling of water resource systems and accurate projection of hydrological parameters are extremely challenging. Although a significant number of modelling techniques (process-based and data-driven) have been developed and adopted in that regard, the field of hydrological modelling is still considered as one that has sluggishly progressed over the past decades. This is majorly as a result of the identification of some degree of uncertainty in the methodologies and results of techniques adopted. In recent times, evolutionary computation (EC) techniques have been developed and introduced in response to the search for efficient and reliable means of providing accurate solutions to hydrological related problems. This paper presents a comprehensive review of the underlying principles, methodological needs and applications of a promising evolutionary computation modelling technique – genetic programming (GP). It examines the specific characteristics of the technique which makes it suitable to solving hydrological modelling problems. It discusses the opportunities inherent in the application of GP in water related-studies such as rainfall estimation, rainfall-runoff modelling, streamflow forecasting, sediment transport modelling, water quality modelling and groundwater modelling among others. Furthermore, the means by which such opportunities could be harnessed in the near future are discussed. In all, a case for total embracement of GP and its variants in hydrological modelling studies is made so as to put in place strategies that would translate into achieving meaningful progress as it relates to modelling of water resource systems, and also positively influence decision-making by relevant stakeholders.

Keywords: Computational modelling, evolutionary algorithms, genetic programming, hydrological modelling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3322
4848 Placement of Implants in Palatum of a Teenager without Maxillary Incisor Teeth

Authors: Luan Mavriqi, Ilma Robo, Emin Kuzimi, Egresa Baca

Abstract:

The process of skeletal growth in an adolescent significantly affects the displacement of implants placed in the palatine suture. The problems caused by this process have an impact on the dental function and aesthetics of the affected data. If fixed prostheses are placed based on implants, the whole structure would impede maxillary growth. This is the significant difference between the maxilla and the mandible, as the lower jaw has no growth process that affects the movement of the implants or the latter to inhibit the growth of the jaw. In a teenager patient an accident occurred accompanied by loss of maxillary central incisors. The main complaint of patients is aesthetics and phonetics. Dental history of patients refers to the presence of a Maryland bridge that was accompanied by dissatisfaction on the part of the patient. Implant placement is not indicated as jaw augmentation may lead to displacement of the implant. The treatment plan includes the placement of implants in the palatum where this bone thickness allows as a procedure.in this article only the first stage of treatment is presented. Implant treatment is ongoing, will be followed by the second phase of treatment when the patient has reached the age of 18 years.

Keywords: Implants, palatum, adolescent, primary incisor teeth.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 385
4847 Project Selection by Using a Fuzzy TOPSIS Technique

Authors: M. Salehi, R. Tavakkoli-Moghaddam

Abstract:

Selection of a project among a set of possible alternatives is a difficult task that the decision maker (DM) has to face. In this paper, by using a fuzzy TOPSIS technique we propose a new method for a project selection problem. After reviewing four common methods of comparing investment alternatives (net present value, rate of return, benefit cost analysis and payback period) we use them as criteria in a TOPSIS technique. First we calculate the weight of each criterion by a pairwise comparison and then we utilize the improved TOPSIS assessment for the project selection.

Keywords: Fuzzy Theory, Pairwise Comparison, ProjectSelection, TOPSIS Technique.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2629
4846 Text Mining Technique for Data Mining Application

Authors: M. Govindarajan

Abstract:

Text Mining is around applying knowledge discovery techniques to unstructured text is termed knowledge discovery in text (KDT), or Text data mining or Text Mining. In decision tree approach is most useful in classification problem. With this technique, tree is constructed to model the classification process. There are two basic steps in the technique: building the tree and applying the tree to the database. This paper describes a proposed C5.0 classifier that performs rulesets, cross validation and boosting for original C5.0 in order to reduce the optimization of error ratio. The feasibility and the benefits of the proposed approach are demonstrated by means of medial data set like hypothyroid. It is shown that, the performance of a classifier on the training cases from which it was constructed gives a poor estimate by sampling or using a separate test file, either way, the classifier is evaluated on cases that were not used to build and evaluate the classifier are both are large. If the cases in hypothyroid.data and hypothyroid.test were to be shuffled and divided into a new 2772 case training set and a 1000 case test set, C5.0 might construct a different classifier with a lower or higher error rate on the test cases. An important feature of see5 is its ability to classifiers called rulesets. The ruleset has an error rate 0.5 % on the test cases. The standard errors of the means provide an estimate of the variability of results. One way to get a more reliable estimate of predictive is by f-fold –cross- validation. The error rate of a classifier produced from all the cases is estimated as the ratio of the total number of errors on the hold-out cases to the total number of cases. The Boost option with x trials instructs See5 to construct up to x classifiers in this manner. Trials over numerous datasets, large and small, show that on average 10-classifier boosting reduces the error rate for test cases by about 25%.

Keywords: C5.0, Error Ratio, text mining, training data, test data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2483
4845 An Approach for Coagulant Dosage Optimization Using Soft Jar Test: A Case Study of Bangkhen Water Treatment Plant

Authors: Ninlawat Phuangchoke, Waraporn Viyanon, Setta Sasananan

Abstract:

The most important process of the water treatment plant process is coagulation, which uses alum and poly aluminum chloride (PACL). Therefore, determining the dosage of alum and PACL is the most important factor to be prescribed. This research applies an artificial neural network (ANN), which uses the Levenberg–Marquardt algorithm to create a mathematical model (Soft Jar Test) for chemical dose prediction, as used for coagulation, such as alum and PACL, with input data consisting of turbidity, pH, alkalinity, conductivity, and, oxygen consumption (OC) of the Bangkhen Water Treatment Plant (BKWTP), under the authority of the Metropolitan Waterworks Authority of Thailand. The data were collected from 1 January 2019 to 31 December 2019 in order to cover the changing seasons of Thailand. The input data of ANN are divided into three groups: training set, test set, and validation set. The coefficient of determination and the mean absolute errors of the alum model are 0.73, 3.18 and the PACL model are 0.59, 3.21, respectively.

Keywords: Soft jar test, jar test, water treatment plant process, artificial neural network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 650
4844 Slow, Wet and Catalytic Pyrolysis of Fowl Manure

Authors: Renzo Carta, Mario Cruccu, Francesco Desogus

Abstract:

This work presents the experimental results obtained at a pilot plant which works with a slow, wet and catalytic pyrolysis process of dry fowl manure. This kind of process mainly consists in the cracking of the organic matrix and in the following reaction of carbon with water, which is either already contained in the organic feed or added, to produce carbon monoxide and hydrogen. Reactions are conducted in a rotating reactor maintained at a temperature of 500°C; the required amount of water is about 30% of the dry organic feed. This operation yields a gas containing about 59% (on a volume basis) of hydrogen, 17% of carbon monoxide and other products such as light hydrocarbons (methane, ethane, propane) and carbon monoxide in lesser amounts. The gas coming from the reactor can be used to produce not only electricity, through internal combustion engines, but also heat, through direct combustion in industrial boilers. Furthermore, as the produced gas is devoid of both solid particles and pollutant species (such as dioxins and furans), the process (in this case applied to fowl manure) can be considered as an optimal way for the disposal and the contemporary energetic valorization of organic materials, in such a way that is not damaging to the environment.

Keywords: Brushwood, fowl manure, kenaf, pilot plant, pyrolysis, pyrolysis gas.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2359
4843 Integration Methods and Processes of Product Design and Flexible Production for Direct Production within the iCIM 3000 System

Authors: Roman Ružarovský, Radovan Holubek, Daynier Rolando Delgado Sobrino

Abstract:

Currently is characterized production engineering together with the integration of industrial automation and robotics such very quick view of to manufacture the products. The production range is continuously changing, expanding and producers have to be flexible in this regard. It means that need to offer production possibilities, which can respond to the quick change. Engineering product development is focused on supporting CAD software, such systems are mainly used for product design. That manufacturers are competitive, it should be kept procured machines made available capable of responding to output flexibility. In response to that problem is the development of flexible manufacturing systems, consisting of various automated systems. The integration of flexible manufacturing systems and subunits together with product design and of engineering is a possible solution for this issue. Integration is possible through the implementation of CIM systems. Such a solution and finding a hyphen between CAD and procurement system ICIM 3000 from Festo Co. is engaged in the research project and this contribution. This can be designed the products in CAD systems and watch the manufacturing process from order to shipping by the development of methods and processes of integration, This can be modeled in CAD systems products and watch the manufacturing process from order to shipping to develop methods and processes of integration, which will improve support for product design parameters by monitoring of the production process, by creating of programs for production using the CAD and therefore accelerates the a total of process from design to implementation.

Keywords: CAD- Computer Aided Design, CAM- Computer Aided Manufacturing, CIM- Computer integrated manufacturing, iCIM 3000, integration, direct production from CAD.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2289
4842 Retrieving Extended High Dynamic Range from Digital Negative Image - An Experiment on Architectural Photo Imaging

Authors: See Zi Siang, Khairul Hazrin Hashim, Harold Thwaites, Lee Xia Sheng, Ooi Wooi Har

Abstract:

The paper explores the development of an optimization of method and apparatus for retrieving extended high dynamic range from digital negative image. Architectural photo imaging can benefit from high dynamic range imaging (HDRI) technique for preserving and presenting sufficient luminance in the shadow and highlight clipping image areas. The HDRI technique that requires multiple exposure images as the source of HDRI rendering may not be effective in terms of time efficiency during the acquisition process and post-processing stage, considering it has numerous potential imaging variables and technical limitations during the multiple exposure process. This paper explores an experimental method and apparatus that aims to expand the dynamic range from digital negative image in HDRI environment. The method and apparatus explored is based on a single source of RAW image acquisition for the use of HDRI post-processing. It will cater the optimization in order to avoid and minimize the conventional HDRI photographic errors caused by different physical conditions during the photographing process and the misalignment of multiple exposed image sequences. The study observes the characteristics and capabilities of RAW image format as digital negative used for the retrieval of extended high dynamic range process in HDRI environment.

Keywords: High Dynamic Range Image, Photography Workflow Optimization, Digital Negative Image, Architectural Image

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1610
4841 An Automatic Model Transformation Methodology Based on Semantic and Syntactic Comparisons and the Granularity Issue Involved

Authors: Tiexin Wang, Sebastien Truptil, Frederick Benaben

Abstract:

Model transformation, as a pivotal aspect of Modeldriven engineering, attracts more and more attentions both from researchers and practitioners. Many domains (enterprise engineering, software engineering, knowledge engineering, etc.) use model transformation principles and practices to serve to their domain specific problems; furthermore, model transformation could also be used to fulfill the gap between different domains: by sharing and exchanging knowledge. Since model transformation has been widely used, there comes new requirement on it: effectively and efficiently define the transformation process and reduce manual effort that involved in. This paper presents an automatic model transformation methodology based on semantic and syntactic comparisons, and focuses particularly on granularity issue that existed in transformation process. Comparing to the traditional model transformation methodologies, this methodology serves to a general purpose: crossdomain methodology. Semantic and syntactic checking measurements are combined into a refined transformation process, which solves the granularity issue. Moreover, semantic and syntactic comparisons are supported by software tool; manual effort is replaced in this way.

Keywords: Automatic model transformation, granularity issue, model-driven engineering, semantic and syntactic comparisons.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2187
4840 Study of Compost Maturity during Humification Process using UV-Spectroscopy

Authors: N. Sanmanee, K. Panishkan, K. Obsuwan, S. Dharmvanij

Abstract:

The increments of aromatic structures are widely used to monitor the degree of humification. Compost derived from mix manures mixed with agricultural wastes was studied. The compost collected at day 0, 7, 14, 21, 28, 35, 49, 77, 91, 105, and 119 was divided into 3 stages, initial stage at day 0, thermophilic stage during day 1-48, and mature stage during day 49-119. The change of highest absorptions at wavelength range between 210-235 nm during day 0- 49 implied that small molecules such as nitrates and carboxylic occurred faster than the aromatic molecules that were found at wavelength around 280 nm. The ratio of electron-transfer band at wavelength 253 nm by the benzonoid band at wavelength 230 nm (E253/E230) also gradually increased during the fermenting period indicating the presence of O-containing functional groups. This was in agreement with the shift change from aliphatic to aromatic structures as shown by the relationship with C/N and H/C ratios (r = - 0.631 and -0.717, p< 0.05) since both were decreasing. Although the amounts of humic acid (HA) were not different much during the humification process, the UV spectral deconvolution showed better qualitative characteristics to help in determining the compost quality. From this study, the compost should be used at day 49 and should not be kept longer than 3 months otherwise the quality of HA would decline regardless of the amounts of HA that might be rising. This implied that other processes, such as mineralization had an influence on the humification process changing HA-s structure and its qualities.

Keywords: Compost maturity, UV spectroscopy, humification, humic acid

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1653
4839 Reduce of Fermentation Time in Composting Process by Using a Special Microbial Consortium

Authors: S.H. Mirdamadian, S.M. Khayam-Nekoui, H. Ghanavati

Abstract:

Composting is the process in which municipal solid waste (MSW) and other organic waste materials such as biosolids and manures are decomposed through the action of bacteria and other microorganisms into a stable granular material which, applied to land, as soil conditioner. Microorganisms, especially those that are able to degrade polymeric organic material have a key role in speed up this process. The aim of this study has been established to isolation of microorganisms with high ability to production extracellular enzymes for degradation of natural polymers that are exists in MSW for decreasing time of degradation phase. Our experimental study for isolation designed in two phases: in first phase we isolated degrading microorganism with selected media that consist a special natural polymer such as cellulose, starch, lipids and etc as sole source of carbon. In second phase we selected microorganism that had high degrading enzyme production with enzymatic assay for seed production. However, our findings in pilot scale have indicated that usage of this microbial consortium had high efficiency for decreasing degradation phase.

Keywords: Biodegradation, Compost, Municipal Solid Waste, Waste Management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2162
4838 Development of a Kinetic Model for the Photodegradation of 4-Chlorophenol using a XeBr Excilamp

Authors: M. Gomez, M. D. Murcia, E. Gomez, J. L. Gomez, N. Christofi

Abstract:

Excilamps are new UV sources with great potential for application in wastewater treatment. In the present work, a XeBr excilamp emitting radiation at 283 nm has been used for the photodegradation of 4-chlorophenol within a range of concentrations from 50 to 500 mg L-1. Total removal of 4-chlorophenol was achieved for all concentrations assayed. The two main photoproduct intermediates formed along the photodegradation process, benzoquinone and hydroquinone, although not being completely removed, remain at very low residual concentrations. Such concentrations are insignificant compared to the 4-chlorophenol initial ones and non-toxic. In order to simulate the process and scaleup, a kinetic model has been developed and validated from the experimental data.

Keywords: 4-chlorophenol, excilamps, kinetic model, photodegradation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1379
4837 Daemon- Based Distributed Deadlock Detection and Resolution

Authors: Z. RahimAlipour, A. T. Haghighat

Abstract:

detecting the deadlock is one of the important problems in distributed systems and different solutions have been proposed for it. Among the many deadlock detection algorithms, Edge-chasing has been the most widely used. In Edge-chasing algorithm, a special message called probe is made and sent along dependency edges. When the initiator of a probe receives the probe back the existence of a deadlock is revealed. But these algorithms are not problem-free. One of the problems associated with them is that they cannot detect some deadlocks and they even identify false deadlocks. A key point not mentioned in the literature is that when the process is waiting to obtain the required resources and its execution has been blocked, how it can actually respond to probe messages in the system. Also the question of 'which process should be victimized in order to achieve a better performance when multiple cycles exist within one single process in the system' has received little attention. In this paper, one of the basic concepts of the operating system - daemon - will be used to solve the problems mentioned. The proposed Algorithm becomes engaged in sending probe messages to the mandatory daemons and collects enough information to effectively identify and resolve multi-cycle deadlocks in distributed systems.

Keywords: Distributed system, distributed deadlock detectionand resolution, daemon, false deadlock.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1930
4836 Intraoperative ICG-NIR Fluorescence Angiography Visualization of Intestinal Perfusion in Primary Pull-Through for Hirschsprung Disease

Authors: Mohammad Emran, Colton Wayne, Shannon M Koehler, P. Stephen Almond, Haroon Patel

Abstract:

Purpose: Assessment of anastomotic perfusion in Hirschsprung disease using Indocyanine Green (ICG)-near-infrared (NIR) fluorescence angiography. Introduction: Anastomotic stricture and leak are well-known complications of Hirschsprung pull-through procedures. Complications are due to tension, infection, and/or poor perfusion. While a surgeon can visually determine and control the amount of tension and contamination, assessment of perfusion is subject to surgeon determination. Intraoperative use of ICG-NIR enhances this decision-making process by illustrating perfusion intensity and adequacy in the pulled-through bowel segment. This technique, proven to reduce anastomotic stricture and leak in adults, has not been studied in children to our knowledge. ICG, an FDA approved, nontoxic, non-immunogenic, intravascular (IV) dye, has been used in adults and children for over 60 years, with few side effects. ICG-NIR was used in this report to demonstrate the adequacy of perfusion during transanal pullthrough for Hirschsprung’s disease. Method: 8 patients with Hirschsprung disease were evaluated with ICG-NIR technology. Levels of affected area ranged from sigmoid to total colonic Hirschsprung disease. After leveling, but prior to anastomosis, ICG was administered at 1.25 mg (< 2 mg/kg) and perfusion visualized using an NIR camera, before and during anastomosis. Video and photo imaging was performed and perfusion of the bowel was compared to surrounding tissues. This showed the degree of perfusion and demarcation of perfused and non-perfused bowel. The anastomosis was completed uneventfully and the patients all did well. Results: There were no complications of stricture or leak. 5 of 8 patients (62.5%) had modification of the plan based on ICG-NIR imaging. Conclusion: Technologies that enhance surgeons’ ability to visualize bowel perfusion prior to anastomosis in Hirschsprung’s patients may help reduce post-operative complications. Further studies are needed to assess the potential benefits.

Keywords: Colonic anastomosis, fluorescence angiography, Hirschsprung disease, pediatric surgery, SPY, ICG, NIR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 661
4835 Minimization Entropic Applied to Rotary Dryers to Reduce the Energy Consumption

Authors: I. O. Nascimento, J. T. Manzi

Abstract:

The drying process is an important operation in the chemical industry and it is widely used in the food, grain industry and fertilizer industry. However, for demanding a considerable consumption of energy, such a process requires a deep energetic analysis in order to reduce operating costs. This paper deals with thermodynamic optimization applied to rotary dryers based on the entropy production minimization, aiming at to reduce the energy consumption. To do this, the mass, energy and entropy balance was used for developing a relationship that represents the rate of entropy production. The use of the Second Law of Thermodynamics is essential because it takes into account constraints of nature. Since the entropy production rate is minimized, optimals conditions of operations can be established and the process can obtain a substantial gain in energy saving. The minimization strategy had been led using classical methods such as Lagrange multipliers and implemented in the MATLAB platform. As expected, the preliminary results reveal a significant energy saving by the application of the optimal parameters found by the procedure of the entropy minimization It is important to say that this method has shown easy implementation and low cost.

Keywords: Drying, entropy minimization, modeling dryers, thermodynamic optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1425
4834 Daily Site Risks Associated with Construction Projects and On-spot Corrective Measurements: Case Study of Revamping Projects in Kuwait Oil Company Fields Area

Authors: Yousef S. Al-Othman

Abstract:

The growth and expansion of the industrial facilities comes proportional to the market increasing demand of products and services. Furthermore, raw material producers such as oil companies usually undergo massive revamping projects to maintain a synchronized supply. These revamping projects are usually delivered through challenging construction projects held and associated with daily site risks related to the construction process. Henceforth, a case study related to these risks and corresponding on-spot corrective measurements has been made on a certain number of construction project contractors at Kuwait Oil Company (KOC) to derive the benefits and overall effectiveness of the on-spot corrective measurements during the construction phase of a project, and how would the same help in avoiding major incidents, ensuring a smooth, cost effective and on time delivery of the project. Findings of this case study shall have an added value to the overall risk management process by minimizing the daily site risks that may affect the project lead time, resulting in an undisturbed on-site construction process.

Keywords: Oil and gas, risk management, construction projects, project lead time.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 877
4833 Harnessing the Potential of Renewable Energy Sources to Reduce Fossil Energy Consumption in the Wastewater Treatment Process

Authors: Hen Friman

Abstract:

Various categories of aqueous solutions are discharged within residential, institutional, commercial, and industrial structures. To safeguard public health and preserve the environment, it is imperative to subject wastewater to treatment processes that eliminate pathogens (such as bacteria and viruses), nutrients (such as nitrogen and phosphorus), and other compounds. Failure to address untreated sewage accumulation can result in an array of adverse consequences. Israel exemplifies a special case in wastewater management. Appropriate wastewater treatment significantly benefits sectors such as agriculture, tourism, horticulture, and industry. Nevertheless, untreated sewage in settlements lacking proper sewage collection or transportation networks remains an ongoing and substantial threat. Notably, the process of wastewater treatment entails substantial energy consumption. Consequently, this study explores the integration of solar energy as a renewable power source within the wastewater treatment framework. By incorporating renewable energy sources into the process, costs can be minimized, and decentralized facilities can be established even in areas lacking adequate infrastructure for traditional treatment methods.

Keywords: Renewable energy, solar energy, decentralized facilities, wastewater treatment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 113
4832 Detecting the Edge of Multiple Images in Parallel

Authors: Prakash K. Aithal, U. Dinesh Acharya, Rajesh Gopakumar

Abstract:

Edge is variation of brightness in an image. Edge detection is useful in many application areas such as finding forests, rivers from a satellite image, detecting broken bone in a medical image etc. The paper discusses about finding edge of multiple aerial images in parallel. The proposed work tested on 38 images 37 colored and one monochrome image. The time taken to process N images in parallel is equivalent to time taken to process 1 image in sequential. Message Passing Interface (MPI) and Open Computing Language (OpenCL) is used to achieve task and pixel level parallelism respectively.

Keywords: Edge detection, multicore, GPU, openCL, MPI.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2334
4831 Optimization of Surface Finish in Milling Operation Using Live Tooling via Taguchi Method

Authors: Harish Kumar Ponnappan, Joseph C. Chen

Abstract:

The main objective of this research is to optimize the surface roughness of a milling operation on AISI 1018 steel using live tooling on a HAAS ST-20 lathe. In this study, Taguchi analysis is used to optimize the milling process by investigating the effect of different machining parameters on surface roughness. The L9 orthogonal array is designed with four controllable factors with three different levels each and an uncontrollable factor, resulting in 18 experimental runs. The optimal parameters determined from Taguchi analysis were feed rate – 76.2 mm/min, spindle speed 1150 rpm, depth of cut – 0.762 mm and 2-flute TiN coated high-speed steel as tool material. The process capability Cp and process capability index Cpk values were improved from 0.62 and -0.44 to 1.39 and 1.24 respectively. The average surface roughness values from the confirmation runs were 1.30 µ, decreasing the defect rate from 87.72% to 0.01%. The purpose of this study is to efficiently utilize the Taguchi design to optimize the surface roughness in a milling operation using live tooling.

Keywords: Live tooling, surface roughness, Taguchi analysis, Computer Numerical Control (CNC) milling operation, CNC turning operation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 764
4830 Bioethanol: Indonesian Macro-Algae as a Renewable Feedstock for Liquid Fuel

Authors: T. Poespowati, E. Marsyahyo, R. Kartika-Dewi

Abstract:

This experimental study aims at studying the conversion of macro-algae into bioethanol under several steps of procedure: preparation, pre-treatment, fermentation, and distillation. The main objective of this work was to investigate the role of buffer’s type as a stabiliser of pH level and fermentation time on the yield of ethanol. For this purpose, experiments were carried out on biomass macro-algae to de-couple the pre-treatment and fermentation processes from those associated with distillation process. β- glucosidase was used as cellulose decomposer during hydrolysis step and yeast was used during fermentation process. The species of macro-algae utilised as energy feedstock was Ulva lactuca and it was harvested from southern coast of Central of Java Island – Indonesia. Experiments were conducted in a simple fermenter over a different buffer: citrate buffer and acetic buffer, and over a range of fermentation times between 5 to 20 days. The ethanol production was found to be significantly affected by both variables. The optimum time of fermentation was 10 days with citrate buffer; result in 0.88458% of ethanol, and the ethanol content after distillation process was shown 0.985015%.

Keywords: Fermentation, ulva-lactuca, buffer, β-glucosidase, bioethanol.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2430