Search results for: wet chemical processing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2579

Search results for: wet chemical processing

449 In Silico Analysis of Pax6 Interacting Proteins Indicates Missing Molecular Links in Development of Brain and Associated Disease

Authors: Ratnakar Tripathi, Rajnikant Mishra

Abstract:

The PAX6, a transcription factor, is essential for the morphogenesis of the eyes, brain, pituitary and pancreatic islets. In rodents, the loss of Pax6 function leads to central nervous system defects, anophthalmia, and nasal hypoplasia. The haplo-insufficiency of Pax6 causes microphthalmia, aggression and other behavioral abnormalities. It is also required in brain patterning and neuronal plasticity. In human, heterozygous mutation of Pax6 causes loss of iris [aniridia], mental retardation and glucose intolerance. The 3- deletion in Pax6 leads to autism and aniridia. The phenotypes are variable in peneterance and expressivity. However, mechanism of function and interaction of PAX6 with other proteins during development and associated disease are not clear. It is intended to explore interactors of PAX6 to elucidated biology of PAX6 function in the tissues where it is expressed and also in the central regulatory pathway. This report describes In-silico approaches to explore interacting proteins of PAX6. The models show several possible proteins interacting with PAX6 like MITF, SIX3, SOX2, SOX3, IPO13, TRIM, and OGT. Since the Pax6 is a critical transcriptional regulator and master control gene of eye and brain development it might be interacting with other protein involved in morphogenesis [TGIF, TGF, Ras etc]. It is also presumed that matricelluar proteins [SPARC, thrombospondin-1 and osteonectin etc] are likely to interact during transport and processing of PAX6 and are somewhere its cascade. The proteins involved in cell survival and cell proliferation can also not be ignored.

Keywords: Interacting Proteins, Pax6, PIP, STRING

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1941
448 A 10 Giga VPN Accelerator Board for Trust Channel Security System

Authors: Ki Hyun Kim, Jang-Hee Yoo, Kyo Il Chung

Abstract:

This paper proposes a VPN Accelerator Board (VPN-AB), a virtual private network (VPN) protocol designed for trust channel security system (TCSS). TCSS supports safety communication channel between security nodes in internet. It furnishes authentication, confidentiality, integrity, and access control to security node to transmit data packets with IPsec protocol. TCSS consists of internet key exchange block, security association block, and IPsec engine block. The internet key exchange block negotiates crypto algorithm and key used in IPsec engine block. Security Association blocks setting-up and manages security association information. IPsec engine block treats IPsec packets and consists of networking functions for communication. The IPsec engine block should be embodied by H/W and in-line mode transaction for high speed IPsec processing. Our VPN-AB is implemented with high speed security processor that supports many cryptographic algorithms and in-line mode. We evaluate a small TCSS communication environment, and measure a performance of VPN-AB in the environment. The experiment results show that VPN-AB gets a performance throughput of maximum 15.645Gbps when we set the IPsec protocol with 3DES-HMAC-MD5 tunnel mode.

Keywords: TCSS(Trust Channel Security System), VPN(VirtualPrivate Network), IPsec, SSL, Security Processor, Securitycommunication.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2075
447 Learning Classifier Systems Approach for Automated Discovery of Crisp and Fuzzy Hierarchical Production Rules

Authors: Suraiya Jabin, Kamal K. Bharadwaj

Abstract:

This research presents a system for post processing of data that takes mined flat rules as input and discovers crisp as well as fuzzy hierarchical structures using Learning Classifier System approach. Learning Classifier System (LCS) is basically a machine learning technique that combines evolutionary computing, reinforcement learning, supervised or unsupervised learning and heuristics to produce adaptive systems. A LCS learns by interacting with an environment from which it receives feedback in the form of numerical reward. Learning is achieved by trying to maximize the amount of reward received. Crisp description for a concept usually cannot represent human knowledge completely and practically. In the proposed Learning Classifier System initial population is constructed as a random collection of HPR–trees (related production rules) and crisp / fuzzy hierarchies are evolved. A fuzzy subsumption relation is suggested for the proposed system and based on Subsumption Matrix (SM), a suitable fitness function is proposed. Suitable genetic operators are proposed for the chosen chromosome representation method. For implementing reinforcement a suitable reward and punishment scheme is also proposed. Experimental results are presented to demonstrate the performance of the proposed system.

Keywords: Hierarchical Production Rule, Data Mining, Learning Classifier System, Fuzzy Subsumption Relation, Subsumption matrix, Reinforcement Learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1431
446 Titania and Cu-Titania Composite Layer on Graphite Substrate as Negative Electrode for Li-Ion Battery

Authors: Fitria Rahmawati, Nuryani, Liviana Wijayanti

Abstract:

This research study the application of the immobilized TiO2 layer and Cu-TiO2 layer on graphite substrate as a negative electrode or anode for Li-ion battery. The titania layer was produced through chemical bath deposition method, meanwhile Cu particles were deposited electrochemically. A material can be used as an electrode as it has capability to intercalates Li ions into its crystal structure. The Li intercalation into TiO2/Graphite and Cu- TiO2/Graphite were analyzed from the changes of its XRD pattern after it was used as electrode during discharging process. The XRD patterns were refined by Le Bail method in order to determine the crystal structure of the prepared materials. A specific capacity and the cycle ability measurement were carried out to study the performance of the prepared materials as negative electrode of the Li-ion battery. The specific capacity was measured during discharging process from fully charged until the cut off voltage. A 300 was used as a load. The result shows that the specific capacity of Li-ion battery with TiO2/Graphite as negative electrode is 230.87 ± 1.70mAh.g-1 which is higher than the specific capacity of Li-ion battery with pure graphite as negative electrode, i.e 140.75 ±0.46mAh.g-1. Meanwhile deposition of Cu onto TiO2 layer does not increase the specific capacity, and the value even lower than the battery with TiO2/Graphite as electrode. The cycle ability of the prepared battery is only two cycles, due to the Li ribbon which was used as cathode became fragile and easily broken.

Keywords: Cu-TiO2, electrode, graphite substrate, Li-ion battery, TiO2 layer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1931
445 Effect of Landfill Leachate on Engineering Properties of Test Soil

Authors: S. A. Nta, M. J. Ayotamuno, I. J. Udom

Abstract:

The work presents result of laboratory analysis of the effects of landfill leachate on engineering properties of test soil. The soil used for the present study was a sandy loam soil and acidic in nature. It was collected at a depth of 0.9 m. The landfill leachate used was collected from a hole dug some meters away from dumped solid waste and analyzed to identify the pollutants and its effect on engineering properties of the test soil. The test soil applied with landfill leachate was collected at 0.25 and 0.50 m radial distances at a depth of 0.15, 0.30, 0.45 and 0.60 m from the point of application of leachate after 50 days were the application of the leachate end and 80 days from the start of the experiment for laboratory analysis. Engineering properties such as particle size distribution, specific gravity, optimum moisture content, maximum dry density, unconfined compressive strength, liquid limit, plastic limit and shrinkage limit were considered. The concentration of various chemicals at 0.25 and 0.50 radial distances and 0.15, 0.30, 0.45 and 0.6 m depth from the point of application of leachate were different. This study founds the effect of landfill leachate on the engineering properties of soil. It can be concluded that, the type of soil, chemical composition of the leachate, infiltration rate, aquifers, ground water table etc., will have a major role on the area of influence zone of the pollutants in a landfill.

Keywords: Engineering properties of test soil, landfill leachate, Municipal solid waste.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 599
444 Modification of Palm Oil Structure to Cocoa Butter Equivalent by Carica papaya Lipase- Catalyzed Interesterification

Authors: P. Pinyaphong, S. Phutrakul

Abstract:

Palm oil could be converted to cocoa butter equivalent by lipase-catalyzed interesterification. The objective of this research was to investigate the structure modification of palm oil to cocoa butter equivalent using Carica papaya lipase –catalyzed interesterification. The study showed that the compositions of cocoa butter equivalent were affected by acyl donor sources, substrate ratio, initial water of enzyme, reaction time, reaction temperature and the amount of enzyme. Among three acyl donors tested (methyl stearate, ethyl stearate and stearic acid), methyl stearate appeared to be the best acyl donor for incorporation to palm oil structure. The best reaction conditions for cocoa butter equivalent production were : substrate ratio (palm oil : methyl stearate, mol/mol) at 1 : 4, water activity of enzyme at 0.11, reaction time at 4 h, reaction temperature at 45 ° C and 18% by weight of the enzyme. The chemical and physical properties of cocoa butter equivalent were 9.75 ± 0.41% free fatty acid, 44.89 ± 0.84 iodine number, 193.19 ± 0.78 sponification value and melting point at 37-39 °C.

Keywords: Carica papaya lipase, cocoa butter equivalent, interesterification, palm oil.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3179
443 A Life Cycle Assessment (LCA) of Aluminum Production Process

Authors: Alaa Al Hawari, Mohammad Khader, Wael El Hasan, Mahmoud Alijla, Ammar Manawi, Abdelbaki Benamour

Abstract:

The production of aluminum alloys and ingots – starting from the processing of alumina to aluminum, and the final cast product – was studied using a Life Cycle Assessment (LCA) approach. The studied aluminum supply chain consisted of a carbon plant, a reduction plant, a casting plant, and a power plant. In the LCA model, the environmental loads of the different plants for the production of 1 ton of aluminum metal were investigated. The impact of the aluminum production was assessed in eight impact categories. The results showed that for all of the impact categories the power plant had the highest impact only in the cases of Human Toxicity Potential (HTP) the reduction plant had the highest impact and in the Marine Aquatic Eco-Toxicity Potential (MAETP) the carbon plant had the highest impact. Furthermore, the impact of the carbon plant and the reduction plant combined was almost the same as the impact of the power plant in the case of the Acidification Potential (AP). The carbon plant had a positive impact on the environment when it come to the Eutrophication Potential (EP) due to the production of clean water in the process. The natural gas based power plant used in the case study had 8.4 times less negative impact on the environment when compared to the heavy fuel based power plant and 10.7 times less negative impact when compared to the hard coal based power plant.

Keywords: Life cycle assessment, aluminum production, Supply chain.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4560
442 Accurate Visualization of Graphs of Functions of Two Real Variables

Authors: Zeitoun D. G., Thierry Dana-Picard

Abstract:

The study of a real function of two real variables can be supported by visualization using a Computer Algebra System (CAS). One type of constraints of the system is due to the algorithms implemented, yielding continuous approximations of the given function by interpolation. This often masks discontinuities of the function and can provide strange plots, not compatible with the mathematics. In recent years, point based geometry has gained increasing attention as an alternative surface representation, both for efficient rendering and for flexible geometry processing of complex surfaces. In this paper we present different artifacts created by mesh surfaces near discontinuities and propose a point based method that controls and reduces these artifacts. A least squares penalty method for an automatic generation of the mesh that controls the behavior of the chosen function is presented. The special feature of this method is the ability to improve the accuracy of the surface visualization near a set of interior points where the function may be discontinuous. The present method is formulated as a minimax problem and the non uniform mesh is generated using an iterative algorithm. Results show that for large poorly conditioned matrices, the new algorithm gives more accurate results than the classical preconditioned conjugate algorithm.

Keywords: Function singularities, mesh generation, point allocation, visualization, collocation least squares method, Augmented Lagrangian method, Uzawa's Algorithm, Preconditioned Conjugate Gradien

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1683
441 Physico-Mechanical Properties of Jute-Coir Fiber Reinforced Hybrid Polypropylene Composites

Authors: Salma Siddika, Fayeka Mansura, Mahbub Hasan

Abstract:

The term hybrid composite refers to the composite containing more than one type of fiber material as reinforcing fillers. It has become attractive structural material due to the ability of providing better combination of properties with respect to single fiber containing composite. The eco-friendly nature as well as processing advantage, light weight and low cost have enhanced the attraction and interest of natural fiber reinforced composite. The objective of present research is to study the mechanical properties of jute-coir fiber reinforced hybrid polypropylene (PP) composite according to filler loading variation. In the present work composites were manufactured by using hot press machine at four levels of fiber loading (5, 10, 15 and 20 wt %). Jute and coir fibers were utilized at a ratio of (1:1) during composite manufacturing. Tensile, flexural, impact and hardness tests were conducted for mechanical characterization. Tensile test of composite showed a decreasing trend of tensile strength and increasing trend of the Young-s modulus with increasing fiber content. During flexural, impact and hardness tests, the flexural strength, flexural modulus, impact strength and hardness were found to be increased with increasing fiber loading. Based on the fiber loading used in this study, 20% fiber reinforced composite resulted the best set of mechanical properties.

Keywords: Mechanical Properties; Coir, Jute, Polypropylene, Hybrid Composite.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3670
440 A Review on Medical Image Registration Techniques

Authors: Shadrack Mambo, Karim Djouani, Yskandar Hamam, Barend van Wyk, Patrick Siarry

Abstract:

This paper discusses the current trends in medical image registration techniques and addresses the need to provide a solid theoretical foundation for research endeavours. Methodological analysis and synthesis of quality literature was done, providing a platform for developing a good foundation for research study in this field which is crucial in understanding the existing levels of knowledge. Research on medical image registration techniques assists clinical and medical practitioners in diagnosis of tumours and lesion in anatomical organs, thereby enhancing fast and accurate curative treatment of patients. Literature review aims to provide a solid theoretical foundation for research endeavours in image registration techniques. Developing a solid foundation for a research study is possible through a methodological analysis and synthesis of existing contributions. Out of these considerations, the aim of this paper is to enhance the scientific community’s understanding of the current status of research in medical image registration techniques and also communicate to them, the contribution of this research in the field of image processing. The gaps identified in current techniques can be closed by use of artificial neural networks that form learning systems designed to minimise error function. The paper also suggests several areas of future research in the image registration.

Keywords: Image registration techniques, medical images, neural networks, optimisation, transformation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1744
439 Studies on the Mechanical Behavior of Bottom Ash for a Sustainable Environment

Authors: B. A. Mir, Asim Malik

Abstract:

Bottom ash is a by-product of the combustion process of coal in furnaces in the production of electricity in thermal power plants. In India, about 75% of total power is produced by using pulverized coal. The coal of India has a high ash content which leads to the generation of a huge quantity of bottom ash per year posing the dual problem of environmental pollution and difficulty in disposal. This calls for establishing strategies to use this industry by-product effectively and efficiently. However, its large-scale utilization is possible only in geotechnical applications, either alone or with soil. In the present investigation, bottom ash was collected from National Capital Power Station Dadri, Uttar Pradesh, India. Test samples of bottom ash admixed with 20% clayey soil were prepared and treated with different cement content by weight and subjected to various laboratory tests for assessing its suitability as an engineered construction material. This study has shown that use of 10% cement content is a viable chemical additive to enhance the mechanical properties of bottom ash, which can be used effectively as an engineered construction material in various geotechnical applications. More importantly, it offers an interesting potential for making use of an industrial waste to overcome challenges posed by bottom ash for a sustainable environment.

Keywords: Bottom ash, environmental pollution, solid waste, sustainable environment, waste utilization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1682
438 Optimal Green Facility Planning - Implementation of Organic Rankine Cycle System for Factory Waste Heat Recovery

Authors: Chun-Wei Lin, Yu-Lin Chen

Abstract:

As global industry developed rapidly, the energy demand also rises simultaneously. In the production process, there’s a lot of energy consumed in the process. Formally, the energy used in generating the heat in the production process. In the total energy consumption, 40% of the heat was used in process heat, mechanical work, chemical energy and electricity. The remaining 50% were released into the environment. It will cause energy waste and environment pollution. There are many ways for recovering the waste heat in factory. Organic Rankine Cycle (ORC) system can produce electricity and reduce energy costs by recovering the waste of low temperature heat in the factory. In addition, ORC is the technology with the highest power generating efficiency in low-temperature heat recycling. However, most of factories executives are still hesitated because of the high implementation cost of the ORC system, even a lot of heat are wasted. Therefore, this study constructs a nonlinear mathematical model of waste heat recovery equipment configuration to maximize profits. A particle swarm optimization algorithm is developed to generate the optimal facility installation plan for the ORC system.

Keywords: Green facility planning, organic rankine cycle, particle swarm optimization, waste heat recovery.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1956
437 Object Recognition on Horse Riding Simulator System

Authors: Kyekyung Kim, Sangseung Kang, Suyoung Chi, Jaehong Kim

Abstract:

In recent years, IT convergence technology has been developed to get creative solution by combining robotics or sports science technology. Object detection and recognition have mainly applied to sports science field that has processed by recognizing face and by tracking human body. But object detection and recognition using vision sensor is challenge task in real world because of illumination. In this paper, object detection and recognition using vision sensor applied to sports simulator has been introduced. Face recognition has been processed to identify user and to update automatically a person athletic recording. Human body has tracked to offer a most accurate way of riding horse simulator. Combined image processing has been processed to reduce illumination adverse affect because illumination has caused low performance in detection and recognition in real world application filed. Face has recognized using standard face graph and human body has tracked using pose model, which has composed of feature nodes generated diverse face and pose images. Face recognition using Gabor wavelet and pose recognition using pose graph is robust to real application. We have simulated using ETRI database, which has constructed on horse riding simulator.

Keywords: Horse riding simulator, Object detection, Object recognition, User identification, Pose recognition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2048
436 Automatic Detection of Breast Tumors in Sonoelastographic Images Using DWT

Authors: A. Sindhuja, V. Sadasivam

Abstract:

Breast Cancer is the most common malignancy in women and the second leading cause of death for women all over the world. Earlier the detection of cancer, better the treatment. The diagnosis and treatment of the cancer rely on segmentation of Sonoelastographic images. Texture features has not considered for Sonoelastographic segmentation. Sonoelastographic images of 15 patients containing both benign and malignant tumorsare considered for experimentation.The images are enhanced to remove noise in order to improve contrast and emphasize tumor boundary. It is then decomposed into sub-bands using single level Daubechies wavelets varying from single co-efficient to six coefficients. The Grey Level Co-occurrence Matrix (GLCM), Local Binary Pattern (LBP) features are extracted and then selected by ranking it using Sequential Floating Forward Selection (SFFS) technique from each sub-band. The resultant images undergo K-Means clustering and then few post-processing steps to remove the false spots. The tumor boundary is detected from the segmented image. It is proposed that Local Binary Pattern (LBP) from the vertical coefficients of Daubechies wavelet with two coefficients is best suited for segmentation of Sonoelastographic breast images among the wavelet members using one to six coefficients for decomposition. The results are also quantified with the help of an expert radiologist. The proposed work can be used for further diagnostic process to decide if the segmented tumor is benign or malignant.

Keywords: Breast Cancer, Segmentation, Sonoelastography, Tumor Detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2176
435 Improvement of Water Distillation Plant by Using Statistical Process Control System

Authors: Qasim Kriri, Harsh B. Desai

Abstract:

Water supply and sanitation in Saudi Arabia is portrayed by difficulties and accomplishments. One of the fundamental difficulties is water shortage. With a specific end goal to beat water shortage, significant ventures have been attempted in sea water desalination, water circulation, sewerage, and wastewater treatment. The motivation behind Statistical Process Control (SPC) is to decide whether the execution of a procedure is keeping up an acceptable quality level [AQL]. SPC is an analytical decision-making method. A fundamental apparatus in the SPC is the Control Charts, which follow the inconstancy in the estimations of the item quality attributes. By utilizing the suitable outline, administration can decide whether changes should be made with a specific end goal to keep the procedure in charge. The two most important quality factors in the distilled water which were taken into consideration were pH (Potential of Hydrogen) and TDS (Total Dissolved Solids). There were three stages at which the quality checks were done. The stages were as follows: (1) Water at the source, (2) water after chemical treatment & (3) water which is sent for packing. The upper specification limit, central limit and lower specification limit are taken as per Saudi water standards. The procedure capacity to accomplish the particulars set for the quality attributes of Berain water Factory chose to be focused by the proposed SPC system.

Keywords: Acceptable quality level, statistical quality control, control charts, process charts.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1016
434 Mobile Robot Control by Von Neumann Computer

Authors: E. V. Larkin, T. A. Akimenko, A. V. Bogomolov, A. N. Privalov

Abstract:

The digital control system of mobile robots (MR) control is considered. It is shown that sequential interpretation of control algorithm operators, unfolding in physical time, suggests the occurrence of time delays between inputting data from sensors and outputting data to actuators. Another destabilizing control factor is presence of backlash in the joints of an actuator with an executive unit. Complex model of control system, which takes into account the dynamics of the MR, the dynamics of the digital controller and backlash in actuators, is worked out. The digital controller model is divided into two parts: the first part describes the control law embedded in the controller in the form of a control program that realizes a polling procedure when organizing transactions to sensors and actuators. The second part of the model describes the time delays that occur in the Von Neumann-type controller when processing data. To estimate time intervals, the algorithm is represented in the form of an ergodic semi-Markov process. For an ergodic semi-Markov process of common form, a method is proposed for estimation a wandering time from one arbitrary state to another arbitrary state. Example shows how the backlash and time delays affect the quality characteristics of the MR control system functioning.

Keywords: Mobile robot, backlash, control algorithm, Von Neumann controller, semi-Markov process, time delay.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 305
433 Effect of Plant Biostimulants on Fruit Set, Yield, and Quality Attributes of “Farbaly” Apricot Cultivar

Authors: A. Tarantino, F. Lops, G. Disciglio, E. Tarantino

Abstract:

Apulia region (southern Italy) is excellent for heavy production of apricot (Prunus armeniaca L.). Fruit quality is a combination of physical, chemical and nutritional characteristics. The present experiment was laid in the commercial orchard in Cerignola (Foggia district, Apulia region, 41°15’49’’N; 15°53’59’’E; 126 a.s.l.) during the 2014-2015 season. The experiment consisted of the use of three biostimulant treatments (Hendophyt®, Ergostim® and Radicon®) compared with untreated control on ‘Farbaly’ apricot cultivar, in order to evaluate the vegeto-productive and fruit qualitative attributes. Foliar spray of biostimulants was applied at different times during the growth season (at red ball, fruit setting and fruit development stages). Experimental data showed some specific differences among the biostimulant treatments, which fruit set, growth and productivity were affected. Moderate influences were found regarding the qualitative attributes of fruits. The soluble solid content was positively affected by Hendophyt® treatment. Antioxidant capacity was significantly higher in Hendophyt® and Radicon® treatments respect to the untreated control.

Keywords: Prunus Armeniaca L., biostimulants, fruit set, fruit quality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1264
432 Non-Overlapping Hierarchical Index Structure for Similarity Search

Authors: Mounira Taileb, Sid Lamrous, Sami Touati

Abstract:

In order to accelerate the similarity search in highdimensional database, we propose a new hierarchical indexing method. It is composed of offline and online phases. Our contribution concerns both phases. In the offline phase, after gathering the whole of the data in clusters and constructing a hierarchical index, the main originality of our contribution consists to develop a method to construct bounding forms of clusters to avoid overlapping. For the online phase, our idea improves considerably performances of similarity search. However, for this second phase, we have also developed an adapted search algorithm. Our method baptized NOHIS (Non-Overlapping Hierarchical Index Structure) use the Principal Direction Divisive Partitioning (PDDP) as algorithm of clustering. The principle of the PDDP is to divide data recursively into two sub-clusters; division is done by using the hyper-plane orthogonal to the principal direction derived from the covariance matrix and passing through the centroid of the cluster to divide. Data of each two sub-clusters obtained are including by a minimum bounding rectangle (MBR). The two MBRs are directed according to the principal direction. Consequently, the nonoverlapping between the two forms is assured. Experiments use databases containing image descriptors. Results show that the proposed method outperforms sequential scan and SRtree in processing k-nearest neighbors.

Keywords: K-nearest neighbour search, multi-dimensional indexing, multimedia databases, similarity search.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1541
431 Numerical Simulation of Natural Gas Dispersion from Low Pressure Pipelines

Authors: Omid Adibi, Nategheh Najafpour, Bijan Farhanieh, Hossein Afshin

Abstract:

Gas release from the pipelines is one of the main factors in the gas industry accidents. Released gas ejects from the pipeline as a free jet and in the growth process, the fuel gets mixed with the ambient air. Accordingly, an accidental spark will release the chemical energy of the mixture with an explosion. Gas explosion damages the equipment and endangers the life of staffs. So due to importance of safety in gas industries, prevision of accident can reduce the number of the casualties. In this paper, natural gas leakages from the low pressure pipelines are studied in two steps: 1) the simulation of mixing process and identification of flammable zones and 2) the simulation of wind effects on the mixing process. The numerical simulations were performed by using the finite volume method and the pressure-based algorithm. Also, for the grid generation the structured method was used. The results show that, in just 6.4 s after accident, released natural gas could penetrate to 40 m in vertical and 20 m in horizontal direction. Moreover, the results show that the wind speed is a key factor in dispersion process. In fact, the wind transports the flammable zones into the downstream. Hence, to improve the safety of the people and human property, it is preferable to construct gas facilities and buildings in the opposite side of prevailing wind direction.

Keywords: Flammable zones, gas pipelines, numerical simulation, wind effects.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1114
430 Static and Dynamic Three-Dimensional Finite Element Analysis of Pelvic Bone

Authors: M. S. El-Asfoury, M. A. El-Hadek

Abstract:

The complex shape of the human pelvic bone was successfully imaged and modeled using finite element FE processing. The bone was subjected to quasi-static and dynamic loading conditions simulating the effect of both weight gain and impact. Loads varying between 500 – 2500 N (~50 – 250 Kg of weight) was used to simulate 3D quasi-static weight gain. Two different 3D dynamic analyses, body free fall at two different heights (1 and 2 m) and forced side impact at two different velocities (20 and 40 Km/hr) were also studied. The computed resulted stresses were compared for the four loading cases, where Von Misses stresses increases linearly with the weight gain increase under quasi-static loading. For the dynamic models, the Von Misses stress history behaviors were studied for the affected area and effected load with respect to time. The normalization Von Misses stresses with respect to the applied load were used for comparing the free fall and the forced impact load results. It was found that under the forced impact loading condition an over lapping behavior was noticed, where as for the free fall the normalized Von Misses stresses behavior was found to nonlinearly different. This phenomenon was explained through the energy dissipation concept. This study will help designers in different specialization in defining the weakest spots for designing different supporting systems.

Keywords: Pelvic Bone, Static and Dynamic Analysis, Three- Dimensional Finite Element Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2103
429 Artificial Intelligence: A Comprehensive and Systematic Literature Review of Applications and Comparative Technologies

Authors: Z. M. Najmi

Abstract:

Over the years, the question around Artificial Intelligence has always been one with many answers. Whether by means of use in business and industry or complicated algorithmic programming, management of these technologies has always been the core focus. More recently, technologies have been questioned in industry and society alike as to whether they have improved human-centred design, assisted choices and objectives, and had a hand in systematic processes across the board. With these questions the answer may lie within AI technologies, and the steps needed in removing common human error. Elements such as Machine Learning, Deep Learning, Recommender Systems and Natural Language Processing will all be features to consider moving forward. Our previous intervention with AI applications has resulted in increased productivity, however, raised concerns for the continuation of traditional human-centred occupations. Emerging technologies such as Augmented Reality and Virtual Reality have all played a part in this during AI’s prominent rise. As mentioned, AI has been constantly under the microscope; the benefits and drawbacks may seem endless is wide, but AI is something we must take notice of and adapt into our everyday lives. The aim of this paper is to give an overview of the technologies surrounding A.I. and its’ related technologies. A comprehensive review has been written as a timeline of the developing events and key points in the history of Artificial Intelligence. This research is gathered entirely from secondary research, academic statements of knowledge and gathered to produce an understanding of the timeline of AI.

Keywords: Artificial Intelligence, Deep Learning, Augmented Reality, Reinforcement Learning, Machine Learning, Supervised Learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 502
428 Processing and Assessment of Quality Characteristics of Composite Baby Foods

Authors: Reihaneh Ahmadzadeh Ghavidel, Mehdi Ghiafeh Davoodi

Abstract:

The usefulness of weaning foods to meet the nutrient needs of children is well recognized, and most of them are precooked roller dried mixtures of cereal and/or legume flours which posses a high viscosity and bulk when reconstituted. The objective of this study was to formulate composite weaning foods using cereals, malted legumes and vegetable powders and analyze them for nutrients, functional properties and sensory attributes. Selected legumes (green gram and lentil) were germinated, dried and dehulled. Roasted wheat, rice, carrot powder and skim milk powder also were used. All the ingredients were mixed in different proportions to get four formulations, made into 30% slurry and dried in roller drier. The products were analyzed for proximate principles, mineral content, functional and sensory qualities. The results of analysis showed following range of constituents per 100g of formulations on dry weight basis, protein, 18.1-18.9 g ; fat, 0.78-1.36 g ; iron, 5.09-6.53 mg; calcium, 265-310 mg. The lowest water absorption capacity was in case of wheat green gram based and the highest was in rice lentil based sample. Overall sensory qualities of all foods were graded as “good" and “very good" with no significant differences. The results confirm that formulated weaning foods were nutritionally superior, functionally appropriate and organoleptically acceptable.

Keywords: malted legumes, weaning foods, nutrition, functional properties

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2050
427 Taguchi-Based Six Sigma Approach to Optimize Surface Roughness for Milling Processes

Authors: Sky Chou, Joseph C. Chen

Abstract:

This paper focuses on using Six Sigma methodologies to improve the surface roughness of a manufactured part produced by the CNC milling machine. It presents a case study where the surface roughness of milled aluminum is required to reduce or eliminate defects and to improve the process capability index Cp and Cpk for a CNC milling process. The six sigma methodology, DMAIC (design, measure, analyze, improve, and control) approach, was applied in this study to improve the process, reduce defects, and ultimately reduce costs. The Taguchi-based six sigma approach was applied to identify the optimized processing parameters that led to the targeted surface roughness specified by our customer. A L9 orthogonal array was applied in the Taguchi experimental design, with four controllable factors and one non-controllable/noise factor. The four controllable factors identified consist of feed rate, depth of cut, spindle speed, and surface roughness. The noise factor is the difference between the old cutting tool and the new cutting tool. The confirmation run with the optimal parameters confirmed that the new parameter settings are correct. The new settings also improved the process capability index. The purpose of this study is that the Taguchi–based six sigma approach can be efficiently used to phase out defects and improve the process capability index of the CNC milling process.

Keywords: CNC machining, Six Sigma, Surface roughness, Taguchi methodology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1026
426 Study on the Derivatization Process Using N-O-bis-(trimethylsilyl)-trifluoroacetamide, N-(tert-butyldimethylsilyl)-N-methyltrifluoroace tamide, Trimethylsilydiazomethane for the Determination of Fecal Sterols by Gas Chromatography-Mass Spectrometry

Authors: Jingming Wu, Ruikang Hu, Junqi Yue, Zhaoguang Yang, Lifeng Zhang

Abstract:

Fecal sterol has been proposed as a chemical indicator of human fecal pollution even when fecal coliform populations have diminished due to water chlorination or toxic effects of industrial effluents. This paper describes an improved derivatization procedure for simultaneous determination of four fecal sterols including coprostanol, epicholestanol, cholesterol and cholestanol using gas chromatography-mass spectrometry (GC-MS), via optimization study on silylation procedures using N-O-bis (trimethylsilyl)-trifluoroacetamide (BSTFA), and N-(tert-butyldimethylsilyl)-N-methyltrifluoroacetamide (MTBSTFA), which lead to the formation of trimethylsilyl (TMS) and tert-butyldimethylsilyl (TBS) derivatives, respectively. Two derivatization processes of injection-port derivatization and water bath derivatization (60 oC, 1h) were inspected and compared. Furthermore, the methylation procedure at 25 oC for 2h with trimethylsilydiazomethane (TMSD) for fecal sterols analysis was also studied. It was found that most of TMS derivatives demonstrated the highest sensitivities, followed by methylated derivatives. For BSTFA or MTBSTFA derivatization processes, the simple injection-port derivatization process could achieve the same efficiency as that in the tedious water bath derivatization procedure.

Keywords: Fecal Sterols, Methylation, Silylation, BSTFA, MTBSTFA, TMSD, GC-MS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2229
425 From Type-I to Type-II Fuzzy System Modeling for Diagnosis of Hepatitis

Authors: Shahabeddin Sotudian, M. H. Fazel Zarandi, I. B. Turksen

Abstract:

Hepatitis is one of the most common and dangerous diseases that affects humankind, and exposes millions of people to serious health risks every year. Diagnosis of Hepatitis has always been a challenge for physicians. This paper presents an effective method for diagnosis of hepatitis based on interval Type-II fuzzy. This proposed system includes three steps: pre-processing (feature selection), Type-I and Type-II fuzzy classification, and system evaluation. KNN-FD feature selection is used as the preprocessing step in order to exclude irrelevant features and to improve classification performance and efficiency in generating the classification model. In the fuzzy classification step, an “indirect approach” is used for fuzzy system modeling by implementing the exponential compactness and separation index for determining the number of rules in the fuzzy clustering approach. Therefore, we first proposed a Type-I fuzzy system that had an accuracy of approximately 90.9%. In the proposed system, the process of diagnosis faces vagueness and uncertainty in the final decision. Thus, the imprecise knowledge was managed by using interval Type-II fuzzy logic. The results that were obtained show that interval Type-II fuzzy has the ability to diagnose hepatitis with an average accuracy of 93.94%. The classification accuracy obtained is the highest one reached thus far. The aforementioned rate of accuracy demonstrates that the Type-II fuzzy system has a better performance in comparison to Type-I and indicates a higher capability of Type-II fuzzy system for modeling uncertainty.

Keywords: Hepatitis disease, medical diagnosis, type-I fuzzy logic, type-II fuzzy logic, feature selection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1611
424 Efficacy of Polyfluoroalkyl Substances Filtration with Low-Cost Organic Fiber Filter

Authors: Gautham Das, Edward Morrone, Erik Treble, Clinton Binder

Abstract:

The purpose of this study was to evaluate the efficacy of a low-cost filter regarding per- and polyfluoroalkyl substances (PFAS). PFAS is a commonly used man-made chemical that can be found in a variety of household and industrial products with deleterious effects on humans. The filter consists of a combination of low-cost materials which could be locally procured. Water testing results for 4 different PFAS contaminants indicated that for Perfluorooctane sulfonic acid (PFOS), the Agency for Toxic Substances and Disease Registry (ATSDR) regulation is 7 ppt, the initial concentration was 15 ppt, and the final concentration was 3.9 ppt. For Perfluorononanoic acid (PFNA), the ATSDR regulation is 10.5 ppt, the initial concentration was 15 ppt, and the final concentration was 3.9 ppt. For Perfluorooctanoic acid (PFOA), the ATSDR regulation is 11 ppt, the initial concentration was 15 ppt, and the final concentration was 3.9 ppt. For Perfluorohexane sulfonic acid (PFHxS), the ATSDR regulation is 70 ppt, the initial concentration was 15 ppt, and the final concentration was 3.9 ppt. The results indicated a 74% reduction in PFAS concentration in filtered samples. Statistical data through regression analysis showed 0.9 validity of the sample data. Initial tests show the efficiency of the proposed filter described could be far greater if tested at a greater scale. It is highly recommended further testing to be conducted to validate the data for an innovative solution to a ubiquitous problem.

Keywords: PFAS, PFOS, PFOA, PFHxS, low-cost filter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 605
423 Study on the Derivatization Process Using N-O-bis-(trimethylsilyl)-trifluoroacetamide,N-(tert-butyldimethylsilyl)-N-methyltrifluoroacetamide, Trimethylsilydiazomethane for the Determination of Fecal Sterols by Gas Chromatography-Mass Spectrometry

Authors: Jingming Wu, Ruikang Hu, Junqi Yue, Zhaoguang Yang, Lifeng Zhang

Abstract:

Fecal sterol has been proposed as a chemical indicator of human fecal pollution even when fecal coliform populations have diminished due to water chlorination or toxic effects of industrial effluents. This paper describes an improved derivatization procedure for simultaneous determination of four fecal sterols including coprostanol, epicholestanol, cholesterol and cholestanol using gas chromatography-mass spectrometry (GC-MS), via optimization study on silylation procedures using N-O-bis (trimethylsilyl)-trifluoroacetamide (BSTFA), and N-(tert-butyldimethylsilyl)-N-methyltrifluoroacetamide (MTBSTFA), which lead to the formation of trimethylsilyl (TMS) and tert-butyldimethylsilyl (TBS) derivatives, respectively. Two derivatization processes of injection-port derivatization and water bath derivatization (60 oC, 1h) were inspected and compared. Furthermore, the methylation procedure at 25 oC for 2h with trimethylsilydiazomethane (TMSD) for fecal sterols analysis was also studied. It was found that most of TMS derivatives demonstrated the highest sensitivities, followed by methylated derivatives. For BSTFA or MTBSTFA derivatization processes, the simple injection-port derivatization process could achieve the same efficiency as that in the tedious water bath derivatization procedure.

Keywords: Fecal Sterols, Methylation, Silylation, BSTFA, MTBSTFA, TMSD, GC-MS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3171
422 Wildfires Assessed by Remote Sense Images and Burned Land Monitoring

Authors: M. C. Proença

Abstract:

The tools described in this paper enable the location of burned areas where took place the annihilation of natural habitats and establishes a baseline for major changes in forest ecosystems during recovery. Moreover, the result allows the follow up of the surface fuel loading, allowing the evaluation and guidance of restoration measures to remote areas by phased time planning. This case study implements the evaluation of burned areas that suffered successive wildfires in Portugal mainland during the summer of 2017, killing more than 60 people. The goal is to show that this evaluation can be done with remote sense data free of charges in a simple laptop, with open-source software, describing the not-so-simple methodology step by step, to make it accessible for local workers in the areas attained, where the availability of information is essential for the immediate planning of mitigation measures, such as restoring road access, allocate funds for the recovery of human dwellings and assess further needs for restoration of the ecological system. Wildfires also devastate forest ecosystems having a direct impact on vegetation cover and killing or driving away the animal population, besides loss of all crops in rural areas that are essential as local resources. The economic interests are also attained, as the pinewood burned becomes useless for the noblest applications, so its value decreases, and resin extraction ends for several years.

Keywords: Image processing, remote sensing, wildfires, burned areas, SENTINEL-2.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1494
421 A Novel SVM-Based OOK Detector in Low SNR Infrared Channels

Authors: J. P. Dubois, O. M. Abdul-Latif

Abstract:

Support Vector Machine (SVM) is a recent class of statistical classification and regression techniques playing an increasing role in applications to detection problems in various engineering problems, notably in statistical signal processing, pattern recognition, image analysis, and communication systems. In this paper, SVM is applied to an infrared (IR) binary communication system with different types of channel models including Ricean multipath fading and partially developed scattering channel with additive white Gaussian noise (AWGN) at the receiver. The structure and performance of SVM in terms of the bit error rate (BER) metric is derived and simulated for these channel stochastic models and the computational complexity of the implementation, in terms of average computational time per bit, is also presented. The performance of SVM is then compared to classical binary signal maximum likelihood detection using a matched filter driven by On-Off keying (OOK) modulation. We found that the performance of SVM is superior to that of the traditional optimal detection schemes used in statistical communication, especially for very low signal-to-noise ratio (SNR) ranges. For large SNR, the performance of the SVM is similar to that of the classical detectors. The implication of these results is that SVM can prove very beneficial to IR communication systems that notoriously suffer from low SNR at the cost of increased computational complexity.

Keywords: Least square-support vector machine, on-off keying, matched filter, maximum likelihood detector, wireless infrared communication.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1925
420 Isolation and Screening of Fungal Strains for β-Galactosidase Production

Authors: Parmjit S. Panesar, Rupinder Kaur, Ram S. Singh

Abstract:

Enzymes are the biocatalysts which catalyze the biochemical processes and thus have a wide variety of applications in the industrial sector. β-Galactosidase (E.C. 3.2.1.23) also known as lactase, is one of the prime enzymes, which has significant potential in the dairy and food processing industries. It has the capability to catalyze both the hydrolytic reaction for the production of lactose hydrolyzed milk and transgalactosylation reaction for the synthesis of prebiotics such as lactulose and galactooligosaccharides. These prebiotics have various nutritional and technological benefits. Although, the enzyme is naturally present in almonds, peaches, apricots and other variety of fruits and animals, the extraction of enzyme from these sources increases the cost of enzyme. Therefore, focus has been shifted towards the production of low cost enzyme from the microorganisms such as bacteria, yeast and fungi. As compared to yeast and bacteria, fungal β-galactosidase is generally preferred as being extracellular and thermostable in nature. Keeping the above in view, the present study was carried out for the isolation of the β-galactosidase producing fungal strain from the food as well as the agricultural wastes. A total of more than 100 fungal cultures were examined for their potential in enzyme production. All the fungal strains were screened using X-gal and IPTG as inducers in the modified Czapek Dox Agar medium. Among the various isolated fungal strains, the strain exhibiting the highest enzyme activity was chosen for further phenotypic and genotypic characterization. The strain was identified as Rhizomucor pusillus on the basis of 5.8s RNA gene sequencing data.

Keywords: β-galactosidase, enzyme, fungus, isolation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2553