Search results for: cassava starch processing wastewater
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4777

Search results for: cassava starch processing wastewater

2317 Comparison Conventional with Microwave-Assisted Drying Method on the Physicochemical Characteristics of Rice Bran Noodle

Authors: Chien-Chun Huang, Yi-U Chiou, Chiun-C.R. Wang

Abstract:

For longer shelf life of noodles, air-dried method is the traditional way for the noodle preparation. Microwave drying has the specific advantage of rapid and uniform heating due to the penetration of microwaves into the body of the product. Microwave-assisted facility offers a quick and energy saving method during food dehydration as compares to the conventional air-dried method. Recently, numerous studies in the rheological characteristics of pasta or spaghetti were carried out with microwave–assisted air driers and many agricultural products were dried successfully. There are few researches about the evaluation of physicochemical characteristics and cooking quality of microwave-assisted air dried salted noodles. The purposes of this study were to compare the difference between conventional and microwave-assisted drying method on the physicochemical properties and eating quality of rice bran noodles. Three different microwave power including 0.5 KW, 0.75 KW and 1.0 KW installing with 50℃ hot air were applied for dehydration of rice bran noodles in this study. Three proportion of rice bran ranging in 0-20% were incorporated into salted noodles processing. The appearance, optimum cooking time, cooking yield and losses, textural profiles analysis, sensory evaluation of rice bran noodles were measured in this study. The results indicated that high power (1.0 KW) microwave facility caused partially burnt and porous on the surface of rice bran noodles. However, no characteristic of noodle was appeared on the surface of noodles preparing by low power (0.5 KW) microwave facility. The optimum cooking time of noodles was decreased as higher power microwave or higher proportion of rice bran was incorporated into noodles preparation. The higher proportion of rice bran (20%) or higher power of microwave-assisted dried noodles obtained the higher color intensity and the higher cooking losses as compared with conventional air dried noodles. The firmness of cooked rice bran noodles slightly decreased in the cooked noodles which were dried by high power microwave-assisted method. The shearing force, tensile strength, elasticity and texture profiles of cooked rice noodles decreased with the progress of the proportion of rice bran. The results of sensory evaluation indicated conventional dried noodles obtained the higher springiness, cohesiveness and acceptability of cooked noodles than high power (1.0 KW) microwave-assisted dried noodles. However, low power (0.5 KW) microwave-assisted dried noodles showed the comparable sensory attributes and acceptability with conventional dried noodles. Moreover, the sensory attributes including firmness, springiness, cohesiveness decreased, but stickiness increased, with the increases of rice bran proportion. These results inferred that incorporation of lower proportion of rice bran and lower power microwave-assisted dried noodles processing could produce faster cooking time and acceptable quality of cooked noodles as compared to conventional dried noodles.

Keywords: microwave-assisted drying method, physicochemical characteristics, rice bran noodles, sensory evaluation

Procedia PDF Downloads 468
2316 Catalytic Pyrolysis of Sewage Sludge for Upgrading Bio-Oil Quality Using Sludge-Based Activated Char as an Alternative to HZSM5

Authors: ALi Zaker, Zhi Chen

Abstract:

Due to the concerns about the depletion of fossil fuel sources and the deteriorating environment, the attempt to investigate the production of renewable energy will play a crucial role as a potential to alleviate the dependency on mineral fuels. In this respect, biofuels are measured as a vital nominee for national energy security and energy sustainability. Sewage sludge (SS), as an alternative source of renewable energy with a complex composition, is a major waste generated during wastewater treatment. Stricter legislation is continuously refining the requirements for the level of removal of various pollutants in treated water, causing continuous growth of sludge production, which has become a global challenge. In general, there are two main procedures for dealing with SS: incineration and landfill. However, there are a variety of limitations in these options (e.g., production of greenhouse gases and restrictive environmental regulations) in regard to negative social and economic impacts. Pyrolysis is a feasible and cost-effective technology that can simultaneously tackle boundaries concerning the current disposal routes while retrieving bioenergy. Pyrolysis of SS has drawn vigorous interest in research due to the ability of high mass yield of pyrolytic liquid production. Nonetheless, the presence of high molecular weight hydrocarbons and oxygenated- and nitrogenated compounds poses a considerable challenge. In this context, catalytic pyrolysis is another attainable route in order to upgrade the bio-oil quality. Among different catalysts (i.e., zeolites) studied for sewage sludge pyrolysis, activated chars are eco-friendly and low-cost alternatives. The beneficial features comprise comparatively large surface area, long-term stability, and enriched surface functional groups. In light of these premises, this research attempts to investigate the catalytic pyrolysis of sewage sludge with a high-performance sludge-based activated char in contrast to HZSM5 from a theoretical and experimental point of view.

Keywords: catalytic pyrolysis, sewage sludge, char, HZSM5, bio-oil.

Procedia PDF Downloads 20
2315 Airport Pavement Crack Measurement Systems and Crack Density for Pavement Evaluation

Authors: Ali Ashtiani, Hamid Shirazi

Abstract:

This paper reviews the status of existing practice and research related to measuring pavement cracking and using crack density as a pavement surface evaluation protocol. Crack density for pavement evaluation is currently not widely used within the airport community and its use by the highway community is limited. However, surface cracking is a distress that is closely monitored by airport staff and significantly influences the development of maintenance, rehabilitation and reconstruction plans for airport pavements. Therefore crack density has the potential to become an important indicator of pavement condition if the type, severity and extent of surface cracking can be accurately measured. A pavement distress survey is an essential component of any pavement assessment. Manual crack surveying has been widely used for decades to measure pavement performance. However, the accuracy and precision of manual surveys can vary depending upon the surveyor and performing surveys may disrupt normal operations. Given the variability of manual surveys, this method has shown inconsistencies in distress classification and measurement. This can potentially impact the planning for pavement maintenance, rehabilitation and reconstruction and the associated funding strategies. A substantial effort has been devoted for the past 20 years to reduce the human intervention and the error associated with it by moving toward automated distress collection methods. The automated methods refer to the systems that identify, classify and quantify pavement distresses through processes that require no or very minimal human intervention. This principally involves the use of a digital recognition software to analyze and characterize pavement distresses. The lack of established protocols for measurement and classification of pavement cracks captured using digital images is a challenge to developing a reliable automated system for distress assessment. Variations in types and severity of distresses, different pavement surface textures and colors and presence of pavement joints and edges all complicate automated image processing and crack measurement and classification. This paper summarizes the commercially available systems and technologies for automated pavement distress evaluation. A comprehensive automated pavement distress survey involves collection, interpretation, and processing of the surface images to identify the type, quantity and severity of the surface distresses. The outputs can be used to quantitatively calculate the crack density. The systems for automated distress survey using digital images reviewed in this paper can assist the airport industry in the development of a pavement evaluation protocol based on crack density. Analysis of automated distress survey data can lead to a crack density index. This index can be used as a means of assessing pavement condition and to predict pavement performance. This can be used by airport owners to determine the type of pavement maintenance and rehabilitation in a more consistent way.

Keywords: airport pavement management, crack density, pavement evaluation, pavement management

Procedia PDF Downloads 176
2314 Social and Educational AI for Diversity: Research on Democratic Values to Develop Artificial Intelligence Tools to Guarantee Access for all to Educational Tools and Public Services

Authors: Roberto Feltrero, Sara Osuna-Acedo

Abstract:

Responsible Research and Innovation have to accomplish one fundamental aim: everybody has to participate in the benefits of innovation, but also innovation has to be democratic; that is to say, everybody may have the possibility to participate in the decisions in the innovation process. Particularly, a democratic and inclusive model of social participation and innovation includes persons with disabilities and people at risk of discrimination. Innovations on Artificial Intelligence for social development have to accomplish the same dual goal: improving equality for accessing fields of public interest like education, training and public services, as well as improving civic and democratic participation in the process of developing such innovations for all. This research aims to develop innovations, policies and policy recommendations to apply and disseminate such artificial intelligence and social model for making educational and administrative processes more accessible. First, designing a citizen participation process to engage citizens in the designing and use of artificial intelligence tools for public services. This will result in improving trust in democratic institutions contributing to enhancing the transparency, effectiveness, accountability and legitimacy of public policy-making and allowing people to participate in the development of ethical standards for the use of such technologies. Second, improving educational tools for lifelong learning with AI models to improve accountability and educational data management. Dissemination, education and social participation will be integrated, measured and evaluated in innovative educational processes to make accessible all the educational technologies and content developed on AI about responsible and social innovation. A particular case will be presented regarding access for all to educational tools and public services. This accessibility requires cognitive adaptability because, many times, legal or administrative language is very complex. Not only for people with cognitive disabilities but also for old people or citizens at risk of educational or social discrimination. Artificial Intelligence natural language processing technologies can provide tools to translate legal, administrative, or educational texts to a more simple language that can be accessible to everybody. Despite technological advances in language processing and machine learning, this becomes a huge project if we really want to respect ethical and legal consequences because that kinds of consequences can only be achieved with civil and democratic engagement in two realms: 1) to democratically select texts that need and can be translated and 2) to involved citizens, experts and nonexperts, to produce and validate real examples of legal texts with cognitive adaptations to feed artificial intelligence algorithms for learning how to translate those texts to a more simple and accessible language, adapted to any kind of population.

Keywords: responsible research and innovation, AI social innovations, cognitive accessibility, public participation

Procedia PDF Downloads 73
2313 Unbalanced Mean-Time and Buffer Effects in Lines Suffering Breakdown

Authors: Sabry Shaaban, Tom McNamara, Sarah Hudson

Abstract:

This article studies the performance of unpaced serial production lines that are subject to breakdown and are imbalanced in terms of both of their processing time means (MTs) and buffer storage capacities (BCs). Simulation results show that the best pattern in terms of throughput is a balanced line with respect to average buffer level; the best configuration is a monotone decreasing MT order, together with an ascending BC arrangement. Statistical analysis shows that BC, patterns of MT and BC imbalance, line length and degree of imbalance all contribute significantly to performance. Results show that unbalanced lines cope well with unreliability.

Keywords: unreliable unpaced serial lines, simulation, unequal mean operation times, uneven buffer capacities, patterns of imbalance, throughput, average buffer level

Procedia PDF Downloads 453
2312 Localization of Mobile Robots with Omnidirectional Cameras

Authors: Tatsuya Kato, Masanobu Nagata, Hidetoshi Nakashima, Kazunori Matsuo

Abstract:

Localization of mobile robots are important tasks for developing autonomous mobile robots. This paper proposes a method to estimate positions of a mobile robot using an omnidirectional camera on the robot. Landmarks for points of references are set up on a field where the robot works. The omnidirectional camera which can obtain 360 [deg] around images takes photographs of these landmarks. The positions of the robots are estimated from directions of these landmarks that are extracted from the images by image processing. This method can obtain the robot positions without accumulative position errors. Accuracy of the estimated robot positions by the proposed method are evaluated through some experiments. The results show that it can obtain the positions with small standard deviations. Therefore the method has possibilities of more accurate localization by tuning of appropriate offset parameters.

Keywords: mobile robots, localization, omnidirectional camera, estimating positions

Procedia PDF Downloads 424
2311 Determination of the Botanical Origin of Honey by the Artificial Neural Network Processing of PARAFAC Scores of Fluorescence Data

Authors: Lea Lenhardt, Ivana Zeković, Tatjana Dramićanin, Miroslav D. Dramićanin

Abstract:

Fluorescence spectroscopy coupled with parallel factor analysis (PARAFAC) and artificial neural networks (ANN) were used for characterization and classification of honey. Excitation emission spectra were obtained for 95 honey samples of different botanical origin (acacia, sunflower, linden, meadow, and fake honey) by recording emission from 270 to 640 nm with excitation in the range of 240-500 nm. Fluorescence spectra were described with a six-component PARAFAC model, and PARAFAC scores were further processed with two types of ANN’s (feed-forward network and self-organizing maps) to obtain algorithms for classification of honey on the basis of their botanical origin. Both ANN’s detected fake honey samples with 100% sensitivity and specificity.

Keywords: honey, fluorescence, PARAFAC, artificial neural networks

Procedia PDF Downloads 937
2310 Anticorrosive Performances of “Methyl Ester Sulfonates” Biodegradable Anionic Synthetized Surfactants on Carbon Steel X 70 in Oilfields

Authors: Asselah Amel, Affif Chaouche M'yassa, Toudji Amira, Tazerouti Amel

Abstract:

This study covers two aspects ; the biodegradability and the performances in corrosion inhibition of a series of synthetized surfactants namely Φ- sodium methyl ester sulfonates (Φ-MES: C₁₂-MES, C₁₄-MES and C₁₆-MES. The biodegradability of these organic compounds was studied using the respirometric method, ‘the standard ISO 9408’. Degradation was followed by analysis of dissolved oxygen using the dissolved oxygen meter over 28 days and the results were compared with that of sodium dodecyl sulphate (SDS). The inoculum used consists of activated sludge taken from the aeration basin of the biological wastewater treatment plant in the city of Boumerdes-Algeria. In addition, the anticorrosive performances of Φ-MES surfactants on a carbon steel "X70" were evaluated in an injection water from a well of Hassi R'mel region- Algeria, known as Baremian water, and are compared to sodium dodecyl sulphate. Two technics, the weight loss and the linear polarization resistance corrosion rate (LPR) are used allowing to investigate the relationships between the concentrations of these synthetized surfactants and their surface properties, surface coverage and inhibition efficiency. Various adsorption isotherm models were used to characterize the nature of adsorption and explain their mechanism. The results show that the MES anionic surfactants was readily biodegradable, degrading faster than SDS, about 88% for C₁₂-MES compared to 66% for the SDS. The length of their carbon chain affects their biodegradability; the longer the chain, the lower the biodegradability. The inhibition efficiency of these surfactants is around 78.4% for C₁₂-MES, 76.60% for C₁₄-MES and 98.19% for C₁₆-MES and increases with their concentration and reaches a maximum value around their critical micelle concentrations ( CMCs). Scanning electron microscopy coupled to energy dispersive X-ray spectroscopy allowed to the visualization of a good adhesion of the protective film formed by the surfactants to the surface of the steel. The studied surfactants show the Langmuirian behavior from which the thermodynamic parameters as adsorption constant (Kads), standard free energy of adsorption (〖∆G〗_ads^0 ) are determined. Interaction of the surfactants with steel surface have involved physisorptions.

Keywords: corrosion, surfactants, adsorption, adsorption isotherems

Procedia PDF Downloads 84
2309 A Research Review on the Presence of Pesticide Residues in Apples Carried out in Poland in the Years 1980-2015

Authors: Bartosz Piechowicz, Stanislaw Sadlo, Przemyslaw Grodzicki, Magdalena Podbielska

Abstract:

Apples are popular fruits. They are eaten freshly and/or after processing. For instance Golden Delicious is an apple variety commonly used in production of foods for babies and toddlers. It is no wonder that complex analyses of the pesticide residue levels in those fruits have been carried out since eighties, and continued for the next years up to now. The results obtained were presented, usually as a teamwork, at the scientific sessions organised by the (IOR) Institute of Plant Protection-National Research Institute in Poznań and published in Scientific Works of the Institute (now Progress in Plant Protection/ Postępy w Ochronie Roślin) or Journal of Plant Protection Research, and in many non-periodical publications. These reports included studies carried out by IOR Laboratories in Poznań, Sośnicowice, Rzeszów and Bialystok. First detailed studies on the presence of pesticide residues in apple fruits by the laboratory in Rzeszów were published in 1991 in the article entitled 'The presence of pesticides in apples of late varieties from the area of south-eastern Poland in the years 1986-1989', in Annals of National Institute of Hygiene in Warsaw. These surveys gave the scientific base for business contacts between the Polish company Alima and the American company Gerber. At the beginning of XXI century, in Poland, systematic and complex studies on the deposition of pesticide residues in apples were initiated. First of all, the levels of active ingredients of plant protection products applied against storage diseases at 2-3 weeks before the harvest were determined. It is known that the above mentioned substances usually generate the highest residue levels. Also, the assessment of the fungicide residues in apples during their storage in controlled atmosphere and during their processing was carried out. Taking into account the need of actualisation the Maximum Residue Levels of pesticides, in force in Poland and in other European countries, and rationalisation of the ways of their determination, a lot of field tests on the behaviour of more important fungicides on the mature fruits just before their harvesting, were carried out. A rate of their disappearance and mathematical equation that showed the relationship between the residue level of any substance and the used dose, have been determined. The two parameters have allowed to evaluate the Maximum Residue Levels (MRLs) of pesticides, which were in force at that time, and to propose a coherent model of their determination in respect to the new substances. The obtained results were assessed in terms of the health risk for adult consumers and children, and to such determination of terms of treatment that mature apples could meet the rigorous level of 0.01 mg/kg.

Keywords: apple, disappearance, health risk, MRL, pesticide residue, research

Procedia PDF Downloads 263
2308 Temperature Contour Detection of Salt Ice Using Color Thermal Image Segmentation Method

Authors: Azam Fazelpour, Saeed Reza Dehghani, Vlastimil Masek, Yuri S. Muzychka

Abstract:

The study uses a novel image analysis based on thermal imaging to detect temperature contours created on salt ice surface during transient phenomena. Thermal cameras detect objects by using their emissivities and IR radiance. The ice surface temperature is not uniform during transient processes. The temperature starts to increase from the boundary of ice towards the center of that. Thermal cameras are able to report temperature changes on the ice surface at every individual moment. Various contours, which show different temperature areas, appear on the ice surface picture captured by a thermal camera. Identifying the exact boundary of these contours is valuable to facilitate ice surface temperature analysis. Image processing techniques are used to extract each contour area precisely. In this study, several pictures are recorded while the temperature is increasing throughout the ice surface. Some pictures are selected to be processed by a specific time interval. An image segmentation method is applied to images to determine the contour areas. Color thermal images are used to exploit the main information. Red, green and blue elements of color images are investigated to find the best contour boundaries. The algorithms of image enhancement and noise removal are applied to images to obtain a high contrast and clear image. A novel edge detection algorithm based on differences in the color of the pixels is established to determine contour boundaries. In this method, the edges of the contours are obtained according to properties of red, blue and green image elements. The color image elements are assessed considering their information. Useful elements proceed to process and useless elements are removed from the process to reduce the consuming time. Neighbor pixels with close intensities are assigned in one contour and differences in intensities determine boundaries. The results are then verified by conducting experimental tests. An experimental setup is performed using ice samples and a thermal camera. To observe the created ice contour by the thermal camera, the samples, which are initially at -20° C, are contacted with a warmer surface. Pictures are captured for 20 seconds. The method is applied to five images ,which are captured at the time intervals of 5 seconds. The study shows the green image element carries no useful information; therefore, the boundary detection method is applied on red and blue image elements. In this case study, the results indicate that proposed algorithm shows the boundaries more effective than other edges detection methods such as Sobel and Canny. Comparison between the contour detection in this method and temperature analysis, which states real boundaries, shows a good agreement. This color image edge detection method is applicable to other similar cases according to their image properties.

Keywords: color image processing, edge detection, ice contour boundary, salt ice, thermal image

Procedia PDF Downloads 302
2307 A New Floating Point Implementation of Base 2 Logarithm

Authors: Ahmed M. Mansour, Ali M. El-Sawy, Ahmed T. Sayed

Abstract:

Logarithms reduce products to sums and powers to products; they play an important role in signal processing, communication and information theory. They are primarily used for hardware calculations, handling multiplications, divisions, powers, and roots effectively. There are three commonly used bases for logarithms; the logarithm with base-10 is called the common logarithm, the natural logarithm with base-e and the binary logarithm with base-2. This paper demonstrates different methods of calculation for log2 showing the complexity of each and finds out the most accurate and efficient besides giving in- sights to their hardware design. We present a new method called Floor Shift for fast calculation of log2, and then we combine this algorithm with Taylor series to improve the accuracy of the output, we illustrate that by using two examples. We finally compare the algorithms and conclude with our remarks.

Keywords: logarithms, log2, floor, iterative, CORDIC, Taylor series

Procedia PDF Downloads 510
2306 Solving Process Planning and Scheduling with Number of Operation Plus Processing Time Due-Date Assignment Concurrently Using a Genetic Search

Authors: Halil Ibrahim Demir, Alper Goksu, Onur Canpolat, Caner Erden, Melek Nur

Abstract:

Traditionally process planning, scheduling and due date assignment are performed sequentially and separately. High interrelation between these functions makes integration very useful. Although there are numerous works on integrated process planning and scheduling and many works on scheduling with due date assignment, there are only a few works on the integration of these three functions. Here we tested the different integration levels of these three functions and found a fully integrated version as the best. We applied genetic search and random search and genetic search was found better compared to the random search. We penalized all earliness, tardiness and due date related costs. Since all these three terms are all undesired, it is better to penalize all of them.

Keywords: process planning, scheduling, due-date assignment, genetic algorithm, random search

Procedia PDF Downloads 362
2305 Data-Mining Approach to Analyzing Industrial Process Information for Real-Time Monitoring

Authors: Seung-Lock Seo

Abstract:

This work presents a data-mining empirical monitoring scheme for industrial processes with partially unbalanced data. Measurement data of good operations are relatively easy to gather, but in unusual special events or faults it is generally difficult to collect process information or almost impossible to analyze some noisy data of industrial processes. At this time some noise filtering techniques can be used to enhance process monitoring performance in a real-time basis. In addition, pre-processing of raw process data is helpful to eliminate unwanted variation of industrial process data. In this work, the performance of various monitoring schemes was tested and demonstrated for discrete batch process data. It showed that the monitoring performance was improved significantly in terms of monitoring success rate of given process faults.

Keywords: data mining, process data, monitoring, safety, industrial processes

Procedia PDF Downloads 383
2304 A New OvS Approach in Assembly Line Balancing Problem

Authors: P. Azimi, B. Behtoiy, A. A. Najafi, H. R. Charmchi

Abstract:

According to the previous studies, one of the most famous techniques which affect the efficiency of a production line is the assembly line balancing (ALB) technique. This paper examines the balancing effect of a whole production line of a real auto glass manufacturer in three steps. In the first step, processing time of each activity in the workstations is generated according to a practical approach. In the second step, the whole production process is simulated and the bottleneck stations have been identified, and finally in the third step, several improvement scenarios are generated to optimize the system throughput, and the best one is proposed. The main contribution of the current research is the proposed framework which combines two famous approaches including Assembly Line Balancing and Optimization via Simulation technique (OvS). The results show that the proposed framework could be applied in practical environments, easily.

Keywords: assembly line balancing problem, optimization via simulation, production planning

Procedia PDF Downloads 509
2303 Optimal Mother Wavelet Function for Shoulder Muscles of Upper Limb Amputees

Authors: Amanpreet Kaur

Abstract:

Wavelet transform (WT) is a powerful statistical tool used in applied mathematics for signal and image processing. The different mother, wavelet basis function, has been compared to select the optimal wavelet function that represents the electromyogram signal characteristics of upper limb amputees. Four different EMG electrode has placed on different location of shoulder muscles. Twenty one wavelet functions from different wavelet families were investigated. These functions included Daubechies (db1-db10), Symlets (sym1-sym5), Coiflets (coif1-coif5) and Discrete Meyer. Using mean square error value, the significance of the mother wavelet functions has been determined for teres, pectorals, and infraspinatus around shoulder muscles. The results show that the best mother wavelet is the db3 from the Daubechies family for efficient classification of the signal.

Keywords: Daubechies, upper limb amputation, shoulder muscles, Symlets, Coiflets

Procedia PDF Downloads 225
2302 CsPbBr₃@MOF-5-Based Single Drop Microextraction for in-situ Fluorescence Colorimetric Detection of Dechlorination Reaction

Authors: Yanxue Shang, Jingbin Zeng

Abstract:

Chlorobenzene homologues (CBHs) are a category of environmental pollutants that can not be ignored. They can stay in the environment for a long period and are potentially carcinogenic. The traditional degradation method of CBHs is dechlorination followed by sample preparation and analysis. This is not only time-consuming and laborious, but the detection and analysis processes are used in conjunction with large-scale instruments. Therefore, this can not achieve rapid and low-cost detection. Compared with traditional sensing methods, colorimetric sensing is simpler and more convenient. In recent years, chromaticity sensors based on fluorescence have attracted more and more attention. Compared with sensing methods based on changes in fluorescence intensity, changes in color gradients are easier to recognize by the naked eye. Accordingly, this work proposes to use single drop microextraction (SDME) technology to solve the above problems. After the dechlorination reaction was completed, the organic droplet extracts Cl⁻ and realizes fluorescence colorimetric sensing at the same time. This method was integrated sample processing and visual in-situ detection, simplifying the detection process. As a fluorescence colorimetric sensor material, CsPbBr₃ was encapsulated in MOF-5 to construct CsPbBr₃@MOF-5 fluorescence colorimetric composite. Then the fluorescence colorimetric sensor was constructed by dispersing the composite in SDME organic droplets. When the Br⁻ in CsPbBr₃ exchanges with Cl⁻ produced by the dechlorination reactions, it is converted into CsPbCl₃. The fluorescence color of the single droplet of SDME will change from green to blue emission, thereby realizing visual observation. Therein, SDME can enhance the concentration and enrichment of Cl⁻ and instead of sample pretreatment. The fluorescence color change of CsPbBr₃@MOF-5 can replace the detection process of large-scale instruments to achieve real-time rapid detection. Due to the absorption ability of MOF-5, it can not only improve the stability of CsPbBr₃, but induce the adsorption of Cl⁻. Simultaneously, accelerate the exchange of Br- and Cl⁻ in CsPbBr₃ and the detection process of Cl⁻. The absorption process was verified by density functional theory (DFT) calculations. This method exhibits exceptional linearity for Cl⁻ in the range of 10⁻² - 10⁻⁶ M (10000 μM - 1 μM) with a limit of detection of 10⁻⁷ M. Whereafter, the dechlorination reactions of different kinds of CBHs were also carried out with this method, and all had satisfactory detection ability. Also verified the accuracy by gas chromatography (GC), and it was found that the SDME we developed in this work had high credibility. In summary, the in-situ visualization method of dechlorination reaction detection was a combination of sample processing and fluorescence colorimetric sensing. Thus, the strategy researched herein represents a promising method for the visual detection of dechlorination reactions and can be extended for applications in environments, chemical industries, and foods.

Keywords: chlorobenzene homologues, colorimetric sensor, metal halide perovskite, metal-organic frameworks, single drop microextraction

Procedia PDF Downloads 131
2301 Efficiently Degradation of Perfluorooctanoic Acid, an Emerging Contaminant, by a Hybrid Process of Membrane Distillation Process and Electro-Fenton

Authors: Afrouz Yousefi, Mohtada Sadrzadeh

Abstract:

The widespread presence of poly- and perfluoroalkyl substances (PFAS) poses a significant concern due to their ability to accumulate in living organisms and their persistence in the environment, thanks to their robust carbon-fluorine (C-F) bonds, which require substantial energy to break (485 kJ/mol). The prevalence of toxic PFAS compounds can be highly detrimental to ecosystems, wildlife, and human health. Ongoing efforts are dedicated to investigating methods for fully breaking down and eliminating PFAS from the environment. Among the various techniques employed, advanced oxidation processes have shown promise in completely breaking down emerging contaminants in wastewater. However, the drawback lies in the relatively slow reaction rates of these processes and the substantial energy input required, which currently impedes their widespread commercial adoption. We developed a hybrid process, comprising electro-Fenton as an advanced oxidation process and membrane distillation, to simultaneously degrade organic PFAS pollutants and extract pure water from the mixture. In this study, environmentally persistent perfluorooctanoic acid (PFOA), as an emerging contaminant, was used to study the effectiveness of the electro-Fenton/membrane distillation hybrid system. The PFOA degradation studies were conducted in two modes: electro-Fenton and electro-Fenton coupled with membrane distillation. High-performance liquid chromatography with ultraviolet detection (HPLC-UV), ion-chromatography (measuring fluoride ion concentration), total organic carbon (TOC) decay, mineralization current efficiency (MCE), and specific energy consumption (SEC) were evaluated for a single EF and hybrid EF-MD processes. In contrast to a single EF reaction, TOC decay improved significantly in the EF-MD process. Overall, the MCE of hybrid processes surpassed 100% while it remained under 50% for a single EF reaction. Calculations of specific energy consumption (SEC) demonstrated a substantial decrease of nearly one-third in energy usage when integrating the EF reaction with the MD process.

Keywords: water treatment, PFAS, membrane distillation, electro-Fenton, advanced oxidation

Procedia PDF Downloads 48
2300 Analysis of Spatial and Temporal Data Using Remote Sensing Technology

Authors: Kapil Pandey, Vishnu Goyal

Abstract:

Spatial and temporal data analysis is very well known in the field of satellite image processing. When spatial data are correlated with time, series analysis it gives the significant results in change detection studies. In this paper the GIS and Remote sensing techniques has been used to find the change detection using time series satellite imagery of Uttarakhand state during the years of 1990-2010. Natural vegetation, urban area, forest cover etc. were chosen as main landuse classes to study. Landuse/ landcover classes within several years were prepared using satellite images. Maximum likelihood supervised classification technique was adopted in this work and finally landuse change index has been generated and graphical models were used to present the changes.

Keywords: GIS, landuse/landcover, spatial and temporal data, remote sensing

Procedia PDF Downloads 417
2299 Functionalized PU Foam for Water Filtration

Authors: Nidal H. Abu-Zahra, Subhashini Gunashekar

Abstract:

Polyurethane foam is functionalized with Sulfonic acid groups to remove lead ions (Pb2+) from drinking water through a action exchange process. The synthesis is based on addition polymerization of the -NCO groups of an isocyanine with the –OH groups of a polio to form the urethane. Toluene-diisocyanateis reacted with Polypropylene glycol to form a linear pre-polymer, which is further polymerized using a chain extender, N, N-bis(2-hydorxyethyl)-2-aminoethane-sulfonic acid (BES). BES acts as a functional group site to exchange Pb2+ ions. A set of experiments was designed to study the effect of various processing parameters on the performance of the synthesized foam. The maximum Pb2+ ion exchange capacity of the foam was found to be 47ppb/g from a 100ppb Pb2+ solution over a period of 60 minutes. A multistage batch filtration process increased the lead removal to 50-54ppb/3g of foam over a period of 90 minutes.

Keywords: adsorption, functionalized, ion exchange, polyurethane, sulfonic

Procedia PDF Downloads 235
2298 Channel Estimation for LTE Downlink

Authors: Rashi Jain

Abstract:

The LTE systems employ Orthogonal Frequency Division Multiplexing (OFDM) as the multiple access technology for the Downlink channels. For enhanced performance, accurate channel estimation is required. Various algorithms such as Least Squares (LS), Minimum Mean Square Error (MMSE) and Recursive Least Squares (RLS) can be employed for the purpose. The paper proposes channel estimation algorithm based on Kalman Filter for LTE-Downlink system. Using the frequency domain pilots, the initial channel response is obtained using the LS criterion. Then Kalman Filter is employed to track the channel variations in time-domain. To suppress the noise within a symbol, threshold processing is employed. The paper draws comparison between the LS, MMSE, RLS and Kalman filter for channel estimation. The parameters for evaluation are Bit Error Rate (BER), Mean Square Error (MSE) and run-time.

Keywords: LTE, channel estimation, OFDM, RLS, Kalman filter, threshold

Procedia PDF Downloads 338
2297 Performance Analysis with the Combination of Visualization and Classification Technique for Medical Chatbot

Authors: Shajida M., Sakthiyadharshini N. P., Kamalesh S., Aswitha B.

Abstract:

Natural Language Processing (NLP) continues to play a strategic part in complaint discovery and medicine discovery during the current epidemic. This abstract provides an overview of performance analysis with a combination of visualization and classification techniques of NLP for a medical chatbot. Sentiment analysis is an important aspect of NLP that is used to determine the emotional tone behind a piece of text. This technique has been applied to various domains, including medical chatbots. In this, we have compared the combination of the decision tree with heatmap and Naïve Bayes with Word Cloud. The performance of the chatbot was evaluated using accuracy, and the results indicate that the combination of visualization and classification techniques significantly improves the chatbot's performance.

Keywords: sentimental analysis, NLP, medical chatbot, decision tree, heatmap, naïve bayes, word cloud

Procedia PDF Downloads 56
2296 Digital Cinema Watermarking State of Art and Comparison

Authors: H. Kelkoul, Y. Zaz

Abstract:

Nowadays, the vigorous popularity of video processing techniques has resulted in an explosive growth of multimedia data illegal use. So, watermarking security has received much more attention. The purpose of this paper is to explore some watermarking techniques in order to observe their specificities and select the finest methods to apply in digital cinema domain against movie piracy by creating an invisible watermark that includes the date, time and the place where the hacking was done. We have studied three principal watermarking techniques in the frequency domain: Spread spectrum, Wavelet transform domain and finally the digital cinema watermarking transform domain. In this paper, a detailed technique is presented where embedding is performed using direct sequence spread spectrum technique in DWT transform domain. Experiment results shows that the algorithm provides high robustness and good imperceptibility.

Keywords: digital cinema, watermarking, wavelet DWT, spread spectrum, JPEG2000 MPEG4

Procedia PDF Downloads 240
2295 Road Vehicle Recognition Using Magnetic Sensing Feature Extraction and Classification

Authors: Xiao Chen, Xiaoying Kong, Min Xu

Abstract:

This paper presents a road vehicle detection approach for the intelligent transportation system. This approach mainly uses low-cost magnetic sensor and associated data collection system to collect magnetic signals. This system can measure the magnetic field changing, and it also can detect and count vehicles. We extend Mel Frequency Cepstral Coefficients to analyze vehicle magnetic signals. Vehicle type features are extracted using representation of cepstrum, frame energy, and gap cepstrum of magnetic signals. We design a 2-dimensional map algorithm using Vector Quantization to classify vehicle magnetic features to four typical types of vehicles in Australian suburbs: sedan, VAN, truck, and bus. Experiments results show that our approach achieves a high level of accuracy for vehicle detection and classification.

Keywords: vehicle classification, signal processing, road traffic model, magnetic sensing

Procedia PDF Downloads 303
2294 Constructing the Density of States from the Parallel Wang Landau Algorithm Overlapping Data

Authors: Arman S. Kussainov, Altynbek K. Beisekov

Abstract:

This work focuses on building an efficient universal procedure to construct a single density of states from the multiple pieces of data provided by the parallel implementation of the Wang Landau Monte Carlo based algorithm. The Ising and Pott models were used as the examples of the two-dimensional spin lattices to construct their densities of states. Sampled energy space was distributed between the individual walkers with certain overlaps. This was made to include the latest development of the algorithm as the density of states replica exchange technique. Several factors of immediate importance for the seamless stitching process have being considered. These include but not limited to the speed and universality of the initial parallel algorithm implementation as well as the data post-processing to produce the expected smooth density of states.

Keywords: density of states, Monte Carlo, parallel algorithm, Wang Landau algorithm

Procedia PDF Downloads 387
2293 Food Safety Aspects of Pesticide Residues in Spice Paprika

Authors: Sz. Klátyik, B. Darvas, M. Mörtl, M. Ottucsák, E. Takács, H. Bánáti, L. Simon, G. Gyurcsó, A. Székács

Abstract:

Environmental and health safety of condiments used for spicing food products in food processing or by culinary means receive relatively low attention, even though possible contamination of spices may affect food quality and safety. Contamination surveys mostly focus on microbial contaminants or their secondary metabolites, mycotoxins. Chemical contaminants, particularly pesticide residues, however, are clearly substantial factors in the case of given condiments in the Capsicum family including spice paprika and chilli. To assess food safety and support the quality of the Hungaricum product spice paprika, the pesticide residue status of spice paprika and chilli is assessed on the basis of reported pesticide contamination cases and non-compliances in the Rapid Alert System for Food and Feed of the European Union since 1998.

Keywords: spice paprika, Capsicum, pesticide residues, RASFF

Procedia PDF Downloads 375
2292 New Approaches for the Handwritten Digit Image Features Extraction for Recognition

Authors: U. Ravi Babu, Mohd Mastan

Abstract:

The present paper proposes a novel approach for handwritten digit recognition system. The present paper extract digit image features based on distance measure and derives an algorithm to classify the digit images. The distance measure can be performing on the thinned image. Thinning is the one of the preprocessing technique in image processing. The present paper mainly concentrated on an extraction of features from digit image for effective recognition of the numeral. To find the effectiveness of the proposed method tested on MNIST database, CENPARMI, CEDAR, and newly collected data. The proposed method is implemented on more than one lakh digit images and it gets good comparative recognition results. The percentage of the recognition is achieved about 97.32%.

Keywords: handwritten digit recognition, distance measure, MNIST database, image features

Procedia PDF Downloads 445
2291 Investigation of the Effects of Visually Disabled and Typical Development Students on Their Multiple Intelligence by Applying Abacus and Right Brain Training

Authors: Sidika Di̇lşad Kaya, Ahmet Seli̇m Kaya, Ibrahi̇m Eri̇k, Havva Yaldiz, Yalçin Kaya

Abstract:

The aim of this study was to reveal the effects of right brain development on reading, comprehension, learning and concentration levels and rapid processing skills in students with low vision and students with standard development, and to explore the effects of right and left brain integration on students' academic success and the permanence of the learned knowledge. A total of 68 students with a mean age of 10.01±0.12 were included in the study, 58 of them with standard development, 9 partially visually impaired and 1 totally visually disabled student. The student with a total visual impairment could not participate in the reading speed test due to her total visual impairment. The following data were measured in the participant students before the project; Reading speed measurement in 1 minute, Reading comprehension questions, Burdon attention test, 50 questions of math quiz timed with a stopwatch. Participants were trained for 3 weeks, 5 days a week, for a total of two hours a day. In this study, right-brain developing exercises were carried out with the use of an abacus, and it was aimed to develop both mathematical and attention of students with questions prepared with numerical data taken from fairy tale activities. Among these problems, the study was supported with multiple-choice, 5W (what, where, who, why, when?), 1H (how?) questions along with true-false and fill-in-the-blank activities. By using memory cards, students' short-term memories were strengthened, photographic memory studies were conducted and their visual intelligence was supported. Auditory intelligence was supported by aiming to make calculations by using the abacus in the minds of the students with the numbers given aurally. When calculating the numbers by touching the real abacus, the development of students' tactile intelligence is enhanced. Research findings were analyzed in SPSS program, Kolmogorov Smirnov test was used for normality analysis. Since the variables did not show normal distribution, Wilcoxon test, one of the non-parametric tests, was used to compare the dependent groups. Statistical significance level was accepted as 0.05. The reading speed of the participants was 83.54±33.03 in the pre-test and 116.25±38.49 in the post-test. Narration pre-test 69.71±25.04 post-test 97.06±6.70; BURDON pretest 84.46±14.35 posttest 95.75±5.67; rapid math processing skills pretest 90.65±10.93, posttest 98.18±2.63 (P<0.05). It was determined that the pre-test and post-test averages of students with typical development and students with low vision were also significant for all four values (p<0.05). As a result of the data obtained from the participants, it is seen that the study was effective in terms of measurement parameters, and the findings were statistically significant. Therefore, it is recommended to use the method widely.

Keywords: Abacus, reading speed, multiple intelligences, right brain training, visually impaired

Procedia PDF Downloads 163
2290 Analysis of Shrinkage Effect during Mercerization on Himalayan Nettle, Cotton and Cotton/Nettle Yarn Blends

Authors: Reena Aggarwal, Neha Kestwal

Abstract:

The Himalayan Nettle (Girardinia diversifolia) has been used for centuries as fibre and food source by Himalayan communities. Himalayan Nettle is a natural cellulosic fibre that can be handled in the same way as other cellulosic fibres. The Uttarakhand Bamboo and Fibre Development Board based in Uttarakhand, India is working extensively with the nettle fibre to explore the potential of nettle for textile production in the region. The fiber is a potential resource for rural enterprise development for some high altitude pockets of the state and traditionally the plant fibre is used for making domestic products like ropes and sacks. Himalayan Nettle is an unconventional natural fiber with functional characteristics of shrink resistance, degree of pathogen and fire resistance and can blend nicely with other fibres. Most importantly, they generate mainly organic wastes and leave residues that are 100% biodegradable. The fabrics may potentially be reused or re-manufactured and can also be used as a source of cellulose feedstock for regenerated cellulosic products. Being naturally bio- degradable, the fibre can be composted if required. Though a lot of research activities and training are directed towards fibre extraction and processing techniques in different craft clusters villagers of different clusters of Uttarkashi, Chamoli and Bageshwar of Uttarakhand like retting and Degumming process, very little is been done to analyse the crucial properties of nettle fiber like shrinkage and wash fastness. These properties are very crucial to obtain desired quality of fibre for further processing of yarn making and weaving and in developing these fibers into fine saleable products. This research therefore is focused towards various on-field experiments which were focused on shrinkage properties conducted on cotton, nettle and cotton/nettle blended yarn samples. The objective of the study was to analyze the scope of the blended fiber for developing into wearable fabrics. For the study, after conducting the initial fiber length and fineness testing, cotton and nettle fibers were mixed in 60:40 ratio and five varieties of yarns were spun in open end spinning mill having yarn count of 3s, 5s, 6s, 7s and 8s. Samples of 100% Nettle 100% cotton fibers in 8s count were also developed for the study. All the six varieties of yarns were tested with shrinkage test and results were critically analyzed as per ASTM method D2259. It was observed that 100% Nettle has a least shrinkage of 3.36% while pure cotton has shrinkage approx. 13.6%. Yarns made of 100% Cotton exhibits four times more shrinkage than 100% Nettle. The results also show that cotton and Nettle blended yarn exhibit lower shrinkage than 100% cotton yarn. It was thus concluded that as the ratio of nettle increases in the samples, the shrinkage decreases in the samples. These results are very crucial for Uttarakhand people who want to commercially exploit the abundant nettle fiber for generating sustainable employment.

Keywords: Himalayan nettle, sustainable, shrinkage, blending

Procedia PDF Downloads 220
2289 Adaptive Filtering in Subbands for Supervised Source Separation

Authors: Bruna Luisa Ramos Prado Vasques, Mariane Rembold Petraglia, Antonio Petraglia

Abstract:

This paper investigates MIMO (Multiple-Input Multiple-Output) adaptive filtering techniques for the application of supervised source separation in the context of convolutive mixtures. From the observation that there is correlation among the signals of the different mixtures, an improvement in the NSAF (Normalized Subband Adaptive Filter) algorithm is proposed in order to accelerate its convergence rate. Simulation results with mixtures of speech signals in reverberant environments show the superior performance of the proposed algorithm with respect to the performances of the NLMS (Normalized Least-Mean-Square) and conventional NSAF, considering both the convergence speed and SIR (Signal-to-Interference Ratio) after convergence.

Keywords: adaptive filtering, multi-rate processing, normalized subband adaptive filter, source separation

Procedia PDF Downloads 416
2288 Supercritical Hydrothermal and Subcritical Glycolysis Conversion of Biomass Waste to Produce Biofuel and High-Value Products

Authors: Chiu-Hsuan Lee, Min-Hao Yuan, Kun-Cheng Lin, Qiao-Yin Tsai, Yun-Jie Lu, Yi-Jhen Wang, Hsin-Yi Lin, Chih-Hua Hsu, Jia-Rong Jhou, Si-Ying Li, Yi-Hung Chen, Je-Lueng Shie

Abstract:

Raw food waste has a high-water content. If it is incinerated, it will increase the cost of treatment. Therefore, composting or energy is usually used. There are mature technologies for composting food waste. Odor, wastewater, and other problems are serious, but the output of compost products is limited. And bakelite is mainly used in the manufacturing of integrated circuit boards. It is hard to directly recycle and reuse due to its hard structure and also difficult to incinerate and produce air pollutants due to incomplete incineration. In this study, supercritical hydrothermal and subcritical glycolysis thermal conversion technology is used to convert biomass wastes of bakelite and raw kitchen wastes to carbon materials and biofuels. Batch carbonization tests are performed under high temperature and pressure conditions of solvents and different operating conditions, including wet and dry base mixed biomass. This study can be divided into two parts. In the first part, bakelite waste is performed as dry-based industrial waste. And in the second part, raw kitchen wastes (lemon, banana, watermelon, and pineapple peel) are used as wet-based biomass ones. The parameters include reaction temperature, reaction time, mass-to-solvent ratio, and volume filling rates. The yield, conversion, and recovery rates of products (solid, gas, and liquid) are evaluated and discussed. The results explore the benefits of synergistic effects in thermal glycolysis dehydration and carbonization on the yield and recovery rate of solid products. The purpose is to obtain the optimum operating conditions. This technology is a biomass-negative carbon technology (BNCT); if it is combined with carbon capture and storage (BECCS), it can provide a new direction for 2050 net zero carbon dioxide emissions (NZCDE).

Keywords: biochar, raw food waste, bakelite, supercritical hydrothermal, subcritical glycolysis, biofuels

Procedia PDF Downloads 160