Search results for: initial detection
649 Using A Hybrid Algorithm to Improve the Quality of Services in Multicast Routing Problem
Authors: Mohammad Reza Karami Nejad
Abstract:
A hybrid learning automata-genetic algorithm (HLGA) is proposed to solve QoS routing optimization problem of next generation networks. The algorithm complements the advantages of the learning Automato Algorithm(LA) and Genetic Algorithm(GA). It firstly uses the good global search capability of LA to generate initial population needed by GA, then it uses GA to improve the Quality of Service(QoS) and acquiring the optimization tree through new algorithms for crossover and mutation operators which are an NP-Complete problem. In the proposed algorithm, the connectivity matrix of edges is used for genotype representation. Some novel heuristics are also proposed for mutation, crossover, and creation of random individuals. We evaluate the performance and efficiency of the proposed HLGA-based algorithm in comparison with other existing heuristic and GA-based algorithms by the result of simulation. Simulation results demonstrate that this paper proposed algorithm not only has the fast calculating speed and high accuracy but also can improve the efficiency in Next Generation Networks QoS routing. The proposed algorithm has overcome all of the previous algorithms in the literature.
Keywords: Routing, Quality of Service, Multicaset, Learning Automata, Genetic, Next Generation Networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1740648 On the Bootstrap P-Value Method in Identifying out of Control Signals in Multivariate Control Chart
Authors: O. Ikpotokin
Abstract:
In any production process, every product is aimed to attain a certain standard, but the presence of assignable cause of variability affects our process, thereby leading to low quality of product. The ability to identify and remove this type of variability reduces its overall effect, thereby improving the quality of the product. In case of a univariate control chart signal, it is easy to detect the problem and give a solution since it is related to a single quality characteristic. However, the problems involved in the use of multivariate control chart are the violation of multivariate normal assumption and the difficulty in identifying the quality characteristic(s) that resulted in the out of control signals. The purpose of this paper is to examine the use of non-parametric control chart (the bootstrap approach) for obtaining control limit to overcome the problem of multivariate distributional assumption and the p-value method for detecting out of control signals. Results from a performance study show that the proposed bootstrap method enables the setting of control limit that can enhance the detection of out of control signals when compared, while the p-value method also enhanced in identifying out of control variables.
Keywords: Bootstrap control limit, p-value method, out-of-control signals, p-value, quality characteristics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1015647 Distributed Estimation Using an Improved Incremental Distributed LMS Algorithm
Authors: Amir Rastegarnia, Mohammad Ali Tinati, Azam Khalili
Abstract:
In this paper we consider the problem of distributed adaptive estimation in wireless sensor networks for two different observation noise conditions. In the first case, we assume that there are some sensors with high observation noise variance (noisy sensors) in the network. In the second case, different variance for observation noise is assumed among the sensors which is more close to real scenario. In both cases, an initial estimate of each sensor-s observation noise is obtained. For the first case, we show that when there are such sensors in the network, the performance of conventional distributed adaptive estimation algorithms such as incremental distributed least mean square (IDLMS) algorithm drastically decreases. In addition, detecting and ignoring these sensors leads to a better performance in a sense of estimation. In the next step, we propose a simple algorithm to detect theses noisy sensors and modify the IDLMS algorithm to deal with noisy sensors. For the second case, we propose a new algorithm in which the step-size parameter is adjusted for each sensor according to its observation noise variance. As the simulation results show, the proposed methods outperforms the IDLMS algorithm in the same condition.
Keywords: Distributes estimation, sensor networks, adaptive filter, IDLMS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1448646 Bearing Capacity of Sheet Hanger Connection to the Trapezoidal Metal Sheet
Authors: Kateřina Jurdová
Abstract:
Hanging to the trapezoidal sheet by decking hanger is a very widespread solution used in civil engineering to lead the distribution of energy, sanitary, air distribution system etc. under the roof or floor structure. The trapezoidal decking hanger is usually a part of the whole installation system for specific distribution medium. The leading companies offer installation systems for each specific distribution e.g. pipe rings, sprinkler systems, installation channels etc. Every specific part is connected to the base connector which is decking hanger. The own connection has three main components: decking hanger, threaded bar with nuts and web of trapezoidal sheet. The aim of this contribution is determinate the failure mechanism of each component in connection. Load bearing capacity of most components in connection could be calculated by formulas in European codes. This contribution is focused on problematic of bearing resistance of threaded bar in web of trapezoidal sheet. This issue is studied by experimental research and numerical modelling. This contribution presented the initial results of experiment which is compared with numerical model of specimen.
Keywords: Decking hanger, concentrated load, connection, load bearing capacity, trapezoidal metal sheet.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2652645 Determination of Water Pollution and Water Quality with Decision Trees
Authors: Çiğdem Bakır, Mecit Yüzkat
Abstract:
With the increasing emphasis on water quality worldwide, the search for and expanding the market for new and intelligent monitoring systems has increased. The current method is the laboratory process, where samples are taken from bodies of water, and tests are carried out in laboratories. This method is time-consuming, a waste of manpower and uneconomical. To solve this problem, we used machine learning methods to detect water pollution in our study. We created decision trees with the Orange3 software used in the study and tried to determine all the factors that cause water pollution. An automatic prediction model based on water quality was developed by taking many model inputs such as water temperature, pH, transparency, conductivity, dissolved oxygen, and ammonia nitrogen with machine learning methods. The proposed approach consists of three stages: Preprocessing of the data used, feature detection and classification. We tried to determine the success of our study with different accuracy metrics and the results were presented comparatively. In addition, we achieved approximately 98% success with the decision tree.
Keywords: Decision tree, water quality, water pollution, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 269644 Automatic Reusability Appraisal of Software Components using Neuro-fuzzy Approach
Authors: Parvinder S. Sandhu, Hardeep Singh
Abstract:
Automatic reusability appraisal could be helpful in evaluating the quality of developed or developing reusable software components and in identification of reusable components from existing legacy systems; that can save cost of developing the software from scratch. But the issue of how to identify reusable components from existing systems has remained relatively unexplored. In this paper, we have mentioned two-tier approach by studying the structural attributes as well as usability or relevancy of the component to a particular domain. Latent semantic analysis is used for the feature vector representation of various software domains. It exploits the fact that FeatureVector codes can be seen as documents containing terms -the idenifiers present in the components- and so text modeling methods that capture co-occurrence information in low-dimensional spaces can be used. Further, we devised Neuro- Fuzzy hybrid Inference System, which takes structural metric values as input and calculates the reusability of the software component. Decision tree algorithm is used to decide initial set of fuzzy rules for the Neuro-fuzzy system. The results obtained are convincing enough to propose the system for economical identification and retrieval of reusable software components.Keywords: Clustering, ID3, LSA, Neuro-fuzzy System, SVD
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1664643 Real-time Laser Monitoring based on Pipe Detective Operation
Authors: Mongkorn Klingajay, Tawatchai Jitson
Abstract:
The pipe inspection operation is the difficult detective performance. Almost applications are mainly relies on a manual recognition of defective areas that have carried out detection by an engineer. Therefore, an automation process task becomes a necessary in order to avoid the cost incurred in such a manual process. An automated monitoring method to obtain a complete picture of the sewer condition is proposed in this work. The focus of the research is the automated identification and classification of discontinuities in the internal surface of the pipe. The methodology consists of several processing stages including image segmentation into the potential defect regions and geometrical characteristic features. Automatic recognition and classification of pipe defects are carried out by means of using an artificial neural network technique (ANN) based on Radial Basic Function (RBF). Experiments in a realistic environment have been conducted and results are presented.Keywords: Artificial neural network, Radial basic function, Curve fitting, CCTV, Image segmentation, Data acquisition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1822642 Detection of Max. Optical Gain by Erbium Doped Fiber Amplifier
Authors: Abdulamgid.T. Bouzed, Suleiman. M. Elhamali
Abstract:
The technical realization of data transmission using glass fiber began after the development of diode laser in year 1962. The erbium doped fiber amplifiers (EDFA's) in high speed networks allow information to be transmitted over longer distances without using of signal amplification repeaters. These kinds of fibers are doped with erbium atoms which have energy levels in its atomic structure for amplifying light at 1550nm. When a carried signal wave at 1550nm enters the erbium fiber, the light stimulates the excited erbium atoms which pumped with laser beam at 980nm as additional light. The wavelength and intensity of the semiconductor lasers depend on the temperature of active zone and the injection current. The present paper shows the effect of the diode lasers temperature and injection current on the optical amplification. From the results of in- and output power one may calculate the max. optical gain by erbium doped fiber amplifier.Keywords: Amplifier, erbium doped fiber, gain, lasers, temperature.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2147641 The Effects of Detector Spacing on Travel Time Prediction on Freeways
Authors: Piyali Chaudhuri, Peter T. Martin, Aleksandar Z. Stevanovic, Chongkai Zhu
Abstract:
Loop detectors report traffic characteristics in real time. They are at the core of traffic control process. Intuitively, one would expect that as density of detection increases, so would the quality of estimates derived from detector data. However, as detector deployment increases, the associated operating and maintenance cost increases. Thus, traffic agencies often need to decide where to add new detectors and which detectors should continue receiving maintenance, given their resource constraints. This paper evaluates the effect of detector spacing on freeway travel time estimation. A freeway section (Interstate-15) in Salt Lake City metropolitan region is examined. The research reveals that travel time accuracy does not necessarily deteriorate with increased detector spacing. Rather, the actual location of detectors has far greater influence on the quality of travel time estimates. The study presents an innovative computational approach that delivers optimal detector locations through a process that relies on Genetic Algorithm formulation.Keywords: Detector, Freeway, Genetic algorithm, Travel timeestimate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1673640 Study of the Antimicrobial Activity of Aminoreductone against Pathogenic Bacteria in Comparison with Other Antibiotics
Authors: Vu Thu Trang, Lam Xuan Thanh, Samira Sarter, Tomoko Shimamura, Hiroaki Takeuchi
Abstract:
Antimicrobial activities of aminoreductone (AR), a product formed in the initial stage of Maillard reaction, were screened against pathogenic bacteria. A significant growth inhibition of AR against all 7 isolates (Staphylococcus aureus ATCC® 25923™, Salmonella typhimurium ATCC® 14028™, Bacillus cereus ATCC® 13061™, Bacillus subtilis ATCC® 11774™, Escherichia coli ATCC® 25922™, Enterococcus faecalis ATCC® 29212™, Listeria innocua ATCC® 33090™) were observed by the standard disc diffusion methods. The inhibition zone for each isolate by AR (2.5 mg) ranged from 15±0mm to 28.3±0.4mm in diameter. The minimum inhibitory concentration (MIC) of AR ranging from 20mM to 26mM was proven in the 7 isolates tested. AR also showed the similar effect of growth inhibition in comparison with antibiotics frequently used for the treatment of infections bacteria, such as amikacin, ciprofloxacin, meropennem and levofloxacin. The results indicated that foods containing AR are valuable sources of bioactive compounds towards pathogenic bacteria.
Keywords: Pathogenic bacteria, aminoreductone, Maillard reaction, antimicrobial activity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2308639 Visualization of Latent Sweat Fingerprints Deposit on Paper by Infrared Radiation and Blue Light
Authors: Xiaochun Huang, Xuejun Zhao, Yun Zou, Feiyu Yang, Wenbin Liu, Nan Deng, Ming Zhang, Nengbin Cai
Abstract:
A simple device termed infrared radiation (IR) was developed for rapid visualization of sweat fingerprints deposit on paper with blue light (450 nm, 11 W). In this approach, IR serves as the pretreatment device before the sweat fingerprints was illuminated by blue light. An annular blue light source was adopted for visualizing latent sweat fingerprints. Sample fingerprints were examined under various conditions after deposition, and experimental results indicate that the recovery rate of the latent sweat fingerprints is in the range of 50%-100% without chemical treatments. A mechanism for the observed visibility is proposed based on transportation and re-impregnation of fluorescer in paper at the region of water. And further exploratory experimental results gave the full support to the visible mechanism. Therefore, such a method as IR-pretreated in detecting latent fingerprints may be better for examination in the case where biological information of samples is needed for consequent testing.
Keywords: Forensic science, visualization, infrared radiation, blue light, latent sweat fingerprints, detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1467638 Biometric Steganography Using Variable Length Embedding
Authors: Souvik Bhattacharyya, Indradip Banerjee, Anumoy Chakraborty, Gautam Sanyal
Abstract:
Recent growth in digital multimedia technologies has presented a lot of facilities in information transmission, reproduction and manipulation. Therefore, the concept of information security is one of the superior articles in the present day situation. The biometric information security is one of the information security mechanisms. It has the advantages as well as disadvantages. The biometric system is at risk to a range of attacks. These attacks are anticipated to bypass the security system or to suspend the normal functioning. Various hazards have been discovered while using biometric system. Proper use of steganography greatly reduces the risks in biometric systems from the hackers. Steganography is one of the fashionable information hiding technique. The goal of steganography is to hide information inside a cover medium like text, image, audio, video etc. through which it is not possible to detect the existence of the secret information. Here in this paper a new security concept has been established by making the system more secure with the help of steganography along with biometric security. Here the biometric information has been embedded to a skin tone portion of an image with the help of proposed steganographic technique.
Keywords: Biometrics, Skin tone detection, Series, Polynomial, Cover Image, Stego Image.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2674637 Genetic Content-Based MP3 Audio Watermarking in MDCT Domain
Authors: N. Moghadam, H. Sadeghi
Abstract:
In this paper a novel scheme for watermarking digital audio during its compression to MPEG-1 Layer III format is proposed. For this purpose we slightly modify some of the selected MDCT coefficients, which are used during MPEG audio compression procedure. Due to the possibility of modifying different MDCT coefficients, there will be different choices for embedding the watermark into audio data, considering robustness and transparency factors. Our proposed method uses a genetic algorithm to select the best coefficients to embed the watermark. This genetic selection is done according to the parameters that are extracted from the perceptual content of the audio to optimize the robustness and transparency of the watermark. On the other hand the watermark security is increased due to the random nature of the genetic selection. The information of the selected MDCT coefficients that carry the watermark bits, are saves in a database for future extraction of the watermark. The proposed method is suitable for online MP3 stores to pursue illegal copies of musical artworks. Experimental results show that the detection ratio of the watermarks at the bitrate of 128kbps remains above 90% while the inaudibility of the watermark is preserved.Keywords: Content-Based Audio Watermarking, Genetic AudioWatermarking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1518636 Binary Decision Diagrams: An Improved Variable Ordering using Graph Representation of Boolean Functions
Authors: P.W. C. Prasad, A. Assi, A. Harb, V.C. Prasad
Abstract:
This paper presents an improved variable ordering method to obtain the minimum number of nodes in Reduced Ordered Binary Decision Diagrams (ROBDD). The proposed method uses the graph topology to find the best variable ordering. Therefore the input Boolean function is converted to a unidirectional graph. Three levels of graph parameters are used to increase the probability of having a good variable ordering. The initial level uses the total number of nodes (NN) in all the paths, the total number of paths (NP) and the maximum number of nodes among all paths (MNNAP). The second and third levels use two extra parameters: The shortest path among two variables (SP) and the sum of shortest path from one variable to all the other variables (SSP). A permutation of the graph parameters is performed at each level for each variable order and the number of nodes is recorded. Experimental results are promising; the proposed method is found to be more effective in finding the variable ordering for the majority of benchmark circuits.
Keywords: Binary decision diagrams, graph representation, Boolean functions representation, variable ordering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2120635 Mechanical Characterization and Impact Study on the Environment of Raw Sediments and Sediments Dehydrated by Addition of Polymer
Authors: A. Kasmi, N. E. Abriak, M. Benzerzour, I. Shahrour
Abstract:
Large volumes of river sediments are dredged each year in Europe in order to maintain harbour activities and prevent floods. The management of this sediment has become increasingly complex. Several European projects were implemented to find environmentally sound solutions for these materials. The main objective of this study is to show the ability of river sediment to be used in road. Since sediments contain a high amount of water, then a dehydrating treatment by addition of the flocculation aid has been used. Firstly, a lot of physical characteristics are measured and discussed for a better identification of the raw sediment and this dehydrated sediment by addition the flocculation aid. The identified parameters are, for example, the initial water content, the density, the organic matter content, the grain size distribution, the liquid limit and plastic limit and geotechnical parameters. The environmental impacts of the used material were evaluated. The results obtained show that there is a slight change on the physical-chemical and geotechnical characteristics of sediment after dehydration by the addition of polymer. However, these sediments cannot be used in road construction.
Keywords: River sediment, dehydration, flocculation aid, characteristics, environmental impacts, road construction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1300634 Sparse Coding Based Classification of Electrocardiography Signals Using Data-Driven Complete Dictionary Learning
Authors: Fuad Noman, Sh-Hussain Salleh, Chee-Ming Ting, Hadri Hussain, Syed Rasul
Abstract:
In this paper, a data-driven dictionary approach is proposed for the automatic detection and classification of cardiovascular abnormalities. Electrocardiography (ECG) signal is represented by the trained complete dictionaries that contain prototypes or atoms to avoid the limitations of pre-defined dictionaries. The data-driven trained dictionaries simply take the ECG signal as input rather than extracting features to study the set of parameters that yield the most descriptive dictionary. The approach inherently learns the complicated morphological changes in ECG waveform, which is then used to improve the classification. The classification performance was evaluated with ECG data under two different preprocessing environments. In the first category, QT-database is baseline drift corrected with notch filter and it filters the 60 Hz power line noise. In the second category, the data are further filtered using fast moving average smoother. The experimental results on QT database confirm that our proposed algorithm shows a classification accuracy of 92%.Keywords: Electrocardiogram, dictionary learning, sparse coding, classification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2098633 Algorithm for Path Recognition in-between Tree Rows for Agricultural Wheeled-Mobile Robots
Authors: Anderson Rocha, Pedro Miguel de Figueiredo Dinis Oliveira Gaspar
Abstract:
Machine vision has been widely used in recent years in agriculture, as a tool to promote the automation of processes and increase the levels of productivity. The aim of this work is the development of a path recognition algorithm based on image processing to guide a terrestrial robot in-between tree rows. The proposed algorithm was developed using the software MATLAB, and it uses several image processing operations, such as threshold detection, morphological erosion, histogram equalization and the Hough transform, to find edge lines along tree rows on an image and to create a path to be followed by a mobile robot. To develop the algorithm, a set of images of different types of orchards was used, which made possible the construction of a method capable of identifying paths between trees of different heights and aspects. The algorithm was evaluated using several images with different characteristics of quality and the results showed that the proposed method can successfully detect a path in different types of environments.
Keywords: Agricultural mobile robot, image processing, path recognition, Hough transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1793632 Oxidation of Selected Pharmaceuticals in Water Matrices by Bromine and Chlorine
Authors: Juan L. Acero, F. Javier Benitez, Francisco J. Real, Gloria Roldan, Francisco Casas
Abstract:
The bromination of five selected pharmaceuticals (metoprolol, naproxen, amoxicillin, hydrochlorotiazide and phenacetin) in ultrapure water and in three water matrices (a groundwater, a surface water from a public reservoir and a secondary effluent from a WWTP) was investigated. The apparent rate constants for the bromination reaction were determined as a function of the pH, and the sequence obtained for the reaction rate was amoxicillin > naproxen >> hydrochlorotiazide ≈ phenacetin ≈ metoprolol. The proposal of a kinetic mechanism, which specifies the dissociation of bromine and each pharmaceutical according to their pKa values and the pH allowed the determination of the intrinsic rate constants for every elementary reaction. The influence of the main operating conditions (pH, initial bromine dose, and the water matrix) on the degradation of pharmaceuticals was established. In addition, the presence of bromide in chlorination experiments was investigated. The presence of bromide in wastewaters and drinking waters in the range of 10 to several hundred μg L-1 accelerated slightly the oxidation of the selected pharmaceuticals during chorine disinfection.Keywords: Pharmaceuticals, bromine, chlorine, apparent andintrinsic rate constants, water matrices, degradation rates
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2334631 Computing Continuous Skyline Queries without Discriminating between Static and Dynamic Attributes
Authors: Ibrahim Gomaa, Hoda M. O. Mokhtar
Abstract:
Although most of the existing skyline queries algorithms focused basically on querying static points through static databases; with the expanding number of sensors, wireless communications and mobile applications, the demand for continuous skyline queries has increased. Unlike traditional skyline queries which only consider static attributes, continuous skyline queries include dynamic attributes, as well as the static ones. However, as skyline queries computation is based on checking the domination of skyline points over all dimensions, considering both the static and dynamic attributes without separation is required. In this paper, we present an efficient algorithm for computing continuous skyline queries without discriminating between static and dynamic attributes. Our algorithm in brief proceeds as follows: First, it excludes the points which will not be in the initial skyline result; this pruning phase reduces the required number of comparisons. Second, the association between the spatial positions of data points is examined; this phase gives an idea of where changes in the result might occur and consequently enables us to efficiently update the skyline result (continuous update) rather than computing the skyline from scratch. Finally, experimental evaluation is provided which demonstrates the accuracy, performance and efficiency of our algorithm over other existing approaches.
Keywords: Continuous query processing, dynamic database, moving object, skyline queries.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1246630 Automatic Detection of Proliferative Cells in Immunohistochemically Images of Meningioma Using Fuzzy C-Means Clustering and HSV Color Space
Authors: Vahid Anari, Mina Bakhshi
Abstract:
Visual search and identification of immunohistochemically stained tissue of meningioma was performed manually in pathologic laboratories to detect and diagnose the cancers type of meningioma. This task is very tedious and time-consuming. Moreover, because of cell's complex nature, it still remains a challenging task to segment cells from its background and analyze them automatically. In this paper, we develop and test a computerized scheme that can automatically identify cells in microscopic images of meningioma and classify them into positive (proliferative) and negative (normal) cells. Dataset including 150 images are used to test the scheme. The scheme uses Fuzzy C-means algorithm as a color clustering method based on perceptually uniform hue, saturation, value (HSV) color space. Since the cells are distinguishable by the human eye, the accuracy and stability of the algorithm are quantitatively compared through application to a wide variety of real images.
Keywords: Positive cell, color segmentation, HSV color space, immunohistochemistry, meningioma, thresholding, fuzzy c-means.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 700629 Modeling Drying and Pyrolysis of Moist Wood Particles at Slow Heating Rates
Authors: Avdhesh K. Sharma
Abstract:
Formulation for drying and pyrolysis process in packed beds at slow heating rates is presented. Drying of biomass particles bed is described by mass diffusion equation and local moisture-vapour-equilibrium relations. In gasifiers, volatilization rate during pyrolysis of biomass is modeled by using apparent kinetic rate expression, while product compositions at slow heating rates is modeled using empirical fitted mass ratios (i.e., CO/CO2, ME/CO2, H2O/CO2) in terms of pyrolysis temperature. The drying module is validated fairly with available chemical kinetics scheme and found that the testing zone in gasifier bed constituted of relatively smaller particles having high airflow with high isothermal temperature expedite the drying process. Further, volatile releases more quickly within the shorter zone height at high temperatures (isothermal). Both, moisture loss and volatile release profiles are found to be sensitive to temperature, although the influence of initial moisture content on volatile release profile is not so sensitive.
Keywords: Modeling downdraft gasifier, drying, pyrolysis, moist woody biomass.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 837628 A Bi-Objective Preventive Healthcare Facility Network Design with Incorporating Cost and Time Saving
Authors: Mehdi Seifbarghy, Keyvan Roshan
Abstract:
Main goal of preventive healthcare problems are at decreasing the likelihood and severity of potentially life-threatening illnesses by protection and early detection. The levels of establishment and staffing costs along with summation of the travel and waiting time that clients spent are considered as objectives functions of the proposed nonlinear integer programming model. In this paper, we have proposed a bi-objective mathematical model for designing a network of preventive healthcare facilities so as to minimize aforementioned objectives, simultaneously. Moreover, each facility acts as M/M/1 queuing system. The number of facilities to be established, the location of each facility, and the level of technology for each facility to be chosen are provided as the main determinants of a healthcare facility network. Finally, to demonstrate performance of the proposed model, four multi-objective decision making techniques are presented to solve the model.Keywords: Preventive healthcare problems, Non-linear integer programming models, Multi-objective decision making techniques
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1773627 Automatic Method for Exudates and Hemorrhages Detection from Fundus Retinal Images
Authors: A. Biran, P. Sobhe Bidari, K. Raahemifar
Abstract:
Diabetic Retinopathy (DR) is an eye disease that leads to blindness. The earliest signs of DR are the appearance of red and yellow lesions on the retina called hemorrhages and exudates. Early diagnosis of DR prevents from blindness; hence, many automated algorithms have been proposed to extract hemorrhages and exudates. In this paper, an automated algorithm is presented to extract hemorrhages and exudates separately from retinal fundus images using different image processing techniques including Circular Hough Transform (CHT), Contrast Limited Adaptive Histogram Equalization (CLAHE), Gabor filter and thresholding. Since Optic Disc is the same color as the exudates, it is first localized and detected. The presented method has been tested on fundus images from Structured Analysis of the Retina (STARE) and Digital Retinal Images for Vessel Extraction (DRIVE) databases by using MATLAB codes. The results show that this method is perfectly capable of detecting hard exudates and the highly probable soft exudates. It is also capable of detecting the hemorrhages and distinguishing them from blood vessels.Keywords: Diabetic retinopathy, fundus, CHT, exudates, hemorrhages.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2650626 Automatic Vehicle Identification by Plate Recognition
Authors: Serkan Ozbay, Ergun Ercelebi
Abstract:
Automatic Vehicle Identification (AVI) has many applications in traffic systems (highway electronic toll collection, red light violation enforcement, border and customs checkpoints, etc.). License Plate Recognition is an effective form of AVI systems. In this study, a smart and simple algorithm is presented for vehicle-s license plate recognition system. The proposed algorithm consists of three major parts: Extraction of plate region, segmentation of characters and recognition of plate characters. For extracting the plate region, edge detection algorithms and smearing algorithms are used. In segmentation part, smearing algorithms, filtering and some morphological algorithms are used. And finally statistical based template matching is used for recognition of plate characters. The performance of the proposed algorithm has been tested on real images. Based on the experimental results, we noted that our algorithm shows superior performance in car license plate recognition.Keywords: Character recognizer, license plate recognition, plate region extraction, segmentation, smearing, template matching.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7592625 Emission of Volatile Organic Compounds from the Residential Combustion of Pyrenean Oak and Black Poplar
Authors: M. Evtyugina, C. A. Alves, A. I. Calvo, T. Nunes, L. Tarelho, M. Duarte, S. O. Prozil
Abstract:
Smoke from domestic wood burning has been identified as a major contributor to air pollution, motivating detailed emission measurements under controlled conditions. A series of experiments was performed to characterise the emissions from wood combustion in a fireplace and in a woodstove of two common species of trees grown in Spain: Pyrenean oak (Quercus pyrenaica) and black poplar (Populus nigra). Volatile organic compounds (VOCs) in the exhaust emissions were collected in Tedlar bags, re-sampled in sorbent tubes and analysed by thermal desorption-gas chromatography-flame ionisation detection. Pyrenean oak presented substantially higher emissions in the woodstove than in the fireplace, for the majority of compounds. The opposite was observed for poplar. Among the 45 identified species, benzene and benzenerelated compounds represent the most abundant group, followed by oxygenated VOCs and aliphatics. Emission factors obtained in this study are generally of the same order than those reported for residential experiments in the USA.Keywords: Fireplace, VOC emissions, woodstove.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1768624 Optimum Time Coordination of Overcurrent Relays using Two Phase Simplex Method
Authors: Prashant P. Bedekar, Sudhir R. Bhide, Vijay S. Kale
Abstract:
Overcurrent (OC) relays are the major protection devices in a distribution system. The operating time of the OC relays are to be coordinated properly to avoid the mal-operation of the backup relays. The OC relay time coordination in ring fed distribution networks is a highly constrained optimization problem which can be stated as a linear programming problem (LPP). The purpose is to find an optimum relay setting to minimize the time of operation of relays and at the same time, to keep the relays properly coordinated to avoid the mal-operation of relays. This paper presents two phase simplex method for optimum time coordination of OC relays. The method is based on the simplex algorithm which is used to find optimum solution of LPP. The method introduces artificial variables to get an initial basic feasible solution (IBFS). Artificial variables are removed using iterative process of first phase which minimizes the auxiliary objective function. The second phase minimizes the original objective function and gives the optimum time coordination of OC relays.Keywords: Constrained optimization, LPP, Overcurrent relaycoordination, Two-phase simplex method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3010623 ADABeV: Automatic Detection of Abnormal Behavior in Video-surveillance
Authors: Nour Charara, Iman Jarkass, Maria Sokhn, Elena Mugellini, Omar Abou Khaled
Abstract:
Intelligent Video-Surveillance (IVS) systems are being more and more popular in security applications. The analysis and recognition of abnormal behaviours in a video sequence has gradually drawn the attention in the field of IVS, since it allows filtering out a large number of useless information, which guarantees the high efficiency in the security protection, and save a lot of human and material resources. We present in this paper ADABeV, an intelligent video-surveillance framework for event recognition in crowded scene to detect the abnormal human behaviour. This framework is attended to be able to achieve real-time alarming, reducing the lags in traditional monitoring systems. This architecture proposal addresses four main challenges: behaviour understanding in crowded scenes, hard lighting conditions, multiple input kinds of sensors and contextual-based adaptability to recognize the active context of the scene.Keywords: Behavior recognition, Crowded scene, Data fusion, Pattern recognition, Video-surveillance
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3640622 New Regression Model and I-Kaz Method for Online Cutting Tool Wear Monitoring
Authors: Jaharah A. Ghani, Muhammad Rizal, Ahmad Sayuti, Mohd Zaki Nuawi, Mohd Nizam Ab. Rahman, Che Hassan Che Haron
Abstract:
This study presents a new method for detecting the cutting tool wear based on the measured cutting force signals using the regression model and I-kaz method. The detection of tool wear was done automatically using the in-house developed regression model and 3D graphic presentation of I-kaz 3D coefficient during machining process. The machining tests were carried out on a CNC turning machine Colchester Master Tornado T4 in dry cutting condition, and Kistler 9255B dynamometer was used to measure the cutting force signals, which then stored and displayed in the DasyLab software. The progression of the cutting tool flank wear land (VB) was indicated by the amount of the cutting force generated. Later, the I-kaz was used to analyze all the cutting force signals from beginning of the cut until the rejection stage of the cutting tool. Results of the IKaz analysis were represented by various characteristic of I-kaz 3D coefficient and 3D graphic presentation. The I-kaz 3D coefficient number decreases when the tool wear increases. This method can be used for real time tool wear monitoring.Keywords: mathematical model, I-kaz method, tool wear
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2403621 Enhancing Protein Incorporation in Calcium Phosphate Coating on Titanium by Rapid Biomimetic Co-Precipitation Technique
Authors: J. Suwanprateeb, F. Thammarakcharoen
Abstract:
Calcium phosphate coating (CaP) has been employed for protein delivery, but the typical direct protein adsorption on the coating led to low incorporation content and fast release of the protein from the coating. By using bovine serum albumin (BSA) as a model protein, rapid biomimetic co-precipitation between calcium phosphate and BSA was employed to control the distribution of BSA within calcium phosphate coating during biomimetic formation on titanium surface for only 6 h at 50oC in an accelerated calcium phosphate solution. As a result, the amount of BSA incorporation and release duration could be increased by using a rapid biomimetic coprecipitation technique. Up to 43 fold increases in the BSA incorporation content and the increase from 6 h to more than 360 h in release duration compared to typical direct adsorption technique were observed depending on the initial BSA concentration used during coprecipitation (1, 10 and 100 μg.ml-1). From x-ray diffraction and Fourier transform infrared spectroscopy studies, the coating composition was not altered with the incorporation of BSA by this rapid biomimetic co-precipitation and mainly comprised octacalcium phosphate and hydroxyapatite. However, the microstructure of calcium phosphate crystals changed from straight, plate-like units to curved, plate-like units with increasing BSA content.
Keywords: Biomimetic, Calcium Phosphate Coating, Protein, Titanium.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2379620 Discussing Embedded versus Central Machine Learning in Wireless Sensor Networks
Authors: Anne-Lena Kampen, Øivind Kure
Abstract:
Machine learning (ML) can be implemented in Wireless Sensor Networks (WSNs) as a central solution or distributed solution where the ML is embedded in the nodes. Embedding improves privacy and may reduce prediction delay. In addition, the number of transmissions is reduced. However, quality factors such as prediction accuracy, fault detection efficiency and coordinated control of the overall system suffer. Here, we discuss and highlight the trade-offs that should be considered when choosing between embedding and centralized ML, especially for multihop networks. In addition, we present estimations that demonstrate the energy trade-offs between embedded and centralized ML. Although the total network energy consumption is lower with central prediction, it makes the network more prone for partitioning due to the high forwarding load on the one-hop nodes. Moreover, the continuous improvements in the number of operations per joule for embedded devices will move the energy balance toward embedded prediction.
Keywords: Central ML, embedded machine learning, energy consumption, local ML, Wireless Sensor Networks, WSN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 839