Search results for: Two combined denitrification/oxic filter
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1454

Search results for: Two combined denitrification/oxic filter

494 Intelligent Heart Disease Prediction System Using CANFIS and Genetic Algorithm

Authors: Latha Parthiban, R. Subramanian

Abstract:

Heart disease (HD) is a major cause of morbidity and mortality in the modern society. Medical diagnosis is an important but complicated task that should be performed accurately and efficiently and its automation would be very useful. All doctors are unfortunately not equally skilled in every sub specialty and they are in many places a scarce resource. A system for automated medical diagnosis would enhance medical care and reduce costs. In this paper, a new approach based on coactive neuro-fuzzy inference system (CANFIS) was presented for prediction of heart disease. The proposed CANFIS model combined the neural network adaptive capabilities and the fuzzy logic qualitative approach which is then integrated with genetic algorithm to diagnose the presence of the disease. The performances of the CANFIS model were evaluated in terms of training performances and classification accuracies and the results showed that the proposed CANFIS model has great potential in predicting the heart disease.

Keywords: CANFIS, genetic algorithms, heart disease, membership function.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3993
493 Unsupervised Segmentation Technique for Acute Leukemia Cells Using Clustering Algorithms

Authors: N. H. Harun, A. S. Abdul Nasir, M. Y. Mashor, R. Hassan

Abstract:

Leukaemia is a blood cancer disease that contributes to the increment of mortality rate in Malaysia each year. There are two main categories for leukaemia, which are acute and chronic leukaemia. The production and development of acute leukaemia cells occurs rapidly and uncontrollable. Therefore, if the identification of acute leukaemia cells could be done fast and effectively, proper treatment and medicine could be delivered. Due to the requirement of prompt and accurate diagnosis of leukaemia, the current study has proposed unsupervised pixel segmentation based on clustering algorithm in order to obtain a fully segmented abnormal white blood cell (blast) in acute leukaemia image. In order to obtain the segmented blast, the current study proposed three clustering algorithms which are k-means, fuzzy c-means and moving k-means algorithms have been applied on the saturation component image. Then, median filter and seeded region growing area extraction algorithms have been applied, to smooth the region of segmented blast and to remove the large unwanted regions from the image, respectively. Comparisons among the three clustering algorithms are made in order to measure the performance of each clustering algorithm on segmenting the blast area. Based on the good sensitivity value that has been obtained, the results indicate that moving kmeans clustering algorithm has successfully produced the fully segmented blast region in acute leukaemia image. Hence, indicating that the resultant images could be helpful to haematologists for further analysis of acute leukaemia.

Keywords: Acute Leukaemia Images, Clustering Algorithms, Image Segmentation, Moving k-Means.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2789
492 Extracts of Cola acuminata, Lupinus arboreus and Bougainvillea spectabilis as Natural Photosensitizers for Dye-Sensitized Solar Cells

Authors: M. L. Akinyemi, T. J. Abodurin, A. O. Boyo, J. A. O. Olugbuyiro

Abstract:

Organic dyes from Cola acuminata (C. acuminata), Lupinus arboreus (L. arboreus) and Bougainvillea spectabilis (B. spectabilis) leaves and their mixtures were used as sensitizers to manufacture dye-sensitized solar cells (DSSC). Photoelectric measurements of C. acuminata showed a short circuit current (Jsc) of 0.027 mA/ cm2, 0.026 mA/ cm2 and 0.018 mA/ cm2 with a mixture of mercury chloride and iodine (Hgcl2 + I); potassium bromide and iodine (KBr + I); and potassium chloride and iodine (KCl + I) respectively. The open circuit voltage (Voc) was 24 mV, 25 mV and 20 mV for the three dyes respectively. L. arboreus had Jsc of 0.034 mA/ cm2, 0.021 mA/ cm2 and 0.013 mA/ cm2; and corresponding Voc of 28 mV, 14.2 mV and 15 mV for the three electrolytes respectively. B. spectabilis recorded Jsc 0.023 mA/ cm2, 0.026 mA/ cm2 and 0.015 mA/ cm2; and corresponding Voc values of 6.2 mV, 14.3 mV and 4.0 mV for the three electrolytes respectively. It was observed that the fill factor (FF) was 0.140 for C. acuminata, 0.3198 for L. arboreus and 0.1138 for B. spectabilis. Internal conversions of 0.096%, 0.056% and 0.063% were recorded for three dyes when combined with (KBr + I) electrolyte. The internal efficiency of C. acuminata DSSC was highest in value.

Keywords: Dye-sensitized Solar Cells, Organic dye, C. acuminate, L. arboreus, B. spectabilis, Dye Mixture.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1438
491 Effect of UV Radiation to Change the Properties of the Composite PA+GF

Authors: Lenka Markovičová, Viera Zatkalíková, Tomasz Garbacz

Abstract:

The development of composite materials and the related design and manufacturing technologies is one of the most important advances in the history of materials. Composites are multifunctional materials having unprecedented mechanical and physical properties that can be tailored to meet the requirements of a particular application. Some composites also exhibit great resistance to high-temperature corrosion, oxidation, and wear. Polymers are widely used indoors and outdoors, therefore they are exposed to a chemical environment which may include atmospheric oxygen, acidic fumes, acidic rain, moisture heat and thermal shock, ultra-violet light, high energy radiation, etc. Different polymers are affected differently by these factors even though the amorphous polymers are more sensitive. Ageing is also important and it is defined as the process of deterioration of engineering materials resulting from the combined effects of atmospheric radiation, heat, oxygen, water, microorganisms and other atmospheric factors.

Keywords: Composites with glass fibres, mechanical properties, polyamides, UV degradation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2181
490 Centralized Cooperative Spectrum Sensing with MIMO in the Reporting Network over κ − μ Fading Channel

Authors: S Hariharan, K Chaitanya, P Muthuchidambaranathan

Abstract:

The IEEE 802.22 working group aims to drive the Digital Video Broadcasting-Terrestrial (DVB-T) bands for data communication to the rural area without interfering the TV broadcast. In this paper, we arrive at a closed-form expression for average detection probability of Fusion center (FC) with multiple antenna over the κ − μ fading channel model. We consider a centralized cooperative multiple antenna network for reporting. The DVB-T samples forwarded by the secondary user (SU) were combined using Maximum ratio combiner at FC, an energy detection is performed to make the decision. The fading effects of the channel degrades the detection probability of the FC, a generalized independent and identically distributed (IID) κ − μ and an additive white Gaussian noise (AWGN) channel is considered for reporting and sensing respectively. The proposed system performance is verified through simulation results.

Keywords: IEEE 802.22, Cooperative spectrum sensing, Multiple antenna, κ − μ .

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5456
489 Extracting Terrain Points from Airborne Laser Scanning Data in Densely Forested Areas

Authors: Ziad Abdeldayem, Jakub Markiewicz, Kunal Kansara, Laura Edwards

Abstract:

Airborne Laser Scanning (ALS) is one of the main technologies for generating high-resolution digital terrain models (DTMs). DTMs are crucial to several applications, such as topographic mapping, flood zone delineation, geographic information systems (GIS), hydrological modelling, spatial analysis, etc. Laser scanning system generates irregularly spaced three-dimensional cloud of points. Raw ALS data are mainly ground points (that represent the bare earth) and non-ground points (that represent buildings, trees, cars, etc.). Removing all the non-ground points from the raw data is referred to as filtering. Filtering heavily forested areas is considered a difficult and challenging task as the canopy stops laser pulses from reaching the terrain surface. This research presents an approach for removing non-ground points from raw ALS data in densely forested areas. Smoothing splines are exploited to interpolate and fit the noisy ALS data. The presented filter utilizes a weight function to allocate weights for each point of the data. Furthermore, unlike most of the methods, the presented filtering algorithm is designed to be automatic. Three different forested areas in the United Kingdom are used to assess the performance of the algorithm. The results show that the generated DTMs from the filtered data are accurate (when compared against reference terrain data) and the performance of the method is stable for all the heavily forested data samples. The average root mean square error (RMSE) value is 0.35 m.

Keywords: Airborne laser scanning, digital terrain models, filtering, forested areas.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 718
488 Face Reconstruction and Camera Pose Using Multi-dimensional Descent

Authors: Varin Chouvatut, Suthep Madarasmi, Mihran Tuceryan

Abstract:

This paper aims to propose a novel, robust, and simple method for obtaining a human 3D face model and camera pose (position and orientation) from a video sequence. Given a video sequence of a face recorded from an off-the-shelf digital camera, feature points used to define facial parts are tracked using the Active- Appearance Model (AAM). Then, the face-s 3D structure and camera pose of each video frame can be simultaneously calculated from the obtained point correspondences. This proposed method is primarily based on the combined approaches of Gradient Descent and Powell-s Multidimensional Minimization. Using this proposed method, temporarily occluded point including the case of self-occlusion does not pose a problem. As long as the point correspondences displayed in the video sequence have enough parallax, these missing points can still be reconstructed.

Keywords: Camera Pose, Face Reconstruction, Gradient Descent, Powell's Multidimensional Minimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1584
487 An Approach to Concerns and Aspects Mining for Web Applications

Authors: Carlo Bellettini, Alessandro Marchetto, Andrea Trentini

Abstract:

Web applications have become very complex and crucial, especially when combined with areas such as CRM (Customer Relationship Management) and BPR (Business Process Reengineering), the scientific community has focused attention to Web applications design, development, analysis, and testing, by studying and proposing methodologies and tools. This paper proposes an approach to automatic multi-dimensional concern mining for Web Applications, based on concepts analysis, impact analysis, and token-based concern identification. This approach lets the user to analyse and traverse Web software relevant to a particular concern (concept, goal, purpose, etc.) via multi-dimensional separation of concerns, to document, understand and test Web applications. This technique was developed in the context of WAAT (Web Applications Analysis and Testing) project. A semi-automatic tool to support this technique is currently under development.

Keywords: Aspect Mining, Concepts Analysis, Concerns Mining, Multi-Dimensional Separation of Concerns, Impact Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1513
486 Effect of Different Methods of Soil Fertility on Grain Yield and Chickpea Quality

Authors: Mohammadi K., Ghalavand A., Aghaalikhani M

Abstract:

In order to evaluation the effects of natural, biological and chemical fertilizers on grain yield and chickpea quality, field experiments were carried out in 2007 and 2008 growing seasons. In this research the effects of different organic, chemical and biological fertilizers were investigated on grain yield and quality of chickpea. Experimental units were arranged in split-split plots based on randomized complete blocks with three replications. The highest amounts of yield and yield components were obtained in G1×N5 interaction. Significant increasing of N, P, K, Fe and Mg content in leaves and grains emphasized on superiority of mentioned treatment because each one of these nutrients has an approved role in chlorophyll synthesis and photosynthesis ability of the crop. The combined application of compost, farmyard manure and chemical phosphorus (N5) had the best grain quality due to high protein, starch and total sugar contents, low crude fiber and reduced cooking time.

Keywords: soil fertility, grain yield, chickpea, natural resources.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2598
485 Large-Deflection Analysis of Automotive Vehicle's Door Wiring Harness System Using Finite Element Method

Authors: Byeong-Sam Kim, Kangsu Lee, Kyoungwoo Park, Samir Ben Chaabane

Abstract:

A Vehicle-s door wireing harness arrangement structure is provided. In vehicle-s door wiring harness(W/H) system is more toward to arrange a passenger compartment than a hinge and a weatherstrip. This article gives some insight into the dimensioning process, with special focus on large deflection analysis of wiring harness(W/H) in vehicle-s door structures for durability problem. An Finite elements analysis for door wiring harness(W/H) are used for residual stresses and dimensional stability with bending flexible. Durability test data for slim test specimens were compared with the numerical predicted fatigue life for verification. The final lifing of the component combines the effects of these microstructural features with the complex stress state arising from the combined service loading and residual stresses.

Keywords: Large deflection, wiring harness system, finite element analysis, vehicle's door.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3315
484 Numerical Analysis of Electrical Interaction between two Axisymmetric Spheroids

Authors: Kuan-Liang Liu, Eric Lee, Jung-Jyh Lee, Jyh-Ping Hsu

Abstract:

The electrical interaction between two axisymmetric spheroidal particles in an electrolyte solution is examined numerically. A Galerkin finite element method combined with a Newton-Raphson iteration scheme is proposed to evaluate the spatial variation in the electrical potential, and the result obtained used to estimate the interaction energy between two particles. We show that if the surface charge density is fixed, the potential gradient is larger at a point, which has a larger curvature, and if surface potential is fixed, surface charge density is proportional to the curvature. Also, if the total interaction energy against closest surface-to-surface curve exhibits a primary maximum, the maximum follows the order (oblate-oblate) > (sphere-sphere)>(oblate-prolate)>(prolate-prolate), and if the curve has a secondary minimum, the absolute value of the minimum follows the same order.

Keywords: interaction energy, interaction force, Poisson-Boltzmann equation, spheroid.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1470
483 Improvement of Load Carrying Capacity of an RCC T-Beam Bridge Longitudinal Girder by Replacing Steel Bars with SMA Bars

Authors: N. K. Paul, S. Saha

Abstract:

An innovative three dimensional finite element model has beed developed and tested under two point loading system to examine the structural behavior of the longitudinal reinforced concrete Tee-beam bridge girder, reinforcing with steel and shape memory alloy bars respectively. 25% of steel bars are replaced with superelastic Shape Memory Alloy bars in this study. Finite element analysis is performed using ANSYS 11.0 program. Experimentally a model of steel reinforced girder has been casted and its load deflection responses are checked with ANSYS analysis. A comparison of load carrying capacity for the model between steel RC girder and the girder combined reinforcement with SMA and steel are also performed.

Keywords: Shape memory alloy, bridge girder, ANSYS, load-deflection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 888
482 Multi-Dimensional Concerns Mining for Web Applications via Concept-Analysis

Authors: Carlo Bellettini, Alessandro Marchetto, Andrea Trentini

Abstract:

Web applications have become very complex and crucial, especially when combined with areas such as CRM (Customer Relationship Management) and BPR (Business Process Reengineering), the scientific community has focused attention to Web applications design, development, analysis, and testing, by studying and proposing methodologies and tools. This paper proposes an approach to automatic multi-dimensional concern mining for Web Applications, based on concepts analysis, impact analysis, and token-based concern identification. This approach lets the user to analyse and traverse Web software relevant to a particular concern (concept, goal, purpose, etc.) via multi-dimensional separation of concerns, to document, understand and test Web applications. This technique was developed in the context of WAAT (Web Applications Analysis and Testing) project. A semi-automatic tool to support this technique is currently under development.

Keywords: Concepts Analysis, Concerns Mining, Multi-Dimensional Separation of Concerns, Impact Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1473
481 U.S. Nuclear Regulatory CommissionTraining for Research and Training Reactor Inspectors

Authors: Gary Marlin Sandquist

Abstract:

Currently, a large number of license activities (Early Site Permits, Combined Operating License, reactor certifications, etc.), are pending for review before the United States Nuclear Regulatory Commission (US NRC). Much of the senior staff at the NRC is now committed to these review and licensing actions. To address this additional workload, the NRC has recruited a large number of new Regulatory Staff for dealing with these and other regulatory actions such as the US Fleet of Research and Test Reactors (RTRs). These reactors pose unusual demands on Regulatory Staff since the US Fleet of RTRs, although few (32 Licensed RTRs as of 2010), they represent a broad range of reactor types, operations, and research and training aspects that nuclear reactor power plants (such as the 104 LWRs) do not pose. The NRC must inspect and regulate all these facilities. This paper addresses selected training topics and regulatory activities providedNRC Inspectors for RTRs.

Keywords: Regulations, Research and Test Reactors, Training, US NRC

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1644
480 A Perceptually Optimized Foveation Based Wavelet Embedded Zero Tree Image Coding

Authors: A. Bajit, M. Nahid, A. Tamtaoui, E. H. Bouyakhf

Abstract:

In this paper, we propose a Perceptually Optimized Foveation based Embedded ZeroTree Image Coder (POEFIC) that introduces a perceptual weighting to wavelet coefficients prior to control SPIHT encoding algorithm in order to reach a targeted bit rate with a perceptual quality improvement with respect to a given bit rate a fixation point which determines the region of interest ROI. The paper also, introduces a new objective quality metric based on a Psychovisual model that integrates the properties of the HVS that plays an important role in our POEFIC quality assessment. Our POEFIC coder is based on a vision model that incorporates various masking effects of human visual system HVS perception. Thus, our coder weights the wavelet coefficients based on that model and attempts to increase the perceptual quality for a given bit rate and observation distance. The perceptual weights for all wavelet subbands are computed based on 1) foveation masking to remove or reduce considerable high frequencies from peripheral regions 2) luminance and Contrast masking, 3) the contrast sensitivity function CSF to achieve the perceptual decomposition weighting. The new perceptually optimized codec has the same complexity as the original SPIHT techniques. However, the experiments results show that our coder demonstrates very good performance in terms of quality measurement.

Keywords: DWT, linear-phase 9/7 filter, Foveation Filtering, CSF implementation approaches, 9/7 Wavelet JND Thresholds and Wavelet Error Sensitivity WES, Luminance and Contrast masking, standard SPIHT, Objective Quality Measure, Probability Score PS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1795
479 An Approach to Image Extraction and Accurate Skin Detection from Web Pages

Authors: Moheb R. Girgis, Tarek M. Mahmoud, Tarek Abd-El-Hafeez

Abstract:

This paper proposes a system to extract images from web pages and then detect the skin color regions of these images. As part of the proposed system, using BandObject control, we built a Tool bar named 'Filter Tool Bar (FTB)' by modifying the Pavel Zolnikov implementation. The Yahoo! Team provides us with the Yahoo! SDK API, which also supports image search and is really useful. In the proposed system, we introduced three new methods for extracting images from the web pages (after loading the web page by using the proposed FTB, before loading the web page physically from the localhost, and before loading the web page from any server). These methods overcome the drawback of the regular expressions method for extracting images suggested by Ilan Assayag. The second part of the proposed system is concerned with the detection of the skin color regions of the extracted images. So, we studied two famous skin color detection techniques. The first technique is based on the RGB color space and the second technique is based on YUV and YIQ color spaces. We modified the second technique to overcome the failure of detecting complex image's background by using the saturation parameter to obtain an accurate skin detection results. The performance evaluation of the efficiency of the proposed system in extracting images before and after loading the web page from localhost or any server in terms of the number of extracted images is presented. Finally, the results of comparing the two skin detection techniques in terms of the number of pixels detected are presented.

Keywords: Browser Helper Object, Color spaces, Image and URL extraction, Skin detection, Web Browser events.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1897
478 Automatic Inspection of Percussion Caps by Means of Combined 2D and 3D Machine Vision Techniques

Authors: A. Tellaeche, R. Arana, I.Maurtua

Abstract:

The exhaustive quality control is becoming more and more important when commercializing competitive products in the world's globalized market. Taken this affirmation as an undeniable truth, it becomes critical in certain sector markets that need to offer the highest restrictions in quality terms. One of these examples is the percussion cap mass production, a critical element assembled in firearm ammunition. These elements, built in great quantities at a very high speed, must achieve a minimum tolerance deviation in their fabrication, due to their vital importance in firing the piece of ammunition where they are built in. This paper outlines a machine vision development for the 100% inspection of percussion caps obtaining data from 2D and 3D simultaneous images. The acquisition speed and precision of these images from a metallic reflective piece as a percussion cap, the accuracy of the measures taken from these images and the multiple fabrication errors detected make the main findings of this work.

Keywords: critical tolerance, high speed decision makingsimultaneous 2D/3D machine vision.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1537
477 Automatic Staging and Subtype Determination for Non-Small Cell Lung Carcinoma Using PET Image Texture Analysis

Authors: Seyhan Karaçavuş, Bülent Yılmaz, Ömer Kayaaltı, Semra İçer, Arzu Taşdemir, Oğuzhan Ayyıldız, Kübra Eset, Eser Kaya

Abstract:

In this study, our goal was to perform tumor staging and subtype determination automatically using different texture analysis approaches for a very common cancer type, i.e., non-small cell lung carcinoma (NSCLC). Especially, we introduced a texture analysis approach, called Law’s texture filter, to be used in this context for the first time. The 18F-FDG PET images of 42 patients with NSCLC were evaluated. The number of patients for each tumor stage, i.e., I-II, III or IV, was 14. The patients had ~45% adenocarcinoma (ADC) and ~55% squamous cell carcinoma (SqCCs). MATLAB technical computing language was employed in the extraction of 51 features by using first order statistics (FOS), gray-level co-occurrence matrix (GLCM), gray-level run-length matrix (GLRLM), and Laws’ texture filters. The feature selection method employed was the sequential forward selection (SFS). Selected textural features were used in the automatic classification by k-nearest neighbors (k-NN) and support vector machines (SVM). In the automatic classification of tumor stage, the accuracy was approximately 59.5% with k-NN classifier (k=3) and 69% with SVM (with one versus one paradigm), using 5 features. In the automatic classification of tumor subtype, the accuracy was around 92.7% with SVM one vs. one. Texture analysis of FDG-PET images might be used, in addition to metabolic parameters as an objective tool to assess tumor histopathological characteristics and in automatic classification of tumor stage and subtype.

Keywords: Cancer stage, cancer cell type, non-small cell lung carcinoma, PET, texture analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 977
476 GA Based Optimal Feature Extraction Method for Functional Data Classification

Authors: Jun Wan, Zehua Chen, Yingwu Chen, Zhidong Bai

Abstract:

Classification is an interesting problem in functional data analysis (FDA), because many science and application problems end up with classification problems, such as recognition, prediction, control, decision making, management, etc. As the high dimension and high correlation in functional data (FD), it is a key problem to extract features from FD whereas keeping its global characters, which relates to the classification efficiency and precision to heavens. In this paper, a novel automatic method which combined Genetic Algorithm (GA) and classification algorithm to extract classification features is proposed. In this method, the optimal features and classification model are approached via evolutional study step by step. It is proved by theory analysis and experiment test that this method has advantages in improving classification efficiency, precision and robustness whereas using less features and the dimension of extracted classification features can be controlled.

Keywords: Classification, functional data, feature extraction, genetic algorithm, wavelet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1555
475 Dynamic Model of Automatic Loom on SimulationX

Authors: A. Jomartov, A. Tuleshov, B. Tultaev

Abstract:

One of the main tasks in the development of textile machinery is to increase the rapidity of automatic looms, and consequently, their productivity. With increasing automatic loom speeds, the dynamic loads on their separate mechanisms and moving joints sharply increase. Dynamic research allows us to determine the weakest mechanisms of the automatic loom. The modern automatic loom consists of a large number of structurally different mechanisms. These are cam, lever, gear, friction and combined cyclic mechanisms. The modern automatic loom contains various mechatronic devices: A device for the automatic removal of faulty weft, electromechanical drive warp yarns, electronic controllers, servos, etc. In the paper, we consider the multibody dynamic model of the automatic loom on the software complex SimulationX. SimulationX is multidisciplinary software for modeling complex physical and technical facilities and systems. The multibody dynamic model of the automatic loom allows consideration of: The transition processes, backlash at the joints and nodes, the force of resistance and electric motor performance.

Keywords: Automatic loom, dynamics, model, multibody, SimulationX.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1494
474 Evaluation of Physicochemical Pretreatment Methods on COD and Ammonia Removal from Landfill Leachate

Authors: M. Poveda, S. Lozecznik, J. Oleszkiewicz, Q. Yuan

Abstract:

The goal of this experiment is to evaluate the effectiveness of different leachate pre-treatment options in terms of COD and ammonia removal. This research focused on the evaluation of physical-chemical methods for pre-treatment of leachate that would be effective and rapid in order to satisfy the requirements of the sewer discharge by-laws. The four pre-treatment options evaluated were: air stripping, chemical coagulation, electrocoagulation and advanced oxidation with sodium ferrate. Chemical coagulation reported the best COD removal rate at 43%, compared to 18% for both air stripping and electro-coagulation, and 20% for oxidation with sodium ferrate. On the other hand, air stripping was far superior to the other treatment options in terms of ammonia removal with 86%. Oxidation with sodium ferrate reached only 16%, while chemical coagulation and electro-coagulation removed less than 10%. When combined, air stripping and chemical coagulation removed up to 50% COD and 85% ammonia.

Keywords: Leachate pretreatment, air stripping, chemical coagulation, electro-coagulation, oxidation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1924
473 Performance Analysis of Chrominance Red and Chrominance Blue in JPEG

Authors: Mamta Garg

Abstract:

While compressing text files is useful, compressing still image files is almost a necessity. A typical image takes up much more storage than a typical text message and without compression images would be extremely clumsy to store and distribute. The amount of information required to store pictures on modern computers is quite large in relation to the amount of bandwidth commonly available to transmit them over the Internet and applications. Image compression addresses the problem of reducing the amount of data required to represent a digital image. Performance of any image compression method can be evaluated by measuring the root-mean-square-error & peak signal to noise ratio. The method of image compression that will be analyzed in this paper is based on the lossy JPEG image compression technique, the most popular compression technique for color images. JPEG compression is able to greatly reduce file size with minimal image degradation by throwing away the least “important" information. In JPEG, both color components are downsampled simultaneously, but in this paper we will compare the results when the compression is done by downsampling the single chroma part. In this paper we will demonstrate more compression ratio is achieved when the chrominance blue is downsampled as compared to downsampling the chrominance red in JPEG compression. But the peak signal to noise ratio is more when the chrominance red is downsampled as compared to downsampling the chrominance blue in JPEG compression. In particular we will use the hats.jpg as a demonstration of JPEG compression using low pass filter and demonstrate that the image is compressed with barely any visual differences with both methods.

Keywords: JPEG, Discrete Cosine Transform, Quantization, Color Space Conversion, Image Compression, Peak Signal to Noise Ratio & Compression Ratio.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1677
472 Understanding the Qualities of Social Spaces of the Indian Neighbourhoods

Authors: Venkata Ravi Kumar Veluru, Tej Kumar Karki

Abstract:

Indian traditional neighbourhoods are socially active and sometimes intrusive communities, which are losing their qualities due to Western influences, undermining the traditional Indian values by blind adaption of neighbourhood concepts since the western scale is not suitable to the Indian context. The main aim is to understand the qualities of Indian traditional neighbourhoods by evaluating a traditional neighbourhood of Jaipur, and comparing it with the modern planned neighbourhood of Chandigarh, (designed by a foreign planner, in the neighbourhood concept of the Western world), to find out the special qualities of traditional Indian neighbourhoods as compared to Western concepts based on social spaces, using methods such as physical observation of selected neighbourhoods and residents structured perception survey. A combined analysis found that social spaces are abundantly available in traditional neighbourhoods which are missing in modern neighbourhoods. The quality of traditional neighbourhoods is interactions that aim toward the formation of social capital. The qualities of traditional neighbourhoods have to be considered while designing new neighbourhoods in India.

Keywords: Indian Neighbourhoods, modern neighbourhoods, neighbourhood planning, social spaces, traditional neighbourhoods.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 164
471 Scatterer Density in Edge and Coherence Enhancing Nonlinear Anisotropic Diffusion for Medical Ultrasound Speckle Reduction

Authors: Ahmed Badawi, J. Michael Johnson, Mohamed Mahfouz

Abstract:

This paper proposes new enhancement models to the methods of nonlinear anisotropic diffusion to greatly reduce speckle and preserve image features in medical ultrasound images. By incorporating local physical characteristics of the image, in this case scatterer density, in addition to the gradient, into existing tensorbased image diffusion methods, we were able to greatly improve the performance of the existing filtering methods, namely edge enhancing (EE) and coherence enhancing (CE) diffusion. The new enhancement methods were tested using various ultrasound images, including phantom and some clinical images, to determine the amount of speckle reduction, edge, and coherence enhancements. Scatterer density weighted nonlinear anisotropic diffusion (SDWNAD) for ultrasound images consistently outperformed its traditional tensor-based counterparts that use gradient only to weight the diffusivity function. SDWNAD is shown to greatly reduce speckle noise while preserving image features as edges, orientation coherence, and scatterer density. SDWNAD superior performances over nonlinear coherent diffusion (NCD), speckle reducing anisotropic diffusion (SRAD), adaptive weighted median filter (AWMF), wavelet shrinkage (WS), and wavelet shrinkage with contrast enhancement (WSCE), make these methods ideal preprocessing steps for automatic segmentation in ultrasound imaging.

Keywords: Nonlinear anisotropic diffusion, ultrasound imaging, speckle reduction, scatterer density estimation, edge based enhancement, coherence enhancement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1906
470 An Efficient Cache Replacement Strategy for the Hybrid Cache Consistency Approach

Authors: Aline Zeitunlian, Ramzi A. Haraty

Abstract:

Caching was suggested as a solution for reducing bandwidth utilization and minimizing query latency in mobile environments. Over the years, different caching approaches have been proposed, some relying on the server to broadcast reports periodically informing of the updated data while others allowed the clients to request for the data whenever needed. Until recently a hybrid cache consistency scheme Scalable Asynchronous Cache Consistency Scheme SACCS was proposed, which combined the two different approaches benefits- and is proved to be more efficient and scalable. Nevertheless, caching has its limitations too, due to the limited cache size and the limited bandwidth, which makes the implementation of cache replacement strategy an important aspect for improving the cache consistency algorithms. In this thesis, we proposed a new cache replacement strategy, the Least Unified Value strategy (LUV) to replace the Least Recently Used (LRU) that SACCS was based on. This paper studies the advantages and the drawbacks of the new proposed strategy, comparing it with different categories of cache replacement strategies.

Keywords: Cache consistency, hybrid algorithm, and mobileenvironments

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2202
469 Combined Beamforming and Channel Estimation in WCDMA Communication Systems

Authors: Nermin A. Mohamed, Mohamed F. Madkour

Abstract:

We address the problem of joint beamforming and multipath channel parameters estimation in Wideband Code Division Multiple Access (WCDMA) communication systems that employ Multiple-Access Interference (MAI) suppression techniques in the uplink (from mobile to base station). Most of the existing schemes rely on time multiplex a training sequence with the user data. In WCDMA, the channel parameters can also be estimated from a code multiplexed common pilot channel (CPICH) that could be corrupted by strong interference resulting in a bad estimate. In this paper, we present new methods to combine interference suppression together with channel estimation when using multiple receiving antennas by using adaptive signal processing techniques. Computer simulation is used to compare between the proposed methods and the existing conventional estimation techniques.

Keywords: Adaptive arrays, channel estimation, interferencecancellation, wideband code division multiple access (WCDMA).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2315
468 Use Cuticular Hydrocarbons as Chemotaxonomic of The Pamphagidae Pamphagus elephas (Insecta, Orthoptera) of Algeria

Authors: M. Bounechada, F. Benia, M. Aiouaz, S. Bouharati, N. Djirar, H. Benamrani

Abstract:

The cuticular hydrocarbons of Pamphagus elephas (Orthoptera: Pamphagidae) has been analysed by gas chromatography and by combined gas chromatograph-mass spectrometry. The following hydrocarbon classes have been identified in insect cuticular hydrocarbons are: n-alkanes and methylalkanes comprising Monomethyl-, dimethyl-and trimethylalkanes. Sexual dimorphism is observed in long chain alkanes (C24-C36) present on male and female. The cuticulars hydrocarbons of P.elephas ranged from 24 to 36 carbons and incluted n-alkanes, Dimethylalkanes and Trimethylalkanes. nalkanes represented by (C24-C36,72,7% on male and 79,2% on female), internally branched Monomethylalkanes identified were (C25, C30-C32,C35-C37;11% on male and 9,4% on female), Dimethylalkanes detected are (C31-C32, C36; 2,2% on male and 2,06% on female) and Trimethylalkanes detected are (C32, C36; 3,1% on male and 4, 97 on female). Larvae male and female (stage 7) showed the same quality of n-alkanes observed in adults. However a difference quantity is noted.

Keywords: Cuticular hydrocarbons, Gas chromatography, Mass spectrometry, Pamphagus elephas, , Sexual dimorphism

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1724
467 Comparison between Associative Classification and Decision Tree for HCV Treatment Response Prediction

Authors: Enas M. F. El Houby, Marwa S. Hassan

Abstract:

Combined therapy using Interferon and Ribavirin is the standard treatment in patients with chronic hepatitis C. However, the number of responders to this treatment is low, whereas its cost and side effects are high. Therefore, there is a clear need to predict patient’s response to the treatment based on clinical information to protect the patients from the bad drawbacks, Intolerable side effects and waste of money. Different machine learning techniques have been developed to fulfill this purpose. From these techniques are Associative Classification (AC) and Decision Tree (DT). The aim of this research is to compare the performance of these two techniques in the prediction of virological response to the standard treatment of HCV from clinical information. 200 patients treated with Interferon and Ribavirin; were analyzed using AC and DT. 150 cases had been used to train the classifiers and 50 cases had been used to test the classifiers. The experiment results showed that the two techniques had given acceptable results however the best accuracy for the AC reached 92% whereas for DT reached 80%.

Keywords: Associative Classification, Data mining, Decision tree, HCV, interferon.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1899
466 A Combined Fuzzy Decision Making Approach to Supply Chain Risk Assessment

Authors: P. Moeinzadeh, A. Hajfathaliha

Abstract:

Many firms implemented various initiatives such as outsourced manufacturing which could make a supply chain (SC) more vulnerable to various types of disruptions. So managing risk has become a critical component of SC management. Different types of SC vulnerability management methodologies have been proposed for managing SC risk, most offer only point-based solutions that deal with a limited set of risks. This research aims to reinforce SC risk management by proposing an integrated approach. SC risks are identified and a risk index classification structure is created. Then we develop a SC risk assessment approach based on the analytic network process (ANP) and the VIKOR methods under the fuzzy environment where the vagueness and subjectivity are handled with linguistic terms parameterized by triangular fuzzy numbers. By using FANP, risks weights are calculated and then inserted to the FVIKOR to rank the SC members and find the most risky partner.

Keywords: Analytic network process (ANP), Fuzzy sets, Supply chain risk management (SCRM), VIšekriterijumsko KOmpromisno Rangiranje (VIKOR)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2928
465 Personalisation of SOA Registry Query Results: Implementation, Performance Analysis and Scalability Evaluation

Authors: Kee-Leong Tan, Karyn Wei-Ju Khoo, Hui-Na Chua

Abstract:

Service discovery is a very important component of Service Oriented Architectures (SOA). This paper presents two alternative approaches to customise the query results of private service registry such as Universal Description, Discovery and Integration (UDDI). The customisation is performed based on some pre-defined and/or real-time changing parameters. This work identifies the requirements, designs and additional mechanisms that must be applied to UDDI in order to support this customisation capability. We also detail the implements of the approaches and examine its performance and scalability. Based on our experimental results, we conclude that both approaches can be used to customise registry query results, but by storing personalization parameters in external resource will yield better performance and but less scalable when size of query results increases. We believe these approaches when combined with semantics enabled service registry will enhance the service discovery methods within a private UDDI registry environment.

Keywords: Service Oriented Architecture (SOA), Web service, Service discovery, registry, UDDI

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1402