Search results for: Matrix Minimization Algorithm
1421 Fracture Toughness Properties and FTIR Analysis of Corn Fiber Green Composites
Authors: Ahmed Mudhafar Hashim, Aseel Mahmood Abdullah
Abstract:
The present work introduced a green composite consisting of corn natural fiber of constant concentration of 10% by weight incorporation with poly methyl methacrylate matrix biomaterial prepared by hand lay-up technique. Corn natural fibers were treated with two concentrations of sodium hydroxide solution (3% and 5%) with different immersed time (1.5 and 3 hours) at room temperature. The fracture toughness test of untreated and alkali treated corn fiber composites were performed. The effect of chemically treated on fracture properties of composites has been analyzed using Fourier transform infrared (FTIR) spectroscopy. The experimental results showed that the alkali treatment improved the fracture properties in terms of plane strain fracture toughness KIC. It was found that the plane strain fracture toughness KIC increased by up to 62% compared to untreated fiber composites. On the other hand, increases in both concentrations of alkali solution and time of soaking to 5% NaOH and 3 hours, respectively reduced the values of KIC lower than the value of the unfilled material.
Keywords: green composites, fracture toughness, corn natural fiber, Bio-PMMA
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5331420 Effect of a Linear-Exponential Penalty Functionon the GA-s Efficiency in Optimization of a Laminated Composite Panel
Authors: A. Abedian, M. H. Ghiasi, B. Dehghan-Manshadi
Abstract:
A stiffened laminated composite panel (1 m length × 0.5m width) was optimized for minimum weight and deflection under several constraints using genetic algorithm. Here, a significant study on the performance of a penalty function with two kinds of static and dynamic penalty factors was conducted. The results have shown that linear dynamic penalty factors are more effective than the static ones. Also, a specially combined linear-exponential function has shown to perform more effective than the previously mentioned penalty functions. This was then resulted in the less sensitivity of the GA to the amount of penalty factor.Keywords: Genetic algorithms, penalty function, stiffenedcomposite panel, finite element method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16781419 Real-time ROI Acquisition for Unsupervised and Touch-less Palmprint
Authors: Yi Feng, Jingwen Li, Lei Huang, Changping Liu
Abstract:
In this paper we proposed a novel method to acquire the ROI (Region of interest) of unsupervised and touch-less palmprint captured from a web camera in real-time. We use Viola-Jones approach and skin model to get the target area in real time. Then an innovative course-to-fine approach to detect the key points on the hand is described. A new algorithm is used to find the candidate key points coarsely and quickly. In finely stage, we verify the hand key points with the shape context descriptor. To make the user much comfortable, it can process the hand image with different poses, even the hand is closed. Experiments show promising result by using the proposed method in various conditions.Keywords: Palmprint recoginition, hand detection, touch-lesspalmprint, ROI localization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17261418 The Effect of Addition of Dioctyl Terephthalate and Calcite on the Tensile Properties of Organoclay/Linear Low Density Polyethylene Nanocomposites
Authors: A. Gürses, Z. Eroğlu, E. Şahin, K. Güneş, Ç. Doğar
Abstract:
In recent years, polymer/clay nanocomposites have generated great interest in the polymer industry as a new type of composite material because of their superior properties, which includes high heat deflection temperature, gas barrier performance, dimensional stability, enhanced mechanical properties, optical clarity and flame retardancy when compared with the pure polymer or conventional composites. The investigation of change of the tensile properties of organoclay/linear low density polyethylene (LLDPE) nanocomposites with the use of Dioctyl terephthalate (DOTP) (as plasticizer) and calcite (as filler) has been aimed. The composites and organoclay synthesized were characterized using the techniques such as XRD, HRTEM and FTIR techniques. The spectroscopic results indicate that platelets of organoclay were well dispersed within the polymeric matrix. The tensile properties of the composites were compared considering the stress-strain curve drawn for each composite and pure polymer. It was observed that the composites prepared by adding the plasticizer at different ratios and a certain amount of calcite exhibited different tensile behaviors compared to pure polymer.
Keywords: Linear low density polyethylene, nanocomposite, organoclay, plasticizer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14461417 An Improved Switching Median filter for Uniformly Distributed Impulse Noise Removal
Authors: Rajoo Pandey
Abstract:
The performance of an image filtering system depends on its ability to detect the presence of noisy pixels in the image. Most of the impulse detection schemes assume the presence of salt and pepper noise in the images and do not work satisfactorily in case of uniformly distributed impulse noise. In this paper, a new algorithm is presented to improve the performance of switching median filter in detection of uniformly distributed impulse noise. The performance of the proposed scheme is demonstrated by the results obtained from computer simulations on various images.Keywords: Switching median filter, Impulse noise, Imagefiltering, Impulse detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19571416 On Improving Breast Cancer Prediction Using GRNN-CP
Authors: Kefaya Qaddoum
Abstract:
The aim of this study is to predict breast cancer and to construct a supportive model that will stimulate a more reliable prediction as a factor that is fundamental for public health. In this study, we utilize general regression neural networks (GRNN) to replace the normal predictions with prediction periods to achieve a reasonable percentage of confidence. The mechanism employed here utilises a machine learning system called conformal prediction (CP), in order to assign consistent confidence measures to predictions, which are combined with GRNN. We apply the resulting algorithm to the problem of breast cancer diagnosis. The results show that the prediction constructed by this method is reasonable and could be useful in practice.
Keywords: Neural network, conformal prediction, cancer classification, regression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8391415 Phenotypical and Genotypical Assessment Techniques for Identification of Some Contagious Mastitis Pathogens
Authors: A. El Behiry, R. N. Zahran, R. Tarabees, E. Marzouk, M. Al-Dubaib
Abstract:
Mastitis is one of the most economic disease affecting dairy cows worldwide. Its classic diagnosis using bacterial culture and biochemical findings is a difficult and prolonged method. In this research, using of matrix-assisted laser desorption/ionization time-of-flight mass spectrometry (MALDI-TOF MS) permitted identification of different microorganisms with high accuracy and rapidity (only 24 hours for microbial growth and analysis). During the application of MALDI-TOF MS, one hundred twenty strains of Staphylococcus and Streptococcus species isolated from milk of cows affected by clinical and subclinical mastitis were identified, and the results were compared with those obtained by traditional methods as API and VITEK 2 Systems. 37 of totality 39 strains (~95%) of Staphylococcus aureus (S. aureus) were exactly detected by MALDI TOF MS and then confirmed by a nuc-based PCR technique, whereas accurate identification was observed in 100% (50 isolates) of the coagulase negative staphylococci (CNS) and Streptococcus agalactiae (31 isolates). In brief, our results demonstrated that MALDI-TOF MS is a fast and truthful technique which has the capability to replace conventional identification of several bacterial strains usually isolated in clinical laboratories of microbiology.
Keywords: Identification, mastitis pathogens, mass spectral, phenotypical.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22741414 Design and Development of Automatic Leveling and Equalizing Hoist Device for Spacecraft
Authors: Fu Hao, Sun Gang, Tang Laiying, Cui Junfeng
Abstract:
To solve the quick and accurate level-adjusting problem in the process of spacecraft precise mating, automatic leveling and equalizing hoist device for spacecraft is developed. Based on lifting point adjustment by utilizing XY-workbench, the leveling and equalizing controller by a self-adaptive control algorithm is proposed. By simulation analysis and lifting test using engineering prototype, validity and reliability of the hoist device is verified, which can meet the precision mating requirements of practical applications for spacecraft.Keywords: automatic leveling and equalizing, hoist device, lifting point adjustment, self-adaptive control
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20201413 Oxidation of Selected Pharmaceuticals in Water Matrices by Bromine and Chlorine
Authors: Juan L. Acero, F. Javier Benitez, Francisco J. Real, Gloria Roldan, Francisco Casas
Abstract:
The bromination of five selected pharmaceuticals (metoprolol, naproxen, amoxicillin, hydrochlorotiazide and phenacetin) in ultrapure water and in three water matrices (a groundwater, a surface water from a public reservoir and a secondary effluent from a WWTP) was investigated. The apparent rate constants for the bromination reaction were determined as a function of the pH, and the sequence obtained for the reaction rate was amoxicillin > naproxen >> hydrochlorotiazide ≈ phenacetin ≈ metoprolol. The proposal of a kinetic mechanism, which specifies the dissociation of bromine and each pharmaceutical according to their pKa values and the pH allowed the determination of the intrinsic rate constants for every elementary reaction. The influence of the main operating conditions (pH, initial bromine dose, and the water matrix) on the degradation of pharmaceuticals was established. In addition, the presence of bromide in chlorination experiments was investigated. The presence of bromide in wastewaters and drinking waters in the range of 10 to several hundred μg L-1 accelerated slightly the oxidation of the selected pharmaceuticals during chorine disinfection.Keywords: Pharmaceuticals, bromine, chlorine, apparent andintrinsic rate constants, water matrices, degradation rates
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23311412 A Visual Control Flow Language and Its Termination Properties
Authors: László Lengyel, Tihamér Levendovszky, Hassan Charaf
Abstract:
This paper presents the visual control flow support of Visual Modeling and Transformation System (VMTS), which facilitates composing complex model transformations out of simple transformation steps and executing them. The VMTS Visual Control Flow Language (VCFL) uses stereotyped activity diagrams to specify control flow structures and OCL constraints to choose between different control flow branches. This work discusses the termination properties of VCFL and provides an algorithm to support the termination analysis of VCFL transformations.
Keywords: Control Flow, Metamodel-Based Visual Model Transformation, OCL, Termination Properties, UML.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20661411 Index t-SNE: Tracking Dynamics of High-Dimensional Datasets with Coherent Embeddings
Authors: G. Candel, D. Naccache
Abstract:
t-SNE is an embedding method that the data science community has widely used. It helps two main tasks: to display results by coloring items according to the item class or feature value; and for forensic, giving a first overview of the dataset distribution. Two interesting characteristics of t-SNE are the structure preservation property and the answer to the crowding problem, where all neighbors in high dimensional space cannot be represented correctly in low dimensional space. t-SNE preserves the local neighborhood, and similar items are nicely spaced by adjusting to the local density. These two characteristics produce a meaningful representation, where the cluster area is proportional to its size in number, and relationships between clusters are materialized by closeness on the embedding. This algorithm is non-parametric. The transformation from a high to low dimensional space is described but not learned. Two initializations of the algorithm would lead to two different embedding. In a forensic approach, analysts would like to compare two or more datasets using their embedding. A naive approach would be to embed all datasets together. However, this process is costly as the complexity of t-SNE is quadratic, and would be infeasible for too many datasets. Another approach would be to learn a parametric model over an embedding built with a subset of data. While this approach is highly scalable, points could be mapped at the same exact position, making them indistinguishable. This type of model would be unable to adapt to new outliers nor concept drift. This paper presents a methodology to reuse an embedding to create a new one, where cluster positions are preserved. The optimization process minimizes two costs, one relative to the embedding shape and the second relative to the support embedding’ match. The embedding with the support process can be repeated more than once, with the newly obtained embedding. The successive embedding can be used to study the impact of one variable over the dataset distribution or monitor changes over time. This method has the same complexity as t-SNE per embedding, and memory requirements are only doubled. For a dataset of n elements sorted and split into k subsets, the total embedding complexity would be reduced from O(n2) to O(n2/k), and the memory requirement from n2 to 2(n/k)2 which enables computation on recent laptops. The method showed promising results on a real-world dataset, allowing to observe the birth, evolution and death of clusters. The proposed approach facilitates identifying significant trends and changes, which empowers the monitoring high dimensional datasets’ dynamics.
Keywords: Concept drift, data visualization, dimension reduction, embedding, monitoring, reusability, t-SNE, unsupervised learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4891410 Axisymmetric Nonlinear Analysis of Point Supported Shallow Spherical Shells
Authors: M. Altekin, R. F. Yükseler
Abstract:
Geometrically nonlinear axisymmetric bending of a shallow spherical shell with a point support at the apex under linearly varying axisymmetric load was investigated numerically. The edge of the shell was assumed to be simply supported or clamped. The solution was obtained by the finite difference and the Newton-Raphson methods. The thickness of the shell was considered to be uniform and the material was assumed to be homogeneous and isotropic. Sensitivity analysis was made for two geometrical parameters. The accuracy of the algorithm was checked by comparing the deflection with the solution of point supported circular plates and good agreement was obtained.
Keywords: Bending, nonlinear, plate, point support, shell.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18821409 Electric Load Forecasting Using Genetic Based Algorithm, Optimal Filter Estimator and Least Error Squares Technique: Comparative Study
Authors: Khaled M. EL-Naggar, Khaled A. AL-Rumaih
Abstract:
This paper presents performance comparison of three estimation techniques used for peak load forecasting in power systems. The three optimum estimation techniques are, genetic algorithms (GA), least error squares (LS) and, least absolute value filtering (LAVF). The problem is formulated as an estimation problem. Different forecasting models are considered. Actual recorded data is used to perform the study. The performance of the above three optimal estimation techniques is examined. Advantages of each algorithms are reported and discussed.
Keywords: Forecasting, Least error squares, Least absolute Value, Genetic algorithms
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27231408 Model Predictive Control of Turbocharged Diesel Engine with Exhaust Gas Recirculation
Authors: U. Yavas, M. Gokasan
Abstract:
Control of diesel engine’s air path has drawn a lot of attention due to its multi input-multi output, closed coupled, non-linear relation. Today, precise control of amount of air to be combusted is a must in order to meet with tight emission limits and performance targets. In this study, passenger car size diesel engine is modeled by AVL Boost RT, and then simulated with standard, industry level PID controllers. Finally, linear model predictive control is designed and simulated. This study shows the importance of modeling and control of diesel engines with flexible algorithm development in computer based systems.Keywords: Predictive control, engine control, engine modeling, PID control, feedforward compensation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18171407 A High Quality Speech Coder at 600 bps
Authors: Yong Zhang, Ruimin Hu
Abstract:
This paper presents a vocoder to obtain high quality synthetic speech at 600 bps. To reduce the bit rate, the algorithm is based on a sinusoidally excited linear prediction model which extracts few coding parameters, and three consecutive frames are grouped into a superframe and jointly vector quantization is used to obtain high coding efficiency. The inter-frame redundancy is exploited with distinct quantization schemes for different unvoiced/voiced frame combinations in the superframe. Experimental results show that the quality of the proposed coder is better than that of 2.4kbps LPC10e and achieves approximately the same as that of 2.4kbps MELP and with high robustness.
Keywords: Speech coding, Vector quantization, linear predicition, Mixed sinusoidal excitation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21881406 Dynamic Admission Control for Quality of Service in IP Networks
Authors: J. Kasigwa, V. Baryamureeba, D. Williams
Abstract:
The goal of admission control is to support the Quality of Service demands of real-time applications via resource reservation in IP networks. In this paper we introduce a novel Dynamic Admission Control (DAC) mechanism for IP networks. The DAC dynamically allocates network resources using the previous network pattern for each path and uses the dynamic admission algorithm to improve bandwidth utilization using bandwidth brokers. We evaluate the performance of the proposed mechanism through trace-driven simulation experiments in view point of blocking probability, throughput and normalized utilization.Keywords: Bandwidth broker, dynamic admission control(DAC), IP networks, quality of service, real-time flows.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12951405 Effect of Volume Fraction of Fibre on the Mechanical Properties of Nanoclay Reinforced E-Glass-Epoxy Composites
Authors: K. Krushnamurty, D. Rasmitha, I. Srikanth, K. Ramji, Ch. Subrahmanyam
Abstract:
E-glass-epoxy laminated composites having different fiber volume fractions (40, 50, 60 and 70) were fabricated with and without the addition of nanoclay. Flexural strength and tensile strength of the composite laminates were determined. It was observed that, with increasing the fiber volume fraction (Vf) of fiber from 40 to 60, the ability of nanoclay to enhance the tensile and flexural strength of E-glass-epoxy composites decreases significantly. At 70Vf, the tensile and flexural strength of the nanoclay reinforced E-glass-epoxy were found to be lowest when compared to the E-glass-epoxy composite made without the addition of nanoclay. Based on the obtained data and microstructure of the tested samples, plausible mechanism for the observed trends has been proposed. The enhanced mechanical properties for nanoclay reinforced E-glass-epoxy composites for 40-60 Vf, due to higher interface toughness coupled with strong interfilament bonding may have ensured the homogeneous load distribution across all the glass fibers. Results in the decrease in mechanical properties at 70Vf, may be due to the inability of the matrix to bind the nanoclay and glass-fibers.Keywords: E-glass-epoxy composite laminates, fiber volume fraction, e-glass fiber, mechanical properties, delamination.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28031404 Proposal of Additional Fuzzy Membership Functions in Smoothing Transition Autoregressive Models
Authors: Ε. Giovanis
Abstract:
In this paper we present, propose and examine additional membership functions for the Smoothing Transition Autoregressive (STAR) models. More specifically, we present the tangent hyperbolic, Gaussian and Generalized bell functions. Because Smoothing Transition Autoregressive (STAR) models follow fuzzy logic approach, more fuzzy membership functions should be tested. Furthermore, fuzzy rules can be incorporated or other training or computational methods can be applied as the error backpropagation or genetic algorithm instead to nonlinear squares. We examine two macroeconomic variables of US economy, the inflation rate and the 6-monthly treasury bills interest rates.Keywords: Forecast , Fuzzy membership functions, Smoothingtransition, Time-series
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15261403 A Parallel Implementation of the Reverse Converter for the Moduli Set {2n, 2n–1, 2n–1–1}
Authors: Mehdi Hosseinzadeh, Amir Sabbagh Molahosseini, Keivan Navi
Abstract:
In this paper, a new reverse converter for the moduli set {2n, 2n–1, 2n–1–1} is presented. We improved a previously introduced conversion algorithm for deriving an efficient hardware design for reverse converter. Hardware architecture of the proposed converter is based on carry-save adders and regular binary adders, without the requirement for modular adders. The presented design is faster than the latest introduced reverse converter for moduli set {2n, 2n–1, 2n–1–1}. Also, it has better performance than the reverse converters for the recently introduced moduli set {2n+1–1, 2n, 2n–1}
Keywords: Residue arithmetic, Residue number system, Residue-to-Binary converter, Reverse converter
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13141402 Iterative Clustering Algorithm for Analyzing Temporal Patterns of Gene Expression
Authors: Seo Young Kim, Jae Won Lee, Jong Sung Bae
Abstract:
Microarray experiments are information rich; however, extensive data mining is required to identify the patterns that characterize the underlying mechanisms of action. For biologists, a key aim when analyzing microarray data is to group genes based on the temporal patterns of their expression levels. In this paper, we used an iterative clustering method to find temporal patterns of gene expression. We evaluated the performance of this method by applying it to real sporulation data and simulated data. The patterns obtained using the iterative clustering were found to be superior to those obtained using existing clustering algorithms.Keywords: Clustering, microarray experiment, temporal pattern of gene expression data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13561401 Unscented Grid Filtering and Smoothing for Nonlinear Time Series Analysis
Authors: Nikolay Nikolaev, Evgueni Smirnov
Abstract:
This paper develops an unscented grid-based filter and a smoother for accurate nonlinear modeling and analysis of time series. The filter uses unscented deterministic sampling during both the time and measurement updating phases, to approximate directly the distributions of the latent state variable. A complementary grid smoother is also made to enable computing of the likelihood. This helps us to formulate an expectation maximisation algorithm for maximum likelihood estimation of the state noise and the observation noise. Empirical investigations show that the proposed unscented grid filter/smoother compares favourably to other similar filters on nonlinear estimation tasks. Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13311400 Using PFA in Feature Analysis and Selection for H.264 Adaptation
Authors: Nora A. Naguib, Ahmed E. Hussein, Hesham A. Keshk, Mohamed I. El-Adawy
Abstract:
Classification of video sequences based on their contents is a vital process for adaptation techniques. It helps decide which adaptation technique best fits the resource reduction requested by the client. In this paper we used the principal feature analysis algorithm to select a reduced subset of video features. The main idea is to select only one feature from each class based on the similarities between the features within that class. Our results showed that using this feature reduction technique the source video features can be completely omitted from future classification of video sequences.
Keywords: Adaptation, feature selection, H.264, Principal Feature Analysis (PFA)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16071399 Effect of Polyvinyl Pyrrolidone and Ethyl Cellulose Concentration on Release Profile and Kinetics of Glibenclamide Extended Release Dosage Form System
Authors: Amit Kumar, Peeyush Sharma, Anil Bhandari
Abstract:
The aim of present work was to optimize the effect of Ethyl Cellulose (EC) and Polyvinyl Pyrrolidone (PVP) concentration in extended release solid dispersion of Glibenclamide using combination of hydrophilic and hydrophobic polymers such as Polyvinyl Pyrrolidone and Ethyl cellulose. The advantage of solid dispersion technique provides a unique approach to particle size reduction and increased rates of dissolution. The compatibility studies of the drug and polymers were studied by TLC and results suggested no interaction between drug and polymers. Solid dispersions of Glibenclamide were prepared by common solvent evaporation method using Polyvinyl Pyrrolidone and Ethyl cellulose. The results indicated that homogeneous or heterogeneous conditions during the preparation methods employed governed the internal structures of the polymer matrices while retaining the drug in an amorphous form. F2 formulation prepared by solid dispersion method, displayed extended drug release followed by Higuchi matrix model indicating diffusion release of GLB from polymer matrices.
Keywords: Ethyl Cellulose, Glibenclamide, Polyvinyl Pyrrolidone, Solid Dispersion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25261398 Statistical Analysis and Predictive Learning of Mechanical Parameters for TiO2 Filled GFRP Composite
Authors: S. Srinivasa Moorthy, K. Manonmani
Abstract:
The new, polymer composites consisting of e-glass fiber reinforcement with titanium oxide filler in the double bonded unsaturated polyester resin matrix were made. The glass fiber and titanium oxide reinforcement composites were made in three different fiber lengths (3cm, 5cm, and 7cm), filler content (2 wt%, 4 wt%, and 6 wt%) and fiber content (20 wt%, 40 wt%, and 60 wt%). 27 different compositions were fabricated and a sequence of experiments were carried out to determine tensile strength and impact strength. The vital influencing factors fiber length, fiber content and filler content were chosen as 3 factors in 3 levels of Taguchi’s L9 orthogonal array. The influences of parameters were determined for tensile strength and impact strength by Analysis of variance (ANOVA) and S/N ratio. Using Artificial Neural Network (ANN) an expert system was devised to predict the properties of hybrid reinforcement GFRP composites. The predict models were experimentally proved with the maximum coincidence.
Keywords: Analysis of variance (ANOVA), Artificial neural network (ANN), Polymer composites, Taguchi’s orthogonal array.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24001397 A High Bitrate Information Hiding Algorithm for Video in Video
Authors: Wang Shou-Dao, Xiao Chuang-Bai, Lin Yu
Abstract:
In high bitrate information hiding techniques, 1 bit is embedded within each 4 x 4 Discrete Cosine Transform (DCT) coefficient block by means of vector quantization, then the hidden bit can be effectively extracted in terminal end. In this paper high bitrate information hiding algorithms are summarized, and the scheme of video in video is implemented. Experimental result shows that the host video which is embedded numerous auxiliary information have little visually quality decline. Peak Signal to Noise Ratio (PSNR)Y of host video only degrades 0.22dB in average, while the hidden information has a high percentage of survives and keeps a high robustness in H.264/AVC compression, the average Bit Error Rate(BER) of hiding information is 0.015%.Keywords: Information Hiding, Embed, Quantification, Extract
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18991396 MC and IC – What Is the Relationship?
Authors: O. V. Missioura
Abstract:
MC (Management Control)& IC (Internal Control) – what is the relationship? (an empirical study into the definitions between MC and IC) based on the wider considerations of Internal Control and Management Control terms, attention is focused not only on the financial aspects but also more on the soft aspects of the business, such as culture, behaviour, standards and values. The limited considerations of Management Control are focused mainly in the hard, financial aspects of business operation. The definitions of Management Control and Internal Control are often used interchangeably and the results of this empirical study reveal that Management Control is part of Internal Control, there is no causal link between the two concepts. Based on the interpretation of the respondents, the term Management Control has moved from a broad term to a more limited term with the soft aspects of the influencing of behaviour, performance measurements, incentives and culture. This paper is an exploratory study based on qualitative research and on a qualitative matrix method analysis of the thematic definition of the terms Management Control and Internal Control.
Keywords: Management Control (MC), Internal Control (IC), definition, causal link, COSO 1992/2004, CoCo (Canadian Institute of Chartered Accountants), Russian CG code (КОДЕКС) , limited and broad concepts MC and IC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17911395 Solving Machine Loading Problem in Flexible Manufacturing Systems Using Particle Swarm Optimization
Authors: S. G. Ponnambalam, Low Seng Kiat
Abstract:
In this paper, a particle swarm optimization (PSO) algorithm is proposed to solve machine loading problem in flexible manufacturing system (FMS), with bicriterion objectives of minimizing system unbalance and maximizing system throughput in the occurrence of technological constraints such as available machining time and tool slots. A mathematical model is used to select machines, assign operations and the required tools. The performance of the PSO is tested by using 10 sample dataset and the results are compared with the heuristics reported in the literature. The results support that the proposed PSO is comparable with the algorithms reported in the literature.Keywords: Machine loading problem, FMS, Particle Swarm Optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21211394 A Content-Based Optimization of Data Stream Television Multiplex
Authors: Jaroslav Polec, Martin Šimek, Michal Martinovič, Elena Šikudová
Abstract:
The television multiplex has reserved capacity and therefore we can use only limited number of videos for propagation of it. Appropriate composition of the multiplex has a major impact on how many videos is spread by multiplex. Therefore in this paper is designed a simple algorithm to optimize capacity utilization multiplex. Significant impact on the number of programs in the multiplex has also the fact from which programs is composed. Content of multiplex can be movies, news, sport, animated stories, documentaries, etc. These types have their own specific characteristics that affect their resulting data stream. In this paper is also done an impact analysis of the composition of the multiplex to use its capacity by video content.
Keywords: Multiplex, content, group of pictures, frame, capacity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14751393 Elimination of Redundant Links in Web Pages– Mathematical Approach
Authors: G. Poonkuzhali, K.Thiagarajan, K.Sarukesi
Abstract:
With the enormous growth on the web, users get easily lost in the rich hyper structure. Thus developing user friendly and automated tools for providing relevant information without any redundant links to the users to cater to their needs is the primary task for the website owners. Most of the existing web mining algorithms have concentrated on finding frequent patterns while neglecting the less frequent one that are likely to contain the outlying data such as noise, irrelevant and redundant data. This paper proposes new algorithm for mining the web content by detecting the redundant links from the web documents using set theoretical(classical mathematics) such as subset, union, intersection etc,. Then the redundant links is removed from the original web content to get the required information by the user..Keywords: Web documents, Web content mining, redundantlink, outliers, set theory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20151392 Comparative Study of Scheduling Algorithms for LTE Networks
Authors: Samia Dardouri, Ridha Bouallegue
Abstract:
Scheduling is the process of dynamically allocating physical resources to User Equipment (UE) based on scheduling algorithms implemented at the LTE base station. Various algorithms have been proposed by network researchers as the implementation of scheduling algorithm which represents an open issue in Long Term Evolution (LTE) standard. This paper makes an attempt to study and compare the performance of PF, MLWDF and EXP/PF scheduling algorithms. The evaluation is considered for a single cell with interference scenario for different flows such as Best effort, Video and VoIP in a pedestrian and vehicular environment using the LTE-Sim network simulator. The comparative study is conducted in terms of system throughput, fairness index, delay, packet loss ratio (PLR) and total cell spectral efficiency.
Keywords: LTE, Multimedia flows, Scheduling algorithms.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4811