Search results for: one-class classification method.
5999 Babbitt Casting and Babbitt Spraying Processes Case Study
Authors: M. Jalali Azizpour, S.Norouzi H. Mohammadi Majd
Abstract:
In this paper, the babbitting of a bearing in boiler feed pump of an electromotor has been studied. These bearings have an important role in reducing the shut down times in the pumps, compressors and turbines. The most conventional method in babbitting is casting as a melting method. The comparison between thermal spray and casting methods in babbitting shows that the thermal spraying babbitt layer has better performance and tribological behavior. The metallurgical and tribological analysis such as SEM, EDS and wet chemical analysis has been made in the Babbitt alloys and worn surfaces. Two type of babbitt materials: tinbase and lead-base babbitt was used. The benefits of thermally sprayed babbitt layers are completely clear especially in large bearings.
Keywords: Thermal spray, Babbitting, Bonding, Bearing, BFP, CWP
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 46545998 A Spatial Information Network Traffic Prediction Method Based on Hybrid Model
Authors: Jingling Li, Yi Zhang, Wei Liang, Tao Cui, Jun Li
Abstract:
Compared with terrestrial network, the traffic of spatial information network has both self-similarity and short correlation characteristics. By studying its traffic prediction method, the resource utilization of spatial information network can be improved, and the method can provide an important basis for traffic planning of a spatial information network. In this paper, considering the accuracy and complexity of the algorithm, the spatial information network traffic is decomposed into approximate component with long correlation and detail component with short correlation, and a time series hybrid prediction model based on wavelet decomposition is proposed to predict the spatial network traffic. Firstly, the original traffic data are decomposed to approximate components and detail components by using wavelet decomposition algorithm. According to the autocorrelation and partial correlation smearing and truncation characteristics of each component, the corresponding model (AR/MA/ARMA) of each detail component can be directly established, while the type of approximate component modeling can be established by ARIMA model after smoothing. Finally, the prediction results of the multiple models are fitted to obtain the prediction results of the original data. The method not only considers the self-similarity of a spatial information network, but also takes into account the short correlation caused by network burst information, which is verified by using the measured data of a certain back bone network released by the MAWI working group in 2018. Compared with the typical time series model, the predicted data of hybrid model is closer to the real traffic data and has a smaller relative root means square error, which is more suitable for a spatial information network.
Keywords: Spatial Information Network, Traffic prediction, Wavelet decomposition, Time series model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6395997 Decreasing Environmental Pollution in Superphosphate Production Using Apatite and Phosphorite Mixture
Authors: R. Guliyev
Abstract:
The enhanced need for food items is receiving more importance due to a gradual increase in the world population and, in this scenario, fertilizers play a very important role in agriculture. In this study, the production of the normal superphosphate was investigated with a continuous chamber method by adding potassium chloride to a mixture of Hibin apatite and Kingisepp phosphorite. In the experiments, the following parameters were selected: The concentration of sulfuric acid (54–66% (w/w)), the stoichiometric norm of sulfuric acid (100, 107, 110, 114% (w/w)), the ratio of apatite/phosphorite in the mixture of phosphate (95/5, 90/10, 85/15, 80/20, 75/25, 70/30, 65/35,60/40, 55/45, 50/50 (w/w)), potassium chloride/the mixture of phosphate (1/50, 2/50, 3/50,4/50, 5/50 (w/w)), and the reaction time (2–8 min). It was observed that by adding potassium chloride to a low-grade phosphorite and using it to substitute a fraction of high-grade apatite in the normal superphosphate production not only resulted in a high-quality product but also eliminated the waiting period for the maturation of superphosphate in the storage. The objective of this study was to produce a normal superphosphate fertilizer by using a continuous chamber method in order to accelerate the production process and to reduce the environmental pollution caused by fluoride gases by eliminating the maturation time in the storage.
Keywords: Continuous chamber method, environmental pollution, fluoride gases.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9025996 Optimization of Assembly and Welding of Complex 3D Structures on the Base of Modeling with Use of Finite Elements Method
Authors: M. N. Zelenin, V. S. Mikhailov, R. P. Zhivotovsky
Abstract:
It is known that residual welding deformations give negative effect to processability and operational quality of welded structures, complicating their assembly and reducing strength. Therefore, selection of optimal technology, ensuring minimum welding deformations, is one of the main goals in developing a technology for manufacturing of welded structures. Through years, JSC SSTC has been developing a theory for estimation of welding deformations and practical activities for reducing and compensating such deformations during welding process. During long time a methodology was used, based on analytic dependence. This methodology allowed defining volumetric changes of metal due to welding heating and subsequent cooling. However, dependences for definition of structures deformations, arising as a result of volumetric changes of metal in the weld area, allowed performing calculations only for simple structures, such as units, flat sections and sections with small curvature. In case of complex 3D structures, estimations on the base of analytic dependences gave significant errors. To eliminate this shortage, it was suggested to use finite elements method for resolving of deformation problem. Here, one shall first calculate volumes of longitudinal and transversal shortenings of welding joints using method of analytic dependences and further, with obtained shortenings, calculate forces, which action is equivalent to the action of active welding stresses. Further, a finiteelements model of the structure is developed and equivalent forces are added to this model. Having results of calculations, an optimal sequence of assembly and welding is selected and special measures to reduce and compensate welding deformations are developed and taken.
Keywords: Finite elements method, modeling, expected welding deformations, welding, assembling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17575995 The Model of Blended Learning and Its Use at Foreign Language Teaching
Authors: A. A. Kudysheva, A. N. Kudyshev
Abstract:
In present article the model of Blended Learning, its advantage at foreign language teaching, and also some problems that can arise during its use are considered. The Blended Learning is a special organization of learning, which allows to combine classroom work and modern technologies in electronic distance teaching environment. Nowadays a lot of European educational institutions and companies use such technology. Through this method: student gets the opportunity to learn in a group (classroom) with a teacher and additionally at home at a convenient time; student himself sets the optimal speed and intensity of the learning process; this method helps student to discipline himself and learn to work independently.
Keywords: Foreign language, information and communication technology (ICT), model of Blended Learning, virtual cool room, technophobia
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 33935994 Pythagorean-Platonic Lattice Method for Finding all Co-Prime Right Angle Triangles
Authors: Anthony Overmars, Sitalakshmi Venkatraman
Abstract:
This paper presents a method for determining all of the co-prime right angle triangles in the Euclidean field by looking at the intersection of the Pythagorean and Platonic right angle triangles and the corresponding lattice that this produces. The co-prime properties of each lattice point representing a unique right angle triangle are then considered. This paper proposes a conjunction between these two ancient disparaging theorists. This work has wide applications in information security where cryptography involves improved ways of finding tuples of prime numbers for secure communication systems. In particular, this paper has direct impact in enhancing the encryption and decryption algorithms in cryptography.
Keywords: Pythagorean triples, platonic triples, right angle triangles, co-prime numbers, cryptography.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12605993 Application of Extreme Learning Machine Method for Time Series Analysis
Authors: Rampal Singh, S. Balasundaram
Abstract:
In this paper, we study the application of Extreme Learning Machine (ELM) algorithm for single layered feedforward neural networks to non-linear chaotic time series problems. In this algorithm the input weights and the hidden layer bias are randomly chosen. The ELM formulation leads to solving a system of linear equations in terms of the unknown weights connecting the hidden layer to the output layer. The solution of this general system of linear equations will be obtained using Moore-Penrose generalized pseudo inverse. For the study of the application of the method we consider the time series generated by the Mackey Glass delay differential equation with different time delays, Santa Fe A and UCR heart beat rate ECG time series. For the choice of sigmoid, sin and hardlim activation functions the optimal values for the memory order and the number of hidden neurons which give the best prediction performance in terms of root mean square error are determined. It is observed that the results obtained are in close agreement with the exact solution of the problems considered which clearly shows that ELM is a very promising alternative method for time series prediction.Keywords: Chaotic time series, Extreme learning machine, Generalization performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35205992 Fuzzy Multi-Criteria Decision-Making Based on Ignatian Discernment Process
Authors: Pathinathan Theresanathan, Ajay Minj
Abstract:
Ignatian Discernment Process (IDP) is an intense decision-making tool to decide on life-issues. Decisions are influenced by various factors outside of the decision maker and inclination within. This paper develops IDP in the context of Fuzzy Multi-criteria Decision Making (FMCDM) process. Extended VIKOR method is a decision-making method which encompasses even conflict situations and accommodates weightage to various issues. Various aspects of IDP, namely three ways of decision making and tactics of inner desires, are observed, analyzed and articulated within the frame work of fuzzy rules. The decision-making situations are broadly categorized into two types. The issues outside of the decision maker influence the person. The inner feeling also plays vital role in coming to a conclusion. IDP integrates both the categories using Extended VIKOR method. Case studies are carried out and analyzed with FMCDM process. Finally, IDP is verified with an illustrative case study and results are interpreted. A confused person who could not come to a conclusion is able to take decision on a concrete way of life through IDP. The proposed IDP model recommends an integrated and committed approach to value-based decision making.
Keywords: Analytical hierarchy process, fuzzy multi-criteria decision making, Ignatian discernment process, Ignatian discernment, multi-criteria decision making, VIKOR.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12735991 Some Application of Random Fuzzy Queueing System Based On Fuzzy Simulation
Authors: Behrouz Fathi-Vajargah, Sara Ghasemalipour
Abstract:
This paper studies a random fuzzy queueing system that the interarrival times of customers arriving at the server and the service times are independent and identically distributed random fuzzy variables. We match the random fuzzy queueing system with the random fuzzy alternating renewal process and we do not use from α-pessimistic and α-optimistic values to estimate the average chance of the event ”random fuzzy queueing system is busy at time t”, we employ the fuzzy simulation method in practical applications. Some theorem is proved and finally we solve a numerical example with fuzzy simulation method.
Keywords: Random fuzzy variables, Fuzzy simulation, Queueing system, Interarrival times.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20845990 Automated Inspection Algorithm for Thick Plate Using Dual Light Switching Lighting Method
Authors: Yong-JuJeon, Doo-chul Choi, Jong Pil Yun, Changhyun Park, Homoon Bae, Sang Woo Kim
Abstract:
This paper presents an automated inspection algorithm for a thick plate. Thick plates typically have various types of surface defects, such as scabs, scratches, and roller marks. These defects have individual characteristics including brightness and shape. Therefore, it is not simple to detect all the defects. In order to solve these problems and to detect defects more effectively, we propose a dual light switching lighting method and a defect detection algorithm based on Gabor filters.Keywords: Thick plate, Defect, Inspection, Gabor filter, Dual Light Switching.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18775989 Teaching Speaking Skills to Adult English Language Learners through ALM
Authors: Wichuda Kunnu, Aungkana Sukwises
Abstract:
Audio-lingual Method (ALM) is a teaching approach that is claimed that ineffective for teaching second/foreign languages. Because some linguists and second/foreign language teachers believe that ALM is a rote learning style. However, this study is done on a belief that ALM will be able to solve Thais’ English speaking problem. This paper aims to report the findings on teaching English speaking to adult learners with an “adapted ALM”, one distinction of which is to use Thai as the medium language of instruction. The participants are consisted of 9 adult learners. They were allowed to speak English more freely using both the materials presented in the class and their background knowledge of English. At the end of the course, they spoke English more fluently, more confidently, to the extent that they applied what they learnt both in and outside the class.
Keywords: Teaching English, Audio Lingual Method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 39515988 STLF Based on Optimized Neural Network Using PSO
Authors: H. Shayeghi, H. A. Shayanfar, G. Azimi
Abstract:
The quality of short term load forecasting can improve the efficiency of planning and operation of electric utilities. Artificial Neural Networks (ANNs) are employed for nonlinear short term load forecasting owing to their powerful nonlinear mapping capabilities. At present, there is no systematic methodology for optimal design and training of an artificial neural network. One has often to resort to the trial and error approach. This paper describes the process of developing three layer feed-forward large neural networks for short-term load forecasting and then presents a heuristic search algorithm for performing an important task of this process, i.e. optimal networks structure design. Particle Swarm Optimization (PSO) is used to develop the optimum large neural network structure and connecting weights for one-day ahead electric load forecasting problem. PSO is a novel random optimization method based on swarm intelligence, which has more powerful ability of global optimization. Employing PSO algorithms on the design and training of ANNs allows the ANN architecture and parameters to be easily optimized. The proposed method is applied to STLF of the local utility. Data are clustered due to the differences in their characteristics. Special days are extracted from the normal training sets and handled separately. In this way, a solution is provided for all load types, including working days and weekends and special days. The experimental results show that the proposed method optimized by PSO can quicken the learning speed of the network and improve the forecasting precision compared with the conventional Back Propagation (BP) method. Moreover, it is not only simple to calculate, but also practical and effective. Also, it provides a greater degree of accuracy in many cases and gives lower percent errors all the time for STLF problem compared to BP method. Thus, it can be applied to automatically design an optimal load forecaster based on historical data.
Keywords: Large Neural Network, Short-Term Load Forecasting, Particle Swarm Optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22245987 Human Face Detection and Segmentation using Eigenvalues of Covariance Matrix, Hough Transform and Raster Scan Algorithms
Authors: J. Prakash, K. Rajesh
Abstract:
In this paper we propose a novel method for human face segmentation using the elliptical structure of the human head. It makes use of the information present in the edge map of the image. In this approach we use the fact that the eigenvalues of covariance matrix represent the elliptical structure. The large and small eigenvalues of covariance matrix are associated with major and minor axial lengths of an ellipse. The other elliptical parameters are used to identify the centre and orientation of the face. Since an Elliptical Hough Transform requires 5D Hough Space, the Circular Hough Transform (CHT) is used to evaluate the elliptical parameters. Sparse matrix technique is used to perform CHT, as it squeeze zero elements, and have only a small number of non-zero elements, thereby having an advantage of less storage space and computational time. Neighborhood suppression scheme is used to identify the valid Hough peaks. The accurate position of the circumference pixels for occluded and distorted ellipses is identified using Bresenham-s Raster Scan Algorithm which uses the geometrical symmetry properties. This method does not require the evaluation of tangents for curvature contours, which are very sensitive to noise. The method has been evaluated on several images with different face orientations.Keywords: Circular Hough Transform, Covariance matrix, Eigenvalues, Elliptical Hough Transform, Face segmentation, Raster Scan Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25175986 Determining a Suitable Maintenance Measure for Gentelligent Components Using Case-Based Reasoning
Authors: M. Winkens, P. Nyhuis
Abstract:
Components with sensory properties such as gentelligent components developed at the Collaborative Research Centre 653 offer a new angle in terms of the full utilization of the remaining service life as well as preventive maintenance. The developed methodology of component status driven maintenance analyzes the stress data obtained during the component's useful life and on the basis of this knowledge assesses the type of maintenance required in this case. The procedure is derived from the case-based reasoning method and will be explained in detail. The method's functionality is demonstrated with real-life data obtained during test runs of a racing car prototype.
Keywords: Gentelligent Components, Preventive Maintenance, Case based Reasoning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19055985 Adaptive Motion Planning for 6-DOF Robots Based on Trigonometric Functions
Authors: Jincan Li, Mingyu Gao, Zhiwei He, Yuxiang Yang, Zhongfei Yu, Yuanyuan Liu
Abstract:
Building an appropriate motion model is crucial for trajectory planning of robots and determines the operational quality directly. An adaptive acceleration and deceleration motion planning based on trigonometric functions for the end-effector of 6-DOF robots in Cartesian coordinate system is proposed in this paper. This method not only achieves the smooth translation motion and rotation motion by constructing a continuous jerk model, but also automatically adjusts the parameters of trigonometric functions according to the variable inputs and the kinematic constraints. The results of computer simulation show that this method is correct and effective to achieve the adaptive motion planning for linear trajectories.
Keywords: 6-DOF robots, motion planning, trigonometric function, kinematic constraints
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9255984 Analysis of Linguistic Disfluencies in Bilingual Children’s Discourse
Authors: Sheena Christabel Pravin, M. Palanivelan
Abstract:
Speech disfluencies are common in spontaneous speech. The primary purpose of this study was to distinguish linguistic disfluencies from stuttering disfluencies in bilingual Tamil–English (TE) speaking children. The secondary purpose was to determine whether their disfluencies are mediated by native language dominance and/or on an early onset of developmental stuttering at childhood. A detailed study was carried out to identify the prosodic and acoustic features that uniquely represent the disfluent regions of speech. This paper focuses on statistical modeling of repetitions, prolongations, pauses and interjections in the speech corpus encompassing bilingual spontaneous utterances from school going children – English and Tamil. Two classifiers including Hidden Markov Models (HMM) and the Multilayer Perceptron (MLP), which is a class of feed-forward artificial neural network, were compared in the classification of disfluencies. The results of the classifiers document the patterns of disfluency in spontaneous speech samples of school-aged children to distinguish between Children Who Stutter (CWS) and Children with Language Impairment CLI). The ability of the models in classifying the disfluencies was measured in terms of F-measure, Recall, and Precision.
Keywords: Bilingual, children who stutter, children with language impairment, Hidden Markov Models, multi-layer perceptron, linguistic disfluencies, stuttering disfluencies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10295983 Optimized Vector Quantization for Bayer Color Filter Array
Authors: M. Lakshmi, J. Senthil Kumar
Abstract:
Digital cameras to reduce cost, use an image sensor to capture color images. Color Filter Array (CFA) in digital cameras permits only one of the three primary (red-green-blue) colors to be sensed in a pixel and interpolates the two missing components through a method named demosaicking. Captured data is interpolated into a full color image and compressed in applications. Color interpolation before compression leads to data redundancy. This paper proposes a new Vector Quantization (VQ) technique to construct a VQ codebook with Differential Evolution (DE) Algorithm. The new technique is compared to conventional Linde- Buzo-Gray (LBG) method.Keywords: Color Filter Array (CFA), Biorthogonal Wavelet, Vector Quantization (VQ), Differential Evolution (DE).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19075982 Entropy Generation for Natural Convection in a Darcy – Brinkman Porous Cavity
Authors: Ali Mchirgui, Nejib Hidouri, Mourad Magherbi, Ammar Ben Brahim
Abstract:
The paper provides a numerical investigation of the entropy generation analysis due to natural convection in an inclined square porous cavity. The coupled equations of mass, momentum, energy and species conservation are solved using the Control Volume Finite-Element Method. Effect of medium permeability and inclination angle on entropy generation is analysed. It was found that according to the Darcy number and the porous thermal Raleigh number values, the entropy generation could be mainly due to heat transfer or to fluid friction irreversibility and that entropy generation reaches extremum values for specific inclination angles.Keywords: Porous media, entropy generation, convection, numerical method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26085981 Ecotoxicity Evaluation and Suggestion of Remediation Method of ZnO Nanoparticles in Aqueous Phase
Authors: Hyunsang Kim, Younghun Kim, Younghee Kim, Sangku Lee
Abstract:
We investigated ecotoxicity and performed experiment for removing ZnO nanoparticles in water. Short term exposure of hatching test using fertilized eggs (O. latipes) showed deformity in 5ppm of ZnO nanoparticles solution. And in 10ppm ZnO nanoparticles solution delayed hatching was observed. Hereine, chemical precipitation method was suggested for removing ZnO nanoparticles in water. The precipitated ZnO nanoparticles showed the form of ZnS after addition of Na2S, and the form of Zn3(PO4)2 for Na2HPO4. The removal efficiency of ZnO nanoparticles in water was closed to 100% for two cases. In ecotoxicity evaluation of as-precipitated ZnS and Zn3(PO4)2, they did not cause any acute toxicity for D. magna. It is noted that this precipitation treatment of ZnO is effective to reduce the potential cytotoxicity.Keywords: ZnO nanoparticles, ZnS, Zn3(PO4)2, ecotoxicity evaluation, chemical precipitation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18695980 Rotation Invariant Face Recognition Based on Hybrid LPT/DCT Features
Authors: Rehab F. Abdel-Kader, Rabab M. Ramadan, Rawya Y. Rizk
Abstract:
The recognition of human faces, especially those with different orientations is a challenging and important problem in image analysis and classification. This paper proposes an effective scheme for rotation invariant face recognition using Log-Polar Transform and Discrete Cosine Transform combined features. The rotation invariant feature extraction for a given face image involves applying the logpolar transform to eliminate the rotation effect and to produce a row shifted log-polar image. The discrete cosine transform is then applied to eliminate the row shift effect and to generate the low-dimensional feature vector. A PSO-based feature selection algorithm is utilized to search the feature vector space for the optimal feature subset. Evolution is driven by a fitness function defined in terms of maximizing the between-class separation (scatter index). Experimental results, based on the ORL face database using testing data sets for images with different orientations; show that the proposed system outperforms other face recognition methods. The overall recognition rate for the rotated test images being 97%, demonstrating that the extracted feature vector is an effective rotation invariant feature set with minimal set of selected features.Keywords: Discrete Cosine Transform, Face Recognition, Feature Extraction, Log Polar Transform, Particle SwarmOptimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18735979 Fabrication of ZnO Nanorods Based Biosensor via Hydrothermal Method
Authors: Muhammad Tariq, Jafar Khan Kasi, Samiullah, Ajab Khan Kasi
Abstract:
Biosensors are playing vital role in industrial, clinical, and chemical analysis applications. Among other techniques, ZnO based biosensor is an easy approach due to its exceptional chemical and electrical properties. ZnO nanorods have positively charged isoelectric point which helps immobilize the negative charge glucose oxides (GOx). Here, we report ZnO nanorods based biosensors for the immobilization of GOx. The ZnO nanorods were grown by hydrothermal method on indium tin oxide substrate (ITO). The fabrication of biosensors was carried through batch processing using conventional photolithography. The buffer solutions of GOx were prepared in phosphate with a pH value of around 7.3. The biosensors effectively immobilized the GOx and result was analyzed by calculation of voltage and current on nanostructures.Keywords: Hydrothermal growth, zinc dioxide, biosensors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10555978 Integrated Subset Split for Balancing Network Utilization and Quality of Routing
Authors: S. V. Kasmir Raja, P. Herbert Raj
Abstract:
The overlay approach has been widely used by many service providers for Traffic Engineering (TE) in large Internet backbones. In the overlay approach, logical connections are set up between edge nodes to form a full mesh virtual network on top of the physical topology. IP routing is then run over the virtual network. Traffic engineering objectives are achieved through carefully routing logical connections over the physical links. Although the overlay approach has been implemented in many operational networks, it has a number of well-known scaling issues. This paper proposes a new approach to achieve traffic engineering without full-mesh overlaying with the help of integrated approach and equal subset split method. Traffic engineering needs to determine the optimal routing of traffic over the existing network infrastructure by efficiently allocating resource in order to optimize traffic performance on an IP network. Even though constraint-based routing [1] of Multi-Protocol Label Switching (MPLS) is developed to address this need, since it is not widely tested or debugged, Internet Service Providers (ISPs) resort to TE methods under Open Shortest Path First (OSPF), which is the most commonly used intra-domain routing protocol. Determining OSPF link weights for optimal network performance is an NP-hard problem. As it is not possible to solve this problem, we present a subset split method to improve the efficiency and performance by minimizing the maximum link utilization in the network via a small number of link weight modifications. The results of this method are compared against results of MPLS architecture [9] and other heuristic methods.
Keywords: Constraint based routing, Link Utilization, Subsetsplit method and Traffic Engineering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13985977 Zero Truncated Strict Arcsine Model
Authors: Y. N. Phang, E. F. Loh
Abstract:
The zero truncated model is usually used in modeling count data without zero. It is the opposite of zero inflated model. Zero truncated Poisson and zero truncated negative binomial models are discussed and used by some researchers in analyzing the abundance of rare species and hospital stay. Zero truncated models are used as the base in developing hurdle models. In this study, we developed a new model, the zero truncated strict arcsine model, which can be used as an alternative model in modeling count data without zero and with extra variation. Two simulated and one real life data sets are used and fitted into this developed model. The results show that the model provides a good fit to the data. Maximum likelihood estimation method is used in estimating the parameters.
Keywords: Hurdle models, maximum likelihood estimation method, positive count data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18575976 Adaptive Transmission Scheme Based on Channel State in Dual-Hop System
Authors: Seung-Jun Yu, Yong-Jun Kim, Jung-In Baik, Hyoung-Kyu Song
Abstract:
In this paper, a dual-hop relay based on channel state is studied. In the conventional relay scheme, a relay uses the same modulation method without reference to channel state. But, a relay uses an adaptive modulation method with reference to channel state. If the channel state is poor, a relay eliminates latter 2 bits and uses Quadrature Phase Shift Keying (QPSK) modulation. If channel state is good, a relay modulates the received symbols with 16-QAM symbols by using 4 bits. The performance of the proposed scheme for Symbol Error Rate (SER) and throughput is analyzed.Keywords: Adaptive transmission, channel state, dual-hop, hierarchical modulation, relay.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18085975 Sensory, Microbiological and Chemical Assessment of Cod (Gadus morhua) Fillets during Chilled Storage as Influenced by Bleeding Methods
Authors: Minh Van Nguyen, Magnea Gudrun Karlsdottir, Adalheidur Olafsdottir, Arnljotur Bjarki Bergsson, Sigurjon Arason
Abstract:
The effects of seawater and slurry ice bleeding methods on the sensory, microbiological and chemical quality changes of cod fillets during chilled storage were examined in this study. The results from sensory evaluation showed that slurry ice bleeding method prolonged the shelf life of cod fillets up to 13-14 days compared to 10-11 days for fish bled in seawater. Slurry ice bleeding method also led to a slower microbial growth and biochemical developments, resulting lower total plate count (TPC), H2S-producing bacteria count, total volatile basic nitrogen (TVB-N), trimethylamine (TMA), free fatty acid (FFA) content and higher phospholipid content (PL) compared to those of samples bled in seawater. The results of principle component analysis revealed that TPC, H2S-producing bacteria, TVB-N, TMA and FFA were in significant correlation. They were also in negative correlation with sensory evaluation (Torry score), PL and water holding capacity (WHC).
Keywords: Bleeding method, chilled storage, microbial growth, sensory evaluation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29855974 Optimizing Forecasting for Indonesia's Coal and Palm Oil Exports: A Comparative Analysis of ARIMA, ANN, and LSTM Methods
Authors: Mochammad Dewo, Sumarsono Sudarto
Abstract:
The Exponential Triple Smoothing Algorithm approach nowadays, which is used to anticipate the export value of Indonesia's two major commodities, coal and palm oil, has a Mean Percentage Absolute Error (MAPE) value of 30-50%, which may be considered as a "reasonable" forecasting mistake. Forecasting errors of more than 30% shall have a domino effect on industrial output, as extra production adds to raw material, manufacturing and storage expenses. Whereas, reaching an "excellent" classification with an error value of less than 10% will provide new investors and exporters with confidence in the commercial development of related sectors. Industrial growth will bring out a positive impact on economic development. It can be applied for other commodities if the forecast error is less than 10%. The purpose of this project is to create a forecasting technique that can produce precise forecasting results with an error of less than 10%. This research analyzes forecasting methods such as ARIMA (Autoregressive Integrated Moving Average), ANN (Artificial Neural Network) and LSTM (Long-Short Term Memory). By providing a MAPE of 1%, this study reveals that ANN is the most successful strategy for forecasting coal and palm oil commodities in Indonesia.
Keywords: ANN, Artificial Neural Network, ARIMA, Autoregressive Integrated Moving Average, export value, forecast, LSTM, Long Short Term Memory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2245973 Feasibility of Leukemia Cancer Treatment (K562) by Atmospheric Pressure Plasma Jet
Authors: Mashayekh Amir Shahriar, Akhlaghi Morteza, Rajaee Hajar, Khani Mohammad Reza, Shokri Babak
Abstract:
A new and novel approach in medicine is the use of cold plasma for various applications such as sterilization blood coagulation and cancer cell treatment. In this paper a pin-to-hole plasma jet suitable for biological applications is investigated and characterized and the possibility and feasibility of cancer cell treatment is evaluated. The characterization includes power consumption via Lissajous method, thermal behavior of plasma using Infra-red camera as a novel method, Optical Emission Spectroscopy (OES) to determine the species that are generated. Treatment of leukemia cancer cells is also implemented and MTT assay is used to evaluate viability.
Keywords: Atmospheric Pressure Plasma Jet (APPJ), Plasma Medicine, Cancer cell treatment, leukemia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22365972 Design Modelling Control and Simulation of DC/DC Power Buck Converter
Authors: H. Abaali
Abstract:
The power buck converter is the most widely used DC/DC converter topology. They have a very large application area such as DC motor drives, photovoltaic power system which require fast transient responses and high efficiency over a wide range of load current. This work proposes, the modelling of DC/DC power buck converter using state-space averaging method and the current-mode control using a proportional-integral controller. The efficiency of the proposed model and control loop are evaluated with operating point changes. The simulation results proved the effectiveness of the linear model of DC/DC power buck converter.Keywords: DC/DC power buck converter, Linear current control, State-space averaging method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34805971 Automatic Segmentation of the Clean Speech Signal
Authors: M. A. Ben Messaoud, A. Bouzid, N. Ellouze
Abstract:
Speech Segmentation is the measure of the change point detection for partitioning an input speech signal into regions each of which accords to only one speaker. In this paper, we apply two features based on multi-scale product (MP) of the clean speech, namely the spectral centroid of MP, and the zero crossings rate of MP. We focus on multi-scale product analysis as an important tool for segmentation extraction. The MP is based on making the product of the speech wavelet transform coefficients (WTC). We have estimated our method on the Keele database. The results show the effectiveness of our method. It indicates that the two features can find word boundaries, and extracted the segments of the clean speech.
Keywords: Speech segmentation, Multi-scale product, Spectral centroid, Zero crossings rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25085970 Computational Method for Annotation of Protein Sequence According to Gene Ontology Terms
Authors: Razib M. Othman, Safaai Deris, Rosli M. Illias
Abstract:
Annotation of a protein sequence is pivotal for the understanding of its function. Accuracy of manual annotation provided by curators is still questionable by having lesser evidence strength and yet a hard task and time consuming. A number of computational methods including tools have been developed to tackle this challenging task. However, they require high-cost hardware, are difficult to be setup by the bioscientists, or depend on time intensive and blind sequence similarity search like Basic Local Alignment Search Tool. This paper introduces a new method of assigning highly correlated Gene Ontology terms of annotated protein sequences to partially annotated or newly discovered protein sequences. This method is fully based on Gene Ontology data and annotations. Two problems had been identified to achieve this method. The first problem relates to splitting the single monolithic Gene Ontology RDF/XML file into a set of smaller files that can be easy to assess and process. Thus, these files can be enriched with protein sequences and Inferred from Electronic Annotation evidence associations. The second problem involves searching for a set of semantically similar Gene Ontology terms to a given query. The details of macro and micro problems involved and their solutions including objective of this study are described. This paper also describes the protein sequence annotation and the Gene Ontology. The methodology of this study and Gene Ontology based protein sequence annotation tool namely extended UTMGO is presented. Furthermore, its basic version which is a Gene Ontology browser that is based on semantic similarity search is also introduced.
Keywords: automatic clustering, bioinformatics tool, gene ontology, protein sequence annotation, semantic similarity search
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3128