Search results for: measurement accuracy
2624 Failure Analysis of Low Relaxation Prestressed High Carbon Steel Wire During Drawing Operation: A Metallurgical Investigation
Authors: Souvik Das, Sandip Bhattacharya, Goutam Mukhopadhyay, Manashi Adhikary
Abstract:
Wires breakages during cold drawing are a complex phenomenon; wire breakages may be induced by improper wire-rod quality, inappropriate heat-treated microstructure, and/or lubrication breakdown on the wire surface. A comprehensive metallurgical investigation of failed/broken wire samples is therefore essential for understanding the origin of failure. Frequent breakage of wires during drawing is a matter of serious concern to the wire drawers as it erodes their already slim margins through reduced productivity and loss in yield. The present paper highlights the failure investigation of wires of Low Relaxation Prestressed High Carbon grade during cold drawing due to entrapment of hard constituents detached from the roller entry guide during rolling operations. The hardness measurement of this entrapped location indicates 54.9 Rockwell Hardness as against the rest portion 33.4 Rockwell Hardness. The microstructure chemical analysis and X-ray mapping analysis data of the entrapment location confirmed complex chromium carbide originated from D2-steel used in entry guide during the rolling process. Since the harder entrapped phase could not be deformed in the same manner as the parent phase, the failure of the wire rod occurs during hot rolling.Keywords: LRPC, D2-steel, chromium carbide, roller guide
Procedia PDF Downloads 1592623 Modeling Slow Crack Growth under Thermal and Chemical Effects for Fitness Predictions of High-Density Polyethylene Material
Authors: Luis Marquez, Ge Zhu, Vikas Srivastava
Abstract:
High-density polyethylene (HDPE) is one of the most commonly used thermoplastic polymer materials for water and gas pipelines. Slow crack growth failure is a well-known phenomenon in high-density polyethylene material and causes brittle failure well below the yield point with no obvious sign. The failure of transportation pipelines can cause catastrophic environmental and economic consequences. Using the non-destructive testing method to predict slow crack growth failure behavior is the primary preventative measurement employed by the pipeline industry but is often costly and time-consuming. Phenomenological slow crack growth models are useful to predict the slow crack growth behavior in the polymer material due to their ability to evaluate slow crack growth under different temperature and loading conditions. We developed a quantitative method to assess the slow crack growth behavior in the high-density polyethylene pipeline material under different thermal conditions based on existing physics-based phenomenological models. We are also working on developing an experimental protocol and quantitative model that can address slow crack growth behavior under different chemical exposure conditions to improve the safety, reliability, and resilience of HDPE-based pipeline infrastructure.Keywords: mechanics of materials, physics-based modeling, civil engineering, fracture mechanics
Procedia PDF Downloads 2062622 Development, Optimization, and Validation of a Synchronous Fluorescence Spectroscopic Method with Multivariate Calibration for the Determination of Amlodipine and Olmesartan Implementing: Experimental Design
Authors: Noha Ibrahim, Eman S. Elzanfaly, Said A. Hassan, Ahmed E. El Gendy
Abstract:
Objectives: The purpose of the study is to develop a sensitive synchronous spectrofluorimetric method with multivariate calibration after studying and optimizing the different variables affecting the native fluorescence intensity of amlodipine and olmesartan implementing an experimental design approach. Method: In the first step, the fractional factorial design used to screen independent factors affecting the intensity of both drugs. The objective of the second step was to optimize the method performance using a Central Composite Face-centred (CCF) design. The optimal experimental conditions obtained from this study were; a temperature of (15°C ± 0.5), the solvent of 0.05N HCl and methanol with a ratio of (90:10, v/v respectively), Δλ of 42 and the addition of 1.48 % surfactant providing a sensitive measurement of amlodipine and olmesartan. The resolution of the binary mixture with a multivariate calibration method has been accomplished mainly by using partial least squares (PLS) model. Results: The recovery percentage for amlodipine besylate and atorvastatin calcium in tablets dosage form were found to be (102 ± 0.24, 99.56 ± 0.10, for amlodipine and Olmesartan, respectively). Conclusion: Method is valid according to some International Conference on Harmonization (ICH) guidelines, providing to be linear over a range of 200-300, 500-1500 ng mL⁻¹ for amlodipine and Olmesartan. The methods were successful to estimate amlodipine besylate and olmesartan in bulk powder and pharmaceutical preparation.Keywords: amlodipine, central composite face-centred design, experimental design, fractional factorial design, multivariate calibration, olmesartan
Procedia PDF Downloads 1512621 Auditory and Language Skills Development after Cochlear Implantation in Children with Multiple Disabilities
Authors: Tamer Mesallam, Medhat Yousef, Ayna Almasaad
Abstract:
BACKGROUND: Cochlear implantation (CI) in children with additional disabilities can be a fundamental and supportive intervention. Although, there may be some positive impacts of CI on children with multiple disabilities such as better outcomes of communication skills, development, and quality of life, the families of those children complain from the post-implant habilitation efforts that considered as a burden. OBJECTIVE: To investigate the outcomes of CI children with different co-disabilities through using the Meaningful Auditory Integration Scale (MAIS) and the Meaningful Use of Speech Scale (MUSS) as outcome measurement tools. METHODS: The study sample comprised 25 hearing-impaired children with co-disability who received cochlear implantation. Age and gender-matched control group of 25 cochlear-implanted children without any other disability has been also included. The participants' auditory skills and speech outcomes were assessed using MAIS and MUSS tests. RESULTS: There was a statistically significant difference in the different outcomes measure between the two groups. However, the outcomes of some multiple disabilities subgroups were comparable to the control group. Around 40% of the participants with co-disabilities experienced advancement in their methods of communication from behavior to oral mode. CONCLUSION: Cochlear-implanted children with multiple disabilities showed variable degrees of auditory and speech outcomes. The degree of benefits depends on the type of the co-disability. Long-term follow-up is recommended for those children.Keywords: children with disabilities, Cochlear implants, hearing impairment, language development
Procedia PDF Downloads 1212620 Feature Weighting Comparison Based on Clustering Centers in the Detection of Diabetic Retinopathy
Authors: Kemal Polat
Abstract:
In this paper, three feature weighting methods have been used to improve the classification performance of diabetic retinopathy (DR). To classify the diabetic retinopathy, features extracted from the output of several retinal image processing algorithms, such as image-level, lesion-specific and anatomical components, have been used and fed them into the classifier algorithms. The dataset used in this study has been taken from University of California, Irvine (UCI) machine learning repository. Feature weighting methods including the fuzzy c-means clustering based feature weighting, subtractive clustering based feature weighting, and Gaussian mixture clustering based feature weighting, have been used and compered with each other in the classification of DR. After feature weighting, five different classifier algorithms comprising multi-layer perceptron (MLP), k- nearest neighbor (k-NN), decision tree, support vector machine (SVM), and Naïve Bayes have been used. The hybrid method based on combination of subtractive clustering based feature weighting and decision tree classifier has been obtained the classification accuracy of 100% in the screening of DR. These results have demonstrated that the proposed hybrid scheme is very promising in the medical data set classification.Keywords: machine learning, data weighting, classification, data mining
Procedia PDF Downloads 3282619 Performance Comparison of Wideband Covariance Matrix Sparse Representation (W-CMSR) with Other Wideband DOA Estimation Methods
Authors: Sandeep Santosh, O. P. Sahu
Abstract:
In this paper, performance comparison of wideband covariance matrix sparse representation (W-CMSR) method with other existing wideband Direction of Arrival (DOA) estimation methods has been made.W-CMSR relies less on a priori information of the incident signal number than the ordinary subspace based methods.Consider the perturbation free covariance matrix of the wideband array output. The diagonal covariance elements are contaminated by unknown noise variance. The covariance matrix of array output is conjugate symmetric i.e its upper right triangular elements can be represented by lower left triangular ones.As the main diagonal elements are contaminated by unknown noise variance,slide over them and align the lower left triangular elements column by column to obtain a measurement vector.Simulation results for W-CMSR are compared with simulation results of other wideband DOA estimation methods like Coherent signal subspace method (CSSM), Capon, l1-SVD, and JLZA-DOA. W-CMSR separate two signals very clearly and CSSM, Capon, L1-SVD and JLZA-DOA fail to separate two signals clearly and an amount of pseudo peaks exist in the spectrum of L1-SVD.Keywords: W-CMSR, wideband direction of arrival (DOA), covariance matrix, electrical and computer engineering
Procedia PDF Downloads 4722618 The Impact of Management Competency, Project Team, and Process Design to Corporate Performance through Implementing the Self-Development ERP
Authors: Zeplin Jiwa Husada Tarigan, Sautma Ronni Basana, Widjojo Suprapto
Abstract:
Manufacturing companies in East Java develop their own ERP system or alter the ERP system which is developed by other companies to suit their needs. To make their own system, the companies mostly assign several employees from various departments to create a project team, and the employees are from the departments that are going to utilize the ERP system as the integrated data. The project team decides the making of the ERP system from the preparation stage until the going live implementation process. In designing the business process, the top management is working together with the project team until the project is accomplished. The completion of the ERP projects depends on the project to be undertaken itself, the strategy chosen to complete the project, the work method selection, the measurement system to monitor the project, the evaluation system of the project, and, in the end, the declaration of 'going live' of the ERP project. There is an increase in the business performance for the companies that have implemented the information technology or ERP as they manage to integrate all management functions within their companies. To investigate, some questionnaires are distributed to 100 manufacturing companies, and 90 questionnaires are returned; however, there are only 46 companies that develop their own ERP system, so the response rate is 46%. The result of data analysis using PLS shows that the management competency brings impacts to the project team and the process design. The process design is adjusted to the real process in order to implement the ERP, but it does not bring direct impacts to the business performance. The implementation of ERP brings positive impacts to the company business performance.Keywords: management competency, project team, process design, ERP implementation, business performance
Procedia PDF Downloads 2192617 A Fuzzy Approach to Liver Tumor Segmentation with Zernike Moments
Authors: Abder-Rahman Ali, Antoine Vacavant, Manuel Grand-Brochier, Adélaïde Albouy-Kissi, Jean-Yves Boire
Abstract:
In this paper, we present a new segmentation approach for liver lesions in regions of interest within MRI (Magnetic Resonance Imaging). This approach, based on a two-cluster Fuzzy C-Means methodology, considers the parameter variable compactness to handle uncertainty. Fine boundaries are detected by a local recursive merging of ambiguous pixels with a sequential forward floating selection with Zernike moments. The method has been tested on both synthetic and real images. When applied on synthetic images, the proposed approach provides good performance, segmentations obtained are accurate, their shape is consistent with the ground truth, and the extracted information is reliable. The results obtained on MR images confirm such observations. Our approach allows, even for difficult cases of MR images, to extract a segmentation with good performance in terms of accuracy and shape, which implies that the geometry of the tumor is preserved for further clinical activities (such as automatic extraction of pharmaco-kinetics properties, lesion characterization, etc).Keywords: defuzzification, floating search, fuzzy clustering, Zernike moments
Procedia PDF Downloads 4532616 Comparison of Efficient Production of Small Module Gears
Authors: Vaclav Musil, Robert Cep, Sarka Malotova, Jiri Hajnys, Frantisek Spalek
Abstract:
The new designs of satellite gears comprising a number of small gears pose high requirements on the precise production of small module gears. The objective of the experimental activity stated in this article was to compare the conventional rolling gear cutting technology with the modern wire electrical discharge machining (WEDM) technology for the production of small module gear m=0.6 mm (thickness of 2.5 mm and material 30CrMoV9). The WEDM technology lies in copying the profile of gearing from the rendered trajectory which is then transferred to the track of a wire electrode. During the experiment, we focused on the comparison of these production methods. Main measured parameters which significantly influence the lifetime and noise was chosen. The first parameter was to compare the precision of gearing profile in respect to the mathematic model. The second monitored parameter was the roughness and surface topology of the gear tooth side. The experiment demonstrated high accuracy of WEDM technology, but a low quality of machined surface.Keywords: precision of gearing, small module gears, surface topology, WEDM technology
Procedia PDF Downloads 2342615 Nonlinear Analysis of Shear Deformable Deep Beam Resting on Nonlinear Two-Parameter Random Soil
Authors: M. Seguini, D. Nedjar
Abstract:
In this paper, the nonlinear analysis of Timoshenko beam undergoing moderate large deflections and resting on nonlinear two-parameter random foundation is presented, taking into account the effects of shear deformation, beam’s properties variation and the spatial variability of soil characteristics. The finite element probabilistic analysis has been performed by using Timoshenko beam theory with the Von Kàrmàn nonlinear strain-displacement relationships combined to Vanmarcke theory and Monte Carlo simulations, which is implemented in a Matlab program. Numerical examples of the newly developed model is conducted to confirm the efficiency and accuracy of this later and the importance of accounting for the foundation second parameter (Winkler-Pasternak). Thus, the results obtained from the developed model are presented and compared with those available in the literature to examine how the consideration of the shear and spatial variability of soil’s characteristics affects the response of the system.Keywords: nonlinear analysis, soil-structure interaction, large deflection, Timoshenko beam, Euler-Bernoulli beam, Winkler foundation, Pasternak foundation, spatial variability
Procedia PDF Downloads 3242614 A Program Based on Artistic and Musical Activities to Acquire Some Educational Concepts for Children with Learning Difficulties
Authors: Ahmed Amin Mousa, Huda Mazeed, Eman Saad
Abstract:
The study aims to identify the extent of the effectiveness of the artistic formation program using some types of pastes to reduce the hyperactivity of the kindergarten child. The researcher has discussed the effectiveness of the artistic program using some types of pastes in reducing the hyperactivity of the kindergarten child. The research sample included 120 children of ages between 5 to 6 years old from the five schools for special needs section learning disability, Cairo Province. The study used the empirical like curriculum which depends on designing one group using the before and after application measurement for the group to validate the fidelity of both the hypothesis and the effectiveness of the program. The variables of the study were specified as follows; artistic formation program using paper Mache as an independent variable and its effect on skills of kindergarten child with learning disabilities as a subsequent variable. The researchers depended on applying a group of artistic formation program using pulp melding skills for kindergarten children with learning disabilities. The tools of the study, designed by the researcher, included: recording card used for recording the Effective program using pulp molding skills for kindergarten children with learning disabilities during practicing the artistic formation activity. In additional, there was a program using pulp molding skills for kindergarten children with learning disabilities. The results proved the effectiveness of the program using pulp molding skills for kindergarten children with learning disabilities.Keywords: artistic program, developing skills, kindergarten, children, learning disabilities
Procedia PDF Downloads 1612613 Intrusion Detection System Using Linear Discriminant Analysis
Authors: Zyad Elkhadir, Khalid Chougdali, Mohammed Benattou
Abstract:
Most of the existing intrusion detection systems works on quantitative network traffic data with many irrelevant and redundant features, which makes detection process more time’s consuming and inaccurate. A several feature extraction methods, such as linear discriminant analysis (LDA), have been proposed. However, LDA suffers from the small sample size (SSS) problem which occurs when the number of the training samples is small compared with the samples dimension. Hence, classical LDA cannot be applied directly for high dimensional data such as network traffic data. In this paper, we propose two solutions to solve SSS problem for LDA and apply them to a network IDS. The first method, reduce the original dimension data using principal component analysis (PCA) and then apply LDA. In the second solution, we propose to use the pseudo inverse to avoid singularity of within-class scatter matrix due to SSS problem. After that, the KNN algorithm is used for classification process. We have chosen two known datasets KDDcup99 and NSLKDD for testing the proposed approaches. Results showed that the classification accuracy of (PCA+LDA) method outperforms clearly the pseudo inverse LDA method when we have large training data.Keywords: LDA, Pseudoinverse, PCA, IDS, NSL-KDD, KDDcup99
Procedia PDF Downloads 2292612 Resolution and Experimental Validation of the Asymptotic Model of a Viscous Laminar Supersonic Flow around a Thin Airfoil
Authors: Eddegdag Nasser, Naamane Azzeddine, Radouani Mohammed, Ensam Meknes
Abstract:
In this study, we are interested in the asymptotic modeling of the two-dimensional stationary supersonic flow of a viscous compressible fluid around wing airfoil. The aim of this article is to solve the partial differential equations of the flow far from the leading edge and near the wall using the triple-deck technique is what brought again in precision according to the principle of least degeneration. In order to validate our theoretical model, these obtained results will be compared with the experimental results. The comparison of the results of our model with experimentation has shown that they are quantitatively acceptable compared to the obtained experimental results. The experimental study was conducted using the AF300 supersonic wind tunnel and a NACA Reduced airfoil model with two pressure Taps on extrados. In this experiment, we have considered the incident upstream supersonic Mach number over a dissymmetric NACA airfoil wing. The validation and the accuracy of the results support our model.Keywords: supersonic, viscous, triple deck technique, asymptotic methods, AF300 supersonic wind tunnel, reduced airfoil model
Procedia PDF Downloads 2432611 DOA Estimation Using Golden Section Search
Authors: Niharika Verma, Sandeep Santosh
Abstract:
DOA technique is a localization technique used in the communication field. Various algorithms have been developed for direction of arrival estimation like MUSIC, ROOT MUSIC, etc. These algorithms depend on various parameters like antenna array elements, number of snapshots and various others. Basically the MUSIC spectrum is evaluated and peaks obtained are considered as the angle of arrivals. The angles evaluated using this process depends on the scanning interval chosen. The accuracy of the results obtained depends on the coarseness of the interval chosen. In this paper, golden section search is applied to the MUSIC algorithm and therefore, more accurate results are achieved. Initially the coarse DOA estimations is done using the MUSIC algorithm in the range -90 to 90 degree at the interval of 10 degree. After the peaks obtained then fine DOA estimation is done using golden section search. Also, the partitioning method is applied to estimate the number of signals incident on the antenna array. Dependency of the algorithm on the number of snapshots is also being explained. Hence, the accurate results are being determined using this algorithm.Keywords: Direction of Arrival (DOA), golden section search, MUSIC, number of snapshots
Procedia PDF Downloads 4482610 Computer Aided Diagnostic System for Detection and Classification of a Brain Tumor through MRI Using Level Set Based Segmentation Technique and ANN Classifier
Authors: Atanu K Samanta, Asim Ali Khan
Abstract:
Due to the acquisition of huge amounts of brain tumor magnetic resonance images (MRI) in clinics, it is very difficult for radiologists to manually interpret and segment these images within a reasonable span of time. Computer-aided diagnosis (CAD) systems can enhance the diagnostic capabilities of radiologists and reduce the time required for accurate diagnosis. An intelligent computer-aided technique for automatic detection of a brain tumor through MRI is presented in this paper. The technique uses the following computational methods; the Level Set for segmentation of a brain tumor from other brain parts, extraction of features from this segmented tumor portion using gray level co-occurrence Matrix (GLCM), and the Artificial Neural Network (ANN) to classify brain tumor images according to their respective types. The entire work is carried out on 50 images having five types of brain tumor. The overall classification accuracy using this method is found to be 98% which is significantly good.Keywords: brain tumor, computer-aided diagnostic (CAD) system, gray-level co-occurrence matrix (GLCM), tumor segmentation, level set method
Procedia PDF Downloads 5142609 Effectiveness of Using Phonemic Awareness Based Activities in Improving Decoding Skills of Third Grade Students Referred for Reading Disabilities in Oman
Authors: Mahmoud Mohamed Emam
Abstract:
In Oman the number of students referred for reading disabilities is on the rise. Schools serve these students by placement in the so-called learning disabilities unit. Recently the author led a strategic project to train teachers on the use of curriculum based measurement to identify students with reading disabilities in Oman. Additional the project involved training teachers to use phonemic awareness based activities to improve reading skills of those students. Phonemic awareness refers to the ability to notice, think about, and work with the individual sounds in words. We know that a student's skill in phonemic awareness is a good predictor of later reading success or difficulty. Using multiple baseline design across four participants the current studies investigated the effectiveness of using phonemic awareness based activities to improve decoding skills of third grade students referred for reading disabilities in Oman. During treatment students received phonemic awareness based activities that were designed to fulfill the idiosyncratic characteristics of Arabic language phonology as well as orthography. Results indicated that the phonemic awareness based activities were effective in substantially increasing the number of correctly decoded word for all four participants. Maintenance of strategy effects was evident for the weeks following the termination of intervention for the four students. In addition, the effects of intervention generalized to decoding novel words for all four participants.Keywords: learning disabilities, phonemic awareness, third graders, Oman
Procedia PDF Downloads 6432608 Cooperative Sensing for Wireless Sensor Networks
Authors: Julien Romieux, Fabio Verdicchio
Abstract:
Wireless Sensor Networks (WSNs), which sense environmental data with battery-powered nodes, require multi-hop communication. This power-demanding task adds an extra workload that is unfairly distributed across the network. As a result, nodes run out of battery at different times: this requires an impractical individual node maintenance scheme. Therefore we investigate a new Cooperative Sensing approach that extends the WSN operational life and allows a more practical network maintenance scheme (where all nodes deplete their batteries almost at the same time). We propose a novel cooperative algorithm that derives a piecewise representation of the sensed signal while controlling approximation accuracy. Simulations show that our algorithm increases WSN operational life and spreads communication workload evenly. Results convey a counterintuitive conclusion: distributing workload fairly amongst nodes may not decrease the network power consumption and yet extend the WSN operational life. This is achieved as our cooperative approach decreases the workload of the most burdened cluster in the network.Keywords: cooperative signal processing, signal representation and approximation, power management, wireless sensor networks
Procedia PDF Downloads 3932607 Design and Implementation of the Embedded Control System for the Electrical Motor Based Cargo Vehicle
Authors: Syed M. Rizvi, Yiqing Meng, Simon Iwnicki
Abstract:
With an increased demand in the land cargo industry, it is predicted that the freight trade will rise to a record $1.1 trillion in revenue and volume in the following years to come. This increase is mainly driven by the e-commerce model ever so popular in the consumer market. Many innovative ideas have stemmed from this demand and change in lifestyle likes of which include e-bike cargo and drones. Rural and urban areas are facing air quality challenges to keep pollution levels in city centre to a minimum. For this purpose, this paper presents the design and implementation of a non-linear PID control system, employing a micro-controller and low cost sensing technique, for controlling an electrical motor based cargo vehicle with various loads, to follow a leading vehicle (bike). Within using this system, the cargo vehicle will have no load influence on the bike rider on different gradient conditions, such as hill climbing. The system is being integrated with a microcontroller to continuously measure several parameters such as relative displacement between bike and the cargo vehicle and gradient of the road, and process these measurements to create a portable controller capable of controlling the performance of electrical vehicle without the need of a PC. As a result, in the case of carrying 180kg of parcel weight, the cargo vehicle can maintain a reasonable spacing over a short length of sensor travel between the bike and itself.Keywords: cargo, e-bike, microcontroller, embedded system, nonlinear pid, self-adaptive, inertial measurement unit (IMU)
Procedia PDF Downloads 2102606 Effectiveness of an Early Intensive Behavioral Intervention Program on Infants with Autism Spectrum Disorder
Authors: Dongjoo Chin
Abstract:
The purpose of this study was to investigate the effectiveness of an Early Intensive Behavioral Intervention (EIBI) program on infants with autism spectrum disorder (ASD) and to explore the factors predicting the effectiveness of the program, focusing on the infant's age, language ability, problem behaviors, and parental stress. 19 pairs of infants aged between 2 and 5 years who have had been diagnosed with ASD, and their parents participated in an EIBI program at a clinic providing evidence-based treatment based on applied behavior analysis. The measurement tools which were administered before and after the EIBI program and compared, included PEP-R, a curriculum evaluation, K-SIB-R, K-Vineland-II, K-CBCL, and PedsQL for the infants, and included PSI-SF and BDI-II for the parents. Statistical analysis was performed using a sample t-test and multiple regression analysis and the results were as follows. The EIBI program showed significant improvements in overall developmental age, curriculum assessment, and quality of life for infants. There was no difference in parenting stress or depression. Furthermore, measures for both children and parents at the start of the program predicted neither PEP-R nor the degree of improvement in curriculum evaluation measured six months later at the end of the program. Based on these results, the authors suggest future directions for developing an effective intensive early intervention (EIBI) program for infants with ASD in Korea, and discuss the implications and limitations of this study.Keywords: applied behavior analysis, autism spectrum disorder, early intensive behavioral intervention, parental stress
Procedia PDF Downloads 1742605 Simulation of Wind Generator with Fixed Wind Turbine under Matlab-Simulink
Authors: Mahdi Motahari, Mojtaba Farzaneh, Armin Parsian Nejad
Abstract:
The rapidly growing wind industry is highly expressing the need for education and training worldwide, particularly on the system level. Modelling and simulating wind generator system using Matlab-Simulink provides expert help in understanding wind systems engineering and system design. Working under Matlab-Simulink we present the integration of the developed WECS model with public electrical grid. A test of the calculated power and Cp related to the experimental equivalent data, using statistical analysis is performed. The statistical indicators of accuracy show better results of the presented method with RMSE: 21%, 22%, MBE : 0.77%, 0.12 % and MAE :3%, 4%.On the other hand we study its behavior when integrated in whole power system. Three level of wind speeds have been chosen: low with 5m/s as the mean value, medium with 8m/s as the mean value and high speed with 12m/s as the mean value. These allowed predicting and supervising the active power produced by the system, characterized respectively by the middle powers of -150 kW, -250kW and -480 kW which will be injected directly into the public electrical grid and the reactive power, characterized respectively by the middle powers of 60 kW, 180 kW and 320 kW and will be consumed by the wind generator.Keywords: modelling, simulation, wind generator, fixed speed wind turbine, Matlab-Simulink
Procedia PDF Downloads 6302604 Comparative Analysis of Predictive Models for Customer Churn Prediction in the Telecommunication Industry
Authors: Deepika Christopher, Garima Anand
Abstract:
To determine the best model for churn prediction in the telecom industry, this paper compares 11 machine learning algorithms, namely Logistic Regression, Support Vector Machine, Random Forest, Decision Tree, XGBoost, LightGBM, Cat Boost, AdaBoost, Extra Trees, Deep Neural Network, and Hybrid Model (MLPClassifier). It also aims to pinpoint the top three factors that lead to customer churn and conducts customer segmentation to identify vulnerable groups. According to the data, the Logistic Regression model performs the best, with an F1 score of 0.6215, 81.76% accuracy, 68.95% precision, and 56.57% recall. The top three attributes that cause churn are found to be tenure, Internet Service Fiber optic, and Internet Service DSL; conversely, the top three models in this article that perform the best are Logistic Regression, Deep Neural Network, and AdaBoost. The K means algorithm is applied to establish and analyze four different customer clusters. This study has effectively identified customers that are at risk of churn and may be utilized to develop and execute strategies that lower customer attrition.Keywords: attrition, retention, predictive modeling, customer segmentation, telecommunications
Procedia PDF Downloads 582603 A Deep Learning-Based Pedestrian Trajectory Prediction Algorithm
Authors: Haozhe Xiang
Abstract:
With the rise of the Internet of Things era, intelligent products are gradually integrating into people's lives. Pedestrian trajectory prediction has become a key issue, which is crucial for the motion path planning of intelligent agents such as autonomous vehicles, robots, and drones. In the current technological context, deep learning technology is becoming increasingly sophisticated and gradually replacing traditional models. The pedestrian trajectory prediction algorithm combining neural networks and attention mechanisms has significantly improved prediction accuracy. Based on in-depth research on deep learning and pedestrian trajectory prediction algorithms, this article focuses on physical environment modeling and learning of historical trajectory time dependence. At the same time, social interaction between pedestrians and scene interaction between pedestrians and the environment were handled. An improved pedestrian trajectory prediction algorithm is proposed by analyzing the existing model architecture. With the help of these improvements, acceptable predicted trajectories were successfully obtained. Experiments on public datasets have demonstrated the algorithm's effectiveness and achieved acceptable results.Keywords: deep learning, graph convolutional network, attention mechanism, LSTM
Procedia PDF Downloads 732602 A Greedy Alignment Algorithm Supporting Medication Reconciliation
Authors: David Tresner-Kirsch
Abstract:
Reconciling patient medication lists from multiple sources is a critical task supporting the safe delivery of patient care. Manual reconciliation is a time-consuming and error-prone process, and recently attempts have been made to develop efficiency- and safety-oriented automated support for professionals performing the task. An important capability of any such support system is automated alignment – finding which medications from a list correspond to which medications from a different source, regardless of misspellings, naming differences (e.g. brand name vs. generic), or changes in treatment (e.g. switching a patient from one antidepressant class to another). This work describes a new algorithmic solution to this alignment task, using a greedy matching approach based on string similarity, edit distances, concept extraction and normalization, and synonym search derived from the RxNorm nomenclature. The accuracy of this algorithm was evaluated against a gold-standard corpus of 681 medication records; this evaluation found that the algorithm predicted alignments with 99% precision and 91% recall. This performance is sufficient to support decision support applications for medication reconciliation.Keywords: clinical decision support, medication reconciliation, natural language processing, RxNorm
Procedia PDF Downloads 2882601 A Clustering Algorithm for Massive Texts
Authors: Ming Liu, Chong Wu, Bingquan Liu, Lei Chen
Abstract:
Internet users have to face the massive amount of textual data every day. Organizing texts into categories can help users dig the useful information from large-scale text collection. Clustering, in fact, is one of the most promising tools for categorizing texts due to its unsupervised characteristic. Unfortunately, most of traditional clustering algorithms lose their high qualities on large-scale text collection. This situation mainly attributes to the high- dimensional vectors generated from texts. To effectively and efficiently cluster large-scale text collection, this paper proposes a vector reconstruction based clustering algorithm. Only the features that can represent the cluster are preserved in cluster’s representative vector. This algorithm alternately repeats two sub-processes until it converges. One process is partial tuning sub-process, where feature’s weight is fine-tuned by iterative process. To accelerate clustering velocity, an intersection based similarity measurement and its corresponding neuron adjustment function are proposed and implemented in this sub-process. The other process is overall tuning sub-process, where the features are reallocated among different clusters. In this sub-process, the features useless to represent the cluster are removed from cluster’s representative vector. Experimental results on the three text collections (including two small-scale and one large-scale text collections) demonstrate that our algorithm obtains high quality on both small-scale and large-scale text collections.Keywords: vector reconstruction, large-scale text clustering, partial tuning sub-process, overall tuning sub-process
Procedia PDF Downloads 4382600 Time-Frequency Feature Extraction Method Based on Micro-Doppler Signature of Ground Moving Targets
Authors: Ke Ren, Huiruo Shi, Linsen Li, Baoshuai Wang, Yu Zhou
Abstract:
Since some discriminative features are required for ground moving targets classification, we propose a new feature extraction method based on micro-Doppler signature. Firstly, the time-frequency analysis of measured data indicates that the time-frequency spectrograms of the three kinds of ground moving targets, i.e., single walking person, two people walking and a moving wheeled vehicle, are discriminative. Then, a three-dimensional time-frequency feature vector is extracted from the time-frequency spectrograms to depict these differences. At last, a Support Vector Machine (SVM) classifier is trained with the proposed three-dimensional feature vector. The classification accuracy to categorize ground moving targets into the three kinds of the measured data is found to be over 96%, which demonstrates the good discriminative ability of the proposed micro-Doppler feature.Keywords: micro-doppler, time-frequency analysis, feature extraction, radar target classification
Procedia PDF Downloads 4072599 Application of Two Stages Adaptive Neuro-Fuzzy Inference System to Improve Dissolved Gas Analysis Interpretation Techniques
Authors: Kharisma Utomo Mulyodinoto, Suwarno, A. Abu-Siada
Abstract:
Dissolved Gas Analysis is one of impressive technique to detect and predict internal fault of transformers by using gas generated by transformer oil sample. A number of methods are used to interpret the dissolved gas from transformer oil sample: Doernenberg Ratio Method, IEC (International Electrotechnical Commission) Ratio Method, and Duval Triangle Method. While the assessment of dissolved gas within transformer oil samples has been standardized over the past two decades, analysis of the results is not always straight forward as it depends on personnel expertise more than mathematical formulas. To get over this limitation, this paper is aimed at improving the interpretation of Doernenberg Ratio Method, IEC Ratio Method, and Duval Triangle Method using Two Stages Adaptive Neuro-Fuzzy Inference System (ANFIS). Dissolved gas analysis data from 520 faulty transformers was analyzed to establish the proposed ANFIS model. Results show that the developed ANFIS model is accurate and can standardize the dissolved gas interpretation process with accuracy higher than 90%.Keywords: ANFIS, dissolved gas analysis, Doernenberg ratio method, Duval triangular method, IEC ratio method, transformer
Procedia PDF Downloads 1502598 Rising of Single and Double Bubbles during Boiling and Effect of Electric Field in This Process
Authors: Masoud Gholam Ale Mohammad, Mojtaba Hafezi Birgani
Abstract:
An experimental study of saturated pool boiling on a single artificial nucleation site without and with the application of an electric field on the boiling surface has been conducted. N-pentane is boiling on a copper surface and is recorded with a high speed camera providing high quality pictures and movies. The accuracy of the visualization allowed establishing an experimental bubble growth law from a large number of experiments. This law shows that the evaporation rate is decreasing during the bubble growth, and underlines the importance of liquid motion induced by the preceding bubble. Bubble rise is therefore studied: once detached, bubbles accelerate vertically until reaching a maximum velocity in good agreement with a correlation from literature. The bubbles then turn to another direction. The effect of applying an electric field on the boiling surface in finally studied. In addition to changes in the bubble shape, changes are also shown in the liquid plume and the convective structures above the surface. Lower maximum rising velocities were measured in the presence of electric fields, especially with a negative polarity.Keywords: single and double bubbles, electric field, boiling, rising
Procedia PDF Downloads 2282597 Comparison of Different Intraocular Lens Power Calculation Formulas in People With Very High Myopia
Authors: Xia Chen, Yulan Wang
Abstract:
purpose: To compare the accuracy of Haigis, SRK/T, T2, Holladay 1, Hoffer Q, Barrett Universal II, Emmetropia Verifying Optical (EVO) and Kane for intraocular lens power calculation in patients with axial length (AL) ≥ 28 mm. Methods: In this retrospective single-center study, 50 eyes of 41 patients with AL ≥ 28 mm that underwent uneventful cataract surgery were enrolled. The actual postoperative refractive results were compared to the predicted refraction calculated with different formulas (Haigis, SRK/T, T2, Holladay 1, Hoffer Q, Barrett Universal II, EVO and Kane). The mean absolute prediction errors (MAE) 1 month postoperatively were compared. Results: The MAE of different formulas were as follows: Haigis (0.509), SRK/T (0.705), T2 (0.999), Holladay 1 (0.714), Hoffer Q (0.583), Barrett Universal II (0.552), EVO (0.463) and Kane (0.441). No significant difference was found among the different formulas (P = .122). The Kane and EVO formulas achieved the lowest level of mean prediction error (PE) and median absolute error (MedAE) (p < 0.05). Conclusion: The Kane and EVO formulas had a better success rate than others in predicting IOL power in high myopic eyes with AL longer than 28 mm in this study.Keywords: cataract, power calculation formulas, intraocular lens, long axial length
Procedia PDF Downloads 882596 Novel Hybrid Ceramic Nanocomposites Fabricated by Rapid Sintering Technology
Authors: Iftikhar Ahmad, Abulhakim Almajid
Abstract:
Alumina (Al2O3) is an attractive structural ceramic however; brittleness turns Al2O3 down for advanced applications. Development of multi-phase phase ceramics systems is promising to curtail the brittleness and the incorporation of strong/elastic graphene, as third phase, into dual phase (Al2O3-SiC) is striking for mechanical upgrading purpose. Thin graphene nanosheets (GNS) were prepared by thermal exfoliation process and reinforced into dual phase ceramic system. The hybrid nanocomposite was consolidated by novel HF-IH (high-frequency induction heating) sintering furnace at 1500 °C under 50 MPa in vacuum conditions. Structural features and grain size of the resulting nanocomposite were analyzed by SEM and TEM whilst the mechanical properties were assessed by microhardness and nanoindentation techniques. The fracture toughness of the hybrid nanocomposites was appraised by direct crack measurement method. Electron microscopic investigations confirmed the preparation of thin (< 10 nm) graphene nanosheets (GNS). HF-IH sintering route condensed the three-phase (GNS-Al2O3-SiC) hybrid nanocomposite system to > 99% relative densities. SEM of the hybrid nanocomposites fractured surfaces revealed even distribution of the nanocomposite constituents and changed in fracture-mode. Structurally, 88% grain reduction into hybrid nanocomposite was also obtained. Mechanically, enhanced fracture toughness (50%) and hardness (53%) were also achieved for hybrid nanocomposites were attained against bench marked monolithic Al2O3.Keywords: alumina, graphene, hybrid nanocomposites, rapid sintering
Procedia PDF Downloads 3812595 Real-Time Observation of Concentration Distribution for Mix Liquids including Water in Micro Fluid Channel with Near-Infrared Spectroscopic Imaging Method
Authors: Hiroki Takiguchi, Masahiro Furuya, Takahiro Arai
Abstract:
In order to quantitatively comprehend thermal flow for some industrial applications such as nuclear and chemical reactors, detailed measurements for temperature and abundance (concentration) of materials at high temporal and spatial resolution are required. Additionally, rigorous evaluation of the size effect is also important for practical realization. This paper introduces a real-time spectroscopic imaging method in micro scale field, which visualizes temperature and concentration distribution of a liquid or mix liquids with near-infrared (NIR) wavelength region. This imaging principle is based on absorption of pre-selected narrow band from absorption spectrum peak or its dependence property of target liquid in NIR region. For example, water has a positive temperature sensitivity in the wavelength at 1905 nm, therefore the temperature of water can be measured using the wavelength band. In the experiment, the real-time imaging observation of concentration distribution in micro channel was demonstrated to investigate the applicability of micro-scale diffusion coefficient and temperature measurement technique using this proposed method. The effect of thermal diffusion and binary mutual diffusion was evaluated with the time-series visualizations of concentration distribution.Keywords: near-infrared spectroscopic imaging, micro fluid channel, concentration distribution, diffusion phenomenon
Procedia PDF Downloads 162