Search results for: Protein Structure Data.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9858

Search results for: Protein Structure Data.

8268 Churn Prediction for Telecommunication Industry Using Artificial Neural Networks

Authors: Ulas Vural, M. Ergun Okay, E. Mesut Yildiz

Abstract:

Telecommunication service providers demand accurate and precise prediction of customer churn probabilities to increase the effectiveness of their customer relation services. The large amount of customer data owned by the service providers is suitable for analysis by machine learning methods. In this study, expenditure data of customers are analyzed by using an artificial neural network (ANN). The ANN model is applied to the data of customers with different billing duration. The proposed model successfully predicts the churn probabilities at 83% accuracy for only three months expenditure data and the prediction accuracy increases up to 89% when the nine month data is used. The experiments also show that the accuracy of ANN model increases on an extended feature set with information of the changes on the bill amounts.

Keywords: Customer relationship management, churn prediction, telecom industry, deep learning, Artificial Neural Networks, ANN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 760
8267 A Technical Perspective on Roadway Safety in Eastern Province: Data Evaluation and Spatial Analysis

Authors: Muhammad Farhan, Sayed Faruque, Amr Mohammed, Sami Osman, Omar Al-Jabari, Abdul Almojil

Abstract:

Saudi Arabia in recent years has seen drastic increase in traffic related crashes. With population of over 29 million, Saudi Arabia is considered as a fast growing and emerging economy. The rapid population increase and economic growth has resulted in rapid expansion of transportation infrastructure, which has led to increase in road crashes. Saudi Ministry of Interior reported more than 7,000 people killed and 68,000 injured in 2011 ranking Saudi Arabia to be one of the worst worldwide in traffic safety. The traffic safety issues in the country also result in distress to road users and cause and economic loss exceeding 3.7 billion Euros annually. Keeping this in view, the researchers in Saudi Arabia are investigating ways to improve traffic safety conditions in the country. This paper presents a multilevel approach to collect traffic safety related data required to do traffic safety studies in the region. Two highway corridors including King Fahd Highway 39 kilometre and Gulf Cooperation Council Highway 42 kilometre long connecting the cities of Dammam and Khobar were selected as a study area. Traffic data collected included traffic counts, crash data, travel time data, and speed data. The collected data was analysed using geographic information system to evaluate any correlation. Further research is needed to investigate the effectiveness of traffic safety related data when collected in a concerted effort.

Keywords: Crash Data, Data Collection, Traffic Safety.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2352
8266 Machine Scoring Model Using Data Mining Techniques

Authors: Wimalin S. Laosiritaworn, Pongsak Holimchayachotikul

Abstract:

this article proposed a methodology for computer numerical control (CNC) machine scoring. The case study company is a manufacturer of hard disk drive parts in Thailand. In this company, sample of parts manufactured from CNC machine are usually taken randomly for quality inspection. These inspection data were used to make a decision to shut down the machine if it has tendency to produce parts that are out of specification. Large amount of data are produced in this process and data mining could be very useful technique in analyzing them. In this research, data mining techniques were used to construct a machine scoring model called 'machine priority assessment model (MPAM)'. This model helps to ensure that the machine with higher risk of producing defective parts be inspected before those with lower risk. If the defective prone machine is identified sooner, defective part and rework could be reduced hence improving the overall productivity. The results showed that the proposed method can be successfully implemented and approximately 351,000 baht of opportunity cost could have saved in the case study company.

Keywords: Computer Numerical Control, Data Mining, HardDisk Drive.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1395
8265 Malaysian Multi-Ethnic Discrimination Scale: Preliminary Factor and Psychometric Analysis

Authors: Chua Bee Seok, Shamsul Amri Baharuddin, Rosnah Ismail, Ferlis Bahari, Jasmine Adela Mutang, Lailawati Madlan, Asong Joseph

Abstract:

The aims of this study were to determine the factor structure and psychometric properties (i.e., reliability and convergent validity) of the Malaysian Multi-Ethnic Discrimination Scale (MMEDS). It consists of 71-items measure experience, strategies used and consequences of ethnic discrimination. A sample of 649 university students from one of the higher education institution in Malaysia was asked to complete MMEDS, as well as Perceived Ethnic and Racial Discrimination. The exploratory factor analysis on ethnic discrimination experience extracted two factors labeled ‘unfair treatment’ (15 items) and ‘Denial of the ethnic right’ (12 items) which accounted for 60.92% of the total variance. The two sub scales demonstrated clear reliability with internal consistency above .70. The convergent validity of the Scale was supported by an expected pattern of correlations (positive and significant correlation) between the score of unfair treatment and denial of the ethnic right and the score of Perceived Ethnic and Racial Discrimination by Peers Scale. The results suggest that the MMEDS is a reliable and valid measure. However, further studies need to be carried out in other groups of sample as to validate the Scale.

Keywords: Factor structure, psychometric properties, exploratory factor analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2495
8264 The Impact of Seasonality on Rainfall Patterns: A Case Study

Authors: Priti Kaushik, Randhir Singh Baghel, Somil Khandelwal

Abstract:

This study uses whole-year data from Rajasthan, India, at the meteorological divisional level to analyze and evaluate long-term spatiotemporal trends in rainfall and looked at the data from each of the thirteen tehsils in the Jaipur district to see how the rainfall pattern has altered over the last 10 years. Data on daily rainfall from the Indian Meteorological Department (IMD) in Jaipur are available for the years 2012 through 2021. We mainly focus on comparing data of tehsil wise in the Jaipur district, Rajasthan, India. Also analyzed is the fact that July and August always see higher rainfall than any other month. Rainfall usually starts to rise around week 25th and peaks in weeks 32nd or 33rd. They showed that on several occasions, 2017 saw the least amount of rainfall during a long span of 10 years. The greatest rain fell between 2012 and 2021 in 2013, 2019, and 2020.

Keywords: Data analysis, extreme events, rainfall, descriptive case studies, precipitation temperature.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 191
8263 Enhance the Power of Sentiment Analysis

Authors: Yu Zhang, Pedro Desouza

Abstract:

Since big data has become substantially more accessible and manageable due to the development of powerful tools for dealing with unstructured data, people are eager to mine information from social media resources that could not be handled in the past. Sentiment analysis, as a novel branch of text mining, has in the last decade become increasingly important in marketing analysis, customer risk prediction and other fields. Scientists and researchers have undertaken significant work in creating and improving their sentiment models. In this paper, we present a concept of selecting appropriate classifiers based on the features and qualities of data sources by comparing the performances of five classifiers with three popular social media data sources: Twitter, Amazon Customer Reviews, and Movie Reviews. We introduced a couple of innovative models that outperform traditional sentiment classifiers for these data sources, and provide insights on how to further improve the predictive power of sentiment analysis. The modeling and testing work was done in R and Greenplum in-database analytic tools.

Keywords: Sentiment Analysis, Social Media, Twitter, Amazon, Data Mining, Machine Learning, Text Mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3518
8262 Effect of Germination on Proximate, Available Phenol and Flavonoid Content, and Antioxidant Activities of African Yam Bean (Sphenostylis stenocarpa)

Authors: Nneka N. Uchegbu, Ndidi F. Amulu

Abstract:

The work studied the effect of germination on proximate, phenol and flavonoid content and antioxidant activities (AOA) of African Yam been (AYB). Germination was done in controlled dark chamber (100% RH, 28oC). The proximate, phenol and flavonoid content and antioxidant activities before and after germination were investigated. The crude protein, moisture, and crude fiber content of germinated AYB were significantly higher (P<0.05) than that of ungermianated seed, while the fat, Ash and carbohydrate content of ungerminated were higher than the germinated seed. Germination increased the phenol and flavoniod content by 19.14% and 14.53% respectively. The results of AOA assay showed that the DPPH, reducing power and FRAP of germinated AYB seed gave high values: 48.92 ±1.22 μg/ml, 0.75± 0.15μg/ml and 98.60±0.04 μmol/g while that of ungerminated seed were: 31.33μ/ml, 0.56±1.52μg/ml and 96.11±1.13μmol/g respectively. Germinated AYB has phytochemicals with potential AOA for disease prevention.

Keywords: Antioxidant, flovonoid, germination, phenol.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2357
8261 Automatic 3D Reconstruction of Coronary Artery Centerlines from Monoplane X-ray Angiogram Images

Authors: Ali Zifan, Panos Liatsis, Panagiotis Kantartzis, Manolis Gavaises, Nicos Karcanias, Demosthenes Katritsis

Abstract:

We present a new method for the fully automatic 3D reconstruction of the coronary artery centerlines, using two X-ray angiogram projection images from a single rotating monoplane acquisition system. During the first stage, the input images are smoothed using curve evolution techniques. Next, a simple yet efficient multiscale method, based on the information of the Hessian matrix, for the enhancement of the vascular structure is introduced. Hysteresis thresholding using different image quantiles, is used to threshold the arteries. This stage is followed by a thinning procedure to extract the centerlines. The resulting skeleton image is then pruned using morphological and pattern recognition techniques to remove non-vessel like structures. Finally, edge-based stereo correspondence is solved using a parallel evolutionary optimization method based on f symbiosis. The detected 2D centerlines combined with disparity map information allow the reconstruction of the 3D vessel centerlines. The proposed method has been evaluated on patient data sets for evaluation purposes.

Keywords: Vessel enhancement, centerline extraction, symbiotic reconstruction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2272
8260 A Comparative Study on the Performance of Viscous and Friction Dampers under Seismic Excitation

Authors: Apetsi K. Ampiah, Zhao Xin

Abstract:

Earthquakes over the years have been known to cause devastating damage on buildings and induced huge loss on human life and properties. It is for this reason that engineers have devised means of protecting buildings and thus protecting human life. Since the invention of devices such as the viscous and friction dampers, scientists/researchers have been able to incorporate these devices into buildings and other engineering structures. The viscous damper is a hydraulic device which dissipates the seismic forces by pushing fluid through an orifice, producing a damping pressure which creates a force. In the friction damper, the force is mainly resisted by converting the kinetic energy into heat by friction. Devices such as viscous and friction dampers are able to absorb almost all the earthquake energy, allowing the structure to remain undamaged (or with some amount of damage) and ready for immediate reuse (with some repair works). Comparing these two devices presents the engineer with adequate information on the merits and demerits of these devices and in which circumstances their use would be highly favorable. This paper examines the performance of both viscous and friction dampers under different ground motions. A two-storey frame installed with both devices under investigation are modeled in commercial computer software and analyzed under different ground motions. The results of the performance of the structure are then tabulated and compared. Also included in this study is the ease of installation and maintenance of these devices.

Keywords: Friction damper, seismic, slip load, viscous damper.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 714
8259 New Multisensor Data Fusion Method Based on Probabilistic Grids Representation

Authors: Zhichao Zhao, Yi Liu, Shunping Xiao

Abstract:

A new data fusion method called joint probability density matrix (JPDM) is proposed, which can associate and fuse measurements from spatially distributed heterogeneous sensors to identify the real target in a surveillance region. Using the probabilistic grids representation, we numerically combine the uncertainty regions of all the measurements in a general framework. The NP-hard multisensor data fusion problem has been converted to a peak picking problem in the grids map. Unlike most of the existing data fusion method, the JPDM method dose not need association processing, and will not lead to combinatorial explosion. Its convergence to the CRLB with a diminishing grid size has been proved. Simulation results are presented to illustrate the effectiveness of the proposed technique.

Keywords: Cramer-Rao lower bound (CRLB), data fusion, probabilistic grids, joint probability density matrix, localization, sensor network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1803
8258 Sampled-Data Model Predictive Tracking Control for Mobile Robot

Authors: Wookyong Kwon, Sangmoon Lee

Abstract:

In this paper, a sampled-data model predictive tracking control method is presented for mobile robots which is modeled as constrained continuous-time linear parameter varying (LPV) systems. The presented sampled-data predictive controller is designed by linear matrix inequality approach. Based on the input delay approach, a controller design condition is derived by constructing a new Lyapunov function. Finally, a numerical example is given to demonstrate the effectiveness of the presented method.

Keywords: Model predictive control, sampled-data control, linear parameter varying systems, LPV.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1278
8257 Talent Management through Integration of Talent Value Chain and Human Capital Analytics Approaches

Authors: Wuttigrai Ngamsirijit

Abstract:

Talent management in today’s modern organizations has become data-driven due to a demand for objective human resource decision making and development of analytics technologies. HR managers have been faced with some obstacles in exploiting data and information to obtain their effective talent management decisions. These include process-based data and records; insufficient human capital-related measures and metrics; lack of capabilities in data modeling in strategic manners; and, time consuming to add up numbers and make decisions. This paper proposes a framework of talent management through integration of talent value chain and human capital analytics approaches. It encompasses key data, measures, and metrics regarding strategic talent management decisions along the organizational and talent value chain. Moreover, specific predictive and prescriptive models incorporating these data and information are recommended to help managers in understanding the state of talent, gaps in managing talent and the organization, and the ways to develop optimized talent strategies.    

Keywords: Decision making, human capital analytics, talent management, talent value chain.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 968
8256 An Improved Adaptive Dot-Shape Beamforming Algorithm Research on Frequency Diverse Array

Authors: Yanping Liao, Zenan Wu, Ruigang Zhao

Abstract:

Frequency diverse array (FDA) beamforming is a technology developed in recent years, and its antenna pattern has a unique angle-distance-dependent characteristic. However, the beam is always required to have strong concentration, high resolution and low sidelobe level to form the point-to-point interference in the concentrated set. In order to eliminate the angle-distance coupling of the traditional FDA and to make the beam energy more concentrated, this paper adopts a multi-carrier FDA structure based on proposed power exponential frequency offset to improve the array structure and frequency offset of the traditional FDA. The simulation results show that the beam pattern of the array can form a dot-shape beam with more concentrated energy, and its resolution and sidelobe level performance are improved. However, the covariance matrix of the signal in the traditional adaptive beamforming algorithm is estimated by the finite-time snapshot data. When the number of snapshots is limited, the algorithm has an underestimation problem, which leads to the estimation error of the covariance matrix to cause beam distortion, so that the output pattern cannot form a dot-shape beam. And it also has main lobe deviation and high sidelobe level problems in the case of limited snapshot. Aiming at these problems, an adaptive beamforming technique based on exponential correction for multi-carrier FDA is proposed to improve beamforming robustness. The steps are as follows: first, the beamforming of the multi-carrier FDA is formed under linear constrained minimum variance (LCMV) criteria. Then the eigenvalue decomposition of the covariance matrix is ​​performed to obtain the diagonal matrix composed of the interference subspace, the noise subspace and the corresponding eigenvalues. Finally, the correction index is introduced to exponentially correct the small eigenvalues ​​of the noise subspace, improve the divergence of small eigenvalues ​​in the noise subspace, and improve the performance of beamforming. The theoretical analysis and simulation results show that the proposed algorithm can make the multi-carrier FDA form a dot-shape beam at limited snapshots, reduce the sidelobe level, improve the robustness of beamforming, and have better performance.

Keywords: Multi-carrier frequency diverse array, adaptive beamforming, correction index, limited snapshot, robust.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 677
8255 Association of G-174C Polymorphism of the Interleukin-6 Gene Promoter with Obesity in Iranian Population

Authors: Rostami F, Haj Hosseini R, Sharifi K, Daneshpour M, Azizi F, Hedayati M

Abstract:

Expression and secretion of inflammation markers are disturbed in obesity. Interleukin-6 reduces body fat mass. The common G-174C polymorphism in the promoter of IL-6 gene has been reported that effects on transcriptional regulation. The objective was to investigate association of the common polymorphism G-174C with obesity in Iranian population. The present study is cross sectional association study that included 242 individuals (110 men and 132 women). Serum IL-6 levels, C-reactive protein, fasting blood glucose and blood lipids profile were measured .BMI and WHR were calculated. Genotyping is carried out by PCR and RFLP. The frequencies of G and C allele were 64.5% and 35.5%, respectively. The G-174C polymorphism was not associated with BMI and WHR. However in obese individual, fasting blood glucose was significantly higher in carrier of C allele compared with the noncarrier. The IL-6 G-174C polymorphism is not a risk factor for obesity in Iranian population.

Keywords: Interleukin 6, Polymorphism genetic, Obesity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1610
8254 Adsorption of Lead(II) and Cadmium(II) Ions from Aqueous Solutions by Adsorption on Activated Carbon Prepared from Cashew Nut Shells

Authors: S. Tangjuank, N. Insuk , J. Tontrakoon , V. Udeye

Abstract:

Cashew nut shells were converted into activated carbon powders using KOH activation plus CO2 gasification at 1027 K. The increase both of impregnation ratio and activation time, there was swiftly the development of mesoporous structure with increasing of mesopore volume ratio from 20-28% and 27-45% for activated carbon with ratio of KOH per char equal to 1 and 4, respectively. Activated carbon derived from KOH/char ratio equal to 1 and CO2 gasification time from 20 to 150 minutes were exhibited the BET surface area increasing from 222 to 627 m2.g-1. And those were derived from KOH/char ratio of 4 with activation time from 20 to 150 minutes exhibited high BET surface area from 682 to 1026 m2.g-1. The adsorption of Lead(II) and Cadmium(II) ion was investigated. This adsorbent exhibited excellent adsorption for Lead(II) and Cadmium(II) ion. Maximum adsorption presented at 99.61% at pH 6.5 and 98.87% at optimum conditions. The experimental data was calculated from Freundlich isotherm and Langmuir isotherm model. The maximum capacity of Pb2+ and Cd2+ ions was found to be 28.90 m2.g-1 and 14.29 m2.g-1, respectively.

Keywords: Activated carbon, cashew nut shell, heavy metals, adsorption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3403
8253 Enhancing K-Means Algorithm with Initial Cluster Centers Derived from Data Partitioning along the Data Axis with the Highest Variance

Authors: S. Deelers, S. Auwatanamongkol

Abstract:

In this paper, we propose an algorithm to compute initial cluster centers for K-means clustering. Data in a cell is partitioned using a cutting plane that divides cell in two smaller cells. The plane is perpendicular to the data axis with the highest variance and is designed to reduce the sum squared errors of the two cells as much as possible, while at the same time keep the two cells far apart as possible. Cells are partitioned one at a time until the number of cells equals to the predefined number of clusters, K. The centers of the K cells become the initial cluster centers for K-means. The experimental results suggest that the proposed algorithm is effective, converge to better clustering results than those of the random initialization method. The research also indicated the proposed algorithm would greatly improve the likelihood of every cluster containing some data in it.

Keywords: Clustering algorithm, K-means algorithm, Datapartitioning, Initial cluster centers.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2866
8252 Semi-Supervised Outlier Detection Using a Generative and Adversary Framework

Authors: Jindong Gu, Matthias Schubert, Volker Tresp

Abstract:

In many outlier detection tasks, only training data belonging to one class, i.e., the positive class, is available. The task is then to predict a new data point as belonging either to the positive class or to the negative class, in which case the data point is considered an outlier. For this task, we propose a novel corrupted Generative Adversarial Network (CorGAN). In the adversarial process of training CorGAN, the Generator generates outlier samples for the negative class, and the Discriminator is trained to distinguish the positive training data from the generated negative data. The proposed framework is evaluated using an image dataset and a real-world network intrusion dataset. Our outlier-detection method achieves state-of-the-art performance on both tasks.

Keywords: Outlier detection, generative adversary networks, semi-supervised learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1074
8251 Methodology of the Turkey’s National Geographic Information System Integration Project

Authors: Buse A. Ataç, Doğan K. Cenan, Arda Çetinkaya, Naz D. Şahin, Köksal Sanlı, Zeynep Koç, Akın Kısa

Abstract:

With its spatial data reliability, interpretation and questioning capabilities, Geographical Information Systems make significant contributions to scientists, planners and practitioners. Geographic information systems have received great attention in today's digital world, growing rapidly, and increasing the efficiency of use. Access to and use of current and accurate geographical data, which are the most important components of the Geographical Information System, has become a necessity rather than a need for sustainable and economic development. This project aims to enable sharing of data collected by public institutions and organizations on a web-based platform. Within the scope of the project, INSPIRE (Infrastructure for Spatial Information in the European Community) data specifications are considered as a road-map. In this context, Turkey's National Geographic Information System (TUCBS) Integration Project supports sharing spatial data within 61 pilot public institutions as complied with defined national standards. In this paper, which is prepared by the project team members in the TUCBS Integration Project, the technical process with a detailed methodology is explained. In this context, the main technical processes of the Project consist of Geographic Data Analysis, Geographic Data Harmonization (Standardization), Web Service Creation (WMS, WFS) and Metadata Creation-Publication. In this paper, the integration process carried out to provide the data produced by 61 institutions to be shared from the National Geographic Data Portal (GEOPORTAL), have been trying to be conveyed with a detailed methodology.

Keywords: Data specification, geoportal, GIS, INSPIRE, TUCBS, Turkey’s National Geographic Information System.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 695
8250 Bioactive Component in Milk and Dairy Product

Authors: Bahareh Hajirostamloo

Abstract:

Recent research has shown that milk proteins can yield bioactive peptides with opioid, mineral binding, cytomodulatory, antihypertensive, immunostimulating, antimicrobial and antioxidative activity in the human body. Bioactive peptides are encrypted in milk proteins and are only released by enzymatic hydrolysis in vivo during gastrointestinal digestion, food processing or by microbial enzymes in fermented products. At present significant research is being undertaken on the health effects of bioactive peptides. A variety of naturally formed bioactive peptides have been found in fermented dairy products, such as yoghurt, sour milk and cheese. In particular, antihypertensive peptides have been identified in fermented milks, whey and ripened cheese. Some of these peptides have been commercialized in the form of fermented milks. Bioactive peptides have the potential to be used in the formulation of health-enhancing nutraceuticals, and as potent drugs with well defined pharmacological effects.

Keywords: Milk protein, Bioactive peptides, Health effects, Dairy product.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3988
8249 The Extension of Monomeric Computational Results to Polymeric Measurable Properties: An Introductory Computational Chemistry Experiment

Authors: Zhao Jing, Bai Yongqing, Shi Qiaofang, Zang Yang, Zhang Huaihao

Abstract:

Advances in software technology enable the computational chemistry to be commonly applied in various research fields, especially in pedagogy. Thus, in order to expand and improve experimental instructions of computational chemistry for undergraduates, we designed an introductory experiment—research on acrylamide molecular structure and physicochemical properties. Initially, students construct molecular models of acrylamide and polyacrylamide in Gaussian and Materials Studio software respectively. Then, the infrared spectral data, atomic charge and molecular orbitals of acrylamide as well as solvation effect of polyacrylamide are calculated to predict their physicochemical performance. At last, rheological experiments are used to validate these predictions. Through the combination of molecular simulation (performed on Gaussian, Materials Studio) with experimental verification (rheology experiment), learners have deeply comprehended the chemical nature of acrylamide and polyacrylamide, achieving good learning outcomes.

Keywords: Upper-division undergraduate, computer-based learning, laboratory instruction, amides, molecular modeling, spectroscopy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 370
8248 Exploring SSD Suitable Allocation Schemes Incompliance with Workload Patterns

Authors: Jae Young Park, Hwansu Jung, Jong Tae Kim

Abstract:

In the Solid-State-Drive (SSD) performance, whether the data has been well parallelized is an important factor. SSD parallelization is affected by allocation scheme and it is directly connected to SSD performance. There are dynamic allocation and static allocation in representative allocation schemes. Dynamic allocation is more adaptive in exploiting write operation parallelism, while static allocation is better in read operation parallelism. Therefore, it is hard to select the appropriate allocation scheme when the workload is mixed read and write operations. We simulated conditions on a few mixed data patterns and analyzed the results to help the right choice for better performance. As the results, if data arrival interval is long enough prior operations to be finished and continuous read intensive data environment static allocation is more suitable. Dynamic allocation performs the best on write performance and random data patterns.

Keywords: Dynamic allocation, NAND Flash based SSD, SSD parallelism, static allocation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1995
8247 Fused Structure and Texture (FST) Features for Improved Pedestrian Detection

Authors: Hussin K. Ragb, Vijayan K. Asari

Abstract:

In this paper, we present a pedestrian detection descriptor called Fused Structure and Texture (FST) features based on the combination of the local phase information with the texture features. Since the phase of the signal conveys more structural information than the magnitude, the phase congruency concept is used to capture the structural features. On the other hand, the Center-Symmetric Local Binary Pattern (CSLBP) approach is used to capture the texture information of the image. The dimension less quantity of the phase congruency and the robustness of the CSLBP operator on the flat images, as well as the blur and illumination changes, lead the proposed descriptor to be more robust and less sensitive to the light variations. The proposed descriptor can be formed by extracting the phase congruency and the CSLBP values of each pixel of the image with respect to its neighborhood. The histogram of the oriented phase and the histogram of the CSLBP values for the local regions in the image are computed and concatenated to construct the FST descriptor. Several experiments were conducted on INRIA and the low resolution DaimlerChrysler datasets to evaluate the detection performance of the pedestrian detection system that is based on the FST descriptor. A linear Support Vector Machine (SVM) is used to train the pedestrian classifier. These experiments showed that the proposed FST descriptor has better detection performance over a set of state of the art feature extraction methodologies.

Keywords: Pedestrian detection, phase congruency, local phase, LBP features, CSLBP features, FST descriptor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1489
8246 Seismic Base Shear Force Depending on Building Fundamental Period and Site Conditions: Deterministic Formulation and Probabilistic Analysis

Authors: S. Dorbani, M. Badaoui, D. Benouar

Abstract:

The aim of this paper is to investigate the effect of the building fundamental period of reinforced concrete buildings of (6, 9, and 12-storey), with different floor plans: Symmetric, mono-symmetric, and unsymmetric. These structures are erected at different epicentral distances. Using the Boumerdes, Algeria (2003) earthquake data, we focused primarily on the establishment of the deterministic formulation linking the base shear force to two parameters: The first one is the fundamental period that represents the numerical fingerprint of the structure, and the second one is the epicentral distance used to represent the impact of the earthquake on this force. In a second step, with a view to highlight the effect of uncertainty in these parameters on the analyzed response, these parameters are modeled as random variables with a log-normal distribution. The variability of the coefficients of variation of the chosen uncertain parameters, on the statistics on the seismic base shear force, showed that the effect of uncertainty on fundamental period on this force statistics is low compared to the epicentral distance uncertainty influence.

Keywords: Base shear force, fundamental period, epicentral distance, uncertainty, lognormal variable, statistics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1302
8245 Adaptive Kernel Principal Analysis for Online Feature Extraction

Authors: Mingtao Ding, Zheng Tian, Haixia Xu

Abstract:

The batch nature limits the standard kernel principal component analysis (KPCA) methods in numerous applications, especially for dynamic or large-scale data. In this paper, an efficient adaptive approach is presented for online extraction of the kernel principal components (KPC). The contribution of this paper may be divided into two parts. First, kernel covariance matrix is correctly updated to adapt to the changing characteristics of data. Second, KPC are recursively formulated to overcome the batch nature of standard KPCA.This formulation is derived from the recursive eigen-decomposition of kernel covariance matrix and indicates the KPC variation caused by the new data. The proposed method not only alleviates sub-optimality of the KPCA method for non-stationary data, but also maintains constant update speed and memory usage as the data-size increases. Experiments for simulation data and real applications demonstrate that our approach yields improvements in terms of both computational speed and approximation accuracy.

Keywords: adaptive method, kernel principal component analysis, online extraction, recursive algorithm

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1552
8244 Effects of Chlorhexidine in Application to Hybrid Layers

Authors: Ilma Robo, Saimir Heta, Edona Hasanaj, Vera Ostreni

Abstract:

The hybrid layer (HL), the way it is created and how it is protected against degradation over time, is the key to the clinical success of a composite restoration. The composite supports the dentinal structure exactly with the realized surface of micro-retention. Thus, this surface is in direct proportion to its size versus the duration of clinical use of composite dental restoration. Micro-retention occurs between dentin or acidified enamel and adhesive resin extensions versus pre-prepared spaces, such as hollow dentinal tubules. The way the adhesive resin binds to the acidified dentinal structure depends on the physical or chemical factors of this interrelationship between two structures with very different characteristics. During the acidification process, a precursor to the placement of the adhesive resin layer, activation of metalloproteinases of dental origin occurs, enzymes which are responsible for the degradation of the HL. These enzymes have expressed activity depending on the presence of Zn2+ or Ca2+ ions. There are several ways to inhibit these enzymes, and consequently, there are several ways to inhibit the degradation process of the HL. The study aim is to evaluate chlorhexidine (CHX) as a solution element, inhibitor of dentin activated metalloproteinases, as a result of the application of acidification. This study aims to look at this solution in advantage or contraindication theories, already published in the literature.

Keywords: Hybrid layer, chlorhexidine, degradation, smear layer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 349
8243 Proposing an Efficient Method for Frequent Pattern Mining

Authors: Vaibhav Kant Singh, Vijay Shah, Yogendra Kumar Jain, Anupam Shukla, A.S. Thoke, Vinay KumarSingh, Chhaya Dule, Vivek Parganiha

Abstract:

Data mining, which is the exploration of knowledge from the large set of data, generated as a result of the various data processing activities. Frequent Pattern Mining is a very important task in data mining. The previous approaches applied to generate frequent set generally adopt candidate generation and pruning techniques for the satisfaction of the desired objective. This paper shows how the different approaches achieve the objective of frequent mining along with the complexities required to perform the job. This paper will also look for hardware approach of cache coherence to improve efficiency of the above process. The process of data mining is helpful in generation of support systems that can help in Management, Bioinformatics, Biotechnology, Medical Science, Statistics, Mathematics, Banking, Networking and other Computer related applications. This paper proposes the use of both upward and downward closure property for the extraction of frequent item sets which reduces the total number of scans required for the generation of Candidate Sets.

Keywords: Data Mining, Candidate Sets, Frequent Item set, Pruning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1684
8242 One-DOF Precision Position Control using the Combined Piezo-VCM Actuator

Authors: Yung-Tien Liu, Chun-Chao Wang

Abstract:

This paper presents the control performance of a high-precision positioning device using the hybrid actuator composed of a piezoelectric (PZT) actuator and a voice-coil motor (VCM). The combined piezo-VCM actuator features two main characteristics: a large operation range due to long stroke of the VCM, and high precision and heavy load positioning ability due to PZT impact force. A one-degree-of-freedom (DOF) experimental setup was configured to examine the fundamental characteristics, and the control performance was effectively demonstrated by using a switching controller. In rough positioning state, an integral variable structure controller (IVSC) was used for the VCM to conduct long range of operation; in precision positioning state, an impact force controller (IFC) for the PZT actuator coupled with presliding states of the sliding table was used to obtain high-precision position control and achieve both forward and backward actuations. The experimental results showed that the sliding table having a mass of 881g and with a preload of 10 N was successfully positioned within the positioning accuracy of 10 nm in both forward and backward position controls.

Keywords: Integral variable structure controller (IVSC), impact force, precision positioning, presliding, PZT actuator, voice-coil motor (VCM).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1937
8241 In silico Analysis of Isoniazid Resistance in Mycobacterium tuberculosis

Authors: A. Nusrath Unissa, Sameer Hassan, Luke Elizabeth Hanna

Abstract:

Altered drug binding may be an important factor in isoniazid (INH) resistance, rather than major changes in the enzyme’s activity as a catalase or peroxidase (KatG). The identification of structural or functional defects in the mutant KatGs responsible for INH resistance remains as an area to be explored. In this connection, the differences in the binding affinity between wild-type (WT) and mutants of KatG were investigated, through the generation of three mutants of KatG, Ser315Thr [S315T], Ser315Asn [S315N], Ser315Arg [S315R] and a WT [S315]) with the help of software-MODELLER. The mutants were docked with INH using the software-GOLD. The affinity is lower for WT than mutant, suggesting the tight binding of INH with the mutant protein compared to WT type. These models provide the in silico evidence for the binding interaction of KatG with INH and implicate the basis for rationalization of INH resistance in naturally occurring KatG mutant strains of Mycobacterium tuberculosis.

Keywords: Mycobacterium tuberculosis, KatG, INH resistance, Mutants, Modeling, Docking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2990
8240 Danger Theory and Intelligent Data Processing

Authors: Anjum Iqbal, Mohd Aizaini Maarof

Abstract:

Artificial Immune System (AIS) is relatively naive paradigm for intelligent computations. The inspiration for AIS is derived from natural Immune System (IS). Classically it is believed that IS strives to discriminate between self and non-self. Most of the existing AIS research is based on this approach. Danger Theory (DT) argues this approach and proposes that IS fights against danger producing elements and tolerates others. We, the computational researchers, are not concerned with the arguments among immunologists but try to extract from it novel abstractions for intelligent computation. This paper aims to follow DT inspiration for intelligent data processing. The approach may introduce new avenue in intelligent processing. The data used is system calls data that is potentially significant in intrusion detection applications.

Keywords: artificial immune system, danger theory, intelligent processing, system calls

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1883
8239 Gas-Liquid Flow on Smooth and Textured Inclined Planes

Authors: J.J. Cooke, S. Gu, L.M. Armstrong, K.H. Luo

Abstract:

Carbon Capture & Storage (CCS) is one of the various methods that can be used to reduce the carbon footprint of the energy sector. This paper focuses on the absorption of CO2 from flue gas using packed columns, whose efficiency is highly dependent on the structure of the liquid films within the column. To study the characteristics of liquid films a CFD solver, OpenFOAM is utilised to solve two-phase, isothermal film flow using the volume-of-fluid (VOF) method. The model was validated using existing experimental data and the Nusselt theory. It was found that smaller plate inclination angles, with respect to the horizontal plane, resulted in larger wetted areas on smooth plates. However, only a slight improvement in the wetted area was observed. Simulations were also performed using a ridged plate and it was observed that these surface textures significantly increase the wetted area of the plate. This was mainly attributed to the channelling effect of the ridges, which helped to oppose the surface tension forces trying to minimise the surface area. Rivulet formations on the ridged plate were also flattened out and spread across a larger proportion of the plate width.

Keywords: CCS, liquid film flow, packed columns, wetted area

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2102