Search results for: fingerprint recognition multiple polynomials.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2361

Search results for: fingerprint recognition multiple polynomials.

381 Satellite Imagery Classification Based on Deep Convolution Network

Authors: Zhong Ma, Zhuping Wang, Congxin Liu, Xiangzeng Liu

Abstract:

Satellite imagery classification is a challenging problem with many practical applications. In this paper, we designed a deep convolution neural network (DCNN) to classify the satellite imagery. The contributions of this paper are twofold — First, to cope with the large-scale variance in the satellite image, we introduced the inception module, which has multiple filters with different size at the same level, as the building block to build our DCNN model. Second, we proposed a genetic algorithm based method to efficiently search the best hyper-parameters of the DCNN in a large search space. The proposed method is evaluated on the benchmark database. The results of the proposed hyper-parameters search method show it will guide the search towards better regions of the parameter space. Based on the found hyper-parameters, we built our DCNN models, and evaluated its performance on satellite imagery classification, the results show the classification accuracy of proposed models outperform the state of the art method.

Keywords: Satellite imagery classification, deep convolution network, genetic algorithm, hyper-parameter optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2299
380 Incorporating Multiple Supervised Learning Algorithms for Effective Intrusion Detection

Authors: Umar Albalawi, Sang C. Suh, Jinoh Kim

Abstract:

As internet continues to expand its usage with an  enormous number of applications, cyber-threats have significantly  increased accordingly. Thus, accurate detection of malicious traffic in  a timely manner is a critical concern in today’s Internet for security.  One approach for intrusion detection is to use Machine Learning (ML)  techniques. Several methods based on ML algorithms have been  introduced over the past years, but they are largely limited in terms of  detection accuracy and/or time and space complexity to run. In this  work, we present a novel method for intrusion detection that  incorporates a set of supervised learning algorithms. The proposed  technique provides high accuracy and outperforms existing techniques  that simply utilizes a single learning method. In addition, our  technique relies on partial flow information (rather than full  information) for detection, and thus, it is light-weight and desirable for  online operations with the property of early identification. With the  mid-Atlantic CCDC intrusion dataset publicly available, we show that  our proposed technique yields a high degree of detection rate over 99%  with a very low false alarm rate (0.4%). 

 

Keywords: Intrusion Detection, Supervised Learning, Traffic Classification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1999
379 Performance Evaluation of Para-virtualization on Modern Mobile Phone Platform

Authors: Yang Xu, Felix Bruns, Elizabeth Gonzalez, Shadi Traboulsi, Klaus Mott, Attila Bilgic

Abstract:

Emergence of smartphones brings to live the concept of converged devices with the availability of web amenities. Such trend also challenges the mobile devices manufactures and service providers in many aspects, such as security on mobile phones, complex and long time design flow, as well as higher development cost. Among these aspects, security on mobile phones is getting more and more attention. Microkernel based virtualization technology will play a critical role in addressing these challenges and meeting mobile market needs and preferences, since virtualization provides essential isolation for security reasons and it allows multiple operating systems to run on one processor accelerating development and cutting development cost. However, virtualization benefits do not come for free. As an additional software layer, it adds some inevitable virtualization overhead to the system, which may decrease the system performance. In this paper we evaluate and analyze the virtualization performance cost of L4 microkernel based virtualization on a competitive mobile phone by comparing the L4Linux, a para-virtualized Linux on top of L4 microkernel, with the native Linux performance using lmbench and a set of typical mobile phone applications.

Keywords: L4 microkernel, virtualization overhead, mobilephone.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1942
378 Online Partial Discharge Source Localization and Characterization Using Non-Conventional Method

Authors: Ammar Anwar Khan, Nissar R. Wani, Nazar Malik, Abdulrehman Al-Arainy, and Saad Alghuwainem

Abstract:

Power cables are vulnerable to failure due to aging or defects that occur with the passage of time under continuous operation and loading stresses. PD detection and characterization provide information on the location, nature, form and extent of the degradation. As a result, PD monitoring has become an important part of condition based maintenance (CBM) program among power utilities. Online partial discharge (PD) localization of defect sources in power cable system is possible using the time of flight method. The information regarding the time difference between the main and reflected pulses and cable length can help in locating the partial discharge source along the cable length. However, if the length of the cable is not known and the defect source is located at the extreme ends of the cable or in the middle of the cable, then double ended measurement is required to indicate the location of PD source. Use of multiple sensors can also help in discriminating the cable PD or local/ external PD. This paper presents the experience and results from online partial discharge measurements conducted in the laboratory and the challenges in partial discharge source localization.

Keywords: Power cables, partial discharge localization, HFCT, condition based monitoring.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2792
377 The Integrated Urban Strategies Based on Deep Urban History and Modern Technology Study: Tourism and Leisure Industries as Driving Force to Reactivate Historical Area

Authors: Cheng Li, Jie Shen, Yutian Tang

Abstract:

Embracing the upcoming era of urbanization with the challenges of limitation of resources, disappearing cultural identities and conflicts among different groups of stakeholders, new integrated approaches are offered in our urban practice to help decision-makers and stakeholders frame and develop well-conceived, practical strategies for urban developing trajectories to approach urban-level sustainability in multiple social, cultural, ecological dimensions. Through bottom-up participation, we take advantage of tourism and leisure industries as driving forces for urbanization in China to promote integrated sustainable systems, with the hope of approaching both historical and ecological aspects of urban sustainability; and also thanks to top-down participation, we have codes, standards and rules established by the governments to strengthen the implementation of ecological urban sustainability. The results are monitored and evaluated experimentally and multidimensionally and the sustainable systems we constructed with local stakeholder groups turned out to be effective. The presentation of our selected projects would indicate our different focuses on urban sustainability.

Keywords: Urban sustainability, integrated urban strategy, tourism and leisure industries, history, modern technology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1546
376 Systematic Identification and Quantification of Substrate Specificity Determinants in Human Protein Kinases

Authors: Manuel A. Alonso-Tarajano, Roberto Mosca, Patrick Aloy

Abstract:

Protein kinases participate in a myriad of cellular processes of major biomedical interest. The in vivo substrate specificity of these enzymes is a process determined by several factors, and despite several years of research on the topic, is still far from being totally understood. In the present work, we have quantified the contributions to the kinase substrate specificity of i) the phosphorylation sites and their surrounding residues in the sequence and of ii) the association of kinases to adaptor or scaffold proteins. We have used position-specific scoring matrices (PSSMs), to represent the stretches of sequences phosphorylated by 93 families of kinases. We have found negative correlations between the number of sequences from which a PSSM is generated and the statistical significance and the performance of that PSSM. Using a subset of 22 statistically significant PSSMs, we have identified specificity determinant residues (SDRs) for 86% of the corresponding kinase families. Our results suggest that different SDRs can function as positive or negative elements of substrate recognition by the different families of kinases. Additionally, we have found that human proteins with known function as adaptors or scaffolds (kAS) tend to interact with a significantly large fraction of the substrates of the kinases to which they associate. Based on this characteristic we have identified a set of 279 potential adaptors/scaffolds (pAS) for human kinases, which is enriched in Pfam domains and functional terms tightly related to the proposed function. Moreover, our results show that for 74.6% of the kinase–pAS association found, the pAS colocalize with the substrates of the kinases they are associated to. Finally, we have found evidence suggesting that the association of kinases to adaptors and scaffolds, may contribute significantly to diminish the in vivo substrate crossed-specificity of protein kinases. In general, our results indicate the relevance of several SDRs for both the positive and negative selection of phosphorylation sites by kinase families and also suggest that the association of kinases to pAS proteins may be an important factor for the localization of the enzymes with their set of substrates.

Keywords: Kinase, phosphorylation, substrate specificity, adaptors, scaffolds, cellular colocalization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1503
375 Aspect Oriented Software Architecture

Authors: Pradip Peter Dey, Ronald F. Gonzales, Gordon W. Romney, Mohammad Amin, Bhaskar Raj Sinha

Abstract:

Natural language processing systems pose a unique challenge for software architectural design as system complexity has increased continually and systems cannot be easily constructed from loosely coupled modules. Lexical, syntactic, semantic, and pragmatic aspects of linguistic information are tightly coupled in a manner that requires separation of concerns in a special way in design, implementation and maintenance. An aspect oriented software architecture is proposed in this paper after critically reviewing relevant architectural issues. For the purpose of this paper, the syntactic aspect is characterized by an augmented context-free grammar. The semantic aspect is composed of multiple perspectives including denotational, operational, axiomatic and case frame approaches. Case frame semantics matured in India from deep thematic analysis. It is argued that lexical, syntactic, semantic and pragmatic aspects work together in a mutually dependent way and their synergy is best represented in the aspect oriented approach. The software architecture is presented with an augmented Unified Modeling Language.

Keywords: Language engineering, parsing, software design, user experience.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1708
374 Testing the Validity of Maturity Model for E-Government Implementation in Indonesia

Authors: Darmawan Napitupulu, Dana Indra Sensuse, Aniati Murni

Abstract:

The research was conducted to empirically validate the proposed maturity model of e-Government implementation, composed of four dimensions, further specified by 54 success factors as attributes. To do so, there are two steps were performed. First, expert’s judgment was conducted to test its content validity. The second, reliability study was performed to evaluate inter-rater agreement by using Fleiss Kappa approach. The kappa statistic (kappa coefficient) is the most commonly used method for testing the consistency among raters. Fleiss Kappa was a generalization of Kappa in extensions to the case of more than two raters (multiple raters) with multi-categorical ratings. Our findings show that most attributes of the proposed model were related to their corresponding dimensions. According to our results, The percentage of agree answers given by the experts was 73.69% in dimension A, 89.76% in B, 81.5% in C and 60.37% in D. This means that more than half of the attributes of each dimensions were appropriate or relevant to the dimensions they were supposed to measure, while 85% of attributes were relevant enough to their corresponding dimensions. Inter-rater reliability coefficient also showed satisfactory result and interpreted as substantial agreement among raters. Therefore, the proposed model in this paper was valid and reliable to measure the maturity of e-Government implementation.

Keywords: E-Government, Model, Maturity, Validity, Reliability Kappa.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2170
373 A Propose of Personnel Assessment Method Including a Two-Way Assessment for Evaluating Evaluators and Employees

Authors: Shunsuke Saito, Kazuho Yoshimoto, Shunichi Ohmori, Sirawadee Arunyanart

Abstract:

In this paper, we suggest a mechanism of assessment that rater and Ratee (or employees) to convince. There are many problems exist in the personnel assessment. In particular, we were focusing on the three. (1) Raters are not sufficiently recognized assessment point. (2) Ratee are not convinced by the mechanism of assessment. (3) Raters (or Evaluators) and ratees have empathy. We suggest 1: Setting of "understanding of the assessment points." 2: Setting of "relative assessment ability." 3: Proposal of two-way assessment mechanism to solve these problems. As a prerequisite, it is assumed that there are multiple raters. This is because has been a growing importance of multi-faceted assessment. In this model, it determines the weight of each assessment point evaluators by the degree of understanding and assessment ability of raters and ratee. We used the ANP (Analytic Network Process) is a theory that an extension of the decision-making technique AHP (Analytic Hierarchy Process). ANP can be to address the problem of forming a network and assessment of Two-Way is possible. We apply this technique personnel assessment, the weights of rater of each point can be reasonably determined. We suggest absolute assessment for Two-Way assessment by ANP. We have verified that the consent of the two approaches is higher than conventional mechanism. Also, human resources consultant we got a comment about the application of the practice.

Keywords: Personnel assessment, ANP (analytic network process), two-way.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 752
372 The Determination of Aflatoxins in Paddy and Milled Fractions of Rice in Guyana: Preliminary Results

Authors: Donna M. Morrison, Lambert Chester, Coretta A. N. Samuels, David R. Ledoux

Abstract:

A survey was conducted in the five rice-growing regions in Guyana to determine the presence of aflatoxins in multiple fractions of rice in June/October 2015 growing season. The fractions were paddy, steamed paddy, cargo rice, white rice and parboiled rice. Samples were analyzed by High Performance Liquid Chromatography. A subset of the samples was further analyzed by enzyme-linked immunosorbent assay (ELISA) for concurrence. All analyses were conducted at the University of Missouri, USA. Of the 186 samples tested, 16 had aflatoxin concentrations greater than 20 ppb the recommended limit for aflatoxins in food according to the United States Food and Drug Administration. An additional three samples had aflatoxin B1 concentrations greater than the European Union Commission maximum levels for aflatoxin B1 in rice at 5 µg/kg and total aflatoxins (B1, B2, G1 and G2) at 10 µg/kg. The survey indicates that there is no widespread aflatoxin problem in rice in Guyana. The incidence of aflatoxins appears to be localized.

Keywords: Aflatoxins, enzyme-linked immunosorbent assay, high-performance liquid chromatography, rice fractions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1534
371 Nonlinear Impact Responses for a Damped Frame Supported by Nonlinear Springs with Hysteresis Using Fast FEA

Authors: T. Yamaguchi, M. Watanabe, M. Sasajima, C. Yuan, S. Maruyama, T. B. Ibrahim, H. Tomita

Abstract:

This paper deals with nonlinear vibration analysis using finite element method for frame structures consisting of elastic and viscoelastic damping layers supported by multiple nonlinear concentrated springs with hysteresis damping. The frame is supported by four nonlinear concentrated springs near the four corners. The restoring forces of the springs have cubic non-linearity and linear component of the nonlinear springs has complex quantity to represent linear hysteresis damping. The damping layer of the frame structures has complex modulus of elasticity. Further, the discretized equations in physical coordinate are transformed into the nonlinear ordinary coupled differential equations using normal coordinate corresponding to linear natural modes. Comparing shares of strain energy of the elastic frame, the damping layer and the springs, we evaluate the influences of the damping couplings on the linear and nonlinear impact responses. We also investigate influences of damping changed by stiffness of the elastic frame on the nonlinear coupling in the damped impact responses.

Keywords: Dynamic response, Nonlinear impact response, Finite Element analysis, Numerical analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1687
370 Impact of Out-of-Plane Stiffness of the Diaphragm on Deflection of Wood Light-Frame Shear Walls

Authors: M. M. Bagheri, G. Doudak, M. Gong

Abstract:

The in-plane rigidity of light frame diaphragms has been investigated by researchers due to the importance of this subsystem regarding lateral force distribution between the lateral force resisting system (LFRS). Where research has lacked is in evaluating the impact of out-of-plane raigidity of the diaphragm on the deflection of shear walls. This study aims at investigating the effect of the diaphragm on the behavior of wood light-frame shear walls, in particular its out-of-plane rigidity was simulated by modeling the floors as beam. The out of plane stiffness of the diaphragm was investigated for idealized (infinitely stiff or flexible) as well as “realistic”. The results showed reductions in the shear wall deflection in the magnitude of approximately 80% considering the out of plane rigidity of the diaphragm. It was also concluded that considering conservative estimates of out-of-plane stiffness might lead to a very significant reduction in deflection and that assuming the floor diaphragm to be infinitely rigid out of plan seems to be reasonable. For diaphragms supported on multiple panels, further reduction in the deflection was observed. More work, particularly at the experimental level, is needed to verify the finding obtained in the numerical investigation related to the effect of out of plane diaphragm stiffness.

Keywords: Deflection of light-frame wood shear walls, out-of-plane stiffness of the diaphragm, initial stiffness.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 731
369 Corporate Governance and Corporate Social Responsibility: Research on the Interconnection of Both Concepts and Its Impact on Non-Profit Organizations

Authors: Helene Eller

Abstract:

The aim of non-profit organizations (NPO) is to provide services and goods for its clientele, with profit being a minor objective. By having this definition as the basic purpose of doing business, it is obvious that the goal of an organisation is to serve several bottom lines and not only the financial one. This approach is underpinned by the non-distribution constraint which means that NPO are allowed to make profits to a certain extent, but not to distribute them. The advantage is that there are no single shareholders who might have an interest in the prosperity of the organisation: there is no pie to divide. The gained profits remain within the organisation and will be reinvested in purposeful projects. Good governance is mandatory to support the aim of NPOs. Looking for a measure of good governance the principals of corporate governance (CG) will come in mind. The purpose of CG is direction and control, and in the field of NPO, CG is enlarged to consider the relationship to all important stakeholders who have an impact on the organisation. The recognition of more relevant parties than the shareholder is the link to corporate social responsibility (CSR). It supports a broader view of the bottom line: It is no longer enough to know how profits are used but rather how they are made. Besides, CSR addresses the responsibility of organisations for their impact on society. When transferring the concept of CSR to the non-profit area it will become obvious that CSR with its distinctive features will match the aims of NPOs. As a consequence, NPOs who apply CG apply also CSR to a certain extent. The research is designed as a comprehensive theoretical and empirical analysis. First, the investigation focuses on the theoretical basis of both concepts. Second, the similarities and differences are outlined and as a result the interconnection of both concepts will show up. The contribution of this research is manifold: The interconnection of both concepts when applied to NPOs has not got any attention in science yet. CSR and governance as integrated concept provides a lot of advantages for NPOs compared to for-profit organisations which are in a steady justification to show the impact they might have on the society. NPOs, however, integrate economic and social aspects as starting point. For NPOs CG is not a mere concept of compliance but rather an enhanced concept integrating a lot of aspects of CSR. There is no “either-nor” between the concepts for NPOs.

Keywords: Business ethics, corporate governance, corporate social responsibility, non-profit organisations, stakeholder theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1913
368 A Cuckoo Search with Differential Evolution for Clustering Microarray Gene Expression Data

Authors: M. Pandi, K. Premalatha

Abstract:

A DNA microarray technology is a collection of microscopic DNA spots attached to a solid surface. Scientists use DNA microarrays to measure the expression levels of large numbers of genes simultaneously or to genotype multiple regions of a genome. Elucidating the patterns hidden in gene expression data offers a tremendous opportunity for an enhanced understanding of functional genomics. However, the large number of genes and the complexity of biological networks greatly increase the challenges of comprehending and interpreting the resulting mass of data, which often consists of millions of measurements. It is handled by clustering which reveals the natural structures and identifying the interesting patterns in the underlying data. In this paper, gene based clustering in gene expression data is proposed using Cuckoo Search with Differential Evolution (CS-DE). The experiment results are analyzed with gene expression benchmark datasets. The results show that CS-DE outperforms CS in benchmark datasets. To find the validation of the clustering results, this work is tested with one internal and one external cluster validation indexes.

Keywords: DNA, Microarray, genomics, Cuckoo Search, Differential Evolution, Gene expression data, Clustering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1448
367 Enhancing the Effectiveness of Air Defense Systems through Simulation Analysis

Authors: F. Felipe

Abstract:

Air Defense Systems contain high-value assets that are expected to fulfill their mission for several years - in many cases, even decades - while operating in a fast-changing, technology-driven environment. Thus, it is paramount that decision-makers can assess how effective an Air Defense System is in the face of new developing threats, as well as to identify the bottlenecks that could jeopardize the security of the airspace of a country. Given the broad extent of activities and the great variety of assets necessary to achieve the strategic objectives, a systems approach was taken in order to delineate the core requirements and the physical architecture of an Air Defense System. Then, value-focused thinking helped in the definition of the measures of effectiveness. Furthermore, analytical methods were applied to create a formal structure that preliminarily assesses such measures. To validate the proposed methodology, a powerful simulation was also used to determine the measures of effectiveness, now in more complex environments that incorporate both uncertainty and multiple interactions of the entities. The results regarding the validity of this methodology suggest that the approach can support decisions aimed at enhancing the capabilities of Air Defense Systems. In conclusion, this paper sheds some light on how consolidated approaches of Systems Engineering and Operations Research can be used as valid techniques for solving problems regarding a complex and yet vital matter.

Keywords: Air defense, effectiveness, system, simulation, decision-support.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 357
366 Coding based Synchronization Algorithm for Secondary Synchronization Channel in WCDMA

Authors: Deng Liao, Dongyu Qiu, Ahmed K. Elhakeem

Abstract:

A new code synchronization algorithm is proposed in this paper for the secondary cell-search stage in wideband CDMA systems. Rather than using the Cyclically Permutable (CP) code in the Secondary Synchronization Channel (S-SCH) to simultaneously determine the frame boundary and scrambling code group, the new synchronization algorithm implements the same function with less system complexity and less Mean Acquisition Time (MAT). The Secondary Synchronization Code (SSC) is redesigned by splitting into two sub-sequences. We treat the information of scrambling code group as data bits and use simple time diversity BCH coding for further reliability. It avoids involved and time-costly Reed-Solomon (RS) code computations and comparisons. Analysis and simulation results show that the Synchronization Error Rate (SER) yielded by the new algorithm in Rayleigh fading channels is close to that of the conventional algorithm in the standard. This new synchronization algorithm reduces system complexities, shortens the average cell-search time and can be implemented in the slot-based cell-search pipeline. By taking antenna diversity and pipelining correlation processes, the new algorithm also shows its flexible application in multiple antenna systems.

Keywords: WCDMA cell-search, synchronization algorithm, secondary synchronization channel, antenna diversity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2351
365 Educational Data Mining: The Case of Department of Mathematics and Computing in the Period 2009-2018

Authors: M. Sitoe, O. Zacarias

Abstract:

University education is influenced by several factors that range from the adoption of strategies to strengthen the whole process to the academic performance improvement of the students themselves. This work uses data mining techniques to develop a predictive model to identify students with a tendency to evasion and retention. To this end, a database of real students’ data from the Department of University Admission (DAU) and the Department of Mathematics and Informatics (DMI) was used. The data comprised 388 undergraduate students admitted in the years 2009 to 2014. The Weka tool was used for model building, using three different techniques, namely: K-nearest neighbor, random forest, and logistic regression. To allow for training on multiple train-test splits, a cross-validation approach was employed with a varying number of folds. To reduce bias variance and improve the performance of the models, ensemble methods of Bagging and Stacking were used. After comparing the results obtained by the three classifiers, Logistic Regression using Bagging with seven folds obtained the best performance, showing results above 90% in all evaluated metrics: accuracy, rate of true positives, and precision. Retention is the most common tendency.

Keywords: Evasion and retention, cross validation, bagging, stacking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 51
364 Effects of a Methanol Fraction of the Leaves of Leonotis leonurus on the Blood Pressure and Heart Rate of Normotensive Male Wistar Rats

Authors: K. Obikeze, P. Mugabo, I. Green, D. Dietrich, A. Burger

Abstract:

Leonotisleonurus a shrub indigenous to Southern Africa is widely used in traditional medicine to treat a variety of conditions ranging from skin diseases and cough to epileptic fits and ‘heart problems’. Studies on the aqueous extract of the leaves have indicated cycloxegenase enzyme inhibitory activity and an antihypertensive effect. Five methanol leaf extract fractions (MLEa - MLEe) of L. leonurus were tested on anaesthetized normotensive male Wistar rats (AWR) and isolated perfused working rat hearts (IWH). Fraction MLEc (0.01mg/kg – 0.05mg/kg) induced significant increases in BP and HR in AWR and positive chronotropic and inotropic effects in IWH (1.0mg/ml – 5.0mg/ml). Pre-administration of atenolol (2.0mg/kg) and prazosin (60μg/kg) significantly inhibited MLEc effect on HR and MAP respectively in vivo, while atenolol (7.0mg/ml) pre-perfusion significantly inhibited MLEc effect in vitro. The hypertensive effect of MLEc is probably via β1agonism. Results also indicate the presence of multiple cardioactive compounds in L. leonurus.

Keywords: Cardiovascular effect, in vitro, in vivo, isolated perfused working heart, Leonotis leonurus, rat.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2062
363 Clustering Mixed Data Using Non-normal Regression Tree for Process Monitoring

Authors: Youngji Yoo, Cheong-Sool Park, Jun Seok Kim, Young-Hak Lee, Sung-Shick Kim, Jun-Geol Baek

Abstract:

In the semiconductor manufacturing process, large amounts of data are collected from various sensors of multiple facilities. The collected data from sensors have several different characteristics due to variables such as types of products, former processes and recipes. In general, Statistical Quality Control (SQC) methods assume the normality of the data to detect out-of-control states of processes. Although the collected data have different characteristics, using the data as inputs of SQC will increase variations of data, require wide control limits, and decrease performance to detect outof- control. Therefore, it is necessary to separate similar data groups from mixed data for more accurate process control. In the paper, we propose a regression tree using split algorithm based on Pearson distribution to handle non-normal distribution in parametric method. The regression tree finds similar properties of data from different variables. The experiments using real semiconductor manufacturing process data show improved performance in fault detecting ability.

Keywords: Semiconductor, non-normal mixed process data, clustering, Statistical Quality Control (SQC), regression tree, Pearson distribution system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1740
362 On the Computation of a Common n-finger Robotic Grasp for a Set of Objects

Authors: Avishai Sintov, Roland Menassa, Amir Shapiro

Abstract:

Industrial robotic arms utilize multiple end-effectors, each for a specific part and for a specific task. We propose a novel algorithm which will define a single end-effector’s configuration able to grasp a given set of objects with different geometries. The algorithm will have great benefit in production lines allowing a single robot to grasp various parts. Hence, reducing the number of endeffectors needed. Moreover, the algorithm will reduce end-effector design and manufacturing time and final product cost. The algorithm searches for a common grasp over the set of objects. The search algorithm maps all possible grasps for each object which satisfy a quality criterion and takes into account possible external wrenches (forces and torques) applied to the object. The mapped grasps are- represented by high-dimensional feature vectors which describes the shape of the gripper. We generate a database of all possible grasps for each object in the feature space. Then we use a search and classification algorithm for intersecting all possible grasps over all parts and finding a single common grasp suitable for all objects. We present simulations of planar and spatial objects to validate the feasibility of the approach.

Keywords: Common Grasping, Search Algorithm, Robotic End-Effector.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1629
361 New Features for Specific JPEG Steganalysis

Authors: Johann Barbier, Eric Filiol, Kichenakoumar Mayoura

Abstract:

We present in this paper a new approach for specific JPEG steganalysis and propose studying statistics of the compressed DCT coefficients. Traditionally, steganographic algorithms try to preserve statistics of the DCT and of the spatial domain, but they cannot preserve both and also control the alteration of the compressed data. We have noticed a deviation of the entropy of the compressed data after a first embedding. This deviation is greater when the image is a cover medium than when the image is a stego image. To observe this deviation, we pointed out new statistic features and combined them with the Multiple Embedding Method. This approach is motivated by the Avalanche Criterion of the JPEG lossless compression step. This criterion makes possible the design of detectors whose detection rates are independent of the payload. Finally, we designed a Fisher discriminant based classifier for well known steganographic algorithms, Outguess, F5 and Hide and Seek. The experiemental results we obtained show the efficiency of our classifier for these algorithms. Moreover, it is also designed to work with low embedding rates (< 10-5) and according to the avalanche criterion of RLE and Huffman compression step, its efficiency is independent of the quantity of hidden information.

Keywords: Compressed frequency domain, Fisher discriminant, specific JPEG steganalysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2129
360 A Multi-Feature Deep Learning Algorithm for Urban Traffic Classification with Limited Labeled Data

Authors: Rohan Putatunda, Aryya Gangopadhyay

Abstract:

Acoustic sensors, if embedded in smart street lights, can help in capturing the activities (car honking, sirens, events, traffic, etc.) in cities. Needless to say, the acoustic data from such scenarios are complex due to multiple audio streams originating from different events, and when decomposed to independent signals, the amount of retrieved data volume is small in quantity which is inadequate to train deep neural networks. So, in this paper, we address the two challenges: a) separating the mixed signals, and b) developing an efficient acoustic classifier under data paucity. So, to address these challenges, we propose an architecture with supervised deep learning, where the initial captured mixed acoustics data are analyzed with Fast Fourier Transformation (FFT), followed by filtering the noise from the signal, and then decomposed to independent signals by fast independent component analysis (Fast ICA). To address the challenge of data paucity, we propose a multi feature-based deep neural network with high performance that is reflected in our experiments when compared to the conventional convolutional neural network (CNN) and multi-layer perceptron (MLP).

Keywords: FFT, ICA, vehicle classification, multi-feature DNN, CNN, MLP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 380
359 Performance Determinants for Convenience Store Suppliers

Authors: Zainah Abdullah, Aznur Hajar Abdullah

Abstract:

This paper examines the impact of information and communication technology (ICT) usage, internal relationship, supplier-retailer relationship, logistics services and inventory management on convenience store suppliers- performance. Data was collected from 275 convenience store managers in Malaysia using a set of questionnaire. The multiple linear regression results indicate that inventory management, supplier-retailer relationship, logistics services and internal relationship are predictors of supplier performance as perceived by convenience store managers. However, ICT usage is not a predictor of supplier performance. The study focuses only on convenience stores and petrol station convenience stores and concentrates only on managers. The results provide insights to suppliers who serve convenience stores and possibly similar retail format on factors to consider in improving their service to retailers. The results also provide insights to government in its aspiration to improve business operations of convenience store to consider ways to enhance the adoption of ICT by retailers and suppliers.

Keywords: Information and communication technology (ICT), internal relationship, inventory management, logistics services, supplier performance, supplier-retailer relationship.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3979
358 Application of Single Subject Experimental Designs in Adapted Physical Activity Research: A Descriptive Analysis

Authors: Jiabei Zhang, Ying Qi

Abstract:

The purpose of this study was to develop a descriptive profile of the adapted physical activity research using single subject experimental designs. All research articles using single subject experimental designs published in the journal of Adapted Physical Activity Quarterly from 1984 to 2013 were employed as the data source. Each of the articles was coded in a subcategory of seven categories: (a) the size of sample; (b) the age of participants; (c) the type of disabilities; (d) the type of data analysis; (e) the type of designs, (f) the independent variable, and (g) the dependent variable. Frequencies, percentages, and trend inspection were used to analyze the data and develop a profile. The profile developed characterizes a small portion of research articles used single subject designs, in which most researchers used a small sample size, recruited children as subjects, emphasized learning and behavior impairments, selected visual inspection with descriptive statistics, preferred a multiple baseline design, focused on effects of therapy, inclusion, and strategy, and measured desired behaviors more often, with a decreasing trend over years.

Keywords: Adapted physical activity research, single subject experimental designs.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1805
357 Evaluation of Fitts’ Law Index of Difficulty Formulation for Screen Size Variations

Authors: Hidehiko Okada, Takayuki Akiba

Abstract:

It is well-known as Fitts’ law that the time for a user to point a target on a GUI screen can be modeled as a linear function of “index of difficulty (ID).” In this paper, the authors investigate whether the traditional ID formulation is appropriate independently of device screen sizes. Result of our experiment reveals that the ID formulation may not consistently capture actual difficulty: users’ pointing performances are not consistent among pointing target variations of which index of difficulty are consistent. The term A/W may not be appropriate because the term causes the observed inconsistency. Based on this finding, the authors then evaluate the applicability of possible models other than Fitts’ one. Multiple regression models are found to be able to appropriately represent the effects of target design variations. The authors next make an attempt to improve the definition of ID in Fitts’ model. Our idea is to raise the size or the distance values depending on the screen size. The modified model is found to fit well to the users’ pointing data, which supports the idea. 

Keywords: Fitts’ law, pointing device, small screen, touch user interface, usability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1592
356 Q-Net: A Novel QoS Aware Routing Algorithm for Future Data Networks

Authors: Maassoumeh Javadi Baygi, Abdul Rahman B Ramli, Borhanuddin Mohd Ali, Syamsiah Mashohor

Abstract:

The expectation of network performance from the early days of ARPANET until now has been changed significantly. Every day, new advancement in technological infrastructure opens the doors for better quality of service and accordingly level of perceived quality of network services have been increased over the time. Nowadays for many applications, late information has no value or even may result in financial or catastrophic loss, on the other hand, demands for some level of guarantee in providing and maintaining quality of service are ever increasing. Based on this history, having a QoS aware routing system which is able to provide today's required level of quality of service in the networks and effectively adapt to the future needs, seems as a key requirement for future Internet. In this work we have extended the traditional AntNet routing system to support QoS with multiple metrics such as bandwidth and delay which is named Q-Net. This novel scalable QoS routing system aims to provide different types of services in the network simultaneously. Each type of service can be provided for a period of time in the network and network nodes do not need to have any previous knowledge about it. When a type of quality of service is requested, Q-Net will allocate required resources for the service and will guarantee QoS requirement of the service, based on target objectives.

Keywords: Quality of Service, Routing, Ant Colony Optimization, Ant-based algorithms.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1299
355 Multi-Robotic Partial Disassembly Line Balancing with Robotic Efficiency Difference via HNSGA-II

Authors: Tao Yin, Zeqiang Zhang, Wei Liang, Yanqing Zeng, Yu Zhang

Abstract:

To accelerate the remanufacturing process of electronic waste products, this study designs a partial disassembly line with the multi-robotic station to effectively dispose of excessive wastes. The multi-robotic partial disassembly line is a technical upgrade to the existing manual disassembly line. Balancing optimization can make the disassembly line smoother and more efficient. For partial disassembly line balancing with the multi-robotic station (PDLBMRS), a mixed-integer programming model (MIPM) considering the robotic efficiency differences is established to minimize cycle time, energy consumption and hazard index and to calculate their optimal global values. Besides, an enhanced NSGA-II algorithm (HNSGA-II) is proposed to optimize PDLBMRS efficiently. Finally, MIPM and HNSGA-II are applied to an actual mixed disassembly case of two types of computers, the comparison of the results solved by GUROBI and HNSGA-II verifies the correctness of the model and excellent performance of the algorithm, and the obtained Pareto solution set provides multiple options for decision-makers.

Keywords: Waste disposal, disassembly line balancing, multi-robot station, robotic efficiency difference, HNSGA-II.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 434
354 Design of IMC-PID Controller Cascaded Filter for Simplified Decoupling Control System

Authors: Le Linh, Truong Nguyen Luan Vu, Le Hieu Giang

Abstract:

In this work, the IMC-PID controller cascaded filter based on Internal Model Control (IMC) scheme is systematically proposed for the simplified decoupling control system. The simplified decoupling is firstly introduced for multivariable processes by using coefficient matching to obtain a stable, proper, and causal simplified decoupler. Accordingly, transfer functions of decoupled apparent processes can be expressed as a set of n equivalent independent processes and then derived as a ratio of the original open-loop transfer function to the diagonal element of the dynamic relative gain array. The IMC-PID controller in series with filter is then directly employed to enhance the overall performance of the decoupling control system while avoiding difficulties arising from properties inherent to simplified decoupling. Some simulation studies are considered to demonstrate the simplicity and effectiveness of the proposed method. Simulations were conducted by tuning various controllers of the multivariate processes with multiple time delays. The results indicate that the proposed method consistently performs well with fast and well-balanced closed-loop time responses.

Keywords: Coefficient matching method, internal model control scheme, PID controller cascaded filter, simplified decoupler.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1443
353 Optimization of Strategies and Models Review for Optimal Technologies - Based On Fuzzy Schemes for Green Architecture

Authors: Ghada Elshafei, Abdelazim Negm

Abstract:

Recently, the green architecture becomes a significant way to a sustainable future. Green building designs involve finding the balance between comfortable homebuilding and sustainable environment. Moreover, the utilization of the new technologies such as artificial intelligence techniques are used to complement current practices in creating greener structures to keep the built environment more sustainable. The most common objectives in green buildings should be designed to minimize the overall impact of the built environment that effect on ecosystems in general and in particularly human health and natural environment. This will lead to protecting occupant health, improving employee productivity, reducing pollution and sustaining the environmental. In green building design, multiple parameters which may be interrelated, contradicting, vague and of qualitative/quantitative nature are broaden to use. This paper presents a comprehensive critical state- ofart- review of current practices based on fuzzy and its combination techniques. Also, presented how green architecture/building can be improved using the technologies that been used for analysis to seek optimal green solutions strategies and models to assist in making the best possible decision out of different alternatives.

Keywords: Green architecture/building, technologies, optimization, strategies, fuzzy techniques and models.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2489
352 Assessment of Energy Use and Energy Efficiency in Two Portuguese Slaughterhouses

Authors: M. Feliciano, F. Rodrigues, A. Gonçalves, J. M. R. C. A. Santos, V. Leite

Abstract:

With the objective of characterizing the profile and performance of energy use by slaughterhouses, surveys and audits were performed in two different facilities located in the northeastern region of Portugal. Energy consumption from multiple energy sources was assessed monthly, along with production and costs, for the same reference year. Gathered data was analyzed to identify and quantify the main consuming processes and to estimate energy efficiency indicators for benchmarking purposes. Main results show differences between the two slaughterhouses concerning energy sources, consumption by source and sector, and global energy efficiency. Electricity is the most used source in both slaughterhouses with a contribution of around 50%, being essentially used for meat processing and refrigeration. Natural gas, in slaughterhouse A, and pellets, in slaughterhouse B, used for heating water take the second place, with a mean contribution of about 45%. On average, a 62 kgoe/t specific energy consumption (SEC) was found, although with differences between slaughterhouses. A prominent negative correlation between SEC and carcass production was found specially in slaughterhouse A. Estimated Specific Energy Cost and Greenhouse Gases Intensity (GHGI) show mean values of about 50 €/t and 1.8 tCO2e/toe, respectively. Main results show that there is a significant margin for improving energy efficiency and therefore lowering costs in this type of non-energy intensive industries. 

Keywords: Meat industry, energy intensity, energy efficiency, GHG emissions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3653