Search results for: super-efficiency data envelopment analysis
12111 Gene Selection Guided by Feature Interdependence
Authors: Hung-Ming Lai, Andreas Albrecht, Kathleen Steinhöfel
Abstract:
Cancers could normally be marked by a number of differentially expressed genes which show enormous potential as biomarkers for a certain disease. Recent years, cancer classification based on the investigation of gene expression profiles derived by high-throughput microarrays has widely been used. The selection of discriminative genes is, therefore, an essential preprocess step in carcinogenesis studies. In this paper, we have proposed a novel gene selector using information-theoretic measures for biological discovery. This multivariate filter is a four-stage framework through the analyses of feature relevance, feature interdependence, feature redundancy-dependence and subset rankings, and having been examined on the colon cancer data set. Our experimental result show that the proposed method outperformed other information theorem based filters in all aspect of classification errors and classification performance.
Keywords: Colon cancer, feature interdependence, feature subset selection, gene selection, microarray data analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 214412110 Sampled-Data Model Predictive Tracking Control for Mobile Robot
Authors: Wookyong Kwon, Sangmoon Lee
Abstract:
In this paper, a sampled-data model predictive tracking control method is presented for mobile robots which is modeled as constrained continuous-time linear parameter varying (LPV) systems. The presented sampled-data predictive controller is designed by linear matrix inequality approach. Based on the input delay approach, a controller design condition is derived by constructing a new Lyapunov function. Finally, a numerical example is given to demonstrate the effectiveness of the presented method.Keywords: Model predictive control, sampled-data control, linear parameter varying systems, LPV.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 128012109 Impact of Vehicle Travel Characteristics on Level of Service: A Comparative Analysis of Rural and Urban Freeways
Authors: Anwaar Ahmed, Muhammad Bilal Khurshid, Samuel Labi
Abstract:
The effect of trucks on the level of service is determined by considering passenger car equivalents (PCE) of trucks. The current version of Highway Capacity Manual (HCM) uses a single PCE value for all tucks combined. However, the composition of truck traffic varies from location to location; therefore, a single PCE value for all trucks may not correctly represent the impact of truck traffic at specific locations. Consequently, present study developed separate PCE values for single-unit and combination trucks to replace the single value provided in the HCM on different freeways. Site specific PCE values, were developed using concept of spatial lagging headways (that is the distance between rear bumpers of two vehicles in a traffic stream) measured from field traffic data. The study used data from four locations on a single urban freeway and three different rural freeways in Indiana. Three-stage-leastsquares (3SLS) regression techniques were used to generate models that predicted lagging headways for passenger cars, single unit trucks (SUT), and combination trucks (CT). The estimated PCE values for single-unit and combination truck for basic urban freeways (level terrain) were: 1.35 and 1.60, respectively. For rural freeways the estimated PCE values for single-unit and combination truck were: 1.30 and 1.45, respectively. As expected, traffic variables such as vehicle flow rates and speed have significant impacts on vehicle headways. Study results revealed that the use of separate PCE values for different truck classes can have significant influence on the LOS estimation.
Keywords: Level of Service, Capacity Analysis, Lagging Headway.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 209612108 Talent Management through Integration of Talent Value Chain and Human Capital Analytics Approaches
Authors: Wuttigrai Ngamsirijit
Abstract:
Talent management in today’s modern organizations has become data-driven due to a demand for objective human resource decision making and development of analytics technologies. HR managers have been faced with some obstacles in exploiting data and information to obtain their effective talent management decisions. These include process-based data and records; insufficient human capital-related measures and metrics; lack of capabilities in data modeling in strategic manners; and, time consuming to add up numbers and make decisions. This paper proposes a framework of talent management through integration of talent value chain and human capital analytics approaches. It encompasses key data, measures, and metrics regarding strategic talent management decisions along the organizational and talent value chain. Moreover, specific predictive and prescriptive models incorporating these data and information are recommended to help managers in understanding the state of talent, gaps in managing talent and the organization, and the ways to develop optimized talent strategies.Keywords: Decision making, human capital analytics, talent management, talent value chain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 96912107 Analysis of Secondary School Students’ Perceptions about Information Technologies through a Word Association Test
Authors: Fetah Eren, Ismail Sahin, Ismail Celik, Ahmet Oguz Akturk
Abstract:
The aim of this study is to discover secondary school students’ perceptions related to information technologies and the connections between concepts in their cognitive structures. A word association test consisting of six concepts related to information technologies is used to collect data from 244 secondary school students. Concept maps that present students’ cognitive structures are drawn with the help of frequency data. Data are analyzed and interpreted according to the connections obtained as a result of the concept maps. It is determined students associate most with these concepts—computer, Internet, and communication of the given concepts, and associate least with these concepts—computer-assisted education and information technologies. These results show the concepts, Internet, communication, and computer, are an important part of students’ cognitive structures. In addition, students mostly answer computer, phone, game, Internet and Facebook as the key concepts. These answers show students regard information technologies as a means for entertainment and free time activity, not as a means for education.
Keywords: Word association test, cognitive structure, information technology, secondary school.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 207912106 Comparison of E-Waste Management in Switzerland and in Australia: A Qualitative Content Analysis
Authors: Md Tasbirul Islam, Pablo Dias, Nazmul Huda
Abstract:
E-waste/Waste electrical and electronic equipment (WEEE) is one of the fastest growing waste streams across the globe. This paper aims to compare the e-waste management system in Switzerland and Australia in terms of four features - legislative initiatives, disposal practice, collection and financial mechanisms. The qualitative content analysis is employed as a research method in the study. Data were collected from various published academic research papers, industry reports, and web sources. In addition, a questionnaire survey is conducted in Australia to understand the public awareness and opinions on the features. The results of the study provide valuable insights to policymakers in Australia developing better e-waste management system in conjunction with the public consensus, and the state-of-the-art operational strategies currently being practiced in Switzerland.Keywords: E-waste management, WEEE, awareness, pro-environmental behavior, Australia, Switzerland.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 175612105 Innovative Teaching in Systems Analysis and Design - an Action Research Project
Authors: Imelda Smit
Abstract:
Systems Analysis and Design is a key subject in Information Technology courses, but students do not find it easy to cope with, since it is not “precise" like programming and not exact like Mathematics. It is a subject working with many concepts, modeling ideas into visual representations and then translating the pictures into a real life system. To complicate matters users who are not necessarily familiar with computers need to give their inputs to ensure that they get the system the need. Systems Analysis and Design also covers two fields, namely Analysis, focusing on the analysis of the existing system and Design, focusing on the design of the new system. To be able to test the analysis and design of a system, it is necessary to develop a system or at least a prototype of the system to test the validity of the analysis and design. The skills necessary in each aspect differs vastly. Project Management Skills, Database Knowledge and Object Oriented Principles are all necessary. In the context of a developing country where students enter tertiary education underprepared and the digital divide is alive and well, students need to be motivated to learn the necessary skills, get an opportunity to test it in a “live" but protected environment – within the framework of a university. The purpose of this article is to improve the learning experience in Systems Analysis and Design through reviewing the underlying teaching principles used, the teaching tools implemented, the observations made and the reflections that will influence future developments in Systems Analysis and Design. Action research principles allows the focus to be on a few problematic aspects during a particular semester.Keywords: Action Research, Project Development, Systems Analysis and Design, Technology in Teaching.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 145212104 Unsupervised Feature Selection Using Feature Density Functions
Authors: Mina Alibeigi, Sattar Hashemi, Ali Hamzeh
Abstract:
Since dealing with high dimensional data is computationally complex and sometimes even intractable, recently several feature reductions methods have been developed to reduce the dimensionality of the data in order to simplify the calculation analysis in various applications such as text categorization, signal processing, image retrieval, gene expressions and etc. Among feature reduction techniques, feature selection is one the most popular methods due to the preservation of the original features. In this paper, we propose a new unsupervised feature selection method which will remove redundant features from the original feature space by the use of probability density functions of various features. To show the effectiveness of the proposed method, popular feature selection methods have been implemented and compared. Experimental results on the several datasets derived from UCI repository database, illustrate the effectiveness of our proposed methods in comparison with the other compared methods in terms of both classification accuracy and the number of selected features.Keywords: Feature, Feature Selection, Filter, Probability Density Function
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 207712103 Enhancing K-Means Algorithm with Initial Cluster Centers Derived from Data Partitioning along the Data Axis with the Highest Variance
Authors: S. Deelers, S. Auwatanamongkol
Abstract:
In this paper, we propose an algorithm to compute initial cluster centers for K-means clustering. Data in a cell is partitioned using a cutting plane that divides cell in two smaller cells. The plane is perpendicular to the data axis with the highest variance and is designed to reduce the sum squared errors of the two cells as much as possible, while at the same time keep the two cells far apart as possible. Cells are partitioned one at a time until the number of cells equals to the predefined number of clusters, K. The centers of the K cells become the initial cluster centers for K-means. The experimental results suggest that the proposed algorithm is effective, converge to better clustering results than those of the random initialization method. The research also indicated the proposed algorithm would greatly improve the likelihood of every cluster containing some data in it.Keywords: Clustering algorithm, K-means algorithm, Datapartitioning, Initial cluster centers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 286612102 A Novel Multiplex Real-Time PCR Assay Using TaqMan MGB Probes for Rapid Detection of Trisomy 21
Authors: Mehrdad Hashemi, Mitra Behrooz Aghdam, Reza Mahdian, Ahmad Reza Kamyab
Abstract:
Cytogenetic analysis still remains the gold standard method for prenatal diagnosis of trisomy 21 (Down syndrome, DS). Nevertheless, the conventional cytogenetic analysis needs live cultured cells and is too time-consuming for clinical application. In contrast, molecular methods such as FISH, QF-PCR, MLPA and quantitative Real-time PCR are rapid assays with results available in 24h. In the present study, we have successfully used a novel MGB TaqMan probe-based real time PCR assay for rapid diagnosis of trisomy 21 status in Down syndrome samples. We have also compared the results of this molecular method with corresponding results obtained by the cytogenetic analysis. Blood samples obtained from DS patients (n=25) and normal controls (n=20) were tested by quantitative Real-time PCR in parallel to standard G-banding analysis. Genomic DNA was extracted from peripheral blood lymphocytes. A high precision TaqMan probe quantitative Real-time PCR assay was developed to determine the gene dosage of DSCAM (target gene on 21q22.2) relative to PMP22 (reference gene on 17p11.2). The DSCAM/PMP22 ratio was calculated according to the formula; ratio=2 -ΔΔCT. The quantitative Real-time PCR was able to distinguish between trisomy 21 samples and normal controls with the gene ratios of 1.49±0.13 and 1.03±0.04 respectively (p value <0.001). These results represent the presence of 3 copies of target gene in DS samples Vs 2 copies in normal controls. The results of quantitative Real-time PCR were in complete agreement with results of cytogenetic analysis. This study confirms previous reports regarding successful implementation of quantitative Real-time PCR for detection of trisomy 21. However, the assay has been improved by using MGB probes and more accurate data analysis. This assay, in particular, when performed in combination with another molecular assay such as QF-PCR or MLPA, can be used as a reliable technique for rapid prenatal diagnosis of trisomy 21.
Keywords: Trisomy 21, Real-time PCR, MGB-TaqMan Probes, Gene Dosage.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 253812101 Semi-Supervised Outlier Detection Using a Generative and Adversary Framework
Authors: Jindong Gu, Matthias Schubert, Volker Tresp
Abstract:
In many outlier detection tasks, only training data belonging to one class, i.e., the positive class, is available. The task is then to predict a new data point as belonging either to the positive class or to the negative class, in which case the data point is considered an outlier. For this task, we propose a novel corrupted Generative Adversarial Network (CorGAN). In the adversarial process of training CorGAN, the Generator generates outlier samples for the negative class, and the Discriminator is trained to distinguish the positive training data from the generated negative data. The proposed framework is evaluated using an image dataset and a real-world network intrusion dataset. Our outlier-detection method achieves state-of-the-art performance on both tasks.Keywords: Outlier detection, generative adversary networks, semi-supervised learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 107412100 An Analysis of Language Borrowing among Algerian University Students Using Online Facebook Conversations
Authors: Messaouda Annab
Abstract:
The rapid development of technology has led to an important context in which different languages and structures are used in the same conversations. This paper investigates the practice of language borrowing within social media platform, namely, Facebook among Algerian Vernacular Arabic (AVA) students. In other words, this study will explore how Algerian students have incorporated lexical English borrowing in their online conversations. This paper will examine the relationships between language, culture and identity among a multilingual group. The main objective is to determine the cultural and linguistic functions that borrowing fulfills in social media and to explain the possible factors underlying English borrowing. The nature of the study entails the use of an online research method that includes ten online Facebook conversations in the form of private messages collected from Bachelor and Masters Algerian students recruited from the English department at the University of Oum El-Bouaghi. The analysis of data revealed that social media platform provided the users with opportunities to shift from one language to another. This practice was noticed in students’ online conversations. English borrowing was the most relevant language performance in accordance with Arabic which is the mother tongue of the chosen sample. The analysis has assumed that participants are skilled in more than one language.Keywords: Borrowing, language performance, linguistic background, social media.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 129712099 Significant Role Analysis of Transmission Control Protocols in 4G Cellular Systems
Authors: Ghassan A. Abed, Bayan M. Sabbar
Abstract:
The society of 3rd Generation Partnership Project (3GPP) is completed developing Long Term Evolution Advanced (LTE-Advanced) systems as a standard 4G cellular system. This generation goals to produce conditions for a new radio-access technology geared to higher data rates, low latency, and better spectral efficiency. LTE-Advanced is an evolutionary step in the continuing development of LTE where the description in this article is based on LTE release 10. This paper provides a model of the traffic links of 4G system represented by LTE-Advanced system with the effect of the Transmission Control Protocols (TCP) and Stream Control Transmission Protocol (SCTP) in term of throughput and packet loss. Furthermore, the article presents the investigation and the analysis the behavior of SCTP and TCP variants over the 4G cellular systems. The traffic model and the scenario of the simulation developed using the network simulator NS-2 using different TCP source variants.
Keywords: LTE-Advanced, LTE, SCTP, TCP, 4G, NS-2.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 194512098 The Vulnerability Analysis of Java Bytecode Based on Points-to Dataflow
Authors: Tang Hong, Zhang Lufeng, Chen Hua, Zhang Jianbo
Abstract:
Today many developers use the Java components collected from the Internet as external LIBs to design and develop their own software. However, some unknown security bugs may exist in these components, such as SQL injection bug may comes from the components which have no specific check for the input string by users. To check these bugs out is very difficult without source code. So a novel method to check the bugs in Java bytecode based on points-to dataflow analysis is in need, which is different to the common analysis techniques base on the vulnerability pattern check. It can be used as an assistant tool for security analysis of Java bytecode from unknown softwares which will be used as extern LIBs.Keywords: Java bytecode, points-to dataflow, vulnerability analysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 176312097 Mechanical Characteristics of Spaghetti Enriched with Whole Soy Flour
Authors: Nasehi, B., Mortazavi, S. A., Razavi, S.
Abstract:
The influence of full-fat soy flour (FFSF) and extrusion conditions on the mechanical characteristics of dry spaghetti were evaluated. Process was performed with screw speed of 10-40rpm and water circulating temperature of 35-70°C. Data analysis using mixture design showed that this enrichment resulted in significant differences in mechanical strength.Keywords: Pasta, Mixture design, Enrichment, Texture.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 265612096 Comparative Analysis of Photovoltaic Systems
Authors: Irtaza M. Syed, Kaamran Raahemifar
Abstract:
This paper presents comparative analysis of photovoltaic systems (PVS) and propose practical techniques to improve operational efficiency of the PVS. The best engineering and construction practices for PVS are identified and field oriented recommendation are made. Comparative analysis of central and string inverter based, as well as 600 and 1000VDC PVS are performed. In addition, direct current (DC) and alternating current (AC) photovoltaic (PV) module based systems are compared. Comparison shows that 1000V DC String Inverters based PVS is the best choice.Keywords: Photovoltaic module, photovoltaic systems, operational efficiency improvement, comparative analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 229812095 Exploring SSD Suitable Allocation Schemes Incompliance with Workload Patterns
Authors: Jae Young Park, Hwansu Jung, Jong Tae Kim
Abstract:
In the Solid-State-Drive (SSD) performance, whether the data has been well parallelized is an important factor. SSD parallelization is affected by allocation scheme and it is directly connected to SSD performance. There are dynamic allocation and static allocation in representative allocation schemes. Dynamic allocation is more adaptive in exploiting write operation parallelism, while static allocation is better in read operation parallelism. Therefore, it is hard to select the appropriate allocation scheme when the workload is mixed read and write operations. We simulated conditions on a few mixed data patterns and analyzed the results to help the right choice for better performance. As the results, if data arrival interval is long enough prior operations to be finished and continuous read intensive data environment static allocation is more suitable. Dynamic allocation performs the best on write performance and random data patterns.
Keywords: Dynamic allocation, NAND Flash based SSD, SSD parallelism, static allocation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 199512094 Sonic Therapeutic Intervention for Preventing Financial Fraud: A Phenomenological Study
Authors: Vasudev Das
Abstract:
The specific problem is that private and public organizational leaders often do not understand the importance of sonic therapeutic intervention in preventing financial fraud. The study aimed to explore sonic therapeutic intervention practitioners' lived experiences regarding the value of sonic therapeutic intervention in preventing financial fraud. The data collection methods were semi-structured interviews of purposeful samples and documentary reviews, which were analyzed thematically. Four themes emerged from the analysis of interview transcription data: Sonic therapeutic intervention enabled self-control, pro-spiritual values, consequentiality mindset, and post-conventional consciousness. The itemized four themes helped non-engagement in financial fraud. Implications for positive social change include enhanced financial fraud management, more significant financial leadership, and result-oriented decision-taking in the financial market. Also, the study results can improve the increased de-escalation of anxiety/stress associated with defrauding.
Keywords: consciousness, consequentiality, rehabilitation, reintegration
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 80712093 Walsh-Hadamard Transform for Facial Feature Extraction in Face Recognition
Authors: M. Hassan, I. Osman, M. Yahia
Abstract:
This Paper proposes a new facial feature extraction approach, Wash-Hadamard Transform (WHT). This approach is based on correlation between local pixels of the face image. Its primary advantage is the simplicity of its computation. The paper compares the proposed approach, WHT, which was traditionally used in data compression with two other known approaches: the Principal Component Analysis (PCA) and the Discrete Cosine Transform (DCT) using the face database of Olivetti Research Laboratory (ORL). In spite of its simple computation, the proposed algorithm (WHT) gave very close results to those obtained by the PCA and DCT. This paper initiates the research into WHT and the family of frequency transforms and examines their suitability for feature extraction in face recognition applications.
Keywords: Face Recognition, Facial Feature Extraction, Principal Component Analysis, and Discrete Cosine Transform, Wash-Hadamard Transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 257112092 WebAppShield: An Approach Exploiting Machine Learning to Detect SQLi Attacks in an Application Layer in Run-Time
Authors: Ahmed Abdulla Ashlam, Atta Badii, Frederic Stahl
Abstract:
In recent years, SQL injection attacks have been identified as being prevalent against web applications. They affect network security and user data, which leads to a considerable loss of money and data every year. This paper presents the use of classification algorithms in machine learning using a method to classify the login data filtering inputs into "SQLi" or "Non-SQLi,” thus increasing the reliability and accuracy of results in terms of deciding whether an operation is an attack or a valid operation. A method as a Web-App is developed for auto-generated data replication to provide a twin of the targeted data structure. Shielding against SQLi attacks (WebAppShield) that verifies all users and prevents attackers (SQLi attacks) from entering and or accessing the database, which the machine learning module predicts as "Non-SQLi", has been developed. A special login form has been developed with a special instance of the data validation; this verification process secures the web application from its early stages. The system has been tested and validated, and up to 99% of SQLi attacks have been prevented.
Keywords: SQL injection, attacks, web application, accuracy, database, WebAppShield.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 44412091 On the Analysis of IP Traffic Distribution in the Network of Suranaree University of Technology
Authors: Paramet Nualmuenwai, Chutima Prommak
Abstract:
This paper presents the IP traffic analysis. The traffic was collected from the network of Suranaree University of Technology using the software based on the Simple Network Management Protocol (SNMP). In particular, we analyze the distribution of the aggregated traffic during the hours of peak load and light load. The traffic profiles including the parameters described the traffic distributions were derived. From the statistical analysis applying three different methods, including the Kolmogorov Smirnov test, Anderson Darling test, and Chi-Squared test, we found that the IP traffic distribution is a non-normal distribution and the distributions during the peak load and the light load are different. The experimental study and analysis show high uncertainty of the IP traffic.Keywords: IP traffic analysis, IP traffic distribution, Traffic uncertainty
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 151712090 Heat Transfer Enhancement Studies in a Circular Tube Fitted with Right-Left Helical Inserts with Spacer
Authors: P. K. Nagarajan, P. Sivashanmugam
Abstract:
Experimental investigation of heat transfer and friction factor characteristics of circular tube fitted with 300 right-left helical screw inserts with 100 mm spacer of different twist ratio has been presented for laminar and turbulent flow.. The experimental data obtained were compared with those obtained from plain tube published data. The heat transfer coefficient enhancement for 300 RL inserts with 100 mm spacer is quite comparable with for 300 R-L inserts. Performance evaluation analysis has been made and found that the performance ratio increases with increasing Reynolds number and decreasing twist ration with the maximum for the twist ratio 2.93. Also, the performance ratio of more than one indicates that the type of twist inserts can be used effectively for heat transfer augmentation.Keywords: Heat transfer augmentation, right-left helical screw inserts with spacer, Twist ratio, Heat Transfer
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 255612089 Parametric Analysis and Optimal Design of Functionally Graded Plates Using Particle Swarm Optimization Algorithm and a Hybrid Meshless Method
Authors: Foad Nazari, Seyed Mahmood Hosseini, Mohammad Hossein Abolbashari, Mohammad Hassan Abolbashari
Abstract:
The present study is concerned with the optimal design of functionally graded plates using particle swarm optimization (PSO) algorithm. In this study, meshless local Petrov-Galerkin (MLPG) method is employed to obtain the functionally graded (FG) plate’s natural frequencies. Effects of two parameters including thickness to height ratio and volume fraction index on the natural frequencies and total mass of plate are studied by using the MLPG results. Then the first natural frequency of the plate, for different conditions where MLPG data are not available, is predicted by an artificial neural network (ANN) approach which is trained by back-error propagation (BEP) technique. The ANN results show that the predicted data are in good agreement with the actual one. To maximize the first natural frequency and minimize the mass of FG plate simultaneously, the weighted sum optimization approach and PSO algorithm are used. However, the proposed optimization process of this study can provide the designers of FG plates with useful data.Keywords: Optimal design, natural frequency, FG plate, hybrid meshless method, MLPG method, ANN approach, particle swarm optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 143312088 Development and Validation of the Response to Stressful Situations Scale in the General Population
Authors: C. Barreto Carvalho, C. da Motta, M. Sousa, J. Cabral, A. L. Carvalho, E. B. Peixoto
Abstract:
The aim of the current study was to develop and validate a Response to Stressful Situations Scale (RSSS) for the Portuguese population. This scale assesses the degree of stress experienced in scenarios that can constitute positive, negative and more neutral stressors, and also describes the physiological, emotional and behavioral reactions to those events according to their intensity. These scenarios include typical stressor scenarios relevant to patients with schizophrenia, which are currently absent from most scales, assessing specific risks that these stressors may bring on subjects, which may prove useful in non-clinical and clinical populations (i.e. Patients with mood or anxiety disorders, schizophrenia). Results from Principal Components Analysis and Confirmatory Factor Analysis of two adult samples from general population allowed to confirm a three-factor model with good fit indices: χ2 (144)= 370.211, p = 0.000; GFI = 0.928; CFI = 0.927; TLI = 0.914, RMSEA = 0.055, P(rmsea ≤0.005) = .096; PCFI = .781. Further data analysis of the scale revealed that RSSS is an adequate assessment tool of stress response in adults to be used in further research and clinical settings, with good psychometric characteristics, adequate divergent and convergent validity, good temporal stability and high internal consistency.Keywords: Assessment, stress events, stress response, stress vulnerability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 212412087 Proposing an Efficient Method for Frequent Pattern Mining
Authors: Vaibhav Kant Singh, Vijay Shah, Yogendra Kumar Jain, Anupam Shukla, A.S. Thoke, Vinay KumarSingh, Chhaya Dule, Vivek Parganiha
Abstract:
Data mining, which is the exploration of knowledge from the large set of data, generated as a result of the various data processing activities. Frequent Pattern Mining is a very important task in data mining. The previous approaches applied to generate frequent set generally adopt candidate generation and pruning techniques for the satisfaction of the desired objective. This paper shows how the different approaches achieve the objective of frequent mining along with the complexities required to perform the job. This paper will also look for hardware approach of cache coherence to improve efficiency of the above process. The process of data mining is helpful in generation of support systems that can help in Management, Bioinformatics, Biotechnology, Medical Science, Statistics, Mathematics, Banking, Networking and other Computer related applications. This paper proposes the use of both upward and downward closure property for the extraction of frequent item sets which reduces the total number of scans required for the generation of Candidate Sets.Keywords: Data Mining, Candidate Sets, Frequent Item set, Pruning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 168412086 An Empirical Analysis of HRM in Different Pharmaceutical Departments of Different Pharmaceutical Industries in Pakistan
Authors: Faisal Ali, Mansoor Shuakat, Lirong Cui, Helena Uhde, Rabia Riasat, Janeth J. Marwa
Abstract:
HR is a department that enhances the power of employee performance in regard with their services, and to make the organization strategic objectives. The main concern of HR department is to organize people, focus on policies and their system. The empirical study shows the relationship between HRM (Human Resource Management practices) and their Job Satisfaction. The Hypothesis is testing on a sample of overall 320 employees of 5 different Pharmaceutical departments of different organizations in Pakistan. The important thing as Relationship of Job satisfaction with HR Practices, Impact on Job Satisfaction with HR Practices, Participation of Staff of Different Departments, HR Practices effects the Job satisfaction, Recruitment or Hiring and Selection effects the Job satisfaction, Training and Development, Performance and Appraisals, Compensation affects the Job satisfaction , and Industrial Relationships affects the Job satisfaction. After finishing all data analysis, the conclusion is that lots of Job related activities raise the confidence of Job satisfaction of employees with their salary and other benefits.Keywords: HRM, HR practices, job satisfaction, TQM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 184012085 An Investigation of Short Circuit Analysis in Komag Sarawak Operations (KSO) Factory
Authors: M. H. Hairi, H. Zainuddin, M.H.N. Talib, A. Khamis, J. Y. Lichun
Abstract:
Short circuit currents plays a vital role in influencing the design and operation of equipment and power system and could not be avoided despite careful planning and design, good maintenance and thorough operation of the system. This paper discusses the short circuit analysis conducted in KSO briefly comprising of its significances, methods and results. A result sample of the analysis based on a single transformer is detailed in this paper. Furthermore, the results of the analysis and its significances were also discussed and commented.
Keywords: Short circuit currents, Transformer fault current
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 240312084 Danger Theory and Intelligent Data Processing
Authors: Anjum Iqbal, Mohd Aizaini Maarof
Abstract:
Artificial Immune System (AIS) is relatively naive paradigm for intelligent computations. The inspiration for AIS is derived from natural Immune System (IS). Classically it is believed that IS strives to discriminate between self and non-self. Most of the existing AIS research is based on this approach. Danger Theory (DT) argues this approach and proposes that IS fights against danger producing elements and tolerates others. We, the computational researchers, are not concerned with the arguments among immunologists but try to extract from it novel abstractions for intelligent computation. This paper aims to follow DT inspiration for intelligent data processing. The approach may introduce new avenue in intelligent processing. The data used is system calls data that is potentially significant in intrusion detection applications.Keywords: artificial immune system, danger theory, intelligent processing, system calls
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 188312083 Using Artificial Neural Network to Forecast Groundwater Depth in Union County Well
Authors: Zahra Ghadampour, Gholamreza Rakhshandehroo
Abstract:
A concern that researchers usually face in different applications of Artificial Neural Network (ANN) is determination of the size of effective domain in time series. In this paper, trial and error method was used on groundwater depth time series to determine the size of effective domain in the series in an observation well in Union County, New Jersey, U.S. different domains of 20, 40, 60, 80, 100, and 120 preceding day were examined and the 80 days was considered as effective length of the domain. Data sets in different domains were fed to a Feed Forward Back Propagation ANN with one hidden layer and the groundwater depths were forecasted. Root Mean Square Error (RMSE) and the correlation factor (R2) of estimated and observed groundwater depths for all domains were determined. In general, groundwater depth forecast improved, as evidenced by lower RMSEs and higher R2s, when the domain length increased from 20 to 120. However, 80 days was selected as the effective domain because the improvement was less than 1% beyond that. Forecasted ground water depths utilizing measured daily data (set #1) and data averaged over the effective domain (set #2) were compared. It was postulated that more accurate nature of measured daily data was the reason for a better forecast with lower RMSE (0.1027 m compared to 0.255 m) in set #1. However, the size of input data in this set was 80 times the size of input data in set #2; a factor that may increase the computational effort unpredictably. It was concluded that 80 daily data may be successfully utilized to lower the size of input data sets considerably, while maintaining the effective information in the data set.Keywords: Neural networks, groundwater depth, forecast.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 251612082 Organizational Data Security in Perspective of Ownership of Mobile Devices Used by Employees for Works
Authors: B. Ferdousi, J. Bari
Abstract:
With advancement of mobile computing, employees are increasingly doing their job-related works using personally owned mobile devices or organization owned devices. The Bring Your Own Device (BYOD) model allows employees to use their own mobile devices for job-related works, while Corporate Owned, Personally Enabled (COPE) model allows both organizations and employees to install applications onto organization-owned mobile devices used for job-related works. While there are many benefits of using mobile computing for job-related works, there are also serious concerns of different levels of threats to the organizational data security. Consequently, it is crucial to know the level of threat to the organizational data security in the BOYD and COPE models. It is also important to ensure that employees comply with the organizational data security policy. This paper discusses the organizational data security issues in perspective of ownership of mobile devices used by employees, especially in BYOD and COPE models. It appears that while the BYOD model has many benefits, there are relatively more data security risks in this model than in the COPE model. The findings also showed that in both BYOD and COPE environments, a more practical approach towards achieving secure mobile computing in organizational setting is through the development of comprehensive cybersecurity policies balancing employees’ need for convenience with organizational data security. The study helps to figure out the compliance and the risks of security breach in BYOD and COPE models.
Keywords: Data security, mobile computing, BYOD, COPE, cybersecurity policy, cybersecurity compliance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 374