Search results for: sequence information
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11460

Search results for: sequence information

11160 Machine Learning-Enabled Classification of Climbing Using Small Data

Authors: Nicholas Milburn, Yu Liang, Dalei Wu

Abstract:

Athlete performance scoring within the climbing do-main presents interesting challenges as the sport does not have an objective way to assign skill. Assessing skill levels within any sport is valuable as it can be used to mark progress while training, and it can help an athlete choose appropriate climbs to attempt. Machine learning-based methods are popular for complex problems like this. The dataset available was composed of dynamic force data recorded during climbing; however, this dataset came with challenges such as data scarcity, imbalance, and it was temporally heterogeneous. Investigated solutions to these challenges include data augmentation, temporal normalization, conversion of time series to the spectral domain, and cross validation strategies. The investigated solutions to the classification problem included light weight machine classifiers KNN and SVM as well as the deep learning with CNN. The best performing model had an 80% accuracy. In conclusion, there seems to be enough information within climbing force data to accurately categorize climbers by skill.

Keywords: classification, climbing, data imbalance, data scarcity, machine learning, time sequence

Procedia PDF Downloads 122
11159 Performance and Limitations of Likelihood Based Information Criteria and Leave-One-Out Cross-Validation Approximation Methods

Authors: M. A. C. S. Sampath Fernando, James M. Curran, Renate Meyer

Abstract:

Model assessment, in the Bayesian context, involves evaluation of the goodness-of-fit and the comparison of several alternative candidate models for predictive accuracy and improvements. In posterior predictive checks, the data simulated under the fitted model is compared with the actual data. Predictive model accuracy is estimated using information criteria such as the Akaike information criterion (AIC), the Bayesian information criterion (BIC), the Deviance information criterion (DIC), and the Watanabe-Akaike information criterion (WAIC). The goal of an information criterion is to obtain an unbiased measure of out-of-sample prediction error. Since posterior checks use the data twice; once for model estimation and once for testing, a bias correction which penalises the model complexity is incorporated in these criteria. Cross-validation (CV) is another method used for examining out-of-sample prediction accuracy. Leave-one-out cross-validation (LOO-CV) is the most computationally expensive variant among the other CV methods, as it fits as many models as the number of observations. Importance sampling (IS), truncated importance sampling (TIS) and Pareto-smoothed importance sampling (PSIS) are generally used as approximations to the exact LOO-CV and utilise the existing MCMC results avoiding expensive computational issues. The reciprocals of the predictive densities calculated over posterior draws for each observation are treated as the raw importance weights. These are in turn used to calculate the approximate LOO-CV of the observation as a weighted average of posterior densities. In IS-LOO, the raw weights are directly used. In contrast, the larger weights are replaced by their modified truncated weights in calculating TIS-LOO and PSIS-LOO. Although, information criteria and LOO-CV are unable to reflect the goodness-of-fit in absolute sense, the differences can be used to measure the relative performance of the models of interest. However, the use of these measures is only valid under specific circumstances. This study has developed 11 models using normal, log-normal, gamma, and student’s t distributions to improve the PCR stutter prediction with forensic data. These models are comprised of four with profile-wide variances, four with locus specific variances, and three which are two-component mixture models. The mean stutter ratio in each model is modeled as a locus specific simple linear regression against a feature of the alleles under study known as the longest uninterrupted sequence (LUS). The use of AIC, BIC, DIC, and WAIC in model comparison has some practical limitations. Even though, IS-LOO, TIS-LOO, and PSIS-LOO are considered to be approximations of the exact LOO-CV, the study observed some drastic deviations in the results. However, there are some interesting relationships among the logarithms of pointwise predictive densities (lppd) calculated under WAIC and the LOO approximation methods. The estimated overall lppd is a relative measure that reflects the overall goodness-of-fit of the model. Parallel log-likelihood profiles for the models conditional on equal posterior variances in lppds were observed. This study illustrates the limitations of the information criteria in practical model comparison problems. In addition, the relationships among LOO-CV approximation methods and WAIC with their limitations are discussed. Finally, useful recommendations that may help in practical model comparisons with these methods are provided.

Keywords: cross-validation, importance sampling, information criteria, predictive accuracy

Procedia PDF Downloads 369
11158 Myanmar Consonants Recognition System Based on Lip Movements Using Active Contour Model

Authors: T. Thein, S. Kalyar Myo

Abstract:

Human uses visual information for understanding the speech contents in noisy conditions or in situations where the audio signal is not available. The primary advantage of visual information is that it is not affected by the acoustic noise and cross talk among speakers. Using visual information from the lip movements can improve the accuracy and robustness of automatic speech recognition. However, a major challenge with most automatic lip reading system is to find a robust and efficient method for extracting the linguistically relevant speech information from a lip image sequence. This is a difficult task due to variation caused by different speakers, illumination, camera setting and the inherent low luminance and chrominance contrast between lip and non-lip region. Several researchers have been developing methods to overcome these problems; the one is lip reading. Moreover, it is well known that visual information about speech through lip reading is very useful for human speech recognition system. Lip reading is the technique of a comprehensive understanding of underlying speech by processing on the movement of lips. Therefore, lip reading system is one of the different supportive technologies for hearing impaired or elderly people, and it is an active research area. The need for lip reading system is ever increasing for every language. This research aims to develop a visual teaching method system for the hearing impaired persons in Myanmar, how to pronounce words precisely by identifying the features of lip movement. The proposed research will work a lip reading system for Myanmar Consonants, one syllable consonants (င (Nga)၊ ည (Nya)၊ မ (Ma)၊ လ (La)၊ ၀ (Wa)၊ သ (Tha)၊ ဟ (Ha)၊ အ (Ah) ) and two syllable consonants ( က(Ka Gyi)၊ ခ (Kha Gway)၊ ဂ (Ga Nge)၊ ဃ (Ga Gyi)၊ စ (Sa Lone)၊ ဆ (Sa Lain)၊ ဇ (Za Gwe) ၊ ဒ (Da Dway)၊ ဏ (Na Gyi)၊ န (Na Nge)၊ ပ (Pa Saug)၊ ဘ (Ba Gone)၊ ရ (Ya Gaug)၊ ဠ (La Gyi) ). In the proposed system, there are three subsystems, the first one is the lip localization system, which localizes the lips in the digital inputs. The next one is the feature extraction system, which extracts features of lip movement suitable for visual speech recognition. And the final one is the classification system. In the proposed research, Two Dimensional Discrete Cosine Transform (2D-DCT) and Linear Discriminant Analysis (LDA) with Active Contour Model (ACM) will be used for lip movement features extraction. Support Vector Machine (SVM) classifier is used for finding class parameter and class number in training set and testing set. Then, experiments will be carried out for the recognition accuracy of Myanmar consonants using the only visual information on lip movements which are useful for visual speech of Myanmar languages. The result will show the effectiveness of the lip movement recognition for Myanmar Consonants. This system will help the hearing impaired persons to use as the language learning application. This system can also be useful for normal hearing persons in noisy environments or conditions where they can find out what was said by other people without hearing voice.

Keywords: feature extraction, lip reading, lip localization, Active Contour Model (ACM), Linear Discriminant Analysis (LDA), Support Vector Machine (SVM), Two Dimensional Discrete Cosine Transform (2D-DCT)

Procedia PDF Downloads 263
11157 Orbital Tuning of Marl-Limestone Alternations (Upper Tithonian to Upper Berriasian) in North-South Axis (Tunisia): Geochronology and Sequence Implications

Authors: Hamdi Omar Omar, Hela Fakhfakh, Chokri Yaich

Abstract:

This work reflects the integration of different techniques, such as field sampling and observations, magnetic susceptibility measurement, cyclostratigaraphy and sequence stratigraphy. The combination of these results allows us to reconstruct the environmental evolution of the Sidi Khalif Formation in the North-South Axis (NOSA), aged of Upper Tithonian, Berriasian and Lower Valanginian. Six sedimentary facies were identified and are primarily influenced by open marine sedimentation receiving increasing terrigenous influx. Spectral analysis, based on MS variation (for the outcropped section) and wireline logging gamma ray (GR) variation (for the sub-area section) show a pervasive dominance of 405-kyr eccentricity cycles with the expression of 100-kyr eccentricity, obliquity and precession. This study provides (for the first time) a precise duration of 2.4 myr for the outcropped Sidi Khalif Formation with a sedimentation rate of 5.4 cm/kyr and the sub-area section to 3.24 myr with a sedimentation rate of 7.64 cm/kyr. We outlined 27 5th-order depositional sequences, 8 Milankovitch depositional sequences and 2 major 3rd-order cycles for the outcropping section, controlled by the long eccentricity (405 kyr) cycles and the precession index cycles. This study has demonstrated the potential of MS and GR to be used as proxies to develop an astronomically calibrated time-scale for the Mesozoic era.

Keywords: Berriasian, magnetic susceptibility, orbital tuning, Sidi Khalif Formation

Procedia PDF Downloads 239
11156 A Pattern Practise for Awareness Educations on Information Security: Information Security Project

Authors: Fati̇h Apaydin

Abstract:

Education technology is an area which constantly changes and creates innovations. As an inevitable part of the changing circumstances, the societies who have a tendency to the improvements keep up with these innovations by using the methods and strategies which have been designed for education technology. At this point, education technology has taken the responsibility to help the individuals improve themselves and teach the effective teaching methods by filling the airs in theoretical information, information security and the practice. The technology which comes to the core of our lives by raising the importance of it day by day and it enforced its position in computer- based environments. As a result, ‘being ready for technological innovations, improvement on computer-based talent, information, ability and attitude’ doctrines have to be given. However, it is today quite hard to deal with the security and reinforcement of this information. The information which is got illegally gives harm to society from every aspect, especially education. This study includes how and to what extent to use these innovative appliances such as computers and the factor of information security of these appliances in computer-based education. As the use of computer is constantly becoming prevalent in our country, both education and computer will never become out of date, so how computer-based education affects our lives and the study of information security for this type of education are important topics.

Keywords: computer, information security, education, technology, development

Procedia PDF Downloads 565
11155 Sequential Pattern Mining from Data of Medical Record with Sequential Pattern Discovery Using Equivalent Classes (SPADE) Algorithm (A Case Study : Bolo Primary Health Care, Bima)

Authors: Rezky Rifaini, Raden Bagus Fajriya Hakim

Abstract:

This research was conducted at the Bolo primary health Care in Bima Regency. The purpose of the research is to find out the association pattern that is formed of medical record database from Bolo Primary health care’s patient. The data used is secondary data from medical records database PHC. Sequential pattern mining technique is the method that used to analysis. Transaction data generated from Patient_ID, Check_Date and diagnosis. Sequential Pattern Discovery Algorithms Using Equivalent Classes (SPADE) is one of the algorithm in sequential pattern mining, this algorithm find frequent sequences of data transaction, using vertical database and sequence join process. Results of the SPADE algorithm is frequent sequences that then used to form a rule. It technique is used to find the association pattern between items combination. Based on association rules sequential analysis with SPADE algorithm for minimum support 0,03 and minimum confidence 0,75 is gotten 3 association sequential pattern based on the sequence of patient_ID, check_Date and diagnosis data in the Bolo PHC.

Keywords: diagnosis, primary health care, medical record, data mining, sequential pattern mining, SPADE algorithm

Procedia PDF Downloads 374
11154 Information Retrieval for Kafficho Language

Authors: Mareye Zeleke Mekonen

Abstract:

The Kafficho language has distinct issues in information retrieval because of its restricted resources and dearth of standardized methods. In this endeavor, with the cooperation and support of linguists and native speakers, we investigate the creation of information retrieval systems specifically designed for the Kafficho language. The Kafficho information retrieval system allows Kafficho speakers to access information easily in an efficient and effective way. Our objective is to conduct an information retrieval experiment using 220 Kafficho text files, including fifteen sample questions. Tokenization, normalization, stop word removal, stemming, and other data pre-processing chores, together with additional tasks like term weighting, were prerequisites for the vector space model to represent each page and a particular query. The three well-known measurement metrics we used for our word were Precision, Recall, and and F-measure, with values of 87%, 28%, and 35%, respectively. This demonstrates how well the Kaffiho information retrieval system performed well while utilizing the vector space paradigm.

Keywords: Kafficho, information retrieval, stemming, vector space

Procedia PDF Downloads 24
11153 Information Overload, Information Literacy and Use of Technology by Students

Authors: Elena Krelja Kurelović, Jasminka Tomljanović, Vlatka Davidović

Abstract:

The development of web technologies and mobile devices makes creating, accessing, using and sharing information or communicating with each other simpler every day. However, while the amount of information constantly increasing it is becoming harder to effectively organize and find quality information despite the availability of web search engines, filtering and indexing tools. Although digital technologies have overall positive impact on students’ lives, frequent use of these technologies and digital media enriched with dynamic hypertext and hypermedia content, as well as multitasking, distractions caused by notifications, calls or messages; can decrease the attention span, make thinking, memorizing and learning more difficult, which can lead to stress and mental exhaustion. This is referred to as “information overload”, “information glut” or “information anxiety”. Objective of this study is to determine whether students show signs of information overload and to identify the possible predictors. Research was conducted using a questionnaire developed for the purpose of this study. The results show that students frequently use technology (computers, gadgets and digital media), while they show moderate level of information literacy. They have sometimes experienced symptoms of information overload. According to the statistical analysis, higher frequency of technology use and lower level of information literacy are correlated with larger information overload. The multiple regression analysis has confirmed that the combination of these two independent variables has statistically significant predictive capacity for information overload. Therefore, the information science teachers should pay attention to improving the level of students’ information literacy and educate them about the risks of excessive technology use.

Keywords: information overload, computers, mobile devices, digital media, information literacy, students

Procedia PDF Downloads 256
11152 Full Characterization of Heterogeneous Antibody Samples under Denaturing and Native Conditions on a Hybrid Quadrupole-Orbitrap Mass Spectrometer

Authors: Rowan Moore, Kai Scheffler, Eugen Damoc, Jennifer Sutton, Aaron Bailey, Stephane Houel, Simon Cubbon, Jonathan Josephs

Abstract:

Purpose: MS analysis of monoclonal antibodies (mAbs) at the protein and peptide levels is critical during development and production of biopharmaceuticals. The compositions of current generation therapeutic proteins are often complex due to various modifications which may affect efficacy. Intact proteins analyzed by MS are detected in higher charge states that also provide more complexity in mass spectra. Protein analysis in native or native-like conditions with zero or minimal organic solvent and neutral or weakly acidic pH decreases charge state value resulting in mAb detection at higher m/z ranges with more spatial resolution. Methods: Three commercially available mAbs were used for all experiments. Intact proteins were desalted online using size exclusion chromatography (SEC) or reversed phase chromatography coupled on-line with a mass spectrometer. For streamlined use of the LC- MS platform we used a single SEC column and alternately selected specific mobile phases to perform separations in either denaturing or native-like conditions: buffer A (20 % ACN, 0.1 % FA) with Buffer B (100 mM ammonium acetate). For peptide analysis mAbs were proteolytically digested with and without prior reduction and alkylation. The mass spectrometer used for all experiments was a commercially available Thermo Scientific™ hybrid Quadrupole-Orbitrap™ mass spectrometer, equipped with the new BioPharma option which includes a new High Mass Range (HMR) mode that allows for improved high mass transmission and mass detection up to 8000 m/z. Results: We have analyzed the profiles of three mAbs under reducing and native conditions by direct infusion with offline desalting and with on-line desalting via size exclusion and reversed phase type columns. The presence of high salt under denaturing conditions was found to influence the observed charge state envelope and impact mass accuracy after spectral deconvolution. The significantly lower charge states observed under native conditions improves the spatial resolution of protein signals and has significant benefits for the analysis of antibody mixtures, e.g. lysine variants, degradants or sequence variants. This type of analysis requires the detection of masses beyond the standard mass range ranging up to 6000 m/z requiring the extended capabilities available in the new HMR mode. We have compared each antibody sample that was analyzed individually with mixtures in various relative concentrations. For this type of analysis, we observed that apparent native structures persist and ESI is benefited by the addition of low amounts of acetonitrile and formic acid in combination with the ammonium acetate-buffered mobile phase. For analyses on the peptide level we analyzed reduced/alkylated, and non-reduced proteolytic digests of the individual antibodies separated via reversed phase chromatography aiming to retrieve as much information as possible regarding sequence coverage, disulfide bridges, post-translational modifications such as various glycans, sequence variants, and their relative quantification. All data acquired were submitted to a single software package for analysis aiming to obtain a complete picture of the molecules analyzed. Here we demonstrate the capabilities of the mass spectrometer to fully characterize homogeneous and heterogeneous therapeutic proteins on one single platform. Conclusion: Full characterization of heterogeneous intact protein mixtures by improved mass separation on a quadrupole-Orbitrap™ mass spectrometer with extended capabilities has been demonstrated.

Keywords: disulfide bond analysis, intact analysis, native analysis, mass spectrometry, monoclonal antibodies, peptide mapping, post-translational modifications, sequence variants, size exclusion chromatography, therapeutic protein analysis, UHPLC

Procedia PDF Downloads 339
11151 A Review of Information Systems Development in Developing Countries

Authors: B. N. Asare, O. A. Ajigini

Abstract:

Information systems (IS) are highly important in the operation of private and public organisations in developing and developed countries. Developing countries are saddled with many project failures during the implementation of information systems. However, successful information systems are greatly needed in developing countries in order to enhance their economies. This paper is highly important in view of the high failure rate of information systems in developing countries which needs to be reduced to minimum acceptable levels by means of recommended interventions. This paper centres on a review of IS development in developing countries. The paper presents evidences of the IS successes and failures in developing countries and posits a model to address the IS failures. The proposed model can then be utilised by developing countries to reduce their IS project implementation failure rate. A comparison is drawn between IS development in developing countries and developed countries. The paper provides valuable information to assist in reducing IS failure, and developing IS models and theories on IS development for developing countries.

Keywords: developing countries, information systems, IS development, information systems failure, information systems success, information systems success model

Procedia PDF Downloads 341
11150 Valorization of Waste Reverse Osmosis Desalination Brine and Crystallization Sequence Approach for Kainite Recovery

Authors: Ayoub Bouazza, Ali Faddouli, Said Amal, Rachid Benhida, Khaoula Khaless

Abstract:

Brine waste generated from reverse osmosis (RO) desalination plants contains various valuable compounds, mainly salts, trace elements, and organic matter. These wastes are up to two times saltier than standard seawater. Therefore, there is a strong economic interest in recovering these salts. The current practice in desalination plants is to reject the brine back to the sea, which affects the marine ecosystem and the environment. Our study aims to bring forth a reliable management solution for the valorisation of waste brines. Natural evaporation, isothermal evaporation at 25°C and 50°C, and evaporation using continuous heating were used to crystallize valuable salts from a reverse osmosis desalination plant brine located on the Moroccan Atlantic coast. The crystallization sequence of the brine was studied in comparison with standard seawater. The X-Ray Diffraction (XRD) of the precipitated solid phases showed similar results, where halite was the main solid phase precipitated from both the brine and seawater. However, Jänecke diagram prediction, along with FREZCHEM simulations, showed that Kainite should crystallize before Epsomite and Carnallite. As the absence of kainite formation in many experiments in the literature has been related to the metastability of kainite and the critical relative humidity conditions, and the precipitation of K–Mg salts is very sensitive to climatic conditions. An evaporation process is proposed as a solution to achieve the predicted crystallization path and to affirm the recovery of Kainite.

Keywords: salts crystallization, reverse osmosis, solar evaporation, frezchem, ZLD

Procedia PDF Downloads 75
11149 Geoelectrical Investigation Around Bomo Area, Kaduna State, Nigeria

Authors: B. S. Jatau, Baba Adama, S. I. Fadele

Abstract:

Electrical resistivity investigation was carried out around Bomo area, Zaria, Kaduna state in order to study the subsurface geologic layer with a view of determining the depth to the bedrock and thickness of the geologic layers. Vertical Electrical Sounding (VES) using Schlumberger array was carried out at fifteen (15) VES stations. ABEM terrameter (SAS 300) was used for the data acquisition. The field data obtained have been analyzed using computer software (IPI2win) which gives an automatic interpretation of the apparent resistivity. The VES results revealed heterogeneous nature of the subsurface geological sequence. The geologic sequence beneath the study area is composed of hard pan top soil (clayey and sandy-lateritic), weathered layer, partly weathered or fractured basement and fresh basement. The resistivity value for the topsoil layer varies from 40Ωm to 450Ωm with thickness ranging from 1.25 to 7.5 m. The weathered basement has resistivity values ranging from 50Ωm to 593Ωm and thickness between 1.37 and 20.1 m. The fractured basement has resistivity values ranging from 218Ωm to 520Ωm and thickness of between 12.9 and 26.3 m. The fresh basement (bedrock) has resistivity values ranging from 1215Ωm to 2150Ωm with infinite depth. However, the depth of the earth’s surface to the bedrock surface varies between 2.63 and 34.99 m. The study further stressed the importance of the findings in civil engineering structures and groundwater prospecting.

Keywords: electrical resistivity, CERT (CT), vertical electrical sounding (VES), top soil (TP), weathered basement (WB), partly weathered basement (PWB), fresh basement (FB)

Procedia PDF Downloads 307
11148 Enterprise Security Architecture: Approaches and a Framework

Authors: Amir Mohtarami, Hadi Kandjani

Abstract:

The amount of business-critical information in enterprises is growing at an extraordinary rate, and the ability to catalog that information and properly protect it using traditional security mechanisms is not keeping pace. Alongside the Information Technology (IT), information security needs a holistic view in enterprise. In other words, a comprehensive architectural approach is required, focusing on the information itself, understanding what the data are, who owns it, and which business and regulatory policies should be applied to the information. Enterprise Architecture Frameworks provide useful tools to grasp different dimensions of IT in organizations. Usually this is done by the layered views on IT architecture, but not requisite security attention has been held in this frameworks. In this paper, after a brief look at the Enterprise Architecture (EA), we discuss the issue of security in the overall enterprise IT architecture. Due to the increasing importance of security, a rigorous EA program in an enterprise should be able to consider security architecture as an integral part of its processes and gives a visible roadmap and blueprint for this aim.

Keywords: enterprise architecture, architecture framework, security architecture, information systems

Procedia PDF Downloads 673
11147 Standardized Description and Modeling Methods of Semiconductor IP Interfaces

Authors: Seongsoo Lee

Abstract:

IP reuse is an effective design methodology for modern SoC design to reduce effort and time. However, description and modeling methods of IP interfaces are different due to different IP designers. In this paper, standardized description and modeling methods of IP interfaces are proposed. It consists of 11 items such as IP information, model provision, data type, description level, interface information, port information, signal information, protocol information, modeling level, modeling information, and source file. The proposed description and modeling methods enables easy understanding, simulation, verification, and modification in IP reuse.

Keywords: interface, standardization, description, modeling, semiconductor IP

Procedia PDF Downloads 475
11146 Earnings-Related Information, Cognitive Bias, and the Disposition Effect

Authors: Chih-Hsiang Chang, Pei-Shan Kao

Abstract:

This paper discusses the reaction of investors in the Taiwan stock market to the most probable unknown earnings-related information and the most probable known earnings-related information. As compared with the previous literature regarding the effect of an official announcement of earnings forecast revision, this paper further analyzes investors’ cognitive bias toward the unknown and known earnings-related information, and the role of media during the investors' reactions to the foresaid information shocks. The empirical results show that both the unknown and known earnings-related information provides useful information content for a stock market. In addition, cognitive bias and disposition effect are the behavioral pitfalls that commonly occur in the process of the investors' reactions to the earnings-related information. Finally, media coverage has a remarkable influence upon the investors' trading decisions.

Keywords: cognitive bias, role of media, disposition effect, earnings-related information, behavioral pitfall

Procedia PDF Downloads 201
11145 Building User Behavioral Models by Processing Web Logs and Clustering Mechanisms

Authors: Madhuka G. P. D. Udantha, Gihan V. Dias, Surangika Ranathunga

Abstract:

Today Websites contain very interesting applications. But there are only few methodologies to analyze User navigations through the Websites and formulating if the Website is put to correct use. The web logs are only used if some major attack or malfunctioning occurs. Web Logs contain lot interesting dealings on users in the system. Analyzing web logs has become a challenge due to the huge log volume. Finding interesting patterns is not as easy as it is due to size, distribution and importance of minor details of each log. Web logs contain very important data of user and site which are not been put to good use. Retrieving interesting information from logs gives an idea of what the users need, group users according to their various needs and improve site to build an effective and efficient site. The model we built is able to detect attacks or malfunctioning of the system and anomaly detection. Logs will be more complex as volume of traffic and the size and complexity of web site grows. Unsupervised techniques are used in this solution which is fully automated. Expert knowledge is only used in validation. In our approach first clean and purify the logs to bring them to a common platform with a standard format and structure. After cleaning module web session builder is executed. It outputs two files, Web Sessions file and Indexed URLs file. The Indexed URLs file contains the list of URLs accessed and their indices. Web Sessions file lists down the indices of each web session. Then DBSCAN and EM Algorithms are used iteratively and recursively to get the best clustering results of the web sessions. Using homogeneity, completeness, V-measure, intra and inter cluster distance and silhouette coefficient as parameters these algorithms self-evaluate themselves to input better parametric values to run the algorithms. If a cluster is found to be too large then micro-clustering is used. Using Cluster Signature Module the clusters are annotated with a unique signature called finger-print. In this module each cluster is fed to Associative Rule Learning Module. If it outputs confidence and support as value 1 for an access sequence it would be a potential signature for the cluster. Then the access sequence occurrences are checked in other clusters. If it is found to be unique for the cluster considered then the cluster is annotated with the signature. These signatures are used in anomaly detection, prevent cyber attacks, real-time dashboards that visualize users, accessing web pages, predict actions of users and various other applications in Finance, University Websites, News and Media Websites etc.

Keywords: anomaly detection, clustering, pattern recognition, web sessions

Procedia PDF Downloads 260
11144 Reverse Engineering of a Secondary Structure of a Helicopter: A Study Case

Authors: Jose Daniel Giraldo Arias, Camilo Rojas Gomez, David Villegas Delgado, Gullermo Idarraga Alarcon, Juan Meza Meza

Abstract:

The reverse engineering processes are widely used in the industry with the main goal to determine the materials and the manufacture used to produce a component. There are a lot of characterization techniques and computational tools that are used in order to get this information. A study case of a reverse engineering applied to a secondary sandwich- hybrid type structure used in a helicopter is presented. The methodology used consists of five main steps, which can be applied to any other similar component: Collect information about the service conditions of the part, disassembly and dimensional characterization, functional characterization, material properties characterization and manufacturing processes characterization, allowing to obtain all the supports of the traceability of the materials and processes of the aeronautical products that ensure their airworthiness. A detailed explanation of each step is covered. Criticality and comprehend the functionalities of each part, information of the state of the art and information obtained from interviews with the technical groups of the helicopter’s operators were analyzed,3D optical scanning technique, standard and advanced materials characterization techniques and finite element simulation allow to obtain all the characteristics of the materials used in the manufacture of the component. It was found that most of the materials are quite common in the aeronautical industry, including Kevlar, carbon, and glass fibers, aluminum honeycomb core, epoxy resin and epoxy adhesive. The stacking sequence and volumetric fiber fraction are a critical issue for the mechanical behavior; a digestion acid method was used for this purpose. This also helps in the determination of the manufacture technique which for this case was Vacuum Bagging. Samples of the material were manufactured and submitted to mechanical and environmental tests. These results were compared with those obtained during reverse engineering, which allows concluding that the materials and manufacture were correctly determined. Tooling for the manufacture was designed and manufactured according to the geometry and manufacture process requisites. The part was manufactured and the mechanical, and environmental tests required were also performed. Finally, a geometric characterization and non-destructive techniques allow verifying the quality of the part.

Keywords: reverse engineering, sandwich-structured composite parts, helicopter, mechanical properties, prototype

Procedia PDF Downloads 387
11143 Mitigating the Unwillingness of e-Forums Members to Engage in Information Exchange

Authors: Dora Triki, Irena Vida, Claude Obadia

Abstract:

Social networks such as e-Forums or dating sites often face the reluctance of key members to participate. Relying on the conation theory, this study investigates this phenomenon and proposes solutions to mitigate the issue. We show that highly experienced e-Forum members refuse to share business information in a peer to peer information exchange forums. However, forums managers can mitigate this behavior by developing a sentiment of belongingness to the network. Furthermore, by selecting only elite forum participants with ample experience, they can reduce the reluctance of key information providers to engage in information exchange. Our hypotheses are tested with PLS structural equations modeling using survey data from members of a French e-Forum dedicated to the exchange of business information about exporting.

Keywords: conation, e-Forum, information exchange, members participation

Procedia PDF Downloads 129
11142 Developing a Systems Dynamics Model for Security Management

Authors: Kuan-Chou Chen

Abstract:

This paper will demonstrate a simulation model of an information security system by using the systems dynamic approach. The relationships in the system model are designed to be simple and functional and do not necessarily represent any particular information security environments. The purpose of the paper aims to develop a generic system dynamic information security system model with implications on information security research. The interrelated and interdependent relationships of five primary sectors in the system dynamic model will be presented in this paper. The integrated information security systems model will include (1) information security characteristics, (2) users, (3) technology, (4) business functions, and (5) policy and management. Environments, attacks, government and social culture will be defined as the external sector. The interactions within each of these sectors will be depicted by system loop map as well. The proposed system dynamic model will not only provide a conceptual framework for information security analysts and designers but also allow information security managers to remove the incongruity between the management of risk incidents and the management of knowledge and further support information security managers and decision makers the foundation for managerial actions and policy decisions.

Keywords: system thinking, information security systems, security management, simulation

Procedia PDF Downloads 402
11141 Pale, Firm and Non-Exudative (PFN): An Emerging Major Broiler Breast Meat Group

Authors: Cintia Midori Kaminishikawahara, Fernanda Jéssica Mendonça, Moisés Grespan, Elza Iouko Ida, Massami Shimokomaki, Adriana Lourenço Soares

Abstract:

The quality of broiler breast meat is changing as a result of continuing emphasis on genetically bird’s selection for efficiently higher meat production. The consumer is experiencing a cooked product that is drier and less juicy when consumed. Breast meat has been classified as PSE (pale, soft, exudative), DFD (dark, firm, dry) and normal color meat. However, recently variations of this color have been observed and they are not in line with the specificity of the meat functional properties. Thus, the objective of this work was to report the finding of a new pale meat color group characterized as Pale, Firm and Non-exudative (PFN) based on its pH, color, meat functional properties and micro structural evaluation. Breast meat fillets samples (n=1045) from commercial line were classified into PSE (pH ≤5.8, L* ≥ 53.0), PFN (pH > 5.8 and L* ≥ 53.0) and Normal (pH >5.8 and L* < 53.0), based on pH and L* values. In sequence, a total of 30 samples of each group were analyzed for the water holding capacity (WHC) and shear force (SF). The incidence was 9.1% for PSE meat, 85.7% for PFN and 5.2% for Normal meat. The PSE meat presented lower values of WHC (P ≤ 0.05) followed in sequence by PFN and Normal samples and also the SF values of fresh PFN was higher than PSE meat (P ≤ 0.05) and similar to Normal samples. Under optical microscopy, the cell diameter was 10% higher for PFN in relation to PSE meat and similar to Normal meat. These preliminary results indicate an emerging group of breast meat and it should be considered that the Pale, Firm and Non-exudative should be considered as an ideal broiler breast meat quality.

Keywords: broiler PSE meat, light microscopy, texture, water holding capacity

Procedia PDF Downloads 317
11140 Explaining Irregularity in Music by Entropy and Information Content

Authors: Lorena Mihelac, Janez Povh

Abstract:

In 2017, we conducted a research study using data consisting of 160 musical excerpts from different musical styles, to analyze the impact of entropy of the harmony on the acceptability of music. In measuring the entropy of harmony, we were interested in unigrams (individual chords in the harmonic progression) and bigrams (the connection of two adjacent chords). In this study, it has been found that 53 musical excerpts out from 160 were evaluated by participants as very complex, although the entropy of the harmonic progression (unigrams and bigrams) was calculated as low. We have explained this by particularities of chord progression, which impact the listener's feeling of complexity and acceptability. We have evaluated the same data twice with new participants in 2018 and with the same participants for the third time in 2019. These three evaluations have shown that the same 53 musical excerpts, found to be difficult and complex in the study conducted in 2017, are exhibiting a high feeling of complexity again. It was proposed that the content of these musical excerpts, defined as “irregular,” is not meeting the listener's expectancy and the basic perceptual principles, creating a higher feeling of difficulty and complexity. As the “irregularities” in these 53 musical excerpts seem to be perceived by the participants without being aware of it, affecting the pleasantness and the feeling of complexity, they have been defined as “subliminal irregularities” and the 53 musical excerpts as “irregular.” In our recent study (2019) of the same data (used in previous research works), we have proposed a new measure of the complexity of harmony, “regularity,” based on the irregularities in the harmonic progression and other plausible particularities in the musical structure found in previous studies. We have in this study also proposed a list of 10 different particularities for which we were assuming that they are impacting the participant’s perception of complexity in harmony. These ten particularities have been tested in this paper, by extending the analysis in our 53 irregular musical excerpts from harmony to melody. In the examining of melody, we have used the computational model “Information Dynamics of Music” (IDyOM) and two information-theoretic measures: entropy - the uncertainty of the prediction before the next event is heard, and information content - the unexpectedness of an event in a sequence. In order to describe the features of melody in these musical examples, we have used four different viewpoints: pitch, interval, duration, scale degree. The results have shown that the texture of melody (e.g., multiple voices, homorhythmic structure) and structure of melody (e.g., huge interval leaps, syncopated rhythm, implied harmony in compound melodies) in these musical excerpts are impacting the participant’s perception of complexity. High information content values were found in compound melodies in which implied harmonies seem to have suggested additional harmonies, affecting the participant’s perception of the chord progression in harmony by creating a sense of an ambiguous musical structure.

Keywords: entropy and information content, harmony, subliminal (ir)regularity, IDyOM

Procedia PDF Downloads 109
11139 Sequence Polymorphism and Haplogroup Distribution of Mitochondrial DNA Control Regions HVS1 and HVS2 in a Southwestern Nigerian Population

Authors: Ogbonnaya O. Iroanya, Samson T. Fakorede, Osamudiamen J. Edosa, Hadiat A. Azeez

Abstract:

The human mitochondrial DNA (mtDNA) is about 17 kbp circular DNA fragments found within the mitochondria together with smaller fragments of 1200 bp known as the control region. Knowledge of variation within populations has been employed in forensic and molecular anthropology studies. The study was aimed at investigating the polymorphic nature of the two hypervariable segments (HVS) of the mtDNA, i.e., HVS1 and HVS2, and to determine the haplogroup distribution among individuals resident in Lagos, Southwestern Nigeria. Peripheral blood samples were obtained from sixty individuals who are not related maternally, followed by DNA extraction and amplification of the extracted DNA using primers specific for the regions under investigation. DNA amplicons were sequenced, and sequenced data were aligned and compared to the revised Cambridge Reference Sequence (rCRS) GenBank Accession number: NC_012920.1) using BioEdit software. Results obtained showed 61 and 52 polymorphic nucleotide positions for HVS1 and HVS2, respectively. While a total of three indels mutation were recorded for HVS1, there were seven for HVS2. Also, transition mutations predominate nucleotide change observed in the study. Genetic diversity (GD) values for HVS1 and HVS2 were estimated to be 84.21 and 90.4%, respectively, while random match probability was 0.17% for HVS1 and 0.89% for HVS2. The study also revealed mixed haplogroups specific to the African (L1-L3) and the Eurasians (U and H) lineages. New polymorphic sites obtained from the study are promising for human identification purposes.

Keywords: hypervariable region, indels, mitochondrial DNA, polymorphism, random match probability

Procedia PDF Downloads 93
11138 Prediction of B-Cell Epitope for 24 Mite Allergens: An in Silico Approach towards Epitope-Based Immune Therapeutics

Authors: Narjes Ebrahimi, Soheila Alyasin, Navid Nezafat, Hossein Esmailzadeh, Younes Ghasemi, Seyed Hesamodin Nabavizadeh

Abstract:

Immunotherapy with allergy vaccines is of great importance in allergen-specific immunotherapy. In recent years, B-cell epitope-based vaccines have attracted considerable attention and the prediction of epitopes is crucial to design these types of allergy vaccines. B-cell epitopes might be linear or conformational. The prerequisite for the identification of conformational epitopes is the information about allergens' tertiary structures. Bioinformatics approaches have paved the way towards the design of epitope-based allergy vaccines through the prediction of tertiary structures and epitopes. Mite allergens are one of the major allergy contributors. Several mite allergens can elicit allergic reactions; however, their structures and epitopes are not well established. So, B-cell epitopes of various groups of mite allergens (24 allergens in 6 allergen groups) were predicted in the present work. Tertiary structures of 17 allergens with unknown structure were predicted and refined with RaptorX and GalaxyRefine servers, respectively. The predicted structures were further evaluated by Rampage, ProSA-web, ERRAT and Verify 3D servers. Linear and conformational B-cell epitopes were identified with Ellipro, Bcepred, and DiscoTope 2 servers. To improve the accuracy level, consensus epitopes were selected. Fifty-four conformational and 133 linear consensus epitopes were predicted. Furthermore, overlapping epitopes in each allergen group were defined, following the sequence alignment of the allergens in each group. The predicted epitopes were also compared with the experimentally identified epitopes. The presented results provide valuable information for further studies about allergy vaccine design.

Keywords: B-cell epitope, Immunotherapy, In silico prediction, Mite allergens, Tertiary structure

Procedia PDF Downloads 131
11137 Strain Based Failure Criterion for Composite Notched Laminates

Authors: Ibrahim A. Elsayed, Mohamed H. Elalfy, Mostafa M. Abdalla

Abstract:

A strain-based failure criterion for composite notched laminates is introduced where the most critical stress concentration factor for the anisotropic notched laminates could be related to the failure of the corresponding quasi-isotropic laminate and the anisotropy ratio of the laminate. The proposed criterion will simplify the design of composites to meet notched failure requirements by eliminating the need for the detailed specifications of the stacking sequence at the preliminary design stage. The designer will be able to design based on the stiffness of the laminate, then at a later stage, select an appropriate stacking sequence to meet the stiffness requirements. The failure strains for the notched laminates are computed using the material’s Omni-strain envelope. The concept of Omni-strain envelope concerns the region of average strain where the laminate is safe regardless of ply orientation. In this work, we use Hashin’s failure criteria and the strains around the hole are computed using Savin’s analytic solution. A progressive damage analysis study has been conducted where the failure loads for the notched laminates are computed using finite element analysis. The failure strains are computed and used to estimate the concentration factor. It is found that the correlation found using Savin’s analytic solution predicts the same ratio of concentration factors between anisotropic and quasi-isotropic laminates as the more expensive progressive failure analysis.

Keywords: anisotropy ratio, failure criteria, notched laminates, Omni-strain envelope, savin’s solution

Procedia PDF Downloads 88
11136 Analysis of Genetic Variations in Camel Breeds (Camelus dromedarius)

Authors: Yasser M. Saad, Amr A. El Hanafy, Saleh A. Alkarim, Hussein A. Almehdar, Elrashdy M. Redwan

Abstract:

Camels are substantial providers of transport, milk, sport, meat, shelter, security and capital in many countries, particularly in Saudi Arabia. Inter simple sequence repeat technique was used to detect the genetic variations among some camel breeds (Majaheim, Safra, Wadah, and Hamara). Actual number of alleles, effective number of alleles, gene diversity, Shannon’s information index and polymorphic bands were calculated for each evaluated camel breed. Neighbor-joining tree that re-constructed for evaluated these camel breeds showed that, Hamara breed is distantly related from the other evaluated camels. In addition, the polymorphic sites, haplotypes and nucleotide diversity were identified for some camelidae cox1 gene sequences (obtained from NCBI). The distance value between C. bactrianus and C. dromedarius (0.072) was relatively low. Analysis of genetic diversity is an important way for conserving Camelus dromedarius genetic resources.

Keywords: camel, genetics, ISSR, neighbor-joining

Procedia PDF Downloads 445
11135 Research on the United Navigation Mechanism of Land, Sea and Air Targets under Multi-Sources Information Fusion

Authors: Rui Liu, Klaus Greve

Abstract:

The navigation information is a kind of dynamic geographic information, and the navigation information system is a kind of special geographic information system. At present, there are many researches on the application of centralized management and cross-integration application of basic geographic information. However, the idea of information integration and sharing is not deeply applied into the research of navigation information service. And the imperfection of navigation target coordination and navigation information sharing mechanism under certain navigation tasks has greatly affected the reliability and scientificity of navigation service such as path planning. Considering this, the project intends to study the multi-source information fusion and multi-objective united navigation information interaction mechanism: first of all, investigate the actual needs of navigation users in different areas, and establish the preliminary navigation information classification and importance level model; and then analyze the characteristics of the remote sensing and GIS vector data, and design the fusion algorithm from the aspect of improving the positioning accuracy and extracting the navigation environment data. At last, the project intends to analyze the feature of navigation information of the land, sea and air navigation targets, and design the united navigation data standard and navigation information sharing model under certain navigation tasks, and establish a test navigation system for united navigation simulation experiment. The aim of this study is to explore the theory of united navigation service and optimize the navigation information service model, which will lay the theory and technology foundation for the united navigation of land, sea and air targets.

Keywords: information fusion, united navigation, dynamic path planning, navigation information visualization

Procedia PDF Downloads 254
11134 A Study on the Treatment of Municipal Waste Water Using Sequencing Batch Reactor

Authors: Bhaven N. Tandel, Athira Rajeev

Abstract:

Sequencing batch reactor process is a suspended growth process operating under non-steady state conditions which utilizes a fill and draw reactor with complete mixing during the batch reaction step (after filling) and where the subsequent steps of aeration and clarification occur in the same tank. All sequencing batch reactor systems have five steps in common, which are carried out in sequence as follows, (1) fill (2) react (3) settle (sedimentation/clarification) (4) draw (decant) and (5) idle. The study was carried out in a sequencing batch reactor of dimensions 44cmx30cmx70cm with a working volume of 40 L. Mechanical stirrer of 100 rpm was used to provide continuous mixing in the react period and oxygen was supplied by fish tank aerators. The duration of a complete cycle of sequencing batch reactor was 8 hours. The cycle period was divided into different phases in sequence as follows-0.25 hours fill phase, 6 hours react period, 1 hour settling phase, 0.5 hours decant period and 0.25 hours idle phase. The study consisted of two runs, run 1 and run 2. Run 1 consisted of 6 hours aerobic react period and run 2 consisted of 3 hours aerobic react period followed by 3 hours anoxic react period. The influent wastewater used for the study had COD, BOD, NH3-N and TKN concentrations of 308.03±48.94 mg/L, 100.36±22.05 mg/L, 14.12±1.18 mg/L, and 24.72±2.21 mg/L respectively. Run 1 had an average COD removal efficiency of 41.28%, BOD removal efficiency of 56.25%, NH3-N removal efficiency of 86.19% and TKN removal efficiency of 54.4%. Run 2 had an average COD removal efficiency of 63.19%, BOD removal efficiency of 73.85%, NH3-N removal efficiency of 90.74% and TKN removal efficiency of 65.25%. It was observed that run 2 gave better performance than run 1 in the removal of COD, BOD and TKN.

Keywords: municipal waste water, aerobic, anoxic, sequencing batch reactor

Procedia PDF Downloads 518
11133 The Role of Information Technology in the Supply Chain Management

Authors: Azar Alizadeh, Mohammad Reza Naserkhaki

Abstract:

The application of the IT systems for collecting and analyzing the data can have a significant effect on the performance of any company. In recent decade, different advancements and achievements in the field of information technology have changed the industry compared to the previous decade. The adoption and application of the information technology are one of the ways to achieve a distinctive competitive personality to the companies and their supply chain. The acceptance of the IT and its proper implementation cam reinforce and improve the cooperation between different parts of the supply chain by rapid transfer and distribution of the precise information and the application of the informational systems, leading to the increase in the supply chain efficiency. The main objective of this research is to study the effects and applications of the information technology on and in the supply chain management and to introduce the effective factors on the acceptance of information technology in the companies. Moreover, in order to understand the subject, we will investigate the role and importance of the information and electronic commerce in the supply chain and the characteristics of the supply chain based on the information flow approach.

Keywords: electronic commerce, industry, information technology, management, supply chain, system

Procedia PDF Downloads 455
11132 Genetic Diversity and Molecular Basis of Carbapenem Resistance in Acinetobacter Baumannii Isolates from Cattle

Authors: Minhas Alam, Muhammad Hidayat Rasool, Mohsin Khurshid, Bilal Aslam

Abstract:

Acinetobacter baumannii is a notorious bacterial pathogen that is an emerging nightmare in clinical settings and is mainly involved in severe nosocomial infections. However, the data related to carbapenem-resistant A. baumannii (CRAB) from veterinary settings is limited, especially in developing countries like Pakistan. To investigate the genetic diversity and molecular basis of carbapenem resistance in Acinetobacter baumannii isolates from Cattle, a total of 1960 samples were collected from cattle from Punjab, Pakistan. The isolates were analyzed by routine microbiological procedures and confirmed by polymerase chain reaction (PCR). The isolates were further screened for antimicrobial susceptibility and the presence of multiple antimicrobial-resistant determinants by PCR. Multilocus sequence typing (MLST) was performed. The results of the current study revealed that the overall prevalence of A. baumannii in cattle was 3.28% (65/1980). Among cattle 27.7% (18/65) were found CRAB strains. The CRAB isolates harbor class D β- lactamases genes, e-g, blaOXA-23 and blaOXA-51, 94.4% (17/18). CRAB isolates carry class B β- lactamases gene blaIMP, and only one isolate carries the blaNDM-1 gene. The MLST results of CRAB isolates from cattle demonstrated 5 STs and one new ST. The commonly found sequence types in CRAB isolates were ST2 (n=10, 55.5%), followed by ST642 (n=5, 27.8%) and ST600 & ST889 (n=1, 5.55%). The presence of CRAB isolates in cattle indicates an alarming situation in Punjab, Pakistan. Immediate control measures should be taken to stop the transmission of CRAB isolates within cattle, to the environment, and to clinical settings.

Keywords: acinetobacter baumannii, carbapenemases, veterinary, drug resistance

Procedia PDF Downloads 29
11131 In silico and in vitro Investigation of the Role of Acinetobacter baumannii in the Pathogenesis of Multiple Sclerosis

Authors: Kieren Luellman, Makenzi Rockwell, Eduardo Callegari, Nichole Haag, Chun Wu

Abstract:

Multiple sclerosis (MS) is an autoimmune disorder that damages the myelin sheath of neurons in the central nervous system. The presence of Acinetobacter bacteria and anti-Acinetobacter antibodies in MS patients has led to the hypothesis that the bacteria may contribute to MS pathogenesis. In this study, the protein sequences of Acinetobacter baumannii were compared to five peptides from three mammalian myelin proteins, i.e., Proteolipid Protein (PLP): PLP 139-151, PLP 178-191, Myelin Basic Protein (MBP): MBP 84-104 and Myelin Oligodendrocyte Glycoprotein (MOG): MOG 35-55 and MOG 92-106 respectively, known to induce experimental autoimmune encephalomyelitis (EAE), a condition similar to MS. We found 11 hits (i.e., with five or more amino acid sequence similarity) in Acinetobacter baumannii, which are identical or similar to PLP139-151, 32 hits to PLP178-191, 35 to MBP 84-104, 41 hits to MOG 35-55 and 26 hits to MOG92-106. In addition, Western blotting was used to assess possible interaction between the bacterial proteins and human anti-MBP, anti-MOG, and anti-PLP antibodies produced in rabbits, corresponding to MBP 84-104, MOG 35-55, and PLP 139-151, respectively. We found that both human Polyclonal anti-MOG antibody and anti-PLP antibody recognized a protein or more proteins of the same molecular mass of around 25 kDa. in Acinetobacter baumannii. The results suggested that this/these protein(s) might potentially serve as antigen(s) to induce anti-MOG antibody and anti-PLP antibody production in mammalian B cells. The proteomic study identified 433 hits, among which the sequence of Acinetobacter baumannii protein 491 subunit A matches a previously published enzyme Acinetobacter 3-Oxoadipate CoA-Transferase, in which a fragment of its peptide was observed to recognize MS patient serum via ELISA method. Our findings might pave the road to understanding one of the pathogenesis mechanisms of MS.

Keywords: multiple sclerosis, pathogenesis, Acinetobacter baumannii, antibody recognition

Procedia PDF Downloads 87