Search results for: information content model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 29209

Search results for: information content model

28669 Comparison of Crossover Types to Obtain Optimal Queries Using Adaptive Genetic Algorithm

Authors: Wafa’ Alma'Aitah, Khaled Almakadmeh

Abstract:

this study presents an information retrieval system of using genetic algorithm to increase information retrieval efficiency. Using vector space model, information retrieval is based on the similarity measurement between query and documents. Documents with high similarity to query are judge more relevant to the query and should be retrieved first. Using genetic algorithms, each query is represented by a chromosome; these chromosomes are fed into genetic operator process: selection, crossover, and mutation until an optimized query chromosome is obtained for document retrieval. Results show that information retrieval with adaptive crossover probability and single point type crossover and roulette wheel as selection type give the highest recall. The proposed approach is verified using (242) proceedings abstracts collected from the Saudi Arabian national conference.

Keywords: genetic algorithm, information retrieval, optimal queries, crossover

Procedia PDF Downloads 284
28668 Influence of κ-Casein Genotype on Milk Productivity of Latvia Local Dairy Breeds

Authors: S. Petrovska, D. Jonkus, D. Smiltiņa

Abstract:

κ-casein is one of milk proteins which are very important for milk processing. Genotypes of κ-casein affect milk yield, fat, and protein content. The main factors which affect local Latvian dairy breed milk yield and composition are analyzed in research. Data were collected from 88 Latvian brown and 82 Latvian blue cows in 2015. AA genotype was 0.557 in Latvian brown and 0.232 in Latvian blue breed. BB genotype was 0.034 in Latvian brown and 0.207 in Latvian blue breed. Highest milk yield was observed in Latvian brown (5131.2 ± 172.01 kg), significantly high fat content and fat yield also was in Latvian brown (p < 0.05). Significant differences between κ-casein genotypes were not found in Latvian brown, but highest milk yield (5057 ± 130.23 kg), protein content (3.42 ± 0.03%), and protein yield (171.9 ± 4.34 kg) were with AB genotype. Significantly high fat content was observed in Latvian blue breed with BB genotype (4.29 ± 0.17%) compared with AA genotypes (3.42 ± 0.19). Similar tendency was found in protein content – 3.27 ± 0.16% with BB genotype and 2.59 ± 0.16% with AA genotype (p < 0.05). Milk yield increases by increasing parity. We did not obtain major tendency of changes of milk fat and protein content according parity.

Keywords: dairy cows, κ-casein, milk productivity, polymorphism

Procedia PDF Downloads 261
28667 Cognitive Model of Analogy Based on Operation of the Brain Cells: Glial, Axons and Neurons

Authors: Ozgu Hafizoglu

Abstract:

Analogy is an essential tool of human cognition that enables connecting diffuse and diverse systems with attributional, deep structural, casual relations that are essential to learning, to innovation in artificial worlds, and to discovery in science. Cognitive Model of Analogy (CMA) leads and creates information pattern transfer within and between domains and disciplines in science. This paper demonstrates the Cognitive Model of Analogy (CMA) as an evolutionary approach to scientific research. The model puts forward the challenges of deep uncertainty about the future, emphasizing the need for flexibility of the system in order to enable reasoning methodology to adapt to changing conditions. In this paper, the model of analogical reasoning is created based on brain cells, their fractal, and operational forms within the system itself. Visualization techniques are used to show correspondences. Distinct phases of the problem-solving processes are divided thusly: encoding, mapping, inference, and response. The system is revealed relevant to brain activation considering each of these phases with an emphasis on achieving a better visualization of the brain cells: glial cells, axons, axon terminals, and neurons, relative to matching conditions of analogical reasoning and relational information. It’s found that encoding, mapping, inference, and response processes in four-term analogical reasoning are corresponding with the fractal and operational forms of brain cells: glial, axons, and neurons.

Keywords: analogy, analogical reasoning, cognitive model, brain and glials

Procedia PDF Downloads 177
28666 Picture of the World by the Second Law of Thermodynamic

Authors: Igor V. Kuzminov

Abstract:

According to its content, the proposed article is a collection of articles with comments and additions. All articles, in one way or another, have a connection with the Second Law of Thermodynamics. The content of the articles is given in a concise form. The articles were published in different journals at different times. Main topics are presented: gravity, biography of the Earth, physics of global warming-cooling cycles, multiverse. The articles are based on the laws of classical physics. Along the way, it should be noted that the Second Law of thermodynamics can be formulated as the Law of Matter Cooling. As it cools down, the processes of condensation, separation, and changes in the aggregate states of matter occur. In accordance with these changes, a picture of the world is being formed. Also, the main driving force of these processes is the inverse temperature dependence of the forces of gravity. As matter cools, the forces of gravity increase. The actions of these phenomena in the compartment form a picture of the world.

Keywords: gravitational forces, cooling of matter, inverse temperature dependence of gravitational forces, planetary model of the atom

Procedia PDF Downloads 234
28665 Energy Enterprise Information System for Strategic Decision-Making

Authors: Woosik Jang, Seung H. Han, Seung Won Baek, Chan Young Park

Abstract:

Natural gas (NG) is a local energy resource that exists in certain countries, and most NG producers operate within unstable governments. Moreover, about 90% of the liquefied natural gas (LNG) market is governed by a small number of international oil companies (IOCs) and national oil companies (NOCs), market entry of second movers is extremely limited. To overcome these barriers, project viability should be assessed based on limited information at the project screening perspective. However, there have been difficulties at the early stages of projects as follows: (1) What factors should be considered? (2) How many experts are needed to make a decision? and (3) How to make an optimal decision with limited information? To answer these questions, this research suggests a LNG project viability assessment model based on the Dempster-Shafer theory (DST). Total of 11 indices for the gas field analysis and 23 indices for the market environment analysis are identified that reflect unique characteristics of LNG industry. Moreover, the proposed model evaluates LNG projects based on questionnaire survey and it provides not only quantified results but also uncertainty level of results based on DST. Consequently, the proposed model as a systematic framework can support the decision-making process from the gas field projects using quantitative results, and it is developed to a stand-alone system to enhance the practical usability. It is expected to improve the decision-making quality and opportunity in LNG projects for enterprise through informed decision.

Keywords: project viability, LNG project, enterprise information system, Dempster-Shafer Theory, strategic decision-making

Procedia PDF Downloads 251
28664 Energy-Efficient Contact Selection Method for CARD in Wireless Ad-Hoc Networks

Authors: Mehdi Assefi, Keihan Hataminezhad

Abstract:

One of the efficient architectures for exploring the resources in wireless ad-hoc networks is contact-based architecture. In this architecture, each node assigns a unique zone for itself and each node keeps all information from inside the zone, as well as some from outside the zone, which is called contact. Reducing the overlap between different zones of a node and its contacts increases its performance, therefore Edge Method (EM) is designed for this purpose. Contacts selected by EM do not have any overlap with their sources, but for choosing the contact a vast amount of information must be transmitted. In this article, we will offer a new protocol for contact selection, which is called PEM. The objective would be reducing the volume of transmitted information, using Non-Uniform Dissemination Probabilistic Protocols. Consumed energy for contact selection is a function of the size of transmitted information between nodes. Therefore, by reducing the content of contact selection message using the PEM will decrease the consumed energy. For evaluation of the PEM we applied the simulation method. Results indicated that PEM consumes less energy compared to EM, and by increasing the number of nodes (level of nodes), performance of PEM will improve in comparison with EM.

Keywords: wireless ad-hoc networks, contact selection, method for CARD, energy-efficient

Procedia PDF Downloads 281
28663 Exploring Factors Affecting the Implementation of Flexible Curriculum in Information Systems Higher Education

Authors: Clement C. Aladi, Zhaoxia Yi

Abstract:

This study investigates factors influencing the implementation of flexible curricula in e-learning in Information Systems (IS) higher education. Drawing from curriculum theorists and contemporary literature, and using the Technology, Pedagogy, and Content Knowledge (TPACK) framework, it explores teacher-related challenges and their impact on curriculum flexibility implementation. By using the PLS-SEM, the study uncovers these factors and hopes to contribute to enhancing curriculum flexibility in delivering online and blended learning in IS higher education.

Keywords: flexible curriculum, online learning, e-learning, technology

Procedia PDF Downloads 39
28662 A Statistical Model for the Geotechnical Parameters of Cement-Stabilised Hightown’s Soft Soil: A Case Stufy of Liverpool, UK

Authors: Hassnen M. Jafer, Khalid S. Hashim, W. Atherton, Ali W. Alattabi

Abstract:

This study investigates the effect of two important parameters (length of curing period and percentage of the added binder) on the strength of soil treated with OPC. An intermediate plasticity silty clayey soil with medium organic content was used in this study. This soft soil was treated with different percentages of a commercially available cement type 32.5-N. laboratory experiments were carried out on the soil treated with 0, 1.5, 3, 6, 9, and 12% OPC by the dry weight to determine the effect of OPC on the compaction parameters, consistency limits, and the compressive strength. Unconfined compressive strength (UCS) test was carried out on cement-treated specimens after exposing them to different curing periods (1, 3, 7, 14, 28, and 90 days). The results of UCS test were used to develop a non-linear multi-regression model to find the relationship between the predicted and the measured maximum compressive strength of the treated soil (qu). The results indicated that there was a significant improvement in the index of plasticity (IP) by treating with OPC; IP was decreased from 20.2 to 14.1 by using 12% of OPC; this percentage was enough to increase the UCS of the treated soil up to 1362 kPa after 90 days of curing. With respect to the statistical model of the predicted qu, the results showed that the regression coefficients (R2) was equal to 0.8534 which indicates a good reproducibility for the constructed model.

Keywords: cement admixtures, soft soil stabilisation, geotechnical parameters, multi-regression model

Procedia PDF Downloads 355
28661 Content-Aware Image Augmentation for Medical Imaging Applications

Authors: Filip Rusak, Yulia Arzhaeva, Dadong Wang

Abstract:

Machine learning based Computer-Aided Diagnosis (CAD) is gaining much popularity in medical imaging and diagnostic radiology. However, it requires a large amount of high quality and labeled training image datasets. The training images may come from different sources and be acquired from different radiography machines produced by different manufacturers, digital or digitized copies of film radiographs, with various sizes as well as different pixel intensity distributions. In this paper, a content-aware image augmentation method is presented to deal with these variations. The results of the proposed method have been validated graphically by plotting the removed and added seams of pixels on original images. Two different chest X-ray (CXR) datasets are used in the experiments. The CXRs in the datasets defer in size, some are digital CXRs while the others are digitized from analog CXR films. With the proposed content-aware augmentation method, the Seam Carving algorithm is employed to resize CXRs and the corresponding labels in the form of image masks, followed by histogram matching used to normalize the pixel intensities of digital radiography, based on the pixel intensity values of digitized radiographs. We implemented the algorithms, resized the well-known Montgomery dataset, to the size of the most frequently used Japanese Society of Radiological Technology (JSRT) dataset and normalized our digital CXRs for testing. This work resulted in the unified off-the-shelf CXR dataset composed of radiographs included in both, Montgomery and JSRT datasets. The experimental results show that even though the amount of augmentation is large, our algorithm can preserve the important information in lung fields, local structures, and global visual effect adequately. The proposed method can be used to augment training and testing image data sets so that the trained machine learning model can be used to process CXRs from various sources, and it can be potentially used broadly in any medical imaging applications.

Keywords: computer-aided diagnosis, image augmentation, lung segmentation, medical imaging, seam carving

Procedia PDF Downloads 206
28660 Kinetics, Equilibrium and Thermodynamic Studies on Adsorption of Reactive Blue 29 from Aqueous Solution Using Activated Tamarind Kernel Powder

Authors: E. D. Paul, A. D. Adams, O. Sunmonu, U. S. Ishiaku

Abstract:

Activated tamarind kernel powder (ATKP) was prepared from tamarind fruit (Tamarindus indica), and utilized for the removal of Reactive Blue 29 (RB29) from its aqueous solution. The powder was activated using 4N nitric acid (HNO₃). The adsorbent was characterised using infrared spectroscopy, bulk density, ash content, pH, moisture content and dry matter content measurements. The effect of various parameters which include; temperature, pH, adsorbent dosage, ion concentration, and contact time were studied. Four different equilibrium isotherm models were tested on the experimental data, but the Temkin isotherm model was best-fitted into the experimental data. The pseudo-first order and pseudo-second-order kinetic models were also fitted into the graphs, but pseudo-second order was best fitted to the experimental data. The thermodynamic parameters showed that the adsorption of Reactive Blue 29 onto activated tamarind kernel powder is a physical process, feasible and spontaneous, exothermic in nature and there is decreased randomness at the solid/solution interphase during the adsorption process. Therefore, activated tamarind kernel powder has proven to be a very good adsorbent for the removal of Reactive Blue 29 dyes from industrial waste water.

Keywords: tamarind kernel powder, reactive blue 29, isotherms, kinetics

Procedia PDF Downloads 238
28659 An Information Matrix Goodness-of-Fit Test of the Conditional Logistic Model for Matched Case-Control Studies

Authors: Li-Ching Chen

Abstract:

The case-control design has been widely applied in clinical and epidemiological studies to investigate the association between risk factors and a given disease. The retrospective design can be easily implemented and is more economical over prospective studies. To adjust effects for confounding factors, methods such as stratification at the design stage and may be adopted. When some major confounding factors are difficult to be quantified, a matching design provides an opportunity for researchers to control the confounding effects. The matching effects can be parameterized by the intercepts of logistic models and the conditional logistic regression analysis is then adopted. This study demonstrates an information-matrix-based goodness-of-fit statistic to test the validity of the logistic regression model for matched case-control data. The asymptotic null distribution of this proposed test statistic is inferred. It needs neither to employ a simulation to evaluate its critical values nor to partition covariate space. The asymptotic power of this test statistic is also derived. The performance of the proposed method is assessed through simulation studies. An example of the real data set is applied to illustrate the implementation of the proposed method as well.

Keywords: conditional logistic model, goodness-of-fit, information matrix, matched case-control studies

Procedia PDF Downloads 281
28658 Sentence vs. Keyword Content Analysis in Intellectual Capital Disclosures Study

Authors: Martin Surya Mulyadi, Yunita Anwar, Rosinta Ria Panggabean

Abstract:

Major transformations in economic activity from an agricultural economy to knowledge economy have led to an increasing focus on intellectual capital (IC) that has been characterized by continuous innovation, the spread of digital and communication technologies, intangible and human factors. IC is defined as the possession of knowledge and experience, professional knowledge and skill, proper relationships and technological capacities, which when applied will give organizations a competitive advantage. All of IC report/disclosure could be captured from the corporate annual report as it is a communication device that allows a corporation to connect with various external and internal stakeholders. This study was conducted using sentence-content analysis of IC disclosure in the annual report. This research aims to analyze whether the keyword-content analysis is reliable research methodology for IC disclosure related research.

Keywords: intellectual capital, intellectual capital disclosure, content analysis, annual report, sentence analysis, keyword analysis

Procedia PDF Downloads 355
28657 Effect of Thermal Treatment on Phenolic Content, Antioxidant, and Alpha-Amylase Inhibition Activities of Moringa stenopetala Leaves

Authors: Daniel Assefa, Engeda Dessalegn, Chetan Chauhan

Abstract:

Moringa stenopetala is a socioeconomic valued tree that is widely available and cultivated in the Southern part of Ethiopia. The leaves have been traditionally used as a food source with high nutritional and medicinal values. The present work was carried out to evaluate the effect of thermal treatment on the total phenolic content, antioxidant and alpha-amylase inhibition activities of aqueous leaf extracts during maceration and different decoction time interval (5, 10 and 15 min). The total phenolic content was determined by the Folin-ciocalteu methods whereas antioxidant activities were determined by 2,2-diphenyl-1-picryl-hydrazyl(DPPH) radical scavenging, reducing power and ferrous ion chelating assays and alpha-amylase inhibition activity was determined using 3,5-dinitrosalicylic acid method. Total phenolic content ranged from 34.35 to 39.47 mgGAE/g. Decoction for 10 min extract showed ferrous ion chelating (92.52), DPPH radical scavenging (91.52%), alpha-amylase inhibition (69.06%) and ferric reducing power (0.765), respectively. DPPH, reducing power and alpha-amylase inhibition activities showed positive linear correlation (R2=0.853, R2= 0.857 and R2=0.930), respectively with total phenolic content but ferrous ion chelating activity was found to be weakly correlated (R2=0.481). Based on the present investigation, it could be concluded that major loss of total phenolic content, antioxidant and alpha-amylase inhibition activities of the crude leaf extracts of Moringa stenopetala leaves were observed at decoction time for 15 min. Therefore, to maintain the total phenolic content, antioxidant, and alpha-amylase inhibition activities of leaves, cooking practice should be at the optimum decoction time (5-10 min).

Keywords: alpha-amylase inhibition, antioxidant, Moringa stenopetala, total phenolic content

Procedia PDF Downloads 345
28656 Multimodal Content: Fostering Students’ Language and Communication Competences

Authors: Victoria L. Malakhova

Abstract:

The research is devoted to multimodal content and its effectiveness in developing students’ linguistic and intercultural communicative competences as an indefeasible constituent of their future professional activity. Description of multimodal content both as a linguistic and didactic phenomenon makes the study relevant. The objective of the article is the analysis of creolized texts and the effect they have on fostering higher education students’ skills and their productivity. The main methods used are linguistic text analysis, qualitative and quantitative methods, deduction, generalization. The author studies texts with full and partial creolization, their features and role in composing multimodal textual space. The main verbal and non-verbal markers and paralinguistic means that enhance the linguo-pragmatic potential of creolized texts are covered. To reveal the efficiency of multimodal content application in English teaching, the author conducts an experiment among both undergraduate students and teachers. This allows specifying main functions of creolized texts in the process of language learning, detecting ways of enhancing students’ competences, and increasing their motivation. The described stages of using creolized texts can serve as an algorithm for work with multimodal content in teaching English as a foreign language. The findings contribute to improving the efficiency of the academic process.

Keywords: creolized text, English language learning, higher education, language and communication competences, multimodal content

Procedia PDF Downloads 104
28655 COVID-19 Case: A Definition of Infodemia through Online Italian Journalism

Authors: Concetta Papapicco

Abstract:

The spreading of new Coronavirus (COVID-19) in addition to becoming a global phenomenon, following the declaration of a pandemic state, has generated excessive access to information, sometimes not thoroughly screened, which makes it difficult to navigate a given topic because of the difficulty of finding reliable sources. As a result, there is a high level of contagion, understood as the spread of the virus, but also as the spread of information in a viral and harmful way, which prompted the World Health Organization to coin the term Infodemia to give 'a name' the phenomenon of excessive information. With neologism 'Infodemia', the World Health Organization (OMS) wanted, in these days when fear of the coronavirus is raging, point out that perhaps the greatest danger of global society in the age of social media. This phenomenon is the distortion of reality in the rumble of echoes and comments of the global community on real or often invented facts. The general purpose of the exploratory study is to investigate how the coronavirus situation is described from journalistic communication. Starting from La Repubblica online, as a reference journalistic magazine, as a specific objective, the research aims to understand the way in which journalistic communication describes the phenomenon of the COVID-19 virus spread, the spread of contagion and restrictive measures of social distancing in the Italian context. The study starts from the hypothesis that if the circulation of information helps to create a social representation of the phenomenon, the excessive accessibility to sources of information (Infodemia) can be modulated by the 'how' the phenomenon is described by the journalists. The methodology proposed, in fact, in the exploratory study is a quanti-qualitative (mixed) method. A Content Analysis with the SketchEngine software is carried out first. In support of the Content Analysis, a Diatextual Analysis was carried out. The Diatextual Analysis is a qualitative analysis useful to detect in the analyzed texts, that is the online articles of La Repubblica on the topic of coronavirus, Subjectivity, Argomentativity, and Mode. The research focuses mainly on 'Mode' or 'How' are the events related to coronavirus in the online articles of La Repubblica about COVID-19 phenomenon. The results show the presence of the contrast vision about COVID-19 situation in Italy.

Keywords: coronavirus, Italian infodemia, La Republica online, mix method

Procedia PDF Downloads 116
28654 Modeling of Tsunami Propagation and Impact on West Vancouver Island, Canada

Authors: S. Chowdhury, A. Corlett

Abstract:

Large tsunamis strike the British Columbia coast every few hundred years. The Cascadia Subduction Zone, which extends along the Pacific coast from Vancouver Island to Northern California is one of the most seismically active regions in Canada. Significant earthquakes have occurred in this region, including the 1700 Cascade Earthquake with an estimated magnitude of 9.2. Based on geological records, experts have predicted a 'great earthquake' of a similar magnitude within this region may happen any time. This earthquake is expected to generate a large tsunami that could impact the coastal communities on Vancouver Island. Since many of these communities are in remote locations, they are more likely to be vulnerable, as the post-earthquake relief efforts would be impacted by the damage to critical road infrastructures. To assess the coastal vulnerability within these communities, a hydrodynamic model has been developed using MIKE-21 software. We have considered a 500 year probabilistic earthquake design criteria including the subsidence in this model. The bathymetry information was collected from Canadian Hydrographic Services (CHS), and National Oceanic Atmospheric and Administration (NOAA). The arial survey was conducted using a Cessna-172 aircraft for the communities, and then the information was converted to generate a topographic digital elevation map. Both survey information was incorporated into the model, and the domain size of the model was about 1000km x 1300km. This model was calibrated with the tsunami occurred off the west coast of Moresby Island on October 28, 2012. The water levels from the model were compared with two tide gauge stations close to the Vancouver Island and the output from the model indicates the satisfactory result. For this study, the design water level was considered as High Water Level plus the Sea Level Rise for 2100 year. The hourly wind speeds from eight directions were collected from different wind stations and used a 200-year return period wind speed in the model for storm events. The regional model was set for 12 hrs simulation period, which takes more than 16 hrs to complete one simulation using double Xeon-E7 CPU computer plus a K-80 GPU. The boundary information for the local model was generated from the regional model. The local model was developed using a high resolution mesh to estimate the coastal flooding for the communities. It was observed from this study that many communities will be effected by the Cascadia tsunami and the inundation maps were developed for the communities. The infrastructures inside the coastal inundation area were identified. Coastal vulnerability planning and resilient design solutions will be implemented to significantly reduce the risk.

Keywords: tsunami, coastal flooding, coastal vulnerable, earthquake, Vancouver, wave propagation

Procedia PDF Downloads 125
28653 Methodology for Obtaining Static Alignment Model

Authors: Lely A. Luengas, Pedro R. Vizcaya, Giovanni Sánchez

Abstract:

In this paper, a methodology is presented to obtain the Static Alignment Model for any transtibial amputee person. The proposed methodology starts from experimental data collected on the Hospital Militar Central, Bogotá, Colombia. The effects of transtibial prosthesis malalignment on amputees were measured in terms of joint angles, center of pressure (COP) and weight distribution. Some statistical tools are used to obtain the model parameters. Mathematical predictive models of prosthetic alignment were created. The proposed models are validated in amputees and finding promising results for the prosthesis Static Alignment. Static alignment process is unique to each subject; nevertheless the proposed methodology can be used in each transtibial amputee.

Keywords: information theory, prediction model, prosthetic alignment, transtibial prosthesis

Procedia PDF Downloads 249
28652 Implementation and Validation of a Damage-Friction Constitutive Model for Concrete

Authors: L. Madouni, M. Ould Ouali, N. E. Hannachi

Abstract:

Two constitutive models for concrete are available in ABAQUS/Explicit, the Brittle Cracking Model and the Concrete Damaged Plasticity Model, and their suitability and limitations are well known. The aim of the present paper is to implement a damage-friction concrete constitutive model and to evaluate the performance of this model by comparing the predicted response with experimental data. The constitutive formulation of this material model is reviewed. In order to have consistent results, the parameter identification and calibration for the model have been performed. Several numerical simulations are presented in this paper, whose results allow for validating the capability of the proposed model for reproducing the typical nonlinear performances of concrete structures under different monotonic and cyclic load conditions. The results of the evaluation will be used for recommendations concerning the application and further improvements of the investigated model.

Keywords: Abaqus, concrete, constitutive model, numerical simulation

Procedia PDF Downloads 354
28651 Email Phishing Detection Using Natural Language Processing and Convolutional Neural Network

Authors: M. Hilani, B. Nassih

Abstract:

Phishing is one of the oldest and best known scams on the Internet. It can be defined as any type of telecommunications fraud that uses social engineering tricks to obtain confidential data from its victims. It’s a cybercrime aimed at stealing your sensitive information. Phishing is generally done via private email, so scammers impersonate large companies or other trusted entities to encourage victims to voluntarily provide information such as login credentials or, worse yet, credit card numbers. The COVID-19 theme is used by cybercriminals in multiple malicious campaigns like phishing. In this environment, messaging filtering solutions have become essential to protect devices that will now be used outside of the secure perimeter. Despite constantly updating methods to avoid these cyberattacks, the end result is currently insufficient. Many researchers are looking for optimal solutions to filter phishing emails, but we still need good results. In this work, we concentrated on solving the problem of detecting phishing emails using the different steps of NLP preprocessing, and we proposed and trained a model using one-dimensional CNN. Our study results show that our model obtained an accuracy of 99.99%, which demonstrates how well our model is working.

Keywords: phishing, e-mail, NLP preprocessing, CNN, e-mail filtering

Procedia PDF Downloads 109
28650 Efficient Fake News Detection Using Machine Learning and Deep Learning Approaches

Authors: Chaima Babi, Said Gadri

Abstract:

The rapid increase in fake news continues to grow at a very fast rate; this requires implementing efficient techniques that allow testing the re-liability of online content. For that, the current research strives to illuminate the fake news problem using deep learning DL and machine learning ML ap-proaches. We have developed the traditional LSTM (Long short-term memory), and the bidirectional BiLSTM model. A such process is to perform a training task on almost of samples of the dataset, validate the model on a subset called the test set to provide an unbiased evaluation of the final model fit on the training dataset, then compute the accuracy of detecting classifica-tion and comparing the results. For the programming stage, we used Tensor-Flow and Keras libraries on Python to support Graphical Processing Units (GPUs) that are being used for developing deep learning applications.

Keywords: machine learning, deep learning, natural language, fake news, Bi-LSTM, LSTM, multiclass classification

Procedia PDF Downloads 77
28649 Forecasting Stock Prices Based on the Residual Income Valuation Model: Evidence from a Time-Series Approach

Authors: Chen-Yin Kuo, Yung-Hsin Lee

Abstract:

Previous studies applying residual income valuation (RIV) model generally use panel data and single-equation model to forecast stock prices. Unlike these, this paper uses Taiwan longitudinal data to estimate multi-equation time-series models such as Vector Autoregressive (VAR), Vector Error Correction Model (VECM), and conduct out-of-sample forecasting. Further, this work assesses their forecasting performance by two instruments. In favor of extant research, the major finding shows that VECM outperforms other three models in forecasting for three stock sectors over entire horizons. It implies that an error correction term containing long-run information contributes to improve forecasting accuracy. Moreover, the pattern of composite shows that at longer horizon, VECM produces the greater reduction in errors, and performs substantially better than VAR.

Keywords: residual income valuation model, vector error correction model, out of sample forecasting, forecasting accuracy

Procedia PDF Downloads 309
28648 Process Assessment Model for Process Capability Determination Based on ISO/IEC 20000-1:2011

Authors: Harvard Najoan, Sarwono Sutikno, Yusep Rosmansyah

Abstract:

Most enterprises are now using information technology services as their assets to support business objectives. These kinds of services are provided by the internal service provider (inside the enterprise) or external service provider (outside enterprise). To deliver quality information technology services, the service provider (which from now on will be called ‘organization’) either internal or external, must have a standard for service management system. At present, the standard that is recognized as best practice for service management system for the organization is international standard ISO/IEC 20000:2011. The most important part of this international standard is the first part or ISO/IEC 20000-1:2011-Service Management System Requirement, because it contains 22 for organization processes as a requirement to be implemented in an organizational environment in order to build, manage and deliver quality service to the customer. Assessing organization management processes is the first step to implementing ISO/IEC 20000:2011 into the organization management processes. This assessment needs Process Assessment Model (PAM) as an assessment instrument. PAM comprises two parts: Process Reference Model (PRM) and Measurement Framework (MF). PRM is built by transforming the 22 process of ISO/IEC 20000-1:2011 and MF is based on ISO/IEC 33020. This assessment instrument was designed to assess the capability of service management process in Divisi Teknologi dan Sistem Informasi (Information Systems and Technology Division) as an internal organization of PT Pos Indonesia. The result of this assessment model can be proposed to improve the capability of service management system.

Keywords: ISO/IEC 20000-1:2011, ISO/IEC 33020:2015, process assessment, process capability, service management system

Procedia PDF Downloads 454
28647 A New Distribution and Application on the Lifetime Data

Authors: Gamze Ozel, Selen Cakmakyapan

Abstract:

We introduce a new model called the Marshall-Olkin Rayleigh distribution which extends the Rayleigh distribution using Marshall-Olkin transformation and has increasing and decreasing shapes for the hazard rate function. Various structural properties of the new distribution are derived including explicit expressions for the moments, generating and quantile function, some entropy measures, and order statistics are presented. The model parameters are estimated by the method of maximum likelihood and the observed information matrix is determined. The potentiality of the new model is illustrated by means of real life data set.

Keywords: Marshall-Olkin distribution, Rayleigh distribution, estimation, maximum likelihood

Procedia PDF Downloads 493
28646 Band Structure Computation of GaMnAs Using the Multiband k.p Theory

Authors: Khadijah B. Alziyadi, Khawlh A. Alzubaidi, Amor M. Alsayari

Abstract:

Recently, GaMnAs diluted magnetic semiconductors(DMSs) have received considerable attention because they combine semiconductor and magnetic properties. GaMnAs has been used as a model DMS and as a test bed for many concepts and functionalities of spintronic devices. In this paper, a theoretical study on the band structure ofGaMnAswill be presented. The model that we used in this study is the 8-band k.p methodwherespin-orbit interaction, spin splitting, and strain are considered. The band structure of GaMnAs will be calculated in different directions in the reciprocal space. The effect of manganese content on the GaMnAs band structure will be discussed. Also, the influence of strain, which varied continuously from tensile to compressive, on the different bands will be studied.

Keywords: band structure, diluted magnetic semiconductor, k.p method, strain

Procedia PDF Downloads 135
28645 Dynamic Determination of Spare Engine Requirements for Air Fighters Integrating Feedback of Operational Information

Authors: Tae Bo Jeon

Abstract:

Korean air force is undertaking a big project to replace prevailing hundreds of old air fighters such as F-4, F-5, KF-16 etc. The task is to develop and produce domestic fighters equipped with 2 complete-type engines each. A large number of engines, however, will be purchased as products from a foreign engine maker. In addition to the fighters themselves, secure the proper number of spare engines serves a significant role in maintaining combat readiness and effectively managing the national defense budget due to high cost. In this paper, we presented a model dynamically updating spare engine requirements. Currently, the military administration purchases all the fighters, engines, and spare engines at acquisition stage and does not have additional procurement processes during the life cycle, 30-40 years. With the assumption that procurement procedure during the operational stage is established, our model starts from the initial estimate of spare engine requirements based on limited information. The model then performs military missions and repair/maintenance works when necessary. During operation, detailed field information - aircraft repair and test, engine repair, planned maintenance, administration time, transportation pipeline between base, field, and depot etc., - should be considered for actual engine requirements. At the end of each year, the performance measure is recorded and proceeds to next year when it shows higher the threshold set. Otherwise, additional engine(s) will be bought and added to the current system. We repeat the process for the life cycle period and compare the results. The proposed model is seen to generate far better results appropriately adding spare engines thus avoiding possible undesirable situations. Our model may well be applied to future air force military operations.

Keywords: DMSMS, operational availability, METRIC, PRS

Procedia PDF Downloads 161
28644 Neural Style Transfer Using Deep Learning

Authors: Shaik Jilani Basha, Inavolu Avinash, Alla Venu Sai Reddy, Bitragunta Taraka Ramu

Abstract:

We can use the neural style transfer technique to build a picture with the same "content" as the beginning image but the "style" of the picture we've chosen. Neural style transfer is a technique for merging the style of one image into another while retaining its original information. The only change is how the image is formatted to give it an additional artistic sense. The content image depicts the plan or drawing, as well as the colors of the drawing or paintings used to portray the style. It is a computer vision programme that learns and processes images through deep convolutional neural networks. To implement software, we used to train deep learning models with the train data, and whenever a user takes an image and a styled image, the output will be as the style gets transferred to the original image, and it will be shown as the output.

Keywords: neural networks, computer vision, deep learning, convolutional neural networks

Procedia PDF Downloads 76
28643 An Investigation of Influential Factors in Adopting the Cloud Computing in Saudi Arabia: An Application of Technology Acceptance Model

Authors: Shayem Saleh ALresheedi, Lu Song Feng, Abdulaziz Abdulwahab M. Fatani

Abstract:

Cloud computing is an emerging concept in the technological sphere. Its development enables many applications to avail information online and on demand. It is becoming an essential element for businesses due to its ability to diminish the costs of IT infrastructure and is being adopted in Saudi Arabia. However, there exist many factors that affect its adoption. Several researchers in the field have ignored the study of the TAM model for identifying the relevant factors and their impact for adopting of cloud computing. This study focuses on evaluating the acceptability of cloud computing and analyzing its impacting factors using Technology Acceptance Model (TAM) of technology adoption in Saudi Arabia. It suggests a model to examine the influential factors of the TAM model along with external factors of technical support in adapting the cloud computing. The proposed model has been tested through the use of multiple hypotheses based on calculation tools and collected data from customers through questionnaires. The findings of the study prove that the TAM model along with external factors can be applied in measuring the expected adoption of cloud computing. The study presents an investigation of influential factors and further recommendation in adopting cloud computing in Saudi Arabia.

Keywords: cloud computing, acceptability, adoption, determinants

Procedia PDF Downloads 182
28642 Modeling of Building a Conceptual Scheme for Multimodal Freight Transportation Information System

Authors: Gia Surguladze, Nino Topuria, Lily Petriashvili, Giorgi Surguladze

Abstract:

Modeling of building processes of a multimodal freight transportation support information system is discussed based on modern CASE technologies. Functional efficiencies of ports in the eastern part of the Black Sea are analyzed taking into account their ecological, seasonal, resource usage parameters. By resources, we mean capacities of berths, cranes, automotive transport, as well as work crews and neighbouring airports. For the purpose of designing database of computer support system for Managerial (Logistics) function, using Object-Role Modeling (ORM) tool (NORMA – Natural ORM Architecture) is proposed, after which Entity Relationship Model (ERM) is generated in automated process. The software is developed based on Process-Oriented and Service-Oriented architecture, in Visual Studio.NET environment.

Keywords: seaport resources, business-processes, multimodal transportation, CASE technology, object-role model, entity relationship model, SOA

Procedia PDF Downloads 417
28641 Tool for Metadata Extraction and Content Packaging as Endorsed in OAIS Framework

Authors: Payal Abichandani, Rishi Prakash, Paras Nath Barwal, B. K. Murthy

Abstract:

Information generated from various computerization processes is a potential rich source of knowledge for its designated community. To pass this information from generation to generation without modifying the meaning is a challenging activity. To preserve and archive the data for future generations it’s very essential to prove the authenticity of the data. It can be achieved by extracting the metadata from the data which can prove the authenticity and create trust on the archived data. Subsequent challenge is the technology obsolescence. Metadata extraction and standardization can be effectively used to resolve and tackle this problem. Metadata can be categorized at two levels i.e. Technical and Domain level broadly. Technical metadata will provide the information that can be used to understand and interpret the data record, but only this level of metadata isn’t sufficient to create trustworthiness. We have developed a tool which will extract and standardize the technical as well as domain level metadata. This paper is about the different features of the tool and how we have developed this.

Keywords: digital preservation, metadata, OAIS, PDI, XML

Procedia PDF Downloads 383
28640 The Asymmetric Proximal Support Vector Machine Based on Multitask Learning for Classification

Authors: Qing Wu, Fei-Yan Li, Heng-Chang Zhang

Abstract:

Multitask learning support vector machines (SVMs) have recently attracted increasing research attention. Given several related tasks, the single-task learning methods trains each task separately and ignore the inner cross-relationship among tasks. However, multitask learning can capture the correlation information among tasks and achieve better performance by training all tasks simultaneously. In addition, the asymmetric squared loss function can better improve the generalization ability of the models on the most asymmetric distributed data. In this paper, we first make two assumptions on the relatedness among tasks and propose two multitask learning proximal support vector machine algorithms, named MTL-a-PSVM and EMTL-a-PSVM, respectively. MTL-a-PSVM seeks a trade-off between the maximum expectile distance for each task model and the closeness of each task model to the general model. As an extension of the MTL-a-PSVM, EMTL-a-PSVM can select appropriate kernel functions for shared information and private information. Besides, two corresponding special cases named MTL-PSVM and EMTLPSVM are proposed by analyzing the asymmetric squared loss function, which can be easily implemented by solving linear systems. Experimental analysis of three classification datasets demonstrates the effectiveness and superiority of our proposed multitask learning algorithms.

Keywords: multitask learning, asymmetric squared loss, EMTL-a-PSVM, classification

Procedia PDF Downloads 101