Search results for: Bayesian information criterion
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11219

Search results for: Bayesian information criterion

10889 e-Learning Security: A Distributed Incident Response Generator

Authors: Bel G Raggad

Abstract:

An e-Learning setting is a distributed computing environment where information resources can be connected to any public network. Public networks are very unsecure which can compromise the reliability of an e-Learning environment. This study is only concerned with the intrusion detection aspect of e-Learning security and how incident responses are planned. The literature reported great advances in intrusion detection system (ids) but neglected to study an important ids weakness: suspected events are detected but an intrusion is not determined because it is not defined in ids databases. We propose an incident response generator (DIRG) that produces incident responses when the working ids system suspects an event that does not correspond to a known intrusion. Data involved in intrusion detection when ample uncertainty is present is often not suitable to formal statistical models including Bayesian. We instead adopt Dempster and Shafer theory to process intrusion data for the unknown event. The DIRG engine transforms data into a belief structure using incident scenarios deduced by the security administrator. Belief values associated with various incident scenarios are then derived and evaluated to choose the most appropriate scenario for which an automatic incident response is generated. This article provides a numerical example demonstrating the working of the DIRG system.

Keywords: decision support system, distributed computing, e-Learning security, incident response, intrusion detection, security risk, statefull inspection

Procedia PDF Downloads 402
10888 Information Seekers vs. Information Providers: New Vistas and New Challenges for the Libraries Today; A Case Study of the Panjab University Library, Chandigarh, India

Authors: Neeru Bhatia

Abstract:

This article presents the results of a case study designed to analyze and deduce Information seekers and the Information Providers in today’s context, wherein we come across a sea of change in the provision of Information services due to the changing electronic environment. The Panjab University Library is one of the biggest libraries of India and was inaugurated in 1963 by Pt. Jawaharlal Nehru, the then Prime Minister of India. The library always thrives to assimilate new technology for the provision of Information services. As we know that the Information seekers today are a whole lot different, they are tech savvy, like to be on their electronic gadgets most of the time, and their Information seeking patterns are also different, the challenge that lies before the libraries is to be always ready for these day to day challenges. The study explores the current status of the Information Services being provided by the Panjab University Library (the Information Providers) vs. the evaluation of these Information services by the users of Library (the Information Seekers). The present study aimed at finding out whether Panjab University Library is able to achieve its mission to be an innovative and user-oriented library by exploring all the new vistas and reach up to the expectations of the information seekers by taking up all the challenges being posed by the ever changing technological scenario.

Keywords: electronic environment, information seekers, information providers, new technology

Procedia PDF Downloads 239
10887 Electronic Resources and Information Literacy in Higher Education Library

Authors: Nirmal Singh, Rajesh Kumar

Abstract:

Abstract- Information literacy aims to develop both critical understanding and active participation in scholars. It enables scholars to interpret and make informed judgments as users of information sources, and it also enables them to become producers of information in their own right, and thereby to become more powerful participants in society. Information literacy is about developing people‘s critical and creative abilities. Digital media – and particularly the Internet – significantly increase the potential for such active participation of the individual, provided scholars have the means and training to effectively access and use them. This paper provides definition, standards and importance of information literacy (IL). Keywords: Information literacy, Digital Media, Training, Communications Technologies.

Keywords: Information literacy, Digital Media, Training, , Communications Technologies

Procedia PDF Downloads 120
10886 Learning the Dynamics of Articulated Tracked Vehicles

Authors: Mario Gianni, Manuel A. Ruiz Garcia, Fiora Pirri

Abstract:

In this work, we present a Bayesian non-parametric approach to model the motion control of ATVs. The motion control model is based on a Dirichlet Process-Gaussian Process (DP-GP) mixture model. The DP-GP mixture model provides a flexible representation of patterns of control manoeuvres along trajectories of different lengths and discretizations. The model also estimates the number of patterns, sufficient for modeling the dynamics of the ATV.

Keywords: Dirichlet processes, gaussian mixture models, learning motion patterns, tracked robots for urban search and rescue

Procedia PDF Downloads 422
10885 A Geographic Information System Mapping Method for Creating Improved Satellite Solar Radiation Dataset Over Qatar

Authors: Sachin Jain, Daniel Perez-Astudillo, Dunia A. Bachour, Antonio P. Sanfilippo

Abstract:

The future of solar energy in Qatar is evolving steadily. Hence, high-quality spatial solar radiation data is of the uttermost requirement for any planning and commissioning of solar technology. Generally, two types of solar radiation data are available: satellite data and ground observations. Satellite solar radiation data is developed by the physical and statistical model. Ground data is collected by solar radiation measurement stations. The ground data is of high quality. However, they are limited to distributed point locations with the high cost of installation and maintenance for the ground stations. On the other hand, satellite solar radiation data is continuous and available throughout geographical locations, but they are relatively less accurate than ground data. To utilize the advantage of both data, a product has been developed here which provides spatial continuity and higher accuracy than any of the data alone. The popular satellite databases: National Solar radiation Data Base, NSRDB (PSM V3 model, spatial resolution: 4 km) is chosen here for merging with ground-measured solar radiation measurement in Qatar. The spatial distribution of ground solar radiation measurement stations is comprehensive in Qatar, with a network of 13 ground stations. The monthly average of the daily total Global Horizontal Irradiation (GHI) component from ground and satellite data is used for error analysis. The normalized root means square error (NRMSE) values of 3.31%, 6.53%, and 6.63% for October, November, and December 2019 were observed respectively when comparing in-situ and NSRDB data. The method is based on the Empirical Bayesian Kriging Regression Prediction model available in ArcGIS, ESRI. The workflow of the algorithm is based on the combination of regression and kriging methods. A regression model (OLS, ordinary least square) is fitted between the ground and NSBRD data points. A semi-variogram is fitted into the experimental semi-variogram obtained from the residuals. The kriging residuals obtained after fitting the semi-variogram model were added to NSRBD data predicted values obtained from the regression model to obtain the final predicted values. The NRMSE values obtained after merging are respectively 1.84%, 1.28%, and 1.81% for October, November, and December 2019. One more explanatory variable, that is the ground elevation, has been incorporated in the regression and kriging methods to reduce the error and to provide higher spatial resolution (30 m). The final GHI maps have been created after merging, and NRMSE values of 1.24%, 1.28%, and 1.28% have been observed for October, November, and December 2019, respectively. The proposed merging method has proven as a highly accurate method. An additional method is also proposed here to generate calibrated maps by using regression and kriging model and further to use the calibrated model to generate solar radiation maps from the explanatory variable only when not enough historical ground data is available for long-term analysis. The NRMSE values obtained after the comparison of the calibrated maps with ground data are 5.60% and 5.31% for November and December 2019 month respectively.

Keywords: global horizontal irradiation, GIS, empirical bayesian kriging regression prediction, NSRDB

Procedia PDF Downloads 65
10884 Spatial Analysis of Flood Vulnerability in Highly Urbanized Area: A Case Study in Taipei City

Authors: Liang Weichien

Abstract:

Without adequate information and mitigation plan for natural disaster, the risk to urban populated areas will increase in the future as populations grow, especially in Taiwan. Taiwan is recognized as the world's high-risk areas, where an average of 5.7 times of floods occur per year should seek to strengthen coherence and consensus in how cities can plan for flood and climate change. Therefore, this study aims at understanding the vulnerability to flooding in Taipei city, Taiwan, by creating indicators and calculating the vulnerability of each study units. The indicators were grouped into sensitivity and adaptive capacity based on the definition of vulnerability of Intergovernmental Panel on Climate Change. The indicators were weighted by using Principal Component Analysis. However, current researches were based on the assumption that the composition and influence of the indicators were the same in different areas. This disregarded spatial correlation that might result in inaccurate explanation on local vulnerability. The study used Geographically Weighted Principal Component Analysis by adding geographic weighting matrix as weighting to get the different main flood impact characteristic in different areas. Cross Validation Method and Akaike Information Criterion were used to decide bandwidth and Gaussian Pattern as the bandwidth weight scheme. The ultimate outcome can be used for the reduction of damage potential by integrating the outputs into local mitigation plan and urban planning.

Keywords: flood vulnerability, geographically weighted principal components analysis, GWPCA, highly urbanized area, spatial correlation

Procedia PDF Downloads 269
10883 A Framework for Information Quality in Accounting Information Systems Adoption

Authors: Wongsim Manirath

Abstract:

In order to implement AIS adoption successfully, it is important to consider the quality of information management and understand Information Quality (IQ) factors influencing AIS adoption. This research aims to explore ways of managing AIS adoption to investigate the adoption of accounting information systems within organisations. The study has led to the development of a framework for understanding the AIS adoption process in an organisation. This research used qualitative, interpretive evidence. This framework was developed from case studies and by collecting qualitative data (interviews). This research has conducted 10 case studies to study how IQ is managed through the accounting information system adoption process. A special focus is placed on determining how organisation size influences the information quality practices. The finding is especially useful to SMEs as many SMEs have the desire to grow bigger. By better dealing with IQ issues, there could be a successful future.

Keywords: data quality, information quality, accounting information system, information management

Procedia PDF Downloads 442
10882 Transformation of Health Communication Literacy in Information Technology during Pandemic in 2019-2022

Authors: K. Y. S. Putri, Heri Fathurahman, Yuki Surisita, Widi Sagita, Kiki Dwi Arviani

Abstract:

Society needs the assistance of academics in understanding and being skilled in health communication literacy. Information technology runs very fast while health communication literacy skills in getting health communication information during the pandemic are not as fast as the development of information technology. The research question is whether there is an influence of health communication on information technology in health information during the pandemic in Indonesia. The purpose of the study is to find out the influence of health communication on information technology in health information during the pandemic in Indonesia. The concepts of health communication literacy and information technology are used this study. Previous research is in support of this study. Quantitative research methods by disseminating questionnaires in this study. The validity and reliability test of this study is positive, so it can proceed to the next statistical analysis. Descriptive results of variable health communication literacy are of positive value in all dimensions. All dimensions of information technology are of positive value. Statistical tests of the influence of health communication literacy on information technology are of great value. Discussion of both variables in the influence of health communication literacy and high-value information technology because health communication literacy has a high effect in information technology. Respondents to this study have high information technology skills. So that health communication literacy in obtaining health information during the 2019-2022 pandemic is needed. Research advice is that academics are still very much needed by the community in the development of society during the pandemic.

Keywords: health information, health information needs, literacy health communication, information technology

Procedia PDF Downloads 104
10881 Failure Inference and Optimization for Step Stress Model Based on Bivariate Wiener Model

Authors: Soudabeh Shemehsavar

Abstract:

In this paper, we consider the situation under a life test, in which the failure time of the test units are not related deterministically to an observable stochastic time varying covariate. In such a case, the joint distribution of failure time and a marker value would be useful for modeling the step stress life test. The problem of accelerating such an experiment is considered as the main aim of this paper. We present a step stress accelerated model based on a bivariate Wiener process with one component as the latent (unobservable) degradation process, which determines the failure times and the other as a marker process, the degradation values of which are recorded at times of failure. Parametric inference based on the proposed model is discussed and the optimization procedure for obtaining the optimal time for changing the stress level is presented. The optimization criterion is to minimize the approximate variance of the maximum likelihood estimator of a percentile of the products’ lifetime distribution.

Keywords: bivariate normal, Fisher information matrix, inverse Gaussian distribution, Wiener process

Procedia PDF Downloads 299
10880 Effect of Anisotropy on Steady Creep in a Whisker Reinforced Functionally Graded Composite Disc

Authors: V. K. Gupta, Tejeet Singh

Abstract:

In many whisker reinforced composites, anisotropy may result due to material flow during processing operations such as forging, extrusion etc. The consequence of anisotropy, introduced during processing of disc material, has been investigated on the steady state creep deformations of the rotating disc. The disc material is assumed to undergo plastic deformations according to Hill’s anisotropic criterion. Steady state creep has been analyzed in a constant thickness rotating disc made of functionally graded 6061Al-SiCw (where the subscript ‘w’ stands for whisker) using Hill’s The content of reinforcement (SiCw) in the disc is assumed to decrease linearly from the inner to outer radius. The stresses and strain rates in the disc are estimated by solving the force equilibrium equation along with the constitutive equations describing multi-axial creep. The results obtained for anisotropic FGM disc have been compared with those estimated for isotropic FGM disc having the same average whisker content. The anisotropic constants, appearing in Hill’s yield criterion, have been obtained from the available experimental results. The results show that the presence of anisotropy reduces the tangential stress in the middle of the disc but near the inner and outer radii the tangential stress is higher when compared to isotropic disc. On the other hand, the steady state creep rates in the anisotropic disc are reduced significantly over the entire disc radius, with the maximum reduction observed at the inner radius. Further, in the presence of anisotropy the distribution of strain rate becomes relatively uniform over the entire disc, which may be responsible for reducing the extent of distortion in the disc.

Keywords: anisotropy, creep, functionally graded composite, rotating disc

Procedia PDF Downloads 364
10879 Supervised-Component-Based Generalised Linear Regression with Multiple Explanatory Blocks: THEME-SCGLR

Authors: Bry X., Trottier C., Mortier F., Cornu G., Verron T.

Abstract:

We address component-based regularization of a Multivariate Generalized Linear Model (MGLM). A set of random responses Y is assumed to depend, through a GLM, on a set X of explanatory variables, as well as on a set T of additional covariates. X is partitioned into R conceptually homogeneous blocks X1, ... , XR , viewed as explanatory themes. Variables in each Xr are assumed many and redundant. Thus, Generalised Linear Regression (GLR) demands regularization with respect to each Xr. By contrast, variables in T are assumed selected so as to demand no regularization. Regularization is performed searching each Xr for an appropriate number of orthogonal components that both contribute to model Y and capture relevant structural information in Xr. We propose a very general criterion to measure structural relevance (SR) of a component in a block, and show how to take SR into account within a Fisher-scoring-type algorithm in order to estimate the model. We show how to deal with mixed-type explanatory variables. The method, named THEME-SCGLR, is tested on simulated data.

Keywords: Component-Model, Fisher Scoring Algorithm, GLM, PLS Regression, SCGLR, SEER, THEME

Procedia PDF Downloads 375
10878 Elaboration Development Strategy and the Analysis of Trends Shaping the Information Economy in Azerbaijan on the Basis of the Experience of Foreign Countries

Authors: Rasim M. Alguliyev, Alovsat G. Aliyev

Abstract:

In the paper information on economic development trends in developed countries are analyzed. The current status of information society and economy of the country is reviewed and some recommendations are given for future development. The problems of Information Society and establishment of its innovative economy are studied. In this turn, development trends information economy in developed countries are analyzed.

Keywords: information economy, ICT sector, ICT infrastructure, innovation, innovation system hi-tech products, antimonopoly policy

Procedia PDF Downloads 315
10877 Logistics Information and Customer Service

Authors: Š. Čemerková, M. Wilczková

Abstract:

The paper deals with the importance of information flow for providing of defined level of customer service in the firms. Setting of the criteria for the selection and implementation of logistics information system is a prerequisite for ensuring of the flow of information in firms. The decision on the selection and implementation of logistics information system is linked to the investment costs and operating costs, which are included in the total logistics costs. The article also deals with the conclusions of the research focused on the logistics information system selection in companies in the Czech Republic.

Keywords: customer service, information system, logistics, research

Procedia PDF Downloads 335
10876 Conceptual Model for Logistics Information System

Authors: Ana María Rojas Chaparro, Cristian Camilo Sarmiento Chaves

Abstract:

Given the growing importance of logistics as a discipline for efficient management of materials flow and information, the adoption of tools that permit to create facilities in making decisions based on a global perspective of the system studied has been essential. The article shows how from a concepts-based model is possible to organize and represent in appropriate way the reality, showing accurate and timely information, features that make this kind of models an ideal component to support an information system, recognizing that information as relevant to establish particularities that allow get a better performance about the evaluated sector.

Keywords: system, information, conceptual model, logistics

Procedia PDF Downloads 467
10875 Investigation into the Suitability of Aggregates for Use in Superpave Design Method

Authors: Ahmad Idris, Armaya`u Suleiman Labo, Ado Yusuf Abdulfatah, Murtala Umar

Abstract:

Super pave is the short form of Superior Performing Asphalt Pavement and represents a basis for specifying component materials, asphalt mixture design and analysis, and pavement performance prediction. This new technology is the result of long research projects conducted by the strategic Highway Research program (SHRP) of the Federal Highway Administration. This research was aimed at examining the suitability of Aggregates found in Kano for used in super pave design method. Aggregates samples were collected from different sources in Kano Nigeria and their Engineering properties, as they relate to the SUPERPAVE design requirements were determined. The average result of Coarse Aggregate Angularity in Kano was found to be 87% and 86% of one fractured face and two or more fractured faces respectively with a standard of 80% and 85% respectively. Fine Aggregate Angularity average result was found to be 47% with a requirement of 45% minimum. A flat and elongated particle which was found to be 10% has a maximum criterion of 10%. Sand equivalent was found to be 51% with the criteria of 45% minimum. Strength tests were also carried out, and the results reflect the requirements of the standards. The tests include Impact value test, Aggregate crushing value and Aggregate Abrasion tests and the results are 27.5%, 26.7% and 13% respectively with a maximum criteria of 30%. Specific gravity was also carried out and the result was found to have an average value of 2.52 with a criterion of 2.6 to 2.9 and Water absorption was found to be 1.41% with maximum criteria of 0.6%. From the study, the result of the tests indicated that the aggregates properties have met the requirements of Super pave design method based on the specifications of ASTMD 5821, ASTM D 4791, AASHTO T176, AASHTO T33 and BS815.

Keywords: aggregates, construction, road design, super pave

Procedia PDF Downloads 216
10874 The Information-Seeking Behaviour of Kuwaiti Judges (KJs)

Authors: Essam Mansour

Abstract:

The key purpose of this study is to show information-seeking behaviour of Kuwaiti Judges (KJs). Being one of the few studies about the information needs and information-seeking behaviour conducted in Arab and developing countries, this study is a pioneer one among many studies conducted in information seeking, especially with this significant group of information users. The authors tried to investigate this seeking behavior in terms of KJs' thoughts, perceptions, motivations, techniques, preferences, tools and barriers met when seeking information. The authors employed a questionnaire, with a response rate 77.2 percent. This study showed that most of KJs were likely to be older, educated and with a work experience ranged from new to old experience. There is a statistically reliable significant difference between KJs' demographic characteristics and some sources of information, such as books, encyclopedias, references and mass media. KJs were using information moderately to make a decision, to be in line with current events, to collect statistics and to make a specific/general research. The office and home were the most frequent location KJs were accessing information from. KJs' efficiency level of the English language is described to be moderately good, and a little number of them confirmed that their efficiency level of French was not bad. The assistance provided by colleagues, followed by consultants, translators, sectaries and librarians were found to be most strong types of assistance needed when seeking information. Mobile apps, followed by PCs, information networks (the Internet) and information databases were the highest technology tool used by KJs. Printed materials, followed by non-printed and audiovisual materials were the most preferred information formats KJs use. The use of languages, the recency of information and the place of information, the deficit role of the library to deliver information were at least significant barriers to KJs when seeking information.

Keywords: information users, information-seeking behaviour, information needs, judges, Kuwait

Procedia PDF Downloads 282
10873 Investigation of Information Security Incident Management Based on International Standard ISO/IEC 27002 in Educational Hospitals in 2014

Authors: Nahid Tavakoli, Asghar Ehteshami, Akbar Hassanzadeh, Fatemeh Amini

Abstract:

Introduction: The Information security incident management guidelines was been developed to help hospitals to meet their information security event and incident management requirements. The purpose of this Study was to investigate on Information Security Incident Management in Isfahan’s educational hospitals in accordance to ISO/IEC 27002 standards. Methods: This was a cross-sectional study to investigate on Information Security Incident Management of educational hospitals in 2014. Based on ISO/IEC 27002 standards, two checklists were applied to check the compliance with standards on Reporting Information Security Events and Weakness and Management of Information Security Incidents and Improvements. One inspector was trained to carry out the assessments in the hospitals. The data was analyzed by SPSS. Findings: In general the score of compliance Information Security Incident Management requirements in two steps; Reporting Information Security Events and Weakness and Management of Information Security Incidents and Improvements was %60. There was the significant difference in various compliance levels among the hospitals (p-valueKeywords: information security incident management, information security management, standards, hospitals

Procedia PDF Downloads 550
10872 The Generalized Pareto Distribution as a Model for Sequential Order Statistics

Authors: Mahdy ‎Esmailian, Mahdi ‎Doostparast, Ahmad ‎Parsian

Abstract:

‎In this article‎, ‎sequential order statistics (SOS) censoring type II samples coming from the generalized Pareto distribution are considered‎. ‎Maximum likelihood (ML) estimators of the unknown parameters are derived on the basis of the available multiple SOS data‎. ‎Necessary conditions for existence and uniqueness of the derived ML estimates are given‎. Due to complexity in the proposed likelihood function‎, ‎a useful re-parametrization is suggested‎. ‎For illustrative purposes‎, ‎a Monte Carlo simulation study is conducted and an illustrative example is analysed‎.

Keywords: bayesian estimation‎, generalized pareto distribution‎, ‎maximum likelihood estimation‎, sequential order statistics

Procedia PDF Downloads 477
10871 The Implementation of Information Security Audits in Public Sector: Perspective from Indonesia

Authors: Nur Imroatun Sholihat, Gresika Bunga Sylvana

Abstract:

Currently, cyber attack became an incredibly serious problem due to its increasing trend all over the world. Therefore, information security becomes prominent for every organization including public sector organization. In Indonesia, unfortunately, Ministry of Finance (MoF) is the only public sector organization that has already formally established procedure to assess its information security adequacy by performing information security audits (November 2017). We assess the implementation of information security audits in the MoF using qualitative data obtained by interviewing IT auditors and by analysis of related documents. For this reason, information security audit practice in the MoF could become the acceptable benchmark for all other public sector organizations in Indonesia. This study is important because, to the best of the author’s knowledge, our research into information security audits practice in Indonesia’s public sector have not been found yet. Results showed that information security audits performed mostly by doing pentest (penetration testing) to MoF’s critical applications.

Keywords: information security audit, information technology, Ministry of Finance of Indonesia, public sector organization

Procedia PDF Downloads 200
10870 A Secure System for Handling Information from Heterogeous Sources

Authors: Shoohira Aftab, Hammad Afzal

Abstract:

Information integration is a well known procedure to provide consolidated view on sets of heterogeneous information sources. It not only provides better statistical analysis of information but also facilitates users to query without any knowledge on the underlying heterogeneous information sources The problem of providing a consolidated view of information can be handled using Semantic data (information stored in such a way that is understandable by machines and integrate-able without manual human intervention). However, integrating information using semantic web technology without any access management enforced, will results in increase of privacy and confidentiality concerns. In this research we have designed and developed a framework that would allow information from heterogeneous formats to be consolidated, thus resolving the issue of interoperability. We have also devised an access control system for defining explicit privacy constraints. We designed and applied our framework on both semantic and non-semantic data from heterogeneous resources. Our approach is validated using scenario based testing.

Keywords: information integration, semantic data, interoperability, security, access control system

Procedia PDF Downloads 321
10869 The Happiness Pulse: A Measure of Individual Wellbeing at a City Scale, Development and Validation

Authors: Rosemary Hiscock, Clive Sabel, David Manley, Sam Wren-Lewis

Abstract:

As part of the Happy City Index Project, Happy City have developed a survey instrument to measure experienced wellbeing: how people are feeling and functioning in their everyday lives. The survey instrument, called the Happiness Pulse, was developed in partnership with the New Economics Foundation (NEF) with the dual aim of collecting citywide wellbeing data and engaging individuals and communities in the measurement and promotion of their own wellbeing. The survey domains and items were selected through a review of the academic literature and a stakeholder engagement process, including local policymakers, community organisations and individuals. The Happiness Pulse was included in the Bristol pilot of the Happy City Index (n=722). The experienced wellbeing items were subjected to factor analysis. A reduced number of items to be included in a revised scale for future data collection were again entered into a factor analysis. These revised factors were tested for reliability and validity. Among items to be included in a revised scale for future data collection three factors emerged: Be, Do and Connect. The Be factor had good reliability, convergent and criterion validity. The Do factor had good discriminant validity. The Connect factor had adequate reliability and good discriminant and criterion validity. Some age, gender and socioeconomic differentiation was found. The properties of a new scale to measure experienced wellbeing, intended for use by municipal authorities, are described. Happiness Pulse data can be combined with local data on wellbeing conditions to determine what matters for peoples wellbeing across a city and why.

Keywords: city wellbeing , community wellbeing, engaging individuals and communities, measuring wellbeing and happiness

Procedia PDF Downloads 231
10868 Residual Life Estimation Based on Multi-Phase Nonlinear Wiener Process

Authors: Hao Chen, Bo Guo, Ping Jiang

Abstract:

Residual life (RL) estimation based on multi-phase nonlinear Wiener process was studied in this paper, which is significant for complicated products with small samples. Firstly, nonlinear Wiener model with random parameter was introduced and multi-phase nonlinear Wiener model was proposed to model degradation process of products that were nonlinear and separated into different phases. Then the multi-phase RL probability density function based on the presented model was derived approximately in a closed form and parameters estimation was achieved with the method of maximum likelihood estimation (MLE). Finally, the method was applied to estimate the RL of high voltage plus capacitor. Compared with the other three different models by log-likelihood function (Log-LF) and Akaike information criterion (AIC), the results show that the proposed degradation model can capture degradation process of high voltage plus capacitors in a better way and provide a more reliable result.

Keywords: multi-phase nonlinear wiener process, residual life estimation, maximum likelihood estimation, high voltage plus capacitor

Procedia PDF Downloads 429
10867 The Role of Online Videos in Undergraduate Casual-Leisure Information Behaviors

Authors: Nei-Ching Yeh

Abstract:

This study describes undergraduate casual-leisure information behaviors relevant to online videos. Diaries and in-depth interviews were used to collect data. Twenty-four undergraduates participated in this study (9 men, 15 women; all were aged 18–22 years). This study presents a model of casual-leisure information behaviors and contributes new insights into user experience in casual-leisure settings, such as online video programs, with implications for other information domains.

Keywords: casual-leisure information behaviors, information behavior, online videos, role

Procedia PDF Downloads 281
10866 Discrimination and Classification of Vestibular Neuritis Using Combined Fisher and Support Vector Machine Model

Authors: Amine Ben Slama, Aymen Mouelhi, Sondes Manoubi, Chiraz Mbarek, Hedi Trabelsi, Mounir Sayadi, Farhat Fnaiech

Abstract:

Vertigo is a sensation of feeling off balance; the cause of this symptom is very difficult to interpret and needs a complementary exam. Generally, vertigo is caused by an ear problem. Some of the most common causes include: benign paroxysmal positional vertigo (BPPV), Meniere's disease and vestibular neuritis (VN). In clinical practice, different tests of videonystagmographic (VNG) technique are used to detect the presence of vestibular neuritis (VN). The topographical diagnosis of this disease presents a large diversity in its characteristics that confirm a mixture of problems for usual etiological analysis methods. In this study, a vestibular neuritis analysis method is proposed with videonystagmography (VNG) applications using an estimation of pupil movements in the case of an uncontrolled motion to obtain an efficient and reliable diagnosis results. First, an estimation of the pupil displacement vectors using with Hough Transform (HT) is performed to approximate the location of pupil region. Then, temporal and frequency features are computed from the rotation angle variation of the pupil motion. Finally, optimized features are selected using Fisher criterion evaluation for discrimination and classification of the VN disease.Experimental results are analyzed using two categories: normal and pathologic. By classifying the reduced features using the Support Vector Machine (SVM), 94% is achieved as classification accuracy. Compared to recent studies, the proposed expert system is extremely helpful and highly effective to resolve the problem of VNG analysis and provide an accurate diagnostic for medical devices.

Keywords: nystagmus, vestibular neuritis, videonystagmographic system, VNG, Fisher criterion, support vector machine, SVM

Procedia PDF Downloads 119
10865 Interoperable Design Coordination Method for Sharing Communication Information Using Building Information Model Collaboration Format

Authors: Jin Gang Lee, Hyun-Soo Lee, Moonseo Park

Abstract:

The utilization of BIM and IFC allows project participants to collaborate across different areas by consistently sharing interoperable product information represented in a model. Comments or markups generated during the coordination process can be categorized as communication information, which can be shared in less standardized manner. It can be difficult to manage and reuse such information compared to the product information in a model. The present study proposes an interoperable coordination method using BCF (the BIM Collaboration Format) for managing and sharing the communication information during BIM based coordination process. A management function for coordination in the BIM collaboration system is developed to assess its ability to share the communication information in BIM collaboration projects. This approach systematically links communication information during the coordination process to the building model and serves as a type of storage system for retrieving knowledge created during BIM collaboration projects.

Keywords: design coordination, building information model, BIM collaboration format, industry foundation classes

Procedia PDF Downloads 396
10864 A Comparative Analysis of Geometric and Exponential Laws in Modelling the Distribution of the Duration of Daily Precipitation

Authors: Mounia El Hafyani, Khalid El Himdi

Abstract:

Precipitation is one of the key variables in water resource planning. The importance of modeling wet and dry durations is a crucial pointer in engineering hydrology. The objective of this study is to model and analyze the distribution of wet and dry durations. For this purpose, the daily rainfall data from 1967 to 2017 of the Moroccan city of Kenitra’s station are used. Three models are implemented for the distribution of wet and dry durations, namely the first-order Markov chain, the second-order Markov chain, and the truncated negative binomial law. The adherence of the data to the proposed models is evaluated using Chi-square and Kolmogorov-Smirnov tests. The Akaike information criterion is applied to assess the most effective model distribution. We go further and study the law of the number of wet and dry days among k consecutive days. The calculation of this law is done through an algorithm that we have implemented based on conditional laws. We complete our work by comparing the observed moments of the numbers of wet/dry days among k consecutive days to the calculated moment of the three estimated models. The study shows the effectiveness of our approach in modeling wet and dry durations of daily precipitation.

Keywords: Markov chain, rainfall, truncated negative binomial law, wet and dry durations

Procedia PDF Downloads 94
10863 The Effect of Information Technologies on Business Performance: An Application on Small Hotels

Authors: Abdullah Karaman, Kursad Sayin

Abstract:

In this research, which information technologies are used in small hotel businesses, and the information technologies-performance perception of the managers are pointed out. During the research, the questionnaire was prepared and the small scale hotel managers were interviewed face to face and they filled out the questionnaire and the answers acquired were evaluated. As the result of the research, it was obtained that the managers do not care much about the information technologies usage in practice even though they accepted that the information technologies are important in terms of performance.

Keywords: information technologies, managers, performance, small hotels

Procedia PDF Downloads 461
10862 Controlling the Expense of Political Contests Using a Modified N-Players Tullock’s Model

Authors: C. Cohen, O. Levi

Abstract:

This work introduces a generalization of the classical Tullock’s model of one-stage contests under complete information with multiple unlimited numbers of contestants. In classical Tullock’s model, the contest winner is not necessarily the highest bidder. Instead, the winner is determined according to a draw in which the winning probabilities are the relative contestants’ efforts. The Tullock modeling fits well political contests, in which the winner is not necessarily the highest effort contestant. This work presents a modified model which uses a simple non-discriminating rule, namely, a parameter to influence the total costs planned for an election, for example, the contest designer can control the contestants' efforts. The winner pays a fee, and the losers are reimbursed the same amount. Our proposed model includes a mechanism that controls the efforts exerted and balances competition, creating a tighter, less predictable and more interesting contest. Additionally, the proposed model follows the fairness criterion in the sense that it does not alter the contestants' probabilities of winning compared to the classic Tullock’s model. We provide an analytic solution for the contestant's optimal effort and expected reward.

Keywords: contests, Tullock's model, political elections, control expenses

Procedia PDF Downloads 116
10861 The Complexities of Designing a Learning Programme in Higher Education with the End-User in Mind

Authors: Andre Bechuke

Abstract:

The quality of every learning programme in Higher Education (HE) is dependent on the planning, design, and development of the curriculum decisions. These curriculum development decisions are highly influenced by the knowledge of the end-user, who are not always just the students. When curriculum experts plan, design and develop learning programmes, they always have the end-users in mind throughout the process. Without proper knowledge of the end-user(s), the design and development of a learning programme might be flawed. Curriculum experts often struggle to determine who the real end-user is. As such, it is even more challenging to establish what needs to be known about the end user that should inform the plan, design, and development of a learning programme. This research sought suggest approaches to guide curriculum experts to identify the end-user(s), taking into consideration the pressure and influence other agencies and structures or stakeholders (industry, students, government, universities context, lecturers, international communities, professional regulatory bodies) have on the design of a learning programme and the graduates of the programmes. Considering the influence of these stakeholders, which is also very important, the task of deciding who the real end-user of the learning programme becomes very challenging. This study makes use of criteria 1 and 18 of the Council on Higher Education criteria for programme accreditation to guide the process of identifying the end-users when developing a learning programme. Criterion 1 suggests that designers must ensure that the programme is consonant with the institution’s mission, forms part of institutional planning and resource allocation, meets national requirements and the needs of students and other stakeholders, and is intellectually credible. According to criterion 18, in designing a learning programme, steps must be taken to enhance the employability of students and alleviate shortages of expertise in relevant fields. In conclusion, there is hardly ever one group of end-users to be considered for developing a learning programme, and the notion that students are the end-users is not true, especially when the graduates are unable to use the qualification for employment.

Keywords: council on higher education, curriculum design and development, higher education, learning programme

Procedia PDF Downloads 55
10860 Health Status Monitoring of COVID-19 Patient's through Blood Tests and Naïve-Bayes

Authors: Carlos Arias-Alcaide, Cristina Soguero-Ruiz, Paloma Santos-Álvarez, Adrián García-Romero, Inmaculada Mora-Jiménez

Abstract:

Analysing clinical data with computers in such a way that have an impact on the practitioners’ workflow is a challenge nowadays. This paper provides a first approach for monitoring the health status of COVID-19 patients through the use of some biomarkers (blood tests) and the simplest Naïve Bayes classifier. Data of two Spanish hospitals were considered, showing the potential of our approach to estimate reasonable posterior probabilities even some days before the event.

Keywords: Bayesian model, blood biomarkers, classification, health tracing, machine learning, posterior probability

Procedia PDF Downloads 187