Search results for: BIM service based cloud computing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 30354

Search results for: BIM service based cloud computing

26004 Impact of Primary Care Telemedicine Consultations On Health Care Resource Utilisation: A Systematic Review

Authors: Anastasia Constantinou, Stephen Morris

Abstract:

Background: The adoption of synchronous and asynchronous telemedicine modalities for primary care consultations has exponentially increased since the COVID-19 pandemic. However, there is limited understanding of how virtual consultations influence healthcare resource utilization and other quality measures including safety, timeliness, efficiency, patient and provider satisfaction, cost-effectiveness and environmental impact. Aim: Quantify the rate of follow-up visits, emergency department visits, hospitalizations, request for investigations and prescriptions and comment on the effect on different quality measures associated with different telemedicine modalities used for primary care services and primary care referrals to secondary care Design and setting: Systematic review in primary care Methods: A systematic search was carried out across three databases (Medline, PubMed and Scopus) between August and November 2023, using terms related to telemedicine, general practice, electronic referrals, follow-up, use and efficiency and supported by citation searching. This was followed by screening according to pre-defined criteria, data extraction and critical appraisal. Narrative synthesis and metanalysis of quantitative data was used to summarize findings. Results: The search identified 2230 studies; 50 studies are included in this review. There was a prevalence of asynchronous modalities in both primary care services (68%) and referrals from primary care to secondary care (83%), and most of the study participants were females (63.3%), with mean age of 48.2. The average follow-up for virtual consultations in primary care was 28.4% (eVisits: 36.8%, secure messages 18.7%, videoconference 23.5%) with no significant difference between them or F2F consultations. There was an average annual reduction of primary care visits by 0.09/patient, an increase in telephone visits by 0.20/patient, an increase in ED encounters by 0.011/patient, an increase in hospitalizations by 0.02/patient and an increase in out of hours visits by 0.019/patient. Laboratory testing was requested on average for 10.9% of telemedicine patients, imaging or procedures for 5.6% and prescriptions for 58.7% of patients. When looking at referrals to secondary care, on average 36.7% of virtual referrals required follow-up visit, with the average rate of follow-up for electronic referrals being higher than for videoconferencing (39.2% vs 23%, p=0.167). Technical failures were reported on average for 1.4% of virtual consultations to primary care. When using carbon footprint estimates, we calculate that the use of telemedicine in primary care services can potentially provide a net decrease in carbon footprint by 0.592kgCO2/patient/year. When follow-up rates are taken into account, we estimate that virtual consultations reduce carbon footprint for primary care services by 2.3 times, and for secondary care referrals by 2.2 times. No major concerns regarding quality of care, or patient satisfaction were identified. 5/7 studies that addressed cost-effectiveness, reported increased savings. Conclusions: Telemedicine provides quality, cost-effective, and environmentally sustainable care for patients in primary care with inconclusive evidence regarding the rates of subsequent healthcare utilization. The evidence is limited by heterogeneous, small-scale studies and lack of prospective comparative studies. Further research to identify the most appropriate telemedicine modality for different patient populations, clinical presentations, service provision (e.g. used to follow-up patients instead of initial diagnosis) as well as further education for patients and providers alike on how to make best use of this service is expected to improve outcomes and influence practice.

Keywords: telemedicine, healthcare utilisation, digital interventions, environmental impact, sustainable healthcare

Procedia PDF Downloads 46
26003 Heart-Rate Resistance Electrocardiogram Identification Based on Slope-Oriented Neural Networks

Authors: Tsu-Wang Shen, Shan-Chun Chang, Chih-Hsien Wang, Te-Chao Fang

Abstract:

For electrocardiogram (ECG) biometrics system, it is a tedious process to pre-install user’s high-intensity heart rate (HR) templates in ECG biometric systems. Based on only resting enrollment templates, it is a challenge to identify human by using ECG with the high-intensity HR caused from exercises and stress. This research provides a heartbeat segment method with slope-oriented neural networks against the ECG morphology changes due to high intensity HRs. The method has overall system accuracy at 97.73% which includes six levels of HR intensities. A cumulative match characteristic curve is also used to compare with other traditional ECG biometric methods.

Keywords: high-intensity heart rate, heart rate resistant, ECG human identification, decision based artificial neural network

Procedia PDF Downloads 416
26002 Genomic Sequence Representation Learning: An Analysis of K-Mer Vector Embedding Dimensionality

Authors: James Jr. Mashiyane, Risuna Nkolele, Stephanie J. Müller, Gciniwe S. Dlamini, Rebone L. Meraba, Darlington S. Mapiye

Abstract:

When performing language tasks in natural language processing (NLP), the dimensionality of word embeddings is chosen either ad-hoc or is calculated by optimizing the Pairwise Inner Product (PIP) loss. The PIP loss is a metric that measures the dissimilarity between word embeddings, and it is obtained through matrix perturbation theory by utilizing the unitary invariance of word embeddings. Unlike in natural language, in genomics, especially in genome sequence processing, unlike in natural language processing, there is no notion of a “word,” but rather, there are sequence substrings of length k called k-mers. K-mers sizes matter, and they vary depending on the goal of the task at hand. The dimensionality of word embeddings in NLP has been studied using the matrix perturbation theory and the PIP loss. In this paper, the sufficiency and reliability of applying word-embedding algorithms to various genomic sequence datasets are investigated to understand the relationship between the k-mer size and their embedding dimension. This is completed by studying the scaling capability of three embedding algorithms, namely Latent Semantic analysis (LSA), Word2Vec, and Global Vectors (GloVe), with respect to the k-mer size. Utilising the PIP loss as a metric to train embeddings on different datasets, we also show that Word2Vec outperforms LSA and GloVe in accurate computing embeddings as both the k-mer size and vocabulary increase. Finally, the shortcomings of natural language processing embedding algorithms in performing genomic tasks are discussed.

Keywords: word embeddings, k-mer embedding, dimensionality reduction

Procedia PDF Downloads 127
26001 Work-Life Balance: A Landscape Mapping of Two Decades of Scholarly Research

Authors: Gertrude I Hewapathirana, Mohamed M. Moustafa, Michel G. Zaitouni

Abstract:

The purposes of this research are: (a) to provide an epistemological and ontological understanding of the WLB theory, practice, and research to illuminate how the WLB evolved between 2000 to 2020 and (b) to analyze peer-reviewed research to identify the gaps, hotspots, underlying dynamics, theoretical and thematic trends, influential authors, research collaborations, geographic networks, and the multidisciplinary nature of the WLB theory to guide future researchers. The research used four-step bibliometric network analysis to explore five research questions. Using keywords such as WLB and associated variants, 1190 peer-reviewed articles were extracted from the Scopus database and transformed to a plain text format for filtering. The analysis was conducted using the R version 4.1 software (R Development Core Team, 2021) and several libraries such as bibliometrics, word cloud, and ggplot2. We used the VOSviewer software (van Eck & Waltman, 2019) for network visualization. The WLB theory has grown into a multifaceted, multidisciplinary field of research. There is a paucity of research between 2000 to 2005 and an exponential growth from 2006 to 2015. The rapid increase of WLB research in the USA, UK, and Australia reflects the increasing workplace stresses due to hyper competitive workplaces, inflexible work systems, and increasing diversity and the emergence of WLB support mechanisms, legal and constitutional mandates to enhance employee and family wellbeing at multilevel social systems. A severe knowledge gap exists due to inadequate publications disseminating the "core" WLB research. "Locally-centralized-globally-discrete" collaboration among researchers indicates a "North-South" divide between developed and developing nations. A shortage in WLB research in developing nations and a lack of research collaboration hinder a global understanding of the WLB as a universal phenomenon. Policymakers and practitioners can use the findings to initiate supporting policies, and innovative work systems. The boundary expansion of the WLB concepts, categories, relations, and properties would facilitate researchers/theoreticians to test a variety of new dimensions. This is the most comprehensive WLB landscape analysis that reveals emerging trends, concepts, networks, underlying dynamics, gaps, and growing theoretical and disciplinary boundaries. It portrays the WLB as a universal theory.

Keywords: work-life balance, co-citation networks; keyword co-occurrence network, bibliometric analysis

Procedia PDF Downloads 185
26000 PTFE Capillary-Based DNA Amplification within an Oscillatory Thermal Cycling Device

Authors: Jyh J. Chen, Fu H. Yang, Ming H. Liao

Abstract:

This study describes a capillary-based device integrated with the heating and cooling modules for polymerase chain reaction (PCR). The device consists of the reaction polytetrafluoroethylene (PTFE) capillary, the aluminum blocks, and is equipped with two cartridge heaters, a thermoelectric (TE) cooler, a fan, and some thermocouples for temperature control. The cartridge heaters are placed into the heating blocks and maintained at two different temperatures to achieve the denaturation and the extension step. Some thermocouples inserted into the capillary are used to obtain the transient temperature profiles of the reaction sample during thermal cycles. A 483-bp DNA template is amplified successfully in the designed system and the traditional thermal cycler. This work should be interesting to persons involved in the high-temperature based reactions and genomics or cell analysis.

Keywords: polymerase chain reaction, thermal cycles, capillary, TE cooler

Procedia PDF Downloads 436
25999 Industrial Process Mining Based on Data Pattern Modeling and Nonlinear Analysis

Authors: Hyun-Woo Cho

Abstract:

Unexpected events may occur with serious impacts on industrial process. This work utilizes a data representation technique to model and to analyze process data pattern for the purpose of diagnosis. In this work, the use of triangular representation of process data is evaluated using simulation process. Furthermore, the effect of using different pre-treatment techniques based on such as linear or nonlinear reduced spaces was compared. This work extracted the fault pattern in the reduced space, not in the original data space. The results have shown that the non-linear technique based diagnosis method produced more reliable results and outperforms linear method.

Keywords: process monitoring, data analysis, pattern modeling, fault, nonlinear techniques

Procedia PDF Downloads 377
25998 Pet Bearing Bio-Based Moities

Authors: Majdi Abid

Abstract:

During the last few decades, great efforts have been made for the development of innovative materials using vegetal biomass. This strategy is understandable for different reasons including the predictable dwindling of the petrochemical feedstock and their price increase as well as the counterbalancing of the environmental problems. As novel bio-based monomers used in polyesters synthesis, two families, namely 1,4:3,6-dianhydrohexitols and furanics were prepared for saccharidic renewable resources. The present work deals with a detail investigation on the synthesis of poly(ethylene-co-isosorbide terephthalate-co-furoate) (PEITF) by melt polycondensation of dimethylterephtalate (DMT), 5,5’-isopropylidene-bis (ethyl 2-furoate) (DEF) ethan-1,2-diol (ED) and isosorbide (IS). Polycondensationwas achieved in two steps: (i) the formation of a hydroxyethylterminated oligomer by reaction of starting diester mixture with excess ED and, (ii) a polycondensation step with elimination of ED was used to obtain high molar mass copolyesters. Copolymers of various compositions were synthesized and characterized by 1H NMR, SEC, DSC and TGA. The resulting materials are amorphous polymers (Tg = 104–127 °C) with good thermal stability.

Keywords: bio-based monomers, furan, isosrbide, polycondensation

Procedia PDF Downloads 274
25997 Approaching Sexual Violence Against People with Disabilities in Colombia from a Qualitative Perspective

Authors: Mariana Calderón, Rocío Murad, Natalia Acevedo, Laura León, Juliana Fonseca, Maria de los Angeles Balaguera Villa

Abstract:

Recently, different countries and international organizations have put on their agenda the elimination of violence against people with disabilities. This research aims to evaluate the social dimensions of sexual violence against people with disabilities, particularly those with psychosocial and cognitive, in Colombia. Results reveal that 55% of people with disabilities that are survivors of sexual violence are younger than 29 years and 20,4 are people with cognitive and psychosocial disabilities. Colombian regions with better social positions presented more cases of sexual violence against people with disabilities. There were found access barriers for health, education and employment among this population, and there was also found poor data quality. Despite Colombia having an important normative framework aimed at preventing and attending to gender-based violence, it does not take into account people with disabilities specific needs. Additionally, it was found an insufficient implementation and appropriation of these norms, negative attitudes, and in general, a lack of service adaptation according to the needs, identities and circumstances of people with disabilities. Furthermore, among the factors that are exposing people with disabilities to sexual violence, it was found that family members tend to be the main aggressors, there are deep gaps in the sex education received by people with disabilities, imaginaries and perceptions about their sexuality are both hypersexualizing and presenting them as asexual. On the other hand, among protective factors, there were found body self-knowledge and conscience, acknowledgment of their sexuality and their sexual and reproductive rights and access to sex ed. Although during the last few years, there has occurred a positive change toward social inclusion of people with disabilities, specifically through their role in the political agenda and the recognition of their rights. More work is needed in order to guarantee their sexual and reproductive rights, particularly for persons with psychosocial and cognitive disabilities. This research results showed the importance of transforming persisting negative imaginaries about their sexuality and also enforcing and promoting their autonomy. In this sense, it is important to acknowledge gaps and barriers faced by them and create strategies to encourage their social inclusion through education, employment, and skill development. Nevertheless, it is necessary to keep contributing new evidence of the social determinants of health that are influencing the occurrence of sexual violence. This research understands sexual violence against people with disabilities in a multidimensional manner and offers the following recommendations: 1- To foment public sensitization and understanding of disabilities. 2- To increase parents, caregivers and officers’ commitment to the prevention and reduction of sexual violence. 3- To focus on the needs, identities and circumstances of people with disabilities.

Keywords: disabilities, sexual and reproductive rights, sexual violence, prevention

Procedia PDF Downloads 64
25996 Suitability of Black Box Approaches for the Reliability Assessment of Component-Based Software

Authors: Anjushi Verma, Tirthankar Gayen

Abstract:

Although, reliability is an important attribute of quality, especially for mission critical systems, yet, there does not exist any versatile model even today for the reliability assessment of component-based software. The existing Black Box models are found to make various assumptions which may not always be realistic and may be quite contrary to the actual behaviour of software. They focus on observing the manner in which the system behaves without considering the structure of the system, the components composing the system, their interconnections, dependencies, usage frequencies, etc.As a result, the entropy (uncertainty) in assessment using these models is much high.Though, there are some models based on operation profile yet sometimes it becomes extremely difficult to obtain the exact operation profile concerned with a given operation. This paper discusses the drawbacks, deficiencies and limitations of Black Box approaches from the perspective of various authors and finally proposes a conceptual model for the reliability assessment of software.

Keywords: black box, faults, failure, software reliability

Procedia PDF Downloads 436
25995 Extent of Applying Evidence Based Practices in Inclusion Programs for Pupils with Intellectual Disability

Authors: Faris Algahtani

Abstract:

The current study aimed to reveal the extent to which evidence-based practices are applied in programs to integrate students with intellectual disabilities from the point of view of their teachers in Yanbu Governorate, and to reveal statistically significant differences in their application of evidence-based practices according to the following variables: gender, educational qualification, experience and training courses. The researcher used the descriptive approach, and accordingly; she designed a questionnaire consisting of 22 phrases applied it to a random sample of (97) teachers of intellectual disability in the integration programs of the Ministry of Education in the government sector in Yanbu Governorate, with (49) male teachers and (48) female teachers. The study showed that teachers of students with intellectual disabilities apply evidence-based practices in programs to integrate students with intellectual disabilities to a large extent. Among the most prominent of these practices came reinforcement in the first place, followed by using visual stimuli/aids, and in the third-place came starting with less complex or challenging skills then moving to more difficult skills. The results also showed no statistically significant differences over the extent of the application attributed to the variables of experience, qualification or training. On the other hand, there were statistically significant differences over the extent of the application attributed to gender in favor of females.

Keywords: evidence-based practices, intellectual disability, inclusion programs, teachers of students with intellectual disabilities

Procedia PDF Downloads 74
25994 Proteomic Analysis of Excretory Secretory Antigen (ESA) from Entamoeba histolytica HM1: IMSS

Authors: N. Othman, J. Ujang, M. N. Ismail, R. Noordin, B. H. Lim

Abstract:

Amoebiasis is caused by the Entamoeba histolytica and still endemic in many parts of the tropical region, worldwide. Currently, there is no available vaccine against amoebiasis. Hence, there is an urgent need to develop a vaccine. The excretory secretory antigen (ESA) of E. histolytica is a suitable biomarker for the vaccine candidate since it can modulate the host immune response. Hence, the objective of this study is to identify the proteome of the ESA towards finding suitable biomarker for the vaccine candidate. The non-gel based and gel-based proteomics analyses were performed to identify proteins. Two kinds of mass spectrometry with different ionization systems were utilized i.e. LC-MS/MS (ESI) and MALDI-TOF/TOF. Then, the functional proteins classification analysis was performed using PANTHER software. Combination of the LC -MS/MS for the non-gel based and MALDI-TOF/TOF for the gel-based approaches identified a total of 273 proteins from the ESA. Both systems identified 29 similar proteins whereby 239 and 5 more proteins were identified by LC-MS/MS and MALDI-TOF/TOF, respectively. Functional classification analysis showed the majority of proteins involved in the metabolic process (24%), primary metabolic process (19%) and protein metabolic process (10%). Thus, this study has revealed the proteome the E. histolytica ESA and the identified proteins merit further investigations as a vaccine candidate.

Keywords: E. histolytica, ESA, proteomics, biomarker

Procedia PDF Downloads 328
25993 Multi-Walled Carbon Nanotube Based Water Filter for Virus Pathogen Removal

Authors: K. Domagala, D. Kata, T. Graule

Abstract:

Diseases caused by contaminated drinking water are the worldwide problem, which leads to the death and severe illnesses for hundreds of millions million people each year. There is an urgent need for efficient water treatment techniques for virus pathogens removal. The aim of the research was to develop safe and economic solution, which help with the water treatment. In this study, the synthesis of copper-based multi-walled carbon nanotube composites is described. Proposed solution utilize combination of a low-cost material with a high active surface area and copper antiviral properties. Removal of viruses from water was possible by adsorption based on electrostatic interactions of negatively charged virus with a positively charged filter material.

Keywords: multi walled carbon nanotubes, water purification, virus removal, water treatment

Procedia PDF Downloads 120
25992 A Case-Based Reasoning-Decision Tree Hybrid System for Stock Selection

Authors: Yaojun Wang, Yaoqing Wang

Abstract:

Stock selection is an important decision-making problem. Many machine learning and data mining technologies are employed to build automatic stock-selection system. A profitable stock-selection system should consider the stock’s investment value and the market timing. In this paper, we present a hybrid system including both engage for stock selection. This system uses a case-based reasoning (CBR) model to execute the stock classification, uses a decision-tree model to help with market timing and stock selection. The experiments show that the performance of this hybrid system is better than that of other techniques regarding to the classification accuracy, the average return and the Sharpe ratio.

Keywords: case-based reasoning, decision tree, stock selection, machine learning

Procedia PDF Downloads 404
25991 Comparing Bestseller Books in Japanese Public Libraries Managed by Outsourcing and by Local Governments

Authors: Yuhiro Mizunuma, Keita Tsuji

Abstract:

In Japan, local governments have long managed public libraries; however, organizations including private enterprises took over their management with the introduction of an outsourcing system in 2003. Now, whether local governments should apply this system to public libraries is under discussion, and many argue that it is inappropriate for public libraries. In this study, to provide basic data for such discussions, the differences between public libraries managed by outsourcing (henceforth, 'outsourcing libraries') and those managed directly by local governments (henceforth 'direct management libraries') were examined, focusing on bestsellers. We chose the 435 books on the top-20 bestseller lists published from 1996 to 2015 inclusive, as our sample bestseller books; and 413 outsourcing libraries' and 2,619 direct management libraries' holdings of them were investigated using the Calil API (a free Web service that can perform cross-library searches on libraries in Japan). Results show that outsourcing libraries tend to hold fewer bestsellers than direct management libraries. It was also found that direct management libraries tend to hold many duplicates of bestsellers, and that, in many cases, many of those were not being borrowed in direct management libraries.

Keywords: bestseller books, book collection, Japanese public libraries, outsourcing

Procedia PDF Downloads 202
25990 Methods of Improving Production Processes Based on Deming Cycle

Authors: Daniel Tochwin

Abstract:

Continuous improvement is an essential part of effective process performance management. In order to achieve continuous quality improvement, each organization must use the appropriate selection of tools and techniques. The basic condition for success is a proper understanding of the business need faced by the company and the selection of appropriate methods to improve a given production process. The main aim of this article is to analyze the methods of conduct which are popular in practice when implementing process improvements and then to determine whether the tested methods include repetitive systematics of the approach, i.e., a similar sequence of the same or similar actions. Based on an extensive literature review, 4 methods of continuous improvement of production processes were selected: A3 report, Gemba Kaizen, PDCA cycle, and Deming cycle. The research shows that all frequently used improvement methods are generally based on the PDCA cycle, and the differences are due to "(re)interpretation" and the need to adapt the continuous improvement approach to the specific business process. The research shows that all the frequently used improvement methods are generally based on the PDCA cycle, and the differences are due to "(re) interpretation" and the need to adapt the continuous improvement approach to the specific business process.

Keywords: continuous improvement, lean methods, process improvement, PDCA

Procedia PDF Downloads 65
25989 Robust Variable Selection Based on Schwarz Information Criterion for Linear Regression Models

Authors: Shokrya Saleh A. Alshqaq, Abdullah Ali H. Ahmadini

Abstract:

The Schwarz information criterion (SIC) is a popular tool for selecting the best variables in regression datasets. However, SIC is defined using an unbounded estimator, namely, the least-squares (LS), which is highly sensitive to outlying observations, especially bad leverage points. A method for robust variable selection based on SIC for linear regression models is thus needed. This study investigates the robustness properties of SIC by deriving its influence function and proposes a robust SIC based on the MM-estimation scale. The aim of this study is to produce a criterion that can effectively select accurate models in the presence of vertical outliers and high leverage points. The advantages of the proposed robust SIC is demonstrated through a simulation study and an analysis of a real dataset.

Keywords: influence function, robust variable selection, robust regression, Schwarz information criterion

Procedia PDF Downloads 128
25988 Model Solutions for Performance-Based Seismic Analysis of an Anchored Sheet Pile Quay Wall

Authors: C. J. W. Habets, D. J. Peters, J. G. de Gijt, A. V. Metrikine, S. N. Jonkman

Abstract:

Conventional seismic designs of quay walls in ports are mostly based on pseudo-static analysis. A more advanced alternative is the Performance-Based Design (PBD) method, which evaluates permanent deformations and amounts of (repairable) damage under seismic loading. The aim of this study is to investigate the suitability of this method for anchored sheet pile quay walls that were not purposely designed for seismic loads. A research methodology is developed in which pseudo-static, permanent-displacement and finite element analysis are employed, calibrated with an experimental reference case that considers a typical anchored sheet pile wall. A reduction factor that accounts for deformation behaviour is determined for pseudo-static analysis. A model to apply traditional permanent displacement analysis on anchored sheet pile walls is proposed. Dynamic analysis is successfully carried out. From the research it is concluded that PBD evaluation can effectively be used for seismic analysis and design of this type of structure.

Keywords: anchored sheet pile quay wall, simplified dynamic analysis, performance-based design, pseudo-static analysis

Procedia PDF Downloads 370
25987 Early Detection of Breast Cancer in Digital Mammograms Based on Image Processing and Artificial Intelligence

Authors: Sehreen Moorat, Mussarat Lakho

Abstract:

A method of artificial intelligence using digital mammograms data has been proposed in this paper for detection of breast cancer. Many researchers have developed techniques for the early detection of breast cancer; the early diagnosis helps to save many lives. The detection of breast cancer through mammography is effective method which detects the cancer before it is felt and increases the survival rate. In this paper, we have purposed image processing technique for enhancing the image to detect the graphical table data and markings. Texture features based on Gray-Level Co-Occurrence Matrix and intensity based features are extracted from the selected region. For classification purpose, neural network based supervised classifier system has been used which can discriminate between benign and malignant. Hence, 68 digital mammograms have been used to train the classifier. The obtained result proved that automated detection of breast cancer is beneficial for early diagnosis and increases the survival rates of breast cancer patients. The proposed system will help radiologist in the better interpretation of breast cancer.

Keywords: medical imaging, cancer, processing, neural network

Procedia PDF Downloads 249
25986 Statistical Inferences for GQARCH-It\^{o} - Jumps Model Based on The Realized Range Volatility

Authors: Fu Jinyu, Lin Jinguan

Abstract:

This paper introduces a novel approach that unifies two types of models: one is the continuous-time jump-diffusion used to model high-frequency data, and the other is discrete-time GQARCH employed to model low-frequency financial data by embedding the discrete GQARCH structure with jumps in the instantaneous volatility process. This model is named “GQARCH-It\^{o} -Jumps mode.” We adopt the realized range-based threshold estimation for high-frequency financial data rather than the realized return-based volatility estimators, which entail the loss of intra-day information of the price movement. Meanwhile, a quasi-likelihood function for the low-frequency GQARCH structure with jumps is developed for the parametric estimate. The asymptotic theories are mainly established for the proposed estimators in the case of finite activity jumps. Moreover, simulation studies are implemented to check the finite sample performance of the proposed methodology. Specifically, it is demonstrated that how our proposed approaches can be practically used on some financial data.

Keywords: It\^{o} process, GQARCH, leverage effects, threshold, realized range-based volatility estimator, quasi-maximum likelihood estimate

Procedia PDF Downloads 141
25985 Self-Regulated Learning: A Required Skill for Web 2.0 Internet-Based Learning

Authors: Pieter Conradie, M. Marina Moller

Abstract:

Web 2.0 Internet-based technologies have intruded all aspects of human life. Presently, this phenomenon is especially evident in the educational context, with increased disruptive Web 2.0 technology infusions dramatically changing educational practice. The most prominent of these Web 2.0 intrusions can be identified as Massive Open Online Courses (Coursera, EdX), video and photo sharing sites (Youtube, Flickr, Instagram), and Web 2.0 online tools utilize to create Personal Learning Environments (PLEs) (Symbaloo (aggregator), Delicious (social bookmarking), PBWorks (collaboration), Google+ (social networks), Wordspress (blogs), Wikispaces (wiki)). These Web 2.0 technologies have supported the realignment from a teacher-based pedagogy (didactic presentation) to a learner-based pedagogy (problem-based learning, project-based learning, blended learning), allowing greater learner autonomy. No longer is the educator the source of knowledge. Instead the educator has become the facilitator and mediator of the learner, involved in developing learner competencies to support life-long learning (continuous learning) in the 21st century. In this study, the self-regulated learning skills of thirty first-year university learners were explored by utilizing the Online Self-regulated Learning Questionnaire. Implementing an action research method, an intervention was affected towards improving the self-regulation skill set of the participants. Statistical significant results were obtained with increased self-regulated learning proficiency, positively impacting learner performance. Goal setting, time management, environment structuring, help seeking, task (learning) strategies and self-evaluation skills were confirmed as determinants of improved learner success.

Keywords: andragogy, online self-regulated learning questionnaire, self-regulated learning, web 2.0

Procedia PDF Downloads 403
25984 Evaluation of Model-Based Code Generation for Embedded Systems–Mature Approach for Development in Evolution

Authors: Nikolay P. Brayanov, Anna V. Stoynova

Abstract:

Model-based development approach is gaining more support and acceptance. Its higher abstraction level brings simplification of systems’ description that allows domain experts to do their best without particular knowledge in programming. The different levels of simulation support the rapid prototyping, verifying and validating the product even before it exists physically. Nowadays model-based approach is beneficial for modelling of complex embedded systems as well as a generation of code for many different hardware platforms. Moreover, it is possible to be applied in safety-relevant industries like automotive, which brings extra automation of the expensive device certification process and especially in the software qualification. Using it, some companies report about cost savings and quality improvements, but there are others claiming no major changes or even about cost increases. This publication demonstrates the level of maturity and autonomy of model-based approach for code generation. It is based on a real live automotive seat heater (ASH) module, developed using The Mathworks, Inc. tools. The model, created with Simulink, Stateflow and Matlab is used for automatic generation of C code with Embedded Coder. To prove the maturity of the process, Code generation advisor is used for automatic configuration. All additional configuration parameters are set to auto, when applicable, leaving the generation process to function autonomously. As a result of the investigation, the publication compares the quality of generated embedded code and a manually developed one. The measurements show that generally, the code generated by automatic approach is not worse than the manual one. A deeper analysis of the technical parameters enumerates the disadvantages, part of them identified as topics for our future work.

Keywords: embedded code generation, embedded C code quality, embedded systems, model-based development

Procedia PDF Downloads 231
25983 Comparative Analysis of Dissimilarity Detection between Binary Images Based on Equivalency and Non-Equivalency of Image Inversion

Authors: Adnan A. Y. Mustafa

Abstract:

Image matching is a fundamental problem that arises frequently in many aspects of robot and computer vision. It can become a time-consuming process when matching images to a database consisting of hundreds of images, especially if the images are big. One approach to reducing the time complexity of the matching process is to reduce the search space in a pre-matching stage, by simply removing dissimilar images quickly. The Probabilistic Matching Model for Binary Images (PMMBI) showed that dissimilarity detection between binary images can be accomplished quickly by random pixel mapping and is size invariant. The model is based on the gamma binary similarity distance that recognizes an image and its inverse as containing the same scene and hence considers them to be the same image. However, in many applications, an image and its inverse are not treated as being the same but rather dissimilar. In this paper, we present a comparative analysis of dissimilarity detection between PMMBI based on the gamma binary similarity distance and a modified PMMBI model based on a similarity distance that does distinguish between an image and its inverse as being dissimilar.

Keywords: binary image, dissimilarity detection, probabilistic matching model for binary images, image mapping

Procedia PDF Downloads 143
25982 Pill-Box Dispenser as a Strategy for Therapeutic Management: A Qualitative Evaluation

Authors: Bruno R. Mendes, Francisco J. Caldeira, Rita S. Luís

Abstract:

Population ageing is directly correlated to an increase in medicine consumption. Beyond the latter and the polymedicated profile of elderly, it is possible to see a need for pharmacotherapeutic monitoring due to cognitive and physical impairment. In this sense, the tracking, organization and administration of medicines become a daily challenge and the pill-box dispenser system a solution. The pill-box dispenser (system) consists in a small compartmentalized container to unit dose organization, which means a container able to correlate the patient’s prescribed dose regimen and the time schedule of intake. In many European countries, this system is part of pharmacist’s role in clinical pharmacy. Despite this simple solution, therapy compliance is only possible if the patient adheres to the system, so it is important to establish a qualitative and quantitative analysis on the perception of the patient on the benefits and risks of the pill-box dispenser as well as the identification of the ideal system. The analysis was conducted through an observational study, based on the application of a standardized questionnaire structured with the numerical scale of Likert (5 levels) and previously validated on the population. The study was performed during a limited period of time and under a randomized sample of 188 participants. The questionnaire consisted of 22 questions: 6 background measures and 16 specific measures. The standards for the final comparative analysis were obtained through the state-of-the-art on the subject. The study carried out using the Likert scale afforded a degree of agreement and discordance between measures (Sample vs. Standard) of 56,25% and 43,75%, respectively. It was concluded that the pill-box dispenser has greater acceptance among a younger population, that was not the initial target of the system. However, this allows us to guarantee a high adherence in the future. Additionally, it was noted that the cost associated with this service is not a limiting factor for its use. The pill-box dispenser system, as currently implemented, demonstrates an important weakness regarding the quality and effectiveness of the medicines, which is not understood by the patient, revealing a significant lack of literacy when it concerns with medicine area. The characteristics of an ideal system remain unchanged, which means that the size, appearance and availability of information in the pill-box continue to be indispensable elements for the compliance with the system. The pill-box dispenser remains unsuitable regarding container size and the type of treatment to which it applies. Despite that, it might be a future standard for clinical pharmacy, allowing a differentiation of the pharmacist role, as well as a wider range of applications to other age groups and treatments.

Keywords: clinical pharmacy, medicines, patient safety, pill-box dispenser

Procedia PDF Downloads 185
25981 Review of Theories and Applications of Genetic Programing in Sediment Yield Modeling

Authors: Adesoji Tunbosun Jaiyeola, Josiah Adeyemo

Abstract:

Sediment yield can be considered to be the total sediment load that leaves a drainage basin. The knowledge of the quantity of sediments present in a river at a particular time can lead to better flood capacity in reservoirs and consequently help to control over-bane flooding. Furthermore, as sediment accumulates in the reservoir, it gradually loses its ability to store water for the purposes for which it was built. The development of hydrological models to forecast the quantity of sediment present in a reservoir helps planners and managers of water resources systems, to understand the system better in terms of its problems and alternative ways to address them. The application of artificial intelligence models and technique to such real-life situations have proven to be an effective approach of solving complex problems. This paper makes an extensive review of literature relevant to the theories and applications of evolutionary algorithms, and most especially genetic programming. The successful applications of genetic programming as a soft computing technique were reviewed in sediment modelling and other branches of knowledge. Some fundamental issues such as benchmark, generalization ability, bloat and over-fitting and other open issues relating to the working principles of GP, which needs to be addressed by the GP community were also highlighted. This review aim to give GP theoreticians, researchers and the general community of GP enough research direction, valuable guide and also keep all stakeholders abreast of the issues which need attention during the next decade for the advancement of GP.

Keywords: benchmark, bloat, generalization, genetic programming, over-fitting, sediment yield

Procedia PDF Downloads 432
25980 Assessment-Assisted and Relationship-Based Financial Advising: Using an Empirical Assessment to Understand Personal Investor Risk Tolerance in Professional Advising Relationships

Authors: Jerry Szatko, Edan L. Jorgensen, Stacia Jorgensen

Abstract:

A crucial component to the success of any financial advising relationship is for the financial professional to understand the perceptions, preferences and thought-processes carried by the financial clients they serve. Armed with this information, financial professionals are more quickly able to understand how they can tailor their approach to best match the individual preferences and needs of each personal investor. Our research explores the use of a quantitative assessment tool in the financial services industry to assist in the identification of the personal investor’s consumer behaviors, especially in terms of financial risk tolerance, as it relates to their financial decision making. Through this process, the Unitifi Consumer Insight Tool (UCIT) was created and refined to capture and categorize personal investor financial behavioral categories and the financial personality tendencies of individuals prior to the initiation of a financial advisement relationship. This paper discusses the use of this tool to place individuals in one of four behavior-based financial risk tolerance categories. Our discoveries and research were aided through administration of a web-based survey to a group of over 1,000 individuals. Our findings indicate that it is possible to use a quantitative assessment tool to assist in predicting the behavioral tendencies of personal consumers when faced with consumer financial risk and decisions.

Keywords: behavior-based advising, financial relationship building, risk capacity based on behavior, risk tolerance, systematic way to assist in financial relationship building

Procedia PDF Downloads 155
25979 Arabic Lexicon Learning to Analyze Sentiment in Microblogs

Authors: Mahmoud B. Rokaya

Abstract:

The study of opinion mining and sentiment analysis includes analysis of opinions, sentiments, evaluations, attitudes, and emotions. The rapid growth of social media, social networks, reviews, forum discussions, microblogs, and Twitter, leads to a parallel growth in the field of sentiment analysis. The field of sentiment analysis tries to develop effective tools to make it possible to capture the trends of people. There are two approaches in the field, lexicon-based and corpus-based methods. A lexicon-based method uses a sentiment lexicon which includes sentiment words and phrases with assigned numeric scores. These scores reveal if sentiment phrases are positive or negative, their intensity, and/or their emotional orientations. Creation of manual lexicons is hard. This brings the need for adaptive automated methods for generating a lexicon. The proposed method generates dynamic lexicons based on the corpus and then classifies text using these lexicons. In the proposed method, different approaches are combined to generate lexicons from text. The proposed method classifies the tweets into 5 classes instead of +ve or –ve classes. The sentiment classification problem is written as an optimization problem, finding optimum sentiment lexicons are the goal of the optimization process. The solution was produced based on mathematical programming approaches to find the best lexicon to classify texts. A genetic algorithm was written to find the optimal lexicon. Then, extraction of a meta-level feature was done based on the optimal lexicon. The experiments were conducted on several datasets. Results, in terms of accuracy, recall and F measure, outperformed the state-of-the-art methods proposed in the literature in some of the datasets. A better understanding of the Arabic language and culture of Arab Twitter users and sentiment orientation of words in different contexts can be achieved based on the sentiment lexicons proposed by the algorithm.

Keywords: social media, Twitter sentiment, sentiment analysis, lexicon, genetic algorithm, evolutionary computation

Procedia PDF Downloads 170
25978 Optimal Planning of Transmission Line Charging Mode During Black Start of a Hydroelectric Unit

Authors: Mohammad Reza Esmaili

Abstract:

After the occurrence of blackouts, the most important subject is how fast the electric service is restored. Power system restoration is an immensely complex issue and there should be a plan to be executed within the shortest time period. This plan has three main stages of black start, network reconfiguration and load restoration. In the black start stage, operators and experts may face several problems, for instance, the unsuccessful connection of the long high-voltage transmission line connected to the electrical source. In this situation, the generator may be tripped because of the unsuitable setting of its line charging mode or high absorbed reactive power. In order to solve this problem, the line charging process is defined as a nonlinear programming problem, and it is optimized by using GAMS software in this paper. The optimized process is performed on a grid that includes a 250 MW hydroelectric unit and a 400 KV transmission system. Simulations and field test results show the effectiveness of optimal planning.

Keywords: power system restoration, black start, line charging mode, nonlinear programming

Procedia PDF Downloads 63
25977 Scalable Learning of Tree-Based Models on Sparsely Representable Data

Authors: Fares Hedayatit, Arnauld Joly, Panagiotis Papadimitriou

Abstract:

Many machine learning tasks such as text annotation usually require training over very big datasets, e.g., millions of web documents, that can be represented in a sparse input space. State-of the-art tree-based ensemble algorithms cannot scale to such datasets, since they include operations whose running time is a function of the input space size rather than a function of the non-zero input elements. In this paper, we propose an efficient splitting algorithm to leverage input sparsity within decision tree methods. Our algorithm improves training time over sparse datasets by more than two orders of magnitude and it has been incorporated in the current version of scikit-learn.org, the most popular open source Python machine learning library.

Keywords: big data, sparsely representable data, tree-based models, scalable learning

Procedia PDF Downloads 252
25976 Combining the Dynamic Conditional Correlation and Range-GARCH Models to Improve Covariance Forecasts

Authors: Piotr Fiszeder, Marcin Fałdziński, Peter Molnár

Abstract:

The dynamic conditional correlation model of Engle (2002) is one of the most popular multivariate volatility models. However, this model is based solely on closing prices. It has been documented in the literature that the high and low price of the day can be used in an efficient volatility estimation. We, therefore, suggest a model which incorporates high and low prices into the dynamic conditional correlation framework. Empirical evaluation of this model is conducted on three datasets: currencies, stocks, and commodity exchange-traded funds. The utilisation of realized variances and covariances as proxies for true variances and covariances allows us to reach a strong conclusion that our model outperforms not only the standard dynamic conditional correlation model but also a competing range-based dynamic conditional correlation model.

Keywords: volatility, DCC model, high and low prices, range-based models, covariance forecasting

Procedia PDF Downloads 169
25975 Ultra Reliable Communication: Availability Analysis in 5G Cellular Networks

Authors: Yosra Benchaabene, Noureddine Boujnah, Faouzi Zarai

Abstract:

To meet the growing demand of users, the fifth generation (5G) will continue to provide services to higher data rates with higher carrier frequencies and wider bandwidths. As part of the 5G communication paradigm, Ultra Reliable Communication (URC) is envisaged as an important technology pillar for providing anywhere and anytime services to end users. Ultra Reliable Communication (URC) is considered an important technology that why it has become an active research topic. In this work, we analyze the availability of a service in the space domain. We characterize spatially available areas consisting of all locations that meet a performance requirement with confidence, and we define cell availability and system availability, individual user availability, and user-oriented system availability. Poisson point process (PPP) and Voronoi tessellation are adopted to model the spatial characteristics of a cell deployment in heterogeneous networks. Numerical results are presented, also highlighting the effect of different system parameters on the achievable link availability.

Keywords: URC, dependability and availability, space domain analysis, Poisson point process, Voronoi Tessellation

Procedia PDF Downloads 107