Search results for: Computing Accreditation Committee
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1543

Search results for: Computing Accreditation Committee

1123 Steps towards the Development of National Health Data Standards in Developing Countries

Authors: Abdullah I. Alkraiji, Thomas W. Jackson, Ian Murray

Abstract:

The proliferation of health data standards today is somewhat overlapping and conflicting, resulting in market confusion and leading to increasing proprietary interests. The government role and support in standardization for health data are thought to be crucial in order to establish credible standards for the next decade, to maximize interoperability across the health sector, and to decrease the risks associated with the implementation of non-standard systems. The normative literature missed out the exploration of the different steps required to be undertaken by the government towards the development of national health data standards. Based on the lessons learned from a qualitative study investigating the different issues to the adoption of health data standards in the major tertiary hospitals in Saudi Arabia and the opinions and feedback from different experts in the areas of data exchange and standards and medical informatics in Saudi Arabia and UK, a list of steps required towards the development of national health data standards was constructed. Main steps are the existence of: a national formal reference for health data standards, an agreed national strategic direction for medical data exchange, a national medical information management plan and a national accreditation body, and more important is the change management at the national and organizational level. The outcome of this study can be used by academics and practitioners to develop the planning of health data standards, and in particular those in developing countries.

Keywords: interoperabilty, medical data exchange, health data standards, case study, Saudi Arabia

Procedia PDF Downloads 338
1122 Roasting Process of Sesame Seeds Modelling Using Gene Expression Programming: A Comparative Analysis with Response Surface Methodology

Authors: Alime Cengiz, Talip Kahyaoglu

Abstract:

Roasting process has the major importance to obtain desired aromatic taste of nuts. In this study, two kinds of roasting process were applied to hulled sesame seeds - vacuum oven and hot air roasting. Efficiency of Gene Expression Programming (GEP), a new soft computing technique of evolutionary algorithm that describes the cause and effect relationships in the data modelling system, and response surface methodology (RSM) were examined in the modelling of roasting processes over a range of temperature (120-180°C) for various times (30-60 min). Color attributes (L*, a*, b*, Browning Index (BI)), textural properties (hardness and fracturability) and moisture content were evaluated and modelled by RSM and GEP. The GEP-based formulations and RSM approach were compared with experimental results and evaluated according to correlation coefficients. The results showed that both GEP and RSM were found to be able to adequately learn the relation between roasting conditions and physical and textural parameters of roasted seeds. However, GEP had better prediction performance than the RSM with the high correlation coefficients (R2 >0.92) for the all quality parameters. This result indicates that the soft computing techniques have better capability for describing the physical changes occuring in sesame seeds during roasting process.

Keywords: genetic expression programming, response surface methodology, roasting, sesame seed

Procedia PDF Downloads 418
1121 Development of Geo-computational Model for Analysis of Lassa Fever Dynamics and Lassa Fever Outbreak Prediction

Authors: Adekunle Taiwo Adenike, I. K. Ogundoyin

Abstract:

Lassa fever is a neglected tropical virus that has become a significant public health issue in Nigeria, with the country having the greatest burden in Africa. This paper presents a Geo-Computational Model for Analysis and Prediction of Lassa Fever Dynamics and Outbreaks in Nigeria. The model investigates the dynamics of the virus with respect to environmental factors and human populations. It confirms the role of the rodent host in virus transmission and identifies how climate and human population are affected. The proposed methodology is carried out on a Linux operating system using the OSGeoLive virtual machine for geographical computing, which serves as a base for spatial ecology computing. The model design uses Unified Modeling Language (UML), and the performance evaluation uses machine learning algorithms such as random forest, fuzzy logic, and neural networks. The study aims to contribute to the control of Lassa fever, which is achievable through the combined efforts of public health professionals and geocomputational and machine learning tools. The research findings will potentially be more readily accepted and utilized by decision-makers for the attainment of Lassa fever elimination.

Keywords: geo-computational model, lassa fever dynamics, lassa fever, outbreak prediction, nigeria

Procedia PDF Downloads 93
1120 Improved Multi–Objective Firefly Algorithms to Find Optimal Golomb Ruler Sequences for Optimal Golomb Ruler Channel Allocation

Authors: Shonak Bansal, Prince Jain, Arun Kumar Singh, Neena Gupta

Abstract:

Recently nature–inspired algorithms have widespread use throughout the tough and time consuming multi–objective scientific and engineering design optimization problems. In this paper, we present extended forms of firefly algorithm to find optimal Golomb ruler (OGR) sequences. The OGRs have their one of the major application as unequally spaced channel–allocation algorithm in optical wavelength division multiplexing (WDM) systems in order to minimize the adverse four–wave mixing (FWM) crosstalk effect. The simulation results conclude that the proposed optimization algorithm has superior performance compared to the existing conventional computing and nature–inspired optimization algorithms to find OGRs in terms of ruler length, total optical channel bandwidth and computation time.

Keywords: channel allocation, conventional computing, four–wave mixing, nature–inspired algorithm, optimal Golomb ruler, lévy flight distribution, optimization, improved multi–objective firefly algorithms, Pareto optimal

Procedia PDF Downloads 320
1119 Participation of Women in the Brazilian Paralympic Sports

Authors: Ana Carolina Felizardo Da Silva

Abstract:

People with disabilities are those who have limitations of a physical, mental, intellectual or sensory nature and who, therefore, should not be excluded or marginalized. In Brazil, the Brazilian Law for the Inclusion of People with Disabilities defines that people with disabilities have the right to culture, sport, tourism and leisure on an equal basis with other people. Sport for people with disabilities, in its genesis, had a character aimed at rehabilitating men and soldiers, that is, the male figure who returned wounded from war and needed care. By gaining practitioners, the marketing issue emerges and, successively, high performance, what we call Paralympic sport. We found that sport for people with disabilities was designed for men, corroborating the social idea that sport is a masculine and masculinizing environment. In this way, the inclusion of women with disabilities in sports becomes a double challenge because they are women and have a disability. From data collected from official documents of the International Paralympic Committee, it is found that the first report on the participation of women in the Paralympic Games was in 1948, in England, in Stoke Mandeville, a championship considered the firstborn of the games, later, became called the “Paralympic Games”. However, due to the lack of information, the return of the appearance of women participating in the Paralympics took place after long 40 years, in 1984, which demonstrates a large gap of records on the official website referring to women in the games. Despite the great challenge, the number of women has been growing substantially. When collecting data from participants of all 16 editions of the Paralympic Games, in its last edition, held in Tokyo, out of 4,400 competing athletes, 1,853 were women, which represents 42% of the total number of athletes. In this same edition, we had the largest delegation of Brazilian women, represented by 96 athletes out of a total of 260 Brazilian athletes. It is estimated that in the next edition, to be taken place in Paris in 2024, the participation of women will equal or surpass that of men. The certain invisibility of women participating in the Paralympic Games is noticed when we access the database of the Brazilian Paralympic Committee website. It is possible to identify all women medalists of a given edition. On the other side, participating female athletes who did not medal are not registered on the site. Regarding the participation of Brazilian women in the Paralympics, there was a considerable growth in the last two editions, in 2012 there were only 69 women participating, going to 102 in 2016 and 96 in 2021. The same happened in relation to the medalists, going from 8 Brazilians in 2012 to 33 in 2016 and 27 in 2021. In this sense, the present study, aims to analyze how Brazilian women participate in the Paralympics, giving visibility and voice to female athletes. Structured interviews are being carried out with the participants of the games, identifying the difficulties and potentialities of participating with athletes in the competition. The analysis will be carried out through Bardin’s content analysis.

Keywords: paralympics, sport for people with disabilities, woman, woman in sport

Procedia PDF Downloads 73
1118 The Agency of Award Systems in Architecture: The Case of Cyprus

Authors: Christakis Chatzichristou, Elias Kranos

Abstract:

Architectural awards, especially if they are given by the state, recognize excellence in the field and, at the same time, strongly contribute to the making of the architectural culture of a place. The present research looks at the houses that have been awarded through such a system in Cyprus in order to discuss the values promoted, directly or not, by such a setup which is quite similar to other prestigious award systems such as the Mies van de Rohe Prize in Europe. In fact, many of the projects signed out through the state award system end up being selected to represent the country for the European awards. The residential architecture encouraged by such systems is quite interesting in that the most public of institutions influence how the most private unit of society is architecturally accommodated. The methodology uses both qualitative as well as quantitative research tools in order to analyze: the official state call for entries to the competition; the final report of the evaluation committee; the spatial characteristics of the houses through the Space Syntax methodology; the statements of the architects regarding their intentions and the final outcome; the feelings of the owners and users of the houses regarding the design process as well as the degree of satisfaction regarding the final product. The above-mentioned analyses allow for a more thorough discussion regarding not only the values promoted explicitly by the system through the brief that describes what the evaluation committee is looking for but also the values that are actually being promoted indirectly through the results of the actual evaluation itself. The findings suggest that: the strong emphasis in brief on bioclimatic design and issues of sustainability weakens significantly, if at all present, in the actual selection process; continuous improvement seems to be fuzzily used as a concept; most of the houses tend to have a similar spatial genotype; most of the houses have similar aesthetic qualities; discrepancies between the proposed lifestyle through the design and the actual use of the spaces do not seem to be acknowledged in the evaluation as an issue; the temporal factor seems to be ignored as the projects are required to be ‘finished projects’ as though the users and their needs do not change through time. The research suggests that, rather than preserving a critical attitude regarding the role of the architect in society, the state award system tends, like any other non-reflective social organism, to simply promote its own unexamined values as well as prejudices. This is perhaps more evident in the shared aesthetic character of the awarded houses and less so in the hidden spatial genotype to which they belong. If the design of houses is indeed a great opportunity for architecture to contribute in a more deliberate manner to the evolution of society, then what the present study shows is that this opportunity seems to be largely missed. The findings may serve better less as a verdict and more as a chance for introspection and discussion.

Keywords: award systems, houses, spatial genotype, aesthetic qualities

Procedia PDF Downloads 70
1117 The Use of Empirical Models to Estimate Soil Erosion in Arid Ecosystems and the Importance of Native Vegetation

Authors: Meshal M. Abdullah, Rusty A. Feagin, Layla Musawi

Abstract:

When humans mismanage arid landscapes, soil erosion can become a primary mechanism that leads to desertification. This study focuses on applying soil erosion models to a disturbed landscape in Umm Nigga, Kuwait, and identifying its predicted change under restoration plans, The northern portion of Umm Nigga, containing both coastal and desert ecosystems, falls within the boundaries of the Demilitarized Zone (DMZ) adjacent to Iraq, and has been fenced off to restrict public access since 1994. The central objective of this project was to utilize GIS and remote sensing to compare the MPSIAC (Modified Pacific South West Inter Agency Committee), EMP (Erosion Potential Method), and USLE (Universal Soil Loss Equation) soil erosion models and determine their applicability for arid regions such as Kuwait. Spatial analysis was used to develop the necessary datasets for factors such as soil characteristics, vegetation cover, runoff, climate, and topography. Results showed that the MPSIAC and EMP models produced a similar spatial distribution of erosion, though the MPSIAC had more variability. For the MPSIAC model, approximately 45% of the land surface ranged from moderate to high soil loss, while 35% ranged from moderate to high for the EMP model. The USLE model had contrasting results and a different spatial distribution of the soil loss, with 25% of area ranging from moderate to high erosion, and 75% ranging from low to very low. We concluded that MPSIAC and EMP were the most suitable models for arid regions in general, with the MPSIAC model best. We then applied the MPSIAC model to identify the amount of soil loss between coastal and desert areas, and fenced and unfenced sites. In the desert area, soil loss was different between fenced and unfenced sites. In these desert fenced sites, 88% of the surface was covered with vegetation and soil loss was very low, while at the desert unfenced sites it was 3% and correspondingly higher. In the coastal areas, the amount of soil loss was nearly similar between fenced and unfenced sites. These results implied that vegetation cover played an important role in reducing soil erosion, and that fencing is much more important in the desert ecosystems to protect against overgrazing. When applying the MPSIAC model predictively, we found that vegetation cover could be increased from 3% to 37% in unfenced areas, and soil erosion could then decrease by 39%. We conclude that the MPSIAC model is best to predict soil erosion for arid regions such as Kuwait.

Keywords: soil erosion, GIS, modified pacific South west inter agency committee model (MPSIAC), erosion potential method (EMP), Universal soil loss equation (USLE)

Procedia PDF Downloads 297
1116 Autism and Mental Health - How Different Individuals Are Impacted

Authors: Kerryn Burgoyne

Abstract:

Statement of the Problem: Women who suffer mental health issues, because of Autism Spectrum Disorder has a significant impact on their lives, especially if they’ve been bullied or discriminated against for the majority of their lives. Autism can impact one's mental health in many ways (child like behaviour), social anxieties or overload. The impact of mental health can also be experienced when the person does not have a good quality of life for themselves (eg employment, independent living skills), or have support from family/friends/society). Mental health issues were also suffered during COVID 19 Lockdown here in Melbourne Australia. It was stated by the Government at the time that people weren’t allowed to travel more than 5 km outside of their residential areas to prevent the spread of COVID to others. Medical appointments were an exception. Kerryn/KTalk will be doing a paper on this topic for the conference if accepted by the committee.

Keywords: Autism, mental health, living & learning, KTalk

Procedia PDF Downloads 36
1115 The Impact of Artificial Intelligence on Food Nutrition

Authors: Antonyous Fawzy Boshra Girgis

Abstract:

Nutrition labels are diet-related health policies. They help individuals improve food-choice decisions and reduce intake of calories and unhealthy food elements, like cholesterol. However, many individuals do not pay attention to nutrition labels or fail to appropriately understand them. According to the literature, thinking and cognitive styles can have significant effects on attention to nutrition labels. According to the author's knowledge, the effect of global/local processing on attention to nutrition labels has not been previously studied. Global/local processing encourages individuals to attend to the whole/specific parts of an object and can have a significant impact on people's visual attention. In this study, this effect was examined with an experimental design using the eye-tracking technique. The research hypothesis was that individuals with local processing would pay more attention to nutrition labels, including nutrition tables and traffic lights. An experiment was designed with two conditions: global and local information processing. Forty participants were randomly assigned to either global or local conditions, and their processing style was manipulated accordingly. Results supported the hypothesis for nutrition tables but not for traffic lights.

Keywords: nutrition, public health, SA Harvest, foodeye-tracking, nutrition labelling, global/local information processing, individual differencesmobile computing, cloud computing, nutrition label use, nutrition management, barcode scanning

Procedia PDF Downloads 40
1114 Creating Energy Sustainability in an Enterprise

Authors: John Lamb, Robert Epstein, Vasundhara L. Bhupathi, Sanjeev Kumar Marimekala

Abstract:

As we enter the new era of Artificial Intelligence (AI) and Cloud Computing, we mostly rely on the Machine and Natural Language Processing capabilities of AI, and Energy Efficient Hardware and Software Devices in almost every industry sector. In these industry sectors, much emphasis is on developing new and innovative methods for producing and conserving energy and sustaining the depletion of natural resources. The core pillars of sustainability are economic, environmental, and social, which is also informally referred to as the 3 P's (People, Planet and Profits). The 3 P's play a vital role in creating a core Sustainability Model in the Enterprise. Natural resources are continually being depleted, so there is more focus and growing demand for renewable energy. With this growing demand, there is also a growing concern in many industries on how to reduce carbon emissions and conserve natural resources while adopting sustainability in corporate business models and policies. In our paper, we would like to discuss the driving forces such as Climate changes, Natural Disasters, Pandemic, Disruptive Technologies, Corporate Policies, Scaled Business Models and Emerging social media and AI platforms that influence the 3 main pillars of Sustainability (3P’s). Through this paper, we would like to bring an overall perspective on enterprise strategies and the primary focus on bringing cultural shifts in adapting energy-efficient operational models. Overall, many industries across the globe are incorporating core sustainability principles such as reducing energy costs, reducing greenhouse gas (GHG) emissions, reducing waste and increasing recycling, adopting advanced monitoring and metering infrastructure, reducing server footprint and compute resources (Shared IT services, Cloud computing, and Application Modernization) with the vision for a sustainable environment.

Keywords: climate change, pandemic, disruptive technology, government policies, business model, machine learning and natural language processing, AI, social media platform, cloud computing, advanced monitoring, metering infrastructure

Procedia PDF Downloads 111
1113 Ecotourism Sites in Central Visayas, Philippines: A Green Business Profile

Authors: Ivy Jumao-As, Randy Lupango, Clifford Villaflores, Marites Khanser

Abstract:

Alongside inadequate implementation of ecotourism standards and other pressing issues on sustainable development is the lack of business plans and formal business structures of various ecotourism sites in the Central Visayas, Philippines, and other parts of the country. Addressing these issues plays a key role to boost ecotourism which is a sustainability tool to the country’s economic development. A three-phase research is designed to investigate the green business practices of selected ecotourism sites in the region in order to propose a business model for ecotourism destinations in the region and outside. This paper reports the initial phase of the study which described the sites’ profile as well as operators of the following selected destinations: Cebu City Protected Landscape and Olango Island Wildlife Bird Sanctuary in Cebu, Rajah Sikatuna Protected Landscape in Bohol. Interview, Self-Administered Questionnaire with key informants and Data Mining were employed in the data collection. Findings highlighted similarities and differences in terms of eco-tourism products, type and number of visitors, manpower composition, cultural and natural resources, complementary services and products, awards and accreditation, peak and off peak seasons, among others. Recommendations based from common issues initially identified in this study are also highlighted.

Keywords: ecotourism, ecotourism sites, green business, sustainability

Procedia PDF Downloads 270
1112 e-Learning Security: A Distributed Incident Response Generator

Authors: Bel G Raggad

Abstract:

An e-Learning setting is a distributed computing environment where information resources can be connected to any public network. Public networks are very unsecure which can compromise the reliability of an e-Learning environment. This study is only concerned with the intrusion detection aspect of e-Learning security and how incident responses are planned. The literature reported great advances in intrusion detection system (ids) but neglected to study an important ids weakness: suspected events are detected but an intrusion is not determined because it is not defined in ids databases. We propose an incident response generator (DIRG) that produces incident responses when the working ids system suspects an event that does not correspond to a known intrusion. Data involved in intrusion detection when ample uncertainty is present is often not suitable to formal statistical models including Bayesian. We instead adopt Dempster and Shafer theory to process intrusion data for the unknown event. The DIRG engine transforms data into a belief structure using incident scenarios deduced by the security administrator. Belief values associated with various incident scenarios are then derived and evaluated to choose the most appropriate scenario for which an automatic incident response is generated. This article provides a numerical example demonstrating the working of the DIRG system.

Keywords: decision support system, distributed computing, e-Learning security, incident response, intrusion detection, security risk, statefull inspection

Procedia PDF Downloads 437
1111 Phenotype Prediction of DNA Sequence Data: A Machine and Statistical Learning Approach

Authors: Mpho Mokoatle, Darlington Mapiye, James Mashiyane, Stephanie Muller, Gciniwe Dlamini

Abstract:

Great advances in high-throughput sequencing technologies have resulted in availability of huge amounts of sequencing data in public and private repositories, enabling a holistic understanding of complex biological phenomena. Sequence data are used for a wide range of applications such as gene annotations, expression studies, personalized treatment and precision medicine. However, this rapid growth in sequence data poses a great challenge which calls for novel data processing and analytic methods, as well as huge computing resources. In this work, a machine and statistical learning approach for DNA sequence classification based on $k$-mer representation of sequence data is proposed. The approach is tested using whole genome sequences of Mycobacterium tuberculosis (MTB) isolates to (i) reduce the size of genomic sequence data, (ii) identify an optimum size of k-mers and utilize it to build classification models, (iii) predict the phenotype from whole genome sequence data of a given bacterial isolate, and (iv) demonstrate computing challenges associated with the analysis of whole genome sequence data in producing interpretable and explainable insights. The classification models were trained on 104 whole genome sequences of MTB isoloates. Cluster analysis showed that k-mers maybe used to discriminate phenotypes and the discrimination becomes more concise as the size of k-mers increase. The best performing classification model had a k-mer size of 10 (longest k-mer) an accuracy, recall, precision, specificity, and Matthews Correlation coeffient of 72.0%, 80.5%, 80.5%, 63.6%, and 0.4 respectively. This study provides a comprehensive approach for resampling whole genome sequencing data, objectively selecting a k-mer size, and performing classification for phenotype prediction. The analysis also highlights the importance of increasing the k-mer size to produce more biological explainable results, which brings to the fore the interplay that exists amongst accuracy, computing resources and explainability of classification results. However, the analysis provides a new way to elucidate genetic information from genomic data, and identify phenotype relationships which are important especially in explaining complex biological mechanisms.

Keywords: AWD-LSTM, bootstrapping, k-mers, next generation sequencing

Procedia PDF Downloads 167
1110 Phenotype Prediction of DNA Sequence Data: A Machine and Statistical Learning Approach

Authors: Darlington Mapiye, Mpho Mokoatle, James Mashiyane, Stephanie Muller, Gciniwe Dlamini

Abstract:

Great advances in high-throughput sequencing technologies have resulted in availability of huge amounts of sequencing data in public and private repositories, enabling a holistic understanding of complex biological phenomena. Sequence data are used for a wide range of applications such as gene annotations, expression studies, personalized treatment and precision medicine. However, this rapid growth in sequence data poses a great challenge which calls for novel data processing and analytic methods, as well as huge computing resources. In this work, a machine and statistical learning approach for DNA sequence classification based on k-mer representation of sequence data is proposed. The approach is tested using whole genome sequences of Mycobacterium tuberculosis (MTB) isolates to (i) reduce the size of genomic sequence data, (ii) identify an optimum size of k-mers and utilize it to build classification models, (iii) predict the phenotype from whole genome sequence data of a given bacterial isolate, and (iv) demonstrate computing challenges associated with the analysis of whole genome sequence data in producing interpretable and explainable insights. The classification models were trained on 104 whole genome sequences of MTB isoloates. Cluster analysis showed that k-mers maybe used to discriminate phenotypes and the discrimination becomes more concise as the size of k-mers increase. The best performing classification model had a k-mer size of 10 (longest k-mer) an accuracy, recall, precision, specificity, and Matthews Correlation coeffient of 72.0 %, 80.5 %, 80.5 %, 63.6 %, and 0.4 respectively. This study provides a comprehensive approach for resampling whole genome sequencing data, objectively selecting a k-mer size, and performing classification for phenotype prediction. The analysis also highlights the importance of increasing the k-mer size to produce more biological explainable results, which brings to the fore the interplay that exists amongst accuracy, computing resources and explainability of classification results. However, the analysis provides a new way to elucidate genetic information from genomic data, and identify phenotype relationships which are important especially in explaining complex biological mechanisms

Keywords: AWD-LSTM, bootstrapping, k-mers, next generation sequencing

Procedia PDF Downloads 159
1109 Tradition and Modernity in Translation Studies: The Case of Undergraduate and Graduate Programs at Unicamp, Brazil

Authors: Erica Lima

Abstract:

In Brazil, considering the (little) age of translation studies, it can be argued that the University of Campinas is traditionally an important place for graduate studies in translation. The story is told from the accreditation for the Masters, in 1987, and the Doctoral program, in 1993, within the Graduate Program in Applied Linguistics. Since the beginning, the program boasted cutting-edge research, with theoretical reflections on various aspects, and with different methodological trends. However, on the one hand, the graduate studies development was continuously growing, but on the other, it is not what was observed in the undergraduate degree program. Currently, there are only a few disciplines of Translation Theory and Practice, which does not seem to respond to student aspirations. The objective of this paper is to present the characteristics of the university’s graduate program as something profitable, considering the concern in relating the research to the historical moment in which we are living, with research conducted in a socially compromised environment and committed to the impact that it will cause ethically and socially, as well as to question the undergraduate program paths. The objective is also to discuss and propose changes, considering the limited scope currently achieved. In light of the information age, in which we have an avalanche of information, we believe that the training of translators in the undergraduate degree should be reviewed, with the goal of retracing current paths and following others that are consistent with our historical period, marked by virtual and real, by the shuffling of borders and languages, the need for new language policies, greater inclusion, and more acceptance of others. We conclude that we need new proposals for the development of the translator in an undergraduate program, and also present suggestions to be implemented in the graduate program.

Keywords: graduate Brazilian program, undergraduate Brazilian program, translator’s education, Unicamp

Procedia PDF Downloads 334
1108 Robust Recognition of Locomotion Patterns via Data-Driven Machine Learning in the Cloud Environment

Authors: Shinoy Vengaramkode Bhaskaran, Kaushik Sathupadi, Sandesh Achar

Abstract:

Human locomotion recognition is important in a variety of sectors, such as robotics, security, healthcare, fitness tracking and cloud computing. With the increasing pervasiveness of peripheral devices, particularly Inertial Measurement Units (IMUs) sensors, researchers have attempted to exploit these advancements in order to precisely and efficiently identify and categorize human activities. This research paper introduces a state-of-the-art methodology for the recognition of human locomotion patterns in a cloud environment. The methodology is based on a publicly available benchmark dataset. The investigation implements a denoising and windowing strategy to deal with the unprocessed data. Next, feature extraction is adopted to abstract the main cues from the data. The SelectKBest strategy is used to abstract optimal features from the data. Furthermore, state-of-the-art ML classifiers are used to evaluate the performance of the system, including logistic regression, random forest, gradient boosting and SVM have been investigated to accomplish precise locomotion classification. Finally, a detailed comparative analysis of results is presented to reveal the performance of recognition models.

Keywords: artificial intelligence, cloud computing, IoT, human locomotion, gradient boosting, random forest, neural networks, body-worn sensors

Procedia PDF Downloads 11
1107 Wheel Diameter and Width Influence in Variability of Brake Data Measurement at Ministry of Transport Facilities

Authors: Carolina Senabre, Sergio Valero, Emilio Velasco

Abstract:

The brake systems of vehicles are tested periodically by a “brake tester” at Ministry of Transport (MOT) stations. This tester measures the effectiveness of vehicle. This parameter is established by the International Committee of Vehicle Inspection (CITA). In this paper, we present an investigation of the influence of the tire size on the measurements of brake force on three MOT brake testers. We performed an analysis of the vehicle braking capacity test at MOT stations. The influence of varying wheel diameter and width on the measurement of braking at MOT stations has been analyzed. Thereby, the MOT brake tester as a verification system for a vehicle has been evaluated.

Keywords: brake tester, ministry of transport facilities, wheel diameter, efficiency

Procedia PDF Downloads 375
1106 Hybrid Genetic Approach for Solving Economic Dispatch Problems with Valve-Point Effect

Authors: Mohamed I. Mahrous, Mohamed G. Ashmawy

Abstract:

Hybrid genetic algorithm (HGA) is proposed in this paper to determine the economic scheduling of electric power generation over a fixed time period under various system and operational constraints. The proposed technique can outperform conventional genetic algorithms (CGAs) in the sense that HGA make it possible to improve both the quality of the solution and reduce the computing expenses. In contrast, any carefully designed GA is only able to balance the exploration and the exploitation of the search effort, which means that an increase in the accuracy of a solution can only occure at the sacrifice of convergent speed, and vice visa. It is unlikely that both of them can be improved simultaneously. The proposed hybrid scheme is developed in such a way that a simple GA is acting as a base level search, which makes a quick decision to direct the search towards the optimal region, and a local search method (pattern search technique) is next employed to do the fine tuning. The aim of the strategy is to achieve the cost reduction within a reasonable computing time. The effectiveness of the proposed hybrid technique is verified on two real public electricity supply systems with 13 and 40 generator units respectively. The simulation results obtained with the HGA for the two real systems are very encouraging with regard to the computational expenses and the cost reduction of power generation.

Keywords: genetic algorithms, economic dispatch, pattern search

Procedia PDF Downloads 444
1105 An Analysis of Innovative Cloud Model as Bridging the Gap between Physical and Virtualized Business Environments: The Customer Perspective

Authors: Asim Majeed, Rehan Bhana, Mak Sharma, Rebecca Goode, Nizam Bolia, Mike Lloyd-Williams

Abstract:

This study aims to investigate and explore the underlying causes of security concerns of customers emerged when WHSmith transformed its physical system to virtualized business model through NetSuite. NetSuite is essentially fully integrated software which helps transforming the physical system to virtualized business model. Modern organisations are moving away from traditional business models to cloud based models and consequently it is expected to have a better, secure and innovative environment for customers. The vital issue of the modern age race is the security when transforming virtualized through cloud based models and designers of interactive systems often misunderstand privacy and even often ignore it, thus causing concerns for users. The content analysis approach is being used to collect the qualitative data from 120 online bloggers including TRUSTPILOT. The results and finding provide useful new insights into the nature and form of security concerns of online users after they have used the WHSmith services offered online through their website. Findings have theoretical as well as practical implications for the successful adoption of cloud computing Business-to-Business model and similar systems.

Keywords: innovation, virtualization, cloud computing, organizational flexibility

Procedia PDF Downloads 384
1104 Development of People's Participation in Environmental Development in Pathumthani Province

Authors: Sakapas Saengchai

Abstract:

Study on the development of people's participation in environmental development was a qualitative research method. Data were collected by participant observation, in-depth interview and discussion group in Pathumthani province. The study indicated that 1) People should be aware of environmental information from government agencies. 2) People in the community should be able to brainstorm information, exchange information about community environment development. 3) People should have a role with community leaders. 4) People in the community should have a role to play in the implementation of projects and activities in the development of the environment and 5) citizens, community leaders, village committee have directed the development of the area. Maintaining a community environment with a shared decision. By emphasizing the process of participation, self-reliance, mutual help, and responsibility for one's own community. Community empowerment strengthens the sustainable spatial development of the environment.

Keywords: people, participation, community, environment

Procedia PDF Downloads 280
1103 International Protection Mechanisms for Refugees

Authors: Djehich Mohamed Yousri

Abstract:

In recent years, the world has witnessed a phenomenon of displacement that is unprecedented in history. The number of refugees has reached record levels, due to wars, persecution, many conflicts and repression in a number of countries. The interest of United Nations bodies and international and regional organizations in the issue of refugees has increased, as they have defined a refugee and thus Determining who is entitled to this legal protection, and the 1951 Convention for the Protection of Refugees defines rights for refugee protection and sets obligations that they must perform. The institutional mechanisms for refugee protection are represented in the various agencies that take care of refugee affairs. At the forefront of these agencies is the United Nations High Commissioner for Refugees, as well as the various efforts provided by the International Committee of the Red Cross and the United Nations Relief and Works Agency for Palestine Refugees in the Middle East (UNRWA).

Keywords: protection, refugees, international, persecution, legal

Procedia PDF Downloads 78
1102 Examination Scheduling System with Proposed Algorithm

Authors: Tabrej Khan

Abstract:

Examination Scheduling System (ESS) is a scheduling system that targets as an exam committee in any academic institute to help them in managing the exams automatically. We present an algorithm for Examination Scheduling System. Nowadays, many universities have challenges with creating examination schedule fast with less confliction compared to hand works. Our aims are to develop a computerized system that can be used in examination scheduling in an academic institute versus available resources (Time, Hall, Invigilator and instructor) with no contradiction and achieve fairness among students. ESS was developed using HTML, C# language, Crystal Report and ASP.NET through Microsoft Visual Studio 2010 as developing tools with integrated SQL server database. This application can produce some benefits such as reducing the time spent in creating an exam schedule and achieving fairness among students

Keywords: examination scheduling system (ESS), algorithm, ASP.NET, crystal report

Procedia PDF Downloads 404
1101 Protocol for Dynamic Load Distributed Low Latency Web-Based Augmented Reality and Virtual Reality

Authors: Rohit T. P., Sahil Athrij, Sasi Gopalan

Abstract:

Currently, the content entertainment industry is dominated by mobile devices. As the trends slowly shift towards Augmented/Virtual Reality applications the computational demands on these devices are increasing exponentially and we are already reaching the limits of hardware optimizations. This paper proposes a software solution to this problem. By leveraging the capabilities of cloud computing we can offload the work from mobile devices to dedicated rendering servers that are way more powerful. But this introduces the problem of latency. This paper introduces a protocol that can achieve high-performance low latency Augmented/Virtual Reality experience. There are two parts to the protocol, 1) In-flight compression The main cause of latency in the system is the time required to transmit the camera frame from client to server. The round trip time is directly proportional to the amount of data transmitted. This can therefore be reduced by compressing the frames before sending. Using some standard compression algorithms like JPEG can result in minor size reduction only. Since the images to be compressed are consecutive camera frames there won't be a lot of changes between two consecutive images. So inter-frame compression is preferred. Inter-frame compression can be implemented efficiently using WebGL but the implementation of WebGL limits the precision of floating point numbers to 16bit in most devices. This can introduce noise to the image due to rounding errors, which will add up eventually. This can be solved using an improved interframe compression algorithm. The algorithm detects changes between frames and reuses unchanged pixels from the previous frame. This eliminates the need for floating point subtraction thereby cutting down on noise. The change detection is also improved drastically by taking the weighted average difference of pixels instead of the absolute difference. The kernel weights for this comparison can be fine-tuned to match the type of image to be compressed. 2) Dynamic Load distribution Conventional cloud computing architectures work by offloading as much work as possible to the servers, but this approach can cause a hit on bandwidth and server costs. The most optimal solution is obtained when the device utilizes 100% of its resources and the rest is done by the server. The protocol balances the load between the server and the client by doing a fraction of the computing on the device depending on the power of the device and network conditions. The protocol will be responsible for dynamically partitioning the tasks. Special flags will be used to communicate the workload fraction between the client and the server and will be updated in a constant interval of time ( or frames ). The whole of the protocol is designed so that it can be client agnostic. Flags are available to the client for resetting the frame, indicating latency, switching mode, etc. The server can react to client-side changes on the fly and adapt accordingly by switching to different pipelines. The server is designed to effectively spread the load and thereby scale horizontally. This is achieved by isolating client connections into different processes.

Keywords: 2D kernelling, augmented reality, cloud computing, dynamic load distribution, immersive experience, mobile computing, motion tracking, protocols, real-time systems, web-based augmented reality application

Procedia PDF Downloads 72
1100 Soft Computing Approach for Diagnosis of Lassa Fever

Authors: Roseline Oghogho Osaseri, Osaseri E. I.

Abstract:

Lassa fever is an epidemic hemorrhagic fever caused by the Lassa virus, an extremely virulent arena virus. This highly fatal disorder kills 10% to 50% of its victims, but those who survive its early stages usually recover and acquire immunity to secondary attacks. One of the major challenges in giving proper treatment is lack of fast and accurate diagnosis of the disease due to multiplicity of symptoms associated with the disease which could be similar to other clinical conditions and makes it difficult to diagnose early. This paper proposed an Adaptive Neuro Fuzzy Inference System (ANFIS) for the prediction of Lass Fever. In the design of the diagnostic system, four main attributes were considered as the input parameters and one output parameter for the system. The input parameters are Temperature on admission (TA), White Blood Count (WBC), Proteinuria (P) and Abdominal Pain (AP). Sixty-one percent of the datasets were used in training the system while fifty-nine used in testing. Experimental results from this study gave a reliable and accurate prediction of Lassa fever when compared with clinically confirmed cases. In this study, we have proposed Lassa fever diagnostic system to aid surgeons and medical healthcare practictionals in health care facilities who do not have ready access to Polymerase Chain Reaction (PCR) diagnosis to predict possible Lassa fever infection.

Keywords: anfis, lassa fever, medical diagnosis, soft computing

Procedia PDF Downloads 269
1099 A Fast Parallel and Distributed Type-2 Fuzzy Algorithm Based on Cooperative Mobile Agents Model for High Performance Image Processing

Authors: Fatéma Zahra Benchara, Mohamed Youssfi, Omar Bouattane, Hassan Ouajji, Mohamed Ouadi Bensalah

Abstract:

The aim of this paper is to present a distributed implementation of the Type-2 Fuzzy algorithm in a parallel and distributed computing environment based on mobile agents. The proposed algorithm is assigned to be implemented on a SPMD (Single Program Multiple Data) architecture which is based on cooperative mobile agents as AVPE (Agent Virtual Processing Element) model in order to improve the processing resources needed for performing the big data image segmentation. In this work we focused on the application of this algorithm in order to process the big data MRI (Magnetic Resonance Images) image of size (n x m). It is encapsulated on the Mobile agent team leader in order to be split into (m x n) pixels one per AVPE. Each AVPE perform and exchange the segmentation results and maintain asynchronous communication with their team leader until the convergence of this algorithm. Some interesting experimental results are obtained in terms of accuracy and efficiency analysis of the proposed implementation, thanks to the mobile agents several interesting skills introduced in this distributed computational model.

Keywords: distributed type-2 fuzzy algorithm, image processing, mobile agents, parallel and distributed computing

Procedia PDF Downloads 428
1098 Decentralized Forest Policy for Natural Sal (Shorea robusta) Forests Management in the Terai Region of Nepal

Authors: Medani Prasad Rijal

Abstract:

The study outlines the impacts of decentralized forest policy on natural Sal (shorea robusta) forests in the Terai region of Nepal. The government has implemented community forestry program to manage the forest resources and improve the livelihood of local people collectively. The forest management authorities such as conserve, manage, develop and use of forest resources were shifted to the local communities, however, the ownership right of the forestland retained by the government. Local communities took the decision on harvesting, distribution, and sell of forest products by fixing the prices independently. The local communities were putting the low value of forest products and distributed among the user households on the name of collective decision. The decision of low valuation is devaluating the worth of forest products. Therefore, the study hypothesized that decision-making capacities are equally prominent next to the decentralized policy and program formulation. To accomplish the study, individual to group level discussions and questionnaire survey methods were applied with executive committee members and user households. The study revealed that the local intuition called Community Forest User Group (CFUG) committee normally took the decisions on consensus basis. Considering to the access and affording capacity of user households having poor economic backgrounds, low pricing mechanism of forest products has been practiced, even though the Sal timber is far expensive in the local market. The local communities thought that low pricing mechanism is accessible to all user households from poor to better off households. However, the analysis of forest products distribution opposed the assumption as most of the Sal timber, which is the most valuable forest product of community forest only purchased by the limited households of better economic conditions. Since the Terai region is heterogeneous by socio-economic conditions, better off households always have higher affording capacity and possibility of taking higher timber benefits because of low price mechanism. On the other hand, the minimum price rate of forest products has poor contribution in community fund collection. Consequently, it has poor support to carry out poverty alleviation activities to poor people. The local communities have been fixed Sal timber price rate around three times cheaper than normal market price, which is a strong evidence of forest product devaluation itself. Finally, the study concluded that the capacity building of local executives as the decision-makers of natural Sal forests is equally indispensable next to the policy and program formulation for effective decentralized forest management. Unilateral decentralized forest policy may devaluate the forest products rather than devolve of power to the local communities and empower to them.

Keywords: community forestry program, decentralized forest policy, Nepal, Sal forests, Terai

Procedia PDF Downloads 333
1097 Integrated Teaching of Hardware Courses for the Undergraduates of Computer Science and Engineering to Attain Focused Outcomes

Authors: Namrata D. Hiremath, Mahalaxmi Bhille, P. G. Sunitha Hiremath

Abstract:

Computer systems play an integral role in all facets of the engineering profession. This calls for an understanding of the processor-level components of computer systems, their design and operation, and their impact on the overall performance of the systems. Systems users are always in need of faster, more powerful, yet cheaper computer systems. The focus of Computer Science engineering graduates is inclined towards software oriented base. To be an efficient programmer there is a need to understand the role of hardware architecture towards the same. It is essential for the students of Computer Science and Engineering to know the basic building blocks of any computing device and how the digital principles can be used to build them. Hence two courses Digital Electronics of 3 credits, which is associated with lab of 1.5 credits and Computer Organization of 5 credits, were introduced at the sophomore level. Activity was introduced with the objective to teach the hardware concepts to the students of Computer science engineering through structured lab. The students were asked to design and implement a component of a computing device using MultiSim simulation tool and build the same using hardware components. The experience of the activity helped the students to understand the real time applications of the SSI and MSI components. The impact of the activity was evaluated and the performance was measured. The paper explains the achievement of the ABET outcomes a, c and k.

Keywords: digital, computer organization, ABET, structured enquiry, course activity

Procedia PDF Downloads 500
1096 Medi-Conf: Conference Management System

Authors: Dishant Kothari, Pankaj Gaur, Priyanshu Sharma, Ratnesh Litoriya, Sachin Solanki, Shimpy Goyal

Abstract:

Web based Conference Management System comprises of all the processes needed for round table conference, research paper publication includes the phases-call for paper, paper submission, paper review, acknowledgement to the author, paper acceptance and payment for publication. It will also help colleges and universities to conduct conferences for research, thus spreading awareness and will contribute to the overall development of students. Web based Conference Management System will streamline the procedure for paper publication by reducing the time and efforts needed in physical (offline mode) submission. A conference can be organized from anywhere and anytime. Authors can easily trace the status of the paper, and the program committee can review them anywhere and provide necessary comments to it.

Keywords: peer review, paper publication, author, chair, reviewer, virtualization, new normal

Procedia PDF Downloads 130
1095 Design of an Ensemble Learning Behavior Anomaly Detection Framework

Authors: Abdoulaye Diop, Nahid Emad, Thierry Winter, Mohamed Hilia

Abstract:

Data assets protection is a crucial issue in the cybersecurity field. Companies use logical access control tools to vault their information assets and protect them against external threats, but they lack solutions to counter insider threats. Nowadays, insider threats are the most significant concern of security analysts. They are mainly individuals with legitimate access to companies information systems, which use their rights with malicious intents. In several fields, behavior anomaly detection is the method used by cyber specialists to counter the threats of user malicious activities effectively. In this paper, we present the step toward the construction of a user and entity behavior analysis framework by proposing a behavior anomaly detection model. This model combines machine learning classification techniques and graph-based methods, relying on linear algebra and parallel computing techniques. We show the utility of an ensemble learning approach in this context. We present some detection methods tests results on an representative access control dataset. The use of some explored classifiers gives results up to 99% of accuracy.

Keywords: cybersecurity, data protection, access control, insider threat, user behavior analysis, ensemble learning, high performance computing

Procedia PDF Downloads 128
1094 SPARK: An Open-Source Knowledge Discovery Platform That Leverages Non-Relational Databases and Massively Parallel Computational Power for Heterogeneous Genomic Datasets

Authors: Thilina Ranaweera, Enes Makalic, John L. Hopper, Adrian Bickerstaffe

Abstract:

Data are the primary asset of biomedical researchers, and the engine for both discovery and research translation. As the volume and complexity of research datasets increase, especially with new technologies such as large single nucleotide polymorphism (SNP) chips, so too does the requirement for software to manage, process and analyze the data. Researchers often need to execute complicated queries and conduct complex analyzes of large-scale datasets. Existing tools to analyze such data, and other types of high-dimensional data, unfortunately suffer from one or more major problems. They typically require a high level of computing expertise, are too simplistic (i.e., do not fit realistic models that allow for complex interactions), are limited by computing power, do not exploit the computing power of large-scale parallel architectures (e.g. supercomputers, GPU clusters etc.), or are limited in the types of analysis available, compounded by the fact that integrating new analysis methods is not straightforward. Solutions to these problems, such as those developed and implemented on parallel architectures, are currently available to only a relatively small portion of medical researchers with access and know-how. The past decade has seen a rapid expansion of data management systems for the medical domain. Much attention has been given to systems that manage phenotype datasets generated by medical studies. The introduction of heterogeneous genomic data for research subjects that reside in these systems has highlighted the need for substantial improvements in software architecture. To address this problem, we have developed SPARK, an enabling and translational system for medical research, leveraging existing high performance computing resources, and analysis techniques currently available or being developed. It builds these into The Ark, an open-source web-based system designed to manage medical data. SPARK provides a next-generation biomedical data management solution that is based upon a novel Micro-Service architecture and Big Data technologies. The system serves to demonstrate the applicability of Micro-Service architectures for the development of high performance computing applications. When applied to high-dimensional medical datasets such as genomic data, relational data management approaches with normalized data structures suffer from unfeasibly high execution times for basic operations such as insert (i.e. importing a GWAS dataset) and the queries that are typical of the genomics research domain. SPARK resolves these problems by incorporating non-relational NoSQL databases that have been driven by the emergence of Big Data. SPARK provides researchers across the world with user-friendly access to state-of-the-art data management and analysis tools while eliminating the need for high-level informatics and programming skills. The system will benefit health and medical research by eliminating the burden of large-scale data management, querying, cleaning, and analysis. SPARK represents a major advancement in genome research technologies, vastly reducing the burden of working with genomic datasets, and enabling cutting edge analysis approaches that have previously been out of reach for many medical researchers.

Keywords: biomedical research, genomics, information systems, software

Procedia PDF Downloads 270