Search results for: user data security
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 27336

Search results for: user data security

24426 Facial Recognition and Landmark Detection in Fitness Assessment and Performance Improvement

Authors: Brittany Richardson, Ying Wang

Abstract:

For physical therapy, exercise prescription, athlete training, and regular fitness training, it is crucial to perform health assessments or fitness assessments periodically. An accurate assessment is propitious for tracking recovery progress, preventing potential injury and making long-range training plans. Assessments include necessary measurements, height, weight, blood pressure, heart rate, body fat, etc. and advanced evaluation, muscle group strength, stability-mobility, and movement evaluation, etc. In the current standard assessment procedures, the accuracy of assessments, especially advanced evaluations, largely depends on the experience of physicians, coaches, and personal trainers. And it is challenging to track clients’ progress in the current assessment. Unlike the tradition assessment, in this paper, we present a deep learning based face recognition algorithm for accurate, comprehensive and trackable assessment. Based on the result from our assessment, physicians, coaches, and personal trainers are able to adjust the training targets and methods. The system categorizes the difficulty levels of the current activity for the client or user, furthermore make more comprehensive assessments based on tracking muscle group over time using a designed landmark detection method. The system also includes the function of grading and correcting the form of the clients during exercise. Experienced coaches and personal trainer can tell the clients' limit based on their facial expression and muscle group movements, even during the first several sessions. Similar to this, using a convolution neural network, the system is trained with people’s facial expression to differentiate challenge levels for clients. It uses landmark detection for subtle changes in muscle groups movements. It measures the proximal mobility of the hips and thoracic spine, the proximal stability of the scapulothoracic region and distal mobility of the glenohumeral joint, as well as distal mobility, and its effect on the kinetic chain. This system integrates data from other fitness assistant devices, including but not limited to Apple Watch, Fitbit, etc. for a improved training and testing performance. The system itself doesn’t require history data for an individual client, but the history data of a client can be used to create a more effective exercise plan. In order to validate the performance of the proposed work, an experimental design is presented. The results show that the proposed work contributes towards improving the quality of exercise plan, execution, progress tracking, and performance.

Keywords: exercise prescription, facial recognition, landmark detection, fitness assessments

Procedia PDF Downloads 118
24425 System of Quality Automation for Documents (SQAD)

Authors: R. Babi Saraswathi, K. Divya, A. Habeebur Rahman, D. B. Hari Prakash, S. Jayanth, T. Kumar, N. Vijayarangan

Abstract:

Document automation is the design of systems and workflows, assembling repetitive documents to meet the specific business needs. In any organization or institution, documenting employee’s information is very important for both employees as well as management. It shows an individual’s progress to the management. Many documents of the employee are in the form of papers, so it is very difficult to arrange and for future reference we need to spend more time in getting the exact document. Also, it is very tedious to generate reports according to our needs. The process gets even more difficult on getting approvals and hence lacks its security aspects. This project overcomes the above-stated issues. By storing the details in the database and maintaining the e-documents, the automation system reduces the manual work to a large extent. Then the approval process of some important documents can be done in a much-secured manner by using Digital Signature and encryption techniques. Details are maintained in the database and e-documents are stored in specific folders and generation of various kinds of reports is possible. Moreover, an efficient search method is implemented is used in the database. Automation supporting document maintenance in many aspects is useful for minimize data entry, reduce the time spent on proof-reading, avoids duplication, and reduce the risks associated with the manual error, etc.

Keywords: e-documents, automation, digital signature, encryption

Procedia PDF Downloads 377
24424 Korean Smart Cities: Strategic Foci, Characteristics and Effects

Authors: Sang Ho Lee, Yountaik Leem

Abstract:

This paper reviews Korean cases of smart cities through the analysis framework of strategic foci, characteristics and effects. Firstly, national strategies including c(cyber), e(electronic), u(ubiquitous) and s(smart) Korea strategies were considered from strategic angles. Secondly, the characteristics of smart cities in Korea were looked through the smart cities examples such as Seoul, Busan, Songdo and Sejong cities etc. from the views on the by STIM (Service, Technology, Infrastructure and Management) analysis. Finally, the effects of smart cities on socio-economies were investigated from industrial perspective using the input-output model and structural path analysis. Korean smart city strategies revealed that there were different kinds of strategic foci. c-Korea strategy focused on information and communications network building and user IT literacy. e-Korea strategy encouraged e-government and e-business through utilizing high-speed information and communications network. u-Korea strategy made ubiquitous service as well as integrated information and communication operations center. s-Korea strategy is propelling 4th industrial platform. Smart cities in Korea showed their own features and trends such as eco-intelligence, high efficiency and low cost oriented IoT, citizen sensored city, big data city. Smart city progress made new production chains fostering ICTs (Information Communication Technologies) and knowledge intermediate inputs to industries.

Keywords: Korean smart cities, Korean smart city strategies, STIM, smart service, infrastructure, technologies, management, effect of smart city

Procedia PDF Downloads 356
24423 Generation of Automated Alarms for Plantwide Process Monitoring

Authors: Hyun-Woo Cho

Abstract:

Earlier detection of incipient abnormal operations in terms of plant-wide process management is quite necessary in order to improve product quality and process safety. And generating warning signals or alarms for operating personnel plays an important role in process automation and intelligent plant health monitoring. Various methodologies have been developed and utilized in this area such as expert systems, mathematical model-based approaches, multivariate statistical approaches, and so on. This work presents a nonlinear empirical monitoring methodology based on the real-time analysis of massive process data. Unfortunately, the big data includes measurement noises and unwanted variations unrelated to true process behavior. Thus the elimination of such unnecessary patterns of the data is executed in data processing step to enhance detection speed and accuracy. The performance of the methodology was demonstrated using simulated process data. The case study showed that the detection speed and performance was improved significantly irrespective of the size and the location of abnormal events.

Keywords: detection, monitoring, process data, noise

Procedia PDF Downloads 234
24422 Meanings and Concepts of Standardization in Systems Medicine

Authors: Imme Petersen, Wiebke Sick, Regine Kollek

Abstract:

In systems medicine, high-throughput technologies produce large amounts of data on different biological and pathological processes, including (disturbed) gene expressions, metabolic pathways and signaling. The large volume of data of different types, stored in separate databases and often located at different geographical sites have posed new challenges regarding data handling and processing. Tools based on bioinformatics have been developed to resolve the upcoming problems of systematizing, standardizing and integrating the various data. However, the heterogeneity of data gathered at different levels of biological complexity is still a major challenge in data analysis. To build multilayer disease modules, large and heterogeneous data of disease-related information (e.g., genotype, phenotype, environmental factors) are correlated. Therefore, a great deal of attention in systems medicine has been put on data standardization, primarily to retrieve and combine large, heterogeneous datasets into standardized and incorporated forms and structures. However, this data-centred concept of standardization in systems medicine is contrary to the debate in science and technology studies (STS) on standardization that rather emphasizes the dynamics, contexts and negotiations of standard operating procedures. Based on empirical work on research consortia that explore the molecular profile of diseases to establish systems medical approaches in the clinic in Germany, we trace how standardized data are processed and shaped by bioinformatics tools, how scientists using such data in research perceive such standard operating procedures and which consequences for knowledge production (e.g. modeling) arise from it. Hence, different concepts and meanings of standardization are explored to get a deeper insight into standard operating procedures not only in systems medicine, but also beyond.

Keywords: data, science and technology studies (STS), standardization, systems medicine

Procedia PDF Downloads 324
24421 Integrated On-Board Diagnostic-II and Direct Controller Area Network Access for Vehicle Monitoring System

Authors: Kavian Khosravinia, Mohd Khair Hassan, Ribhan Zafira Abdul Rahman, Syed Abdul Rahman Al-Haddad

Abstract:

The CAN (controller area network) bus is introduced as a multi-master, message broadcast system. The messages sent on the CAN are used to communicate state information, referred as a signal between different ECUs, which provides data consistency in every node of the system. OBD-II Dongles that are based on request and response method is the wide-spread solution for extracting sensor data from cars among researchers. Unfortunately, most of the past researches do not consider resolution and quantity of their input data extracted through OBD-II technology. The maximum feasible scan rate is only 9 queries per second which provide 8 data points per second with using ELM327 as well-known OBD-II dongle. This study aims to develop and design a programmable, and latency-sensitive vehicle data acquisition system that improves the modularity and flexibility to extract exact, trustworthy, and fresh car sensor data with higher frequency rates. Furthermore, the researcher must break apart, thoroughly inspect, and observe the internal network of the vehicle, which may cause severe damages to the expensive ECUs of the vehicle due to intrinsic vulnerabilities of the CAN bus during initial research. Desired sensors data were collected from various vehicles utilizing Raspberry Pi3 as computing and processing unit with using OBD (request-response) and direct CAN method at the same time. Two types of data were collected for this study. The first, CAN bus frame data that illustrates data collected for each line of hex data sent from an ECU and the second type is the OBD data that represents some limited data that is requested from ECU under standard condition. The proposed system is reconfigurable, human-readable and multi-task telematics device that can be fitted into any vehicle with minimum effort and minimum time lag in the data extraction process. The standard operational procedure experimental vehicle network test bench is developed and can be used for future vehicle network testing experiment.

Keywords: CAN bus, OBD-II, vehicle data acquisition, connected cars, telemetry, Raspberry Pi3

Procedia PDF Downloads 184
24420 Big Data in Construction Project Management: The Colombian Northeast Case

Authors: Sergio Zabala-Vargas, Miguel Jiménez-Barrera, Luz VArgas-Sánchez

Abstract:

In recent years, information related to project management in organizations has been increasing exponentially. Performance data, management statistics, indicator results have forced the collection, analysis, traceability, and dissemination of project managers to be essential. In this sense, there are current trends to facilitate efficient decision-making in emerging technology projects, such as: Machine Learning, Data Analytics, Data Mining, and Big Data. The latter is the most interesting in this project. This research is part of the thematic line Construction methods and project management. Many authors present the relevance that the use of emerging technologies, such as Big Data, has taken in recent years in project management in the construction sector. The main focus is the optimization of time, scope, budget, and in general mitigating risks. This research was developed in the northeastern region of Colombia-South America. The first phase was aimed at diagnosing the use of emerging technologies (Big-Data) in the construction sector. In Colombia, the construction sector represents more than 50% of the productive system, and more than 2 million people participate in this economic segment. The quantitative approach was used. A survey was applied to a sample of 91 companies in the construction sector. Preliminary results indicate that the use of Big Data and other emerging technologies is very low and also that there is interest in modernizing project management. There is evidence of a correlation between the interest in using new data management technologies and the incorporation of Building Information Modeling BIM. The next phase of the research will allow the generation of guidelines and strategies for the incorporation of technological tools in the construction sector in Colombia.

Keywords: big data, building information modeling, tecnology, project manamegent

Procedia PDF Downloads 116
24419 An Empirical Study on Employees’ Theft Behavior in Insurance Industry

Authors: B. Khorsandi Talab, M. Kordi

Abstract:

It is highly deplorable that every year, theft behavior among employees of the insurance industry is growing throughout the world. A very significant source of contraction (despite many costly technological and widespread security measures) needs to be addressed and prevented. Employee and agent theft cannot be ignored as it causes significant losses to employers. This study investigates the workplace factors that affect the insurance employee and agent theft behavior. Although identifying theft is difficult, this study will help employers to further understand employees’ theft behavior. This study was conducted in two service small and medium organizations (two branches of insurance companies) in ALBORZ’s capital city, KARAJ. Data has been collected via questionnaire from 30 employees and agents consisting employees and supervisors of branches and agencies. According to the results, it must be acknowledged that compensation, organizational justice, internal control systems, penalties and personal characteristics were associated with employees' theft behavior, it is despite the fact that, no effect could be assumed for organizational ethics and requirement in this case. Nevertheless, poor financial status cannot be considered as the driving factor in pushing employees to steal property as well as increasing their theft behavior. As mentioned earlier, the purpose of this study was to determine the factors contributing to employees’ theft (insurance employees and agencies) behavior in insurance organizations in Karaj.

Keywords: service theft, employee theft behavior, work theft, insurance agency, SMEs

Procedia PDF Downloads 423
24418 Integrating Road Safety into Mainstreaming Education and Other Initiatives with Holistic Approach in the State: A Case Study of Madhya Pradesh, India

Authors: Yogesh Mahor, Subhash Nigam, Abhai Khare

Abstract:

Road safety education is a composite subject which should be viewed holistically if taken into accoubehavior change communication, safe road infrastructure and low enforcement. Specific and customized road safety education is crucial for each type of road user and learners in the formal and informal teaching and various kind of training programs directly sponsored by state and center government, as they are active contributors to shaping a community and responsible citizens. The aim of this discussion article is to explore a strategy to integrate road safety education into the formal curriculum of schools, higher education institutions, driving schools, skill development centers, various government funded urban and rural development training institutions and their work plans as standing agenda. By applying the desktop research method, the article conceptualizes what the possible focus of road safety education and training should be. The article then explores international common practices in road safety education and training, and considers the necessary synergy between education, road engineering and low enforcement. The article uses secondary data collected from documents which are then analysed in a sectoral way. A well-designed road safety strategy for mainstreaming education and government-sponsored training is urgently needed, facilitating partnerships in various sectors to implement such education in the students and learners in multidisciplinary ways.

Keywords: road safety education, curriculum-based road safety education, behavior change communication, low enforcement, road engineering, safe system approach, infrastructure development consultants

Procedia PDF Downloads 109
24417 Analysis of Policy Issues on Computer-Based Testing in Nigeria

Authors: Samuel Oye Bandele

Abstract:

A policy is a system of principles to guide activities and strategic decisions of an organisation in order to achieve stated objectives and meeting expected outcomes. A Computer Based Test (CBT) policy is therefore a statement of intent to drive the CBT programmes, and should be implemented as a procedure or protocol. Policies are hence generally adopted by an organization or a nation. The concern here, in this paper, is the consideration and analysis of issues that are significant to evolving the acceptable policy that will drive the new CBT innovation in Nigeria. Public examinations and internal examinations in higher educational institutions in Nigeria are gradually making a radical shift from Paper Based or Paper-Pencil to Computer-Based Testing. The need to make an objective and empirical analysis of Policy issues relating to CBT became expedient. The following are some of the issues on CBT evolution in Nigeria that were identified as requiring policy backing. Prominent among them are requirements for establishing CBT centres, purpose of CBT, types and acquisition of CBT equipment, qualifications of staff: professional, technical and regular, security plans and curbing of cheating during examinations, among others. The descriptive research design was employed based on a population consisting of Principal Officers (Policymakers), Staff (Teaching and non-Teaching-Policy implementors), and CBT staff ( Technical and Professional- Policy supports) and candidates (internal and external). A fifty-item researcher-constructed questionnaire on policy issues was employed to collect data from 600 subjects drawn from higher institutions in South West Nigeria, using the purposive and stratified random sampling techniques. Data collected were analysed using descriptive (frequency counts, means and standard deviation) and inferential (t-test, ANOVA, regression and Factor analysis) techniques. Findings from this study showed, among others, that the factor loadings had significantly weights on the organizational and National policy issues on CBT innovation in Nigeria.

Keywords: computer-based testing, examination, innovation, paper-based testing, paper pencil based testing, policy issues

Procedia PDF Downloads 231
24416 Minimum Data of a Speech Signal as Special Indicators of Identification in Phonoscopy

Authors: Nazaket Gazieva

Abstract:

Voice biometric data associated with physiological, psychological and other factors are widely used in forensic phonoscopy. There are various methods for identifying and verifying a person by voice. This article explores the minimum speech signal data as individual parameters of a speech signal. Monozygotic twins are believed to be genetically identical. Using the minimum data of the speech signal, we came to the conclusion that the voice imprint of monozygotic twins is individual. According to the conclusion of the experiment, we can conclude that the minimum indicators of the speech signal are more stable and reliable for phonoscopic examinations.

Keywords: phonogram, speech signal, temporal characteristics, fundamental frequency, biometric fingerprints

Procedia PDF Downloads 123
24415 Hidro-IA: An Artificial Intelligent Tool Applied to Optimize the Operation Planning of Hydrothermal Systems with Historical Streamflow

Authors: Thiago Ribeiro de Alencar, Jacyro Gramulia Junior, Patricia Teixeira Leite

Abstract:

The area of the electricity sector that deals with energy needs by the hydroelectric in a coordinated manner is called Operation Planning of Hydrothermal Power Systems (OPHPS). The purpose of this is to find a political operative to provide electrical power to the system in a given period, with reliability and minimal cost. Therefore, it is necessary to determine an optimal schedule of generation for each hydroelectric, each range, so that the system meets the demand reliably, avoiding rationing in years of severe drought, and that minimizes the expected cost of operation during the planning, defining an appropriate strategy for thermal complementation. Several optimization algorithms specifically applied to this problem have been developed and are used. Although providing solutions to various problems encountered, these algorithms have some weaknesses, difficulties in convergence, simplification of the original formulation of the problem, or owing to the complexity of the objective function. An alternative to these challenges is the development of techniques for simulation optimization and more sophisticated and reliable, it can assist the planning of the operation. Thus, this paper presents the development of a computational tool, namely Hydro-IA for solving optimization problem identified and to provide the User an easy handling. Adopted as intelligent optimization technique is Genetic Algorithm (GA) and programming language is Java. First made the modeling of the chromosomes, then implemented the function assessment of the problem and the operators involved, and finally the drafting of the graphical interfaces for access to the User. The results with the Genetic Algorithms were compared with the optimization technique nonlinear programming (NLP). Tests were conducted with seven hydroelectric plants interconnected hydraulically with historical stream flow from 1953 to 1955. The results of comparison between the GA and NLP techniques shows that the cost of operating the GA becomes increasingly smaller than the NLP when the number of hydroelectric plants interconnected increases. The program has managed to relate a coherent performance in problem resolution without the need for simplification of the calculations together with the ease of manipulating the parameters of simulation and visualization of output results.

Keywords: energy, optimization, hydrothermal power systems, artificial intelligence and genetic algorithms

Procedia PDF Downloads 407
24414 A Non-parametric Clustering Approach for Multivariate Geostatistical Data

Authors: Francky Fouedjio

Abstract:

Multivariate geostatistical data have become omnipresent in the geosciences and pose substantial analysis challenges. One of them is the grouping of data locations into spatially contiguous clusters so that data locations within the same cluster are more similar while clusters are different from each other, in some sense. Spatially contiguous clusters can significantly improve the interpretation that turns the resulting clusters into meaningful geographical subregions. In this paper, we develop an agglomerative hierarchical clustering approach that takes into account the spatial dependency between observations. It relies on a dissimilarity matrix built from a non-parametric kernel estimator of the spatial dependence structure of data. It integrates existing methods to find the optimal cluster number and to evaluate the contribution of variables to the clustering. The capability of the proposed approach to provide spatially compact, connected and meaningful clusters is assessed using bivariate synthetic dataset and multivariate geochemical dataset. The proposed clustering method gives satisfactory results compared to other similar geostatistical clustering methods.

Keywords: clustering, geostatistics, multivariate data, non-parametric

Procedia PDF Downloads 467
24413 Big Data in Telecom Industry: Effective Predictive Techniques on Call Detail Records

Authors: Sara ElElimy, Samir Moustafa

Abstract:

Mobile network operators start to face many challenges in the digital era, especially with high demands from customers. Since mobile network operators are considered a source of big data, traditional techniques are not effective with new era of big data, Internet of things (IoT) and 5G; as a result, handling effectively different big datasets becomes a vital task for operators with the continuous growth of data and moving from long term evolution (LTE) to 5G. So, there is an urgent need for effective Big data analytics to predict future demands, traffic, and network performance to full fill the requirements of the fifth generation of mobile network technology. In this paper, we introduce data science techniques using machine learning and deep learning algorithms: the autoregressive integrated moving average (ARIMA), Bayesian-based curve fitting, and recurrent neural network (RNN) are employed for a data-driven application to mobile network operators. The main framework included in models are identification parameters of each model, estimation, prediction, and final data-driven application of this prediction from business and network performance applications. These models are applied to Telecom Italia Big Data challenge call detail records (CDRs) datasets. The performance of these models is found out using a specific well-known evaluation criteria shows that ARIMA (machine learning-based model) is more accurate as a predictive model in such a dataset than the RNN (deep learning model).

Keywords: big data analytics, machine learning, CDRs, 5G

Procedia PDF Downloads 124
24412 A Data Mining Approach for Analysing and Predicting the Bank's Asset Liability Management Based on Basel III Norms

Authors: Nidhin Dani Abraham, T. K. Sri Shilpa

Abstract:

Asset liability management is an important aspect in banking business. Moreover, the today’s banking is based on BASEL III which strictly regulates on the counterparty default. This paper focuses on prediction and analysis of counter party default risk, which is a type of risk occurs when the customers fail to repay the amount back to the lender (bank or any financial institutions). This paper proposes an approach to reduce the counterparty risk occurring in the financial institutions using an appropriate data mining technique and thus predicts the occurrence of NPA. It also helps in asset building and restructuring quality. Liability management is very important to carry out banking business. To know and analyze the depth of liability of bank, a suitable technique is required. For that a data mining technique is being used to predict the dormant behaviour of various deposit bank customers. Various models are implemented and the results are analyzed of saving bank deposit customers. All these data are cleaned using data cleansing approach from the bank data warehouse.

Keywords: data mining, asset liability management, BASEL III, banking

Procedia PDF Downloads 536
24411 Dynamic Process Model for Designing Smart Spaces Based on Context-Awareness and Computational Methods Principles

Authors: Heba M. Jahin, Ali F. Bakr, Zeyad T. Elsayad

Abstract:

As smart spaces can be defined as any working environment which integrates embedded computers, information appliances and multi-modal sensors to remain focused on the interaction between the users, their activity, and their behavior in the space; hence, smart space must be aware of their contexts and automatically adapt to their changing context-awareness, by interacting with their physical environment through natural and multimodal interfaces. Also, by serving the information used proactively. This paper suggests a dynamic framework through the architectural design process of the space based on the principles of computational methods and context-awareness principles to help in creating a field of changes and modifications. It generates possibilities, concerns about the physical, structural and user contexts. This framework is concerned with five main processes: gathering and analyzing data to generate smart design scenarios, parameters, and attributes; which will be transformed by coding into four types of models. Furthmore, connecting those models together in the interaction model which will represent the context-awareness system. Then, transforming that model into a virtual and ambient environment which represents the physical and real environments, to act as a linkage phase between the users and their activities taking place in that smart space . Finally, the feedback phase from users of that environment to be sure that the design of that smart space fulfill their needs. Therefore, the generated design process will help in designing smarts spaces that can be adapted and controlled to answer the users’ defined goals, needs, and activity.

Keywords: computational methods, context-awareness, design process, smart spaces

Procedia PDF Downloads 307
24410 Development of the Drug Abuse Health Information System in Thai Community

Authors: Waraporn Boonchieng, Ekkarat Boonchieng, Sivaporn Aungwattana, Decha Tamdee, Wongamporn Pinyavong

Abstract:

Drug addiction represents one of the most important public health issues in both developed and developing countries. The purpose of this study was to develop a drug abuse health information in a community in Northern Thailand using developmental research design. The developmental researchers performed four phases to develop drug abuse health information, including 1) synthesizing knowledge related to drug abuse prevention and identifying the components of drug abuse health information; 2) developing the system in mobile application and website; 3) implementing drug abuse health information in the rural community; and 4) evaluating the feasibility of drug abuse health information. Data collection involved both qualitative and quantitative procedures. The qualitative data and quantitative data were analyzed using content analysis and descriptive statistics, respectively. The findings of this study showed that drug abuse health information consisted of five sections, including drug-related prevention knowledge for teens, drug-related knowledge for adults and professionals, the database for drug dependence treatment centers, self-administered questionnaires, and supportive counseling sections. First, in drug-related prevention knowledge for teens, the developmental researchers designed four infographics and animation to provide drug-related prevention knowledge, including types of illegal drugs, causes of drug abuse, consequences of drug abuse, drug abuse diagnosis and treatment, and drug abuse prevention. Second, in drug-related knowledge for adults and professionals, the developmental researchers developed many documents in a form of PDF file to provide drug-related knowledge, including types of illegal drugs, causes of drug abuse, drug abuse prevention, and relapse prevention guideline. Third, database for drug dependence treatment centers included the place, direction map, operation time, and the way for contacting all drug dependence treatment centers in Thailand. Fourth, self-administered questionnaires comprised preventive drugs behavior questionnaire, drug abuse knowledge questionnaire, the stages of change readiness and treatment eagerness to drug use scale, substance use behaviors questionnaire, tobacco use behaviors questionnaire, stress screening, and depression screening. Finally, for supportive counseling, the developmental researchers designed chatting box through which each user could write and send their concerns to counselors individually. Results from evaluation process showed that 651 participants used drug abuse health information via mobile application and website. Among all users, 48.8% were males and 51.2% were females. More than half (55.3%) were 15-20 years old and most of them (88.0%) were Buddhists. Most users reported ever getting knowledge related to drugs (86.1%), and drinking alcohol (94.2%) while some of them (6.9%) reported ever using tobacco. For satisfaction with using the drug abuse health information, more than half of users reflected that the contents of drug abuse health information were interesting (59%), up-to date (61%), and highly useful to their self-study (59%) at high level. In addition, half of them were satisfied with the design in terms of infographics (54%) and animation (51%). Thus, this drug abuse health information can be adopted to explore drug abuse situation and serves as a tool to prevent drug abuse and addiction among Thai community people.

Keywords: drug addiction, health informatics, big data, development research

Procedia PDF Downloads 96
24409 The Two Layers of Food Safety and GMOs in the Hungarian Agricultural Law

Authors: Gergely Horváth

Abstract:

The study presents the complexity of food safety dividing it into two layers. Beyond the basic layer of requirements, there is a more demanding higher level linked with quality and purity aspects. It would be important to give special prominence to both layers, given that massive illnesses are caused by foods even though officially licensed. Then the study discusses an exciting safety challenge stemming from the risks of genetically modified organisms (GMOs). Furthermore, it features legal case examples that illustrate how certain liability questions are solved or not yet decided in connection with the production of genetically modified crops. In addition, a special kind of land grabbing, more precisely land grabbing from non-GMO farming systems can also be noticed as well as a new phenomenon eroding food sovereignty. Coexistence, the state where organic, conventional, and GM farming systems are standing alongside each other is an unsuitable experiment that cannot be successful, because of biophysical reasons (such as cross-pollination). Agricultural and environmental lawyers both try to find the optimal solution. Agri-environmental measures are introduced as a special subfield of law maintaining also food safety. The important steps of agri-environmental legislation are aiming at the protection of natural values, the environmental media and strengthening food safety as well, practically the quality of agricultural products intended for human consumption. The major findings of the study focus on searching for the appropriate approach capable of solving the security and safety problems of food production. The most interesting concepts of the Hungarian national and EU food law legislation are analyzed in more detail with descriptive, analytic and comparative methods.

Keywords: food law, food safety, food security, GMO, Genetically Modified Organisms, agri-environmental measures

Procedia PDF Downloads 423
24408 Vulnerability and Risk Assessment, and Preparedness to Natural Disasters of Schools in Southern Leyte, Philippines

Authors: Lorifel Hinay

Abstract:

Natural disasters have increased in frequency and severity in the Philippines over the years resulting to detrimental impacts in school properties and lives of learners. The topography of the Province of Southern Leyte is a hotspot for inevitable natural disaster-causing hazards that could affect schools, cripple the educational system and cause environmental, cultural and social detrimental impacts making Disaster Risk Reduction and Management (DRRM) an indispensable platform to keep learners safe, secure and resilient. This study determined the schools’ vulnerability and risk assessment to earthquake, landslide, flood, storm surge and tsunami hazards, and its relationship to status in disaster preparedness. Descriptive-correlational research design was used where the respondents were School DRRM Coordinators/School Administrators and Municipal DRRM Officers. It was found that schools’ vulnerability and risk were high in landslide, medium in earthquake, and low in flood, storm surge and tsunami. Though schools were moderately prepared in disasters across all hazards, they were less accomplished in group organization and property security. Less planning preparation and less implementation of DRRM measures were observed in schools highly at risk of earthquake and landslide. Also, schools vulnerable to landslide and flood have very high property security. Topography and location greatly contributed to schools’ vulnerability to hazards, thus, a school-based disaster preparedness plan is hoped to help ensure that hazard-exposed schools can build a culture of safety, disaster resiliency and education continuity.

Keywords: disaster risk reduction and management, earthquake, flood, landslide, storm surge, tsunami

Procedia PDF Downloads 105
24407 A Study to Examine the Use of Traditional Agricultural Practices to Fight the Effects of Climate Change

Authors: Rushva Parihar, Anushka Barua

Abstract:

The negative repercussions of a warming planet are already visible, with biodiversity loss, water scarcity, and extreme weather events becoming ever so frequent. The agriculture sector is perhaps the most impacted, and modern agriculture has failed to defend farmers from the effects of climate change. This, coupled with the added pressure of higher demands for food production caused due to population growth, has only compounded the impact. Traditional agricultural practices that are routed in indigenous knowledge have long safeguarded the delicate balance of the ecosystem through sustainable production techniques. This paper uses secondary data to explore these traditional processes (like Beejamrita, Jeevamrita, sheep penning, earthen bunding, and others) from around the world that have been developed over centuries and focuses on how they can be used to tackle contemporary issues arising from climate change (such as nutrient and water loss, soil degradation, increased incidences of pests). Finally, the resulting framework has been applied to the context of Indian agriculture as a means to combat climate change and improve food security, all while encouraging documentation and transfer of local knowledge as a shared resource among farmers.

Keywords: sustainable food systems, traditional agricultural practices, climate smart agriculture, climate change, indigenous knowledge

Procedia PDF Downloads 112
24406 Thresholding Approach for Automatic Detection of Pseudomonas aeruginosa Biofilms from Fluorescence in situ Hybridization Images

Authors: Zonglin Yang, Tatsuya Akiyama, Kerry S. Williamson, Michael J. Franklin, Thiruvarangan Ramaraj

Abstract:

Pseudomonas aeruginosa is an opportunistic pathogen that forms surface-associated microbial communities (biofilms) on artificial implant devices and on human tissue. Biofilm infections are difficult to treat with antibiotics, in part, because the bacteria in biofilms are physiologically heterogeneous. One measure of biological heterogeneity in a population of cells is to quantify the cellular concentrations of ribosomes, which can be probed with fluorescently labeled nucleic acids. The fluorescent signal intensity following fluorescence in situ hybridization (FISH) analysis correlates to the cellular level of ribosomes. The goals here are to provide computationally and statistically robust approaches to automatically quantify cellular heterogeneity in biofilms from a large library of epifluorescent microscopy FISH images. In this work, the initial steps were developed toward these goals by developing an automated biofilm detection approach for use with FISH images. The approach allows rapid identification of biofilm regions from FISH images that are counterstained with fluorescent dyes. This methodology provides advances over other computational methods, allowing subtraction of spurious signals and non-biological fluorescent substrata. This method will be a robust and user-friendly approach which will enable users to semi-automatically detect biofilm boundaries and extract intensity values from fluorescent images for quantitative analysis of biofilm heterogeneity.

Keywords: image informatics, Pseudomonas aeruginosa, biofilm, FISH, computer vision, data visualization

Procedia PDF Downloads 119
24405 Unsupervised Text Mining Approach to Early Warning System

Authors: Ichihan Tai, Bill Olson, Paul Blessner

Abstract:

Traditional early warning systems that alarm against crisis are generally based on structured or numerical data; therefore, a system that can make predictions based on unstructured textual data, an uncorrelated data source, is a great complement to the traditional early warning systems. The Chicago Board Options Exchange (CBOE) Volatility Index (VIX), commonly referred to as the fear index, measures the cost of insurance against market crash, and spikes in the event of crisis. In this study, news data is consumed for prediction of whether there will be a market-wide crisis by predicting the movement of the fear index, and the historical references to similar events are presented in an unsupervised manner. Topic modeling-based prediction and representation are made based on daily news data between 1990 and 2015 from The Wall Street Journal against VIX index data from CBOE.

Keywords: early warning system, knowledge management, market prediction, topic modeling.

Procedia PDF Downloads 322
24404 The Role of Synthetic Data in Aerial Object Detection

Authors: Ava Dodd, Jonathan Adams

Abstract:

The purpose of this study is to explore the characteristics of developing a machine learning application using synthetic data. The study is structured to develop the application for the purpose of deploying the computer vision model. The findings discuss the realities of attempting to develop a computer vision model for practical purpose, and detail the processes, tools, and techniques that were used to meet accuracy requirements. The research reveals that synthetic data represents another variable that can be adjusted to improve the performance of a computer vision model. Further, a suite of tools and tuning recommendations are provided.

Keywords: computer vision, machine learning, synthetic data, YOLOv4

Procedia PDF Downloads 208
24403 A Miniaturized Circular Patch Antenna Based on Metamaterial for WI-FI Applications

Authors: Fatima Zahra Moussa, Yamina Belhadef, Souheyla Ferouani

Abstract:

In this work, we present a new form of miniature circular patch antenna based on CSRR metamaterials with an extended bandwidth proposed for 5 GHz Wi-Fiapplications. A reflection coefficient of -35 dB and a radiation pattern of 7.47 dB are obtained when simulating the initial proposed antenna with the CST microwave studio simulation software. The notch insertion technique in the radiating element was used for matching the antenna to the desired frequency in the frequency band [5150-5875] MHz.An extension of the bandwidth from 332 MHz to 1423 MHz was done by the DGS (defected ground structure) technique to meet the user's requirement in the 5 GHz Wi-Fi frequency band.

Keywords: patch antenna, miniaturisation, CSRR, notches, wifi, DGS

Procedia PDF Downloads 100
24402 Juridically Secure Trade Mechanisms for Alternative Dispute Resolution in Transnational Business Negotiations

Authors: Linda Frazer

Abstract:

A pluralistic methodology focuses on promoting an understanding that an alternative juridical framework for the regulation of transnational business negotiations (TBN) between private business parties is fundamentally required. This paper deals with the evolving assessment of the doctoral research of the author which demonstrated that due to insufficient juridical tools, negotiations are commonly misunderstood within the complexity of pluralistic and conflicting legal regimes. This inadequacy causes uncertainty in the enforcement of legal remedies, leaving business parties surprised. Consequently, parties cannot sufficiently anticipate when and how legal rights and obligations are created, often counting on oral or incomplete agreements which may lead to the misinterpretation of the extent of their legal rights and obligations. This uncertainty causes threats to business parties for fear of creating unintended legal obligations or, conversely, that law will not enforce intended agreements for failure to pass the tests of contractual validity. A need to find a manner to set default standards of communications and standards of conduct to monitor our evolving global trade would aid law to provide the security, predictability and foreseeability during alternative dispute resolution required by TBN parties. The conclusion of this study includes a proposal of new trade mechanisms, termed 'Bills of Negotiations' (BON) to enhance party autonomy and promote the ability for TBN parties to self-regulate within the boundaries of law. BON will be guided by a secure juridical institutionalized setting that caters to guiding communications during TBN and resolving disputes that arise along the negotiation processes on a fast track basis.

Keywords: alternative resolution disputes, ADR, good faith, good faith, juridical security, legal regulation, trade mechanisms, transnational business negotiations

Procedia PDF Downloads 132
24401 Perception-Oriented Model Driven Development for Designing Data Acquisition Process in Wireless Sensor Networks

Authors: K. Indra Gandhi

Abstract:

Wireless Sensor Networks (WSNs) have always been characterized for application-specific sensing, relaying and collection of information for further analysis. However, software development was not considered as a separate entity in this process of data collection which has posed severe limitations on the software development for WSN. Software development for WSN is a complex process since the components involved are data-driven, network-driven and application-driven in nature. This implies that there is a tremendous need for the separation of concern from the software development perspective. A layered approach for developing data acquisition design based on Model Driven Development (MDD) has been proposed as the sensed data collection process itself varies depending upon the application taken into consideration. This work focuses on the layered view of the data acquisition process so as to ease the software point of development. A metamodel has been proposed that enables reusability and realization of the software development as an adaptable component for WSN systems. Further, observing users perception indicates that proposed model helps in improving the programmer's productivity by realizing the collaborative system involved.

Keywords: data acquisition, model-driven development, separation of concern, wireless sensor networks

Procedia PDF Downloads 419
24400 Comparative Analysis of Data Gathering Protocols with Multiple Mobile Elements for Wireless Sensor Network

Authors: Bhat Geetalaxmi Jairam, D. V. Ashoka

Abstract:

Wireless Sensor Networks are used in many applications to collect sensed data from different sources. Sensed data has to be delivered through sensors wireless interface using multi-hop communication towards the sink. The data collection in wireless sensor networks consumes energy. Energy consumption is the major constraints in WSN .Reducing the energy consumption while increasing the amount of generated data is a great challenge. In this paper, we have implemented two data gathering protocols with multiple mobile sinks/elements to collect data from sensor nodes. First, is Energy-Efficient Data Gathering with Tour Length-Constrained Mobile Elements in Wireless Sensor Networks (EEDG), in which mobile sinks uses vehicle routing protocol to collect data. Second is An Intelligent Agent-based Routing Structure for Mobile Sinks in WSNs (IAR), in which mobile sinks uses prim’s algorithm to collect data. Authors have implemented concepts which are common to both protocols like deployment of mobile sinks, generating visiting schedule, collecting data from the cluster member. Authors have compared the performance of both protocols by taking statistics based on performance parameters like Delay, Packet Drop, Packet Delivery Ratio, Energy Available, Control Overhead. Authors have concluded this paper by proving EEDG is more efficient than IAR protocol but with few limitations which include unaddressed issues likes Redundancy removal, Idle listening, Mobile Sink’s pause/wait state at the node. In future work, we plan to concentrate more on these limitations to avail a new energy efficient protocol which will help in improving the life time of the WSN.

Keywords: aggregation, consumption, data gathering, efficiency

Procedia PDF Downloads 476
24399 Status and Results from EXO-200

Authors: Ryan Maclellan

Abstract:

EXO-200 has provided one of the most sensitive searches for neutrinoless double-beta decay utilizing 175 kg of enriched liquid xenon in an ultra-low background time projection chamber. This detector has demonstrated excellent energy resolution and background rejection capabilities. Using the first two years of data, EXO-200 has set a limit of 1.1x10^25 years at 90% C.L. on the neutrinoless double-beta decay half-life of Xe-136. The experiment has experienced a brief hiatus in data taking during a temporary shutdown of its host facility: the Waste Isolation Pilot Plant. EXO-200 expects to resume data taking in earnest this fall with upgraded detector electronics. Results from the analysis of EXO-200 data and an update on the current status of EXO-200 will be presented.

Keywords: double-beta, Majorana, neutrino, neutrinoless

Procedia PDF Downloads 398
24398 Condensation of Moist Air in Heat Exchanger Using CFD

Authors: Jan Barak, Karel Frana, Joerg Stiller

Abstract:

This work presents results of moist air condensation in heat exchanger. It describes theoretical knowledge and definition of moist air. Model with geometry of square canal was created for better understanding and post processing of condensation phenomena. Different approaches were examined on this model to find suitable software and model. Obtained knowledge was applied to geometry of real heat exchanger and results from experiment were compared with numerical results. One of the goals is to solve this issue without creating any user defined function in the applied code. It also contains summary of knowledge and outlook for future work.

Keywords: condensation, exchanger, experiment, validation

Procedia PDF Downloads 386
24397 Remaining Useful Life (RUL) Assessment Using Progressive Bearing Degradation Data and ANN Model

Authors: Amit R. Bhende, G. K. Awari

Abstract:

Remaining useful life (RUL) prediction is one of key technologies to realize prognostics and health management that is being widely applied in many industrial systems to ensure high system availability over their life cycles. The present work proposes a data-driven method of RUL prediction based on multiple health state assessment for rolling element bearings. Bearing degradation data at three different conditions from run to failure is used. A RUL prediction model is separately built in each condition. Feed forward back propagation neural network models are developed for prediction modeling.

Keywords: bearing degradation data, remaining useful life (RUL), back propagation, prognosis

Procedia PDF Downloads 417