Search results for: internet data centre
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 25916

Search results for: internet data centre

24656 Identity Verification Using k-NN Classifiers and Autistic Genetic Data

Authors: Fuad M. Alkoot

Abstract:

DNA data have been used in forensics for decades. However, current research looks at using the DNA as a biometric identity verification modality. The goal is to improve the speed of identification. We aim at using gene data that was initially used for autism detection to find if and how accurate is this data for identification applications. Mainly our goal is to find if our data preprocessing technique yields data useful as a biometric identification tool. We experiment with using the nearest neighbor classifier to identify subjects. Results show that optimal classification rate is achieved when the test set is corrupted by normally distributed noise with zero mean and standard deviation of 1. The classification rate is close to optimal at higher noise standard deviation reaching 3. This shows that the data can be used for identity verification with high accuracy using a simple classifier such as the k-nearest neighbor (k-NN). 

Keywords: biometrics, genetic data, identity verification, k nearest neighbor

Procedia PDF Downloads 243
24655 A Review on Intelligent Systems for Geoscience

Authors: R Palson Kennedy, P.Kiran Sai

Abstract:

This article introduces machine learning (ML) researchers to the hurdles that geoscience problems present, as well as the opportunities for improvement in both ML and geosciences. This article presents a review from the data life cycle perspective to meet that need. Numerous facets of geosciences present unique difficulties for the study of intelligent systems. Geosciences data is notoriously difficult to analyze since it is frequently unpredictable, intermittent, sparse, multi-resolution, and multi-scale. The first half addresses data science’s essential concepts and theoretical underpinnings, while the second section contains key themes and sharing experiences from current publications focused on each stage of the data life cycle. Finally, themes such as open science, smart data, and team science are considered.

Keywords: Data science, intelligent system, machine learning, big data, data life cycle, recent development, geo science

Procedia PDF Downloads 125
24654 ASEAN Citizenship in the Internationalization of Thai Higher Education

Authors: Bella Llego

Abstract:

This research aims to study on “ASEAN Citizenship in the Internationalization of Thai Higher Education.” The purposes of this research are (1) to examine the Thai academics and scholars defined in the concept of internationalization of higher education, (2) to know how Thailand tries to fulfill its internationalization on education goal, (3) to find out the advantages and disadvantages of Thailand hub for higher education in Asia. Sequential mixed methods, qualitative and quantitative research methods were utilized to gather the data collected. By using a qualitative method (individual interviews from key Thai administrators and educators in the international higher education sector), a quantitative method (survey) was utilized to draw upon and to elaborate the recurring themes present during the interviews. The study found that many aspects of Thai international higher education programs received heavy influence from both the American and European higher education systems. Thailand’s role and leadership in the creation and launch of the ASEAN Economic Community (AEC) by 2015 gives its unique context for its internationalization efforts. English is being designated as the language of all Thai international programs; its influence further strengthened being the current language of academia, international business, and the internet, having global influence.

Keywords: ASEAN citizenship, internationalization, Thai higher education, academics

Procedia PDF Downloads 254
24653 Improving Search Engine Performance by Removing Indexes to Malicious URLs

Authors: Durga Toshniwal, Lokesh Agrawal

Abstract:

As the web continues to play an increasing role in information exchange, and conducting daily activities, computer users have become the target of miscreants which infects hosts with malware or adware for financial gains. Unfortunately, even a single visit to compromised web site enables the attacker to detect vulnerabilities in the user’s applications and force the downloading of multitude of malware binaries. We provide an approach to effectively scan the so-called drive-by downloads on the Internet. Drive-by downloads are result of URLs that attempt to exploit their visitors and cause malware to be installed and run automatically. To scan the web for malicious pages, the first step is to use a crawler to collect URLs that live on the Internet, and then to apply fast prefiltering techniques to reduce the amount of pages that are needed to be examined by precise, but slower, analysis tools (such as honey clients or antivirus programs). Although the technique is effective, it requires a substantial amount of resources. A main reason is that the crawler encounters many pages on the web that are legitimate and needs to be filtered. In this paper, to characterize the nature of this rising threat, we present implementation of a web crawler on Python, an approach to search the web more efficiently for pages that are likely to be malicious, filtering benign pages and passing remaining pages to antivirus program for detection of malwares. Our approaches starts from an initial seed of known, malicious web pages. Using these seeds, our system generates search engines queries to identify other malicious pages that are similar to the ones in the initial seed. By doing so, it leverages the crawling infrastructure of search engines to retrieve URLs that are much more likely to be malicious than a random page on the web. The results shows that this guided approach is able to identify malicious web pages more efficiently when compared to random crawling-based approaches.

Keywords: web crawler, malwares, seeds, drive-by-downloads, security

Procedia PDF Downloads 219
24652 Packet Fragmentation Caused by Encryption and Using It as a Security Method

Authors: Said Rabah Azzam, Andrew Graham

Abstract:

Fragmentation of packets caused by encryption applied on the network layer of the IOS model in Internet Protocol version 4 (IPv4) networks as well as the possibility of using fragmentation and Access Control Lists (ACLs) as a method of restricting network access to certain hosts or areas of a network.Using default settings, fragmentation is expected to occur and each fragment to be reassembled at the other end. If this does not occur then a high number of ICMP messages should be generated back towards the source host indicating that the packet is too large and that it needs to be made smaller. This result is also expected when the MTU is changed for certain links between devices.When using ACLs and packet fragments to restrict access to hosts or network segments it is possible that ACLs cannot be set up in this way. If ACLs cannot be setup to allow only fragments then it is a limitation of the hardware’s firmware holding back this particular method. If the ACL on the restricted switch can be set up in such a way to allow only fragments then a connection that forces packets to fragment should be allowed to pass through the ACL. This should then make a network connection to the destination machine allowing data to be sent to and from the destination machine. ICMP messages from the restricted access switch and host should also be blocked from being sent back across the link which will be shown in an SSH session into the switch.

Keywords: fragmentation, encryption, security, switch

Procedia PDF Downloads 317
24651 Use of Galileo Advanced Features in Maritime Domain

Authors: Olivier Chaigneau, Damianos Oikonomidis, Marie-Cecile Delmas

Abstract:

GAMBAS (Galileo Advanced features for the Maritime domain: Breakthrough Applications for Safety and security) is a project funded by the European Space Program Agency (EUSPA) aiming at identifying the search-and-rescue and ship security alert system needs for maritime users (including operators and fishing stakeholders) and developing operational concepts to answer these needs. The general objective of the GAMBAS project is to support the deployment of Galileo exclusive features in the maritime domain in order to improve safety and security at sea, detection of illegal activities and associated surveillance means, resilience to natural and human-induced emergency situations, and develop, integrate, demonstrate, standardize and disseminate these new associated capabilities. The project aims to demonstrate: improvement of the SAR (Search And Rescue) and SSAS (Ship Security Alert System) detection and response to maritime distress through the integration of new features into the beacon for SSAS in terms of cost optimization, user-friendly aspects, integration of Galileo and OS NMA (Open Service Navigation Message Authentication) reception for improved authenticated localization performance and reliability, and at sea triggering capabilities, optimization of the responsiveness of RCCs (Rescue Co-ordination Centre) towards the distress situations affecting vessels, the adaptation of the MCCs (Mission Control Center) and MEOLUT (Medium Earth Orbit Local User Terminal) to the data distribution of SSAS alerts.

Keywords: Galileo new advanced features, maritime, safety, security

Procedia PDF Downloads 84
24650 An Exploration of Cyberspace Security, Strategy for a New Era

Authors: Laxmi R. Kasaraneni

Abstract:

The Internet connects all the networks, including the nation’s critical infrastructure that are used extensively by not only a nation’s government and military to protect sensitive information and execute missions, but also the primary infrastructure that provides services that enable modern conveniences such as education, potable water, electricity, natural gas, and financial transactions. It has become the central nervous system for the government, the citizens, and the industries. When it is attacked, the effects can ripple far and wide impacts not only to citizens’ well-being but nation’s economy, civil infrastructure, and national security. As such, these critical services may be targeted by malicious hackers during cyber warfare, it is imperative to not only protect them and mitigate any immediate or potential threats, but to also understand the current or potential impacts beyond the IT networks or the organization. The Nation’s IT infrastructure which is now vital for communication, commerce, and control of our physical infrastructure, is highly vulnerable to attack. While existing technologies can address some vulnerabilities, fundamentally new architectures and technologies are needed to address the larger structural insecurities of an infrastructure developed in a more trusting time when mass cyber attacks were not foreseen. This research is intended to improve the core functions of the Internet and critical-sector information systems by providing a clear path to create a safe, secure, and resilient cyber environment that help stakeholders at all levels of government, and the private sector work together to develop the cybersecurity capabilities that are key to our economy, national security, and public health and safety. This research paper also emphasizes the present and future cyber security threats, the capabilities and goals of cyber attackers, a strategic concept and steps to implement cybersecurity for maximum effectiveness, enabling technologies, some strategic assumptions and critical challenges, and the future of cyberspace.

Keywords: critical challenges, critical infrastructure, cyber security, enabling technologies, national security

Procedia PDF Downloads 286
24649 Implementation of Smart Card Automatic Fare Collection Technology in Small Transit Agencies for Standards Development

Authors: Walter E. Allen, Robert D. Murray

Abstract:

Many large transit agencies have adopted RFID technology and electronic automatic fare collection (AFC) or smart card systems, but small and rural agencies remain tied to obsolete manual, cash-based fare collection. Small countries or transit agencies can benefit from the implementation of smart card AFC technology with the promise of increased passenger convenience, added passenger satisfaction and improved agency efficiency. For transit agencies, it reduces revenue loss, improves passenger flow and bus stop data. For countries, further implementation into security, distribution of social services or currency transactions can provide greater benefits. However, small countries or transit agencies cannot afford expensive proprietary smart card solutions typically offered by the major system suppliers. Deployment of Contactless Fare Media System (CFMS) Standard eliminates the proprietary solution, ultimately lowering the cost of implementation. Acumen Building Enterprise, Inc. chose the Yuma County Intergovernmental Public Transportation Authority (YCIPTA) existing proprietary YCAT smart card system to implement CFMS. The revised system enables the purchase of fare product online with prepaid debit or credit cards using the Payment Gateway Processor. Open and interoperable smart card standards for transit have been developed. During the 90-day Pilot Operation conducted, the transit agency gathered the data from the bus AcuFare 200 Card Reader, loads (copies) the data to a USB Thumb Drive and uploads the data to the Acumen Host Processing Center for consolidation of the data into the transit agency master data file. The transition from the existing proprietary smart card data format to the new CFMS smart card data format was transparent to the transit agency cardholders. It was proven that open standards and interoperability design can work and reduce both implementation and operational costs for small transit agencies or countries looking to expand smart card technology. Acumen was able to avoid the implementation of the Payment Card Industry (PCI) Data Security Standards (DSS) which is expensive to develop and costly to operate on a continuing basis. Due to the substantial additional complexities of implementation and the variety of options presented to the transit agency cardholder, Acumen chose to implement only the Directed Autoload. To improve the implementation efficiency and the results for a similar undertaking, it should be considered that some passengers lack credit cards and are averse to technology. There are more than 1,300 small and rural agencies in the United States. This grows by 10 fold when considering small countries or rural locations throughout Latin American and the world. Acumen is evaluating additional countries, sites or transit agency that can benefit from the smart card systems. Frequently, payment card systems require extensive security procedures for implementation. The Project demonstrated the ability to purchase fare value, rides and passes with credit cards on the internet at a reasonable cost without highly complex security requirements.

Keywords: automatic fare collection, near field communication, small transit agencies, smart cards

Procedia PDF Downloads 270
24648 Data Quality as a Pillar of Data-Driven Organizations: Exploring the Benefits of Data Mesh

Authors: Marc Bachelet, Abhijit Kumar Chatterjee, José Manuel Avila

Abstract:

Data quality is a key component of any data-driven organization. Without data quality, organizations cannot effectively make data-driven decisions, which often leads to poor business performance. Therefore, it is important for an organization to ensure that the data they use is of high quality. This is where the concept of data mesh comes in. Data mesh is an organizational and architectural decentralized approach to data management that can help organizations improve the quality of data. The concept of data mesh was first introduced in 2020. Its purpose is to decentralize data ownership, making it easier for domain experts to manage the data. This can help organizations improve data quality by reducing the reliance on centralized data teams and allowing domain experts to take charge of their data. This paper intends to discuss how a set of elements, including data mesh, are tools capable of increasing data quality. One of the key benefits of data mesh is improved metadata management. In a traditional data architecture, metadata management is typically centralized, which can lead to data silos and poor data quality. With data mesh, metadata is managed in a decentralized manner, ensuring accurate and up-to-date metadata, thereby improving data quality. Another benefit of data mesh is the clarification of roles and responsibilities. In a traditional data architecture, data teams are responsible for managing all aspects of data, which can lead to confusion and ambiguity in responsibilities. With data mesh, domain experts are responsible for managing their own data, which can help provide clarity in roles and responsibilities and improve data quality. Additionally, data mesh can also contribute to a new form of organization that is more agile and adaptable. By decentralizing data ownership, organizations can respond more quickly to changes in their business environment, which in turn can help improve overall performance by allowing better insights into business as an effect of better reports and visualization tools. Monitoring and analytics are also important aspects of data quality. With data mesh, monitoring, and analytics are decentralized, allowing domain experts to monitor and analyze their own data. This will help in identifying and addressing data quality problems in quick time, leading to improved data quality. Data culture is another major aspect of data quality. With data mesh, domain experts are encouraged to take ownership of their data, which can help create a data-driven culture within the organization. This can lead to improved data quality and better business outcomes. Finally, the paper explores the contribution of AI in the coming years. AI can help enhance data quality by automating many data-related tasks, like data cleaning and data validation. By integrating AI into data mesh, organizations can further enhance the quality of their data. The concepts mentioned above are illustrated by AEKIDEN experience feedback. AEKIDEN is an international data-driven consultancy that has successfully implemented a data mesh approach. By sharing their experience, AEKIDEN can help other organizations understand the benefits and challenges of implementing data mesh and improving data quality.

Keywords: data culture, data-driven organization, data mesh, data quality for business success

Procedia PDF Downloads 122
24647 New Model of Immersive Experiential Branding for International Universities

Authors: Kakhaber Djakeli

Abstract:

For market leadership, iconic brands already start to establish their unique digital avatars into Metaverse and offer Non Fungible Tokens to their fans. Metaverse can be defined as an evolutionary step of Internet development. So if companies and brands use the internet, logically, they can find new solutions for them and their customers in Metaverse. Marketing and Management today must learn how to combine physical world activities with those either entitled as digital, virtual, and immersive. A “Phygital” Solution uniting physical and digital competitive activities of the company covering the questions about how to use virtual worlds for Brand Development and Non Fungible Tokens for more attractiveness soon will be most relevant question for Branding. Thinking comprehensively, we can entitle this type of branding as an Immersive one. As we see, the Immersive Brands give customers more mesmerizing feelings than traditional ones. Accordingly, the Branding can be divided by the company in its own understanding into two models: traditional and immersive. Immersive Branding being more directed to Sensorial challenges of Humans will be big job for International Universities in near future because they target the Generation - Z. To try to help those International Universities opening the door to the mesmerizing, immersive branding, the Marketing Research have been undertaken. The main goal of the study was to establish the model for Immersive Branding at International Universities and answer on many questions what logically arises in university life. The type of Delphi Surveys entitled as an Expert Studies was undertaken for one great mission, to help International Universities to open the opportunities to Phygital activities with reliable knowledge with Model of Immersive Branding. The Questionnaire sent to Experts of Education were covering professional type of questions from education to segmentation of customers, branding, attitude to students, and knowledge to Immersive Marketing. The research results being very interesting and encouraging enough to make author to establish the New Model of Immersive Experiential Branding for International Universities.

Keywords: branding, immersive marketing, students, university

Procedia PDF Downloads 72
24646 Development of a Novel Clinical Screening Tool, Using the BSGE Pain Questionnaire, Clinical Examination and Ultrasound to Predict the Severity of Endometriosis Prior to Laparoscopic Surgery

Authors: Marlin Mubarak

Abstract:

Background: Endometriosis is a complex disabling disease affecting young females in the reproductive period mainly. The aim of this project is to generate a diagnostic model to predict severity and stage of endometriosis prior to Laparoscopic surgery. This will help to improve the pre-operative diagnostic accuracy of stage 3 & 4 endometriosis and as a result, refer relevant women to a specialist centre for complex Laparoscopic surgery. The model is based on the British Society of Gynaecological Endoscopy (BSGE) pain questionnaire, clinical examination and ultrasound scan. Design: This is a prospective, observational, study, in which women completed the BSGE pain questionnaire, a BSGE requirement. Also, as part of the routine preoperative assessment patient had a routine ultrasound scan and when recto-vaginal and deep infiltrating endometriosis was suspected an MRI was performed. Setting: Luton & Dunstable University Hospital. Patients: Symptomatic women (n = 56) scheduled for laparoscopy due to pelvic pain. The age ranged between 17 – 52 years of age (mean 33.8 years, SD 8.7 years). Interventions: None outside the recognised and established endometriosis centre protocol set up by BSGE. Main Outcome Measure(s): Sensitivity and specificity of endometriosis diagnosis predicted by symptoms based on BSGE pain questionnaire, clinical examinations and imaging. Findings: The prevalence of diagnosed endometriosis was calculated to be 76.8% and the prevalence of advanced stage was 55.4%. Deep infiltrating endometriosis in various locations was diagnosed in 32/56 women (57.1%) and some had DIE involving several locations. Logistic regression analysis was performed on 36 clinical variables to create a simple clinical prediction model. After creating the scoring system using variables with P < 0.05, the model was applied to the whole dataset. The sensitivity was 83.87% and specificity 96%. The positive likelihood ratio was 20.97 and the negative likelihood ratio was 0.17, indicating that the model has a good predictive value and could be useful in predicting advanced stage endometriosis. Conclusions: This is a hypothesis-generating project with one operator, but future proposed research would provide validation of the model and establish its usefulness in the general setting. Predictive tools based on such model could help organise the appropriate investigation in clinical practice, reduce risks associated with surgery and improve outcome. It could be of value for future research to standardise the assessment of women presenting with pelvic pain. The model needs further testing in a general setting to assess if the initial results are reproducible.

Keywords: deep endometriosis, endometriosis, minimally invasive, MRI, ultrasound.

Procedia PDF Downloads 342
24645 The Growth of E-Commerce and Online Dispute Resolution in Developing Nations: An Analysis

Authors: Robin V. Cupido

Abstract:

Online dispute resolution has been identified in many countries as a viable alternative for resolving conflicts which have arisen in the so-called digital age. This system of dispute resolution is developing alongside the Internet, and as new types of transactions are made possible by our increased connectivity, new ways of resolving disputes must be explored. Developed nations, such as the United States of America and the European Union, have been involved in creating these online dispute resolution mechanisms from the outset, and currently have sophisticated systems in place to deal with conflicts arising in a number of different fields, such as e-commerce, domain name disputes, labour disputes and conflicts arising from family law. Specifically, in the field of e-commerce, the Internet’s borderless nature has served as a way to promote cross-border trade, and has created a global marketplace. Participation in this marketplace boosts a country’s economy, as new markets are now available, and consumers can transact from anywhere in the world. It would be especially advantageous for developing nations to be a part of this global marketplace, as it could stimulate much-needed investment in these nations, and encourage international co-operation and trade. However, for these types of transactions to proliferate, an effective system for resolving the inevitable disputes arising from such an increase in e-commerce is needed. Online dispute resolution scholarship and practice is flourishing in developed nations, and it is clear that the gap is widening between developed and developing nations in this regard. The potential for implementing online dispute resolution in developing countries has been discussed, but there are a number of obstacles that have thus far prevented its continued development. This paper aims to evaluate the various political, infrastructural and socio-economic challenges faced in developing nations, and to question how these have impacted the acceptance and development of online dispute resolution, scholarship and training of online dispute resolution practitioners and, ultimately, developing nations’ readiness to participate in cross-border e-commerce.

Keywords: developing countries, feasibility, online dispute resolution, progress

Procedia PDF Downloads 294
24644 Optimizing Resource Allocation and Indoor Location Using Bluetooth Low Energy

Authors: Néstor Álvarez-Díaz, Pino Caballero-Gil, Héctor Reboso-Morales, Francisco Martín-Fernández

Abstract:

The recent tendency of "Internet of Things" (IoT) has developed in the last years, causing the emergence of innovative communication methods among multiple devices. The appearance of Bluetooth Low Energy (BLE) has allowed a push to IoT in relation to smartphones. In this moment, a set of new applications related to several topics like entertainment and advertisement has begun to be developed but not much has been done till now to take advantage of the potential that these technologies can offer on many business areas and in everyday tasks. In the present work, the application of BLE technology and smartphones is proposed on some business areas related to the optimization of resource allocation in huge facilities like airports. An indoor location system has been developed through triangulation methods with the use of BLE beacons. The described system can be used to locate all employees inside the building in such a way that any task can be automatically assigned to a group of employees. It should be noted that this system cannot only be used to link needs with employees according to distances, but it also takes into account other factors like occupation level or category. In addition, it has been endowed with a security system to manage business and personnel sensitive data. The efficiency of communications is another essential characteristic that has been taken into account in this work.

Keywords: bluetooth low energy, indoor location, resource assignment, smartphones

Procedia PDF Downloads 380
24643 Developing Fault Tolerance Metrics of Web and Mobile Applications

Authors: Ahmad Mohsin, Irfan Raza Naqvi, Syda Fatima Usamn

Abstract:

Applications with higher fault tolerance index are considered more reliable and trustworthy to drive quality. In recent years application development has been shifted from traditional desktop and web to native and hybrid application(s) for the web and mobile platforms. With the emergence of Internet of things IOTs, cloud and big data trends, the need for measuring Fault Tolerance for these complex nature applications has increased to evaluate their performance. There is a phenomenal gap between fault tolerance metrics development and measurement. Classic quality metric models focused on metrics for traditional systems ignoring the essence of today’s applications software, hardware & deployment characteristics. In this paper, we have proposed simple metrics to measure fault tolerance considering general requirements for Web and Mobile Applications. We have aligned factors – subfactors, using GQM for metrics development considering the nature of mobile we apps. Systematic Mathematical formulation is done to measure metrics quantitatively. Three web mobile applications are selected to measure Fault Tolerance factors using formulated metrics. Applications are then analysed on the basis of results from observations in a controlled environment on different mobile devices. Quantitative results are presented depicting Fault tolerance in respective applications.

Keywords: web and mobile applications, reliability, fault tolerance metric, quality metrics, GQM based metrics

Procedia PDF Downloads 330
24642 A Reading Light That Can Adjust Indoor Light Intensity According to the Activity and Person for Improve Indoor Visual Comfort of Occupants and Tested using Post-occupancy Evaluation Techniques for Sri Lankan Population

Authors: R.T.P. De Silva, T. K. Wijayasiriwardhane, B. Jayawardena

Abstract:

Most people nowadays spend their time indoor environment. Because of that, a quality indoor environment needs for them. This study was conducted to identify how to improve indoor visual comfort using a personalized light system. Light intensity, light color, glare, and contrast are the main facts that affect visual comfort. The light intensity which needs to perform a task is changed according to the task. Using necessary light intensity and we can improve the visual comfort of occupants. The hue can affect the emotions of occupants. The preferred light colors and intensity change according to the occupant's age and gender. The research was conducted to identify is there any relationship between personalization and visual comfort. To validate this designed an Internet of Things-based reading light. This light can work according to the standard light levels and personalized light levels. It also can measure the current light intensity of the environment and maintain continuous light levels according to the task. The test was conducted by using 25 undergraduates, and 5school students, and 5 adults. The feedbacks are gathered using Post-occupancy evaluation (POE) techniques. Feedbacks are gathered in three steps, It was done without any light control, with standard light level, and with personalized light level Users had to spend 10 minutes under each condition. After finishing each step, collected their feedbacks. According to the result gathered, 94% of participants rated a personalized light system as comfort for them. The feedbacks show stay under continuous light level help to keep their concentrate. Future research can be conducted on how the color of indoor light can affect for indoor visual comfort of occupants using a personalized light system. Further proposed IoT based can improve to change the light colors according to the user's preference.

Keywords: indoor environment quality, internet of things based light system, post occupancy evaluation, visual comfort

Procedia PDF Downloads 148
24641 State of the Science: Digital Therapies in Pediatric Mental Health

Authors: Billy Zou

Abstract:

Statement of the Problem: The burden of mental illness and problem behaviors in adolescence has risen worldwide. While less than 50% of teens have access to traditional mental health care, more than 73% have smartphones. Internet-based interventions offer advantages such as cost-effectiveness, availability, and flexibility. Methodology & Theoretical Orientation: A literature review was done using a PubMed search with the words mental health app yielding 2113 results. 103 articles that met inclusion criteria were reviewed, and findings were then described and synthesized. Findings: 1. Computer-based CBT was found to be effective for OCD, depression, social phobia, and panic disorder. 2. Web-based psychoeducation reduced problem behavior and improved parental well-being. 3. There is limited evidence for mobile-phone-based apps, but preliminary results suggest computer-based interventions are transferrable to mobile apps. 4. Adherence to app-based treatment was correlated with impressions about the user interface Conclusion & Significance: There is evidence for the effectiveness of computer-based programs in filling the significant gaps that currently exist in mental health delivery in the United States and internationally. There is also potential and theoretical validity for mobile-based apps to do the same, though more data is needed.

Keywords: children's mental health, mental health app, child and adolecent psychiatry, digital therapy

Procedia PDF Downloads 64
24640 Big Data Analysis with RHadoop

Authors: Ji Eun Shin, Byung Ho Jung, Dong Hoon Lim

Abstract:

It is almost impossible to store or analyze big data increasing exponentially with traditional technologies. Hadoop is a new technology to make that possible. R programming language is by far the most popular statistical tool for big data analysis based on distributed processing with Hadoop technology. With RHadoop that integrates R and Hadoop environment, we implemented parallel multiple regression analysis with different sizes of actual data. Experimental results showed our RHadoop system was much faster as the number of data nodes increases. We also compared the performance of our RHadoop with lm function and big lm packages available on big memory. The results showed that our RHadoop was faster than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases.

Keywords: big data, Hadoop, parallel regression analysis, R, RHadoop

Procedia PDF Downloads 424
24639 A Mutually Exclusive Task Generation Method Based on Data Augmentation

Authors: Haojie Wang, Xun Li, Rui Yin

Abstract:

In order to solve the memorization overfitting in the meta-learning MAML algorithm, a method of generating mutually exclusive tasks based on data augmentation is proposed. This method generates a mutex task by corresponding one feature of the data to multiple labels, so that the generated mutex task is inconsistent with the data distribution in the initial dataset. Because generating mutex tasks for all data will produce a large number of invalid data and, in the worst case, lead to exponential growth of computation, this paper also proposes a key data extraction method, that only extracts part of the data to generate the mutex task. The experiments show that the method of generating mutually exclusive tasks can effectively solve the memorization overfitting in the meta-learning MAML algorithm.

Keywords: data augmentation, mutex task generation, meta-learning, text classification.

Procedia PDF Downloads 82
24638 Efficient Positioning of Data Aggregation Point for Wireless Sensor Network

Authors: Sifat Rahman Ahona, Rifat Tasnim, Naima Hassan

Abstract:

Data aggregation is a helpful technique for reducing the data communication overhead in wireless sensor network. One of the important tasks of data aggregation is positioning of the aggregator points. There are a lot of works done on data aggregation. But, efficient positioning of the aggregators points is not focused so much. In this paper, authors are focusing on the positioning or the placement of the aggregation points in wireless sensor network. Authors proposed an algorithm to select the aggregators positions for a scenario where aggregator nodes are more powerful than sensor nodes.

Keywords: aggregation point, data communication, data aggregation, wireless sensor network

Procedia PDF Downloads 146
24637 Spatial Econometric Approaches for Count Data: An Overview and New Directions

Authors: Paula Simões, Isabel Natário

Abstract:

This paper reviews a number of theoretical aspects for implementing an explicit spatial perspective in econometrics for modelling non-continuous data, in general, and count data, in particular. It provides an overview of the several spatial econometric approaches that are available to model data that are collected with reference to location in space, from the classical spatial econometrics approaches to the recent developments on spatial econometrics to model count data, in a Bayesian hierarchical setting. Considerable attention is paid to the inferential framework, necessary for structural consistent spatial econometric count models, incorporating spatial lag autocorrelation, to the corresponding estimation and testing procedures for different assumptions, to the constrains and implications embedded in the various specifications in the literature. This review combines insights from the classical spatial econometrics literature as well as from hierarchical modeling and analysis of spatial data, in order to look for new possible directions on the processing of count data, in a spatial hierarchical Bayesian econometric context.

Keywords: spatial data analysis, spatial econometrics, Bayesian hierarchical models, count data

Procedia PDF Downloads 580
24636 The Role of Trust in Intention to Use Prescribed and Non-prescribed Connected Devices

Authors: Jean-michel Sahut, Lubica Hikkerova, Wissal Ben Arfi

Abstract:

The Internet of Things (IoT) emerged over the last few decades in many fields. Healthcare can significantly benefit from IoT. This study aims to examine factors influencing the adoption of IoT in eHealth. To do so, an innovative framework has been developed which applies both the Technology Acceptance Model (TAM) and the United Theory of Acceptance and Use of Technology (UTAUT) model and builds on them by analyzing trust and perceived-risk dimensions to predict intention to use IoT in eHealth. In terms of methodology, a Partial Least Approach Structural Equation Modelling was carried out on a sample of 267 French users. The findings of this research support the significant positive effect of constructs set out in the TAM (perceived ease of use) on predicting behavioral intention by adding the effects identified for UTAUT variables. This research also demonstrates how perceived risk and trust are significant factors for models examining behavioral intentions to use IoT. Perceived risk enhanced by the trust has a significant effect on patients’ behavioral intentions. Moreover, the results highlight the key role of prescription as a moderator of IoT adoption in eHealth. Depending on whether an individual has a prescription to use connected devices or not, ease of use has a stronger impact on adoption, while trust has a negative impact on adoption for users without a prescription. In accordance with the empirical results, several practical implications can be proposed. All connected devices applied in a medical context should be divided into groups according to their functionality: whether they are essential for the patient’s health and whether they require a prescription or not. Devices used with a prescription are easily accepted because the intention to use them is moderated by the medical trust (discussed above). For users without a prescription, ease of use is a more significant factor than for users who have a prescription. This suggests that currently, connected e-Health devices and online healthcare systems have to take this factor into account to better meet the needs and expectations of end-users.

Keywords: internet of things, Healthcare, trust, consumer acceptance

Procedia PDF Downloads 131
24635 Fuzzy Optimization Multi-Objective Clustering Ensemble Model for Multi-Source Data Analysis

Authors: C. B. Le, V. N. Pham

Abstract:

In modern data analysis, multi-source data appears more and more in real applications. Multi-source data clustering has emerged as a important issue in the data mining and machine learning community. Different data sources provide information about different data. Therefore, multi-source data linking is essential to improve clustering performance. However, in practice multi-source data is often heterogeneous, uncertain, and large. This issue is considered a major challenge from multi-source data. Ensemble is a versatile machine learning model in which learning techniques can work in parallel, with big data. Clustering ensemble has been shown to outperform any standard clustering algorithm in terms of accuracy and robustness. However, most of the traditional clustering ensemble approaches are based on single-objective function and single-source data. This paper proposes a new clustering ensemble method for multi-source data analysis. The fuzzy optimized multi-objective clustering ensemble method is called FOMOCE. Firstly, a clustering ensemble mathematical model based on the structure of multi-objective clustering function, multi-source data, and dark knowledge is introduced. Then, rules for extracting dark knowledge from the input data, clustering algorithms, and base clusterings are designed and applied. Finally, a clustering ensemble algorithm is proposed for multi-source data analysis. The experiments were performed on the standard sample data set. The experimental results demonstrate the superior performance of the FOMOCE method compared to the existing clustering ensemble methods and multi-source clustering methods.

Keywords: clustering ensemble, multi-source, multi-objective, fuzzy clustering

Procedia PDF Downloads 173
24634 Morphological Rules of Bangla Repetition Words for UNL Based Machine Translation

Authors: Nawab Yousuf Ali, S. Golam, A. Ameer, Ashok Toru Roy

Abstract:

This paper develops new morphological rules suitable for Bangla repetition words to be incorporated into an inter lingua representation called Universal Networking Language (UNL). The proposed rules are to be used to combine verb roots and their inflexions to produce words which are then combined with other similar types of words to generate repetition words. This paper outlines the format of morphological rules for different types of repetition words that come from verb roots based on the framework of UNL provided by the UNL centre of the Universal Networking Digital Language (UNDL) foundation.

Keywords: Universal Networking Language (UNL), universal word (UW), head word (HW), Bangla-UNL Dictionary, morphological rule, enconverter (EnCo)

Procedia PDF Downloads 304
24633 Efficiency of Information Technology Based Learning and Teaching in Higher Educations

Authors: Mahalingam Palaniandi

Abstract:

Higher education plays vital role in the nation building process for a country and the rest of world. The higher education sector develops the change-agents for the various fields which will help the human-kind wheel to run further. Conventional and traditional class-room based learning and teaching was followed in many decades which is one-to-one and one-to-many. In a way, these are simplest form of learners to be assembled in a class room wherein the teacher used the blackboard to demonstrate the theory and laboratories used for practical. As the technology evolved tremendously for the last 40 years, the teaching and learning environment changed slowly, wherein, the learning community will be anywhere in the world and teacher deliver the content through internet based tools such as video conferencing, web based conferencing tools or E-learning platforms such as Blackboard or noodle. Present day, the mobile technologies plays an important tool to deliver the teaching content on-the-go. Both PC based and mobile based learning technology brought the learning and teaching community together in various aspects. However, as the learning technology also brought various hurdles for learning processes such as plagiarism and not using the reference books entirely as most of the students wants the information instantaneously using internet without actually going to the library to take the notes from the millions of the books which are not available online as e-books which result lack of fundamental knowledge of the concepts complex theories. However, technology is inseparable in human life, now-a-days and every part of it contains piece of information technology right from computers to home appliances. To make use of the IT based learning and teaching at most efficiency, we should have a proper framework and recommendations laid to the learning community in order to derive the maximum efficiency from the IT based teaching and leaning. This paper discusses various IT based tools available for the learning community, efficiency from its usage and recommendations for the suitable framework that needs to be implemented at higher education institutions which makes the learners stronger in both theory as well as real-time knowledge of their studies that is going to be used in their future for the better world.

Keywords: higher education, e-learning, teaching learning, eLearning tools

Procedia PDF Downloads 415
24632 Evaluation of Virtual Reality for the Rehabilitation of Athlete Lower Limb Musculoskeletal Injury: A Method for Obtaining Practitioner’s Viewpoints through Observation and Interview

Authors: Hannah K. M. Tang, Muhammad Ateeq, Mark J. Lake, Badr Abdullah, Frederic A. Bezombes

Abstract:

Based on a theoretical assessment of current literature, virtual reality (VR) could help to treat sporting injuries in a number of ways. However, it is important to obtain rehabilitation specialists’ perspectives in order to design, develop and validate suitable content for a VR application focused on treatment. Subsequently, a one-day observation and interview study focused on the use of VR for the treatment of lower limb musculoskeletal conditions in athletes was conducted at St George’s Park England National Football Centre with rehabilitation specialists. The current paper established the methods suitable for obtaining practitioner’s viewpoints through observation and interview in this context. Particular detail was provided regarding the method of qualitatively processing interview results using the qualitative data analysis software tool NVivo, in order to produce a narrative of overarching themes. The observations and overarching themes identified could be used as a framework and success criteria of a VR application developed in future research. In conclusion, this work explained the methods deemed suitable for obtaining practitioner’s viewpoints through observation and interview. This was required in order to highlight characteristics and features of a VR application designed to treat lower limb musculoskeletal injury of athletes and could be built upon to direct future work.

Keywords: athletes, lower-limb musculoskeletal injury, rehabilitation, return-to-sport, virtual reality

Procedia PDF Downloads 245
24631 Impact Analysis of a School-Based Oral Health Program in Brazil

Authors: Fabio L. Vieira, Micaelle F. C. Lemos, Luciano C. Lemos, Rafaela S. Oliveira, Ian A. Cunha

Abstract:

Brazil has some challenges ahead related to population oral health, most of them associated with the need of expanding into the local level its promotion and prevention activities, offer equal access to services and promote changes in the lifestyle of the population. The program implemented an oral health initiative in public schools in the city of Salvador, Bahia. The mission was to improve oral health among students on primary and secondary education, from 2 to 15 years old, using the school as a pathway to increase access to healthcare. The main actions consisted of a team's visit to the schools with educational sessions for dental cavity prevention and individual assessment. The program incorporated a clinical surveillance component through a dental evaluation of every student searching for dental disease and caries, standardization of the dentists’ team to reach uniform classification on the assessments, and the use of an online platform to register data directly from the schools. Sequentially, the students with caries were referred for free clinical treatment on the program’s Health Centre. The primary purpose of this study was to analyze the effects and outcomes of this school-based oral health program. The study sample was composed by data of a period of 3 years - 2015 to 2017 - from 13 public schools on the suburb of the city of Salvador with a total number of assessments of 9,278 on this period. From the data collected the prevalence of children with decay on permanent teeth was chosen as the most reliable indicator. The prevalence was calculated for each one of the 13 schools using the number of children with 1 or more dental caries on permanent teeth divided by the total number of students assessed for school each year. Then the percentage change per year was calculated for each school. Some schools presented a higher variation on the total number of assessments in one of the three years, so for these, the percentage change calculation was done using the two years with less variation. The results show that 10 of the 13 schools presented significative improvements for the indicator of caries in permanent teeth. The mean for the number of students with caries percentage reduction on the 13 schools was 26.8%, and the median was 32.2% caries in permanent teeth institution. The highest percentage of improvement reached a decrease of 65.6% on the indicator. Three schools presented a rise in caries prevalence (8.9, 18.9 and 37.2% increase) that, on an initial analysis, seems to be explained with the students’ cohort rotation among other schools, as well as absenteeism on the treatment. In conclusion, the program shows a relevant impact on the reduction of caries in permanent teeth among students and the need for the continuity and expansion of this integrated healthcare approach. It has also been evident the significative of the articulation between health and educational systems representing a fundamental approach to improve healthcare access for children especially in scenarios such as presented in Brazil.

Keywords: primary care, public health, oral health, school-based oral health, data management

Procedia PDF Downloads 123
24630 The Impact of the Atypical Crisis on Educational Migration: Economic and Policy Challenges

Authors: Manana Lobzhanidze, Marine Kobalava, Lali Chikviladze

Abstract:

The global pandemic crisis has had a significant impact on educational migration, substantially limiting young people’s access to education abroad. Therefore, it became necessary to study the economic, demographic, social, cultural and other factors associated with educational migration, to identify the economic and political challenges of educational migration and to develop recommendations. The aim of the research is to study the effects of the atypical crisis on educational migration and to make recommendations on effective migration opportunities based on the identification of economic and policy challenges in this area. Bibliographic research is used to assess the effects of the impact of the atypical crisis on educational migration presented in the papers of various scholars. Against the background of the restrictions imposed during the COVID19 pandemic, migration rates have been analyzed, endogenous and exogenous factors affecting educational migration have been identified. Quantitative and qualitative research of students and graduates of TSU Economics and Business Faculty is conducted, the results have been processed by SPSS program, the factors hindering educational migration and the challenges have been identified. The Internet and digital technologies have been shown to play a vital role in alleviating the challenges posed by the COVID-19 pandemic, however, lack of Internet access and limited financial resources have played a disruptive role in the educational migration process. The analysis of quantitative research materials revealed the problems of educational migration caused by the atypical crisis, while some issues were clarified during the focus group meetings. The following theoretical-methodological approaches were used during the research: a bibliographic research, analysis, synthesis, comparison, selection-grouping are used; Quantitative and qualitative research has been carried out, the results have been processed by SPSS program. The article presents the consequences of the atypical crisis for educational migration, identifies the main economic and policy challenges in the field of educational migration, and develops appropriate recommendations to overcome them.

Keywords: educational migration, atypical crisis, economic-political challenges, educational migration factors

Procedia PDF Downloads 134
24629 Visualization of Taiwan's Religious Social Networking Sites

Authors: Jia-Jane Shuai

Abstract:

Purpose of this research aims to improve understanding of the nature of online religion by examining the religious social websites. What motivates individual users to use the online religious social websites, and which factors affect those motivations. We survey various online religious social websites provided by different religions, especially the Taiwanese folk religion. Based on the theory of the Content Analysis and Social Network Analysis, religious social websites and religious web activities are examined. This research examined the folk religion websites’ presentation and contents that promote the religious use of the Internet in Taiwan. The difference among different religions and religious websites also be compared. First, this study used keywords to examine what types of messages gained the most clicks of “Like”, “Share” and comments on Facebook. Dividing the messages into four media types, namely, text, link, video, and photo, reveal which category receive more likes and comments than the others. Meanwhile, this study analyzed the five dialogic principles of religious websites accessed from mobile phones and also assessed their mobile readiness. Using the five principles of dialogic theory as a basis, do a general survey on the websites with elements of online religion. Second, the project analyzed the characteristics of Taiwanese participants for online religious activities. Grounded by social network analysis and text mining, this study comparatively explores the network structure, interaction pattern, and geographic distribution of users involved in communication networks of the folk religion in social websites and mobile sites. We studied the linkage preference of different religious groups. The difference among different religions and religious websites also be compared. We examined the reasons for the success of these websites, as well as reasons why young users accept new religious media. The outcome of the research will be useful for online religious service providers and non-profit organizations to manage social websites and internet marketing.

Keywords: content analysis, online religion, social network analysis, social websites

Procedia PDF Downloads 156
24628 Modeling Activity Pattern Using XGBoost for Mining Smart Card Data

Authors: Eui-Jin Kim, Hasik Lee, Su-Jin Park, Dong-Kyu Kim

Abstract:

Smart-card data are expected to provide information on activity pattern as an alternative to conventional person trip surveys. The focus of this study is to propose a method for training the person trip surveys to supplement the smart-card data that does not contain the purpose of each trip. We selected only available features from smart card data such as spatiotemporal information on the trip and geographic information system (GIS) data near the stations to train the survey data. XGboost, which is state-of-the-art tree-based ensemble classifier, was used to train data from multiple sources. This classifier uses a more regularized model formalization to control the over-fitting and show very fast execution time with well-performance. The validation results showed that proposed method efficiently estimated the trip purpose. GIS data of station and duration of stay at the destination were significant features in modeling trip purpose.

Keywords: activity pattern, data fusion, smart-card, XGboost

Procedia PDF Downloads 231
24627 Design and Implementation of Medium Access Control Based Routing on Real Wireless Sensor Networks Testbed

Authors: Smriti Agarwal, Ashish Payal, B. V. R. Reddy

Abstract:

IEEE 802.15.4 is a Low Rate Wireless Personal Area Networks (LR-WPAN) standard combined with ZigBee, which is going to enable new applications in Wireless Sensor Networks (WSNs) and Internet of Things (IoT) domain. In recent years, it has become a popular standard for WSNs. Wireless communication among sensor motes, enabled by IEEE 802.15.4 standard, is extensively replacing the existing wired technology in a wide range of monitoring and control applications. Researchers have proposed a routing framework and mechanism that interacts with the IEEE 802.15.4 standard using software platform. In this paper, we have designed and implemented MAC based routing (MBR) based on IEEE 802.15.4 standard using a hardware platform “SENSEnuts”. The experimental results include data through light and temperature sensors obtained from communication between PAN coordinator and source node through coordinator, MAC address of some modules used in the experimental setup, topology of the network created for simulation and the remaining battery power of the source node. Our experimental effort on a WSN Testbed has helped us in bridging the gap between theoretical and practical aspect of implementing IEEE 802.15.4 for WSNs applications.

Keywords: IEEE 802.15.4, routing, WSN, ZigBee

Procedia PDF Downloads 392