Search results for: R data science
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 26130

Search results for: R data science

25530 Big Data Applications for Transportation Planning

Authors: Antonella Falanga, Armando Cartenì

Abstract:

"Big data" refers to extremely vast and complex sets of data, encompassing extraordinarily large and intricate datasets that require specific tools for meaningful analysis and processing. These datasets can stem from diverse origins like sensors, mobile devices, online transactions, social media platforms, and more. The utilization of big data is pivotal, offering the chance to leverage vast information for substantial advantages across diverse fields, thereby enhancing comprehension, decision-making, efficiency, and fostering innovation in various domains. Big data, distinguished by its remarkable attributes of enormous volume, high velocity, diverse variety, and significant value, represent a transformative force reshaping the industry worldwide. Their pervasive impact continues to unlock new possibilities, driving innovation and advancements in technology, decision-making processes, and societal progress in an increasingly data-centric world. The use of these technologies is becoming more widespread, facilitating and accelerating operations that were once much more complicated. In particular, big data impacts across multiple sectors such as business and commerce, healthcare and science, finance, education, geography, agriculture, media and entertainment and also mobility and logistics. Within the transportation sector, which is the focus of this study, big data applications encompass a wide variety, spanning across optimization in vehicle routing, real-time traffic management and monitoring, logistics efficiency, reduction of travel times and congestion, enhancement of the overall transportation systems, but also mitigation of pollutant emissions contributing to environmental sustainability. Meanwhile, in public administration and the development of smart cities, big data aids in improving public services, urban planning, and decision-making processes, leading to more efficient and sustainable urban environments. Access to vast data reservoirs enables deeper insights, revealing hidden patterns and facilitating more precise and timely decision-making. Additionally, advancements in cloud computing and artificial intelligence (AI) have further amplified the potential of big data, enabling more sophisticated and comprehensive analyses. Certainly, utilizing big data presents various advantages but also entails several challenges regarding data privacy and security, ensuring data quality, managing and storing large volumes of data effectively, integrating data from diverse sources, the need for specialized skills to interpret analysis results, ethical considerations in data use, and evaluating costs against benefits. Addressing these difficulties requires well-structured strategies and policies to balance the benefits of big data with privacy, security, and efficient data management concerns. Building upon these premises, the current research investigates the efficacy and influence of big data by conducting an overview of the primary and recent implementations of big data in transportation systems. Overall, this research allows us to conclude that big data better provide to enhance rational decision-making for mobility choices and is imperative for adeptly planning and allocating investments in transportation infrastructures and services.

Keywords: big data, public transport, sustainable mobility, transport demand, transportation planning

Procedia PDF Downloads 48
25529 Need of Trained Clinical Research Professionals Globally to Conduct Clinical Trials

Authors: Tambe Daniel Atem

Abstract:

Background: Clinical Research is an organized research on human beings intended to provide adequate information on the drug use as a therapeutic agent on its safety and efficacy. The significance of the study is to educate the global health and life science graduates in Clinical Research in depth to perform better as it involves testing drugs on human beings. Objectives: to provide an overall understanding of the scientific approach to the evaluation of new and existing medical interventions and to apply ethical and regulatory principles appropriate to any individual research. Methodology: It is based on – Primary data analysis and Secondary data analysis. Primary data analysis: means the collection of data from journals, the internet, and other online sources. Secondary data analysis: a survey was conducted with a questionnaire to interview the Clinical Research Professionals to understand the need of training to perform clinical trials globally. The questionnaire consisted details of the professionals working with the expertise. It also included the areas of clinical research which needed intense training before entering into hardcore clinical research domain. Results: The Clinical Trials market worldwide worth over USD 26 billion and the industry has employed an estimated 2,10,000 people in the US and over 70,000 in the U.K, and they form one-third of the total research and development staff. There are more than 2,50,000 vacant positions globally with salary variations in the regions for a Clinical Research Coordinator. R&D cost on new drug development is estimated at US$ 70-85 billion. The cost of doing clinical trials for a new drug is US$ 200-250 million. Due to an increase trained Clinical Research Professionals India has emerged as a global hub for clinical research. The Global Clinical Trial outsourcing opportunity in India in the pharmaceutical industry increased to more than $2 billion in 2014 due to increased outsourcing from U.S and Europe to India. Conclusion: Assessment of training need is recommended for newer Clinical Research Professionals and trial sites, especially prior the conduct of larger confirmatory clinical trials.

Keywords: clinical research, clinical trials, clinical research professionals

Procedia PDF Downloads 441
25528 Imputation of Incomplete Large-Scale Monitoring Count Data via Penalized Estimation

Authors: Mohamed Dakki, Genevieve Robin, Marie Suet, Abdeljebbar Qninba, Mohamed A. El Agbani, Asmâa Ouassou, Rhimou El Hamoumi, Hichem Azafzaf, Sami Rebah, Claudia Feltrup-Azafzaf, Nafouel Hamouda, Wed a.L. Ibrahim, Hosni H. Asran, Amr A. Elhady, Haitham Ibrahim, Khaled Etayeb, Essam Bouras, Almokhtar Saied, Ashrof Glidan, Bakar M. Habib, Mohamed S. Sayoud, Nadjiba Bendjedda, Laura Dami, Clemence Deschamps, Elie Gaget, Jean-Yves Mondain-Monval, Pierre Defos Du Rau

Abstract:

In biodiversity monitoring, large datasets are becoming more and more widely available and are increasingly used globally to estimate species trends and con- servation status. These large-scale datasets challenge existing statistical analysis methods, many of which are not adapted to their size, incompleteness and heterogeneity. The development of scalable methods to impute missing data in incomplete large-scale monitoring datasets is crucial to balance sampling in time or space and thus better inform conservation policies. We developed a new method based on penalized Poisson models to impute and analyse incomplete monitoring data in a large-scale framework. The method al- lows parameterization of (a) space and time factors, (b) the main effects of predic- tor covariates, as well as (c) space–time interactions. It also benefits from robust statistical and computational capability in large-scale settings. The method was tested extensively on both simulated and real-life waterbird data, with the findings revealing that it outperforms six existing methods in terms of missing data imputation errors. Applying the method to 16 waterbird species, we estimated their long-term trends for the first time at the entire North African scale, a region where monitoring data suffer from many gaps in space and time series. This new approach opens promising perspectives to increase the accuracy of species-abundance trend estimations. We made it freely available in the r package ‘lori’ (https://CRAN.R-project.org/package=lori) and recommend its use for large- scale count data, particularly in citizen science monitoring programmes.

Keywords: biodiversity monitoring, high-dimensional statistics, incomplete count data, missing data imputation, waterbird trends in North-Africa

Procedia PDF Downloads 136
25527 Application of Artificial Neural Network Technique for Diagnosing Asthma

Authors: Azadeh Bashiri

Abstract:

Introduction: Lack of proper diagnosis and inadequate treatment of asthma leads to physical and financial complications. This study aimed to use data mining techniques and creating a neural network intelligent system for diagnosis of asthma. Methods: The study population is the patients who had visited one of the Lung Clinics in Tehran. Data were analyzed using the SPSS statistical tool and the chi-square Pearson's coefficient was the basis of decision making for data ranking. The considered neural network is trained using back propagation learning technique. Results: According to the analysis performed by means of SPSS to select the top factors, 13 effective factors were selected, in different performances, data was mixed in various forms, so the different models were made for training the data and testing networks and in all different modes, the network was able to predict correctly 100% of all cases. Conclusion: Using data mining methods before the design structure of system, aimed to reduce the data dimension and the optimum choice of the data, will lead to a more accurate system. Therefore, considering the data mining approaches due to the nature of medical data is necessary.

Keywords: asthma, data mining, Artificial Neural Network, intelligent system

Procedia PDF Downloads 260
25526 Interpreting Privacy Harms from a Non-Economic Perspective

Authors: Christopher Muhawe, Masooda Bashir

Abstract:

With increased Internet Communication Technology(ICT), the virtual world has become the new normal. At the same time, there is an unprecedented collection of massive amounts of data by both private and public entities. Unfortunately, this increase in data collection has been in tandem with an increase in data misuse and data breach. Regrettably, the majority of data breach and data misuse claims have been unsuccessful in the United States courts for the failure of proof of direct injury to physical or economic interests. The requirement to express data privacy harms from an economic or physical stance negates the fact that not all data harms are physical or economic in nature. The challenge is compounded by the fact that data breach harms and risks do not attach immediately. This research will use a descriptive and normative approach to show that not all data harms can be expressed in economic or physical terms. Expressing privacy harms purely from an economic or physical harm perspective negates the fact that data insecurity may result into harms which run counter the functions of privacy in our lives. The promotion of liberty, selfhood, autonomy, promotion of human social relations and the furtherance of the existence of a free society. There is no economic value that can be placed on these functions of privacy. The proposed approach addresses data harms from a psychological and social perspective.

Keywords: data breach and misuse, economic harms, privacy harms, psychological harms

Procedia PDF Downloads 181
25525 Machine Learning Analysis of Student Success in Introductory Calculus Based Physics I Course

Authors: Chandra Prayaga, Aaron Wade, Lakshmi Prayaga, Gopi Shankar Mallu

Abstract:

This paper presents the use of machine learning algorithms to predict the success of students in an introductory physics course. Data having 140 rows pertaining to the performance of two batches of students was used. The lack of sufficient data to train robust machine learning models was compensated for by generating synthetic data similar to the real data. CTGAN and CTGAN with Gaussian Copula (Gaussian) were used to generate synthetic data, with the real data as input. To check the similarity between the real data and each synthetic dataset, pair plots were made. The synthetic data was used to train machine learning models using the PyCaret package. For the CTGAN data, the Ada Boost Classifier (ADA) was found to be the ML model with the best fit, whereas the CTGAN with Gaussian Copula yielded Logistic Regression (LR) as the best model. Both models were then tested for accuracy with the real data. ROC-AUC analysis was performed for all the ten classes of the target variable (Grades A, A-, B+, B, B-, C+, C, C-, D, F). The ADA model with CTGAN data showed a mean AUC score of 0.4377, but the LR model with the Gaussian data showed a mean AUC score of 0.6149. ROC-AUC plots were obtained for each Grade value separately. The LR model with Gaussian data showed consistently better AUC scores compared to the ADA model with CTGAN data, except in two cases of the Grade value, C- and A-.

Keywords: machine learning, student success, physics course, grades, synthetic data, CTGAN, gaussian copula CTGAN

Procedia PDF Downloads 30
25524 The Sustainable Blue Economy Innovation and Growth: Data Based on China for 2006-2015 Years

Authors: Mingbao Chen

Abstract:

The blue economy is a new comprehensive marine economy integrated with resources, industries, and regions, and is an upgraded version of the marine economy. The blue economy attaches great importance to the coordinated development of the ecological environment and the economy, which is an emerging economic form advocated by all countries in the world. This paper constructs the model including four variables:natural capital, economic capital, intellectual capital, cultural capital. Theoretically, this paper deduces the function mechanism of variables on economic growth, and empirically calculates the driving force and influence of the blue economy on the national economy by using data of China's 2006-2015 year. The results show that natural capital and economic capital remain the main factors of blue growth in the blue economy. And with the development of economic society and technological progress, the role of intellectual capital and cultural capital is bigger and bigger. Therefore, promoting the development of marine science and technology and culture is the focus of the future blue economic development.

Keywords: blue growth, natural capital, intellectual capital, cultural capital

Procedia PDF Downloads 144
25523 Biomechanical Analysis and Interpretation of Pitching Sequences for Enhanced Performance Programming

Authors: Corey F. Fitzgerald

Abstract:

This study provides a comprehensive examination of the biomechanical sequencing inherent in pitching motions, coupled with an advanced methodology for interpreting gathered data to inform programming strategies. The analysis is conducted utilizing state-of-the-art biomechanical laboratory equipment capable of detecting subtle changes and deviations, facilitating highly informed decision-making processes. Through this presentation, the intricate dynamics of pitching sequences are meticulously discussed to highlight the complex movement patterns accessible and actionable for performance enhancement purposes in the weight room.

Keywords: sport science, applied biomechanics, strength and conditioning, applied research

Procedia PDF Downloads 42
25522 Attitude towards Biology among Academic Talented Students in Term of Gender: Case Study of ASASIpintar Pre–University Programme

Authors: Sy Ing Ong, N. Norazman, H. W. You, A. Zahidi, R. Ahmad Faisal

Abstract:

The decline in students’ involvement in the fields of Science, Technology, Engineering and Mathematics (STEM) worldwide is alarming. In Malaysia, this scenario also draws the attention of Ministry of Education due to the high demand for professionals in STEM industries and for the sake of country development. Intensive researches have been done worldwide to identify the best solution to improve the enrolment of students in science studies. Attitude is being identified as one of the key factors that will influence students’ interest and achievement in academic. Male students are always associated with negative attitudes compared to female students towards the study of Biological science. Hence, this study investigates the attitudes of students towards Biology in the setting of an academic talented institution in Malaysia namely ASASIpintar Pre-University Programme. A total of 84 students were randomly selected from 127 students as the samples of this study. The instrument of Biology Attitude Scale (BAS) from Russell and Hollander (1975) was used to identify the attitudes of samples. Based on the analysis, there was no significant difference in the students’ mean attitude towards Biology subject in this institution between genders with a significant level of p = 0.05. This indicated that gender is not a key factor that influences students’ attitude towards Biology in this study. Future research can be done on other factors that will contribute to the attitude of students towards biology in Malaysia, especially for academically talented students.

Keywords: academic talented, attitude, biology, gender

Procedia PDF Downloads 209
25521 Rationalizing the Utilization of Interactive Engagement Strategies in Teaching Specialized Science Courses of STEM and GA Strands in the Academic Track of Philippine Senior High School Curriculum

Authors: Raul G. Angeles

Abstract:

The Philippine government instituted major reforms in its educational system. The Department of Education pushes the K to 12 program that makes kindergarten mandatory and adds two years of senior high school to the country's basic education. In essence, the students’ stay in basic education particularly those who are supposedly going to college is extended. The majority of the students expressed that they will be taking the Academic Track of the Senior High School curriculum specifically the Science, Technology, Engineering and Mathematics (STEM) and General Academic (GA) strands. Almost certainly, instruction should match the students' styles and thus through this descriptive study a city survey was conducted to explore the teaching strategies preferences of junior high school students and teachers who will be promoted to senior high school during the Academic Year 2016-2017. This study was conducted in selected public and private secondary schools in Metro Manila. Questionnaires were distributed to students and teachers; and series of follow-up interviews were also carried out to generate additional information. Preferences of students are centered on employing innovations such as technology, cooperative and problem-based learning. While the students will still be covered by basic education their interests in science are sparking to a point where the usual teaching styles may no longer work to them and for that cause, altering the teaching methods is recommended to create a teacher-student style matching. Other effective strategies must likewise be implemented.

Keywords: curriculum development, effective teaching strategies, problem-based learning, senior high school, science education, technology

Procedia PDF Downloads 247
25520 Data Access, AI Intensity, and Scale Advantages

Authors: Chuping Lo

Abstract:

This paper presents a simple model demonstrating that ceteris paribus countries with lower barriers to accessing global data tend to earn higher incomes than other countries. Therefore, large countries that inherently have greater data resources tend to have higher incomes than smaller countries, such that the former may be more hesitant than the latter to liberalize cross-border data flows to maintain this advantage. Furthermore, countries with higher artificial intelligence (AI) intensity in production technologies tend to benefit more from economies of scale in data aggregation, leading to higher income and more trade as they are better able to utilize global data.

Keywords: digital intensity, digital divide, international trade, scale of economics

Procedia PDF Downloads 53
25519 Secured Transmission and Reserving Space in Images Before Encryption to Embed Data

Authors: G. R. Navaneesh, E. Nagarajan, C. H. Rajam Raju

Abstract:

Nowadays the multimedia data are used to store some secure information. All previous methods allocate a space in image for data embedding purpose after encryption. In this paper, we propose a novel method by reserving space in image with a boundary surrounded before encryption with a traditional RDH algorithm, which makes it easy for the data hider to reversibly embed data in the encrypted images. The proposed method can achieve real time performance, that is, data extraction and image recovery are free of any error. A secure transmission process is also discussed in this paper, which improves the efficiency by ten times compared to other processes as discussed.

Keywords: secure communication, reserving room before encryption, least significant bits, image encryption, reversible data hiding

Procedia PDF Downloads 400
25518 Identity Verification Using k-NN Classifiers and Autistic Genetic Data

Authors: Fuad M. Alkoot

Abstract:

DNA data have been used in forensics for decades. However, current research looks at using the DNA as a biometric identity verification modality. The goal is to improve the speed of identification. We aim at using gene data that was initially used for autism detection to find if and how accurate is this data for identification applications. Mainly our goal is to find if our data preprocessing technique yields data useful as a biometric identification tool. We experiment with using the nearest neighbor classifier to identify subjects. Results show that optimal classification rate is achieved when the test set is corrupted by normally distributed noise with zero mean and standard deviation of 1. The classification rate is close to optimal at higher noise standard deviation reaching 3. This shows that the data can be used for identity verification with high accuracy using a simple classifier such as the k-nearest neighbor (k-NN). 

Keywords: biometrics, genetic data, identity verification, k nearest neighbor

Procedia PDF Downloads 238
25517 Intelligent Earthquake Prediction System Based On Neural Network

Authors: Emad Amar, Tawfik Khattab, Fatma Zada

Abstract:

Predicting earthquakes is an important issue in the study of geography. Accurate prediction of earthquakes can help people to take effective measures to minimize the loss of personal and economic damage, such as large casualties, destruction of buildings and broken of traffic, occurred within a few seconds. United States Geological Survey (USGS) science organization provides reliable scientific information of Earthquake Existed throughout history & Preliminary database from the National Center Earthquake Information (NEIC) show some useful factors to predict an earthquake in a seismic area like Aleutian Arc in the U.S. state of Alaska. The main advantage of this prediction method that it does not require any assumption, it makes prediction according to the future evolution of object's time series. The article compares between simulation data result from trained BP and RBF neural network versus actual output result from the system calculations. Therefore, this article focuses on analysis of data relating to real earthquakes. Evaluation results show better accuracy and higher speed by using radial basis functions (RBF) neural network.

Keywords: BP neural network, prediction, RBF neural network, earthquake

Procedia PDF Downloads 484
25516 Patient Tracking Challenges During Disasters and Emergencies

Authors: Mohammad H. Yarmohammadian, Reza Safdari, Mahmoud Keyvanara, Nahid Tavakoli

Abstract:

One of the greatest challenges in disaster and emergencies is patient tracking. The concept of tracking has different denotations. One of the meanings refers to tracking patients’ physical locations and the other meaning refers to tracking patients ‘medical needs during emergency services. The main goal of patient tracking is to provide patient safety during disaster and emergencies and manage the flow of patient and information in different locations. In most of cases, there are not sufficient and accurate data regarding the number of injuries, medical conditions and their accommodation and transference. The objective of the present study is to survey on patient tracking issue in natural disaster and emergencies. Methods: This was a narrative study in which the population was E-Journals and the electronic database such as PubMed, Proquest, Science direct, Elsevier, etc. Data was gathered by Extraction Form. All data were analyzed via content analysis. Results: In many countries there is no appropriate and rapid method for tracking patients and transferring victims after the occurrence of incidents. The absence of reliable data of patients’ transference and accommodation, even in the initial hours and days after the occurrence of disasters, and coordination for appropriate resource allocation, have faced challenges for evaluating needs and services challenges. Currently, most of emergency services are based on paper systems, while these systems do not act appropriately in great disasters and incidents and this issue causes information loss. Conclusion: Patient tracking system should update the location of patients or evacuees and information related to their states. Patients’ information should be accessible for authorized users to continue their treatment, accommodation and transference. Also it should include timely information of patients’ location as soon as they arrive somewhere and leave therein such a way that health care professionals can be able to provide patients’ proper medical treatment.

Keywords: patient tracking, challenges, disaster, emergency

Procedia PDF Downloads 292
25515 Data Quality as a Pillar of Data-Driven Organizations: Exploring the Benefits of Data Mesh

Authors: Marc Bachelet, Abhijit Kumar Chatterjee, José Manuel Avila

Abstract:

Data quality is a key component of any data-driven organization. Without data quality, organizations cannot effectively make data-driven decisions, which often leads to poor business performance. Therefore, it is important for an organization to ensure that the data they use is of high quality. This is where the concept of data mesh comes in. Data mesh is an organizational and architectural decentralized approach to data management that can help organizations improve the quality of data. The concept of data mesh was first introduced in 2020. Its purpose is to decentralize data ownership, making it easier for domain experts to manage the data. This can help organizations improve data quality by reducing the reliance on centralized data teams and allowing domain experts to take charge of their data. This paper intends to discuss how a set of elements, including data mesh, are tools capable of increasing data quality. One of the key benefits of data mesh is improved metadata management. In a traditional data architecture, metadata management is typically centralized, which can lead to data silos and poor data quality. With data mesh, metadata is managed in a decentralized manner, ensuring accurate and up-to-date metadata, thereby improving data quality. Another benefit of data mesh is the clarification of roles and responsibilities. In a traditional data architecture, data teams are responsible for managing all aspects of data, which can lead to confusion and ambiguity in responsibilities. With data mesh, domain experts are responsible for managing their own data, which can help provide clarity in roles and responsibilities and improve data quality. Additionally, data mesh can also contribute to a new form of organization that is more agile and adaptable. By decentralizing data ownership, organizations can respond more quickly to changes in their business environment, which in turn can help improve overall performance by allowing better insights into business as an effect of better reports and visualization tools. Monitoring and analytics are also important aspects of data quality. With data mesh, monitoring, and analytics are decentralized, allowing domain experts to monitor and analyze their own data. This will help in identifying and addressing data quality problems in quick time, leading to improved data quality. Data culture is another major aspect of data quality. With data mesh, domain experts are encouraged to take ownership of their data, which can help create a data-driven culture within the organization. This can lead to improved data quality and better business outcomes. Finally, the paper explores the contribution of AI in the coming years. AI can help enhance data quality by automating many data-related tasks, like data cleaning and data validation. By integrating AI into data mesh, organizations can further enhance the quality of their data. The concepts mentioned above are illustrated by AEKIDEN experience feedback. AEKIDEN is an international data-driven consultancy that has successfully implemented a data mesh approach. By sharing their experience, AEKIDEN can help other organizations understand the benefits and challenges of implementing data mesh and improving data quality.

Keywords: data culture, data-driven organization, data mesh, data quality for business success

Procedia PDF Downloads 119
25514 The Coauthorship Network Analysis of the Norwegian School of Economics

Authors: Ivan Belik, Kurt Jornsten

Abstract:

We construct the coauthorship network based on the scientific collaboration between the faculty members at the Norwegian School of Economics (NHH) and based on their international academic publication experience. The network structure is based on the NHH faculties’ publications recognized by the ISI Web of Science for the period 1950 – Spring, 2014. The given network covers the publication activities of the NHH faculty members (over six departments) based on the information retrieved from the ISI Web of Science in Spring, 2014. In this paper we analyse the constructed coauthorship network in different aspects of the theory of social networks analysis.

Keywords: coauthorship networks, social networks analysis, Norwegian School of Economics, ISI

Procedia PDF Downloads 419
25513 Ripple Effect Analysis of Government Investment for Research and Development by the Artificial Neural Networks

Authors: Hwayeon Song

Abstract:

The long-term purpose of research and development (R&D) programs is to strengthen national competitiveness by developing new knowledge and technologies. Thus, it is important to determine a proper budget for government programs to maintain the vigor of R&D when the total funding is tight due to the national deficit. In this regard, a ripple effect analysis for the budgetary changes in R&D programs is necessary as well as an investigation of the current status. This study proposes a new approach using Artificial Neural Networks (ANN) for both tasks. It particularly focuses on R&D programs related to Construction and Transportation (C&T) technology in Korea. First, key factors in C&T technology are explored to draw impact indicators in three areas: economy, society, and science and technology (S&T). Simultaneously, ANN is employed to evaluate the relationship between data variables. From this process, four major components in R&D including research personnel, expenses, management, and equipment are assessed. Then the ripple effect analysis is performed to see the changes in the hypothetical future by modifying current data. Any research findings can offer an alternative strategy about R&D programs as well as a new analysis tool.

Keywords: Artificial Neural Networks, construction and transportation technology, Government Research and Development, Ripple Effect

Procedia PDF Downloads 229
25512 Effect of Ausubel's Advance Organizer Model to Enhancing Meta-Cognition of Students at Secondary Level

Authors: Qaisara Parveen, M. Imran Yousuf

Abstract:

The purpose of this study was to find the effectiveness of the use of advance organizer model for enhancing meta-cognition of students in the subject of science. It was hypothesized that the students of experimental group taught through advance organizer model would show the better cognition than the students of control group taught through traditional teaching. The population of the study consisted of all secondary school students studying in government high school located in Rawalpindi. The sample of the study consisted of 50 students of 9th class of humanities group. The sample was selected on the basis of their pretest scores through matching, and the groups were randomly assigned for the treatment. The experimental group was taught through advance organizer model while the control group was taught through traditional teaching. The self-developed achievement test was used for the purpose of pretest and posttest. After collecting the pre-test score and post-test score, the data was analyzed and interpreted by use of descriptive statistics as mean and standard deviation and inferential statistics t-test. The findings indicate that students taught using advance organizers had a higher level of meta-cognition as compared to control group. Further, meta cognition level of boys was found higher than that of girls students. This study also revealed the fact that though the students at different meta-cognition level approached learning situations in a different manner, Advance organizer model is far superior to Traditional method of teaching.

Keywords: descriptive, experimental, humanities, meta-cognition, statistics, science

Procedia PDF Downloads 294
25511 Big Data Analysis with RHadoop

Authors: Ji Eun Shin, Byung Ho Jung, Dong Hoon Lim

Abstract:

It is almost impossible to store or analyze big data increasing exponentially with traditional technologies. Hadoop is a new technology to make that possible. R programming language is by far the most popular statistical tool for big data analysis based on distributed processing with Hadoop technology. With RHadoop that integrates R and Hadoop environment, we implemented parallel multiple regression analysis with different sizes of actual data. Experimental results showed our RHadoop system was much faster as the number of data nodes increases. We also compared the performance of our RHadoop with lm function and big lm packages available on big memory. The results showed that our RHadoop was faster than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases.

Keywords: big data, Hadoop, parallel regression analysis, R, RHadoop

Procedia PDF Downloads 422
25510 A Mutually Exclusive Task Generation Method Based on Data Augmentation

Authors: Haojie Wang, Xun Li, Rui Yin

Abstract:

In order to solve the memorization overfitting in the meta-learning MAML algorithm, a method of generating mutually exclusive tasks based on data augmentation is proposed. This method generates a mutex task by corresponding one feature of the data to multiple labels, so that the generated mutex task is inconsistent with the data distribution in the initial dataset. Because generating mutex tasks for all data will produce a large number of invalid data and, in the worst case, lead to exponential growth of computation, this paper also proposes a key data extraction method, that only extracts part of the data to generate the mutex task. The experiments show that the method of generating mutually exclusive tasks can effectively solve the memorization overfitting in the meta-learning MAML algorithm.

Keywords: data augmentation, mutex task generation, meta-learning, text classification.

Procedia PDF Downloads 80
25509 Efficient Positioning of Data Aggregation Point for Wireless Sensor Network

Authors: Sifat Rahman Ahona, Rifat Tasnim, Naima Hassan

Abstract:

Data aggregation is a helpful technique for reducing the data communication overhead in wireless sensor network. One of the important tasks of data aggregation is positioning of the aggregator points. There are a lot of works done on data aggregation. But, efficient positioning of the aggregators points is not focused so much. In this paper, authors are focusing on the positioning or the placement of the aggregation points in wireless sensor network. Authors proposed an algorithm to select the aggregators positions for a scenario where aggregator nodes are more powerful than sensor nodes.

Keywords: aggregation point, data communication, data aggregation, wireless sensor network

Procedia PDF Downloads 144
25508 Spatial Econometric Approaches for Count Data: An Overview and New Directions

Authors: Paula Simões, Isabel Natário

Abstract:

This paper reviews a number of theoretical aspects for implementing an explicit spatial perspective in econometrics for modelling non-continuous data, in general, and count data, in particular. It provides an overview of the several spatial econometric approaches that are available to model data that are collected with reference to location in space, from the classical spatial econometrics approaches to the recent developments on spatial econometrics to model count data, in a Bayesian hierarchical setting. Considerable attention is paid to the inferential framework, necessary for structural consistent spatial econometric count models, incorporating spatial lag autocorrelation, to the corresponding estimation and testing procedures for different assumptions, to the constrains and implications embedded in the various specifications in the literature. This review combines insights from the classical spatial econometrics literature as well as from hierarchical modeling and analysis of spatial data, in order to look for new possible directions on the processing of count data, in a spatial hierarchical Bayesian econometric context.

Keywords: spatial data analysis, spatial econometrics, Bayesian hierarchical models, count data

Procedia PDF Downloads 576
25507 Modifying Assessment Modes in the Science Classroom as a Solution to Examination Malpractice

Authors: Catherine Omole

Abstract:

Examination malpractice includes acts that temper with collecting accurate results during the conduct of an examination, thereby giving undue advantage to a student over his colleagues. Even though examination malpractice has been a lingering problem, examinations may not be easy to do away with completely as it is an important feedback tool in the learning process with several other functions e.g for the purpose of selection, placement, certification and promotion. Examination malpractice has created a lot of problems such as a relying on a weak work force based on false assessment results. The question is why is this problem still persisting, despite measures that have been taken to curb this ugly trend over the years? This opinion paper has identified modifications that could help relieve the student of the examination stress and thus increase the student’s effort towards effective learning and discourage examination malpractice in the long run.

Keywords: assessment, examination malpractice, learning, science classroom

Procedia PDF Downloads 243
25506 A NoSQL Based Approach for Real-Time Managing of Robotics's Data

Authors: Gueidi Afef, Gharsellaoui Hamza, Ben Ahmed Samir

Abstract:

This paper deals with the secret of the continual progression data that new data management solutions have been emerged: The NoSQL databases. They crossed several areas like personalization, profile management, big data in real-time, content management, catalog, view of customers, mobile applications, internet of things, digital communication and fraud detection. Nowadays, these database management systems are increasing. These systems store data very well and with the trend of big data, a new challenge’s store demands new structures and methods for managing enterprise data. The new intelligent machine in the e-learning sector, thrives on more data, so smart machines can learn more and faster. The robotics are our use case to focus on our test. The implementation of NoSQL for Robotics wrestle all the data they acquire into usable form because with the ordinary type of robotics; we are facing very big limits to manage and find the exact information in real-time. Our original proposed approach was demonstrated by experimental studies and running example used as a use case.

Keywords: NoSQL databases, database management systems, robotics, big data

Procedia PDF Downloads 334
25505 Innovation Culture TV “Stars of Science”: 15 Seasons Case Study

Authors: Fouad Mrad, Viviane Zaccour

Abstract:

The accelerated developments in the political, economic, environmental, security, health, and social folders are exhausting planners across the world, especially in Arab countries. The impact of the tension is multifaceted and has resulted in conflicts, wars, migration, and human insecurity. The potential cross-cutting role that science, innovation and technology can play in supporting Arab societies to address these pressing challenges is a serious, unique chance for the people of the region. This opportunity is based on the existing capacity of educated youth and inaccessible talents in the local universities and research centers. It has been accepted that Arab countries have achieved major advancements in the economy, education and social wellbeing since the 70s of the 20th Century. Mainly direct outcome of the oil and other natural resources. The UN Secretary-General, during the Education Summit in Sep 2022, stressed that “Learning continues to underplay skills, including problem-solving, critical thinking and empathy.” Stars of Science by Qatar Foundation was launched in 2009 and has been sustained through 2023. Consistent mission from the start: To mobilize a new generation of Pan-Arab innovators and problem solvers by encouraging youth participation and interest in Science, Technology and Entrepreneurship throughout the Arab world via the program and its social media activities. To make science accessible and attractive to mass audiences by de-mystifying the process of innovation. Harnessing best practices within reality TV to show that science, engineering, and innovation are important in everyday life and can be fun.” Thousands of Participants learned unforgettable lessons; winners changed their lives forever as they learned and earned seed capital; they became drivers of change in their countries and families; millions of viewers were exposed to an innovative experimental process, and culturally, several relevant national institutions adopted the SOS track in their national initiatives. The program exhibited experientially youth self-efficacy as the most distinct core property of human agency, which is an individual's belief in his or her capacity to execute behaviors necessary to produce specific performance attainments. In addition, the program proved that innovations are performed by networks of people with different sets of technological, useful knowledge, skills and competencies introduced by socially shared technological knowledge as a main determinant of economic activities in any economy.

Keywords: science, invention, innovation, Qatar foundation, QSTP, prototyping

Procedia PDF Downloads 65
25504 Fuzzy Optimization Multi-Objective Clustering Ensemble Model for Multi-Source Data Analysis

Authors: C. B. Le, V. N. Pham

Abstract:

In modern data analysis, multi-source data appears more and more in real applications. Multi-source data clustering has emerged as a important issue in the data mining and machine learning community. Different data sources provide information about different data. Therefore, multi-source data linking is essential to improve clustering performance. However, in practice multi-source data is often heterogeneous, uncertain, and large. This issue is considered a major challenge from multi-source data. Ensemble is a versatile machine learning model in which learning techniques can work in parallel, with big data. Clustering ensemble has been shown to outperform any standard clustering algorithm in terms of accuracy and robustness. However, most of the traditional clustering ensemble approaches are based on single-objective function and single-source data. This paper proposes a new clustering ensemble method for multi-source data analysis. The fuzzy optimized multi-objective clustering ensemble method is called FOMOCE. Firstly, a clustering ensemble mathematical model based on the structure of multi-objective clustering function, multi-source data, and dark knowledge is introduced. Then, rules for extracting dark knowledge from the input data, clustering algorithms, and base clusterings are designed and applied. Finally, a clustering ensemble algorithm is proposed for multi-source data analysis. The experiments were performed on the standard sample data set. The experimental results demonstrate the superior performance of the FOMOCE method compared to the existing clustering ensemble methods and multi-source clustering methods.

Keywords: clustering ensemble, multi-source, multi-objective, fuzzy clustering

Procedia PDF Downloads 169
25503 A Study to Explore the Views of Students regarding E-Learning as an Instructional Tool at University Level

Authors: Zafar Iqbal

Abstract:

This study involved students of 6th semester enrolled in a Bachelor of Computer Science Program at university level. In this era of science and technology, e-learning can be helpful for grassroots in providing them access to education tenant in less developed areas. It is a potential substitute of face-to-face teaching being used in different countries. The purpose of the study was to explore the views of students about e-learning (Facebook) as an instructional tool. By using purposive sampling technique an intact class of 30 students included both male and female were selected where e-learning was used as an instructional tool. The views of students were explored through qualitative approach by using focus group interviews. The approach was helpful to develop comprehensive understanding of students’ views towards e- learning. In addition, probing questions were also asked and recorded. Data was transcribed, generated nodes and then coded text against these nodes. For this purpose and further analysis, NVivo 10 software was used. Themes were generated and tangibly presented through cluster analysis. Findings were interesting and provide sufficient evidence that face book is a subsequent e-learning source for students of higher education. Students acknowledged it as best source of learning and it was aligned with their academic and social behavior. It was not time specific and therefore, feasible for students who work day time and can get on line access to the material when they got free time. There were some distracters (time wasters) reported by the students but can be minimized by little effort. In short, e-learning is need of the day and potential learning source for every individual who have access to internet living at any part of the globe.

Keywords: e-learning, facebook, instructional tool, higher education

Procedia PDF Downloads 363
25502 Modeling Activity Pattern Using XGBoost for Mining Smart Card Data

Authors: Eui-Jin Kim, Hasik Lee, Su-Jin Park, Dong-Kyu Kim

Abstract:

Smart-card data are expected to provide information on activity pattern as an alternative to conventional person trip surveys. The focus of this study is to propose a method for training the person trip surveys to supplement the smart-card data that does not contain the purpose of each trip. We selected only available features from smart card data such as spatiotemporal information on the trip and geographic information system (GIS) data near the stations to train the survey data. XGboost, which is state-of-the-art tree-based ensemble classifier, was used to train data from multiple sources. This classifier uses a more regularized model formalization to control the over-fitting and show very fast execution time with well-performance. The validation results showed that proposed method efficiently estimated the trip purpose. GIS data of station and duration of stay at the destination were significant features in modeling trip purpose.

Keywords: activity pattern, data fusion, smart-card, XGboost

Procedia PDF Downloads 228
25501 Managing Fake News for Sustainable Democracy in Enugu State, Nigeria

Authors: Gloria Ebere Amadi, Emeka Promise Ugwunwotti

Abstract:

The study was carried out to determine the strategies for managing fake news for sustainable democracy in Enugu State, Nigeria. Two research questions and two null hypotheses guided the study. A survey research design was used for the study. The population for the study consisted of 100 respondents (from Enugu state House of Assembly). Of the entire population, 24 elected law makers and 76 staff were used; hence there was no sampling since the population was manageable. A 28-item structured questionnaire developed by the researcher was used for data collection. The instrument entitled Managing Fake News Questionnaire (MFNQ) was validated by three experts, two from the Department of Computer Science and one from the Department of Maths and Statistics, all from Enugu State University of Science and Technology. Cronbach Alpha was used to determine the reliability coefficient of the two sections of the instrument, and they are 0.67 and 0.82, while the reliability coefficient of the whole instrument gave a value of 0.81. Mean with standard deviation was used to answer research questions, while the null hypotheses at 0.5 level of significance at 98 degrees of freedom were tested with a t-test. The findings of the study revealed that the respondents agreed that government and citizens-related strategies improve the management of fake news for sustainable democracy in Enugu State. Again, there was no significant difference between the mean response of the lawmakers and staff on government and citizens-related strategies for managing fake news for sustainable democracy in Enugu State. Based on the findings, it was recommended, among others, that there should be regular workshops on the management of fake news for citizens.

Keywords: fake news, sustainability, democracy, management

Procedia PDF Downloads 55