Search results for: ArcGIS data analysis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 42046

Search results for: ArcGIS data analysis

40696 A Review on 3D Smart City Platforms Using Remotely Sensed Data to Aid Simulation and Urban Analysis

Authors: Slim Namouchi, Bruno Vallet, Imed Riadh Farah

Abstract:

3D urban models provide powerful tools for decision making, urban planning, and smart city services. The accuracy of this 3D based systems is directly related to the quality of these models. Since manual large-scale modeling, such as cities or countries is highly time intensive and very expensive process, a fully automatic 3D building generation is needed. However, 3D modeling process result depends on the input data, the proprieties of the captured objects, and the required characteristics of the reconstructed 3D model. Nowadays, producing 3D real-world model is no longer a problem. Remotely sensed data had experienced a remarkable increase in the recent years, especially data acquired using unmanned aerial vehicles (UAV). While the scanning techniques are developing, the captured data amount and the resolution are getting bigger and more precise. This paper presents a literature review, which aims to identify different methods of automatic 3D buildings extractions either from LiDAR or the combination of LiDAR and satellite or aerial images. Then, we present open source technologies, and data models (e.g., CityGML, PostGIS, Cesiumjs) used to integrate these models in geospatial base layers for smart city services.

Keywords: CityGML, LiDAR, remote sensing, SIG, Smart City, 3D urban modeling

Procedia PDF Downloads 134
40695 The Relations between Seismic Results and Groundwater near the Gokpinar Damp Area, Denizli, Turkey

Authors: Mahmud Gungor, Ali Aydin, Erdal Akyol, Suat Tasdelen

Abstract:

The understanding of geotechnical characteristics of near-surface material and the effects of the groundwater is very important problem in such as site studies. For showing the relations between seismic data and groundwater we selected about 25 km2 as the study area. It has been presented which is a detailed work of seismic data and groundwater depths of Gokpinar Damp area. Seismic waves velocity (Vp and Vs) are very important parameters showing the soil properties. The seismic records were used the method of the multichannel analysis of surface waves near area of Gokpinar Damp area. Sixty sites in this area have been investigated with survey lines about 60 m in length. MASW (Multichannel analysis of surface wave) method has been used to generate one-dimensional shear wave velocity profile at locations. These shear wave velocities are used to estimate equivalent shear wave velocity in the study area at every 2 and 5 m intervals up to a depth of 45 m. Levels of equivalent shear wave velocity of soil are used the classified of the study area. After the results of the study, it must be considered as components of urban planning and building design of Gokpinar Damp area, Denizli and the application and use of these results should be required and enforced by municipal authorities.

Keywords: seismic data, Gokpinar Damp, urban planning, Denizli

Procedia PDF Downloads 286
40694 Storage System Validation Study for Raw Cocoa Beans Using Minitab® 17 and R (R-3.3.1)

Authors: Anthony Oppong Kyekyeku, Sussana Antwi-Boasiako, Emmanuel De-Graft Johnson Owusu Ansah

Abstract:

In this observational study, the performance of a known conventional storage system was tested and evaluated for fitness for its intended purpose. The system has a scope extended for the storage of dry cocoa beans. System sensitivity, reproducibility and uncertainties are not known in details. This study discusses the system performance in the context of existing literature on factors that influence the quality of cocoa beans during storage. Controlled conditions were defined precisely for the system to give reliable base line within specific established procedures. Minitab® 17 and R statistical software (R-3.3.1) were used for the statistical analyses. The approach to the storage system testing was to observe and compare through laboratory test methods the quality of the cocoa beans samples before and after storage. The samples were kept in Kilner jars and the temperature of the storage environment controlled and monitored over a period of 408 days. Standard test methods use in international trade of cocoa such as the cut test analysis, moisture determination with Aqua boy KAM III model and bean count determination were used for quality assessment. The data analysis assumed the entire population as a sample in order to establish a reliable baseline to the data collected. The study concluded a statistically significant mean value at 95% Confidence Interval (CI) for the performance data analysed before and after storage for all variables observed. Correlational graphs showed a strong positive correlation for all variables investigated with the exception of All Other Defect (AOD). The weak relationship between the before and after data for AOD had an explained variability of 51.8% with the unexplained variability attributable to the uncontrolled condition of hidden infestation before storage. The current study concluded with a high-performance criterion for the storage system.

Keywords: benchmarking performance data, cocoa beans, hidden infestation, storage system validation

Procedia PDF Downloads 173
40693 A Case Comparative Study of Infant Mortality Rate in North-West Nigeria

Authors: G. I. Onwuka, A. Danbaba, S. U. Gulumbe

Abstract:

This study investigated of Infant Mortality Rate as observed at a general hospital in Kaduna-South, Kaduna State, North West Nigeria. The causes of infant Mortality were examined. The data used for this analysis were collected at the statistics unit of the Hospital. The analysis was carried out on the data using Multiple Linear regression Technique and this showed that there is linear relationship between the dependent variable (death) and the independent variables (malaria, measles, anaemia, and coronary heart disease). The resultant model also revealed that a unit increment in each of these diseases would result to a unit increment in death recorded, 98.7% of the total variation in mortality is explained by the given model. The highest number of mortality was recorded in July, 2005 and the lowest mortality recorded in October, 2009.Recommendations were however made based on the results of the study.

Keywords: infant mortality rate, multiple linear regression, diseases, serial correlation

Procedia PDF Downloads 326
40692 Data Access, AI Intensity, and Scale Advantages

Authors: Chuping Lo

Abstract:

This paper presents a simple model demonstrating that ceteris paribus countries with lower barriers to accessing global data tend to earn higher incomes than other countries. Therefore, large countries that inherently have greater data resources tend to have higher incomes than smaller countries, such that the former may be more hesitant than the latter to liberalize cross-border data flows to maintain this advantage. Furthermore, countries with higher artificial intelligence (AI) intensity in production technologies tend to benefit more from economies of scale in data aggregation, leading to higher income and more trade as they are better able to utilize global data.

Keywords: digital intensity, digital divide, international trade, scale of economics

Procedia PDF Downloads 66
40691 Secured Transmission and Reserving Space in Images Before Encryption to Embed Data

Authors: G. R. Navaneesh, E. Nagarajan, C. H. Rajam Raju

Abstract:

Nowadays the multimedia data are used to store some secure information. All previous methods allocate a space in image for data embedding purpose after encryption. In this paper, we propose a novel method by reserving space in image with a boundary surrounded before encryption with a traditional RDH algorithm, which makes it easy for the data hider to reversibly embed data in the encrypted images. The proposed method can achieve real time performance, that is, data extraction and image recovery are free of any error. A secure transmission process is also discussed in this paper, which improves the efficiency by ten times compared to other processes as discussed.

Keywords: secure communication, reserving room before encryption, least significant bits, image encryption, reversible data hiding

Procedia PDF Downloads 410
40690 Identity Verification Using k-NN Classifiers and Autistic Genetic Data

Authors: Fuad M. Alkoot

Abstract:

DNA data have been used in forensics for decades. However, current research looks at using the DNA as a biometric identity verification modality. The goal is to improve the speed of identification. We aim at using gene data that was initially used for autism detection to find if and how accurate is this data for identification applications. Mainly our goal is to find if our data preprocessing technique yields data useful as a biometric identification tool. We experiment with using the nearest neighbor classifier to identify subjects. Results show that optimal classification rate is achieved when the test set is corrupted by normally distributed noise with zero mean and standard deviation of 1. The classification rate is close to optimal at higher noise standard deviation reaching 3. This shows that the data can be used for identity verification with high accuracy using a simple classifier such as the k-nearest neighbor (k-NN). 

Keywords: biometrics, genetic data, identity verification, k nearest neighbor

Procedia PDF Downloads 253
40689 Rejoinders to the Expression of Reprimand among Jordanian Youth: A Pragmatic Study

Authors: Nisreen Al-Khawaldeh

Abstract:

The study investigates the expressions voiced by Jordanian youth as rejoinders to the expressions of reprimands. It also explores the impact sociocultural variables exert on such types of rejoinders. To our best knowledge, this study is the first of its kind. Despite the significance and sensitivity of such type of communicative act, there is a scarcity of research on it, and it has not been investigated in the Jordanian context. Data collected from observation of naturally occurring data. Data have been qualitatively and quantitatively analyzed in light of the rapport management approach (RMA). The analysis revealed different types of rejoinders, among which was the expression of apology, admitting responsibility, and trying to manage and fix the situation were the most used strategies. Variation in the types of strategies was attributed to the influence of the sociocultural variables. Promising ideas were recommended for future research.

Keywords: gender, rejoinder to reprimand, Jordanian youth, rapport management approach

Procedia PDF Downloads 195
40688 Failure Statistics Analysis of China’s Spacecraft in Full-Life

Authors: Xin-Yan Ji

Abstract:

The historical failures data of the spacecraft is very useful to improve the spacecraft design and the test philosophies and reduce the spacecraft flight risk. A study of spacecraft failures data was performed, which is the most comprehensive statistics of spacecrafts in China. 2593 on-orbit failures data and 1298 ground data that occurred on 150 spacecraft launched from 2000 to 2016 were identified and collected, which covered the navigation satellites, communication satellites, remote sensing deep space exploration manned spaceflight platforms. In this paper, the failures were analyzed to compare different spacecraft subsystem and estimate their impact on the mission, then the development of spacecraft in China was evaluated from design, software, workmanship, management, parts, and materials. Finally, the lessons learned from the past years show that electrical and mechanical failures are responsible for the largest parts, and the key solution to reduce in-orbit failures is improving design technology, enough redundancy, adequate space environment protection measures, and adequate ground testing.

Keywords: spacecraft anomalies, anomalies mechanism, failure cause, spacecraft testing

Procedia PDF Downloads 115
40687 Prevalence of Visual Impairment among School Children in Ethiopia: A Systematic Review and Meta-Analysis

Authors: Merkineh Markos Lorato, Gedefaw Diress Alene

Abstract:

Introduction: Visual impairment is any condition of the eye or visual system that results in loss/reduction of visual functioning. It significantly influences the academic routine and social activities of children, and the effect is severe for low-income countries like Ethiopia. So, this study aimed to determine the pooled prevalence of visual impairment among school children in Ethiopia. Methods: Databases such as Medical Literature Analysis and Retrieval System Online, Excerpta Medica dataBASE, World Wide Web of Science, and Cochrane Library searched to retrieve eligible articles. In addition, Google Scholar and a reference list of the retrieved eligible articles were addressed. Studies that reported the prevalence of visual impairment were included to estimate the pooled prevalence. Data were extracted using a standardized data extraction format prepared in Microsoft Excel and analysis was held using STATA 11 statistical software. I² was used to assess the heterogeneity. Because of considerable heterogeneity, a random effect meta-analysis model was used to estimate the pooled prevalence of visual impairment among school children in Ethiopia. Results: The result of 9 eligible studies showed that the pooled prevalence of visual impairment among school children in Ethiopia was 7.01% (95% CI: 5.46, 8.56%). In the subgroup analysis, the highest prevalence was reported in South Nations Nationalities and Tigray region together (7.99%; 3.63, 12.35), while the lowest prevalence was reported in Addis Ababa (5.73%; 3.93, 7.53). Conclusion: The prevalence of visual impairment among school children is significantly high in Ethiopia. If it is not detected and intervened early, it will cause a lifetime threat to visually impaired school children, so that school vision screening program plan and its implementation may cure the life quality of future generations in Ethiopia.

Keywords: visual impairment, school children, Ethiopia, prevalence

Procedia PDF Downloads 35
40686 Globalization and Women's Social Identity in Iran: A Case Study of Educated Women in the 'World City' of Yazd

Authors: Mohammad Tefagh

Abstract:

The process of globalization has transformed many social and cultural phenomena and has entered the world into a new era and arena. This phenomenon has introduced new methods, ideas, and identity interactions to human beings and has caused great changes in individual and social identity. Women have also been affected by globalization. Globalization has made the presence of women more and more effective and has caused identity changes and changes in the dimensions of identity in them. The purpose of this study is to investigate the impact of globalization of culture on changes in the social identity of educated women in the global city of Yazd. This study will discuss identity change and identity reconstruction due to globalization. The method of this study is qualitative, and the research data is obtained through in-depth interviews with 15 Yazdi-educated women at the Ph.D. level. The method of data analysis is thematic analysis. Findings of the research show that educated Yazdi women have changed their identity due to new communication processes and globalization, including faster, easier, and cheaper communication with other women in the world near and far. Women's social identity has also changed in the face of elements of globalization in various dimensions such as national, gender, religious, and group identities. The analysis of the interviews revealed the confronting elements such as using new cultural goods and communication technologies, membership in social networks, and increasing awareness of environmental change.

Keywords: globalization, social identity, educated women, Yazd

Procedia PDF Downloads 331
40685 Analysis of Maternal Death Surveillance and Response: Causes and Contributing Factors in Addis Ababa, Ethiopia, 2022

Authors: Sisay Tiroro Salato

Abstract:

Background: Ethiopia has been implementing the maternal death surveillance and response system to provide real-time actionable information, including causes of death and contributing factors. Analysis of maternal mortality surveillance data was conducted to identify the causes and underlying factors in Addis Ababa, Ethiopia. Methods: We carried out a retrospective surveillance data analysis of 324 maternal deaths reported in Addis Ababa, Ethiopia, from 2017 to 2021. The data were extracted from the national maternal death surveillance and response database, including information from case investigation, verbal autopsy, and facility extraction forms. The data were analyzed by computing frequency and presented in numbers, proportions, and ratios. Results: Of 324 maternal deaths, 92% died in the health facilities, 6.2% in transit, and 1.5% at home. The mean age at death was 28 years, ranging from 17 to 45. The maternal mortality ratio per 100,000 live births was 77for the five years, ranging from 126 in 2017 to 21 in 2021. The direct and indirect causes of death were responsible for 87% and 13%, respectively. The direct causes included obstetric haemorrhage, hypertensive disorders in pregnancy, puerperal sepsis, embolism, obstructed labour, and abortion. The third delay (delay in receiving care after reaching health facilities) accounted for 57% of deaths, while the first delay (delay in deciding to seek health care) and the second delay (delay in reaching health facilities) and accounted for 34% and 24%, respectively. Late arrival to the referral facility, delayed management after admission, andnon-recognition of danger signs were underlying factors. Conclusion: Over 86% of maternal deaths were attributed by avoidable direct causes. The majority of women do try to reach health services when an emergency occurs, but the third delays present a major problem. Improving the quality of care at the healthcare facility level will help to reduce maternal death.

Keywords: maternal death, surveillance, delays, factors

Procedia PDF Downloads 110
40684 A Review on Intelligent Systems for Geoscience

Authors: R Palson Kennedy, P.Kiran Sai

Abstract:

This article introduces machine learning (ML) researchers to the hurdles that geoscience problems present, as well as the opportunities for improvement in both ML and geosciences. This article presents a review from the data life cycle perspective to meet that need. Numerous facets of geosciences present unique difficulties for the study of intelligent systems. Geosciences data is notoriously difficult to analyze since it is frequently unpredictable, intermittent, sparse, multi-resolution, and multi-scale. The first half addresses data science’s essential concepts and theoretical underpinnings, while the second section contains key themes and sharing experiences from current publications focused on each stage of the data life cycle. Finally, themes such as open science, smart data, and team science are considered.

Keywords: Data science, intelligent system, machine learning, big data, data life cycle, recent development, geo science

Procedia PDF Downloads 133
40683 Study on Pedestrian Street Reconstruction under Comfortable Continuous View: Take the Walking Streets of Zhengzhou City as an Example

Authors: Liu Mingxin

Abstract:

Streets act as the organizers of each image element on the urban spatial route, and the spatial continuity of urban streets is the basis for people to perceive the overall image of the city. This paper takes the walking space of Zhengzhou city as the research object, conducts investigation and analysis through questionnaire interviews, and selects typical walking space for in-depth study. Through the analysis of questionnaire data, the investigation and analysis of the current situation of walking space, and the analysis of pedestrian psychological behavior activities, the paper summarizes the construction suggestions of urban walking space continuity from the three aspects of the composition of walking street, the bottom interface and side interface, and the service facilities of walking space. The walking space is not only the traffic space but also the comfortable experience and the continuity of the space.

Keywords: walking space, spatial continuity, walking psychology, space reconstruction

Procedia PDF Downloads 44
40682 Assessing Project Performance through Work Sampling and Earned Value Analysis

Authors: Shobha Ramalingam

Abstract:

The majority of the infrastructure projects are affected by time overrun, resulting in project delays and subsequently cost overruns. Time overrun may vary from a few months to as high as five or more years, placing the project viability at risk. One of the probable reasons noted in the literature for this outcome in projects is due to poor productivity. Researchers contend that productivity in construction has only marginally increased over the years. While studies in the literature have extensively focused on time and cost parameters in projects, there are limited studies that integrate time and cost with productivity to assess project performance. To this end, a study was conducted to understand the project delay factors concerning cost, time and productivity. A case-study approach was adopted to collect rich data from a nuclear power plant project site for two months through observation, interviews and document review. The data were analyzed using three different approaches for a comprehensive understanding. Foremost, a root-cause analysis was performed on the data using Ishikawa’s fish-bone diagram technique to identify the various factors impacting the delay concerning time. Based on it, a questionnaire was designed and circulated to concerned executives, including project engineers and contractors to determine the frequency of occurrence of the delay, which was then compiled and presented to the management for a possible solution to mitigate. Second, a productivity analysis was performed on select activities, including rebar bending and concreting through a time-motion study to analyze product performance. Third, data on cost of construction for three years allowed analyzing the cost performance using earned value management technique. All three techniques allowed to systematically and comprehensively identify the key factors that deter project performance and productivity loss in the construction of the nuclear power plant project. The findings showed that improper planning and coordination between multiple trades, concurrent operations, improper workforce and material management, fatigue due to overtime were some of the key factors that led to delays and poor productivity. The findings are expected to act as a stepping stone for further research and have implications for practitioners.

Keywords: earned value analysis, time performance, project costs, project delays, construction productivity

Procedia PDF Downloads 95
40681 Customer Preference in the Textile Market: Fabric-Based Analysis

Authors: Francisca Margarita Ocran

Abstract:

Underwear, and more particularly bras and panties, are defined as intimate clothing. Strictly speaking, they enhance the place of women in the public or private satchel. Therefore, women's lingerie is a complex garment with a high involvement profile, motivating consumers to buy it not only by its functional utility but also by the multisensory experience it provides them. Customer behavior models are generally based on customer data mining, and each model is designed to answer questions at a specific time. Predicting the customer experience is uncertain and difficult. Thus, knowledge of consumers' tastes in lingerie deserves to be treated as an experiential product, where the dimensions of the experience motivating consumers to buy a lingerie product and to remain faithful to it must be analyzed in detail by the manufacturers and retailers to engage and retain consumers, which is why this research aims to identify the variables that push consumers to choose their lingerie product, based on an in-depth analysis of the types of fabrics used to make lingerie. The data used in this study comes from online purchases. Machine learning approach with the use of Python programming language and Pycaret gives us a precision of 86.34%, 85.98%, and 84.55% for the three algorithms to use concerning the preference of a buyer in front of a range of lingerie. Gradient Boosting, random forest, and K Neighbors were used in this study; they are very promising and rich in the classification of preference in the textile industry.

Keywords: consumer behavior, data mining, lingerie, machine learning, preference

Procedia PDF Downloads 88
40680 Prospective Teachers’ Metacognitive Awareness and Goal Orientation as Predictors of Academic Success

Authors: Gidado Lawal Likko

Abstract:

The study examined the relationship of achievement goals, metacognitive awareness and academic success among students of colleges of education in North Western Nigeria. The study was guided by three objectives. The first two were to find out whether students’ achievement goals and metacognitive awareness correlate with their academic success. 358 students comprising 242 males (67.6%) and 116 females (32.4%) were studied. Correlation survey was employed in the conduct of the study. The instruments used to collect data were students’ bio data form, achievement goals inventory (Roedel, Schraw and Plake, 1994), metacognitive awareness inventory (Schraw & Dennison, 1994) and students’ CGPA (NCCE minimum standard, 2013) was used as the index of academic success. Pearson Product Moment and regression analysis were the statistical techniques used to analyze the data. Results of the analysis indicated that students’ achievement goals (r=0.554, p=0.004) and metacognitive awareness (r= 0.67, p=0.001) positively correlated with their academic success. Similarly, significant relationship exists between achievement goals and metacognitive awareness (r=0.77, p=0.000). Part of the recommendations is the need for the management of all colleges of education to have educational interventions aimed at developing students’ metacognitive awareness which will foster purposeful self-regulation of their learning. This could be achieved by periodic assessment of students’ metacognitive awareness which will serve as feedback as they move from one educational level to another.

Keywords: academic success, goal orientation, metacognitive awareness, prospective teachers

Procedia PDF Downloads 225
40679 The Results of Longitudinal Water Quality Monitoring of the Brandywine River, Chester County, Pennsylvania by High School Students

Authors: Dina L. DiSantis

Abstract:

Strengthening a sense of responsibility while relating global sustainability concepts such as water quality and pollution to a local water system can be achieved by teaching students to conduct and interpret water quality monitoring tests. When students conduct their own research, they become better stewards of the environment. Providing outdoor learning and place-based opportunities for students helps connect them to the natural world. By conducting stream studies and collecting data, students are able to better understand how the natural environment is a place where everything is connected. Students have been collecting physical, chemical and biological data along the West and East Branches of the Brandywine River, in Pennsylvania for over ten years. The stream studies are part of the advanced placement environmental science and aquatic science courses that are offered as electives to juniors and seniors at the Downingtown High School West Campus in Downingtown, Pennsylvania. Physical data collected includes: temperature, turbidity, width, depth, velocity, and volume of flow or discharge. The chemical tests conducted are: dissolved oxygen, carbon dioxide, pH, nitrates, alkalinity and phosphates. Macroinvertebrates are collected with a kick net, identified and then released. Students collect the data from several locations while traveling by canoe. In the classroom, students prepare a water quality data analysis and interpretation report based on their collected data. The summary of the results from longitudinal water quality data collection by students, as well as the strengths and weaknesses of student data collection will be presented.

Keywords: place-based, student data collection, sustainability, water quality monitoring

Procedia PDF Downloads 155
40678 Machine Learning Strategies for Data Extraction from Unstructured Documents in Financial Services

Authors: Delphine Vendryes, Dushyanth Sekhar, Baojia Tong, Matthew Theisen, Chester Curme

Abstract:

Much of the data that inform the decisions of governments, corporations and individuals are harvested from unstructured documents. Data extraction is defined here as a process that turns non-machine-readable information into a machine-readable format that can be stored, for instance, in a database. In financial services, introducing more automation in data extraction pipelines is a major challenge. Information sought by financial data consumers is often buried within vast bodies of unstructured documents, which have historically required thorough manual extraction. Automated solutions provide faster access to non-machine-readable datasets, in a context where untimely information quickly becomes irrelevant. Data quality standards cannot be compromised, so automation requires high data integrity. This multifaceted task is broken down into smaller steps: ingestion, table parsing (detection and structure recognition), text analysis (entity detection and disambiguation), schema-based record extraction, user feedback incorporation. Selected intermediary steps are phrased as machine learning problems. Solutions leveraging cutting-edge approaches from the fields of computer vision (e.g. table detection) and natural language processing (e.g. entity detection and disambiguation) are proposed.

Keywords: computer vision, entity recognition, finance, information retrieval, machine learning, natural language processing

Procedia PDF Downloads 109
40677 De-Novo Structural Elucidation from Mass/NMR Spectra

Authors: Ismael Zamora, Elisabeth Ortega, Tatiana Radchenko, Guillem Plasencia

Abstract:

The structure elucidation based on Mass Spectra (MS) data of unknown substances is an unresolved problem that affects many different fields of application. The recent overview of software available for structure elucidation of small molecules has shown the demand for efficient computational tool that will be able to perform structure elucidation of unknown small molecules and peptides. We developed an algorithm for De-Novo fragment analysis based on MS data that proposes a set of scored and ranked structures that are compatible with the MS and MSMS spectra. Several different algorithms were developed depending on the initial set of fragments and the structure building processes. Also, in all cases, several scores for the final molecule ranking were computed. They were validated with small and middle databases (DB) with the eleven test set compounds. Similar results were obtained from any of the databases that contained the fragments of the expected compound. We presented an algorithm. Or De-Novo fragment analysis based on only mass spectrometry (MS) data only that proposed a set of scored/ranked structures that was validated on different types of databases and showed good results as proof of concept. Moreover, the solutions proposed by Mass Spectrometry were submitted to the prediction of NMR spectra in order to elucidate which of the proposed structures was compatible with the NMR spectra collected.

Keywords: De Novo, structure elucidation, mass spectrometry, NMR

Procedia PDF Downloads 294
40676 Implementation of a Low-Cost Instrumentation for an Open Cycle Wind Tunnel to Evaluate Pressure Coefficient

Authors: Cristian P. Topa, Esteban A. Valencia, Victor H. Hidalgo, Marco A. Martinez

Abstract:

Wind tunnel experiments for aerodynamic profiles display numerous advantages, such as: clean steady laminar flow, controlled environmental conditions, streamlines visualization, and real data acquisition. However, the experiment instrumentation usually is expensive, and hence, each test implies a incremented in design cost. The aim of this work is to select and implement a low-cost static pressure data acquisition system for a NACA 2412 airfoil in an open cycle wind tunnel. This work compares wind tunnel experiment with Computational Fluid Dynamics (CFD) simulation and parametric analysis. The experiment was evaluated at Reynolds of 1.65 e5, with increasing angles from -5° to 15°. The comparison between the approaches show good enough accuracy, between the experiment and CFD, additional parametric analysis results differ widely from the other methods, which complies with the lack of accuracy of the lateral approach due its simplicity.

Keywords: wind tunnel, low cost instrumentation, experimental testing, CFD simulation

Procedia PDF Downloads 175
40675 Enhancing Teachers’ Professional Development Programmes by the Implementation of Flipped Learning Instruction: A Qualitative Study

Authors: Badriah Algarni

Abstract:

The pedagogy of ‘flipped learning’ is a form of blended instruction which is gaining widespread attention throughout the world. However, there is a lack of research concerning teachers’ professional development (TPD) in teachers who use flipping. The aim of this study was, therefore, to identify teachers’ perspectives on their experience of flipped PD. The study used a qualitative approach. Purposive sampling recruited nineteen teachers who participated in semi-structured, in-depth interviews. Thematic analysis was used to analyse the interview data. Overall, the teachers reported feeling more confident in their knowledge and skills after participating in flipped TPD. The analysis of the interview data revealed five overarching themes:1) increased engagement with the content; 2) better use of resources; 3) a social, collaborative environment; 4) exchange of practices and experiences; and 5) valuable online activities. These findings can encourage educators, policymakers, and trainers to consider flipped TPD as a form of PD to promote the building of teachers’ knowledge and stimulate reflective practices to improve teaching and learning practices.

Keywords: engagement, flipped learning, teachers’ professional development, collaboration

Procedia PDF Downloads 94
40674 Film Therapy on Adolescent Body Image: A Pilot Study

Authors: Sonia David, Uma Warrier

Abstract:

Background: Film therapy is the use of commercial or non-commercial films to enhance healing for therapeutic purposes. Objectives: The mixed-method study aims to evaluate the effect of film-based counseling on body image dissatisfaction among adolescents to precisely ascertain the cause of the alteration in body image dissatisfaction due to the said intervention. Method: The one group pre-test post-test research design study using inferential statistics and thematic analysis is based on a pre-test post-test design conducted on 44 school-going adolescents between 13 and 17. The Body Shape Questionnaire (BSQ- 34) was used as a pre-test and post-test measure. The film-based counseling intervention model was used through individual counseling sessions. The analysis involved paired sample t-test used to examine the data quantitatively, and thematic analysis was used to evaluate qualitative data. Findings: The results indicated that there is a significant difference between the pre-test and post-test means. Since t(44)= 9.042 is significant at a 99% confidence level, it is ascertained that film-based counseling intervention reduces body image dissatisfaction. The five distinct themes from the thematic analysis are “acceptance, awareness, empowered to change, empathy, and reflective.” Novelty: The paper originally contributes to the repertoire of research on film therapy as a successful counseling intervention for addressing the challenges of body image dissatisfaction. This study also opens avenues for considering alteration of teaching pedagogy to include video-based learning in various subjects.

Keywords: body image dissatisfaction, adolescents, film-based counselling, film therapy, acceptance and commitment therapy

Procedia PDF Downloads 294
40673 Anxiety and Depression in Caregivers of Autistic Children

Authors: Mou Juliet Rebeiro, S. M. Abul Kalam Azad

Abstract:

This study was carried out to see the anxiety and depression in caregivers of autistic children. The objectives of the research were to assess depression and anxiety among caregivers of autistic children and to find out the experience of caregivers. For this purpose, the research was conducted on a sample of 39 caregivers of autistic children. Participants were taken from a special school. To collect data for this study each of the caregivers were administered questionnaire comprising scales to measure anxiety and depression and some responses of the participants were taken through interview based on a topic guide. Obtained quantitative data were analyzed by using statistical analysis and qualitative data were analyzed according to themes. Mean of the anxiety score (55.85) and depression score (108.33) is above the cutoff point. Results showed that anxiety and depression is clinically present in caregivers of autistic children. Most of the caregivers experienced behavior, emotional, cognitive and social problems of their child that is linked with anxiety and depression.

Keywords: anxiety, autism, caregiver, depression

Procedia PDF Downloads 301
40672 Classification of EEG Signals Based on Dynamic Connectivity Analysis

Authors: Zoran Šverko, Saša Vlahinić, Nino Stojković, Ivan Markovinović

Abstract:

In this article, the classification of target letters is performed using data from the EEG P300 Speller paradigm. Neural networks trained with the results of dynamic connectivity analysis between different brain regions are used for classification. Dynamic connectivity analysis is based on the adaptive window size and the imaginary part of the complex Pearson correlation coefficient. Brain dynamics are analysed using the relative intersection of confidence intervals for the imaginary component of the complex Pearson correlation coefficient method (RICI-imCPCC). The RICI-imCPCC method overcomes the shortcomings of currently used dynamical connectivity analysis methods, such as the low reliability and low temporal precision for short connectivity intervals encountered in constant sliding window analysis with wide window size and the high susceptibility to noise encountered in constant sliding window analysis with narrow window size. This method overcomes these shortcomings by dynamically adjusting the window size using the RICI rule. This method extracts information about brain connections for each time sample. Seventy percent of the extracted brain connectivity information is used for training and thirty percent for validation. Classification of the target word is also done and based on the same analysis method. As far as we know, through this research, we have shown for the first time that dynamic connectivity can be used as a parameter for classifying EEG signals.

Keywords: dynamic connectivity analysis, EEG, neural networks, Pearson correlation coefficients

Procedia PDF Downloads 213
40671 Evaluating Models Through Feature Selection Methods Using Data Driven Approach

Authors: Shital Patil, Surendra Bhosale

Abstract:

Cardiac diseases are the leading causes of mortality and morbidity in the world, from recent few decades accounting for a large number of deaths have emerged as the most life-threatening disorder globally. Machine learning and Artificial intelligence have been playing key role in predicting the heart diseases. A relevant set of feature can be very helpful in predicting the disease accurately. In this study, we proposed a comparative analysis of 4 different features selection methods and evaluated their performance with both raw (Unbalanced dataset) and sampled (Balanced) dataset. The publicly available Z-Alizadeh Sani dataset have been used for this study. Four feature selection methods: Data Analysis, minimum Redundancy maximum Relevance (mRMR), Recursive Feature Elimination (RFE), Chi-squared are used in this study. These methods are tested with 8 different classification models to get the best accuracy possible. Using balanced and unbalanced dataset, the study shows promising results in terms of various performance metrics in accurately predicting heart disease. Experimental results obtained by the proposed method with the raw data obtains maximum AUC of 100%, maximum F1 score of 94%, maximum Recall of 98%, maximum Precision of 93%. While with the balanced dataset obtained results are, maximum AUC of 100%, F1-score 95%, maximum Recall of 95%, maximum Precision of 97%.

Keywords: cardio vascular diseases, machine learning, feature selection, SMOTE

Procedia PDF Downloads 116
40670 Comparing Performance of Neural Network and Decision Tree in Prediction of Myocardial Infarction

Authors: Reza Safdari, Goli Arji, Robab Abdolkhani Maryam zahmatkeshan

Abstract:

Background and purpose: Cardiovascular diseases are among the most common diseases in all societies. The most important step in minimizing myocardial infarction and its complications is to minimize its risk factors. The amount of medical data is increasingly growing. Medical data mining has a great potential for transforming these data into information. Using data mining techniques to generate predictive models for identifying those at risk for reducing the effects of the disease is very helpful. The present study aimed to collect data related to risk factors of heart infarction from patients’ medical record and developed predicting models using data mining algorithm. Methods: The present work was an analytical study conducted on a database containing 350 records. Data were related to patients admitted to Shahid Rajaei specialized cardiovascular hospital, Iran, in 2011. Data were collected using a four-sectioned data collection form. Data analysis was performed using SPSS and Clementine version 12. Seven predictive algorithms and one algorithm-based model for predicting association rules were applied to the data. Accuracy, precision, sensitivity, specificity, as well as positive and negative predictive values were determined and the final model was obtained. Results: five parameters, including hypertension, DLP, tobacco smoking, diabetes, and A+ blood group, were the most critical risk factors of myocardial infarction. Among the models, the neural network model was found to have the highest sensitivity, indicating its ability to successfully diagnose the disease. Conclusion: Risk prediction models have great potentials in facilitating the management of a patient with a specific disease. Therefore, health interventions or change in their life style can be conducted based on these models for improving the health conditions of the individuals at risk.

Keywords: decision trees, neural network, myocardial infarction, Data Mining

Procedia PDF Downloads 429
40669 Analysis of Factors Affecting Public Awareness in Paying Zakat

Authors: Roikhan Mochamad Aziz

Abstract:

This study aims to analze the interdependence of several variables simultaneously in order to simplify the form of the relationship between some of the variables studied a number of factors less than the variable studied which means it can also describe the data structure of a research. Based 100 respondents from the public, such as the people of South Tangerang, this study used factor analysis tool. The results of this study indicate that the studied variables being formed into nine factors, namely faith factors, community factors, factors of social care, confidence factor, factor income, educational factors, self-satisfaction factors, factors work, and knowledge factor. Total variance of the 9 factors is 67,30% means that all nine of these factors are factors that can contribute too paying zakat of muzakki consciousness of 67,30% while the remaining 32,70% is supported by other factors outside the 9 factors.

Keywords: zakat, analysis factor, faith, education, knowledge

Procedia PDF Downloads 277
40668 Development of a Technology Assessment Model by Patents and Customers' Review Data

Authors: Kisik Song, Sungjoo Lee

Abstract:

Recent years have seen an increasing number of patent disputes due to excessive competition in the global market and a reduced technology life-cycle; this has increased the risk of investment in technology development. While many global companies have started developing a methodology to identify promising technologies and assess for decisions, the existing methodology still has some limitations. Post hoc assessments of the new technology are not being performed, especially to determine whether the suggested technologies turned out to be promising. For example, in existing quantitative patent analysis, a patent’s citation information has served as an important metric for quality assessment, but this analysis cannot be applied to recently registered patents because such information accumulates over time. Therefore, we propose a new technology assessment model that can replace citation information and positively affect technological development based on post hoc analysis of the patents for promising technologies. Additionally, we collect customer reviews on a target technology to extract keywords that show the customers’ needs, and we determine how many keywords are covered in the new technology. Finally, we construct a portfolio (based on a technology assessment from patent information) and a customer-based marketability assessment (based on review data), and we use them to visualize the characteristics of the new technologies.

Keywords: technology assessment, patents, citation information, opinion mining

Procedia PDF Downloads 465
40667 Data Quality as a Pillar of Data-Driven Organizations: Exploring the Benefits of Data Mesh

Authors: Marc Bachelet, Abhijit Kumar Chatterjee, José Manuel Avila

Abstract:

Data quality is a key component of any data-driven organization. Without data quality, organizations cannot effectively make data-driven decisions, which often leads to poor business performance. Therefore, it is important for an organization to ensure that the data they use is of high quality. This is where the concept of data mesh comes in. Data mesh is an organizational and architectural decentralized approach to data management that can help organizations improve the quality of data. The concept of data mesh was first introduced in 2020. Its purpose is to decentralize data ownership, making it easier for domain experts to manage the data. This can help organizations improve data quality by reducing the reliance on centralized data teams and allowing domain experts to take charge of their data. This paper intends to discuss how a set of elements, including data mesh, are tools capable of increasing data quality. One of the key benefits of data mesh is improved metadata management. In a traditional data architecture, metadata management is typically centralized, which can lead to data silos and poor data quality. With data mesh, metadata is managed in a decentralized manner, ensuring accurate and up-to-date metadata, thereby improving data quality. Another benefit of data mesh is the clarification of roles and responsibilities. In a traditional data architecture, data teams are responsible for managing all aspects of data, which can lead to confusion and ambiguity in responsibilities. With data mesh, domain experts are responsible for managing their own data, which can help provide clarity in roles and responsibilities and improve data quality. Additionally, data mesh can also contribute to a new form of organization that is more agile and adaptable. By decentralizing data ownership, organizations can respond more quickly to changes in their business environment, which in turn can help improve overall performance by allowing better insights into business as an effect of better reports and visualization tools. Monitoring and analytics are also important aspects of data quality. With data mesh, monitoring, and analytics are decentralized, allowing domain experts to monitor and analyze their own data. This will help in identifying and addressing data quality problems in quick time, leading to improved data quality. Data culture is another major aspect of data quality. With data mesh, domain experts are encouraged to take ownership of their data, which can help create a data-driven culture within the organization. This can lead to improved data quality and better business outcomes. Finally, the paper explores the contribution of AI in the coming years. AI can help enhance data quality by automating many data-related tasks, like data cleaning and data validation. By integrating AI into data mesh, organizations can further enhance the quality of their data. The concepts mentioned above are illustrated by AEKIDEN experience feedback. AEKIDEN is an international data-driven consultancy that has successfully implemented a data mesh approach. By sharing their experience, AEKIDEN can help other organizations understand the benefits and challenges of implementing data mesh and improving data quality.

Keywords: data culture, data-driven organization, data mesh, data quality for business success

Procedia PDF Downloads 133