Search results for: incomplete data extrapolation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24471

Search results for: incomplete data extrapolation

24171 Synthesis of Balanced 3-RRR Planar Parallel Manipulators

Authors: Arakelian Vigen, Geng Jing, Le Baron Jean-Paul

Abstract:

The paper deals with the design of parallel manipulators with balanced inertia forces and moments. The balancing of the resultant of the inertia forces of 3-RRR planar parallel manipulators is carried out through mass redistribution and centre of mass acceleration minimization. The proposed balancing technique is achieved in two steps: at first, optimal redistribution of the masses of input links is accomplished, which ensures the similarity of the end-effector trajectory and the manipulator’s common centre of mass trajectory, then, optimal trajectory planning of the end-effector by 'bang-bang' profile is reached. In such a way, the minimization of the magnitude of the acceleration of the centre of mass of the manipulator brings about a minimization of shaking force. To minimize the resultant of the inertia moments (shaking moment), the active balancing via inertia flywheel is applied. However, in this case, the active balancing is quite different from previous applications because it provides only a partial cancellation of the shaking moment due to the incomplete balancing of shaking force.

Keywords: dynamic balancing, inertia force minimization, inertia moment minimization, 3-RRR planar parallel manipulator

Procedia PDF Downloads 435
24170 Machine Learning Analysis of Student Success in Introductory Calculus Based Physics I Course

Authors: Chandra Prayaga, Aaron Wade, Lakshmi Prayaga, Gopi Shankar Mallu

Abstract:

This paper presents the use of machine learning algorithms to predict the success of students in an introductory physics course. Data having 140 rows pertaining to the performance of two batches of students was used. The lack of sufficient data to train robust machine learning models was compensated for by generating synthetic data similar to the real data. CTGAN and CTGAN with Gaussian Copula (Gaussian) were used to generate synthetic data, with the real data as input. To check the similarity between the real data and each synthetic dataset, pair plots were made. The synthetic data was used to train machine learning models using the PyCaret package. For the CTGAN data, the Ada Boost Classifier (ADA) was found to be the ML model with the best fit, whereas the CTGAN with Gaussian Copula yielded Logistic Regression (LR) as the best model. Both models were then tested for accuracy with the real data. ROC-AUC analysis was performed for all the ten classes of the target variable (Grades A, A-, B+, B, B-, C+, C, C-, D, F). The ADA model with CTGAN data showed a mean AUC score of 0.4377, but the LR model with the Gaussian data showed a mean AUC score of 0.6149. ROC-AUC plots were obtained for each Grade value separately. The LR model with Gaussian data showed consistently better AUC scores compared to the ADA model with CTGAN data, except in two cases of the Grade value, C- and A-.

Keywords: machine learning, student success, physics course, grades, synthetic data, CTGAN, gaussian copula CTGAN

Procedia PDF Downloads 22
24169 Data Access, AI Intensity, and Scale Advantages

Authors: Chuping Lo

Abstract:

This paper presents a simple model demonstrating that ceteris paribus countries with lower barriers to accessing global data tend to earn higher incomes than other countries. Therefore, large countries that inherently have greater data resources tend to have higher incomes than smaller countries, such that the former may be more hesitant than the latter to liberalize cross-border data flows to maintain this advantage. Furthermore, countries with higher artificial intelligence (AI) intensity in production technologies tend to benefit more from economies of scale in data aggregation, leading to higher income and more trade as they are better able to utilize global data.

Keywords: digital intensity, digital divide, international trade, scale of economics

Procedia PDF Downloads 42
24168 Secured Transmission and Reserving Space in Images Before Encryption to Embed Data

Authors: G. R. Navaneesh, E. Nagarajan, C. H. Rajam Raju

Abstract:

Nowadays the multimedia data are used to store some secure information. All previous methods allocate a space in image for data embedding purpose after encryption. In this paper, we propose a novel method by reserving space in image with a boundary surrounded before encryption with a traditional RDH algorithm, which makes it easy for the data hider to reversibly embed data in the encrypted images. The proposed method can achieve real time performance, that is, data extraction and image recovery are free of any error. A secure transmission process is also discussed in this paper, which improves the efficiency by ten times compared to other processes as discussed.

Keywords: secure communication, reserving room before encryption, least significant bits, image encryption, reversible data hiding

Procedia PDF Downloads 381
24167 Identity Verification Using k-NN Classifiers and Autistic Genetic Data

Authors: Fuad M. Alkoot

Abstract:

DNA data have been used in forensics for decades. However, current research looks at using the DNA as a biometric identity verification modality. The goal is to improve the speed of identification. We aim at using gene data that was initially used for autism detection to find if and how accurate is this data for identification applications. Mainly our goal is to find if our data preprocessing technique yields data useful as a biometric identification tool. We experiment with using the nearest neighbor classifier to identify subjects. Results show that optimal classification rate is achieved when the test set is corrupted by normally distributed noise with zero mean and standard deviation of 1. The classification rate is close to optimal at higher noise standard deviation reaching 3. This shows that the data can be used for identity verification with high accuracy using a simple classifier such as the k-nearest neighbor (k-NN). 

Keywords: biometrics, genetic data, identity verification, k nearest neighbor

Procedia PDF Downloads 228
24166 Insect Diversity Potential in Olive Trees in Two Orchards Differently Managed Under an Arid Climate in the Western Steppe Land, Algeria

Authors: Samir Ali-arous, Mohamed Beddane, Khaled Djelouah

Abstract:

This study investigated the insect diversity of olive (Olea europaea Linnaeus (Oleaceae)) groves grown in an arid climate in Algeria. In this context, several sampling methods were used within two orchards differently managed. Fifty arthropod species belonging to diverse orders and families were recorded. Hymenopteran species were quantitatively the most abundant, followed by species associated with Heteroptera, Aranea, Coleoptera and Homoptera orders. Regarding functional feeding groups, phytophagous species were dominant in the weeded and the unweeded orchard; however, higher abundance was recorded in the weeded site. Predators were ranked second, and pollinators were more frequent in the unweeded olive orchard. Two-factor Anova with repeated measures had revealed high significant effect of the weed management system, measures repetition and interaction with measurement repetition on arthropod’s abundances (P < 0.05). Likewise, generalized linear models showed that N/S ratio varied significantly between the two weed management approaches, in contrast, the remaining diversity indices including the Shannon index H’ had no significant correlation. Moreover, diversity parameters of arthropod’s communities in each agro-system highlighted multiples significant correlations (P <0.05). Rarefaction and extrapolation (R/E) sampling curves, evidenced that the survey and monitoring carried out in both sites had a optimum coverage of entomofauna present including scarce and transient species. Overall, calculated diversity and similarity indices were greater in the unweeded orchard than in the weeded orchard, demonstrating spontaneous flora's key role in entomofaunal diversity. Principal Component Analysis (PCA) has defined correlations between arthropod’s abundances and naturally occurring plants in olive orchards, including beneficials.

Keywords: Algeria, olive, insects, diversity, wild plants

Procedia PDF Downloads 43
24165 Hybrid Method for Smart Suggestions in Conversations for Online Marketplaces

Authors: Yasamin Rahimi, Ali Kamandi, Abbas Hoseini, Hesam Haddad

Abstract:

Online/offline chat is a convenient approach in the electronic markets of second-hand products in which potential customers would like to have more information about the products to fill the information gap between buyers and sellers. Online peer in peer market is trying to create artificial intelligence-based systems that help customers ask more informative questions in an easier way. In this article, we introduce a method for the question/answer system that we have developed for the top-ranked electronic market in Iran called Divar. When it comes to secondhand products, incomplete product information in a purchase will result in loss to the buyer. One way to balance buyer and seller information of a product is to help the buyer ask more informative questions when purchasing. Also, the short time to start and achieve the desired result of the conversation was one of our main goals, which was achieved according to A/B tests results. In this paper, we propose and evaluate a method for suggesting questions and answers in the messaging platform of the e-commerce website Divar. Creating such systems is to help users gather knowledge about the product easier and faster, All from the Divar database. We collected a dataset of around 2 million messages in Persian colloquial language, and for each category of product, we gathered 500K messages, of which only 2K were Tagged, and semi-supervised methods were used. In order to publish the proposed model to production, it is required to be fast enough to process 10 million messages daily on CPU processors. In order to reach that speed, in many subtasks, faster and simplistic models are preferred over deep neural models. The proposed method, which requires only a small amount of labeled data, is currently used in Divar production on CPU processors, and 15% of buyers and seller’s messages in conversations is directly chosen from our model output, and more than 27% of buyers have used this model suggestions in at least one daily conversation.

Keywords: smart reply, spell checker, information retrieval, intent detection, question answering

Procedia PDF Downloads 159
24164 A Review on Intelligent Systems for Geoscience

Authors: R Palson Kennedy, P.Kiran Sai

Abstract:

This article introduces machine learning (ML) researchers to the hurdles that geoscience problems present, as well as the opportunities for improvement in both ML and geosciences. This article presents a review from the data life cycle perspective to meet that need. Numerous facets of geosciences present unique difficulties for the study of intelligent systems. Geosciences data is notoriously difficult to analyze since it is frequently unpredictable, intermittent, sparse, multi-resolution, and multi-scale. The first half addresses data science’s essential concepts and theoretical underpinnings, while the second section contains key themes and sharing experiences from current publications focused on each stage of the data life cycle. Finally, themes such as open science, smart data, and team science are considered.

Keywords: Data science, intelligent system, machine learning, big data, data life cycle, recent development, geo science

Procedia PDF Downloads 117
24163 Data Quality as a Pillar of Data-Driven Organizations: Exploring the Benefits of Data Mesh

Authors: Marc Bachelet, Abhijit Kumar Chatterjee, José Manuel Avila

Abstract:

Data quality is a key component of any data-driven organization. Without data quality, organizations cannot effectively make data-driven decisions, which often leads to poor business performance. Therefore, it is important for an organization to ensure that the data they use is of high quality. This is where the concept of data mesh comes in. Data mesh is an organizational and architectural decentralized approach to data management that can help organizations improve the quality of data. The concept of data mesh was first introduced in 2020. Its purpose is to decentralize data ownership, making it easier for domain experts to manage the data. This can help organizations improve data quality by reducing the reliance on centralized data teams and allowing domain experts to take charge of their data. This paper intends to discuss how a set of elements, including data mesh, are tools capable of increasing data quality. One of the key benefits of data mesh is improved metadata management. In a traditional data architecture, metadata management is typically centralized, which can lead to data silos and poor data quality. With data mesh, metadata is managed in a decentralized manner, ensuring accurate and up-to-date metadata, thereby improving data quality. Another benefit of data mesh is the clarification of roles and responsibilities. In a traditional data architecture, data teams are responsible for managing all aspects of data, which can lead to confusion and ambiguity in responsibilities. With data mesh, domain experts are responsible for managing their own data, which can help provide clarity in roles and responsibilities and improve data quality. Additionally, data mesh can also contribute to a new form of organization that is more agile and adaptable. By decentralizing data ownership, organizations can respond more quickly to changes in their business environment, which in turn can help improve overall performance by allowing better insights into business as an effect of better reports and visualization tools. Monitoring and analytics are also important aspects of data quality. With data mesh, monitoring, and analytics are decentralized, allowing domain experts to monitor and analyze their own data. This will help in identifying and addressing data quality problems in quick time, leading to improved data quality. Data culture is another major aspect of data quality. With data mesh, domain experts are encouraged to take ownership of their data, which can help create a data-driven culture within the organization. This can lead to improved data quality and better business outcomes. Finally, the paper explores the contribution of AI in the coming years. AI can help enhance data quality by automating many data-related tasks, like data cleaning and data validation. By integrating AI into data mesh, organizations can further enhance the quality of their data. The concepts mentioned above are illustrated by AEKIDEN experience feedback. AEKIDEN is an international data-driven consultancy that has successfully implemented a data mesh approach. By sharing their experience, AEKIDEN can help other organizations understand the benefits and challenges of implementing data mesh and improving data quality.

Keywords: data culture, data-driven organization, data mesh, data quality for business success

Procedia PDF Downloads 103
24162 Porous Titanium Scaffolds Fabricated by Metal Injection Moulding Using Potassium-Chloride and Space Holder

Authors: Ali Dehghan Manshadi, David H. StJohn, Matthew S. Dargusch, M. Qian

Abstract:

Biocompatible, highly porous titanium scaffolds were manufactured by metal injection moulding of spherical titanium powder (powder size: -45 µm) with potassium chloride (powder size: -250 µm) as a space holder. Property evaluation of scaffolds confirmed a high level of compatibility between their mechanical properties and those of human cortical bone. The optimum sintering temperature was found to be 1250°C producing scaffolds with more than 90% interconnected pores in the size range of 200-250 µm, yield stress of 220 MPa and Young’s modulus of 7.80 GPa, all of which are suitable for bone tissue engineering. Increasing the sintering temperature to 1300°C increased the Young’s modulus to 22.0 GPa while reducing the temperature to 1150°C reduced the yield stress to 120 MPa due to incomplete sintering. The residual potassium chloride was determined vs. sintering temperature. A comparison was also made between the porous titanium scaffolds fabricated in this study and the additively manufactured titanium lattices of similar porosity reported in the literature.

Keywords: titanium, metal injection moulding, mechanical properties, scaffolds

Procedia PDF Downloads 185
24161 Big Data Analysis with RHadoop

Authors: Ji Eun Shin, Byung Ho Jung, Dong Hoon Lim

Abstract:

It is almost impossible to store or analyze big data increasing exponentially with traditional technologies. Hadoop is a new technology to make that possible. R programming language is by far the most popular statistical tool for big data analysis based on distributed processing with Hadoop technology. With RHadoop that integrates R and Hadoop environment, we implemented parallel multiple regression analysis with different sizes of actual data. Experimental results showed our RHadoop system was much faster as the number of data nodes increases. We also compared the performance of our RHadoop with lm function and big lm packages available on big memory. The results showed that our RHadoop was faster than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases.

Keywords: big data, Hadoop, parallel regression analysis, R, RHadoop

Procedia PDF Downloads 412
24160 A Mutually Exclusive Task Generation Method Based on Data Augmentation

Authors: Haojie Wang, Xun Li, Rui Yin

Abstract:

In order to solve the memorization overfitting in the meta-learning MAML algorithm, a method of generating mutually exclusive tasks based on data augmentation is proposed. This method generates a mutex task by corresponding one feature of the data to multiple labels, so that the generated mutex task is inconsistent with the data distribution in the initial dataset. Because generating mutex tasks for all data will produce a large number of invalid data and, in the worst case, lead to exponential growth of computation, this paper also proposes a key data extraction method, that only extracts part of the data to generate the mutex task. The experiments show that the method of generating mutually exclusive tasks can effectively solve the memorization overfitting in the meta-learning MAML algorithm.

Keywords: data augmentation, mutex task generation, meta-learning, text classification.

Procedia PDF Downloads 70
24159 Efficient Positioning of Data Aggregation Point for Wireless Sensor Network

Authors: Sifat Rahman Ahona, Rifat Tasnim, Naima Hassan

Abstract:

Data aggregation is a helpful technique for reducing the data communication overhead in wireless sensor network. One of the important tasks of data aggregation is positioning of the aggregator points. There are a lot of works done on data aggregation. But, efficient positioning of the aggregators points is not focused so much. In this paper, authors are focusing on the positioning or the placement of the aggregation points in wireless sensor network. Authors proposed an algorithm to select the aggregators positions for a scenario where aggregator nodes are more powerful than sensor nodes.

Keywords: aggregation point, data communication, data aggregation, wireless sensor network

Procedia PDF Downloads 132
24158 Spatial Econometric Approaches for Count Data: An Overview and New Directions

Authors: Paula Simões, Isabel Natário

Abstract:

This paper reviews a number of theoretical aspects for implementing an explicit spatial perspective in econometrics for modelling non-continuous data, in general, and count data, in particular. It provides an overview of the several spatial econometric approaches that are available to model data that are collected with reference to location in space, from the classical spatial econometrics approaches to the recent developments on spatial econometrics to model count data, in a Bayesian hierarchical setting. Considerable attention is paid to the inferential framework, necessary for structural consistent spatial econometric count models, incorporating spatial lag autocorrelation, to the corresponding estimation and testing procedures for different assumptions, to the constrains and implications embedded in the various specifications in the literature. This review combines insights from the classical spatial econometrics literature as well as from hierarchical modeling and analysis of spatial data, in order to look for new possible directions on the processing of count data, in a spatial hierarchical Bayesian econometric context.

Keywords: spatial data analysis, spatial econometrics, Bayesian hierarchical models, count data

Procedia PDF Downloads 563
24157 A NoSQL Based Approach for Real-Time Managing of Robotics's Data

Authors: Gueidi Afef, Gharsellaoui Hamza, Ben Ahmed Samir

Abstract:

This paper deals with the secret of the continual progression data that new data management solutions have been emerged: The NoSQL databases. They crossed several areas like personalization, profile management, big data in real-time, content management, catalog, view of customers, mobile applications, internet of things, digital communication and fraud detection. Nowadays, these database management systems are increasing. These systems store data very well and with the trend of big data, a new challenge’s store demands new structures and methods for managing enterprise data. The new intelligent machine in the e-learning sector, thrives on more data, so smart machines can learn more and faster. The robotics are our use case to focus on our test. The implementation of NoSQL for Robotics wrestle all the data they acquire into usable form because with the ordinary type of robotics; we are facing very big limits to manage and find the exact information in real-time. Our original proposed approach was demonstrated by experimental studies and running example used as a use case.

Keywords: NoSQL databases, database management systems, robotics, big data

Procedia PDF Downloads 325
24156 Management of Gastrointestinal Metastasis of Invasive Lobular Carcinoma

Authors: Sally Shepherd, Richard De Boer, Craig Murphy

Abstract:

Background: Invasive lobular carcinoma (ILC) can metastasize to atypical sites within the peritoneal cavity, gastrointestinal, or genitourinary tract. Management varies depending on the symptom presentation, extent of disease burden, particularly if the primary disease is occult, and patient wishes. Case Series: 6 patients presented with general surgical presentations of ILC, including incomplete large bowel obstruction, cholecystitis, persistent lower abdominal pain, and faecal incontinence. 3 were diagnosed with their primary and metastatic disease in the same presentation, whilst 3 patients developed metastasis from 5 to 8 years post primary diagnosis of ILC. Management included resection of the metastasis (laparoscopic cholecystectomy), excision of the primary (mastectomy and axillary clearance), followed by a combination of aromatase inhibitors, biologic therapy, and chemotherapy. Survival post diagnosis of metastasis ranged from 3 weeks to 7 years. Conclusion: Metastatic ILC must be considered with any gastrointestinal or genitourinary symptoms in patients with a current or past history of ILC. Management may not be straightforward to chemotherapy if the acute pathology is resulting in a surgically resectable disease.

Keywords: breast cancer, gastrointestinal metastasis, invasive lobular carcinoma, metastasis

Procedia PDF Downloads 120
24155 Fuzzy Optimization Multi-Objective Clustering Ensemble Model for Multi-Source Data Analysis

Authors: C. B. Le, V. N. Pham

Abstract:

In modern data analysis, multi-source data appears more and more in real applications. Multi-source data clustering has emerged as a important issue in the data mining and machine learning community. Different data sources provide information about different data. Therefore, multi-source data linking is essential to improve clustering performance. However, in practice multi-source data is often heterogeneous, uncertain, and large. This issue is considered a major challenge from multi-source data. Ensemble is a versatile machine learning model in which learning techniques can work in parallel, with big data. Clustering ensemble has been shown to outperform any standard clustering algorithm in terms of accuracy and robustness. However, most of the traditional clustering ensemble approaches are based on single-objective function and single-source data. This paper proposes a new clustering ensemble method for multi-source data analysis. The fuzzy optimized multi-objective clustering ensemble method is called FOMOCE. Firstly, a clustering ensemble mathematical model based on the structure of multi-objective clustering function, multi-source data, and dark knowledge is introduced. Then, rules for extracting dark knowledge from the input data, clustering algorithms, and base clusterings are designed and applied. Finally, a clustering ensemble algorithm is proposed for multi-source data analysis. The experiments were performed on the standard sample data set. The experimental results demonstrate the superior performance of the FOMOCE method compared to the existing clustering ensemble methods and multi-source clustering methods.

Keywords: clustering ensemble, multi-source, multi-objective, fuzzy clustering

Procedia PDF Downloads 151
24154 Strategic Decision Making Practice in Croatia: Which Decision Making Style is More Effective?

Authors: Ivana Bulog

Abstract:

Decision making is a vital part of the business world and any other field of human endeavor. Which way a business organization will take, and where that way will lead it, depends on broad range of decisions made by managers in the managerial structure. Strategic decisions are of the greatest importance for organizational success. Although much empirical research has been done trying to describe and explain its nature and effectiveness, knowledge about strategic decision making is still incomplete. This paper explores the nature of strategic decision making in particular setting - in Croatian companies. The main focus of this research is on the style that decision makers on strategic management level are following when making decisions of life importance for their companies. Two main decision making style that explain the way decision maker collects and processes available information and performs all the activities in strategic decision making process were empirical tested: rational and intuitive one. Besides analyzing their existence on strategic management level in Croatian companies, their effectiveness is analyzed as well. Results showed that decision makers at strategic management level are following both styles somewhat equally in order to function effectively, and that intuitive style is more effective when considering decisions outcomes.

Keywords: decision making style, decision making effectiveness, strategic decisions, management sciences

Procedia PDF Downloads 354
24153 Measuring the Full Impact of Culture: Social Indicators and Canadian Cultural Policy

Authors: Steven Wright

Abstract:

This paper argues that there is an opportunity for PCH to further expand its relevance within the Canadian policy context by taking advantage of the growing international trend of using social indicators for public policy evaluation. Within the mandate and vision of PCH, there is an incomplete understanding of the value that the arts and culture provide for Canadians, specifically with regard to four social indicators: community development, civic engagement, life satisfaction, and work-life balance. As will be shown, culture and the arts have a unique role to play in such quality of life indicators, and there is an opportunity for PCH to aid in the development of a comprehensive national framework that includes these indicators. This paper lays out approach to understanding how social indicators may be included in the Canadian context by first illustrating recent trends in policy evaluation on a national and international scale. From there, a theoretical analysis of the connection between cultural policy and social indicators is provided. The second half of the paper is dedicated to explaining the shortcomings of Canadian cultural policy evaluation in terms of its tendency to justify expenditures related to arts and cultural activities in purely economic terms, and surveying how other governments worldwide are leading the charge in this regard.

Keywords: social indicators, evaluation, cultural policy, arts

Procedia PDF Downloads 271
24152 Modeling Activity Pattern Using XGBoost for Mining Smart Card Data

Authors: Eui-Jin Kim, Hasik Lee, Su-Jin Park, Dong-Kyu Kim

Abstract:

Smart-card data are expected to provide information on activity pattern as an alternative to conventional person trip surveys. The focus of this study is to propose a method for training the person trip surveys to supplement the smart-card data that does not contain the purpose of each trip. We selected only available features from smart card data such as spatiotemporal information on the trip and geographic information system (GIS) data near the stations to train the survey data. XGboost, which is state-of-the-art tree-based ensemble classifier, was used to train data from multiple sources. This classifier uses a more regularized model formalization to control the over-fitting and show very fast execution time with well-performance. The validation results showed that proposed method efficiently estimated the trip purpose. GIS data of station and duration of stay at the destination were significant features in modeling trip purpose.

Keywords: activity pattern, data fusion, smart-card, XGboost

Procedia PDF Downloads 219
24151 A Mutually Exclusive Task Generation Method Based on Data Augmentation

Authors: Haojie Wang, Xun Li, Rui Yin

Abstract:

In order to solve the memorization overfitting in the model-agnostic meta-learning MAML algorithm, a method of generating mutually exclusive tasks based on data augmentation is proposed. This method generates a mutex task by corresponding one feature of the data to multiple labels so that the generated mutex task is inconsistent with the data distribution in the initial dataset. Because generating mutex tasks for all data will produce a large number of invalid data and, in the worst case, lead to an exponential growth of computation, this paper also proposes a key data extraction method that only extract part of the data to generate the mutex task. The experiments show that the method of generating mutually exclusive tasks can effectively solve the memorization overfitting in the meta-learning MAML algorithm.

Keywords: mutex task generation, data augmentation, meta-learning, text classification.

Procedia PDF Downloads 108
24150 Revolutionizing Traditional Farming Using Big Data/Cloud Computing: A Review on Vertical Farming

Authors: Milind Chaudhari, Suhail Balasinor

Abstract:

Due to massive deforestation and an ever-increasing population, the organic content of the soil is depleting at a much faster rate. Due to this, there is a big chance that the entire food production in the world will drop by 40% in the next two decades. Vertical farming can help in aiding food production by leveraging big data and cloud computing to ensure plants are grown naturally by providing the optimum nutrients sunlight by analyzing millions of data points. This paper outlines the most important parameters in vertical farming and how a combination of big data and AI helps in calculating and analyzing these millions of data points. Finally, the paper outlines how different organizations are controlling the indoor environment by leveraging big data in enhancing food quantity and quality.

Keywords: big data, IoT, vertical farming, indoor farming

Procedia PDF Downloads 147
24149 Data Challenges Facing Implementation of Road Safety Management Systems in Egypt

Authors: A. Anis, W. Bekheet, A. El Hakim

Abstract:

Implementing a Road Safety Management System (SMS) in a crowded developing country such as Egypt is a necessity. Beginning a sustainable SMS requires a comprehensive reliable data system for all information pertinent to road crashes. In this paper, a survey for the available data in Egypt and validating it for using in an SMS in Egypt. The research provides some missing data, and refer to the unavailable data in Egypt, looking forward to the contribution of the scientific society, the authorities, and the public in solving the problem of missing or unreliable crash data. The required data for implementing an SMS in Egypt are divided into three categories; the first is available data such as fatality and injury rates and it is proven in this research that it may be inconsistent and unreliable, the second category of data is not available, but it may be estimated, an example of estimating vehicle cost is available in this research, the third is not available and can be measured case by case such as the functional and geometric properties of a facility. Some inquiries are provided in this research for the scientific society, such as how to improve the links among stakeholders of road safety in order to obtain a consistent, non-biased, and reliable data system.

Keywords: road safety management system, road crash, road fatality, road injury

Procedia PDF Downloads 92
24148 Predictors of Lost to Follow-Up among HIV Patients Attending Anti-Retroviral Therapy Treatment Centers in Nigeria

Authors: Oluwasina Folajinmi, Kate Ssamulla, Penninah Lutung, Daniel Reijer

Abstract:

Background: Despite of well-verified benefits of anti-retroviral therapy (ART) in prolonging life expectancy being lost to follow-up (LTFU) presents a challenge to the success of ART programs in resource limited countries like Nigeria. In several studies of ART programs in developing countries, researchers have reported that there has been a high rate of LTFU among patients receiving care and treatment at ART treatment centers. This study seeks to determine the cause of LTFU among HIV clients. Method: A descriptive cross sectional study focused on a population of 9,280 persons living with HIV/AIDS who were enrolled in nine treatment centers in Nigeria (both pre-ART and ART patients were included). Out of the total population, 1752 (18.9%) were found to be LTFU. Of this group we randomly selected 1200 clients (68.5%) their d patients’ information was generated through a database. Data on demographics and CD4 counts, causes of LTFU were analyzed and summarized. Results: Out of 1200 LTFU clients selected, 462 (38.5%) were on ART; 341 clients (73.8%) had CD4 level < 500cell/µL and 738 (61.5%) on pre-ART had CD4 level >500/µL. In our records we found telephone number for 675 (56.1%) of these clients. 675 (56.1%) were owners of a phone. The majority of the client’s 731 (60.9%) were living at not more than 25km away from the ART center. A majority were females (926 or 77.2%) while 274 (22.8%) were male. 675 (56.1%) clients were reported traced via telephone and home address. 326 (27.2%) of clients phone numbers were not reachable; 173 (14.4%) of telephone numbers were incomplete. 71 (5.9%) had relocated due to communal crises and expert client trackers reported that some patient could not afford transportation to ART centers. Conclusion: This study shows that, low health education levels, poverty, relocations and lack of reliable phone contact were major predictors of LTFU. Periodic updates of home addresses, telephone contacts including at least two next of kin, phone text messages and home visits may improve follow up. Early and consistent tracking of missed appointments is crucial. Creation of more ART decentralized centres are needed to avoid long distances.

Keywords: anti-retroviral therapy, HIV/AIDS, predictors, lost to follow up

Procedia PDF Downloads 284
24147 Big Data-Driven Smart Policing: Big Data-Based Patrol Car Dispatching in Abu Dhabi, UAE

Authors: Oualid Walid Ben Ali

Abstract:

Big Data has become one of the buzzwords today. The recent explosion of digital data has led the organization, either private or public, to a new era towards a more efficient decision making. At some point, business decided to use that concept in order to learn what make their clients tick with phrases like ‘sales funnel’ analysis, ‘actionable insights’, and ‘positive business impact’. So, it stands to reason that Big Data was viewed through green (read: money) colored lenses. Somewhere along the line, however someone realized that collecting and processing data doesn’t have to be for business purpose only, but also could be used for other purposes to assist law enforcement or to improve policing or in road safety. This paper presents briefly, how Big Data have been used in the fields of policing order to improve the decision making process in the daily operation of the police. As example, we present a big-data driven system which is sued to accurately dispatch the patrol cars in a geographic environment. The system is also used to allocate, in real-time, the nearest patrol car to the location of an incident. This system has been implemented and applied in the Emirate of Abu Dhabi in the UAE.

Keywords: big data, big data analytics, patrol car allocation, dispatching, GIS, intelligent, Abu Dhabi, police, UAE

Procedia PDF Downloads 464
24146 Mining Multicity Urban Data for Sustainable Population Relocation

Authors: Xu Du, Aparna S. Varde

Abstract:

In this research, we propose to conduct diagnostic and predictive analysis about the key factors and consequences of urban population relocation. To achieve this goal, urban simulation models extract the urban development trends as land use change patterns from a variety of data sources. The results are treated as part of urban big data with other information such as population change and economic conditions. Multiple data mining methods are deployed on this data to analyze nonlinear relationships between parameters. The result determines the driving force of population relocation with respect to urban sprawl and urban sustainability and their related parameters. Experiments so far reveal that data mining methods discover useful knowledge from the multicity urban data. This work sets the stage for developing a comprehensive urban simulation model for catering to specific questions by targeted users. It contributes towards achieving sustainability as a whole.

Keywords: data mining, environmental modeling, sustainability, urban planning

Procedia PDF Downloads 273
24145 Fracture Dislocation of Upper Sacrum in an Adolescent: Case Report and Review of Literature

Authors: S. Alireza Mirghasemi, Narges Rahimi Gabaran

Abstract:

Although sacral fractures in children are rare due to the fact that the occurrence of pelvic fracture is not common in childhood. Sacral fractures present a high risk of neurological damage. This kind of fracture is often missed because the routine pelvic X-rays imaging scarcely show this fracture. Also, the treatment is controversial, and it ranges from fine reduction to conservative treatments without any try to reduce the dislocation. In this article, a case of fracture dislocation of S1 and S2 along with a suggested diagnostic test and treatment based on similar cases are presented. The case investigates a 14-year-old boy who entered the hospital one week after a car accident that knocked him to the ground in crawling position and a rack fell down on his body. Pain and tenderness in the sacral region and a fracture in the left leg were notable--we detected incomplete bilateral palsy of L5, S1 and S2 roots. In radiographs of the spine fracture dislocation of S1, the sacral fracture was seen. The treatment included a skeletal traction with a halo over the patient’s head and two femoral pins. After one week, another surgery was performed in order to stabilize and reduce the fracture, and we employed a posterior approach with CD and a pedicular screw. After two years of follow-up, the fracture is completely cured without any loss of reduction.

Keywords: adolescent, fracture in adolescent, fracture dislocation, sacrum

Procedia PDF Downloads 271
24144 Model Order Reduction for Frequency Response and Effect of Order of Method for Matching Condition

Authors: Aref Ghafouri, Mohammad javad Mollakazemi, Farhad Asadi

Abstract:

In this paper, model order reduction method is used for approximation in linear and nonlinearity aspects in some experimental data. This method can be used for obtaining offline reduced model for approximation of experimental data and can produce and follow the data and order of system and also it can match to experimental data in some frequency ratios. In this study, the method is compared in different experimental data and influence of choosing of order of the model reduction for obtaining the best and sufficient matching condition for following the data is investigated in format of imaginary and reality part of the frequency response curve and finally the effect and important parameter of number of order reduction in nonlinear experimental data is explained further.

Keywords: frequency response, order of model reduction, frequency matching condition, nonlinear experimental data

Procedia PDF Downloads 373
24143 Multi-Objective Multi-Mode Resource-Constrained Project Scheduling Problem by Preemptive Fuzzy Goal Programming

Authors: Busaba Phurksaphanrat

Abstract:

This research proposes a pre-emptive fuzzy goal programming model for multi-objective multi-mode resource constrained project scheduling problem. The objectives of the problem are minimization of the total time and the total cost of the project. Objective in a multi-mode resource-constrained project scheduling problem is often a minimization of make-span. However, both time and cost should be considered at the same time with different level of important priorities. Moreover, all elements of cost functions in a project are not included in the conventional cost objective function. Incomplete total project cost causes an error in finding the project scheduling time. In this research, pre-emptive fuzzy goal programming is presented to solve the multi-objective multi-mode resource constrained project scheduling problem. It can find the compromise solution of the problem. Moreover, it is also flexible in adjusting to find a variety of alternative solutions.

Keywords: multi-mode resource constrained project scheduling problem, fuzzy set, goal programming, pre-emptive fuzzy goal programming

Procedia PDF Downloads 406
24142 An Empirical Study of the Impacts of Big Data on Firm Performance

Authors: Thuan Nguyen

Abstract:

In the present time, data to a data-driven knowledge-based economy is the same as oil to the industrial age hundreds of years ago. Data is everywhere in vast volumes! Big data analytics is expected to help firms not only efficiently improve performance but also completely transform how they should run their business. However, employing the emergent technology successfully is not easy, and assessing the roles of big data in improving firm performance is even much harder. There was a lack of studies that have examined the impacts of big data analytics on organizational performance. This study aimed to fill the gap. The present study suggested using firms’ intellectual capital as a proxy for big data in evaluating its impact on organizational performance. The present study employed the Value Added Intellectual Coefficient method to measure firm intellectual capital, via its three main components: human capital efficiency, structural capital efficiency, and capital employed efficiency, and then used the structural equation modeling technique to model the data and test the models. The financial fundamental and market data of 100 randomly selected publicly listed firms were collected. The results of the tests showed that only human capital efficiency had a significant positive impact on firm profitability, which highlighted the prominent human role in the impact of big data technology.

Keywords: big data, big data analytics, intellectual capital, organizational performance, value added intellectual coefficient

Procedia PDF Downloads 216