Search results for: algorithms and data structure
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 32286

Search results for: algorithms and data structure

31506 A Supervised Learning Data Mining Approach for Object Recognition and Classification in High Resolution Satellite Data

Authors: Mais Nijim, Rama Devi Chennuboyina, Waseem Al Aqqad

Abstract:

Advances in spatial and spectral resolution of satellite images have led to tremendous growth in large image databases. The data we acquire through satellites, radars and sensors consists of important geographical information that can be used for remote sensing applications such as region planning, disaster management. Spatial data classification and object recognition are important tasks for many applications. However, classifying objects and identifying them manually from images is a difficult task. Object recognition is often considered as a classification problem, this task can be performed using machine-learning techniques. Despite of many machine-learning algorithms, the classification is done using supervised classifiers such as Support Vector Machines (SVM) as the area of interest is known. We proposed a classification method, which considers neighboring pixels in a region for feature extraction and it evaluates classifications precisely according to neighboring classes for semantic interpretation of region of interest (ROI). A dataset has been created for training and testing purpose; we generated the attributes by considering pixel intensity values and mean values of reflectance. We demonstrated the benefits of using knowledge discovery and data-mining techniques, which can be on image data for accurate information extraction and classification from high spatial resolution remote sensing imagery.

Keywords: remote sensing, object recognition, classification, data mining, waterbody identification, feature extraction

Procedia PDF Downloads 342
31505 The Effect of Multimedia Use on Students’ Academic Achievement and Course-Oriented Self-Efficacy

Authors: Hasan Coruk, Recep Cakir

Abstract:

This study aimed at investigating the effect of multimedia containing ‘the structure and properties of matter’ unit on students’ academic achievement level and self-efficacy relating to science and technology course. The study used an experimental design with pre-test and post-test groups. The data collection tools were ‘Science and Technology Course Achievement Test’ and ‘Science and Technology Self-Efficacy Scale’. The sample of the study consisted of 8th grade students at a primary school in Tokat Province. The study was carried out with 42 students from two classes, 21 (8 males, 13 females) from experimental group and 21 (13 males and 8 females) from control group. The data were analyzed in SPSS.18 software. The findings of the study indicated that the use of multimedia increased the students’ academic achievement in science and technology course in comparison with traditional teaching methods. It was also determined that there was not a significant difference in students’ course-oriented self-efficacy levels regarding the two methods. Necessary and feasible suggestions were put forward for whom it concerns.

Keywords: multimedia learning, science and technology, the structure-properties of matter, self-efficacy, academic achievement

Procedia PDF Downloads 457
31504 Analysis of Steel Beam-Column Joints Under Seismic Loads

Authors: Mizam Doğan

Abstract:

Adapazarı railway car factory, the only railway car factory of Turkey, was constructed in 1950. It was a steel design and it had filled beam sections and truss beam systems. Columns were steel profiles and box sections. The factory was damaged heavily on Izmit Earthquake and closed. In this earthquake 90% of damaged structures are reinforced concrete, the others are %7 prefabricated and 3% steel construction. As can be seen in statistical data, damaged industrial buildings in this earthquake were generally reinforced concrete and prefabricated structures. Adapazari railway car factory is the greatest steel structure damaged in the earthquake. This factory has 95% of the total damaged steel structure area. In this paper; earthquake damages on beams and columns of the factory are studied by considering TS648 'Turkish Standard Building Code for Steel Structures' and also damaged connection elements as welds, rivets and bolts are examined. A model similar to the damaged system is made and high-stress zones are searched. These examinations, conclusions, suggestions are explained by damage photos and details.

Keywords: column-beam connection, seismic analysis, seismic load, steel structure

Procedia PDF Downloads 279
31503 Dividend Payout and Capital Structure: A Family Firm Perspective

Authors: Abhinav Kumar Rajverma, Arun Kumar Misra, Abhijeet Chandra

Abstract:

Family involvement in business is universal across countries, with varying characteristics. Firms of developed economies have diffused ownership structure; however, that of emerging markets have concentrated ownership structure, having resemblance with that of family firms. Optimization of dividend payout and leverage are very crucial for firm’s valuation. This paper studies dividend paying behavior of National Stock Exchange listed Indian firms from financial year 2007 to 2016. The final sample consists of 422 firms and of these more than 49% (207) are family firms. Results reveal that family firms pay lower dividend and are more leveraged compared to non-family firms. This unique data set helps to understand dividend behavior and capital structure of sample firms over a long-time period and across varying family ownership concentration. Using panel regression models, this paper examines factors affecting dividend payout and capital structure and establishes a link between the two using Two-stage Least Squares regression model. Profitability shows a positive impact on dividend and negative impact on leverage, confirming signaling and pecking order theory. Further, findings support bankruptcy theory as firm size has a positive relation with dividend and leverage and volatility shows a negative relation with both dividend and leverage. Findings are also consistent with agency theory, family ownership concentration has negative relation with both dividend payments and leverage. Further, the impact of family ownership control confirms the similar finding. The study further reveals that firms with high family ownership concentration (family control) do have an impact on determining the level of private benefits. Institutional ownership is not significant for dividend payments. However, it shows significant negative relation with leverage for both family and non-family firms. Dividend payout and leverage show mixed association with each other. This paper provides evidence of how varying level of family ownership concentration and ownership control influences the dividend policy and capital structure of firms in an emerging market like India and it can have significant contribution towards understanding and formulating corporate dividend policy decisions and capital structure for emerging economies, where majority of firms exhibit behavior of family firm.

Keywords: dividend, family firms, leverage, ownership structure

Procedia PDF Downloads 282
31502 Location Privacy Preservation of Vehicle Data In Internet of Vehicles

Authors: Ying Ying Liu, Austin Cooke, Parimala Thulasiraman

Abstract:

Internet of Things (IoT) has attracted a recent spark in research on Internet of Vehicles (IoV). In this paper, we focus on one research area in IoV: preserving location privacy of vehicle data. We discuss existing location privacy preserving techniques and provide a scheme for evaluating these techniques under IoV traffic condition. We propose a different strategy in applying Differential Privacy using k-d tree data structure to preserve location privacy and experiment on real world Gowalla data set. We show that our strategy produces differentially private data, good preservation of utility by achieving similar regression accuracy to the original dataset on an LSTM (Long Term Short Term Memory) neural network traffic predictor.

Keywords: differential privacy, internet of things, internet of vehicles, location privacy, privacy preservation scheme

Procedia PDF Downloads 183
31501 Characterization of Nickel Based Metallic Superconducting Materials

Authors: Y. Benmalem , A. Abbad, W. Benstaali, T. Lantri

Abstract:

Density functional theory is used to investigate the.the structural, electronic, and magnetic properties of the cubic anti-perovskites InNNi3 and ZnNNi3. The structure of antiperovskite also called (perovskite-inverse) identical to the perovskite structure of the general formula ABX3, where A is a main group (III–V) element or a metallic element, B is carbon or nitrogen, and X is a transition metal, displays a wide range of interesting physical properties, such as giant magnetoresistance. Elastic and electronic properties were determined using generalized gradient approximation (GGA), and local spin density approximation (LSDA) approaches, ), as implemented in the Wien2k computer package. The results show that the two compounds are strong ductile and satisfy the Born-Huang criteria, so they are mechanically stable at normal conditions. Electronic properties show that the two compounds studied are metallic and non-magnetic. The studies of these compounds have confirmed the effectiveness of the two approximations and the ground-state properties are in good agreement with experimental data and theoretical results available.

Keywords: anti-perovskites, elastic anisotropy, electronic band structure, first-principles calculations

Procedia PDF Downloads 288
31500 Mathematical Properties of the Viscous Rotating Stratified Fluid Counting with Salinity and Heat Transfer in a Layer

Authors: A. Giniatoulline

Abstract:

A model of the mathematical fluid dynamics which describes the motion of a three-dimensional viscous rotating fluid in a homogeneous gravitational field with the consideration of the salinity and heat transfer is considered in a vertical finite layer. The model is a generalization of the linearized Navier-Stokes system with the addition of the Coriolis parameter and the equations for changeable density, salinity, and heat transfer. An explicit solution is constructed and the proof of the existence and uniqueness theorems is given. The localization and the structure of the spectrum of inner waves is also investigated. The results may be used, in particular, for constructing stable numerical algorithms for solutions of the considered models of fluid dynamics of the Atmosphere and the Ocean.

Keywords: Fourier transform, generalized solutions, Navier-Stokes equations, stratified fluid

Procedia PDF Downloads 257
31499 Matrix Completion with Heterogeneous Cost

Authors: Ilqar Ramazanli

Abstract:

The matrix completion problem has been studied broadly under many underlying conditions. The problem has been explored under adaptive or non-adaptive, exact or estimation, single-phase or multi-phase, and many other categories. In most of these cases, the observation cost of each entry is uniform and has the same cost across the columns. However, in many real-life scenarios, we could expect elements from distinct columns or distinct positions to have a different cost. In this paper, we explore this generalization under adaptive conditions. We approach the problem under two different cost models. The first one is that entries from different columns have different observation costs, but within the same column, each entry has a uniform cost. The second one is any two entry has different observation cost, despite being the same or different columns. We provide complexity analysis of our algorithms and provide tightness guarantees.

Keywords: matroid optimization, matrix completion, linear algebra, algorithms

Procedia PDF Downloads 111
31498 Comparative Analysis of Classification Methods in Determining Non-Active Student Characteristics in Indonesia Open University

Authors: Dewi Juliah Ratnaningsih, Imas Sukaesih Sitanggang

Abstract:

Classification is one of data mining techniques that aims to discover a model from training data that distinguishes records into the appropriate category or class. Data mining classification methods can be applied in education, for example, to determine the classification of non-active students in Indonesia Open University. This paper presents a comparison of three methods of classification: Naïve Bayes, Bagging, and C.45. The criteria used to evaluate the performance of three methods of classification are stratified cross-validation, confusion matrix, the value of the area under the ROC Curve (AUC), Recall, Precision, and F-measure. The data used for this paper are from the non-active Indonesia Open University students in registration period of 2004.1 to 2012.2. Target analysis requires that non-active students were divided into 3 groups: C1, C2, and C3. Data analyzed are as many as 4173 students. Results of the study show: (1) Bagging method gave a high degree of classification accuracy than Naïve Bayes and C.45, (2) the Bagging classification accuracy rate is 82.99 %, while the Naïve Bayes and C.45 are 80.04 % and 82.74 % respectively, (3) the result of Bagging classification tree method has a large number of nodes, so it is quite difficult in decision making, (4) classification of non-active Indonesia Open University student characteristics uses algorithms C.45, (5) based on the algorithm C.45, there are 5 interesting rules which can describe the characteristics of non-active Indonesia Open University students.

Keywords: comparative analysis, data mining, clasiffication, Bagging, Naïve Bayes, C.45, non-active students, Indonesia Open University

Procedia PDF Downloads 318
31497 Dual Set Point Governor Control Structure with Common Optimum Temporary Droop Settings for both Islanded and Grid Connected Modes

Authors: Deepen Sharma, Eugene F. Hill

Abstract:

For nearly 100 years, hydro-turbine governors have operated with only a frequency set point. This natural governor action means that the governor responds with changing megawatt output to disturbances in system frequency. More and more, power system managers are demanding that governors operate with constant megawatt output. One way of doing this is to introduce a second set point in the control structure called a power set point. The control structure investigated and analyzed in this paper is unique in the way that it utilizes a power reference set point in addition to the conventional frequency reference set point. An optimum set of temporary droop parameters derived based on the turbine-generator inertia constant and the penstock water start time for stable islanded operation are shown to be also equally applicable for a satisfactory rate of generator loading during its grid connected mode. A theoretical development shows why this is the case. The performance of the control structure has been investigated and established based on the simulation study made in MATLAB/Simulink as well as through testing the real time controller performance on a 15 MW Kaplan Turbine and generator. Recordings have been made using the labVIEW data acquisition platform. The hydro-turbine governor control structure and its performance investigated in this paper thus eliminates the need to have a separate set of temporary droop parameters, one valid for islanded mode and the other for interconnected operations mode.

Keywords: frequency set point, hydro governor, interconnected operation, isolated operation, power set point

Procedia PDF Downloads 369
31496 A Formal Approach for Instructional Design Integrated with Data Visualization for Learning Analytics

Authors: Douglas A. Menezes, Isabel D. Nunes, Ulrich Schiel

Abstract:

Most Virtual Learning Environments do not provide support mechanisms for the integrated planning, construction and follow-up of Instructional Design supported by Learning Analytic results. The present work aims to present an authoring tool that will be responsible for constructing the structure of an Instructional Design (ID), without the data being altered during the execution of the course. The visual interface aims to present the critical situations present in this ID, serving as a support tool for the course follow-up and possible improvements, which can be made during its execution or in the planning of a new edition of this course. The model for the ID is based on High-Level Petri Nets and the visualization forms are determined by the specific kind of the data generated by an e-course, a population of students generating sequentially dependent data.

Keywords: educational data visualization, high-level petri nets, instructional design, learning analytics

Procedia PDF Downloads 248
31495 Design and Development of an Algorithm to Predict Fluctuations of Currency Rates

Authors: Nuwan Kuruwitaarachchi, M. K. M. Peiris, C. N. Madawala, K. M. A. R. Perera, V. U. N Perera

Abstract:

Dealing with businesses with the foreign market always took a special place in a country’s economy. Political and social factors came into play making currency rate changes fluctuate rapidly. Currency rate prediction has become an important factor for larger international businesses since large amounts of money exchanged between countries. This research focuses on comparing the accuracy of mainly three models; Autoregressive Integrated Moving Average (ARIMA), Artificial Neural Networks(ANN) and Support Vector Machines(SVM). series of data import, export, USD currency exchange rate respect to LKR has been selected for training using above mentioned algorithms. After training the data set and comparing each algorithm, it was able to see that prediction in SVM performed better than other models. It was improved more by combining SVM and SVR models together.

Keywords: ARIMA, ANN, FFNN, RMSE, SVM, SVR

Procedia PDF Downloads 215
31494 Discrete PID and Discrete State Feedback Control of a Brushed DC Motor

Authors: I. Valdez, J. Perdomo, M. Colindres, N. Castro

Abstract:

Today, digital servo systems are extensively used in industrial manufacturing processes, robotic applications, vehicles and other areas. In such control systems, control action is provided by digital controllers with different compensation algorithms, which are designed to meet specific requirements for a given application. Due to the constant search for optimization in industrial processes, it is of interest to design digital controllers that offer ease of realization, improved computational efficiency, affordable return rates, and ease of tuning that ultimately improve the performance of the controlled actuators. There is a vast range of options of compensation algorithms that could be used, although in the industry, most controllers used are based on a PID structure. This research article compares different types of digital compensators implemented in a servo system for DC motor position control. PID compensation is evaluated on its two most common architectures: PID position form (1 DOF), and PID speed form (2 DOF). State feedback algorithms are also evaluated, testing two modern control theory techniques: discrete state observer for non-measurable variables tracking, and a linear quadratic method which allows a compromise between the theoretical optimal control and the realization that most closely matches it. The compared control systems’ performance is evaluated through simulations in the Simulink platform, in which it is attempted to model accurately each of the system’s hardware components. The criteria by which the control systems are compared are reference tracking and disturbance rejection. In this investigation, it is considered that the accurate tracking of the reference signal for a position control system is particularly important because of the frequency and the suddenness in which the control signal could change in position control applications, while disturbance rejection is considered essential because the torque applied to the motor shaft due to sudden load changes can be modeled as a disturbance that must be rejected, ensuring reference tracking. Results show that 2 DOF PID controllers exhibit high performance in terms of the benchmarks mentioned, as long as they are properly tuned. As for controllers based on state feedback, due to the nature and the advantage which state space provides for modelling MIMO, it is expected that such controllers evince ease of tuning for disturbance rejection, assuming that the designer of such controllers is experienced. An in-depth multi-dimensional analysis of preliminary research results indicate that state feedback control method is more satisfactory, but PID control method exhibits easier implementation in most control applications.

Keywords: control, DC motor, discrete PID, discrete state feedback

Procedia PDF Downloads 269
31493 A Non-parametric Clustering Approach for Multivariate Geostatistical Data

Authors: Francky Fouedjio

Abstract:

Multivariate geostatistical data have become omnipresent in the geosciences and pose substantial analysis challenges. One of them is the grouping of data locations into spatially contiguous clusters so that data locations within the same cluster are more similar while clusters are different from each other, in some sense. Spatially contiguous clusters can significantly improve the interpretation that turns the resulting clusters into meaningful geographical subregions. In this paper, we develop an agglomerative hierarchical clustering approach that takes into account the spatial dependency between observations. It relies on a dissimilarity matrix built from a non-parametric kernel estimator of the spatial dependence structure of data. It integrates existing methods to find the optimal cluster number and to evaluate the contribution of variables to the clustering. The capability of the proposed approach to provide spatially compact, connected and meaningful clusters is assessed using bivariate synthetic dataset and multivariate geochemical dataset. The proposed clustering method gives satisfactory results compared to other similar geostatistical clustering methods.

Keywords: clustering, geostatistics, multivariate data, non-parametric

Procedia PDF Downloads 481
31492 Mobile Crowdsensing Scheme by Predicting Vehicle Mobility Using Deep Learning Algorithm

Authors: Monojit Manna, Arpan Adhikary

Abstract:

In Mobile cloud sensing across the globe, an emerging paradigm is selected by the user to compute sensing tasks. In urban cities current days, Mobile vehicles are adapted to perform the task of data sensing and data collection for universality and mobility. In this work, we focused on the optimality and mobile nodes that can be selected in order to collect the maximum amount of data from urban areas and fulfill the required data in the future period within a couple of minutes. We map out the requirement of the vehicle to configure the maximum data optimization problem and budget. The Application implementation is basically set up to generalize a realistic online platform in which real-time vehicles are moving apparently in a continuous manner. The data center has the authority to select a set of vehicles immediately. A deep learning-based scheme with the help of mobile vehicles (DLMV) will be proposed to collect sensing data from the urban environment. From the future time perspective, this work proposed a deep learning-based offline algorithm to predict mobility. Therefore, we proposed a greedy approach applying an online algorithm step into a subset of vehicles for an NP-complete problem with a limited budget. Real dataset experimental extensive evaluations are conducted for the real mobility dataset in Rome. The result of the experiment not only fulfills the efficiency of our proposed solution but also proves the validity of DLMV and improves the quantity of collecting the sensing data compared with other algorithms.

Keywords: mobile crowdsensing, deep learning, vehicle recruitment, sensing coverage, data collection

Procedia PDF Downloads 81
31491 Sequential Pattern Mining from Data of Medical Record with Sequential Pattern Discovery Using Equivalent Classes (SPADE) Algorithm (A Case Study : Bolo Primary Health Care, Bima)

Authors: Rezky Rifaini, Raden Bagus Fajriya Hakim

Abstract:

This research was conducted at the Bolo primary health Care in Bima Regency. The purpose of the research is to find out the association pattern that is formed of medical record database from Bolo Primary health care’s patient. The data used is secondary data from medical records database PHC. Sequential pattern mining technique is the method that used to analysis. Transaction data generated from Patient_ID, Check_Date and diagnosis. Sequential Pattern Discovery Algorithms Using Equivalent Classes (SPADE) is one of the algorithm in sequential pattern mining, this algorithm find frequent sequences of data transaction, using vertical database and sequence join process. Results of the SPADE algorithm is frequent sequences that then used to form a rule. It technique is used to find the association pattern between items combination. Based on association rules sequential analysis with SPADE algorithm for minimum support 0,03 and minimum confidence 0,75 is gotten 3 association sequential pattern based on the sequence of patient_ID, check_Date and diagnosis data in the Bolo PHC.

Keywords: diagnosis, primary health care, medical record, data mining, sequential pattern mining, SPADE algorithm

Procedia PDF Downloads 404
31490 Non-Parametric Regression over Its Parametric Couterparts with Large Sample Size

Authors: Jude Opara, Esemokumo Perewarebo Akpos

Abstract:

This paper is on non-parametric linear regression over its parametric counterparts with large sample size. Data set on anthropometric measurement of primary school pupils was taken for the analysis. The study used 50 randomly selected pupils for the study. The set of data was subjected to normality test, and it was discovered that the residuals are not normally distributed (i.e. they do not follow a Gaussian distribution) for the commonly used least squares regression method for fitting an equation into a set of (x,y)-data points using the Anderson-Darling technique. The algorithms for the nonparametric Theil’s regression are stated in this paper as well as its parametric OLS counterpart. The use of a programming language software known as “R Development” was used in this paper. From the analysis, the result showed that there exists a significant relationship between the response and the explanatory variable for both the parametric and non-parametric regression. To know the efficiency of one method over the other, the Akaike Information Criterion (AIC), Bayesian Information Criterion (BIC) are used, and it is discovered that the nonparametric regression performs better than its parametric regression counterparts due to their lower values in both the AIC and BIC. The study however recommends that future researchers should study a similar work by examining the presence of outliers in the data set, and probably expunge it if detected and re-analyze to compare results.

Keywords: Theil’s regression, Bayesian information criterion, Akaike information criterion, OLS

Procedia PDF Downloads 310
31489 Electronic, Structure and Magnetic Properties of KXF3(X= Fe, Co, Mn, V) from Ab Initio Calculations

Authors: M. Ibrir, S. Berri, S. Lakel, D. Maouche And Y. Medkour

Abstract:

We have performed first-principle calculations of the structural, electronic and magnetic properties of KFeF3, KCoF3, KMnF3, KVF3, using full-potential linearized augmented plane-wave (FP-LAPW) scheme within GGA. Features such as the lattice constant, bulk modulus and its pressure derivative are reported. Also, we have presented our results of the band structure and the density of states. The magnetic moments of KFeF3, KCoF3, KMnF3, KVF3 compounds are in most came from the exchange-splitting of X-3d orbital.

Keywords: Ab initio calculations, electronic structure, magnetic materials

Procedia PDF Downloads 422
31488 Modeling of Historical Lime Masonry Structure in Abaqus

Authors: Ram Narayan Khare, Adhyatma Khare, Aradhna Shrivastava

Abstract:

In this study, numerical modeling of ‘Lime Surkhi’ masonry building has been carried out for a prototype ancient building situated at seismic zone III using the Finite Element Method by Abaqus software. The model is designed in order to get the failure envelope and then decide the best method of retrofitting the structure so that the structure is made to withstand more decades, given its historical background. Previously, due to a lack of technologies, it was difficult to determine the mode of failure. Present technological development can predict the mode of failure, and subsequently, the structure can be refabricated accordingly. The study makes an important addition to the understanding of retrofitting ancient and old buildings based on the results of FEM modeling.

Keywords: seismic retrofitting, Abaqus, FEM, historic building, Lime Surkhi masonry

Procedia PDF Downloads 40
31487 Gradient Boosted Trees on Spark Platform for Supervised Learning in Health Care Big Data

Authors: Gayathri Nagarajan, L. D. Dhinesh Babu

Abstract:

Health care is one of the prominent industries that generate voluminous data thereby finding the need of machine learning techniques with big data solutions for efficient processing and prediction. Missing data, incomplete data, real time streaming data, sensitive data, privacy, heterogeneity are few of the common challenges to be addressed for efficient processing and mining of health care data. In comparison with other applications, accuracy and fast processing are of higher importance for health care applications as they are related to the human life directly. Though there are many machine learning techniques and big data solutions used for efficient processing and prediction in health care data, different techniques and different frameworks are proved to be effective for different applications largely depending on the characteristics of the datasets. In this paper, we present a framework that uses ensemble machine learning technique gradient boosted trees for data classification in health care big data. The framework is built on Spark platform which is fast in comparison with other traditional frameworks. Unlike other works that focus on a single technique, our work presents a comparison of six different machine learning techniques along with gradient boosted trees on datasets of different characteristics. Five benchmark health care datasets are considered for experimentation, and the results of different machine learning techniques are discussed in comparison with gradient boosted trees. The metric chosen for comparison is misclassification error rate and the run time of the algorithms. The goal of this paper is to i) Compare the performance of gradient boosted trees with other machine learning techniques in Spark platform specifically for health care big data and ii) Discuss the results from the experiments conducted on datasets of different characteristics thereby drawing inference and conclusion. The experimental results show that the accuracy is largely dependent on the characteristics of the datasets for other machine learning techniques whereas gradient boosting trees yields reasonably stable results in terms of accuracy without largely depending on the dataset characteristics.

Keywords: big data analytics, ensemble machine learning, gradient boosted trees, Spark platform

Procedia PDF Downloads 243
31486 Effectiveness of Reinforcement Learning (RL) for Autonomous Energy Management Solutions

Authors: Tesfaye Mengistu

Abstract:

This thesis aims to investigate the effectiveness of Reinforcement Learning (RL) for Autonomous Energy Management solutions. The study explores the potential of Model Free RL approaches, such as Monte Carlo RL and Q-learning, to improve energy management by autonomously adjusting energy management strategies to maximize efficiency. The research investigates the implementation of RL algorithms for optimizing energy consumption in a single-agent environment. The focus is on developing a framework for the implementation of RL algorithms, highlighting the importance of RL for enabling autonomous systems to adapt quickly to changing conditions and make decisions based on previous experiences. Moreover, the paper proposes RL as a novel energy management solution to address nations' CO2 emission goals. Reinforcement learning algorithms are well-suited to solving problems with sequential decision-making patterns and can provide accurate and immediate outputs to ease the planning and decision-making process. This research provides insights into the challenges and opportunities of using RL for energy management solutions and recommends further studies to explore its full potential. In conclusion, this study provides valuable insights into how RL can be used to improve the efficiency of energy management systems and supports the use of RL as a promising approach for developing autonomous energy management solutions in residential buildings.

Keywords: artificial intelligence, reinforcement learning, monte carlo, energy management, CO2 emission

Procedia PDF Downloads 90
31485 Microtomographic Analysis of Friction Materials Used in the Brakes of Railway Vehicles

Authors: Mikołaj Szyca

Abstract:

Friction elements of rail vehicle brakes are more and more often made of composite materials that displace cast iron. Materials are tested primarily in terms of their dynamic abilities, but the material structure of brake pads and linings changes during operation. In connection with the above, the changes taking place in the tested rubbing materials were analyzed using X-ray computed tomography in order to obtain data on changes in the structure of the material immediately after production and after a certain number of operating cycles. The implementation of microtomography research for experimental work on new friction materials may result in increasing the potential for the production of new composites by eliminating unfavorable material factors and, consequently, improving the dynamic parameters.

Keywords: composite materials, friction pair, X-ray computed microtomography, railway

Procedia PDF Downloads 80
31484 Unpacking Tourist Experience: A Case Study of Chinese Tourists Visiting the UK

Authors: Guanhao Tong, Li Li, Ben David

Abstract:

This study aims to provide an explanatory account of how the leisure tourist experience emerges from tourists and their surroundings through a critical realist lens. This was achieved by applying Archer’s realist social theory as the underlying theoretical ground to unpack the interplays between the external (tourism system or structure) and the internal (tourists or agency). This theory argues that social phenomena can be analyzed in three domains - structure, agency, and culture (SAC), and along three phases – structure conditioning, sociocultural interactions, and structure elaboration. From the realist perspective, the world is an open system; events and discourses are irreducible to present individuals and collectivities. Therefore, identifying the processes or mechanisms is key to help researchers understand how social reality is brought about. Based on the contextual nature of the tourist experience, the research focuses on Chinese tourists (from mainland China) to London as a destination and British culture conveyed through the concept of the destination image. This study uses an intensive approach based on Archer’s M/M approach to discover the mechanisms/processes of the emergence of the tourist experience. Individual interviews were conducted to reveal the underlying causes of lived experiences of the tourists. Secondary data was also collected to understand how British destinations are portrayed to Chinese tourists.

Keywords: Chinese tourists, destination image, M/M approach, realist social theory, social mechanisms, tourist experience

Procedia PDF Downloads 77
31483 An Adiabatic Quantum Optimization Approach for the Mixed Integer Nonlinear Programming Problem

Authors: Maxwell Henderson, Tristan Cook, Justin Chan Jin Le, Mark Hodson, YoungJung Chang, John Novak, Daniel Padilha, Nishan Kulatilaka, Ansu Bagchi, Sanjoy Ray, John Kelly

Abstract:

We present a method of using adiabatic quantum optimization (AQO) to solve a mixed integer nonlinear programming (MINLP) problem instance. The MINLP problem is a general form of a set of NP-hard optimization problems that are critical to many business applications. It requires optimizing a set of discrete and continuous variables with nonlinear and potentially nonconvex constraints. Obtaining an exact, optimal solution for MINLP problem instances of non-trivial size using classical computation methods is currently intractable. Current leading algorithms leverage heuristic and divide-and-conquer methods to determine approximate solutions. Creating more accurate and efficient algorithms is an active area of research. Quantum computing (QC) has several theoretical benefits compared to classical computing, through which QC algorithms could obtain MINLP solutions that are superior to current algorithms. AQO is a particular form of QC that could offer more near-term benefits compared to other forms of QC, as hardware development is in a more mature state and devices are currently commercially available from D-Wave Systems Inc. It is also designed for optimization problems: it uses an effect called quantum tunneling to explore all lowest points of an energy landscape where classical approaches could become stuck in local minima. Our work used a novel algorithm formulated for AQO to solve a special type of MINLP problem. The research focused on determining: 1) if the problem is possible to solve using AQO, 2) if it can be solved by current hardware, 3) what the currently achievable performance is, 4) what the performance will be on projected future hardware, and 5) when AQO is likely to provide a benefit over classical computing methods. Two different methods, integer range and 1-hot encoding, were investigated for transforming the MINLP problem instance constraints into a mathematical structure that can be embedded directly onto the current D-Wave architecture. For testing and validation a D-Wave 2X device was used, as well as QxBranch’s QxLib software library, which includes a QC simulator based on simulated annealing. Our results indicate that it is mathematically possible to formulate the MINLP problem for AQO, but that currently available hardware is unable to solve problems of useful size. Classical general-purpose simulated annealing is currently able to solve larger problem sizes, but does not scale well and such methods would likely be outperformed in the future by improved AQO hardware with higher qubit connectivity and lower temperatures. If larger AQO devices are able to show improvements that trend in this direction, commercially viable solutions to the MINLP for particular applications could be implemented on hardware projected to be available in 5-10 years. Continued investigation into optimal AQO hardware architectures and novel methods for embedding MINLP problem constraints on to those architectures is needed to realize those commercial benefits.

Keywords: adiabatic quantum optimization, mixed integer nonlinear programming, quantum computing, NP-hard

Procedia PDF Downloads 531
31482 Geophysical Approach in the Geological Characterization of a Dam Site: Case of the Chebabta-Dam, Meskiana, Oum El-Bouaghi

Authors: Benhammadi Hocine, Djamel Boubaya, Chaffai Hicham

Abstract:

Meskiana Area is characterized by a semi-arid climate where the water supply for irrigation and industry is not sufficient as the priority goes for domestic use. To meet the increasing population growth and development, the authorities have considered building a new water retaining structure on some major temporary water streams. For this purpose Chebabta site on Oued Meskiana was chosen as the future dam site. It is large enough to store the desired volume of water. This study comes to investigate the conditions of the site and the adequacy of the ground as a foundation for the projected dam. The conditions of the site include the geological structure and mainly the presence of discontinuities in the formation on which the dam will be built, the nature of the lithologies under the foundation and the future lake, and the presence of any hazard. This site characterization is usually carried out using different methods in order to highlight any underground buried problematic structure. In this context, the different geophysical technics remain the most used ones. Three geophysical methods were used in the case of the Chebabta dam site, namely, electric survey, seismic refraction, and tomography. The choice of the technics and the location of the scan line was made on the basis of the available geological data. In this sense, profiles have been established on both banks of Oued Meskiana. The obtained results have allowed a better characterization of the geological structure, defining the limit between the surface cover and the bedrock, which is, in other words, the limit between the weathered zone and the bedrock. Their respective thicknesses were also determined by seismic refraction and electrical resistivity sounding. However, the tomography imaging technic has succeeded in positioning a fault structure passing through the right bank of the wadi.

Keywords: dam site, fault, geophysic, investigation, Meskiana

Procedia PDF Downloads 92
31481 A Fresh Look at the Tense-Aspect System of the Qashqaie Dialect of Turkish Language

Authors: Mohammad Sharifi Bohlouli, Elnaz Sharifi Bohlouli

Abstract:

Turkish language with many dialects is native or official language of great number of people all around the world. The Qashqaie dialect of Turkish language is spoken by the Qashqaie tribe mostly scattered in the southern part of Iran. This paper aims at analyzing the tense system of this dialect to detect the type and number of tense and aspects available to its speakers. To collect a reliable data, a group of 50 old native speakers were randomly chosen as the informants and different techniques such as; Shuy et al interviews, selective listening ,and eavesdropping were used. The results of data analysis showed that the tense system in the Qashqaie dialect of Turkish language includes 3 absolute tenses, 6 aspectual, and 2 subjunctive ones. The interesting part of the study is that Qashqaie dialect enables its speakers to make a kind of aspectual opposition through verb structure which seems to be almost impossible through verb forms in any other nonturkish languages. For example in the following examples sentences 1&2 and 3&4 have the same translation In English although they are different in both meaning and structure. 1. Ali ensha yazirdi. 2. Ali ensha yazirmush. (Ali was writing a composition.) 3. Ali yadmishdi. 4. Ali yadmishimish. (Ali had slept.). The changes in the verb structure in Qashqaie dialect enables its speakers to say that whether the doer of the action remembers the process of doing the action or not. So, it presents a new aspectual opposition as Observed /nonobserved. The research findings reveal many other regularities and linguistic features that can be useful for linguists interested in Turkish in general and for those interested in tense and aspect and also they can be helpful for different pedagogical purposes including teaching and translating.

Keywords: qashqaie dialect, tense, aspect, linguistics, Turkish language

Procedia PDF Downloads 487
31480 Imputing Missing Data in Electronic Health Records: A Comparison of Linear and Non-Linear Imputation Models

Authors: Alireza Vafaei Sadr, Vida Abedi, Jiang Li, Ramin Zand

Abstract:

Missing data is a common challenge in medical research and can lead to biased or incomplete results. When the data bias leaks into models, it further exacerbates health disparities; biased algorithms can lead to misclassification and reduced resource allocation and monitoring as part of prevention strategies for certain minorities and vulnerable segments of patient populations, which in turn further reduce data footprint from the same population – thus, a vicious cycle. This study compares the performance of six imputation techniques grouped into Linear and Non-Linear models on two different realworld electronic health records (EHRs) datasets, representing 17864 patient records. The mean absolute percentage error (MAPE) and root mean squared error (RMSE) are used as performance metrics, and the results show that the Linear models outperformed the Non-Linear models in terms of both metrics. These results suggest that sometimes Linear models might be an optimal choice for imputation in laboratory variables in terms of imputation efficiency and uncertainty of predicted values.

Keywords: EHR, machine learning, imputation, laboratory variables, algorithmic bias

Procedia PDF Downloads 88
31479 Towards the Modeling of Lost Core Viability in High-Pressure Die Casting: A Fluid-Structure Interaction Model with 2-Phase Flow Fluid Model

Authors: Sebastian Kohlstädt, Michael Vynnycky, Stephan Goeke, Jan Jäckel, Andreas Gebauer-Teichmann

Abstract:

This paper summarizes the progress in the latest computational fluid dynamics research towards the modeling in of lost core viability in high-pressure die casting. High-pressure die casting is a process that is widely employed in the automotive and neighboring industries due to its advantages in casting quality and cost efficiency. The degrees of freedom are however somewhat limited as it has been so far difficult to use lost cores in the process. This is right now changing and the deployment of lost cores is considered a future growth potential for high-pressure die casting companies. The use of this technology itself is difficult though. The strength of the core material, as chiefly salt is used, is limited and experiments have shown that the cores will not hold under all circumstances and process designs. For this purpose, the publicly available CFD library foam-extend (OpenFOAM) is used, and two additional fluid models for incompressible and compressible two-phase flow are implemented as fluid solver models into the FSI library. For this purpose, the volume-of-fluid (VOF) methodology is used. The necessity for the fluid-structure interaction (FSI) approach is shown by a simple CFD model geometry. The model is benchmarked against analytical models and experimental data. Sufficient agreement is found with the analytical models and good agreement with the experimental data. An outlook on future developments concludes the paper.

Keywords: CFD, fluid-structure interaction, high-pressure die casting, multiphase flow

Procedia PDF Downloads 335
31478 Seafloor and Sea Surface Modelling in the East Coast Region of North America

Authors: Magdalena Idzikowska, Katarzyna Pająk, Kamil Kowalczyk

Abstract:

Seafloor topography is a fundamental issue in geological, geophysical, and oceanographic studies. Single-beam or multibeam sonars attached to the hulls of ships are used to emit a hydroacoustic signal from transducers and reproduce the topography of the seabed. This solution provides relevant accuracy and spatial resolution. Bathymetric data from ships surveys provides National Centers for Environmental Information – National Oceanic and Atmospheric Administration. Unfortunately, most of the seabed is still unidentified, as there are still many gaps to be explored between ship survey tracks. Moreover, such measurements are very expensive and time-consuming. The solution is raster bathymetric models shared by The General Bathymetric Chart of the Oceans. The offered products are a compilation of different sets of data - raw or processed. Indirect data for the development of bathymetric models are also measurements of gravity anomalies. Some forms of seafloor relief (e.g. seamounts) increase the force of the Earth's pull, leading to changes in the sea surface. Based on satellite altimetry data, Sea Surface Height and marine gravity anomalies can be estimated, and based on the anomalies, it’s possible to infer the structure of the seabed. The main goal of the work is to create regional bathymetric models and models of the sea surface in the area of the east coast of North America – a region of seamounts and undulating seafloor. The research includes an analysis of the methods and techniques used, an evaluation of the interpolation algorithms used, model thickening, and the creation of grid models. Obtained data are raster bathymetric models in NetCDF format, survey data from multibeam soundings in MB-System format, and satellite altimetry data from Copernicus Marine Environment Monitoring Service. The methodology includes data extraction, processing, mapping, and spatial analysis. Visualization of the obtained results was carried out with Geographic Information System tools. The result is an extension of the state of the knowledge of the quality and usefulness of the data used for seabed and sea surface modeling and knowledge of the accuracy of the generated models. Sea level is averaged over time and space (excluding waves, tides, etc.). Its changes, along with knowledge of the topography of the ocean floor - inform us indirectly about the volume of the entire water ocean. The true shape of the ocean surface is further varied by such phenomena as tides, differences in atmospheric pressure, wind systems, thermal expansion of water, or phases of ocean circulation. Depending on the location of the point, the higher the depth, the lower the trend of sea level change. Studies show that combining data sets, from different sources, with different accuracies can affect the quality of sea surface and seafloor topography models.

Keywords: seafloor, sea surface height, bathymetry, satellite altimetry

Procedia PDF Downloads 84
31477 Searching k-Nearest Neighbors to be Appropriate under Gaming Environments

Authors: Jae Moon Lee

Abstract:

In general, algorithms to find continuous k-nearest neighbors have been researched on the location based services, monitoring periodically the moving objects such as vehicles and mobile phone. Those researches assume the environment that the number of query points is much less than that of moving objects and the query points are not moved but fixed. In gaming environments, this problem is when computing the next movement considering the neighbors such as flocking, crowd and robot simulations. In this case, every moving object becomes a query point so that the number of query point is same to that of moving objects and the query points are also moving. In this paper, we analyze the performance of the existing algorithms focused on location based services how they operate under gaming environments.

Keywords: flocking behavior, heterogeneous agents, similarity, simulation

Procedia PDF Downloads 306