Search results for: motion data acquisition
24602 The Effect of Institutions on Economic Growth: An Analysis Based on Bayesian Panel Data Estimation
Authors: Mohammad Anwar, Shah Waliullah
Abstract:
This study investigated panel data regression models. This paper used Bayesian and classical methods to study the impact of institutions on economic growth from data (1990-2014), especially in developing countries. Under the classical and Bayesian methodology, the two-panel data models were estimated, which are common effects and fixed effects. For the Bayesian approach, the prior information is used in this paper, and normal gamma prior is used for the panel data models. The analysis was done through WinBUGS14 software. The estimated results of the study showed that panel data models are valid models in Bayesian methodology. In the Bayesian approach, the effects of all independent variables were positively and significantly affected by the dependent variables. Based on the standard errors of all models, we must say that the fixed effect model is the best model in the Bayesian estimation of panel data models. Also, it was proved that the fixed effect model has the lowest value of standard error, as compared to other models.Keywords: Bayesian approach, common effect, fixed effect, random effect, Dynamic Random Effect Model
Procedia PDF Downloads 6924601 Diagnosis of the Heart Rhythm Disorders by Using Hybrid Classifiers
Authors: Sule Yucelbas, Gulay Tezel, Cuneyt Yucelbas, Seral Ozsen
Abstract:
In this study, it was tried to identify some heart rhythm disorders by electrocardiography (ECG) data that is taken from MIT-BIH arrhythmia database by subtracting the required features, presenting to artificial neural networks (ANN), artificial immune systems (AIS), artificial neural network based on artificial immune system (AIS-ANN) and particle swarm optimization based artificial neural network (PSO-NN) classifier systems. The main purpose of this study is to evaluate the performance of hybrid AIS-ANN and PSO-ANN classifiers with regard to the ANN and AIS. For this purpose, the normal sinus rhythm (NSR), atrial premature contraction (APC), sinus arrhythmia (SA), ventricular trigeminy (VTI), ventricular tachycardia (VTK) and atrial fibrillation (AF) data for each of the RR intervals were found. Then these data in the form of pairs (NSR-APC, NSR-SA, NSR-VTI, NSR-VTK and NSR-AF) is created by combining discrete wavelet transform which is applied to each of these two groups of data and two different data sets with 9 and 27 features were obtained from each of them after data reduction. Afterwards, the data randomly was firstly mixed within themselves, and then 4-fold cross validation method was applied to create the training and testing data. The training and testing accuracy rates and training time are compared with each other. As a result, performances of the hybrid classification systems, AIS-ANN and PSO-ANN were seen to be close to the performance of the ANN system. Also, the results of the hybrid systems were much better than AIS, too. However, ANN had much shorter period of training time than other systems. In terms of training times, ANN was followed by PSO-ANN, AIS-ANN and AIS systems respectively. Also, the features that extracted from the data affected the classification results significantly.Keywords: AIS, ANN, ECG, hybrid classifiers, PSO
Procedia PDF Downloads 44624600 Topic Modelling Using Latent Dirichlet Allocation and Latent Semantic Indexing on SA Telco Twitter Data
Authors: Phumelele Kubheka, Pius Owolawi, Gbolahan Aiyetoro
Abstract:
Twitter is one of the most popular social media platforms where users can share their opinions on different subjects. As of 2010, The Twitter platform generates more than 12 Terabytes of data daily, ~ 4.3 petabytes in a single year. For this reason, Twitter is a great source for big mining data. Many industries such as Telecommunication companies can leverage the availability of Twitter data to better understand their markets and make an appropriate business decision. This study performs topic modeling on Twitter data using Latent Dirichlet Allocation (LDA). The obtained results are benchmarked with another topic modeling technique, Latent Semantic Indexing (LSI). The study aims to retrieve topics on a Twitter dataset containing user tweets on South African Telcos. Results from this study show that LSI is much faster than LDA. However, LDA yields better results with higher topic coherence by 8% for the best-performing model represented in Table 1. A higher topic coherence score indicates better performance of the model.Keywords: big data, latent Dirichlet allocation, latent semantic indexing, telco, topic modeling, twitter
Procedia PDF Downloads 15324599 Enhance the Power of Sentiment Analysis
Authors: Yu Zhang, Pedro Desouza
Abstract:
Since big data has become substantially more accessible and manageable due to the development of powerful tools for dealing with unstructured data, people are eager to mine information from social media resources that could not be handled in the past. Sentiment analysis, as a novel branch of text mining, has in the last decade become increasingly important in marketing analysis, customer risk prediction and other fields. Scientists and researchers have undertaken significant work in creating and improving their sentiment models. In this paper, we present a concept of selecting appropriate classifiers based on the features and qualities of data sources by comparing the performances of five classifiers with three popular social media data sources: Twitter, Amazon Customer Reviews, and Movie Reviews. We introduced a couple of innovative models that outperform traditional sentiment classifiers for these data sources, and provide insights on how to further improve the predictive power of sentiment analysis. The modelling and testing work was done in R and Greenplum in-database analytic tools.Keywords: sentiment analysis, social media, Twitter, Amazon, data mining, machine learning, text mining
Procedia PDF Downloads 35324598 Real-Time Big-Data Warehouse a Next-Generation Enterprise Data Warehouse and Analysis Framework
Authors: Abbas Raza Ali
Abstract:
Big Data technology is gradually becoming a dire need of large enterprises. These enterprises are generating massively large amount of off-line and streaming data in both structured and unstructured formats on daily basis. It is a challenging task to effectively extract useful insights from the large scale datasets, even though sometimes it becomes a technology constraint to manage transactional data history of more than a few months. This paper presents a framework to efficiently manage massively large and complex datasets. The framework has been tested on a communication service provider producing massively large complex streaming data in binary format. The communication industry is bound by the regulators to manage history of their subscribers’ call records where every action of a subscriber generates a record. Also, managing and analyzing transactional data allows service providers to better understand their customers’ behavior, for example, deep packet inspection requires transactional internet usage data to explain internet usage behaviour of the subscribers. However, current relational database systems limit service providers to only maintain history at semantic level which is aggregated at subscriber level. The framework addresses these challenges by leveraging Big Data technology which optimally manages and allows deep analysis of complex datasets. The framework has been applied to offload existing Intelligent Network Mediation and relational Data Warehouse of the service provider on Big Data. The service provider has 50+ million subscriber-base with yearly growth of 7-10%. The end-to-end process takes not more than 10 minutes which involves binary to ASCII decoding of call detail records, stitching of all the interrogations against a call (transformations) and aggregations of all the call records of a subscriber.Keywords: big data, communication service providers, enterprise data warehouse, stream computing, Telco IN Mediation
Procedia PDF Downloads 17724597 Programming with Grammars
Authors: Peter M. Maurer Maurer
Abstract:
DGL is a context free grammar-based tool for generating random data. Many types of simulator input data require some computation to be placed in the proper format. For example, it might be necessary to generate ordered triples in which the third element is the sum of the first two elements, or it might be necessary to generate random numbers in some sorted order. Although DGL is universal in computational power, generating these types of data is extremely difficult. To overcome this problem, we have enhanced DGL to include features that permit direct computation within the structure of a context free grammar. The features have been implemented as special types of productions, preserving the context free flavor of DGL specifications.Keywords: DGL, Enhanced Context Free Grammars, Programming Constructs, Random Data Generation
Procedia PDF Downloads 15024596 A Model Architecture Transformation with Approach by Modeling: From UML to Multidimensional Schemas of Data Warehouses
Authors: Ouzayr Rabhi, Ibtissam Arrassen
Abstract:
To provide a complete analysis of the organization and to help decision-making, leaders need to have relevant data; Data Warehouses (DW) are designed to meet such needs. However, designing DW is not trivial and there is no formal method to derive a multidimensional schema from heterogeneous databases. In this article, we present a Model-Driven based approach concerning the design of data warehouses. We describe a multidimensional meta-model and also specify a set of transformations starting from a Unified Modeling Language (UML) metamodel. In this approach, the UML metamodel and the multidimensional one are both considered as a platform-independent model (PIM). The first meta-model is mapped into the second one through transformation rules carried out by the Query View Transformation (QVT) language. This proposal is validated through the application of our approach to generating a multidimensional schema of a Balanced Scorecard (BSC) DW. We are interested in the BSC perspectives, which are highly linked to the vision and the strategies of an organization.Keywords: data warehouse, meta-model, model-driven architecture, transformation, UML
Procedia PDF Downloads 16124595 Overlaps and Intersections: An Alternative Look at Choreography
Authors: Ashlie Latiolais
Abstract:
Architecture, as a discipline, is on a trajectory of extension beyond the boundaries of buildings and, more increasingly, is coupled with research that connects to alternative and typically disjointed disciplines. A “both/and” approach and (expanded) definition of architecture, as depicted here, expands the margins that contain the profession. Figuratively, architecture is a series of edges, events, and occurrences that establishes a choreography or stage by which humanity exists. The way in which architecture controls and suggests the movement through these spaces, being within a landscape, city, or building, can be viewed as a datum by which the “dance” of everyday life occurs. This submission views the realm of architecture through the lens of movement and dance as a cross-fertilizer of collaboration, tectonic, and spatial geometry investigations. “Designing on digital programs puts architects at a distance from the spaces they imagine. While this has obvious advantages, it also means that they lose the lived, embodied experience of feeling what is needed in space—meaning that some design ideas that work in theory ultimately fail in practice.” By studying the body in motion through real-time performance, a more holistic understanding of architectural space surfaces and new prospects for theoretical teaching pedagogies emerge. The atypical intersection rethinks how architecture is considered, created, and tested, similar to how “dance artists often do this by thinking through the body, opening pathways and possibilities that might not otherwise be accessible” –this is the essence of this poster submission as explained through unFOLDED, a creative performance work. A new languageismaterialized through unFOLDED, a dynamic occupiable installation by which architecture is investigated through dance, movement, and body analysis. The entry unfolds a collaboration of an architect, dance choreographer, musicians, video artist, and lighting designers to re-create one of the first documented avant-garde performing arts collaborations (Matisse, Satie, Massine, Picasso) from the Ballet Russes in 1917, entitled Parade. Architecturally, this interdisciplinary project orients and suggests motion through structure, tectonic, lightness, darkness, and shadow as it questions the navigation of the dark space (stage) surrounding the installation. Artificial light via theatrical lighting and video graphics brought the blank canvas to life – where the sensitive mix of musicality coordinated with the structure’s movement sequencing was certainly a challenge. The upstage light from the video projections created both flickered contextual imagery and shadowed figures. When the dancers were either upstage or downstage of the structure, both silhouetted figures and revealed bodies are experienced as dancer-controlled installation manipulations occurred throughout the performance. The experimental performance, through structure, prompted moving (dancing) bodies in space, where the architecture served as a key component to the choreography itself. The tectonic of the delicate steel structure allowed for the dancers to interact with the installation, which created a variety of spatial conditions – the contained box of three-dimensional space, to a wall, and various abstracted geometries in between. The development of this research unveils the new role of an Architect as a Choreographer of the built environment.Keywords: dance, architecture, choreography, installation, architect, choreographer, space
Procedia PDF Downloads 9324594 Secured Embedding of Patient’s Confidential Data in Electrocardiogram Using Chaotic Maps
Authors: Butta Singh
Abstract:
This paper presents a chaotic map based approach for secured embedding of patient’s confidential data in electrocardiogram (ECG) signal. The chaotic map generates predefined locations through the use of selective control parameters. The sample value difference method effectually hides the confidential data in ECG sample pairs at these predefined locations. Evaluation of proposed method on all 48 records of MIT-BIH arrhythmia ECG database demonstrates that the embedding does not alter the diagnostic features of cover ECG. The secret data imperceptibility in stego-ECG is evident through various statistical and clinical performance measures. Statistical metrics comprise of Percentage Root Mean Square Difference (PRD) and Peak Signal to Noise Ratio (PSNR). Further, a comparative analysis between proposed method and existing approaches was also performed. The results clearly demonstrated the superiority of proposed method.Keywords: chaotic maps, ECG steganography, data embedding, electrocardiogram
Procedia PDF Downloads 19924593 Small Scale Mobile Robot Auto-Parking Using Deep Learning, Image Processing, and Kinematics-Based Target Prediction
Authors: Mingxin Li, Liya Ni
Abstract:
Autonomous parking is a valuable feature applicable to many robotics applications such as tour guide robots, UV sanitizing robots, food delivery robots, and warehouse robots. With auto-parking, the robot will be able to park at the charging zone and charge itself without human intervention. As compared to self-driving vehicles, auto-parking is more challenging for a small-scale mobile robot only equipped with a front camera due to the camera view limited by the robot’s height and the narrow Field of View (FOV) of the inexpensive camera. In this research, auto-parking of a small-scale mobile robot with a front camera only was achieved in a four-step process: Firstly, transfer learning was performed on the AlexNet, a popular pre-trained convolutional neural network (CNN). It was trained with 150 pictures of empty parking slots and 150 pictures of occupied parking slots from the view angle of a small-scale robot. The dataset of images was divided into a group of 70% images for training and the remaining 30% images for validation. An average success rate of 95% was achieved. Secondly, the image of detected empty parking space was processed with edge detection followed by the computation of parametric representations of the boundary lines using the Hough Transform algorithm. Thirdly, the positions of the entrance point and center of available parking space were predicted based on the robot kinematic model as the robot was driving closer to the parking space because the boundary lines disappeared partially or completely from its camera view due to the height and FOV limitations. The robot used its wheel speeds to compute the positions of the parking space with respect to its changing local frame as it moved along, based on its kinematic model. Lastly, the predicted entrance point of the parking space was used as the reference for the motion control of the robot until it was replaced by the actual center when it became visible again by the robot. The linear and angular velocities of the robot chassis center were computed based on the error between the current chassis center and the reference point. Then the left and right wheel speeds were obtained using inverse kinematics and sent to the motor driver. The above-mentioned four subtasks were all successfully accomplished, with the transformed learning, image processing, and target prediction performed in MATLAB, while the motion control and image capture conducted on a self-built small scale differential drive mobile robot. The small-scale robot employs a Raspberry Pi board, a Pi camera, an L298N dual H-bridge motor driver, a USB power module, a power bank, four wheels, and a chassis. Future research includes three areas: the integration of all four subsystems into one hardware/software platform with the upgrade to an Nvidia Jetson Nano board that provides superior performance for deep learning and image processing; more testing and validation on the identification of available parking space and its boundary lines; improvement of performance after the hardware/software integration is completed.Keywords: autonomous parking, convolutional neural network, image processing, kinematics-based prediction, transfer learning
Procedia PDF Downloads 13324592 System for Electromyography Signal Emulation Through the Use of Embedded Systems
Authors: Valentina Narvaez Gaitan, Laura Valentina Rodriguez Leguizamon, Ruben Dario Hernandez B.
Abstract:
This work describes a physiological signal emulation system that uses electromyography (EMG) signals obtained from muscle sensors in the first instance. These signals are used to extract their characteristics to model and emulate specific arm movements. The main objective of this effort is to develop a new biomedical software system capable of generating physiological signals through the use of embedded systems by establishing the characteristics of the acquired signals. The acquisition system used was Biosignals, which contains two EMG electrodes used to acquire signals from the forearm muscles placed on the extensor and flexor muscles. Processing algorithms were implemented to classify the signals generated by the arm muscles when performing specific movements such as wrist flexion extension, palmar grip, and wrist pronation-supination. Matlab software was used to condition and preprocess the signals for subsequent classification. Subsequently, the mathematical modeling of each signal is performed to be generated by the embedded system, with a validation of the accuracy of the obtained signal using the percentage of cross-correlation, obtaining a precision of 96%. The equations are then discretized to be emulated in the embedded system, obtaining a system capable of generating physiological signals according to the characteristics of medical analysis.Keywords: classification, electromyography, embedded system, emulation, physiological signals
Procedia PDF Downloads 11324591 MHD Non-Newtonian Nanofluid Flow over a Permeable Stretching Sheet with Heat Generation and Velocity Slip
Authors: Rama Bhargava, Mania Goyal
Abstract:
The problem of magnetohydrodynamics boundary layer flow and heat transfer on a permeable stretching surface in a second grade nanofluid under the effect of heat generation and partial slip is studied theoretically. The Brownian motion and thermophoresis effects are also considered. The boundary layer equations governed by the PDE’s are transformed into a set of ODE’s with the help of local similarity transformations. The differential equations are solved by variational finite element method. The effects of different controlling parameters on the flow field and heat transfer characteristics are examined. The numerical results for the dimensionless velocity, temperature and nanoparticle volume fraction as well as the reduced Nusselt and Sherwood number have been presented graphically. The comparison confirmed excellent agreement. The present study is of great interest in coating and suspensions, cooling of metallic plate, oils and grease, paper production, coal water or coal-oil slurries, heat exchangers technology, materials processing exploiting.Keywords: viscoelastic nanofluid, partial slip, stretching sheet, heat generation/absorption, MHD flow, FEM
Procedia PDF Downloads 31424590 Detection Efficient Enterprises via Data Envelopment Analysis
Authors: S. Turkan
Abstract:
In this paper, the Turkey’s Top 500 Industrial Enterprises data in 2014 were analyzed by data envelopment analysis. Data envelopment analysis is used to detect efficient decision-making units such as universities, hospitals, schools etc. by using inputs and outputs. The decision-making units in this study are enterprises. To detect efficient enterprises, some financial ratios are determined as inputs and outputs. For this reason, financial indicators related to productivity of enterprises are considered. The efficient foreign weighted owned capital enterprises are detected via super efficiency model. According to the results, it is said that Mercedes-Benz is the most efficient foreign weighted owned capital enterprise in Turkey.Keywords: data envelopment analysis, super efficiency, logistic regression, financial ratios
Procedia PDF Downloads 32824589 Intelligent Process Data Mining for Monitoring for Fault-Free Operation of Industrial Processes
Authors: Hyun-Woo Cho
Abstract:
The real-time fault monitoring and diagnosis of large scale production processes is helpful and necessary in order to operate industrial process safely and efficiently producing good final product quality. Unusual and abnormal events of the process may have a serious impact on the process such as malfunctions or breakdowns. This work try to utilize process measurement data obtained in an on-line basis for the safe and some fault-free operation of industrial processes. To this end, this work evaluated the proposed intelligent process data monitoring framework based on a simulation process. The monitoring scheme extracts the fault pattern in the reduced space for the reliable data representation. Moreover, this work shows the results of using linear and nonlinear techniques for the monitoring purpose. It has shown that the nonlinear technique produced more reliable monitoring results and outperforms linear methods. The adoption of the qualitative monitoring model helps to reduce the sensitivity of the fault pattern to noise.Keywords: process data, data mining, process operation, real-time monitoring
Procedia PDF Downloads 64124588 Combined Model Predictive Controller Technique for Enhancing NAO Gait Stabilization
Authors: Brahim Brahmi, Mohammed Hamza Laraki, Mohammad Habibur Rahman, Islam M. Rasedul, M. Assad Uz-Zaman
Abstract:
The humanoid robot, specifically the NAO robot must be able to provide a highly dynamic performance on the soccer field. Maintaining the balance of the humanoid robot during the required motion is considered as one of a challenging problems especially when the robot is subject to external disturbances, as contact with other robots. In this paper, a dynamic controller is proposed in order to ensure a robust walking (stabilization) and to improve the dynamic balance of the robot during its contact with the environment (external disturbances). The generation of the trajectory of the center of mass (CoM) is done by a model predictive controller (MPC) conjoined with zero moment point (ZMP) technique. Taking into account the properties of the rotational dynamics of the whole-body system, a modified previous control mixed with feedback control is employed to manage the angular momentum and the CoM’s acceleration, respectively. This latter is dedicated to provide a robust gait of the robot in the presence of the external disturbances. Simulation results are presented to show the feasibility of the proposed strategy.Keywords: preview control, Nao robot, model predictive control
Procedia PDF Downloads 12824587 Numerical Study of Sloshing in a Flexible Tank
Authors: Wissem Tighidet, Faïçal Naït Bouda, Moussa Allouche
Abstract:
The numerical study of the Fluid-Structure Interaction (FSI) in a partially filled flexible tank submitted to a horizontal harmonic excitation motion. It is investigated by using two-way Fluid-Structure Interaction (FSI) in a flexible tank by Coupling between the Transient Structural (Mechanical) and Fluid Flow (Fluent) in ANSYS-Workbench Student version. The Arbitrary Lagrangian-Eulerian (ALE) formulation is adopted to solve with the finite volume method, the Navier-Stokes equations in two phases in a moving domain. The Volume of Fluid (VOF) method is applied to track the free surface. However, the equations of the dynamics of the structure are solved with the finite element method assuming a linear elastic behavior. To conclude, the Fluid-Structure Interaction (IFS) has a vital role in the analysis of the dynamic behavior of the rectangular tank. The results indicate that the flexibility of the tank walls has a significant impact on the amplitude of tank sloshing and the deformation of the free surface as well as the effect of liquid sloshing on wall deformation.Keywords: arbitrary lagrangian-eulerian, fluid-structure interaction, sloshing, volume of fluid
Procedia PDF Downloads 10724586 Learners as Consultants: Knowledge Acquisition and Client Organisations-A Student as Producer Case Study
Authors: Barry Ardley, Abi Hunt, Nick Taylor
Abstract:
As a theoretical and practical framework, this study uses the student-as-producer approach to learning in higher education, as adopted by the Lincoln International Business School, University of Lincoln, UK. Students as producer positions learners as skilled and capable agents, able to participate as partners with tutors in live research projects. To illuminate the nature of this approach to learning and to highlight its critical issues, the authors report on two guided student consultancy projects. These were set up with the assistance of two local organisations in the city of Lincoln, UK. Using the student as a producer model to deliver the projects enabled learners to acquire and develop a range of key skills and knowledge not easily accessible in more traditional educational settings. This paper presents a systematic case study analysis of the eight organising principles of the student-as-producer model, as adopted by university tutors. The experience of tutors implementing students as producers suggests that the model can be widely applied to benefit not only the learning and teaching experiences of higher education students and staff but additionally a university’s research programme and its community partners.Keywords: consultancy, learning, student as producer, research
Procedia PDF Downloads 8024585 Effects of X and + Tail-Body Configurations on Hydrodynamic Performance and Stability of an Underwater Vehicle
Authors: Kadri Koçer, Sezer Kefeli
Abstract:
This paper proposes a comparison of hydrodynamic performance and stability characteristic for an underwater vehicle which has two type of tail design, namely X and +tail-body configurations. The effects of these configurations on the underwater vehicle’s hydrodynamic performance and maneuvering characteristic will be investigated comprehensively. Hydrodynamic damping coefficients for modeling the motion of the underwater vehicles will be predicted. Additionally, forces and moments due to control surfaces will be compared using computational fluid dynamics methods. In the aviation, the X tail-body configuration is widely used for high maneuverability requirements. However, in the underwater, the + tail-body configuration is more commonly used than the X tail-body configuration for its stability characteristics. Thus it is important to see the effect and differences of the tail designs in the underwater world. For CFD analysis, the incompressible, three-dimensional, and steady Navier-Stokes equations will be used to simulate the flows. Also, k-ε Realizable turbulence model with enhanced wall treatment will be taken. Numerical results is verified with experimental results for verification. The overall goal of this study is to present the advantages and disadvantages of hydrodynamic performance and stability characteristic for X and + tail-body configurations of the underwater vehicle.Keywords: maneuverability, stability, CFD, tail configuration, hydrodynamic design
Procedia PDF Downloads 18724584 A Soft System Approach to Explore Ill-Defined Issues in Distance Education System - A Case of Saudi Arabia
Authors: Sulafah Basahel
Abstract:
Nowadays, Higher Education Institutions (HEIs) around the world are attempting to utilize Information and Communication Technologies (ICTs) to enhance learning process and strategies of knowledge delivery for students through Distance Education (DE) system. Stakeholders in DE system face a complex situation of different ill-defined and related issues that influence decision making process. In this study system thinking as a body of knowledge is used to explore the emergent properties that produced from these connections between issues and could have either positive or negative outcomes for the DE development. Checkland Soft System Methodology (SSM) - Mode 2 is employed in a cultural context of Saudi Arabia for more knowledge acquisition purposes among multiple stakeholders in DE rather than solving problems to achieve an overall development of DE system. This paper will discuss some political, cultural issues and connections between them that impact on effectiveness of stakeholders’ activities and relations. This study will significantly contribute to both system thinking and education fields by leading decision makers in DE to reconsider future plans, strategies and right actions for more successful educational practices.Keywords: distance education, higher education institutions, ill-defined issues, soft system methodology-Mode 2
Procedia PDF Downloads 27024583 Statistically Accurate Synthetic Data Generation for Enhanced Traffic Predictive Modeling Using Generative Adversarial Networks and Long Short-Term Memory
Authors: Srinivas Peri, Siva Abhishek Sirivella, Tejaswini Kallakuri, Uzair Ahmad
Abstract:
Effective traffic management and infrastructure planning are crucial for the development of smart cities and intelligent transportation systems. This study addresses the challenge of data scarcity by generating realistic synthetic traffic data using the PeMS-Bay dataset, improving the accuracy and reliability of predictive modeling. Advanced synthetic data generation techniques, including TimeGAN, GaussianCopula, and PAR Synthesizer, are employed to produce synthetic data that replicates the statistical and structural characteristics of real-world traffic. Future integration of Spatial-Temporal Generative Adversarial Networks (ST-GAN) is planned to capture both spatial and temporal correlations, further improving data quality and realism. The performance of each synthetic data generation model is evaluated against real-world data to identify the best models for accurately replicating traffic patterns. Long Short-Term Memory (LSTM) networks are utilized to model and predict complex temporal dependencies within traffic patterns. This comprehensive approach aims to pinpoint areas with low vehicle counts, uncover underlying traffic issues, and inform targeted infrastructure interventions. By combining GAN-based synthetic data generation with LSTM-based traffic modeling, this study supports data-driven decision-making that enhances urban mobility, safety, and the overall efficiency of city planning initiatives.Keywords: GAN, long short-term memory, synthetic data generation, traffic management
Procedia PDF Downloads 2924582 Fold and Thrust Belts Seismic Imaging and Interpretation
Authors: Sunjay
Abstract:
Plate tectonics is of very great significance as it represents the spatial relationships of volcanic rock suites at plate margins, the distribution in space and time of the conditions of different metamorphic facies, the scheme of deformation in mountain belts, or orogens, and the association of different types of economic deposit. Orogenic belts are characterized by extensive thrust faulting, movements along large strike-slip fault zones, and extensional deformation that occur deep within continental interiors. Within oceanic areas there also are regions of crustal extension and accretion in the backarc basins that are located on the landward sides of many destructive plate margins.Collisional orogens develop where a continent or island arc collides with a continental margin as a result of subduction. collisional and noncollisional orogens can be explained by differences in the strength and rheology of the continental lithosphere and by processes that influence these properties during orogenesis.Seismic Imaging Difficulties-In triangle zones, several factors reduce the effectiveness of seismic methods. The topography in the central part of the triangle zone is usually rugged and is associated with near-surface velocity inversions which degrade the quality of the seismic image. These characteristics lead to low signal-to-noise ratio, inadequate penetration of energy through overburden, poor geophone coupling with the surface and wave scattering. Depth Seismic Imaging Techniques-Seismic processing relates to the process of altering the seismic data to suppress noise, enhancing the desired signal (higher signal-to-noise ratio) and migrating seismic events to their appropriate location in space and depth. Processing steps generally include analysis of velocities, static corrections, moveout corrections, stacking and migration. Exploration seismology Bow-tie effect -Shadow Zones-areas with no reflections (dead areas). These are called shadow zones and are common in the vicinity of faults and other discontinuous areas in the subsurface. Shadow zones result when energy from a reflector is focused on receivers that produce other traces. As a result, reflectors are not shown in their true positions. Subsurface Discontinuities-Diffractions occur at discontinuities in the subsurface such as faults and velocity discontinuities (as at “bright spot” terminations). Bow-tie effect caused by the two deep-seated synclines. Seismic imaging of thrust faults and structural damage-deepwater thrust belts, Imaging deformation in submarine thrust belts using seismic attributes,Imaging thrust and fault zones using 3D seismic image processing techniques, Balanced structural cross sections seismic interpretation pitfalls checking, The seismic pitfalls can originate due to any or all of the limitations of data acquisition, processing, interpretation of the subsurface geology,Pitfalls and limitations in seismic attribute interpretation of tectonic features, Seismic attributes are routinely used to accelerate and quantify the interpretation of tectonic features in 3D seismic data. Coherence (or variance) cubes delineate the edges of megablocks and faulted strata, curvature delineates folds and flexures, while spectral components delineate lateral changes in thickness and lithology. Carbon capture and geological storage leakage surveillance because fault behave as a seal or a conduit for hydrocarbon transportation to a trap,etc.Keywords: tectonics, seismic imaging, fold and thrust belts, seismic interpretation
Procedia PDF Downloads 7024581 A Machine Learning Approach for the Leakage Classification in the Hydraulic Final Test
Authors: Christian Neunzig, Simon Fahle, Jürgen Schulz, Matthias Möller, Bernd Kuhlenkötter
Abstract:
The widespread use of machine learning applications in production is significantly accelerated by improved computing power and increasing data availability. Predictive quality enables the assurance of product quality by using machine learning models as a basis for decisions on test results. The use of real Bosch production data based on geometric gauge blocks from machining, mating data from assembly and hydraulic measurement data from final testing of directional valves is a promising approach to classifying the quality characteristics of workpieces.Keywords: machine learning, classification, predictive quality, hydraulics, supervised learning
Procedia PDF Downloads 21524580 Analysis of Cyber Activities of Potential Business Customers Using Neo4j Graph Databases
Authors: Suglo Tohari Luri
Abstract:
Data analysis is an important aspect of business performance. With the application of artificial intelligence within databases, selecting a suitable database engine for an application design is also very crucial for business data analysis. The application of business intelligence (BI) software into some relational databases such as Neo4j has proved highly effective in terms of customer data analysis. Yet what remains of great concern is the fact that not all business organizations have the neo4j business intelligence software applications to implement for customer data analysis. Further, those with the BI software lack personnel with the requisite expertise to use it effectively with the neo4j database. The purpose of this research is to demonstrate how the Neo4j program code alone can be applied for the analysis of e-commerce website customer visits. As the neo4j database engine is optimized for handling and managing data relationships with the capability of building high performance and scalable systems to handle connected data nodes, it will ensure that business owners who advertise their products at websites using neo4j as a database are able to determine the number of visitors so as to know which products are visited at routine intervals for the necessary decision making. It will also help in knowing the best customer segments in relation to specific goods so as to place more emphasis on their advertisement on the said websites.Keywords: data, engine, intelligence, customer, neo4j, database
Procedia PDF Downloads 19424579 Construction of a Dynamic Model of Cerebral Blood Circulation for Future Integrated Control of Brain State
Authors: Tomohiko Utsuki
Abstract:
Currently, brain resuscitation becomes increasingly important due to revising various clinical guidelines pertinent to emergency care. In brain resuscitation, the control of brain temperature (BT), intracranial pressure (ICP), and cerebral blood flow (CBF) is required for stabilizing physiological state of brain, and is described as the essential treatment points in many guidelines of disorder and/or disease such as brain injury, stroke, and encephalopathy. Thus, an integrated control system of BT, ICP, and CBF will greatly contribute to alleviating the burden on medical staff and improving treatment effect in brain resuscitation. In order to develop such a control system, models related to BT, ICP, and CBF are required for control simulation, because trial and error experiments using patients are not ethically allowed. A static model of cerebral blood circulation from intracranial arteries and vertebral artery to jugular veins has already constructed and verified. However, it is impossible to represent the pooling of blood in blood vessels, which is one cause of cerebral hypertension in this model. And, it is also impossible to represent the pulsing motion of blood vessels caused by blood pressure change which can have an affect on the change of cerebral tissue pressure. Thus, a dynamic model of cerebral blood circulation is constructed in consideration of the elasticity of the blood vessel and the inertia of the blood vessel wall. The constructed dynamic model was numerically analyzed using the normal data, in which each arterial blood flow in cerebral blood circulation, the distribution of blood pressure in the Circle of Willis, and the change of blood pressure along blood flow were calculated for verifying against physiological knowledge. As the result, because each calculated numerical value falling within the generally known normal range, this model has no problem in representing at least the normal physiological state of the brain. It is the next task to verify the accuracy of the present model in the case of disease or disorder. Currently, the construction of a migration model of extracellular fluid and a model of heat transfer in cerebral tissue are in progress for making them parts of an integrated model of brain physiological state, which is necessary for developing an future integrated control system of BT, ICP and CBF. The present model is applicable to constructing the integrated model representing at least the normal condition of brain physiological state by uniting with such models.Keywords: dynamic model, cerebral blood circulation, brain resuscitation, automatic control
Procedia PDF Downloads 15424578 Analyzing Electromagnetic and Geometric Characterization of Building Insulation Materials Using the Transient Radar Method (TRM)
Authors: Ali Pourkazemi
Abstract:
The transient radar method (TRM) is one of the non-destructive methods that was introduced by authors a few years ago. The transient radar method can be classified as a wave-based non destructive testing (NDT) method that can be used in a wide frequency range. Nevertheless, it requires a narrow band, ranging from a few GHz to a few THz, depending on the application. As a time-of-flight and real-time method, TRM can measure the electromagnetic properties of the sample under test not only quickly and accurately, but also blindly. This means that it requires no prior knowledge of the sample under test. For multi-layer structures, TRM is not only able to detect changes related to any parameter within the multi-layer structure but can also measure the electromagnetic properties of each layer and its thickness individually. Although the temperature, humidity, and general environmental conditions may affect the sample under test, they do not affect the accuracy of the Blind TRM algorithm. In this paper, the electromagnetic properties as well as the thickness of the individual building insulation materials - as a single-layer structure - are measured experimentally. Finally, the correlation between the reflection coefficients and some other technical parameters such as sound insulation, thermal resistance, thermal conductivity, compressive strength, and density is investigated. The sample to be studied is 30 cm x 50 cm and the thickness of the samples varies from a few millimeters to 6 centimeters. This experiment is performed with both biostatic and differential hardware at 10 GHz. Since it is a narrow-band system, high-speed computation for analysis, free-space application, and real-time sensor, it has a wide range of potential applications, e.g., in the construction industry, rubber industry, piping industry, wind energy industry, automotive industry, biotechnology, food industry, pharmaceuticals, etc. Detection of metallic, plastic pipes wires, etc. through or behind the walls are specific applications for the construction industry.Keywords: transient radar method, blind electromagnetic geometrical parameter extraction technique, ultrafast nondestructive multilayer dielectric structure characterization, electronic measurement systems, illumination, data acquisition performance, submillimeter depth resolution, time-dependent reflected electromagnetic signal blind analysis method, EM signal blind analysis method, time domain reflectometer, microwave, milimeter wave frequencies
Procedia PDF Downloads 6924577 Decision Making System for Clinical Datasets
Authors: P. Bharathiraja
Abstract:
Computer Aided decision making system is used to enhance diagnosis and prognosis of diseases and also to assist clinicians and junior doctors in clinical decision making. Medical Data used for decision making should be definite and consistent. Data Mining and soft computing techniques are used for cleaning the data and for incorporating human reasoning in decision making systems. Fuzzy rule based inference technique can be used for classification in order to incorporate human reasoning in the decision making process. In this work, missing values are imputed using the mean or mode of the attribute. The data are normalized using min-ma normalization to improve the design and efficiency of the fuzzy inference system. The fuzzy inference system is used to handle the uncertainties that exist in the medical data. Equal-width-partitioning is used to partition the attribute values into appropriate fuzzy intervals. Fuzzy rules are generated using Class Based Associative rule mining algorithm. The system is trained and tested using heart disease data set from the University of California at Irvine (UCI) Machine Learning Repository. The data was split using a hold out approach into training and testing data. From the experimental results it can be inferred that classification using fuzzy inference system performs better than trivial IF-THEN rule based classification approaches. Furthermore it is observed that the use of fuzzy logic and fuzzy inference mechanism handles uncertainty and also resembles human decision making. The system can be used in the absence of a clinical expert to assist junior doctors and clinicians in clinical decision making.Keywords: decision making, data mining, normalization, fuzzy rule, classification
Procedia PDF Downloads 51924576 Estimating Bridge Deterioration for Small Data Sets Using Regression and Markov Models
Authors: Yina F. Muñoz, Alexander Paz, Hanns De La Fuente-Mella, Joaquin V. Fariña, Guilherme M. Sales
Abstract:
The primary approach for estimating bridge deterioration uses Markov-chain models and regression analysis. Traditional Markov models have problems in estimating the required transition probabilities when a small sample size is used. Often, reliable bridge data have not been taken over large periods, thus large data sets may not be available. This study presents an important change to the traditional approach by using the Small Data Method to estimate transition probabilities. The results illustrate that the Small Data Method and traditional approach both provide similar estimates; however, the former method provides results that are more conservative. That is, Small Data Method provided slightly lower than expected bridge condition ratings compared with the traditional approach. Considering that bridges are critical infrastructures, the Small Data Method, which uses more information and provides more conservative estimates, may be more appropriate when the available sample size is small. In addition, regression analysis was used to calculate bridge deterioration. Condition ratings were determined for bridge groups, and the best regression model was selected for each group. The results obtained were very similar to those obtained when using Markov chains; however, it is desirable to use more data for better results.Keywords: concrete bridges, deterioration, Markov chains, probability matrix
Procedia PDF Downloads 33724575 Enhancing Human Mobility Exoskeleton Comfort Using Admittance Controller
Authors: Alexandre Rabaseda, Emelie Seguin, Marc Doumit
Abstract:
Human mobility exoskeletons have been in development for several years and are becoming increasingly efficient. Unfortunately, user comfort was not always a priority design criterion throughout their development. To further improve this technology, exoskeletons should operate and deliver assistance without causing discomfort to the user. For this, improvements are necessary from an ergonomic point of view. The device’s control method is important when endeavoring to enhance user comfort. Exoskeleton or rehabilitation device controllers use methods of control called interaction controls (admittance and impedance controls). This paper proposes an extended version of an admittance controller to enhance user comfort. The control method used consists of adding an inner loop that is controlled by a proportional-integral-derivative (PID) controller. This allows the interaction force to be kept as close as possible to the desired force trajectory. The force-tracking admittance controller modifies the actuation force of the system in order to follow both the desired motion trajectory and the desired relative force between the user and the exoskeleton.Keywords: mobility assistive device, exoskeleton, force-tracking admittance controller, user comfort
Procedia PDF Downloads 15724574 Using Genre Analysis to Teach Contract Negotiation Discourse Practices
Authors: Anthony Townley
Abstract:
Contract negotiation is fundamental to commercial law practice. For this study, genre and discourse analytical methodology was used to examine the legal negotiation of a Merger & Acquisition (M&A) deal undertaken by legal and business professionals in English across different jurisdictions in Europe. While some of the most delicate negotiations involved in this process were carried on face-to-face or over the telephone, these were generally progressed more systematically – and on the record – in the form of emails, email attachments, and as comments and amendments recorded in successive ‘marked-up’ versions of the contracts under negotiation. This large corpus of textual data was originally obtained by the author, in 2012, for the purpose of doctoral research. For this study, the analysis is particularly concerned with the use of emails and covering letters to exchange legal advice about the negotiations. These two genres help to stabilize and progress the negotiation process and account for negotiation activities. Swalesian analysis of functional Moves and Steps was able to identify structural similarities and differences between these text types and to identify certain salient discursive features within them. The analytical findings also indicate how particular linguistic strategies are more appropriately and more effectively associated with one legal genre rather than another. The concept of intertextuality is an important dimension of contract negotiation discourse and this study also examined how the discursive relationships between the different texts influence the way that texts are constructed. In terms of materials development, the research findings can contribute to more authentic English for Legal & Business Purposes pedagogies for students and novice lawyers and business professionals. The findings can first be used to design discursive maps that provide learners with a coherent account of the intertextual nature of the contract negotiation process. These discursive maps can then function as a framework in which to present detailed findings about the textual and structural features of the text types by applying the Swalesian genre analysis. Based on this acquired knowledge of the textual nature of contract negotiation, the authentic discourse materials can then be used to provide learners with practical opportunities to role-play negotiation activities and experience professional ways of thinking and using language in preparation for the written discourse challenges they will face in this important area of legal and business practice.Keywords: English for legal and business purposes, discourse analysis, genre analysis, intertextuality, pedagogical materials
Procedia PDF Downloads 15024573 Validation of Visibility Data from Road Weather Information Systems by Comparing Three Data Resources: Case Study in Ohio
Authors: Fan Ye
Abstract:
Adverse weather conditions, particularly those with low visibility, are critical to the driving tasks. However, the direct relationship between visibility distances and traffic flow/roadway safety is uncertain due to the limitation of visibility data availability. The recent growth of deployment of Road Weather Information Systems (RWIS) makes segment-specific visibility information available which can be integrated with other Intelligent Transportation System, such as automated warning system and variable speed limit, to improve mobility and safety. Before applying the RWIS visibility measurements in traffic study and operations, it is critical to validate the data. Therefore, an attempt was made in the paper to examine the validity and viability of RWIS visibility data by comparing visibility measurements among RWIS, airport weather stations, and weather information recorded by police in crash reports, based on Ohio data. The results indicated that RWIS visibility measurements were significantly different from airport visibility data in Ohio, but no conclusion regarding the reliability of RWIS visibility could be drawn in the consideration of no verified ground truth in the comparisons. It was suggested that more objective methods are needed to validate the RWIS visibility measurements, such as continuous in-field measurements associated with various weather events using calibrated visibility sensors.Keywords: RWIS, visibility distance, low visibility, adverse weather
Procedia PDF Downloads 252