Search results for: Data Centric Approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11074

Search results for: Data Centric Approach

8434 Impact of Hard Limited Clipping Crest Factor Reduction Technique on Bit Error Rate in OFDM Based Systems

Authors: Theodore Grosch, Felipe Koji Godinho Hoshino

Abstract:

In wireless communications, 3GPP LTE is one of the solutions to meet the greater transmission data rate demand. One issue inherent to this technology is the PAPR (Peak-to-Average Power Ratio) of OFDM (Orthogonal Frequency Division Multiplexing) modulation. This high PAPR affects the efficiency of power amplifiers. One approach to mitigate this effect is the Crest Factor Reduction (CFR) technique. In this work, we simulate the impact of Hard Limited Clipping Crest Factor Reduction technique on BER (Bit Error Rate) in OFDM based Systems. In general, the results showed that CFR has more effects on higher digital modulation schemes, as expected. More importantly, we show the worst-case degradation due to CFR on QPSK, 16QAM, and 64QAM signals in a linear system. For example, hard clipping of 9 dB results in a 2 dB increase in signal to noise energy at a 1% BER for 64-QAM modulation.

Keywords: Bit error rate, crest factor reduction, OFDM, physical layer simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2140
8433 The Sequential Estimation of the Seismoacoustic Source Energy in C-OTDR Monitoring Systems

Authors: Andrey V. Timofeev, Dmitry V. Egorov

Abstract:

The practical efficient approach is suggested for estimation of the seismoacoustic sources energy in C-OTDR monitoring systems. This approach is represents the sequential plan for confidence estimation both the seismoacoustic sources energy, as well the absorption coefficient of the soil. The sequential plan delivers the non-asymptotic guaranteed accuracy of obtained estimates in the form of non-asymptotic confidence regions with prescribed sizes. These confidence regions are valid for a finite sample size when the distributions of the observations are unknown. Thus, suggested estimates are non-asymptotic and nonparametric, and also these estimates guarantee the prescribed estimation accuracy in form of prior prescribed size of confidence regions, and prescribed confidence coefficient value.

Keywords: C-OTDR-system, guaranteed estimates, nonparametric estimation, sequential confidence estimation, multichannel monitoring systems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2099
8432 Classroom Teacher Candidates' Definitions and Beliefs about Technology Integration

Authors: Ahmet Baytak, Cenk Akbıyık

Abstract:

The purpose of this paper is to present teacher candidates- beliefs about technology integration in their field of study, which is classroom teaching in this case. The study was conducted among the first year students in college of education in Turkey. This study is based on both quantitative and qualitative data. For the quantitative data- Likert scale was used and for the qualitative data pattern matching was employed. The primary findings showed that students defined educational technology as technologies that improve learning with their visual, easily accessible, and productive features. They also believe these technologies could affect their future students- learning positively.

Keywords: Educational technology, classroom teacher candidates, technology integration, teacher education.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1926
8431 A Quality Optimization Approach: An Application on Next Generation Networks

Authors: Gülfem I. Alptekin, S. Emre Alptekin

Abstract:

The next generation wireless systems, especially the cognitive radio networks aim at utilizing network resources more efficiently. They share a wide range of available spectrum in an opportunistic manner. In this paper, we propose a quality management model for short-term sub-lease of unutilized spectrum bands to different service providers. We built our model on competitive secondary market architecture. To establish the necessary conditions for convergent behavior, we utilize techniques from game theory. Our proposed model is based on potential game approach that is suitable for systems with dynamic decision making. The Nash equilibrium point tells the spectrum holders the ideal price values where profit is maximized at the highest level of customer satisfaction. Our numerical results show that the price decisions of the network providers depend on the price and QoS of their own bands as well as the prices and QoS levels of their opponents- bands.

Keywords: cognitive radio networks, game theory, nextgeneration wireless networks, spectrum management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1512
8430 TTCN-3 Based Conformance Testing of a Node Monitoring Protocol for MANETs

Authors: Mallikarjun B. Channappagoudar, Pallapa Venkataram

Abstract:

As a node monitoring protocol, which is a part of network management, operates in distributed manner, conformance testing of such protocols is more tedious than testing a peer-to-peer protocol. Various works carried out to give the methodology to do conformance testing of distributed protocol. In this paper, we have presented a formal approach for conformance testing of a Node Monitoring Protocol, which uses both static and mobile agents, for MANETs. First, we use SDL to obtain MSCs, which represent the scenario descriptions by sequence diagrams, which in turn generate test sequences and test cases. Later, Testing and Test Control Notation Version-3 (TTCN-3) is used to execute test cases with respect to generated test sequences to know the conformance of protocol against the given specification. This approach shows, the effective conformance testing of the distributed protocols for the network with varying node density and complex behavior. Experimental results for the protocol scenario represent the effectiveness of the method used.

Keywords: Conformance Testing, FSM, Mobile agent, TTCN, Test sequence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2332
8429 Damping Mechanism in Welded Structures

Authors: B.Singh, B.K.Nanda

Abstract:

Response surface methodology with Box–Benhken (BB) design of experiment approach has been utilized to study the mechanism of interface slip damping in layered and jointed tack welded beams with varying surface roughness. The design utilizes the initial amplitude of excitation, tack length and surface roughness at the interfaces to develop the model for the logarithmic damping decrement of the layered and jointed welded structures. Statistically designed experiments have been performed to estimate the coefficients in the mathematical model, predict the response, and check the adequacy of the model. Comparison of predicted and experimental response values outside the design conditions have shown good correspondence, implying that empirical model derived from response surface approach can be effectively used to describe the mechanism of interface slip damping in layered and jointed tack welded structures.

Keywords: Interface slip damping, welded joint, surface roughness, amplitude, tack length, response surface methodology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1830
8428 Reconfigurable Autonomous Mini Robot Design using CPLD's

Authors: Aditya K, Dinesh P, Ramesh Bhakthavatchalu

Abstract:

This paper explains a project based learning method where autonomous mini-robots are developed for research, education and entertainment purposes. In case of remote systems wireless sensors are developed in critical areas, which would collect data at specific time intervals, send the data to the central wireless node based on certain preferred information would make decisions to turn on or off a switch or control unit. Such information transfers hardly sums up to a few bytes and hence low data rates would suffice for such implementations. As a robot is a multidisciplinary platform, the interfacing issues involved are discussed in this paper. The paper is mainly focused on power supply, grounding and decoupling issues.

Keywords: CPLD, power supply, decoupling, grounding.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1912
8427 Structuring and Visualizing Healthcare Claims Data Using Systems Architecture Methodology

Authors: Inas S. Khayal, Weiping Zhou, Jonathan Skinner

Abstract:

Healthcare delivery systems around the world are in crisis. The need to improve health outcomes while decreasing healthcare costs have led to an imminent call to action to transform the healthcare delivery system. While Bioinformatics and Biomedical Engineering have primarily focused on biological level data and biomedical technology, there is clear evidence of the importance of the delivery of care on patient outcomes. Classic singular decomposition approaches from reductionist science are not capable of explaining complex systems. Approaches and methods from systems science and systems engineering are utilized to structure healthcare delivery system data. Specifically, systems architecture is used to develop a multi-scale and multi-dimensional characterization of the healthcare delivery system, defined here as the Healthcare Delivery System Knowledge Base. This paper is the first to contribute a new method of structuring and visualizing a multi-dimensional and multi-scale healthcare delivery system using systems architecture in order to better understand healthcare delivery.

Keywords: Health informatics, systems thinking, systems architecture, healthcare delivery system, data analytics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1138
8426 Emotion Classification using Adaptive SVMs

Authors: P. Visutsak

Abstract:

The study of the interaction between humans and computers has been emerging during the last few years. This interaction will be more powerful if computers are able to perceive and respond to human nonverbal communication such as emotions. In this study, we present the image-based approach to emotion classification through lower facial expression. We employ a set of feature points in the lower face image according to the particular face model used and consider their motion across each emotive expression of images. The vector of displacements of all feature points input to the Adaptive Support Vector Machines (A-SVMs) classifier that classify it into seven basic emotions scheme, namely neutral, angry, disgust, fear, happy, sad and surprise. The system was tested on the Japanese Female Facial Expression (JAFFE) dataset of frontal view facial expressions [7]. Our experiments on emotion classification through lower facial expressions demonstrate the robustness of Adaptive SVM classifier and verify the high efficiency of our approach.

Keywords: emotion classification, facial expression, adaptive support vector machines, facial expression classifier.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2224
8425 The Development of the Multi-Agent Classification System (MACS) in Compliance with FIPA Specifications

Authors: Mohamed R. Mhereeg

Abstract:

The paper investigates the feasibility of constructing a software multi-agent based monitoring and classification system and utilizing it to provide an automated and accurate classification of end users developing applications in the spreadsheet domain. The agents function autonomously to provide continuous and periodic monitoring of excels spreadsheet workbooks. Resulting in, the development of the MultiAgent classification System (MACS) that is in compliance with the specifications of the Foundation for Intelligent Physical Agents (FIPA). However, different technologies have been brought together to build MACS. The strength of the system is the integration of the agent technology with the FIPA specifications together with other technologies that are Windows Communication Foundation (WCF) services, Service Oriented Architecture (SOA), and Oracle Data Mining (ODM). The Microsoft's .NET widows service based agents were utilized to develop the monitoring agents of MACS, the .NET WCF services together with SOA approach allowed the distribution and communication between agents over the WWW that is in order to satisfy the monitoring and classification of the multiple developer aspect. ODM was used to automate the classification phase of MACS.

Keywords: Autonomous, Classification, MACS, Multi-Agent, SOA, WCF.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1589
8424 Multiple Criteria Decision Making for Turkish Air Force Stealth Fighter Aircraft Selection

Authors: C. Ardil

Abstract:

Neutrosophic logic decision analysis is proposed as a method of stealth fighter aircraft selection for Turkish Air Force. The opinion of experts is employed to rank the alternatives across a set of criteria. The analyst uses neutrosophic logic numbers to describe the experts' preferences. This approach can handle the situation in the case of unavailability of precise data, which is most commonly the case in stealth fighter aircraft selection. Neutrosophic logic numbers can consider the imprecision of the factors affecting decision making such as stealth analysis, survivability analysis, and performance analysis. Neutrosophic logic ranking is achieved using weighted arithmetic operator and weighted geometric operator and the alternatives are ranked from best to worst. An example is also presented to illustrate the applicability and effectiveness of the proposed method. 

Keywords: Neutrosophic set theory, stealth fighter aircraft selection, multiple criteria decision-making, neutrosophic logic decision making, Turkish Air Force, MCDM

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 497
8423 Mobile Robot Path Planning in a 2-Dimentional Mesh

Authors: Doraid Dalalah

Abstract:

A topologically oriented neural network is very efficient for real-time path planning for a mobile robot in changing environments. When using a recurrent neural network for this purpose and with the combination of the partial differential equation of heat transfer and the distributed potential concept of the network, the problem of obstacle avoidance of trajectory planning for a moving robot can be efficiently solved. The related dimensional network represents the state variables and the topology of the robot's working space. In this paper two approaches to problem solution are proposed. The first approach relies on the potential distribution of attraction distributed around the moving target, acting as a unique local extreme in the net, with the gradient of the state variables directing the current flow toward the source of the potential heat. The second approach considers two attractive and repulsive potential sources to decrease the time of potential distribution. Computer simulations have been carried out to interrogate the performance of the proposed approaches.

Keywords: Mobile robot, Path Planning, Mesh, Potential field.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1926
8422 Modeling and Simulating Human Arm Movement Using a 2 Dimensional 3 Segments Coupled Pendulum System

Authors: Loay A. Al-Zu'be, Asma A. Al-Tamimi, Thakir D. Al-Momani, Ayat J. Alkarala, Maryam A. Alzawahreh

Abstract:

A two dimensional three segments coupled pendulum system that mathematically models human arm configuration was developed along with constructing and solving the equations of motions for this model using the energy (work) based approach of Lagrange. The equations of motion of the model were solved iteratively both as an initial value problem and as a two point boundary value problem. In the initial value problem solutions, both the initial system configuration (segment angles) and initial system velocity (segment angular velocities) were used as inputs, whereas, in the two point boundary value problem solutions initial and final configurations and time were used as inputs to solve for the trajectory of motion. The results suggest that the model solutions are sensitive to small changes in the dynamic forces applied to the system as well as to the initial and boundary conditions used. To overcome the system sensitivity a new approach is suggested.

Keywords: Body Configurations, Equations of Motion, Mathematical Modeling, Movement Trajectories.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2157
8421 Risk Factors of Becoming NEET Youth in Iran: A Machine Learning Approach

Authors: Hamed Rahmani, Wim Groot

Abstract:

The term "youth not in employment, education or training (NEET)" refers to a combination of youth unemployment and school dropout. This study investigates the variables that increase the risk of becoming NEET in Iran. A selection bias-adjusted Probit model was employed using machine learning to identify these risk factors. We used cross-sectional data obtained from the Statistical Center of Iran and the Ministry of Cooperatives Labor and Social Welfare that are taken from the labor force survey conducted in the spring of 2021. We look at years of education, work experience, housework, the number of children under the age of 6 years in the home, family education, birthplace, and the amount of land owned by households. Results show that hours spent performing domestic chores enhance the likelihood of youth becoming NEET, and years of education, years of potential work experience decrease the chance of being NEET. The findings also show that female youth born in cities were less likely than those born in rural regions to become NEET.

Keywords: NEET youth, probit, CART, machine learning, unemployment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 350
8420 Building an Integrated Relational Database from Swiss Nutrition National Survey and Swiss Health Datasets for Data Mining Purposes

Authors: Ilona Mewes, Helena Jenzer, Farshideh Einsele

Abstract:

Objective: The objective of the study was to integrate two big databases from Swiss nutrition national survey (menuCH) and Swiss health national survey 2012 for data mining purposes. Each database has a demographic base data. An integrated Swiss database is built to later discover critical food consumption patterns linked with lifestyle diseases known to be strongly tied with food consumption. Design: Swiss nutrition national survey (menuCH) with approx. 2000 respondents from two different surveys, one by Phone and the other by questionnaire along with Swiss health national survey 2012 with 21500 respondents were pre-processed, cleaned and finally integrated to a unique relational database. Results: The result of this study is an integrated relational database from the Swiss nutritional and health databases.

Keywords: Health informatics, data mining, nutritional and health databases, nutritional and chronical databases.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1725
8419 The DAQ Debugger for iFDAQ of the COMPASS Experiment

Authors: Y. Bai, M. Bodlak, V. Frolov, S. Huber, V. Jary, I. Konorov, D. Levit, J. Novy, D. Steffen, O. Subrt, M. Virius

Abstract:

In general, state-of-the-art Data Acquisition Systems (DAQ) in high energy physics experiments must satisfy high requirements in terms of reliability, efficiency and data rate capability. This paper presents the development and deployment of a debugging tool named DAQ Debugger for the intelligent, FPGA-based Data Acquisition System (iFDAQ) of the COMPASS experiment at CERN. Utilizing a hardware event builder, the iFDAQ is designed to be able to readout data at the average maximum rate of 1.5 GB/s of the experiment. In complex softwares, such as the iFDAQ, having thousands of lines of code, the debugging process is absolutely essential to reveal all software issues. Unfortunately, conventional debugging of the iFDAQ is not possible during the real data taking. The DAQ Debugger is a tool for identifying a problem, isolating the source of the problem, and then either correcting the problem or determining a way to work around it. It provides the layer for an easy integration to any process and has no impact on the process performance. Based on handling of system signals, the DAQ Debugger represents an alternative to conventional debuggers provided by most integrated development environments. Whenever problem occurs, it generates reports containing all necessary information important for a deeper investigation and analysis. The DAQ Debugger was fully incorporated to all processes in the iFDAQ during the run 2016. It helped to reveal remaining software issues and improved significantly the stability of the system in comparison with the previous run. In the paper, we present the DAQ Debugger from several insights and discuss it in a detailed way.

Keywords: DAQ debugger, data acquisition system, FPGA, system signals, Qt framework.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 893
8418 Estimation of Geotechnical Parameters by Comparing Monitoring Data with Numerical Results: Case Study of Arash–Esfandiar-Niayesh Under-Passing Tunnel, Africa Tunnel, Tehran, Iran

Authors: Aliakbar Golshani, Seyyed Mehdi Poorhashemi, Mahsa Gharizadeh

Abstract:

The under passing tunnels are strongly influenced by the soils around. There are some complexities in the specification of real soil behavior, owing to the fact that lots of uncertainties exist in soil properties, and additionally, inappropriate soil constitutive models. Such mentioned factors may cause incompatible settlements in numerical analysis with the obtained values in actual construction. This paper aims to report a case study on a specific tunnel constructed by NATM. The tunnel has a depth of 11.4 m, height of 12.2 m, and width of 14.4 m with 2.5 lanes. The numerical modeling was based on a 2D finite element program. The soil material behavior was modeled by hardening soil model. According to the field observations, the numerical estimated settlement at the ground surface was approximately four times more than the measured one, after the entire installation of the initial lining, indicating that some unknown factors affect the values. Consequently, the geotechnical parameters are accurately revised by a numerical back-analysis using laboratory and field test data and based on the obtained monitoring data. The obtained result confirms that typically, the soil parameters are conservatively low-estimated. And additionally, the constitutive models cannot be applied properly for all soil conditions.

Keywords: NATM tunnel, initial lining, field test data, laboratory test data, monitoring data, numerical back-analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 729
8417 Criteria Analysis of Residential Location Preferences: An Urban Dwellers’ Perspective

Authors: Arati Siddharth Petkar, Joel E. M. Macwan

Abstract:

Preferences for residential location are of a diverse nature. Primarily they are based on the socio-economic, socio-cultural, socio-demographic characteristics of the household. It also depends on character, and the growth potential of different areas in a city. In the present study, various criteria affecting residential location preferences from the Urban Dwellers’ perspective have been analyzed. The household survey has been conducted in two parts: Existing Buyers’ survey and Future Buyers’ survey. The analysis reveals that workplace location is the most governing criterion in deciding residential location from the majority of the urban dwellers perspective. For analyzing the importance of varied criteria, Analytical Hierarchy Process approach has been explored. The suggested approach will be helpful for urban planners, decision makers and developers, while designating a new residential area or redeveloping an existing one.

Keywords: Analytical hierarchy process, household, preferences, residential location preferences, residential land use, urban dwellers.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1258
8416 Application of Neural Networks in Financial Data Mining

Authors: Defu Zhang, Qingshan Jiang, Xin Li

Abstract:

This paper deals with the application of a well-known neural network technique, multilayer back-propagation (BP) neural network, in financial data mining. A modified neural network forecasting model is presented, and an intelligent mining system is developed. The system can forecast the buying and selling signs according to the prediction of future trends to stock market, and provide decision-making for stock investors. The simulation result of seven years to Shanghai Composite Index shows that the return achieved by this mining system is about three times as large as that achieved by the buy and hold strategy, so it is advantageous to apply neural networks to forecast financial time series, the different investors could benefit from it.

Keywords: Data mining, neural network, stock forecasting.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3590
8415 Selecting Stealth Aircraft Using Determinate Fuzzy Preference Programming in Multiple Criteria Decision Making

Authors: C. Ardil

Abstract:

This paper investigates the application of the determinate fuzzy preference programming method for a more nuanced and comprehensive evaluation of stealth aircraft. Traditional methods often struggle to incorporate subjective factors and uncertainties inherent in complex systems like stealth aircraft. Determinate fuzzy preference programming addresses this limitation by leveraging the strengths of determinate fuzzy sets. The proposed novel multiple criteria decision-making algorithm integrates these concepts to consider aspects and criteria influencing aircraft performance. This approach aims to provide a more holistic assessment by enabling decision-makers to observe positive and negative outranking flows simultaneously. By demonstrating the validity and effectiveness of this approach through a practical example of selecting a stealth aircraft, this paper aims to establish the determinate fuzzy preference programming method as a valuable tool for informed decision-making in this critical domain.

Keywords: Determinate fuzzy set, stealth aircraft selection, distance function, decision making, uncertainty, preference programming. MCDM

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 145
8414 MIMO System Order Reduction Using Real-Coded Genetic Algorithm

Authors: Swadhin Ku. Mishra, Sidhartha Panda, Simanchala Padhy, C. Ardil

Abstract:

In this paper, real-coded genetic algorithm (RCGA) optimization technique has been applied for large-scale linear dynamic multi-input-multi-output (MIMO) system. The method is based on error minimization technique where the integral square error between the transient responses of original and reduced order models has been minimized by RCGA. The reduction procedure is simple computer oriented and the approach is comparable in quality with the other well-known reduction techniques. Also, the proposed method guarantees stability of the reduced model if the original high-order MIMO system is stable. The proposed approach of MIMO system order reduction is illustrated with the help of an example and the results are compared with the recently published other well-known reduction techniques to show its superiority.

Keywords: Multi-input-multi-output (MIMO) system.Modelorder reduction. Integral squared error (ISE). Real-coded geneticalgorithm

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2262
8413 Investigation of Maritime Accidents with Exploratory Data Analysis in the Strait of Çanakkale (Dardanelles)

Authors: Gizem Kodak

Abstract:

The Strait of Çanakkale (Dardanelles), together with the Strait of Istanbul and the Sea of Marmara, form the Turkish Straits System. In other words, the Strait of Çanakkale is the southern gate of the system that connects the Black Sea countries with the other countries of the world. Due to the heavy maritime traffic, it is important to scientifically examine the accident characteristics in the region. In particular, the results indicated by the descriptive statistics are of critical importance in order to strengthen the safety of navigation. At this point, exploratory data analysis offers strategic outputs in terms of defining the problem and knowing the strengths and weaknesses against possible accident risk. The study aims to determine the accident characteristics in the Strait of Çanakkale with temporal and spatial analysis of historical data, using Exploratory Data Analysis (EDA) as the research method. The study's results will reveal the general characteristics of maritime accidents in the region and form the infrastructure for future studies. Therefore, the text provides a clear description of the research goals and methodology, and the study's contributions are well-defined.

Keywords: Maritime Accidents, EDA, Strait of Çanakkale, navigational safety.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 130
8412 Feature Point Detection by Combining Advantages of Intensity-based Approach and Edge-based Approach

Authors: Sungho Kim, Chaehoon Park, Yukyung Choi, Soon Kwon, In So Kweon

Abstract:

In this paper, a novel corner detection method is presented to stably extract geometrically important corners. Intensity-based corner detectors such as the Harris corner can detect corners in noisy environments but has inaccurate corner position and misses the corners of obtuse angles. Edge-based corner detectors such as Curvature Scale Space can detect structural corners but show unstable corner detection due to incomplete edge detection in noisy environments. The proposed image-based direct curvature estimation can overcome limitations in both inaccurate structural corner detection of the Harris corner detector (intensity-based) and the unstable corner detection of Curvature Scale Space caused by incomplete edge detection. Various experimental results validate the robustness of the proposed method.

Keywords: Feature, intensity, contour, hybrid.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1831
8411 A Generic Middleware to Instantly Sync Intensive Writes of Heterogeneous Massive Data via Internet

Authors: Haitao Yang, Zhenjiang Ruan, Fei Xu, Lanting Xia

Abstract:

Industry data centers often need to sync data changes reliably and instantly from a large-scale of heterogeneous autonomous relational databases accessed via the not-so-reliable Internet, for which a practical generic sync middleware of low maintenance and operation costs is most wanted. To this demand, this paper presented a generic sync middleware system (GSMS), which has been developed, applied and optimized since 2006, holding the principles or advantages that it must be SyncML-compliant and transparent to data application layer logic without referring to implementation details of databases synced, does not rely on host computer operating systems deployed, and its construction is light weighted and hence of low cost. Regarding these hard commitments of developing GSMS, in this paper we stressed the significant optimization breakthrough of GSMS sync delay being well below a fraction of millisecond per record sync. A series of ultimate tests with GSMS sync performance were conducted for a persuasive example, in which the source relational database underwent a broad range of write loads (from one thousand to one million intensive writes within a few minutes). All these tests showed that the performance of GSMS is competent and smooth even under ultimate write loads.

Keywords: Heterogeneous massive data, instantly sync intensive writes, Internet generic middleware design, optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 454
8410 Blood Cell Dynamics in a Simple Shear Flow using an Implicit Fluid-Structure Interaction Method Based on the ALE Approach

Authors: Choeng-Ryul Choi, Chang-Nyung Kim, Tae-Hyub Hong

Abstract:

A numerical method is developed for simulating the motion of particles with arbitrary shapes in an effectively infinite or bounded viscous flow. The particle translational and angular motions are numerically investigated using a fluid-structure interaction (FSI) method based on the Arbitrary-Lagrangian-Eulerian (ALE) approach and the dynamic mesh method (smoothing and remeshing) in FLUENT ( ANSYS Inc., USA). Also, the effects of arbitrary shapes on the dynamics are studied using the FSI method which could be applied to the motions and deformations of a single blood cell and multiple blood cells, and the primary thrombogenesis caused by platelet aggregation. It is expected that, combined with a sophisticated large-scale computational technique, the simulation method will be useful for understanding the overall properties of blood flow from blood cellular level (microscopic) to the resulting rheological properties of blood as a mass (macroscopic).

Keywords: Blood Flow, Fluid-Structure Interaction (FSI), Micro-Channels, Arbitrary Shapes, Red Blood Cells (RBCs)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2310
8409 Data and Control Flow Analysis of VDMµ Specifications

Authors: Mubina Nazmeen, Iram Rubab

Abstract:

Formal Specification languages are being widely used for system specification and testing. Highly critical systems such as real time systems, avionics, and medical systems are represented using Formal specification languages. Formal specifications based testing is mostly performed using black box testing approaches thus testing only the set of inputs and outputs of the system. The formal specification language such as VDMµ can be used for white box testing as they provide enough constructs as any other high level programming language. In this work, we perform data and control flow analysis of VDMµ class specifications. The proposed work is discussed with an example of SavingAccount.

Keywords: VDM-SL, VDMµ, data flow graph, control flowgraph, testing, formal specification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4377
8408 Analyses and Optimization of Physical and Mechanical Properties of Direct Recycled Aluminium Alloy (AA6061) Wastes by ANOVA Approach

Authors: Mohammed H. Rady, Mohd Sukri Mustapa, S Shamsudin, M. A. Lajis, A. Wagiman

Abstract:

The present study is aimed at investigating microhardness and density of aluminium alloy chips when subjected to various settings of preheating temperature and preheating time. Three values of preheating temperature were taken as 450 °C, 500 °C, and 550 °C. On the other hand, three values of preheating time were chosen (1, 2, 3) hours. The influences of the process parameters (preheating temperature and time) were analyzed using Design of Experiments (DOE) approach whereby full factorial design with center point analysis was adopted. The total runs were 11 and they comprise of two factors of full factorial design with 3 center points. The responses were microhardness and density. The results showed that the density and microhardness increased with decreasing the preheating temperature. The results also found that the preheating temperature is more important to be controlled rather than the preheating time in microhardness analysis while both the preheating temperature and preheating time are important in density analysis. It can be concluded that setting temperature at 450 °C for 1 hour resulted in the optimum responses.

Keywords: AA6061, density, DOE, hot extrusion, microhardness.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 704
8407 Using Data Mining in Automotive Safety

Authors: Carine Cridelich, Pablo Juesas Cano, Emmanuel Ramasso, Noureddine Zerhouni, Bernd Weiler

Abstract:

Safety is one of the most important considerations when buying a new car. While active safety aims at avoiding accidents, passive safety systems such as airbags and seat belts protect the occupant in case of an accident. In addition to legal regulations, organizations like Euro NCAP provide consumers with an independent assessment of the safety performance of cars and drive the development of safety systems in automobile industry. Those ratings are mainly based on injury assessment reference values derived from physical parameters measured in dummies during a car crash test. The components and sub-systems of a safety system are designed to achieve the required restraint performance. Sled tests and other types of tests are then carried out by car makers and their suppliers to confirm the protection level of the safety system. A Knowledge Discovery in Databases (KDD) process is proposed in order to minimize the number of tests. The KDD process is based on the data emerging from sled tests according to Euro NCAP specifications. About 30 parameters of the passive safety systems from different data sources (crash data, dummy protocol) are first analysed together with experts opinions. A procedure is proposed to manage missing data and validated on real data sets. Finally, a procedure is developed to estimate a set of rough initial parameters of the passive system before testing aiming at reducing the number of tests.

Keywords: KDD process, passive safety systems, sled test, dummy injury assessment reference values, frontal impact

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2844
8406 Two Dimensional Simulation of Fluid Flow and Heat Transfer in the Transition Flow Regime using a Lattice Boltzmann Approach

Authors: Mehdi Shamshiri, Mahmud Ashrafizaadeh

Abstract:

The significant effects of the interactions between the system boundaries and the near wall molecules in miniaturized gaseous devices lead to the formation of the Knudsen layer in which the Navier-Stokes-Fourier (NSF) equations fail to predict the correct associated phenomena. In this paper, the well-known lattice Boltzmann method (LBM) is employed to simulate the fluid flow and heat transfer processes in rarefied gaseous micro media. Persuaded by the problematic deficiency of the LBM in capturing the Knudsen layer phenomena, present study tends to concentrate on the effective molecular mean free path concept the main essence of which is to compensate the incapability of this mesoscopic method in dealing with the momentum and energy transport within the above mentioned kinetic boundary layer. The results show qualitative and quantitative accuracy comparable to the solutions of the linearized Boltzmann equation or the DSMC data for the Knudsen numbers of O (1) .

Keywords: Fluid flow and Heat transfer, Knudsen layer, Lattice Boltzmann method (LBM), Micro-scale numerical simulation, Transition regime.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1762
8405 A System Dynamic Based DSS for Ecological Urban Management in Alexandria, Egypt

Authors: Mona M. Salem, Khaled S. Al-Hagla, Hany M. Ayad

Abstract:

The concept of urban metabolism has increasingly been employed in a diverse range of disciplines as a mean to analyze and theorize the city. Urban ecology has a particular focus on the implications of applying the metabolism concept to the urban realm. This approach has been developed by a few researchers, though it has rarely if ever been used in policy development for city planning. The aim of this research is to use ecologically informed urban planning interventions to increase the sustainability of urban metabolism; with special focus on land stock as a most important city resource by developing a system dynamic based DSS. This model identifies two critical management strategy variables for the Strategic Urban Plan Alexandria SUP 2032. As a result, this comprehensive and precise quantitative approach is needed to monitor, measure, evaluate and observe dynamic urban changes working as a decision support system (DSS) for policy making.

Keywords: Alexandria SUP 2032, DSS, ecology, land resource, LULCC, management, metabolism, model, scenarios, System dynamics, urban development.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1194