Search results for: Data cutting and sorting method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 13805

Search results for: Data cutting and sorting method

12185 Data Centers’ Temperature Profile Simulation Optimized by Finite Elements and Discretization Methods

Authors: José Alberto García Fernández, Zhimin Du, Xinqiao Jin

Abstract:

Nowadays, data center industry faces strong challenges for increasing the speed and data processing capacities while at the same time is trying to keep their devices a suitable working temperature without penalizing that capacity. Consequently, the cooling systems of this kind of facilities use a large amount of energy to dissipate the heat generated inside the servers, and developing new cooling techniques or perfecting those already existing would be a great advance in this type of industry. The installation of a temperature sensor matrix distributed in the structure of each server would provide the necessary information for collecting the required data for obtaining a temperature profile instantly inside them. However, the number of temperature probes required to obtain the temperature profiles with sufficient accuracy is very high and expensive. Therefore, other less intrusive techniques are employed where each point that characterizes the server temperature profile is obtained by solving differential equations through simulation methods, simplifying data collection techniques but increasing the time to obtain results. In order to reduce these calculation times, complicated and slow computational fluid dynamics simulations are replaced by simpler and faster finite element method simulations which solve the Burgers‘ equations by backward, forward and central discretization techniques after simplifying the energy and enthalpy conservation differential equations. The discretization methods employed for solving the first and second order derivatives of the obtained Burgers‘ equation after these simplifications are the key for obtaining results with greater or lesser accuracy regardless of the characteristic truncation error.

Keywords: Burgers’ equations, CFD simulation, data center, discretization methods, FEM simulation, temperature profile.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 516
12184 Conduction Accompanied With Transient Radiative Heat Transfer Using Finite Volume Method

Authors: A. Ashok, K.Satapathy, B. Prerana Nashine

Abstract:

The objective of this research work is to investigate for one dimensional transient radiative transfer equations with conduction using finite volume method. Within the infrastructure of finite-volume, we obtain the conservative discretization of the terms in order to preserve the overall conservative property of finitevolume schemes. Coupling of conductive and radiative equation resulting in fluxes is governed by the magnitude of emissivity, extinction coefficient, and temperature of the medium as well as geometry of the problem. The problem under consideration has been solved, for a slab dominating radiation coupled with transient conduction based on finite volume method. The boundary conditions are also chosen so as to give a good model of the discretized form of radiation transfer equation. The important feature of the present method is flexibility in specifying the control angles in the FVM, while keeping the simplicity in the solution procedure. Effects of various model parameters are examined on the distributions of temperature, radiative and conductive heat fluxes and incident radiation energy etc. The finite volume method is considered to effectively evaluate the propagation of radiation intensity through a participating medium.

Keywords: Radiative transfer equation, finite volume method, conduction, transient radiation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1544
12183 Improvements in Edge Detection Based on Mathematical Morphology and Wavelet Transform using Fuzzy Rules

Authors: Masrour Dowlatabadi, Jalil Shirazi

Abstract:

In this paper, an improved edge detection algorithm based on fuzzy combination of mathematical morphology and wavelet transform is proposed. The combined method is proposed to overcome the limitation of wavelet based edge detection and mathematical morphology based edge detection in noisy images. Experimental results show superiority of the proposed method, as compared to the traditional Prewitt, wavelet based and morphology based edge detection methods. The proposed method is an effective edge detection method for noisy image and keeps clear and continuous edges.

Keywords: Edge detection, Wavelet transform, Mathematical morphology, Fuzzy logic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2403
12182 A Sociocybernetics Data Analysis Using Causality in Tourism Networks

Authors: M. Lloret-Climent, J. Nescolarde-Selva

Abstract:

The aim of this paper is to propose a mathematical model to determine invariant sets, set covering, orbits and, in particular, attractors in the set of tourism variables. Analysis was carried out based on a pre-designed algorithm and applying our interpretation of chaos theory developed in the context of General Systems Theory. This article sets out the causal relationships associated with tourist flows in order to enable the formulation of appropriate strategies. Our results can be applied to numerous cases. For example, in the analysis of tourist flows, these findings can be used to determine whether the behaviour of certain groups affects that of other groups and to analyse tourist behaviour in terms of the most relevant variables. Unlike statistical analyses that merely provide information on current data, our method uses orbit analysis to forecast, if attractors are found, the behaviour of tourist variables in the immediate future.

Keywords: Attractor, invariant set, orbits, tourist variables.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1745
12181 Mathematical Modeling of Human Cardiovascular System: A Lumped Parameter Approach and Simulation

Authors: Ketan Naik, P. H. Bhathawala

Abstract:

The purpose of this work is to develop a mathematical model of Human Cardiovascular System using lumped parameter method. The model is divided in three parts: Systemic Circulation, Pulmonary Circulation and the Heart. The established mathematical model has been simulated by MATLAB software. The innovation of this study is in describing the system based on the vessel diameters and simulating mathematical equations with active electrical elements. Terminology of human physical body and required physical data like vessel’s radius, thickness etc., which are required to calculate circuit parameters like resistance, inductance and capacitance, are proceeds from well-known medical books. The developed model is useful to understand the anatomic of human cardiovascular system and related syndromes. The model is deal with vessel’s pressure and blood flow at certain time.

Keywords: Cardiovascular system, lumped parameter method, mathematical modeling, simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3361
12180 The Comparison Study of Harmonic Detection Methods for Shunt Active Power Filters

Authors: K-L. Areerak, K-N. Areerak

Abstract:

The paper deals with the comparison study of harmonic detection methods for a shunt active power filter. The %THD and the power factor value at the PCC point after compensation are considered for the comparison. There are three harmonic detection methods used in the paper that are synchronous reference frame method, synchronous detection method, and DQ axis with Fourier method. In addition, the ideal current source is used to represent the active power filter by assuming an infinitely fast controller action of the active power filter. The simulation results show that the DQ axis with Fourier method provides the minimum %THD after compensation compared with other methods. However, the power factor value at the PCC point after compensation is slightly lower than that of synchronous detection method.

Keywords: Harmonic detection, shunt active power filter, DQaxis with Fourier, power factor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3293
12179 Active Tendons for Seismic Control of Buildings

Authors: S. M. Nigdeli, M. H. Boduroglu

Abstract:

In this study, active tendons with Proportional Integral Derivation type controllers were applied to a SDOF and a MDOF building model. Physical models of buildings were constituted with virtual springs, dampers and rigid masses. After that, equations of motion of all degrees of freedoms were obtained. Matlab Simulink was utilized to obtain the block diagrams for these equations of motion. Parameters for controller actions were found by using a trial method. After earthquake acceleration data were applied to the systems, building characteristics such as displacements, velocities, accelerations and transfer functions were analyzed for all degrees of freedoms. Comparisons on displacement vs. time, velocity vs. time, acceleration vs. time and transfer function (Db) vs. frequency (Hz) were made for uncontrolled and controlled buildings. The results show that the method seems feasible.

Keywords: Active Tendons, Proportional Integral DerivationType Controllers, SDOF, MDOF, Earthquake, Building.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3328
12178 Robust Face Recognition using AAM and Gabor Features

Authors: Sanghoon Kim, Sun-Tae Chung, Souhwan Jung, Seoungseon Jeon, Jaemin Kim, Seongwon Cho

Abstract:

In this paper, we propose a face recognition algorithm using AAM and Gabor features. Gabor feature vectors which are well known to be robust with respect to small variations of shape, scaling, rotation, distortion, illumination and poses in images are popularly employed for feature vectors for many object detection and recognition algorithms. EBGM, which is prominent among face recognition algorithms employing Gabor feature vectors, requires localization of facial feature points where Gabor feature vectors are extracted. However, localization method employed in EBGM is based on Gabor jet similarity and is sensitive to initial values. Wrong localization of facial feature points affects face recognition rate. AAM is known to be successfully applied to localization of facial feature points. In this paper, we devise a facial feature point localization method which first roughly estimate facial feature points using AAM and refine facial feature points using Gabor jet similarity-based facial feature localization method with initial points set by the rough facial feature points obtained from AAM, and propose a face recognition algorithm using the devised localization method for facial feature localization and Gabor feature vectors. It is observed through experiments that such a cascaded localization method based on both AAM and Gabor jet similarity is more robust than the localization method based on only Gabor jet similarity. Also, it is shown that the proposed face recognition algorithm using this devised localization method and Gabor feature vectors performs better than the conventional face recognition algorithm using Gabor jet similarity-based localization method and Gabor feature vectors like EBGM.

Keywords: Face Recognition, AAM, Gabor features, EBGM.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2206
12177 Retrieving Similar Segmented Objects Using Motion Descriptors

Authors: Konstantinos C. Kartsakalis, Angeliki Skoura, Vasileios Megalooikonomou

Abstract:

The fuzzy composition of objects depicted in images acquired through MR imaging or the use of bio-scanners has often been a point of controversy for field experts attempting to effectively delineate between the visualized objects. Modern approaches in medical image segmentation tend to consider fuzziness as a characteristic and inherent feature of the depicted object, instead of an undesirable trait. In this paper, a novel technique for efficient image retrieval in the context of images in which segmented objects are either crisp or fuzzily bounded is presented. Moreover, the proposed method is applied in the case of multiple, even conflicting, segmentations from field experts. Experimental results demonstrate the efficiency of the suggested method in retrieving similar objects from the aforementioned categories while taking into account the fuzzy nature of the depicted data.

Keywords: Fuzzy Object, Fuzzy Image Segmentation, Motion Descriptors, MRI Imaging, Object-Based Image Retrieval.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2302
12176 Solar Radiation Studies for Dubai and Sharjah, UAE

Authors: Muhammed A. Ahmed, Sidra A. Shaikh

Abstract:

Global Solar Radiation (H) for Dubai and Sharjah, Latitude 25.25oN, Longitude 55oE and 25.29oN, Longitude 55oE respectively have been studied using sunshine hour data (n) of the areas using various methods. These calculated global solar radiation values are then compared to the measured values presented by NASA. Furthermore, the extraterrestrial (H0), diffuse (Hd) and beam radiation (Hb) are also calculated. The diffuse radiation is calculated using methods proposed by Page and Liu and Jordan (L-J). Diffuse Radiation from the Page method is higher than the L-J method. Moreover, the clearness index (KT) signifies a clear sky almost all year round. Rainy days are hardly a few in a year and limited in the months December to March. The temperature remains between 25oC in winter to 44oC in summer and is desirable for thermal applications of solar energy. From the estimated results, it appears that solar radiation can be utilized very efficiently throughout the year for photovoltaic and thermal applications.

Keywords: Dubai, Sharjah, Global Solar Radiation, Diffuse Radiation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9299
12175 Speech Recognition Using Scaly Neural Networks

Authors: Akram M. Othman, May H. Riadh

Abstract:

This research work is aimed at speech recognition using scaly neural networks. A small vocabulary of 11 words were established first, these words are “word, file, open, print, exit, edit, cut, copy, paste, doc1, doc2". These chosen words involved with executing some computer functions such as opening a file, print certain text document, cutting, copying, pasting, editing and exit. It introduced to the computer then subjected to feature extraction process using LPC (linear prediction coefficients). These features are used as input to an artificial neural network in speaker dependent mode. Half of the words are used for training the artificial neural network and the other half are used for testing the system; those are used for information retrieval. The system components are consist of three parts, speech processing and feature extraction, training and testing by using neural networks and information retrieval. The retrieve process proved to be 79.5-88% successful, which is quite acceptable, considering the variation to surrounding, state of the person, and the microphone type.

Keywords: Feature extraction, Liner prediction coefficients, neural network, Speech Recognition, Scaly ANN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1737
12174 Time/Temperature-Dependent Finite Element Model of Laminated Glass Beams

Authors: Alena Zemanová, Jan Zeman, Michal Šejnoha

Abstract:

The polymer foil used for manufacturing of laminated glass members behaves in a viscoelastic manner with temperature dependance. This contribution aims at incorporating the time/temperature-dependent behavior of interlayer to our earlier elastic finite element model for laminated glass beams. The model is based on a refined beam theory: each layer behaves according to the finite-strain shear deformable formulation by Reissner and the adjacent layers are connected via the Lagrange multipliers ensuring the inter-layer compatibility of a laminated unit. The time/temperature-dependent behavior of the interlayer is accounted for by the generalized Maxwell model and by the time-temperature superposition principle due to the Williams, Landel, and Ferry. The resulting system is solved by the Newton method with consistent linearization and the viscoelastic response is determined incrementally by the exponential algorithm. By comparing the model predictions against available experimental data, we demonstrate that the proposed formulation is reliable and accurately reproduces the behavior of the laminated glass units.

Keywords: Laminated glass, finite element method, finite-strain Reissner model, Lagrange multipliers, generalized Maxwell model, Williams-Landel-Ferry equation, Newton method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1685
12173 The Harada Method – A Method for Employee Development during Production Ramp Up

Authors: M. Goerke, J. Gehrmann

Abstract:

Caused by shorter product life cycles and higher product variety the importance of production ramp ups is increasing. Even though companies are aware of that fact, up to 40% of the ramp up projects still miss technical and economical requirements. The success of a ramp up depends on the planning of human factors, organizational aspects and technological solutions. Since only partly considered in scientific literature, this paper lays its focus on the human factor during production ramp up. There are only incoherent methods which address the problems in this area. A systematic and holistic method to improve the capabilities of the employees during ramp up is missing. The Harada Method is a relatively young approach for developing highly-skilled workers. It consists of different worksheets which help employees to set guidelines and reach overall objectives. This approach is going to be transferred into a tool for ramp up management.

Keywords: Employee Development, Harada, Production Ramp Up.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2278
12172 Agile Methodology for Modeling and Design of Data Warehouses -AM4DW-

Authors: Nieto Bernal Wilson, Carmona Suarez Edgar

Abstract:

The organizations have structured and unstructured information in different formats, sources, and systems. Part of these come from ERP under OLTP processing that support the information system, however these organizations in OLAP processing level, presented some deficiencies, part of this problematic lies in that does not exist interesting into extract knowledge from their data sources, as also the absence of operational capabilities to tackle with these kind of projects.  Data Warehouse and its applications are considered as non-proprietary tools, which are of great interest to business intelligence, since they are repositories basis for creating models or patterns (behavior of customers, suppliers, products, social networks and genomics) and facilitate corporate decision making and research. The following paper present a structured methodology, simple, inspired from the agile development models as Scrum, XP and AUP. Also the models object relational, spatial data models, and the base line of data modeling under UML and Big data, from this way sought to deliver an agile methodology for the developing of data warehouses, simple and of easy application. The methodology naturally take into account the application of process for the respectively information analysis, visualization and data mining, particularly for patterns generation and derived models from the objects facts structured.

Keywords: Data warehouse, model data, big data, object fact, object relational fact, process developed data warehouse.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1478
12171 The Study on Evaluation System and Method of Legacy System

Authors: Chao Qi, Fuyang Peng, Bo Deng, Xiaoyan Su

Abstract:

In the upgrade process of enterprise information systems, how to deal with and utilize those legacy systems affects the efficiency of construction and development of the new system. We propose an evaluation system, which comprehensively describes the capacity of legacy information systems in five aspects. Then we propose a practical legacy systems evaluation method. Base on the evaluation result, we can determine the current state of legacy system which was evaluated.

Keywords: Legacy Information Systems, Evaluation IndexSystem, Evaluation Method, Evaluation Level

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1620
12170 Study of the Electromagnetic Resonances of a Cavity with an Aperture Using Numerical Method and Equivalent Circuit Method

Authors: Ming-Chu Yin, Ping-An Du

Abstract:

The shielding ability of a shielding cavity with an aperture will be greatly degraded at resonance frequencies, and the resonance modes and frequencies are affected by aperture resonances and aperture-cavity coupling, which are closely related with aperture sizes. The equivalent circuit method and numerical method of Transmission Line Matrix (TLM) are used to analyze the effects of aperture resonances and aperture-cavity coupling on the electromagnetic resonances of a cavity with an aperture in this paper. Both analytical and numerical results show that the resonance modes of a shielding cavity with an aperture consist of cavity resonance modes and aperture resonance modes, and the resonance frequencies will shift with the change of the aperture sizes because of the aperture resonances and aperture-cavity coupling. Variation rules of electromagnetic resonances with aperture sizes for a cavity with an aperture are given, which will be useful for design of shielding cavities.

Keywords: Aperture-cavity coupling, equivalent circuit method, resonances, shielding equipment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1598
12169 Postbuckling Analysis of End Supported Rods under Self-Weight Using Intrinsic Coordinate Finite Elements

Authors: C. Juntarasaid, T. Pulngern, S. Chucheepsakul

Abstract:

A formulation of postbuckling analysis of end supported rods under self-weight has been presented by the variational method. The variational formulation involving the strain energy due to bending and the potential energy of the self-weight, are expressed in terms of the intrinsic coordinates. The variational formulation is accomplished by introducing the Lagrange multiplier technique to impose the boundary conditions. The finite element method is used to derive a system of nonlinear equations resulting from the stationary of the total potential energy and then Newton-Raphson iterative procedure is applied to solve this system of equations. The numerical results demonstrate the postbluckled configurations of end supported rods under self-weight. This finite element method based on variational formulation expressed in term of intrinsic coordinate is highly recommended for postbuckling analysis of end-supported rods under self-weight.

Keywords: Variational method, postbuckling, finite element method, intrinsic coordinate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 848
12168 The Design of a Vehicle Traffic Flow Prediction Model for a Gauteng Freeway Based on an Ensemble of Multi-Layer Perceptron

Authors: Tebogo Emma Makaba, Barnabas Ndlovu Gatsheni

Abstract:

The cities of Johannesburg and Pretoria both located in the Gauteng province are separated by a distance of 58 km. The traffic queues on the Ben Schoeman freeway which connects these two cities can stretch for almost 1.5 km. Vehicle traffic congestion impacts negatively on the business and the commuter’s quality of life. The goal of this paper is to identify variables that influence the flow of traffic and to design a vehicle traffic prediction model, which will predict the traffic flow pattern in advance. The model will unable motorist to be able to make appropriate travel decisions ahead of time. The data used was collected by Mikro’s Traffic Monitoring (MTM). Multi-Layer perceptron (MLP) was used individually to construct the model and the MLP was also combined with Bagging ensemble method to training the data. The cross—validation method was used for evaluating the models. The results obtained from the techniques were compared using predictive and prediction costs. The cost was computed using combination of the loss matrix and the confusion matrix. The predicted models designed shows that the status of the traffic flow on the freeway can be predicted using the following parameters travel time, average speed, traffic volume and day of month. The implications of this work is that commuters will be able to spend less time travelling on the route and spend time with their families. The logistics industry will save more than twice what they are currently spending.

Keywords: Bagging ensemble methods, confusion matrix, multi-layer perceptron, vehicle traffic flow.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1777
12167 Representing Data without Lost Compression Properties in Time Series: A Review

Authors: Nabilah Filzah Mohd Radzuan, Zalinda Othman, Azuraliza Abu Bakar, Abdul Razak Hamdan

Abstract:

Uncertain data is believed to be an important issue in building up a prediction model. The main objective in the time series uncertainty analysis is to formulate uncertain data in order to gain knowledge and fit low dimensional model prior to a prediction task. This paper discusses the performance of a number of techniques in dealing with uncertain data specifically those which solve uncertain data condition by minimizing the loss of compression properties.

Keywords: Compression properties, uncertainty, uncertain time series, mining technique, weather prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1620
12166 The Reliability of Wireless Sensor Network

Authors: B. Juhasova, I. Halenar, M. Juhas

Abstract:

The wireless communication is one of the widely used methods of data transfer at the present days. The benefit of this communication method is the partial independence of the infrastructure and the possibility of mobility. In some special applications it is the only way how to connect. This paper presents some problems in the implementation of a sensor network connection for measuring environmental parameters in the area of manufacturing plants.

Keywords: Network, communication, reliability, sensors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1535
12165 Parametric Approach for Reserve Liability Estimate in Mortgage Insurance

Authors: Rajinder Singh, Ram Valluru

Abstract:

Chain Ladder (CL) method, Expected Loss Ratio (ELR) method and Bornhuetter-Ferguson (BF) method, in addition to more complex transition-rate modeling, are commonly used actuarial reserving methods in general insurance. There is limited published research about their relative performance in the context of Mortgage Insurance (MI). In our experience, these traditional techniques pose unique challenges and do not provide stable claim estimates for medium to longer term liabilities. The relative strengths and weaknesses among various alternative approaches revolve around: stability in the recent loss development pattern, sufficiency and reliability of loss development data, and agreement/disagreement between reported losses to date and ultimate loss estimate. CL method results in volatile reserve estimates, especially for accident periods with little development experience. The ELR method breaks down especially when ultimate loss ratios are not stable and predictable. While the BF method provides a good tradeoff between the loss development approach (CL) and ELR, the approach generates claim development and ultimate reserves that are disconnected from the ever-to-date (ETD) development experience for some accident years that have more development experience. Further, BF is based on subjective a priori assumption. The fundamental shortcoming of these methods is their inability to model exogenous factors, like the economy, which impact various cohorts at the same chronological time but at staggered points along their life-time development. This paper proposes an alternative approach of parametrizing the loss development curve and using logistic regression to generate the ultimate loss estimate for each homogeneous group (accident year or delinquency period). The methodology was tested on an actual MI claim development dataset where various cohorts followed a sigmoidal trend, but levels varied substantially depending upon the economic and operational conditions during the development period spanning over many years. The proposed approach provides the ability to indirectly incorporate such exogenous factors and produce more stable loss forecasts for reserving purposes as compared to the traditional CL and BF methods.

Keywords: Actuarial loss reserving techniques, logistic regression, parametric function, volatility.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 416
12164 Are XBRL-based Financial Reports Better than Non-XBRL Reports? A Quality Assessment

Authors: Zhenkun Wang, Simon S. Gao

Abstract:

Using a scoring system, this paper provides a comparative assessment of the quality of data between XBRL formatted financial reports and non-XBRL financial reports. It shows a major improvement in the quality of data of XBRL formatted financial reports. Although XBRL formatted financial reports do not show much advantage in the quality at the beginning, XBRL financial reports lately display a large improvement in the quality of data in almost all aspects. With the improved XBRL web data managing, presentation and analysis applications, XBRL formatted financial reports have a much better accessibility, are more accurate and better in timeliness.

Keywords: Data Quality; Financial Report; Information; XBRL

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2566
12163 T-Wave Detection Based on an Adjusted Wavelet Transform Modulus Maxima

Authors: Samar Krimi, Kaïs Ouni, Noureddine Ellouze

Abstract:

The method described in this paper deals with the problems of T-wave detection in an ECG. Determining the position of a T-wave is complicated due to the low amplitude, the ambiguous and changing form of the complex. A wavelet transform approach handles these complications therefore a method based on this concept was developed. In this way we developed a detection method that is able to detect T-waves with a sensitivity of 93% and a correct-detection ratio of 93% even with a serious amount of baseline drift and noise.

Keywords: ECG, Modulus Maxima Wavelet Transform, Performance, T-wave detection

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1853
12162 Evaluation of a PSO Approach for Optimum Design of a First-Order Controllers for TCP/AQM Systems

Authors: Sana Testouri, Karim Saadaoui, Mohamed Benrejeb

Abstract:

This paper presents a Particle Swarm Optimization (PSO) method for determining the optimal parameters of a first-order controller for TCP/AQM system. The model TCP/AQM is described by a second-order system with time delay. First, the analytical approach, based on the D-decomposition method and Lemma of Kharitonov, is used to determine the stabilizing regions of a firstorder controller. Second, the optimal parameters of the controller are obtained by the PSO algorithm. Finally, the proposed method is implemented in the Network Simulator NS-2 and compared with the PI controller.

Keywords: AQM, first-order controller, time delay, stability, PSO.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1763
12161 Modeling of Random Variable with Digital Probability Hyper Digraph: Data-Oriented Approach

Authors: A. Habibizad Navin, M. Naghian Fesharaki, M. Mirnia, M. Kargar

Abstract:

In this paper we introduce Digital Probability Hyper Digraph for modeling random variable as the hierarchical data-oriented model.

Keywords: Data-Oriented Models, Data Structure, DigitalProbability Hyper Digraph, Random Variable, Statistic andProbability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1273
12160 Corporate Cultures Management towards the Retention of Employees: Case Study Company in Thailand

Authors: Duangsamorn Rungsawanpho

Abstract:

The objectives of this paper are to explore the corporate cultures management as determinants of employee retention company in Thailand. This study using mixed method methodology. Data collection using questionnaires and in-depth interviews. The statistics used for data analysis were percentage, mean, standard deviation and inferential statistics will include. The results show that the corporate management culture is perfect for any organization but it depends on the business and the industry because the situations or circumstances that corporate executives are met is different. Because the finding explained that the employees of the company determine the achievement of value-oriented by the corporate culture and international relations is perceived most value for their organizations. In additional we found the employees perceiving with participation can be interpreted as a positive example, many employees feel that they are part of management because they care about their opinions or ideas related with their work.

Keywords: Corporate culture, employee retention, retention of employees, management approaches.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 719
12159 The Simulation and Experimental Investigation to Study the Strain Distribution Pattern during the Closed Die Forging Process

Authors: D. B. Gohil

Abstract:

Closed die forging is a very complex process, and measurement of actual forces for real material is difficult and time consuming. Hence, the modelling technique has taken the advantage of carrying out the experimentation with the proper model material which needs lesser forces and relatively low temperature. The results of experiments on the model material then may be correlated with the actual material by using the theory of similarity. There are several methods available to resolve the complexity involved in the closed die forging process. Finite Element Method (FEM) and Finite Difference Method (FDM) are relatively difficult as compared to the slab method. The slab method is very popular and very widely used by the people working on shop floor because it is relatively easy to apply and reasonably accurate for most of the common forging load requirement computations.

Keywords: Experimentation, forging, process modeling, strain distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1365
12158 Wireless Transmission of Big Data Using Novel Secure Algorithm

Authors: K. Thiagarajan, K. Saranya, A. Veeraiah, B. Sudha

Abstract:

This paper presents a novel algorithm for secure, reliable and flexible transmission of big data in two hop wireless networks using cooperative jamming scheme. Two hop wireless networks consist of source, relay and destination nodes. Big data has to transmit from source to relay and from relay to destination by deploying security in physical layer. Cooperative jamming scheme determines transmission of big data in more secure manner by protecting it from eavesdroppers and malicious nodes of unknown location. The novel algorithm that ensures secure and energy balance transmission of big data, includes selection of data transmitting region, segmenting the selected region, determining probability ratio for each node (capture node, non-capture and eavesdropper node) in every segment, evaluating the probability using binary based evaluation. If it is secure transmission resume with the two- hop transmission of big data, otherwise prevent the attackers by cooperative jamming scheme and transmit the data in two-hop transmission.

Keywords: Big data, cooperative jamming, energy balance, physical layer, two-hop transmission, wireless security.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2180
12157 Using Interval Trees for Approximate Indexing of Instances

Authors: Khalil el Hindi

Abstract:

This paper presents a simple and effective method for approximate indexing of instances for instance based learning. The method uses an interval tree to determine a good starting search point for the nearest neighbor. The search stops when an early stopping criterion is met. The method proved to be very effective especially when only the first nearest neighbor is required.

Keywords: Instance based learning, interval trees, the knn algorithm, machine learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1510
12156 Increasing Value Added of Recycling Business Management: A Case of Thailand

Authors: Yananda Siraphatthada

Abstract:

This policy participation action research explores the roles of Thai government units during its 2010 fiscal year on how to create value added to recycling business in the central part of Thailand. The research aims to a) study how the government plays a role to support the business, and its problems and obstacles on supporting the business, b) to design a strategic action – short, medium, and long term plans -- to create value added to the recycling business, particularly in local full-loop companies/organizations licensed by Wongpanit Waste Separation Plant as well as those licensed by the Department of Provincial Administration. Mixed method research design, i.e., a combination of quantitative and qualitative methods is utilized in the present study in both data collection and analysis procedures. Quantitative data was analyzed by frequency, percent value, mean scores, and standard deviation, and aimed to note trend and generalizations. Qualitative data was collected via semi-structured interviews/focus group interviews to explore in-depth views of the operators. The sampling included 1,079 operators in eight provinces in the central part of Thailand.

Keywords: Management, Recycling Business, Value Added.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2758