Search results for: block typechannel estimation.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1507

Search results for: block typechannel estimation.

1147 Dead-Reckoning Error Calibration using Celling Looking Vision Camera

Authors: Jae-Young Choi, Sung-Gaun Kim

Abstract:

This paper suggests a calibration method to reduce errors occurring due to mobile robot sliding during location estimation using the Dead-reckoning. Due to sliding of the mobile robot caused between its wheels and the road surface while on free run, location estimation can be erroneous. Sliding especially occurs during cornering of mobile robot. Therefore, in order to reduce these frequent sliding errors in cornering, we calibrated the mobile robot-s heading values using a vision camera and templates of the ceiling.

Keywords: Dead-reckoning, Localization, Odomerty, Vision Camera

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1752
1146 Exponential State Estimation for Neural Networks with Leakage, Discrete and Distributed Delays

Authors: Liyuan Wang, Shouming Zhong

Abstract:

In this paper, the design problem of state estimator for neural networks with the mixed time-varying delays are investigated by constructing appropriate Lyapunov-Krasovskii functionals and using some effective mathematical techniques. In order to derive several conditions to guarantee the estimation error systems to be globally exponential stable, we transform the considered systems into the neural-type time-delay systems. Then with a set of linear inequalities(LMIs), we can obtain the stable criteria. Finally, three numerical examples are given to show the effectiveness and less conservatism of the proposed criterion.

Keywords: State estimator, Neural networks, Globally exponential stability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1629
1145 Hybrid Genetic-Simulated Annealing Approach for Fractal Image Compression

Authors: Y.Chakrapani, K.Soundera Rajan

Abstract:

In this paper a hybrid technique of Genetic Algorithm and Simulated Annealing (HGASA) is applied for Fractal Image Compression (FIC). With the help of this hybrid evolutionary algorithm effort is made to reduce the search complexity of matching between range block and domain block. The concept of Simulated Annealing (SA) is incorporated into Genetic Algorithm (GA) in order to avoid pre-mature convergence of the strings. One of the image compression techniques in the spatial domain is Fractal Image Compression but the main drawback of FIC is that it involves more computational time due to global search. In order to improve the computational time along with acceptable quality of the decoded image, HGASA technique has been proposed. Experimental results show that the proposed HGASA is a better method than GA in terms of PSNR for Fractal image Compression.

Keywords: Fractal Image Compression, Genetic Algorithm, HGASA, Simulated Annealing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1640
1144 Estimation of Shock Velocity and Pressure of Detonations and Finding Their Flow Parameters

Authors: Mahmoud Zarrini, R. N. Pralhad

Abstract:

In this paper, mathematical modeling of detonation in the ground is studied. Estimation of flow parameters such as velocity, maximum velocity, acceleration, maximum acceleration, shock pressure as a result of an explosion in the ground have been computed in an appropriate dynamic model approach. The variation of these parameters with the diameter of detonation place (L), density of earth or stone (¤ü), time decay of detonation (T), peak pressure (Pm), and time (t) have been analyzed. The model has been developed from the concept of underwater explosions [Refs. [1]-[3]] with appropriate changes to the present model requirements.

Keywords: Shock velocity, detonation, shock acceleration, shock pressure.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1264
1143 Optimizing Hadoop Block Placement Policy and Cluster Blocks Distribution

Authors: Nchimbi Edward Pius, Liu Qin, Fion Yang, Zhu Hong Ming

Abstract:

The current Hadoop block placement policy do not fairly and evenly distributes replicas of blocks written to datanodes in a Hadoop cluster.

This paper presents a new solution that helps to keep the cluster in a balanced state while an HDFS client is writing data to a file in Hadoop cluster. The solution had been implemented, and test had been conducted to evaluate its contribution to Hadoop distributed file system.

It has been found that, the solution has lowered global execution time taken by Hadoop balancer to 22 percent. It also has been found that, Hadoop balancer respectively over replicate 1.75 and 3.3 percent of all re-distributed blocks in the modified and original Hadoop clusters.

The feature that keeps the cluster in a balanced state works as a core part to Hadoop system and not just as a utility like traditional balancer. This is one of the significant achievements and uniqueness of the solution developed during the course of this research work.

Keywords: Balancer, Datanode, Distributed file system, Hadoop, Replicas.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4909
1142 Zero Inflated Strict Arcsine Regression Model

Authors: Y. N. Phang, E. F. Loh

Abstract:

Zero inflated strict arcsine model is a newly developed model which is found to be appropriate in modeling overdispersed count data. In this study, we extend zero inflated strict arcsine model to zero inflated strict arcsine regression model by taking into consideration the extra variability caused by extra zeros and covariates in count data. Maximum likelihood estimation method is used in estimating the parameters for this zero inflated strict arcsine regression model.

Keywords: Overdispersed count data, maximum likelihood estimation, simulated annealing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1720
1141 Design and Research of a New Kind Balance Adjusting System of Centrifuge

Authors: Li Xinyao, Hong Jianzhong, Wu Wenkai

Abstract:

In order to make environmental test centrifuge balance automatically and accurately, reduce unbalance centrifugal force, balance adjusting system of centrifuge is designed. The new balance adjusting system comprises motor-reducer, timing belt, screw pair, slider-guideway and four rocker force sensors. According to information obtained by the four rocker force sensors, unbalanced value at both ends of the big arm is computed and heavy block is moved to achieve balance adjusting. In this paper, motor power and torque to move the heavy block is calculated. In full load running progress of centrifuge, the stress-strain of screw pair composed by adjusting nut and big arm are analyzed. A successful application of the balance adjusting system is also put forwarded. The results show that the balance adjusting system can satisfy balance require of environmental test centrifuge.

Keywords: balance adjusting system, centrifuge, screw pair, timing belt

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1342
1140 Design of an Authentication Protocol for Secure Electronic Seals

Authors: Seongsoo Park, Mun-Kyu Lee, Dong Kyue Kim, Kunsoo Park, Yousung Kang, Sokjoon Lee, Howon Kim, Kyoil Chung

Abstract:

Electronic seal is an electronic device to check the authenticity and integrity of freight containers at the point of arrival. While RFID-based eSeals are gaining more acceptances and there are also some standardization processes for these devices, a recent research revealed that the current RFID-based eSeals are vulnerable to various attacks. In this paper, we provide a feasible solution to enhance the security of active RFID-based eSeals. Our approach is to use an authentication and key agreement protocol between eSeal and reader device, enabling data encryption and integrity check. Our protocol is based on the use of block cipher AES, which is reasonable since a block cipher can also be used for many other security purposes including data encryption and pseudo-random number generation. Our protocol is very simple, and it is applicable to low-end active RFID eSeals.

Keywords: Authentication, Container Security, Electronic seal, RFID

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1868
1139 Optimal Image Representation for Linear Canonical Transform Multiplexing

Authors: Navdeep Goel, Salvador Gabarda

Abstract:

Digital images are widely used in computer applications. To store or transmit the uncompressed images requires considerable storage capacity and transmission bandwidth. Image compression is a means to perform transmission or storage of visual data in the most economical way. This paper explains about how images can be encoded to be transmitted in a multiplexing time-frequency domain channel. Multiplexing involves packing signals together whose representations are compact in the working domain. In order to optimize transmission resources each 4 × 4 pixel block of the image is transformed by a suitable polynomial approximation, into a minimal number of coefficients. Less than 4 × 4 coefficients in one block spares a significant amount of transmitted information, but some information is lost. Different approximations for image transformation have been evaluated as polynomial representation (Vandermonde matrix), least squares + gradient descent, 1-D Chebyshev polynomials, 2-D Chebyshev polynomials or singular value decomposition (SVD). Results have been compared in terms of nominal compression rate (NCR), compression ratio (CR) and peak signal-to-noise ratio (PSNR) in order to minimize the error function defined as the difference between the original pixel gray levels and the approximated polynomial output. Polynomial coefficients have been later encoded and handled for generating chirps in a target rate of about two chirps per 4 × 4 pixel block and then submitted to a transmission multiplexing operation in the time-frequency domain.

Keywords: Chirp signals, Image multiplexing, Image transformation, Linear canonical transform, Polynomial approximation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2096
1138 Recursive Filter for Coastal Displacement Estimation

Authors: Efstratios Doukakis, Nikolaos Petrelis

Abstract:

All climate models agree that the temperature in Greece will increase in the range of 1° to 2°C by the year 2030 and mean sea level in Mediterranean is expected to rise at the rate of 5 cm/decade. The aim of the present paper is the estimation of the coastline displacement driven by the climate change and sea level rise. In order to achieve that, all known statistical and non-statistical computational methods are employed on some Greek coastal areas. Furthermore, Kalman filtering techniques are for the first time introduced, formulated and tested. Based on all the above, shoreline change signals and noises are computed and an inter-comparison between the different methods can be deduced to help evaluating which method is most promising as far as the retrieve of shoreline change rate is concerned.

Keywords: Climate Change, Coastal Displacement, KalmanFilter

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1376
1137 A Simulation for Estimation of the Blood Pressure using Arterial Pressure-volume Model

Authors: Gye-rok Jeon, Jae-hee Jung, In-cheol Kim, Ah-young Jeon, Sang-hwa Yoon, Jung-man Son, Jae-hyung Kim, Soo-young Ye, Jung-hoon Ro, Dong-hyun Kim, Chul-han Kim

Abstract:

A analysis on the conventional the blood pressure estimation method using an oscillometric sphygmomanometer was performed through a computer simulation using an arterial pressure-volume (APV) model. Traditionally, the maximum amplitude algorithm (MAP) was applied on the oscillation waveforms of the APV model to obtain the mean arterial pressure and the characteristic ratio. The estimation of mean arterial pressure and characteristic ratio was significantly affected with the shape of the blood pressure waveforms and the cutoff frequency of high-pass filter (HPL) circuitry. Experimental errors are due to these effects when estimating blood pressure. To find out an algorithm independent from the influence of waveform shapes and parameters of HPL, the volume oscillation of the APV model and the phase shift of the oscillation with fast fourier transform (FFT) were testified while increasing the cuff pressure from 1 mmHg to 200 mmHg (1 mmHg per second). The phase shift between the ranges of volume oscillation was then only observed between the systolic and the diastolic blood pressures. The same results were also obtained from the simulations performed on two different the arterial blood pressure waveforms and one hyperthermia waveform.

Keywords: Arterial blood pressure, oscillometric method

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3297
1136 Human Action Recognition System Based on Silhouette

Authors: S. Maheswari, P. Arockia Jansi Rani

Abstract:

Human action is recognized directly from the video sequences. The objective of this work is to recognize various human actions like run, jump, walk etc. Human action recognition requires some prior knowledge about actions namely, the motion estimation, foreground and background estimation. Region of interest (ROI) is extracted to identify the human in the frame. Then, optical flow technique is used to extract the motion vectors. Using the extracted features similarity measure based classification is done to recognize the action. From experimentations upon the Weizmann database, it is found that the proposed method offers a high accuracy.

Keywords: Background subtraction, human silhouette, optical flow, classification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 962
1135 A Study on the Planning Criteria of Block-Unit Redevelopment to Improve Residential Environment - Focused on Redevelopment Project in Seoul -

Authors: Hong-Nam Choi, Hyeong-Wook Song, Sungwan Hong, Hong-Kyu Kim

Abstract:

In Korea, elements that decide the quality of residential environment are not only diverse, but show deviation as well. However, people do not consider these elements and instead, they try to settle the uniformed style of residential environment, which focuses on the construction development of apartment housing and business based plans. Recently, block-unit redevelopment is becoming the standout alternative plan of standardize redevelopment projects, but constructions become inefficient because of indefinite planning criteria.

 In conclusion, the following research is about analyzing and categorizing the development method and legal ground of redevelopment project district and plan determinant and applicable standard; the purpose of this study is to become a basis in compatible analysis of planning standards that will happen in the future.

Keywords: Shape Restrictions, Improvement of Reagulation, Diversity of Residential Environment, Classification of Redevelopment Project, Planning Criteria of Redevelopment, Special Architectural District (SAD).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1288
1134 Coherence Analysis between Respiration and PPG Signal by Bivariate AR Model

Authors: Yue-Der Lin, Wei-Ting Liu, Ching-Che Tsai, Wen-Hsiu Chen

Abstract:

PPG is a potential tool in clinical applications. Among such, the relationship between respiration and PPG signal has attracted attention in past decades. In this research, a bivariate AR spectral estimation method was utilized for the coherence analysis between these two signals. Ten healthy subjects participated in this research with signals measured at different respiratory rates. The results demonstrate that high coherence exists between respiration and PPG signal, whereas the coherence disappears in breath-holding experiments. These results imply that PPG signal reveals the respiratory information. The utilized method may provide an attractive alternative approach for the related researches.

Keywords: Coherence analysis, photoplethysmography (PPG), bivariate AR spectral estimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2557
1133 Particle Swarm Optimization Algorithm vs. Genetic Algorithm for Image Watermarking Based Discrete Wavelet Transform

Authors: Omaima N. Ahmad AL-Allaf

Abstract:

Over communication networks, images can be easily copied and distributed in an illegal way. The copyright protection for authors and owners is necessary. Therefore, the digital watermarking techniques play an important role as a valid solution for authority problems. Digital image watermarking techniques are used to hide watermarks into images to achieve copyright protection and prevent its illegal copy. Watermarks need to be robust to attacks and maintain data quality. Therefore, we discussed in this paper two approaches for image watermarking, first is based on Particle Swarm Optimization (PSO) and the second approach is based on Genetic Algorithm (GA). Discrete wavelet transformation (DWT) is used with the two approaches separately for embedding process to cover image transformation. Each of PSO and GA is based on co-relation coefficient to detect the high energy coefficient watermark bit in the original image and then hide the watermark in original image. Many experiments were conducted for the two approaches with different values of PSO and GA parameters. From experiments, PSO approach got better results with PSNR equal 53, MSE equal 0.0039. Whereas GA approach got PSNR equal 50.5 and MSE equal 0.0048 when using population size equal to 100, number of iterations equal to 150 and 3×3 block. According to the results, we can note that small block size can affect the quality of image watermarking based PSO/GA because small block size can increase the search area of the watermarking image. Better PSO results were obtained when using swarm size equal to 100.

Keywords: Image watermarking, genetic algorithm, particle swarm optimization, discrete wavelet transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1115
1132 Unscented Grid Filtering and Smoothing for Nonlinear Time Series Analysis

Authors: Nikolay Nikolaev, Evgueni Smirnov

Abstract:

This paper develops an unscented grid-based filter and a smoother for accurate nonlinear modeling and analysis of time series. The filter uses unscented deterministic sampling during both the time and measurement updating phases, to approximate directly the distributions of the latent state variable. A complementary grid smoother is also made to enable computing of the likelihood. This helps us to formulate an expectation maximisation algorithm for maximum likelihood estimation of the state noise and the observation noise. Empirical investigations show that the proposed unscented grid filter/smoother compares favourably to other similar filters on nonlinear estimation tasks.

Keywords:

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1297
1131 Application of Seismic Wave Method in Early Estimation of Wencheng Earthquake

Authors: Wenlong Liu, Yucheng Liu

Abstract:

This paper introduces the application of seismic wave method in earthquake prediction and early estimation. The advantages of the seismic wave method over the traditional earthquake prediction method are demonstrated. An example is presented in this study to show the accuracy and efficiency of using the seismic wave method in predicting a medium-sized earthquake swarm occurred in Wencheng, Zhejiang, China. By applying this method, correct predictions were made on the day after this earthquake swarm started and the day the maximum earthquake occurred, which provided scientific bases for governmental decision-making.

Keywords: earthquake prediction, earthquake swarm, seismicactivity method, seismic wave method, Wencheng earthquake

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1615
1130 Time-Delay Estimation Using Cross-ΨB-Energy Operator

Authors: Z. Saidi, A.O. Boudraa, J.C. Cexus, S. Bourennane

Abstract:

In this paper, a new time-delay estimation technique based on the cross IB-energy operator [5] is introduced. This quadratic energy detector measures how much a signal is present in another one. The location of the peak of the energy operator, corresponding to the maximum of interaction between the two signals, is the estimate of the delay. The method is a fully data-driven approach. The discrete version of the continuous-time form of the cross IBenergy operator, for its implementation, is presented. The effectiveness of the proposed method is demonstrated on real underwater acoustic signals arriving from targets and the results compared to the cross-correlation method.

Keywords: Teager-Kaiser energy operator, Cross-energyoperator, Time-Delay, Underwater acoustic signals.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5619
1129 Orthogonal Polynomial Density Estimates: Alternative Representation and Degree Selection

Authors: Serge B. Provost, Min Jiang

Abstract:

The density estimates considered in this paper comprise a base density and an adjustment component consisting of a linear combination of orthogonal polynomials. It is shown that, in the context of density approximation, the coefficients of the linear combination can be determined either from a moment-matching technique or a weighted least-squares approach. A kernel representation of the corresponding density estimates is obtained. Additionally, two refinements of the Kronmal-Tarter stopping criterion are proposed for determining the degree of the polynomial adjustment. By way of illustration, the density estimation methodology advocated herein is applied to two data sets.

Keywords: kernel density estimation, orthogonal polynomials, moment-based methodologies, density approximation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2338
1128 Fusing Local Binary Patterns with Wavelet Features for Ethnicity Identification

Authors: S. Hma Salah, H. Du, N. Al-Jawad

Abstract:

Ethnicity identification of face images is of interest in many areas of application, but existing methods are few and limited. This paper presents a fusion scheme that uses block-based uniform local binary patterns and Haar wavelet transform to combine local and global features. In particular, the LL subband coefficients of the whole face are fused with the histograms of uniform local binary patterns from block partitions of the face. We applied the principal component analysis on the fused features and managed to reduce the dimensionality of the feature space from 536 down to around 15 without sacrificing too much accuracy. We have conducted a number of preliminary experiments using a collection of 746 subject face images. The test results show good accuracy and demonstrate the potential of fusing global and local features. The fusion approach is robust, making it easy to further improve the identification at both feature and score levels.

Keywords: Ethnicity identification, fusion, local binary patterns, wavelet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2959
1127 Code-Aided Turbo Channel Estimation for OFDM Systems with NB-LDPC Codes

Authors: Ł. Januszkiewicz, G. Bacci, H. Gierszal, M. Luise

Abstract:

In this paper channel estimation techniques are considered as the support methods for OFDM transmission systems based on Non Binary LDPC (Low Density Parity Check) codes. Standard frequency domain pilot aided LS (Least Squares) and LMMSE (Linear Minimum Mean Square Error) estimators are investigated. Furthermore, an iterative algorithm is proposed as a solution exploiting the NB-LDPC channel decoder to improve the performance of the LMMSE estimator. Simulation results of signals transmitted through fading mobile channels are presented to compare the performance of the proposed channel estimators.

Keywords: LDPC codes, LMMSE, OFDM, turbo channelestimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1629
1126 Very Large Scale Integration Architecture of Finite Impulse Response Filter Implementation Using Retiming Technique

Authors: S. Jalaja, A. M. Vijaya Prakash

Abstract:

Recursive combination of an algorithm based on Karatsuba multiplication is exploited to design a generalized transpose and parallel Finite Impulse Response (FIR) Filter. Mid-range Karatsuba multiplication and Carry Save adder based on Karatsuba multiplication reduce time complexity for higher order multiplication implemented up to n-bit. As a result, we design modified N-tap Transpose and Parallel Symmetric FIR Filter Structure using Karatsuba algorithm. The mathematical formulation of the FFA Filter is derived. The proposed architecture involves significantly less area delay product (APD) then the existing block implementation. By adopting retiming technique, hardware cost is reduced further. The filter architecture is designed by using 90 nm technology library and is implemented by using cadence EDA Tool. The synthesized result shows better performance for different word length and block size. The design achieves switching activity reduction and low power consumption by applying with and without retiming for different combination of the circuit. The proposed structure achieves more than a half of the power reduction by adopting with and without retiming techniques compared to the earlier design structure. As a proof of the concept for block size 16 and filter length 64 for CKA method, it achieves a 51% as well as 70% less power by applying retiming technique, and for CSA method it achieves a 57% as well as 77% less power by applying retiming technique compared to the previously proposed design.

Keywords: Carry save adder Karatsuba multiplication, mid-range Karatsuba multiplication, modified FFA, transposed filter, retiming.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 876
1125 Estimation of Critical Period for Weed Control in Corn in Iran

Authors: Sohrab Mahmoodi, Ali Rahimi

Abstract:

The critical period for weed control (CPWC) is the period in the crop growth cycle during which weeds must be controlled to prevent unacceptable yield losses. Field studies were conducted in 2005 and 2006 in the University of Birjand at the south east of Iran to determine CPWC of corn using a randomized complete block design with 14 treatments and four replications. The treatments consisted of two different periods of weed interference, a critical weed-free period and a critical time of weed removal, were imposed at V3, V6, V9, V12, V15, and R1 (based on phonological stages of corn development) with a weedy check and a weed-free check. The CPWC was determined with the use of 2.5, 5, 10, 15 and 20% acceptable yield loss levels by non-linear Regression method and fitting Logistic and Gompertz nonlinear equations to relative yield data. The CPWC of corn was from 5- to 15-leaf stage (19-55 DAE) to prevent yield losses of 5%. This period to prevent yield losses of 2.5, 10 and 20% was 4- to 17-leaf stage (14-59 DAE), 6- to 12-leaf stage (25-47 DAE) and 8- to 9-leaf stage (31-36 DAE) respectively. The height and leaf area index of corn were significantly decreased by weed competition in both weed free and weed infested treatments (P<0.01). Results also showed that there was a significant positive correlation between yield and LAI of corn at silk stage when competing with weeds (r= 0.97).

Keywords: Corn, Critical period, Gompertz, Logistic, Weed control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1997
1124 Fingerprint Image Encryption Using a 2D Chaotic Map and Elliptic Curve Cryptography

Authors: D. M. S. Bandara, Yunqi Lei, Ye Luo

Abstract:

Fingerprints are suitable as long-term markers of human identity since they provide detailed and unique individual features which are difficult to alter and durable over life time. In this paper, we propose an algorithm to encrypt and decrypt fingerprint images by using a specially designed Elliptic Curve Cryptography (ECC) procedure based on block ciphers. In addition, to increase the confusing effect of fingerprint encryption, we also utilize a chaotic-behaved method called Arnold Cat Map (ACM) for a 2D scrambling of pixel locations in our method. Experimental results are carried out with various types of efficiency and security analyses. As a result, we demonstrate that the proposed fingerprint encryption/decryption algorithm is advantageous in several different aspects including efficiency, security and flexibility. In particular, using this algorithm, we achieve a margin of about 0.1% in the test of Number of Pixel Changing Rate (NPCR) values comparing to the-state-of-the-art performances.

Keywords: Arnold cat map, biometric encryption, block cipher, elliptic curve cryptography, fingerprint encryption, Koblitz’s Encoding.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1044
1123 Development of a Speed Sensorless IM Drives

Authors: Dj. Cherifi, Y. Miloud, A. Tahri

Abstract:

The primary objective of this paper is to elimination of the problem of sensitivity to parameter variation of induction motor drive. The proposed sensorless strategy is based on an algorithm permitting a better simultaneous estimation of the rotor speed and the stator resistance including an adaptive mechanism based on the lyaponov theory. To study the reliability and the robustness of the sensorless technique to abnormal operations, some simulation tests have been performed under several cases.

The proposed sensorless vector control scheme showed a good performance behavior in the transient and steady states, with an excellent disturbance rejection of the load torque.

Keywords: Induction Motor Drive, field-oriented control, adaptive speed observer, stator resistance estimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1985
1122 Maximum Likelihood Estimation of Burr Type V Distribution under Left Censored Samples

Authors: N. Feroze, M. Aslam

Abstract:

The paper deals with the maximum likelihood estimation of the parameters of the Burr type V distribution based on left censored samples. The maximum likelihood estimators (MLE) of the parameters have been derived and the Fisher information matrix for the parameters of the said distribution has been obtained explicitly. The confidence intervals for the parameters have also been discussed. A simulation study has been conducted to investigate the performance of the point and interval estimates.

Keywords: Fisher information matrix, confidence intervals, censoring.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1678
1121 Building Information Modeling-Based Approach for Automatic Quantity Take-off and Cost Estimation

Authors: Lo Kar Yin, Law Ka Mei

Abstract:

Architectural, engineering, construction and operations (AECO) industry practitioners have been well adapting to the dynamic construction market from the fundamental training of its disciplines. As further triggered by the pandemic since 2019, great steps are taken in virtual environment and the best collaboration is strived with project teams without boundaries. With adoption of Building Information Modeling-based approach and qualitative analysis, this paper is to review quantity take-off (QTO) and cost estimation process through modeling techniques in liaison with suppliers, fabricators, subcontractors, contractors, designers, consultants and services providers in the construction industry value chain for automatic project cost budgeting, project cost control and cost evaluation on design options of in-situ reinforced-concrete construction and Modular Integrated Construction (MiC) at design stage, variation of works and cash flow/spending analysis at construction stage as far as practicable, with a view to sharing the findings for enhancing mutual trust and co-operation among AECO industry practitioners. It is to foster development through a common prototype of design and build project delivery method in NEC4 Engineering and Construction Contract (ECC) Options A and C.

Keywords: Building Information Modeling, cost estimation, quantity take-off, modeling techniques.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 638
1120 Fast Algorithm of Infrared Point Target Detection in Fluctuant Background

Authors: Yang Weiping, Zhang Zhilong, Li Jicheng, Chen Zengping, He Jun

Abstract:

The background estimation approach using a small window median filter is presented on the bases of analyzing IR point target, noise and clutter model. After simplifying the two-dimensional filter, a simple method of adopting one-dimensional median filter is illustrated to make estimations of background according to the characteristics of IR scanning system. The adaptive threshold is used to segment canceled image in the background. Experimental results show that the algorithm achieved good performance and satisfy the requirement of big size image-s real-time processing.

Keywords: Point target, background estimation, median filter, adaptive threshold, target detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1799
1119 Long Term Examination of the Profitability Estimation Focused on Benefits

Authors: Stephan Printz, Kristina Lahl, René Vossen, Sabina Jeschke

Abstract:

Strategic investment decisions are characterized by high innovation potential and long-term effects on the competitiveness of enterprises. Due to the uncertainty and risks involved in this complex decision making process, the need arises for well-structured support activities. A method that considers cost and the long-term added value is the cost-benefit effectiveness estimation. One of those methods is the “profitability estimation focused on benefits – PEFB”-method developed at the Institute of Management Cybernetics at RWTH Aachen University. The method copes with the challenges associated with strategic investment decisions by integrating long-term non-monetary aspects whilst also mapping the chronological sequence of an investment within the organization’s target system. Thus, this method is characterized as a holistic approach for the evaluation of costs and benefits of an investment. This participation-oriented method was applied to business environments in many workshops. The results of the workshops are a library of more than 96 cost aspects, as well as 122 benefit aspects. These aspects are preprocessed and comparatively analyzed with regards to their alignment to a series of risk levels. For the first time, an accumulation and a distribution of cost and benefit aspects regarding their impact and probability of occurrence are given. The results give evidence that the PEFB-method combines precise measures of financial accounting with the incorporation of benefits. Finally, the results constitute the basics for using information technology and data science for decision support when applying within the PEFB-method.

Keywords: Cost-benefit analysis, multi-criteria decision, profitability estimation focused on benefits, risk and uncertainty analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1459
1118 A Self Configuring System for Object Recognition in Color Images

Authors: Michela Lecca

Abstract:

System MEMORI automatically detects and recognizes rotated and/or rescaled versions of the objects of a database within digital color images with cluttered background. This task is accomplished by means of a region grouping algorithm guided by heuristic rules, whose parameters concern some geometrical properties and the recognition score of the database objects. This paper focuses on the strategies implemented in MEMORI for the estimation of the heuristic rule parameters. This estimation, being automatic, makes the system a highly user-friendly tool.

Keywords: Automatic object recognition, clustering, content based image retrieval system, image segmentation, region adjacency graph, region grouping.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1379