Search results for: inverse analysis
8682 Electromagnetic Imaging of Inhomogeneous Dielectric Cylinders Buried in a Slab Mediumby TE Wave Illumination
Authors: Chung-Hsin Huang, Chien-Ching Chiu, Chun Jen Lin
Abstract:
The electromagnetic imaging of inhomogeneous dielectric cylinders buried in a slab medium by transverse electric (TE) wave illumination is investigated. Dielectric cylinders of unknown permittivities are buried in second space and scattered a group of unrelated waves incident from first space where the scattered field is recorded. By proper arrangement of the various unrelated incident fields, the difficulties of ill-posedness and nonlinearity are circumvented, and the permittivity distribution can be reconstructed through simple matrix operations. The algorithm is based on the moment method and the unrelated illumination method. Numerical results are given to demonstrate the capability of the inverse algorithm. Good reconstruction is obtained even in the presence of additive Gaussian random noise in measured data. In addition, the effect of noise on the reconstruction result is also investigated.Keywords: Slab Medium, Unrelated Illumination Method, TEWave Illumination, Inhomogeneous Cylinders.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12038681 Modified Scaling-Free CORDIC Based Pipelined Parallel MDC FFT and IFFT Architecture for Radix 2^2 Algorithm
Authors: C. Paramasivam, K. B. Jayanthi
Abstract:
An innovative approach to develop modified scaling free CORDIC based two parallel pipelined Multipath Delay Commutator (MDC) FFT and IFFT architectures for radix 22 FFT algorithm is presented. Multipliers and adders are the most important data paths in FFT and IFFT architectures. Multipliers occupy high area and consume more power. In order to optimize the area and power overhead, modified scaling-free CORDIC based complex multiplier is utilized in the proposed design. In general twiddle factor values are stored in RAM block. In the proposed work, modified scaling-free CORDIC based twiddle factor generator unit is used to generate the twiddle factor and efficient switching units are used. In addition to this, four point FFT operations are performed without complex multiplication which helps to reduce area and power in the last two stages of the pipelined architectures. The design proposed in this paper is based on multipath delay commutator method. The proposed design can be extended to any radix 2n based FFT/IFFT algorithm to improve the throughput. The work is synthesized using Synopsys design Compiler using TSMC 90-nm library. The proposed method proves to be better compared to the reference design in terms of area, throughput and power consumption. The comparative analysis of the proposed design with Xilinx FPGA platform is also discussed in the paper.Keywords: Coordinate Rotational Digital Computer(CORDIC), Complex multiplier, Fast Fourier transform (FFT), Inverse fast Fourier transform (IFFT), Multipath delay Commutator (MDC), modified scaling free CORDIC, complex multiplier, pipelining, parallel processing, radix-2^2.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18178680 Zero Inflated Models for Overdispersed Count Data
Authors: Y. N. Phang, E. F. Loh
Abstract:
The zero inflated models are usually used in modeling count data with excess zeros where the existence of the excess zeros could be structural zeros or zeros which occur by chance. These type of data are commonly found in various disciplines such as finance, insurance, biomedical, econometrical, ecology, and health sciences which involve sex and health dental epidemiology. The most popular zero inflated models used by many researchers are zero inflated Poisson and zero inflated negative binomial models. In addition, zero inflated generalized Poisson and zero inflated double Poisson models are also discussed and found in some literature. Recently zero inflated inverse trinomial model and zero inflated strict arcsine models are advocated and proven to serve as alternative models in modeling overdispersed count data caused by excessive zeros and unobserved heterogeneity. The purpose of this paper is to review some related literature and provide a variety of examples from different disciplines in the application of zero inflated models. Different model selection methods used in model comparison are discussed.
Keywords: Overdispersed count data, model selection methods, likelihood ratio, AIC, BIC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 45318679 Evaluating the Understanding of the University Students (Basic Sciences and Engineering) about the Numerical Representation of the Average Rate of Change
Authors: Saeid Haghjoo, Ebrahim Reyhani, Fahimeh Kolahdouz
Abstract:
The present study aimed to evaluate the understanding of the students in Tehran universities (Iran) about the numerical representation of the average rate of change based on the Structure of Observed Learning Outcomes (SOLO). In the present descriptive-survey research, the statistical population included undergraduate students (basic sciences and engineering) in the universities of Tehran. The samples were 604 students selected by random multi-stage clustering. The measurement tool was a task whose face and content validity was confirmed by math and mathematics education professors. Using Cronbach's Alpha criterion, the reliability coefficient of the task was obtained 0.95, which verified its reliability. The collected data were analyzed by descriptive statistics and inferential statistics (chi-squared and independent t-tests) under SPSS-24 software. According to the SOLO model in the prestructural, unistructural, and multistructural levels, basic science students had a higher percentage of understanding than that of engineering students, although the outcome was inverse at the relational level. However, there was no significant difference in the average understanding of both groups. The results indicated that students failed to have a proper understanding of the numerical representation of the average rate of change, in addition to missconceptions when using physics formulas in solving the problem. In addition, multiple solutions were derived along with their dominant methods during the qualitative analysis. The current research proposed to focus on the context problems with approximate calculations and numerical representation, using software and connection common relations between math and physics in the teaching process of teachers and professors.
Keywords: Average rate of change, context problems, derivative, numerical representation, SOLO taxonomy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7598678 Multi Band Frequency Synthesizer Based on ISPD PLL with Adapted LC Tuned VCO
Authors: Bilel Gassara, Mahmoud Abdellaoui, Nouri Masmoud
Abstract:
The 4G front-end transceiver needs a high performance which can be obtained mainly with an optimal architecture and a multi-band Local Oscillator. In this study, we proposed and presented a new architecture of multi-band frequency synthesizer based on an Inverse Sine Phase Detector Phase Locked Loop (ISPD PLL) without any filters and any controlled gain block and associated with adapted multi band LC tuned VCO using a several numeric controlled capacitive branches but not binary weighted. The proposed architecture, based on 0.35μm CMOS process technology, supporting Multi-band GSM/DCS/DECT/ UMTS/WiMax application and gives a good performances: a phase noise @1MHz -127dBc and a Factor Of Merit (FOM) @ 1MHz - 186dB and a wide band frequency range (from 0.83GHz to 3.5GHz), that make the proposed architecture amenable for monolithic integration and 4G multi-band application.Keywords: GSM/DCS/DECT/UMTS/WiMax, ISPD PLL, keep and capture range, Multi-Band, Synthesizer, Wireless.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19998677 Advanced Robust PDC Fuzzy Control of Nonlinear Systems
Authors: M. Polanský
Abstract:
This paper introduces a new method called ARPDC (Advanced Robust Parallel Distributed Compensation) for automatic control of nonlinear systems. This method improves a quality of robust control by interpolating of robust and optimal controller. The weight of each controller is determined by an original criteria function for model validity and disturbance appreciation. ARPDC method is based on nonlinear Takagi-Sugeno (T-S) fuzzy systems and Parallel Distributed Compensation (PDC) control scheme. The relaxed stability conditions of ARPDC control of nominal system have been derived. The advantages of presented method are demonstrated on the inverse pendulum benchmark problem. From comparison between three different controllers (robust, optimal and ARPDC) follows, that ARPDC control is almost optimal with the robustness close to the robust controller. The results indicate that ARPDC algorithm can be a good alternative not only for a robust control, but in some cases also to an adaptive control of nonlinear systems.
Keywords: Robust control, optimal control, Takagi–Sugeno (TS) fuzzy models, linear matrix inequality (LMI), observer, Advanced Robust Parallel Distributed Compensation (ARPDC).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15748676 Magnetic Properties and Cytotoxicity of Ga-Mn Magnetic Ferrites Synthesized by the Citrate Sol-Gel Method
Authors: Javier Sánchez, Laura Elena De León Prado, Dora Alicia Cortés Hernández
Abstract:
Magnetic spinel ferrites are materials that possess size, magnetic properties and heating ability adequate for their potential use in biomedical applications. The Mn0.5Ga0.5Fe2O4 magnetic nanoparticles (MNPs) were synthesized by sol-gel method using citric acid as chelating agent of metallic precursors. The synthesized samples were identified by X-Ray Diffraction (XRD) as an inverse spinel structure with no secondary phases. Saturation magnetization (Ms) of crystalline powders was 45.9 emu/g, which was higher than those corresponding to GaFe2O4 (14.2 emu/g) and MnFe2O4 (40.2 emu/g) synthesized under similar conditions, while the coercivity field (Hc) was 27.9 Oe. The average particle size was 18 ± 7 nm. The heating ability of the MNPs was enough to increase the surrounding temperature up to 43.5 °C in 7 min when a quantity of 4.5 mg of MNPs per mL of liquid medium was tested. Cytotoxic effect (hemolysis assay) of MNPs was determined and the results showed hemolytic values below 1% in all tested cases. According to the results obtained, these synthesized nanoparticles can be potentially used as thermoseeds for hyperthermia therapy.
Keywords: Cytotoxicity, heating ability, manganese-gallium ferrite, magnetic hyperthermia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13158675 Effect of Zidovudine on Hematological and Virologic Parameters among Female Sex Workers Receiving Antiretroviral Therapy (ART) in North – Western Nigeria
Authors: N. M. Sani, E. D. Jatau, O. S. Olonitola, M. Y. Gwarzo, P. Moodley, N. S. Mujahid
Abstract:
Hemoglobin (HB) indicates anemia level and by extension may reflect the nutritional level and perhaps the immunity of an individual. Some antiretroviral drugs like Zidovudine are known to cause anemia in people living with HIV/AIDS (PLWHA). A cross sectional study using demographic data and blood specimen from 218 female commercial sex workers attending antiretroviral therapy (ART) clinics was conducted between December, 2009 and July, 2011 to assess the effect of zidovudine on hematologic, and RNA viral load of female sex workers receiving antiretroviral treatment in north western Nigeria. Anemia is a common and serious complication of both HIV infection and its treatment. In the setting of HIV infection, anemia has been associated with decreased quality of life, functional status, and survival. Antiretroviral therapy, particularly the highly active antiretroviral therapy (HAART), has been associated with a decrease in the incidence and severity of anemia in HIV-infected patients who have received a HAART regimen for at least 1 year. In this study, result has shown that of the 218 patients, 26 with hemoglobin count between 5.1 – 10g/dl were observed to have the highest viral load count of 300,000 – 350,000copies/ml. It was also observed that most patients (190) with HB of 10.1 – 15.0g/dl had viral load count of 200,000 – 250,000 copies /ml. An inverse relationship therefore exists i.e. the lower the hemoglobin level, the higher the viral load count even though the test statistics did not show any significance between the two (P = 0.206). This shows that multivariate logistic regression analysis demonstrated that anemia was associated with a CD4 + cell count below 50/μL, female sex workers with a viral load above 100,000 copies/mL, who use zidovudine. Severe anemia was less prevalent in this study population than in historical comparators; however, mild to moderate anemia rates remain high. The study therefore recommends that hematological and virologic parameters be monitored closely in patients receiving first line ART regimen.Keywords: Female sex worker, Zidovudine, Hemoglobin, Anemia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17628674 Variable Step-Size Affine Projection Algorithm With a Weighted and Regularized Projection Matrix
Authors: Tao Dai, Andy Adler, Behnam Shahrrava
Abstract:
This paper presents a forgetting factor scheme for variable step-size affine projection algorithms (APA). The proposed scheme uses a forgetting processed input matrix as the projection matrix of pseudo-inverse to estimate system deviation. This method introduces temporal weights into the projection matrix, which is typically a better model of the real error's behavior than homogeneous temporal weights. The regularization overcomes the ill-conditioning introduced by both the forgetting process and the increasing size of the input matrix. This algorithm is tested by independent trials with coloured input signals and various parameter combinations. Results show that the proposed algorithm is superior in terms of convergence rate and misadjustment compared to existing algorithms. As a special case, a variable step size NLMS with forgetting factor is also presented in this paper.
Keywords: Adaptive signal processing, affine projection algorithms, variable step-size adaptive algorithms, regularization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16308673 Identifying the Kinematic Parameters of Hexapod Machine Tool
Authors: M. M. Agheli, M. J. Nategh
Abstract:
Hexapod Machine Tool (HMT) is a parallel robot mostly based on Stewart platform. Identification of kinematic parameters of HMT is an important step of calibration procedure. In this paper an algorithm is presented for identifying the kinematic parameters of HMT using inverse kinematics error model. Based on this algorithm, the calibration procedure is simulated. Measurement configurations with maximum observability are decided as the first step of this algorithm for a robust calibration. The errors occurring in various configurations are illustrated graphically. It has been shown that the boundaries of the workspace should be searched for the maximum observability of errors. The importance of using configurations with sufficient observability in calibrating hexapod machine tools is verified by trial calibration with two different groups of randomly selected configurations. One group is selected to have sufficient observability and the other is in disregard of the observability criterion. Simulation results confirm the validity of the proposed identification algorithm.Keywords: Calibration, Hexapod Machine Tool (HMT), InverseKinematics Error Model, Observability, Parallel Robot, ParameterIdentification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23668672 Biomechanics Analysis When Delivering Baby
Authors: Kristyanto B.
Abstract:
Plenty of analyses based on Biomechanics were carried out on many jobs in manufactures or services. Now Biomechanics analysis is being applied on mothers who are giving birth. The analysis conducted in terms of normal condition of the birth process without Gyn Bed (Obstetric Bed). The aim of analysis is to study whether it is risky or not when choosing the position of mother’s postures when delivering the baby. This investigation was applied on two positions that generally appear in common birth process. Results will show the analysis of both positions to support the birth process based on the Biomechanics analysis (Ergonomic approaches).
Keywords: Biomechanics analysis, Birth process, Position of postures analysis, Ergonomic approaches.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23018671 Neural Network Based Icing Identification and Fault Tolerant Control of a 340 Aircraft
Authors: F. Caliskan
Abstract:
This paper presents a Neural Network (NN) identification of icing parameters in an A340 aircraft and a reconfiguration technique to keep the A/C performance close to the performance prior to icing. Five aircraft parameters are assumed to be considerably affected by icing. The off-line training for identifying the clear and iced dynamics is based on the Levenberg-Marquard Backpropagation algorithm. The icing parameters are located in the system matrix. The physical locations of the icing are assumed at the right and left wings. The reconfiguration is based on the technique known as the control mixer approach or pseudo inverse technique. This technique generates the new control input vector such that the A/C dynamics is not much affected by icing. In the simulations, the longitudinal and lateral dynamics of an Airbus A340 aircraft model are considered, and the stability derivatives affected by icing are identified. The simulation results show the successful NN identification of the icing parameters and the reconfigured flight dynamics having the similar performance before the icing. In other words, the destabilizing icing affect is compensated.Keywords: Aircraft Icing, Stability Derivatives, Neural NetworkIdentification, Reconfiguration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17018670 Bond Graph Modeling of Mechanical Dynamics of an Excavator for Hydraulic System Analysis and Design
Authors: Mutuku Muvengei, John Kihiu
Abstract:
This paper focuses on the development of bond graph dynamic model of the mechanical dynamics of an excavating mechanism previously designed to be used with small tractors, which are fabricated in the Engineering Workshops of Jomo Kenyatta University of Agriculture and Technology. To develop a mechanical dynamics model of the manipulator, forward recursive equations similar to those applied in iterative Newton-Euler method were used to obtain kinematic relationships between the time rates of joint variables and the generalized cartesian velocities for the centroids of the links. Representing the obtained kinematic relationships in bondgraphic form, while considering the link weights and momenta as the elements led to a detailed bond graph model of the manipulator. The bond graph method was found to reduce significantly the number of recursive computations performed on a 3 DOF manipulator for a mechanical dynamic model to result, hence indicating that bond graph method is more computationally efficient than the Newton-Euler method in developing dynamic models of 3 DOF planar manipulators. The model was verified by comparing the joint torque expressions of a two link planar manipulator to those obtained using Newton- Euler and Lagrangian methods as analyzed in robotic textbooks. The expressions were found to agree indicating that the model captures the aspects of rigid body dynamics of the manipulator. Based on the model developed, actuator sizing and valve sizing methodologies were developed and used to obtain the optimal sizes of the pistons and spool valve ports respectively. It was found that using the pump with the sized flow rate capacity, the engine of the tractor is able to power the excavating mechanism in digging a sandy-loom soil.Keywords: Actuators, bond graphs, inverse dynamics, recursive equations, quintic polynomial trajectory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28978669 Joint Use of Factor Analysis (FA) and Data Envelopment Analysis (DEA) for Ranking of Data Envelopment Analysis
Authors: Reza Nadimi, Fariborz Jolai
Abstract:
This article combines two techniques: data envelopment analysis (DEA) and Factor analysis (FA) to data reduction in decision making units (DMU). Data envelopment analysis (DEA), a popular linear programming technique is useful to rate comparatively operational efficiency of decision making units (DMU) based on their deterministic (not necessarily stochastic) input–output data and factor analysis techniques, have been proposed as data reduction and classification technique, which can be applied in data envelopment analysis (DEA) technique for reduction input – output data. Numerical results reveal that the new approach shows a good consistency in ranking with DEA.Keywords: Effectiveness, Decision Making, Data EnvelopmentAnalysis, Factor Analysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24248668 Movement Optimization of Robotic Arm Movement Using Soft Computing
Authors: V. K. Banga
Abstract:
Robots are now playing a very promising role in industries. Robots are commonly used in applications in repeated operations or where operation by human is either risky or not feasible. In most of the industrial applications, robotic arm manipulators are widely used. Robotic arm manipulator with two link or three link structures is commonly used due to their low degrees-of-freedom (DOF) movement. As the DOF of robotic arm increased, complexity increases. Instrumentation involved with robotics plays very important role in order to interact with outer environment. In this work, optimal control for movement of various DOFs of robotic arm using various soft computing techniques has been presented. We have discussed about different robotic structures having various DOF robotics arm movement. Further stress is on kinematics of the arm structures i.e. forward kinematics and inverse kinematics. Trajectory planning of robotic arms using soft computing techniques is demonstrating the flexibility of this technique. The performance is optimized for all possible input values and results in optimized movement as resultant output. In conclusion, soft computing has been playing very important role for achieving optimized movement of robotic arm. It also requires very limited knowledge of the system to implement soft computing techniques.
Keywords: Artificial intelligence, kinematics, robotic arm, neural networks, fuzzy logic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17768667 Passenger Seat Vibration Control of Quarter Car System with MR Shock Absorber
Authors: Devdutt, M. L. Aggarwal
Abstract:
Semi-active Fuzzy control of quarter car system having three degrees of freedom and assembled with magneto-rheological (MR) shock absorber is studied in present paper. First, experimental work was performed on an MR shock absorber under different excitation conditions to obtain force-displacement and force-velocity curves. Then, for the application of experimental data in semi-active quarter car system, a polynomial model was selected. Finally, Fuzzy logic controller was designed having the combination of Forward fuzzy controller and Inverse fuzzy controller for integration in secondary suspension system of concerned model. The proposed controlled quarter car model was compared with uncontrolled system using simulation work under bump type of road excitation. Results obtained by simulation work shows the effectiveness of fuzzy controlled suspension system in improving the ride comfort and safety of travelling passengers compared to uncontrolled suspension system.
Keywords: MR shock absorber, three degrees of freedom, quarter car model, fuzzy controller.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32958666 Performance Evaluation of Refinement Method for Wideband Two-Beams Formation
Authors: C. Bunsanit
Abstract:
This paper presents the refinement method for two beams formation of wideband smart antenna. The refinement method for weighting coefficients is based on Fully Spatial Signal Processing by taking Inverse Discrete Fourier Transform (IDFT), and its simulation results are presented using MATLAB. The radiation pattern is created by multiplying the incoming signal with real weights and then summing them together. These real weighting coefficients are computed by IDFT method; however, the range of weight values is relatively wide. Therefore, for reducing this range, the refinement method is used. The radiation pattern concerns with five input parameters to control. These parameters are maximum weighting coefficient, wideband signal, direction of mainbeam, beamwidth, and maximum of minor lobe level. Comparison of the obtained simulation results between using refinement method and taking only IDFT shows that the refinement method works well for wideband two beams formation.
Keywords: Fully spatial signal processing, beam forming, refinement method, smart antenna, weighting coefficient, wideband.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10778665 A Sparse Representation Speech Denoising Method Based on Adapted Stopping Residue Error
Authors: Qianhua He, Weili Zhou, Aiwu Chen
Abstract:
A sparse representation speech denoising method based on adapted stopping residue error was presented in this paper. Firstly, the cross-correlation between the clean speech spectrum and the noise spectrum was analyzed, and an estimation method was proposed. In the denoising method, an over-complete dictionary of the clean speech power spectrum was learned with the K-singular value decomposition (K-SVD) algorithm. In the sparse representation stage, the stopping residue error was adaptively achieved according to the estimated cross-correlation and the adjusted noise spectrum, and the orthogonal matching pursuit (OMP) approach was applied to reconstruct the clean speech spectrum from the noisy speech. Finally, the clean speech was re-synthesised via the inverse Fourier transform with the reconstructed speech spectrum and the noisy speech phase. The experiment results show that the proposed method outperforms the conventional methods in terms of subjective and objective measure.
Keywords: Speech denoising, sparse representation, K-singular value decomposition, orthogonal matching pursuit.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10138664 Image Dehazing Using Dark Channel Prior and Fast Guided Filter in Daubechies Lifting Wavelet Transform Domain
Authors: Harpreet Kaur, Sudipta Majumdar
Abstract:
In this paper a method for image dehazing is proposed in lifting wavelet transform domain. Lifting Daubechies (D4) wavelet has been used to obtain the approximate image and detail images. As the haze is contained in low frequency part, only the approximate image is used for further processing. This region is processed by dehazing algorithm based on dark channel prior (DCP). The dehazed approximate image is then recombined with the detail images using inverse lifting wavelet transform. Implementation of lifting wavelet transform has the advantage of auxiliary memory saving, fast implementation and simplicity. Also, the proposed method deals with near white scene problem, blue horizon issue and localized light sources in a way to enhance image quality and makes the algorithm robust. Simulation results present improvement in terms of visual quality, parameters such as root mean square (RMS) contrast, structural similarity index (SSIM), entropy and execution time.
Keywords: Dark channel prior, image dehazing, lifting wavelet transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11228663 A Normalization-based Robust Image Watermarking Scheme Using SVD and DCT
Authors: Say Wei Foo, Qi Dong
Abstract:
Digital watermarking is one of the techniques for copyright protection. In this paper, a normalization-based robust image watermarking scheme which encompasses singular value decomposition (SVD) and discrete cosine transform (DCT) techniques is proposed. For the proposed scheme, the host image is first normalized to a standard form and divided into non-overlapping image blocks. SVD is applied to each block. By concatenating the first singular values (SV) of adjacent blocks of the normalized image, a SV block is obtained. DCT is then carried out on the SV blocks to produce SVD-DCT blocks. A watermark bit is embedded in the highfrequency band of a SVD-DCT block by imposing a particular relationship between two pseudo-randomly selected DCT coefficients. An adaptive frequency mask is used to adjust local watermark embedding strength. Watermark extraction involves mainly the inverse process. The watermark extracting method is blind and efficient. Experimental results show that the quality degradation of watermarked image caused by the embedded watermark is visually transparent. Results also show that the proposed scheme is robust against various image processing operations and geometric attacks.Keywords: Image watermarking, Image normalization, Singularvalue decomposition, Discrete cosine transform, Robustness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20958662 Multimedia Data Fusion for Event Detection in Twitter by Using Dempster-Shafer Evidence Theory
Authors: Samar M. Alqhtani, Suhuai Luo, Brian Regan
Abstract:
Data fusion technology can be the best way to extract useful information from multiple sources of data. It has been widely applied in various applications. This paper presents a data fusion approach in multimedia data for event detection in twitter by using Dempster-Shafer evidence theory. The methodology applies a mining algorithm to detect the event. There are two types of data in the fusion. The first is features extracted from text by using the bag-ofwords method which is calculated using the term frequency-inverse document frequency (TF-IDF). The second is the visual features extracted by applying scale-invariant feature transform (SIFT). The Dempster - Shafer theory of evidence is applied in order to fuse the information from these two sources. Our experiments have indicated that comparing to the approaches using individual data source, the proposed data fusion approach can increase the prediction accuracy for event detection. The experimental result showed that the proposed method achieved a high accuracy of 0.97, comparing with 0.93 with texts only, and 0.86 with images only.Keywords: Data fusion, Dempster-Shafer theory, data mining, event detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17988661 Codebook Generation for Vector Quantization on Orthogonal Polynomials based Transform Coding
Authors: R. Krishnamoorthi, N. Kannan
Abstract:
In this paper, a new algorithm for generating codebook is proposed for vector quantization (VQ) in image coding. The significant features of the training image vectors are extracted by using the proposed Orthogonal Polynomials based transformation. We propose to generate the codebook by partitioning these feature vectors into a binary tree. Each feature vector at a non-terminal node of the binary tree is directed to one of the two descendants by comparing a single feature associated with that node to a threshold. The binary tree codebook is used for encoding and decoding the feature vectors. In the decoding process the feature vectors are subjected to inverse transformation with the help of basis functions of the proposed Orthogonal Polynomials based transformation to get back the approximated input image training vectors. The results of the proposed coding are compared with the VQ using Discrete Cosine Transform (DCT) and Pairwise Nearest Neighbor (PNN) algorithm. The new algorithm results in a considerable reduction in computation time and provides better reconstructed picture quality.
Keywords: Orthogonal Polynomials, Image Coding, Vector Quantization, TSVQ, Binary Tree Classifier
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21488660 Feasibility Analysis Studies on New National R&D Programs in Korea
Authors: Seongmin Yim, Hyun-Kyu Kang
Abstract:
As a part of evaluation system for R&D program, the Korean government has applied feasibility analysis since 2008. Various professionals put forth a great effort in order to catch up the high degree of freedom of R&D programs, and make contributions to evolving the feasibility analysis. We analyze diverse R&D programs from various viewpoints, such as technology, policy, and Economics, integrate the separate analysis, and finally arrive at a definite result; whether a program is feasible or unfeasible. This paper describes the concept and method of the feasibility analysis as a decision making tool. The analysis unit and content of each criterion, which are key elements in a comprehensive decision making structure, are examinedKeywords: Decision Making of New Government R&D Program, Feasibility Analysis Study
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14078659 A Self Organized Map Method to Classify Auditory-Color Synesthesia from Frontal Lobe Brain Blood Volume
Authors: Takashi Kaburagi, Takamasa Komura, Yosuke Kurihara
Abstract:
Absolute pitch is the ability to identify a musical note without a reference tone. Training for absolute pitch often occurs in preschool education. It is necessary to clarify how well the trainee can make use of synesthesia in order to evaluate the effect of the training. To the best of our knowledge, there are no existing methods for objectively confirming whether the subject is using synesthesia. Therefore, in this study, we present a method to distinguish the use of color-auditory synesthesia from the separate use of color and audition during absolute pitch training. This method measures blood volume in the prefrontal cortex using functional Near-infrared spectroscopy (fNIRS) and assumes that the cognitive step has two parts, a non-linear step and a linear step. For the linear step, we assume a second order ordinary differential equation. For the non-linear part, it is extremely difficult, if not impossible, to create an inverse filter of such a complex system as the brain. Therefore, we apply a method based on a self-organizing map (SOM) and are guided by the available data. The presented method was tested using 15 subjects, and the estimation accuracy is reported.
Keywords: Absolute pitch, functional near-infrared spectroscopy, prefrontal cortex, synesthesia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9768658 Instability of Ties in Compression
Authors: T. Cornelius
Abstract:
Masonry cavity walls are loaded by wind pressure and vertical load from upper floors. These loads results in bending moments and compression forces in the ties connecting the outer and the inner wall in a cavity wall. Large cavity walls are furthermore loaded by differential movements from the temperature gradient between the outer and the inner wall, which results in critical increase of the bending moments in the ties. Since the ties are loaded by combined compression and moment forces, the loadbearing capacity is derived from instability equilibrium equations. Most of them are iterative, since exact instability solutions are complex to derive, not to mention the extra complexity introducing dimensional instability from the temperature gradients. Using an inverse variable substitution and comparing an exact theory with an analytical instability solution a method to design tie-connectors in cavity walls was developed. The method takes into account constraint conditions limiting the free length of the wall tie, and the instability in case of pure compression which gives an optimal load bearing capacity. The model is illustrated with examples from praxis.
Keywords: Masonry, tie connectors, cavity wall, instability, differential movements, combined bending and compression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16948657 Efficient Semi-Systolic Finite Field Multiplier Using Redundant Basis
Authors: Hyun-Ho Lee, Kee-Won Kim
Abstract:
The arithmetic operations over GF(2m) have been extensively used in error correcting codes and public-key cryptography schemes. Finite field arithmetic includes addition, multiplication, division and inversion operations. Addition is very simple and can be implemented with an extremely simple circuit. The other operations are much more complex. The multiplication is the most important for cryptosystems, such as the elliptic curve cryptosystem, since computing exponentiation, division, and computing multiplicative inverse can be performed by computing multiplication iteratively. In this paper, we present a parallel computation algorithm that operates Montgomery multiplication over finite field using redundant basis. Also, based on the multiplication algorithm, we present an efficient semi-systolic multiplier over finite field. The multiplier has less space and time complexities compared to related multipliers. As compared to the corresponding existing structures, the multiplier saves at least 5% area, 50% time, and 53% area-time (AT) complexity. Accordingly, it is well suited for VLSI implementation and can be easily applied as a basic component for computing complex operations over finite field, such as inversion and division operation.Keywords: Finite field, Montgomery multiplication, systolic array, cryptography.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16458656 Optimization of the Characteristic Straight Line Method by a “Best Estimate“ of Observed, Normal Orthometric Elevation Differences
Authors: Mahmoud M. S. Albattah
Abstract:
In this paper, to optimize the “Characteristic Straight Line Method" which is used in the soil displacement analysis, a “best estimate" of the geodetic leveling observations has been achieved by taking in account the concept of 'Height systems'. This concept has been discussed in detail and consequently the concept of “height". In landslides dynamic analysis, the soil is considered as a mosaic of rigid blocks. The soil displacement has been monitored and analyzed by using the “Characteristic Straight Line Method". Its characteristic components have been defined constructed from a “best estimate" of the topometric observations. In the measurement of elevation differences, we have used the most modern leveling equipment available. Observational procedures have also been designed to provide the most effective method to acquire data. In addition systematic errors which cannot be sufficiently controlled by instrumentation or observational techniques are minimized by applying appropriate corrections to the observed data: the level collimation correction minimizes the error caused by nonhorizontality of the leveling instrument's line of sight for unequal sight lengths, the refraction correction is modeled to minimize the refraction error caused by temperature (density) variation of air strata, the rod temperature correction accounts for variation in the length of the leveling rod' s Invar/LO-VAR® strip which results from temperature changes, the rod scale correction ensures a uniform scale which conforms to the international length standard and the introduction of the concept of the 'Height systems' where all types of height (orthometric, dynamic, normal, gravity correction, and equipotential surface) have been investigated. The “Characteristic Straight Line Method" is slightly more convenient than the “Characteristic Circle Method". It permits to evaluate a displacement of very small magnitude even when the displacement is of an infinitesimal quantity. The inclination of the landslide is given by the inverse of the distance reference point O to the “Characteristic Straight Line". Its direction is given by the bearing of the normal directed from point O to the Characteristic Straight Line (Fig..6). A “best estimate" of the topometric observations was used to measure the elevation of points carefully selected, before and after the deformation. Gross errors have been eliminated by statistical analyses and by comparing the heights within local neighborhoods. The results of a test using an area where very interesting land surface deformation occurs are reported. Monitoring with different options and qualitative comparison of results based on a sufficient number of check points are presented.
Keywords: Characteristic straight line method, dynamic height, landslides, orthometric height, systematic errors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15658655 Recent Trends in Nonlinear Methods of HRV Analysis: A Review
Authors: Ramesh K. Sunkaria
Abstract:
The linear methods of heart rate variability analysis such as non-parametric (e.g. fast Fourier transform analysis) and parametric methods (e.g. autoregressive modeling) has become an established non-invasive tool for marking the cardiac health, but their sensitivity and specificity were found to be lower than expected with positive predictive value <30%. This may be due to considering the RR-interval series as stationary and re-sampling them prior to their use for analysis, whereas actually it is not. This paper reviews the non-linear methods of HRV analysis such as correlation dimension, largest Lyupnov exponent, power law slope, fractal analysis, detrended fluctuation analysis, complexity measure etc. which are currently becoming popular as these uses the actual RR-interval series. These methods are expected to highly accurate cardiac health prognosis.Keywords: chaos, nonlinear dynamics, sample entropy, approximate entropy, detrended fluctuation analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23508654 A Comparison of Adaline and MLP Neural Network based Predictors in SIR Estimation in Mobile DS/CDMA Systems
Authors: Nahid Ardalani, Ahmadreza Khoogar, H. Roohi
Abstract:
In this paper we compare the response of linear and nonlinear neural network-based prediction schemes in prediction of received Signal-to-Interference Power Ratio (SIR) in Direct Sequence Code Division Multiple Access (DS/CDMA) systems. The nonlinear predictor is Multilayer Perceptron MLP and the linear predictor is an Adaptive Linear (Adaline) predictor. We solve the problem of complexity by using the Minimum Mean Squared Error (MMSE) principle to select the optimal predictors. The optimized Adaline predictor is compared to optimized MLP by employing noisy Rayleigh fading signals with 1.8 GHZ carrier frequency in an urban environment. The results show that the Adaline predictor can estimates SIR with the same error as MLP when the user has the velocity of 5 km/h and 60 km/h but by increasing the velocity up-to 120 km/h the mean squared error of MLP is two times more than Adaline predictor. This makes the Adaline predictor (with lower complexity) more suitable than MLP for closed-loop power control where efficient and accurate identification of the time-varying inverse dynamics of the multi path fading channel is required.Keywords: Power control, neural networks, DS/CDMA mobilecommunication systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25148653 Oxidation of Amitriptyline by Bromamine-T in Acidic Buffer Medium: A Kinetic and Mechanistic Approach
Authors: Chandrashekar, R. T. Radhika, B. M. Venkatesha, S. Ananda, Shivalingegowda, T. S. Shashikumar, H. Ramachandra
Abstract:
The kinetics of the oxidation of amitriptyline (AT) by sodium N-bromotoluene sulphonamide (C6H5SO2NBrNa) has been studied in an acidic buffer medium of pH 1.2 at 303 K. The oxidation reaction of AT was followed spectrophotometrically at maximum wavelength, 410 nm. The reaction rate shows a first order dependence each on concentration of AT and concentration of sodium N-bromotoluene sulphonamide. The reaction also shows an inverse fractional order dependence at low or high concentration of HCl. The dielectric constant of the solvent shows negative effect on the rate of reaction. The addition of halide ions and the reduction product of BAT have no significant effect on the rate. The rate is unchanged with the variation in the ionic strength (NaClO4) of the medium. Addition of reaction mixtures to be aqueous acrylamide solution did not initiate polymerization, indicating the absence of free radical species. The stoichiometry of the reaction was found to be 1:1 and oxidation product of AT is identified. The Michaelis-Menton type of kinetics has been proposed. The CH3C6H5SO2NHBr has been assumed to be the reactive oxidizing species. Thermodynamical parameters were computed by studying the reactions at different temperatures. A mechanism consistent with observed kinetics is presented.
Keywords: Amitriptyline, bromamine-T, kinetics, oxidation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1448