Search results for: Quality Index Method (QIM)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11054

Search results for: Quality Index Method (QIM)

9344 State Estimation Method Based on Unscented Kalman Filter for Vehicle Nonlinear Dynamics

Authors: Wataru Nakamura, Tomoaki Hashimoto, Liang-Kuang Chen

Abstract:

This paper provides a state estimation method for automatic control systems of nonlinear vehicle dynamics. A nonlinear tire model is employed to represent the realistic behavior of a vehicle. In general, all the state variables of control systems are not precisedly known, because those variables are observed through output sensors and limited parts of them might be only measurable. Hence, automatic control systems must incorporate some type of state estimation. It is needed to establish a state estimation method for nonlinear vehicle dynamics with restricted measurable state variables. For this purpose, unscented Kalman filter method is applied in this study for estimating the state variables of nonlinear vehicle dynamics. The objective of this paper is to propose a state estimation method using unscented Kalman filter for nonlinear vehicle dynamics. The effectiveness of the proposed method is verified by numerical simulations.

Keywords: State estimation, control systems, observer systems, unscented Kalman filter, nonlinear vehicle dynamics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 613
9343 An Approximation Method for Exact Boundary Controllability of Euler-Bernoulli System

Authors: Abdelaziz Khernane, Naceur Khelil, Leila Djerou

Abstract:

The aim of this work is to study the numerical implementation of the Hilbert Uniqueness Method for the exact boundary controllability of Euler-Bernoulli beam equation. This study may be difficult. This will depend on the problem under consideration (geometry, control and dimension) and the numerical method used. Knowledge of the asymptotic behaviour of the control governing the system at time T may be useful for its calculation. This idea will be developed in this study. We have characterized as a first step, the solution by a minimization principle and proposed secondly a method for its resolution to approximate the control steering the considered system to rest at time T.

Keywords: Boundary control, exact controllability, finite difference methods, functional optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1487
9342 An Adaptive Mammographic Image Enhancement in Orthogonal Polynomials Domain

Authors: R. Krishnamoorthy, N. Amudhavalli, M.K. Sivakkolunthu

Abstract:

X-ray mammography is the most effective method for the early detection of breast diseases. However, the typical diagnostic signs such as microcalcifications and masses are difficult to detect because mammograms are of low-contrast and noisy. In this paper, a new algorithm for image denoising and enhancement in Orthogonal Polynomials Transformation (OPT) is proposed for radiologists to screen mammograms. In this method, a set of OPT edge coefficients are scaled to a new set by a scale factor called OPT scale factor. The new set of coefficients is then inverse transformed resulting in contrast improved image. Applications of the proposed method to mammograms with subtle lesions are shown. To validate the effectiveness of the proposed method, we compare the results to those obtained by the Histogram Equalization (HE) and the Unsharp Masking (UM) methods. Our preliminary results strongly suggest that the proposed method offers considerably improved enhancement capability over the HE and UM methods.

Keywords: mammograms, image enhancement, orthogonalpolynomials, contrast improvement

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2011
9341 HSV Image Watermarking Scheme Based on Visual Cryptography

Authors: Rawan I. Zaghloul, Enas F. Al-Rawashdeh

Abstract:

In this paper a simple watermarking method for color images is proposed. The proposed method is based on watermark embedding for the histograms of the HSV planes using visual cryptography watermarking. The method has been proved to be robust for various image processing operations such as filtering, compression, additive noise, and various geometrical attacks such as rotation, scaling, cropping, flipping, and shearing.

Keywords: Histogram, HSV image, Visual Cryptography, Watermark.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1970
9340 Easy Shopping by Electronic Credit

Authors: M. Kargar, A. Isazadeh, F. Fartash, T. Saderi

Abstract:

In this paper we suggest a method for setting electronic credits for the customers. In this method banks and market-sites help each other to make doing large shopping through internet so easy. By developing this system, the people who have less money to buy most of the things they want, become able to buy all of them just through a credit. This credit is given by market-sites through a banking control on it. The method suggested can stop being imprisoned because of banking debts.

Keywords: E-Business, E-Credit, Market-site, Buy-site, Bank, E-Commerce.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1486
9339 Regularization of the Trajectories of Dynamical Systems by Adjusting Parameters

Authors: Helle Hein, Ülo Lepik

Abstract:

A gradient learning method to regulate the trajectories of some nonlinear chaotic systems is proposed. The method is motivated by the gradient descent learning algorithms for neural networks. It is based on two systems: dynamic optimization system and system for finding sensitivities. Numerical results of several examples are presented, which convincingly illustrate the efficiency of the method.

Keywords: Chaos, Dynamical Systems, Learning, Neural Networks

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1366
9338 Fast 3D Collision Detection Algorithm using 2D Intersection Area

Authors: Taehyun Yoon, Keechul Jung

Abstract:

There are many researches to detect collision between real object and virtual object in 3D space. In general, these techniques are need to huge computing power. So, many research and study are constructed by using cloud computing, network computing, and distribute computing. As a reason of these, this paper proposed a novel fast 3D collision detection algorithm between real and virtual object using 2D intersection area. Proposed algorithm uses 4 multiple cameras and coarse-and-fine method to improve accuracy and speed performance of collision detection. In the coarse step, this system examines the intersection area between real and virtual object silhouettes from all camera views. The result of this step is the index of virtual sensors which has a possibility of collision in 3D space. To decide collision accurately, at the fine step, this system examines the collision detection in 3D space by using the visual hull algorithm. Performance of the algorithm is verified by comparing with existing algorithm. We believe proposed algorithm help many other research, study and application fields such as HCI, augmented reality, intelligent space, and so on.

Keywords: Collision Detection, Computer Vision, Human Computer Interaction, Visual Hull

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2406
9337 Computation of the Filtering Properties of Photonic Crystal Waveguide Discontinuities Using the Mode Matching Method

Authors: Athanasios Theoharidis, Thomas Kamalakis, Ioannis Neokosmidis, Thomas Sphicopoulos

Abstract:

In this paper, the application of the Mode Matching (MM) method in the case of photonic crystal waveguide discontinuities is presented. The structure under consideration is divided into a number of cells, which supports a number of guided and evanescent modes. These modes can be calculated numerically by an alternative formulation of the plane wave expansion method for each frequency. A matrix equation is then formed relating the modal amplitudes at the beginning and at the end of the structure. The theory is highly efficient and accurate and can be applied to study the transmission sensitivity of photonic crystal devices due to fabrication tolerances. The accuracy of the MM method is compared to the Finite Difference Frequency Domain (FDFD) and the Adjoint Variable Method (AVM) and good agreement is observed.

Keywords: Optical Communications, Integrated Optics, Photonic Crystals, Optical Waveguide Discontinuities.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1578
9336 A Robust Image Steganography Method Using PMM in Bit Plane Domain

Authors: Souvik Bhattacharyya, Aparajita Khan, Indradip Banerjee, Gautam Sanyal

Abstract:

Steganography is the art and science that hides the information in an appropriate cover carrier like image, text, audio and video media. In this work the authors propose a new image based steganographic method for hiding information within the complex bit planes of the image. After slicing into bit planes the cover image is analyzed to extract the most complex planes in decreasing order based on their bit plane complexity. The complexity function next determines the complex noisy blocks of the chosen bit plane and finally pixel mapping method (PMM) has been used to embed secret bits into those regions of the bit plane. The novel approach of using pixel mapping method (PMM) in bit plane domain adaptively embeds data on most complex regions of image, provides high embedding capacity, better imperceptibility and resistance to steganalysis attack.

Keywords: PMM (Pixel Mapping Method), Bit Plane, Steganography, SSIM, KL-Divergence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2867
9335 Index t-SNE: Tracking Dynamics of High-Dimensional Datasets with Coherent Embeddings

Authors: G. Candel, D. Naccache

Abstract:

t-SNE is an embedding method that the data science community has widely used. It helps two main tasks: to display results by coloring items according to the item class or feature value; and for forensic, giving a first overview of the dataset distribution. Two interesting characteristics of t-SNE are the structure preservation property and the answer to the crowding problem, where all neighbors in high dimensional space cannot be represented correctly in low dimensional space. t-SNE preserves the local neighborhood, and similar items are nicely spaced by adjusting to the local density. These two characteristics produce a meaningful representation, where the cluster area is proportional to its size in number, and relationships between clusters are materialized by closeness on the embedding. This algorithm is non-parametric. The transformation from a high to low dimensional space is described but not learned. Two initializations of the algorithm would lead to two different embedding. In a forensic approach, analysts would like to compare two or more datasets using their embedding. A naive approach would be to embed all datasets together. However, this process is costly as the complexity of t-SNE is quadratic, and would be infeasible for too many datasets. Another approach would be to learn a parametric model over an embedding built with a subset of data. While this approach is highly scalable, points could be mapped at the same exact position, making them indistinguishable. This type of model would be unable to adapt to new outliers nor concept drift. This paper presents a methodology to reuse an embedding to create a new one, where cluster positions are preserved. The optimization process minimizes two costs, one relative to the embedding shape and the second relative to the support embedding’ match. The embedding with the support process can be repeated more than once, with the newly obtained embedding. The successive embedding can be used to study the impact of one variable over the dataset distribution or monitor changes over time. This method has the same complexity as t-SNE per embedding, and memory requirements are only doubled. For a dataset of n elements sorted and split into k subsets, the total embedding complexity would be reduced from O(n2) to O(n2/k), and the memory requirement from n2 to 2(n/k)2 which enables computation on recent laptops. The method showed promising results on a real-world dataset, allowing to observe the birth, evolution and death of clusters. The proposed approach facilitates identifying significant trends and changes, which empowers the monitoring high dimensional datasets’ dynamics.

Keywords: Concept drift, data visualization, dimension reduction, embedding, monitoring, reusability, t-SNE, unsupervised learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 489
9334 A Laplace Transform Dual-Reciprocity Boundary Element Method for Axisymmetric Elastodynamic Problems

Authors: B. I. Yun

Abstract:

A dual-reciprocity boundary element method is presented for the numerical solution of a class of axisymmetric elastodynamic problems. The domain integrals that arise in the integrodifferential formulation are converted to line integrals by using the dual-reciprocity method together suitably constructed interpolating functions. The second order time derivatives of the displacement in the governing partial differential equations are suppressed by using Laplace transformation. In the Laplace transform domain, the problem under consideration is eventually reduced to solving a system of linear algebraic equations. Once the linear algebraic equations are solved, the displacement and stress fields in the physical domain can be recovered by using a numerical technique for inverting Laplace transforms.

Keywords: Axisymmetric elasticity, boundary element method, dual-reciprocity method, Laplace transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1671
9333 Enhance Indoor Environment in Buildings and Its Effect on Improving Occupant's Health

Authors: Imad M. Assali

Abstract:

Recently, the world main problem is a global warming and climate change affecting both outdoor and indoor environments, especially the air quality (AQ) as a result of vast migration of people from rural areas to urban areas. Therefore, cities became more crowded and denser from an irregular population increase, along with increasing urbanization caused many problems for the environment such as increasing the land prices, changes in life style, and the new buildings are not adapted to the climate producing uncomfortable and unhealthy indoor building conditions. As interior environments are the places that create the most intimate relationship with the user. Consequently, the indoor environment quality (IEQ) for buildings became uncomfortable and unhealthy for its occupants. The symptoms commonly associated with poor indoor environment such as itchy, headache, fatigue, and respiratory complaints such as cough and congestion, etc. The symptoms tend to improve over time or even disappear when people are away from the building. Therefore, designing a healthy indoor environment to fulfill human needs is the main concern for architects and interior designer. However, this research explores how occupant expectations and environmental attitudes may influence occupant health and satisfaction within the context of the indoor environment. In doing so, it reviews and contributes to the methods and tools used to evaluate only the indoor environment quality (IEQ) components of building performance. Its main aim is to review the literature on indoor human comfort. This is followed by a review of previous papers published related to human comfort. Finally, this paper will provide possible approaches in design level of healthy buildings.

Keywords: Sustainable building, indoor environment quality (IEQ), occupant's health, active system, sick building syndrome (SBS).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1904
9332 A Comparative Analysis of Fuzzy, Neuro-Fuzzy and Fuzzy-GA Based Approaches for Software Reusability Evaluation

Authors: Parvinder Singh Sandhu, Dalwinder Singh Salaria, Hardeep Singh

Abstract:

Software Reusability is primary attribute of software quality. There are metrics for identifying the quality of reusable components but the function that makes use of these metrics to find reusability of software components is still not clear. These metrics if identified in the design phase or even in the coding phase can help us to reduce the rework by improving quality of reuse of the component and hence improve the productivity due to probabilistic increase in the reuse level. In this paper, we have devised the framework of metrics that uses McCabe-s Cyclometric Complexity Measure for Complexity measurement, Regularity Metric, Halstead Software Science Indicator for Volume indication, Reuse Frequency metric and Coupling Metric values of the software component as input attributes and calculated reusability of the software component. Here, comparative analysis of the fuzzy, Neuro-fuzzy and Fuzzy-GA approaches is performed to evaluate the reusability of software components and Fuzzy-GA results outperform the other used approaches. The developed reusability model has produced high precision results as expected by the human experts.

Keywords: Software Reusability, Software Metrics, Neural Networks, Genetic Algorithm, Fuzzy Logic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1816
9331 Optimal and Critical Path Analysis of State Transportation Network Using Neo4J

Authors: Pallavi Bhogaram, Xiaolong Wu, Min He, Onyedikachi Okenwa

Abstract:

A transportation network is a realization of a spatial network, describing a structure which permits either vehicular movement or flow of some commodity. Examples include road networks, railways, air routes, pipelines, and many more. The transportation network plays a vital role in maintaining the vigor of the nation’s economy. Hence, ensuring the network stays resilient all the time, especially in the face of challenges such as heavy traffic loads and large scale natural disasters, is of utmost importance. In this paper, we used the Neo4j application to develop the graph. Neo4j is the world's leading open-source, NoSQL, a native graph database that implements an ACID-compliant transactional backend to applications. The Southern California network model is developed using the Neo4j application and obtained the most critical and optimal nodes and paths in the network using centrality algorithms. The edge betweenness centrality algorithm calculates the critical or optimal paths using Yen's k-shortest paths algorithm, and the node betweenness centrality algorithm calculates the amount of influence a node has over the network. The preliminary study results confirm that the Neo4j application can be a suitable tool to study the important nodes and the critical paths for the major congested metropolitan area.

Keywords: Transportation network, critical path, connectivity reliability, network model, Neo4J application, optimal path, critical path, edge betweenness centrality index, node betweenness centrality index, Yen’s k-shortest paths.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 853
9330 Discrete and Stationary Adaptive Sub-Band Threshold Method for Improving Image Resolution

Authors: P. Joyce Beryl Princess, Y. Harold Robinson

Abstract:

Image Processing is a structure of Signal Processing for which the input is the image and the output is also an image or parameter of the image. Image Resolution has been frequently referred as an important aspect of an image. In Image Resolution Enhancement, images are being processed in order to obtain more enhanced resolution. To generate highly resoluted image for a low resoluted input image with high PSNR value. Stationary Wavelet Transform is used for Edge Detection and minimize the loss occurs during Downsampling. Inverse Discrete Wavelet Transform is to get highly resoluted image. Highly resoluted output is generated from the Low resolution input with high quality. Noisy input will generate output with low PSNR value. So Noisy resolution enhancement technique has been used for adaptive sub-band thresholding is used. Downsampling in each of the DWT subbands causes information loss in the respective subbands. SWT is employed to minimize this loss. Inverse Discrete wavelet transform (IDWT) is to convert the object which is downsampled using DWT into a highly resoluted object. Used Image denoising and resolution enhancement techniques will generate image with high PSNR value. Our Proposed method will improve Image Resolution and reached the optimized threshold.

Keywords: Image Processing, Inverse Discrete wavelet transform, PSNR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1790
9329 High Performance Liquid Chromatographic Method for Determination of Colistin Sulfate and its Application in Medicated Premixand Animal Feed

Authors: S.Choosakoonkriang, S. Supaluknari, P. Puangkaew

Abstract:

The aim of the present study was to develop and validate an inexpensive and simple high performance liquid chromatographic (HPLC) method for the determination of colistin sulfate. Separation of colistin sulfate was achieved on a ZORBAX Eclipse XDB-C18 column using UV detection at λ=215 nm. The mobile phase was 30 mM sulfate buffer (pH 2.5):acetonitrile(76:24). An excellent linearity (r2=0.998) was found in the concentration range of 25 - 400 μg/mL. Intra- day and inter-day precisions of method (%RSD, n=3) were less than 7.9%.The developed and validated method was applied to determination of the content of colistin sulfate in medicated premix and animal feed sample.The recovery of colistin from animal feed was satisfactorily ranged from 90.92 to 93.77%. The results demonstrated that the HPLC method developed in this work is appropriate for direct determination of colistin sulfate in commercial medicated premixes and animal feed.

Keywords: Colistin sulfate, HPLC, medicated premix, animal feed

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8165
9328 Efficient Iris Recognition Method for Human Identification

Authors: A. Basit, M. Y. Javed, M. A. Anjum

Abstract:

In this paper, an efficient method for personal identification based on the pattern of human iris is proposed. It is composed of image acquisition, image preprocessing to make a flat iris then it is converted into eigeniris and decision is carried out using only reduction of iris in one dimension. By comparing the eigenirises it is determined whether two irises are similar. The results show that proposed method is quite effective.

Keywords: Biometrics, Canny Operator, Eigeniris, Iris Recognition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1543
9327 Bubble Point Pressures of CO2+Ethyl Palmitate by a Cubic Equation of State and the Wong-Sandler Mixing Rule

Authors: M. A. Sedghamiz, S. Raeissi

Abstract:

This study presents three different approaches to estimate bubble point pressures for the binary system of CO2 and ethyl palmitate fatty acid ethyl ester. The first method involves the Peng-Robinson (PR) Equation of State (EoS) with the conventional mixing rule of Van der Waals. The second approach involves the PR EOS together with the Wong Sandler (WS) mixing rule, coupled with the UNIQUAC GE model. In order to model the bubble point pressures with this approach, the volume and area parameter for ethyl palmitate were estimated by the Hansen group contribution method. The last method involved the Peng-Robinson, combined with the Wong-Sandler method, but using NRTL as the GE model. Results using the Van der Waals mixing rule clearly indicated that this method has the largest errors among all three methods, with errors in the range of 3.96-6.22%. The PR-WS-UNIQUAC method exhibited small errors, with average absolute deviations between 0.95 to 1.97 percent. The PR-WS-NRTL method led to the least errors, where average absolute deviations ranged between 0.65-1.7%.

Keywords: Bubble pressure, Gibbs excess energy model, mixing rule, CO2 solubility, ethyl palmitate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1853
9326 The Construction of Interactive Computer Multimedia Instruction on “Basic Japanese Vocabulary“

Authors: Kongrit Jittangthammagul, Sakesun Yampinij, Thapanee Endoo, Nattapong Kramwong

Abstract:

The study entitled “The Construction of Interactive Computer Multimedia Instruction on Basic Japanese Vocabulary" was aimed: 1) To construct the interactive computer multimedia instruction on Basic Japanese Vocabulary, 2) To find out multimedia-s quality, 3) To examine the student-s satisfaction and 4) To study the learning achievement in Basic Japanese vocabulary. The sampling group used in this study was composed of 40 1st year student in Educational Communications and Technology Department, Faculty of Industrial Education and Technology, King Mongkut-s University of Technology Thonburi, in the academic year 2553 B.E. (2010). According to research results, we found that 1). The quality assessment by 3 mass media experts was at 4.72 on average or at high level. 2) In terms of contents, the evaluation by 3 experts was at 4.81 on average or at high level. 3) In terms of achievement, there was a statistical significance between before and after the treatment at the .05 level. 4) The satisfaction of students towards the interactive computer multimedia Instruction on “Basic Japanese Vocabulary" was 4.35 on average, or at high level.

Keywords: Interactive Computer Multimedia on Basic Japanese Vocabulary, Learning Achievement, Quality

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1496
9325 Vibration Transmission across Junctions of Walls and Floors in an Apartment Building: An Experimental Investigation

Authors: Hugo Sampaio Libero, Max de Castro Magalhaes

Abstract:

The perception of sound radiated from a building floor is greatly influenced by the rooms in which it is immersed and by the position of both listener and source. The main question that remains unanswered is related to the influence of the source position on the sound power radiated by a complex wall-floor system in buildings. This research is concerned with the investigation of vibration transmission across walls and floors in buildings. It is primarily based on the determination of vibration reduction index via experimental tests. Knowledge of this parameter may help in predicting noise and vibration propagation in building components. First, the physical mechanisms involving vibration transmission across structural junctions is described. An experimental set-up is performed to aid this investigation. The experimental tests have showed that the vibration generation in the walls and floors are directed related to their size and boundary conditions. It is also shown that the vibration source position can affect the overall vibration spectrum significantly. Second, the characteristics of the noise spectra inside the rooms due to an impact source (tapping machine) are also presented. Conclusions are drawn for the general trend of vibration and noise spectrum of the structural components and rooms respectively. In summary, the aim of this paper is to investigate the vibro-acoustical behavior of building floors and walls under floor impact excitation. The impact excitation was at distinct positions on the slab. The analysis has highlighted the main physical characteristics of the vibration transmission mechanism.

Keywords: Vibration transmission, Vibration Reduction Index, Impact excitation, building acoustics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 334
9324 Study of Effect of Gear Tooth Accuracy on Transmission Mount Vibration

Authors: Kalyan Deepak Kolla, Ketan Paua, Rajkumar Bhagate

Abstract:

Transmission dynamics occupy major role in customer perception of the product in both senses of touch and quality of sound. The quantity and quality of sound perceived is more concerned with the whine noise of the gears engaged. Whine noise is tonal in nature and tonal noises cause fatigue and irritation to customers, which in turn affect the quality of the product. Transmission error is the usual suspect for whine noise, which can be caused due to misalignments, tolerances, manufacturing variabilities. In-cabin noise is also more sensitive to the gear design. As the details of the gear tooth design and manufacturing are in microns, anything out of the tolerance zone, either in design or manufacturing, will cause a whine noise. This will also cause high variation in stress and deformation due to change in the load and leads to the fatigue failure of the gears. Hence gear design and development take priority in the transmission development process. This paper aims to study such variability by considering five pairs of helical spur gears and their effect on the transmission error, contact pattern and vibration level on the transmission.

Keywords: Gears, whine noise, manufacturing variability, mount vibration variability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 645
9323 Self-protection Method for Flying Robots to Avoid Collision

Authors: Guosheng Wu, Luning Wang, Changyuan Fan, Xi Zhu

Abstract:

This paper provides a new approach to solve the motion planning problems of flying robots in uncertain 3D dynamic environments. The robots controlled by this method can adaptively choose the fast way to avoid collision without information about the shapes and trajectories of obstacles. Based on sphere coordinates the new method accomplishes collision avoidance of flying robots without any other auxiliary positioning systems. The Self-protection System gives robots self-protection abilities to work in uncertain 3D dynamic environments. Simulations illustrate the validity of the proposed method.

Keywords: Collision avoidance, Mobile robots, Motion-planning, Sphere coordinates, Self-protection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1581
9322 Recognition-based Segmentation in Persian Character Recognition

Authors: Mohsen Zand, Ahmadreza Naghsh Nilchi, S. Amirhassan Monadjemi

Abstract:

Optical character recognition of cursive scripts presents a number of challenging problems in both segmentation and recognition processes in different languages, including Persian. In order to overcome these problems, we use a newly developed Persian word segmentation method and a recognition-based segmentation technique to overcome its segmentation problems. This method is robust as well as flexible. It also increases the system-s tolerances to font variations. The implementation results of this method on a comprehensive database show a high degree of accuracy which meets the requirements for commercial use. Extended with a suitable pre and post-processing, the method offers a simple and fast framework to develop a full OCR system.

Keywords: OCR, Persian, Recognition, Segmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1840
9321 Delineating Concern Ground in Block Caving – Underground Mine Using Ground Penetrating Radar

Authors: Eric Sitorus, Septian Prahastudhi, Turgod Nainggolan, Erwin Riyanto

Abstract:

Mining by block or panel caving is a mining method that takes advantage of fractures within an ore body, coupled with gravity, to extract material from a predetermined column of ore. The caving column is weakened from beneath through the use of undercutting, after which the ore breaks up and is extracted from below in a continuous cycle. The nature of this method induces cyclical stresses on the pillars of excavations as stress is built up and released over time, which has a detrimental effect on both the installed ground support and the rock mass itself. Ground support capacity, especially on the production where excavation void ratio is highest, is subjected to heavy loading. Strain above threshold of the elongation of support capacity can yield resulting in damage to excavations. Geotechnical engineers must evaluate not only the remnant capacity of ground support systems but also investigate depth of rock mass yield within pillars, backs and floors. Ground Penetrating Radar (GPR) is a geophysical method that has the ability to evaluate rock mass damage using electromagnetic waves. This paper illustrates a case study from the Grasberg mining complex where non-invasive information on the depth of damage and condition of the remaining rock mass was required. GPR with 100 MHz antenna resolution was used to obtain images of the subsurface to determine rehabilitation requirements prior to recommencing production activities. The GPR surveys were used to calibrate the reflection coefficient response of varying rock mass conditions to known Rock Quality Designation (RQD) parameters observed at the mine. The calibrated GPR survey allowed site engineers to map subsurface conditions and plan rehabilitation accordingly.

Keywords: Block caving, ground penetrating radar, reflectivity, RQD.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 669
9320 Contour Estimation in Synthetic and Real Weld Defect Images based on Maximum Likelihood

Authors: M. Tridi, N. Nacereddine, N. Oucief

Abstract:

This paper describes a novel method for automatic estimation of the contours of weld defect in radiography images. Generally, the contour detection is the first operation which we apply in the visual recognition system. Our approach can be described as a region based maximum likelihood formulation of parametric deformable contours. This formulation provides robustness against the poor image quality, and allows simultaneous estimation of the contour parameters together with other parameters of the model. Implementation is performed by a deterministic iterative algorithm with minimal user intervention. Results testify for the very good performance of the approach especially in synthetic weld defect images.

Keywords: Contour, gaussian, likelihood, rayleigh.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1661
9319 A Sub Pixel Resolution Method

Authors: S. Khademi, A. Darudi, Z. Abbasi

Abstract:

One of the main limitations for the resolution of optical instruments is the size of the sensor-s pixels. In this paper we introduce a new sub pixel resolution algorithm to enhance the resolution of images. This method is based on the analysis of multiimages which are fast recorded during the fine relative motion of image and pixel arrays of CCDs. It is shown that by applying this method for a sample noise free image one will enhance the resolution with 10-14 order of error.

Keywords: Sub Pixel Resolution, Moving Pixels, CCD, Image, Optical Instrument.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1995
9318 The Analysis of Printing Quality of Offset - Printing Ink with Coconut Oil Base

Authors: Wat Ploysri

Abstract:

The objectives of this research are to produce prototype coconut oil based solvent offset printing inks and to analyze a basic quality of printing work derived from coconut oil based solvent offset printing inks, by mean of bringing coconut oil for producing varnish and bringing such varnish to produce black offset printing inks. Then, analysis of qualities i.e. CIELAB value, density value, and dot gain value of printing work from coconut oil based solvent offset printing inks which printed on gloss-coated woodfree paper weighs 130 grams were done. The research result of coconut oil based solvent offset printing inks indicated that the suitable varnish formulation is using 51% of coconut oil, 36% of phenolic resin, and 14% of solvent oil 14%, while the result of producing black offset ink displayed that the suitable formula of printing ink is using varnish mixed with 20% of coconut oil, and the analyzing printing work of coconut oil based solvent offset printing inks which printed on paper, the results were as follows: CIELAB value of black offset printing ink is at L* = 31.90, a* = 0.27, and b* = 1.86, density value is at 1.27 and dot gain value was high at mid tone area of image area.

Keywords: Offset Printing, Coconut Oil, Printing Ink, Printing Quality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3828
9317 Bone Mineral Density and Frequency of Low-Trauma Fractures in Ukrainian Women with Metabolic Syndrome

Authors: Vladyslav Povoroznyuk, Larysa Martynyuk, Iryna Syzonenko, Liliya Martynyuk

Abstract:

Osteoporosis is one of the important problems in postmenopausal women due to an increased risk of sudden and unexpected fractures. This study is aimed to determine the connection between bone mineral density (BMD) and trabecular bone score (TBS) in Ukrainian women suffering from metabolic syndrome. Participating in the study, 566 menopausal women aged 50-79 year-old were examined and divided into two groups: Group A included 336 women with no obesity (BMI ≤ 29.9 kg/m2), and Group B – 230 women with metabolic syndrome (diagnosis according to IDF criteria, 2005). Dual-energy X-ray absorptiometry was used for measuring of lumbar spine (L1-L4), femoral neck, total body and forearm BMD and bone quality indexes (last according to Med-Imaps installation). Data were analyzed using Statistical Package 6.0. A significant increase of lumbar spine (L1-L4), femoral neck, total body and ultradistal radius BMD was found in women with metabolic syndrome compared to those without obesity (p < 0.001) both in their totality and in groups of 50-59 years, 60-69 years, and 70-79 years. TBS was significantly higher in non-obese women compared to metabolic syndrome patients of 50-59 years and in the general sample (p < 0.05). Analysis showed significant positive correlation between body mass index (BMI) and BMD at all levels. Significant negative correlation between BMI and TBS (L1-L4) was established. Despite the fact that BMD indexes were significantly higher in women with metabolic syndrome, the frequency of vertebral and non-vertebral fractures did not differ significantly in the groups of patients.

Keywords: Bone mineral density, trabecular bone score, metabolic syndrome, fracture.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1000
9316 Perceived Determinants of Obesity among Primary School Pupils in Eti Osa Local Government Area of Lagos State, Nigeria

Authors: B. O. Diyaolu, E. A. Okebanjo

Abstract:

Children in today’s world need attention and care even with their physique as obesity is also at the increased. Several factors can be responsible for obesity in children and adequate attention is paramount in other not to accommodate it into adolescent period. This study investigated perceived determinants of obesity among primary school pupils in Eti Osa Local Government area of Lagos State. Descriptive survey research design was used and population was all obese pupils in Eti Osa Local Government Area of Lagos State. 92 pupils were selected from randomly picked 12 primary schools while purposive sampling technique was used to pick primary 4-6 pupils. With the aid of body mass index (BMI) and age percentile chart the obese pupils were selected. The instrument for the study was a self-developed and structured questionnaire on perceived determinant of obesity. The questionnaire was divided into three sections. The Cronbach’s Alpha reliability coefficient of 0.74 was obtained. The hypotheses were tested at 0.05 significant levels. The completed questionnaire was collated coded and analyzed using descriptive statistics of frequency counts and percentage and inferential statistics of chi-square (X2). Findings of this study revealed that physical activities and parental influences were determinant of obesity. Physical activity is essential in reducing the rate of obesity in Eti Osa Local Government Area both at home and within the school environment. Primary schools need to create more playing ground for pupils to exercise themselves. Parents need to cater for their children diet ensuring not just the quantity but the quality as well.

Keywords: Feeding pattern, obese pupils, parental influence, physical activities.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 610
9315 Exp-Function Method for Finding Some Exact Solutions of Rosenau Kawahara and Rosenau Korteweg-de Vries Equations

Authors: Ehsan Mahdavi

Abstract:

In this paper, we apply the Exp-function method to Rosenau-Kawahara and Rosenau-KdV equations. Rosenau-Kawahara equation is the combination of the Rosenau and standard Kawahara equations and Rosenau-KdV equation is the combination of the Rosenau and standard KdV equations. These equations are nonlinear partial differential equations (NPDE) which play an important role in mathematical physics. Exp-function method is easy, succinct and powerful to implement to nonlinear partial differential equations arising in mathematical physics. We mainly try to present an application of Exp-function method and offer solutions for common errors wich occur during some of the recent works.

Keywords: Exp-function method, Rosenau Kawahara equation, Rosenau Korteweg-de Vries equation, nonlinear partial differential equation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2058