Search results for: time domain reflectometry measurement technique
9841 Turbulence Modeling of Source and Sink Flows
Authors: Israt Jahan Eshita
Abstract:
Flows developed between two parallel disks have many engineering applications. Two types of non-swirling flows can be generated in such a domain. One is purely source flow in disc type domain (outward flow). Other is purely sink flow in disc type domain (inward flow). This situation often appears in some turbo machinery components such as air bearings, heat exchanger, radial diffuser, vortex gyroscope, disc valves, and viscosity meters. The main goal of this paper is to show the mesh convergence, because mesh convergence saves time, and economical to run and increase the efficiency of modeling for both sink and source flow. Then flow field is resolved using a very fine mesh near-wall, using enhanced wall treatment. After that we are going to compare this flow using standard k-epsilon, RNG k-epsilon turbulence models. Lastly compare some experimental data with numerical solution for sink flow. The good agreement of numerical solution with the experimental works validates the current modeling.
Keywords: Hydraulic diameter, k-epsilon model, meshes convergence, Reynolds number, RNG model, sink flow, source flow and wall y+.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25359840 Surface Charge Based Rapid Method for Detection of Microbial Contamination in Drinking Water and Food Products
Authors: Kandpal M. , Gundampati R. K , Debnath M.
Abstract:
Microbial contamination, most of which are fecal born in drinking water and food industry is a serious threat to humans. Escherichia coli is one of the most common and prevalent among them. We have developed a sensor for rapid and an early detection of contaminants, taking E.coli as a threat indicator organism. The sensor is based on co-polymerizations of aniline and formaldehyde in form of thin film over glass surface using the vacuum deposition technique. The particular doping combination of thin film with Fe-Al and Fe-Cu in different concentrations changes its non conducting properties to p- type semi conductor. This property is exploited to detect the different contaminants, believed to have the different surface charge. It was found through experiments that different microbes at same OD (0.600 at 600 nm) have different conductivity in solution. Also the doping concentration is found to be specific for attracting microbes on the basis of surface charge. This is a simple, cost effective and quick detection method which not only decreases the measurement time but also gives early warnings for highly contaminated samples.
Keywords: Sensor, Vacuum deposition technique, thin film, E.coli detection, doping concentration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15929839 Multidimensional Performance Management
Authors: David Wiese
Abstract:
In order to maximize efficiency of an information management platform and to assist in decision making, the collection, storage and analysis of performance-relevant data has become of fundamental importance. This paper addresses the merits and drawbacks provided by the OLAP paradigm for efficiently navigating large volumes of performance measurement data hierarchically. The system managers or database administrators navigate through adequately (re)structured measurement data aiming to detect performance bottlenecks, identify causes for performance problems or assessing the impact of configuration changes on the system and its representative metrics. Of particular importance is finding the root cause of an imminent problem, threatening availability and performance of an information system. Leveraging OLAP techniques, in contrast to traditional static reporting, this is supposed to be accomplished within moderate amount of time and little processing complexity. It is shown how OLAP techniques can help improve understandability and manageability of measurement data and, hence, improve the whole Performance Analysis process.
Keywords: Data Warehousing, OLAP, Multidimensional Navigation, Performance Diagnosis, Performance Management, Performance Tuning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21369838 Simulation and Measurement the Radiation of an Antenna inside a Metallic Case using FDTD
Authors: Shabnam Ladan, M. S. Abrishamian
Abstract:
In this paper we have developed a FDTD simulation code which can treat wave propagation of a monopole antenna in a metallic case which covers with PML, and performed a series of three dimensional FDTD simulations of electromagnetic wave propagation in this space .We also provide a measurement set up in antenna lab and fortunately the simulations and measurements show good agreement. According to simulation and measurement results, we confirmed that the computer program which had been written in FORTRAN, works correctly.Keywords: FDTD, EMC, monopole antenna.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15359837 Performance Complexity Measurement of Tightening Equipment Based on Kolmogorov Entropy
Authors: Guoliang Fan, Aiping Li, Xuemei Liu, Liyun Xu
Abstract:
The performance of the tightening equipment will decline with the working process in manufacturing system. The main manifestations are the randomness and discretization degree increasing of the tightening performance. To evaluate the degradation tendency of the tightening performance accurately, a complexity measurement approach based on Kolmogorov entropy is presented. At first, the states of performance index are divided for calibrating the discrete degree. Then the complexity measurement model based on Kolmogorov entropy is built. The model describes the performance degradation tendency of tightening equipment quantitatively. At last, a study case is applied for verifying the efficiency and validity of the approach. The research achievement shows that the presented complexity measurement can effectively evaluate the degradation tendency of the tightening equipment. It can provide theoretical basis for preventive maintenance and life prediction of equipment.
Keywords: Complexity measurement, Kolmogorov entropy, manufacturing system, performance evaluation, tightening equipment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9649836 Measuring Process Component Design on Achieving Managerial Goals
Authors: Eakong Atiptamvaree, Twittie Senivongse
Abstract:
Process-oriented software development is a new software development paradigm in which software design is modeled by a business process which is in turn translated into a process execution language for execution. The building blocks of this paradigm are software units that are composed together to work according to the flow of the business process. This new paradigm still exhibits the characteristic of the applications built with the traditional software component technology. This paper discusses an approach to apply a traditional technique for software component fabrication to the design of process-oriented software units, called process components. These process components result from decomposing a business process of a particular application domain into subprocesses, and these process components can be reused to design the business processes of other application domains. The decomposition considers five managerial goals, namely cost effectiveness, ease of assembly, customization, reusability, and maintainability. The paper presents how to design or decompose process components from a business process model and measure some technical features of the design that would affect the managerial goals. A comparison between the measurement values from different designs can tell which process component design is more appropriate for the managerial goals that have been set. The proposed approach can be applied in Web Services environment which accommodates process-oriented software development.Keywords: Business Process Model, Managerial Goals, ProcessComponent.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15139835 Fourier Galerkin Approach to Wave Equation with Absorbing Boundary Conditions
Authors: Alexandra Leukauf, Alexander Schirrer, Emir Talic
Abstract:
Numerical computation of wave propagation in a large domain usually requires significant computational effort. Hence, the considered domain must be truncated to a smaller domain of interest. In addition, special boundary conditions, which absorb the outward travelling waves, need to be implemented in order to describe the system domains correctly. In this work, the linear one dimensional wave equation is approximated by utilizing the Fourier Galerkin approach. Furthermore, the artificial boundaries are realized with absorbing boundary conditions. Within this work, a systematic work flow for setting up the wave problem, including the absorbing boundary conditions, is proposed. As a result, a convenient modal system description with an effective absorbing boundary formulation is established. Moreover, the truncated model shows high accuracy compared to the global domain.Keywords: Absorbing boundary conditions, boundary control, Fourier Galerkin approach, modal approach, wave equation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8889834 Accurate Fault Classification and Section Identification Scheme in TCSC Compensated Transmission Line using SVM
Authors: Pushkar Tripathi, Abhishek Sharma, G. N. Pillai, Indira Gupta
Abstract:
This paper presents a new approach for the protection of Thyristor-Controlled Series Compensator (TCSC) line using Support Vector Machine (SVM). One SVM is trained for fault classification and another for section identification. This method use three phase current measurement that results in better speed and accuracy than other SVM based methods which used single phase current measurement. This makes it suitable for real-time protection. The method was tested on 10,000 data instances with a very wide variation in system conditions such as compensation level, source impedance, location of fault, fault inception angle, load angle at source bus and fault resistance. The proposed method requires only local current measurement.Keywords: Fault Classification, Section Identification, Feature Selection, Support Vector Machine (SVM), Thyristor-Controlled Series Compensator (TCSC)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25279833 Conducting Flow Measurement Laboratory Test Work
Authors: M. B. Kime
Abstract:
Mass flow measurement is the basis of most technoeconomic formulations in the chemical industry. This calls for reliable and accurate detection of mass flow. Flow measurement laboratory experiments were conducted using various instruments. These consisted of orifice plates, various sized rotameters, wet gas meter and soap bubble meter. This work was aimed at evaluating appropriate operating conditions and accuracy of the aforementioned devices. The experimental data collected were compared to theoretical predictions from Bernoulli’s equation and calibration curves supplied by the instrument’s manufacturers. The results obtained showed that rotameters were more reliable for measuring high and low flow rates; while soap-bubble meters and wet-gas meters were found to be suitable for measuring low flow rates. The laboratory procedures and findings of the actual work can assist engineering students and professionals in conducting their flow measurement laboratory test work.
Keywords: Flow measurement, orifice plates, rotameters, wet gas meter, soap bubble meter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 49449832 Particle Image Velocimetry for Measuring Water Flow Velocity
Authors: King Kuok Kuok, Po Chan Chiu
Abstract:
Floods are natural phenomena, which may turn into disasters causing widespread damage, health problems and even deaths. Nowadays, floods had become more serious and more frequent due to climatic changes. During flooding, discharge measurement still can be taken by standing on the bridge across the river using portable measurement instrument. However, it is too dangerous to get near to the river especially during high flood. Therefore, this study employs Particle Image Velocimetry (PIV) as a tool to measure the surface flow velocity. PIV is a image processing technique to track the movement of water from one point to another. The PIV codes are developed using Matlab. In this study, 18 ping pong balls were scattered over the surface of the drain and images were taken with a digital SLR camera. The images obtained were analyzed using the PIV code. Results show that PIV is able to produce the flow velocity through analyzing the series of images captured.
Keywords: Particle Image Velocimetry, flow velocity, surface flow.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28459831 Improving the Effectiveness of Software Testing through Test Case Reduction
Authors: R. P. Mahapatra, Jitendra Singh
Abstract:
This paper proposes a new technique for improving the efficiency of software testing, which is based on a conventional attempt to reduce test cases that have to be tested for any given software. The approach utilizes the advantage of Regression Testing where fewer test cases would lessen time consumption of the testing as a whole. The technique also offers a means to perform test case generation automatically. Compared to one of the techniques in the literature where the tester has no option but to perform the test case generation manually, the proposed technique provides a better option. As for the test cases reduction, the technique uses simple algebraic conditions to assign fixed values to variables (Maximum, minimum and constant variables). By doing this, the variables values would be limited within a definite range, resulting in fewer numbers of possible test cases to process. The technique can also be used in program loops and arrays.Keywords: Software Testing, Test Case Generation, Test CaseReduction
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30199830 A Real-Time Signal Processing Technique for MIDI Generation
Authors: Farshad Arvin, Shyamala Doraisamy
Abstract:
This paper presents a new hardware interface using a microcontroller which processes audio music signals to standard MIDI data. A technique for processing music signals by extracting note parameters from music signals is described. An algorithm to convert the voice samples for real-time processing without complex calculations is proposed. A high frequency microcontroller as the main processor is deployed to execute the outlined algorithm. The MIDI data generated is transmitted using the EIA-232 protocol. The analyses of data generated show the feasibility of using microcontrollers for real-time MIDI generation hardware interface.Keywords: Signal processing, MIDI, Microcontroller, EIA-232.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21279829 Mining News Sites to Create Special Domain News Collections
Authors: David B. Bracewell, Fuji Ren, Shingo Kuroiwa
Abstract:
We present a method to create special domain collections from news sites. The method only requires a single sample article as a seed. No prior corpus statistics are needed and the method is applicable to multiple languages. We examine various similarity measures and the creation of document collections for English and Japanese. The main contributions are as follows. First, the algorithm can build special domain collections from as little as one sample document. Second, unlike other algorithms it does not require a second “general" corpus to compute statistics. Third, in our testing the algorithm outperformed others in creating collections made up of highly relevant articles.Keywords: Information Retrieval, News, Special DomainCollections,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14889828 A Development of the Multiple Intelligences Measurement of Elementary Students
Authors: Chaiwat Waree
Abstract:
This research aims at development of the Multiple Intelligences Measurement of Elementary Students. The structural accuracy test and normality establishment are based on the Multiple Intelligences Theory of Gardner. This theory consists of eight aspects namely linguistics, logic and mathematics, visual-spatial relations, body and movement, music, human relations, self-realization/selfunderstanding and nature. The sample used in this research consists of elementary school students (aged between 5-11 years). The size of the sample group was determined by Yamane Table. The group has 2,504 students. Multistage Sampling was used. Basic statistical analysis and construct validity testing were done using confirmatory factor analysis. The research can be summarized as follows; 1. Multiple Intelligences Measurement consisting of 120 items is content-accurate. Internal consistent reliability according to the method of Kuder-Richardson of the whole Multiple Intelligences Measurement equals .91. The difficulty of the measurement test is between .39-.83. Discrimination is between .21-.85. 2). The Multiple Intelligences Measurement has construct validity in a good range, that is 8 components and all 120 test items have statistical significance level at .01. Chi-square value equals 4357.7; p=.00 at the degree of freedom of 244 and Goodness of Fit Index equals 1.00. Adjusted Goodness of Fit Index equals .92. Comparative Fit Index (CFI) equals .68. Root Mean Squared Residual (RMR) equals 0.064 and Root Mean Square Error of Approximation equals 0.82. 3). The normality of the Multiple Intelligences Measurement is categorized into 3 levels. Those with high intelligence are those with percentiles of more than 78. Those with moderate/medium intelligence are those with percentiles between 24 and 77.9. Those with low intelligence are those with percentiles from 23.9 downwards.
Keywords: Multiple Intelligences, Measurement, Elementary Students.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29589827 Applying Energy Consumption Schedule and Comparing It with Load Shifting Technique in Residential Load
Authors: Amira M. Attia, Karim H. Youssef, Nabil H. Abbasy
Abstract:
Energy consumption schedule (ECS) technique shifts usage of loads from on peak hours and redistributes them throughout the day according to residents’ operating time preferences. This technique is used as form of indirect control from utility to improve the load curve and hence its load factor and reduce customer’s total electric bill as well. Similarly, load shifting technique achieves ECS purposes but as direct control form applied from utility. In this paper, ECS is simulated twice as optimal constrained mathematical formula, solved by using CVX program in MATLAB® R2013b. First, it is utilized for single residential building with ten apartments to determine max allowable energy consumption per hour for each residential apartment. Then, it is used for single apartment with number of shiftable domestic devices, where operating schedule is deduced using previous simulation output results as constraints. The paper ends by giving differences between ECS technique and load shifting technique via literature and simulation. Based on results assessment, it will be shown whether using ECS or load shifting is more beneficial to both customer and utility.Keywords: Energy consumption schedule, load shifting technique, comparison.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11149826 A Watermarking System Using the Wavelet Technique for Satellite Images
Authors: I. R. Farah, I. B. Ismail, M. B. Ahmed
Abstract:
The huge development of new technologies and the apparition of open communication system more and more sophisticated create a new challenge to protect digital content from piracy. Digital watermarking is a recent research axis and a new technique suggested as a solution to these problems. This technique consists in inserting identification information (watermark) into digital data (audio, video, image, databases...) in an invisible and indelible manner and in such a way not to degrade original medium-s quality. Moreover, we must be able to correctly extract the watermark despite the deterioration of the watermarked medium (i.e attacks). In this paper we propose a system for watermarking satellite images. We chose to embed the watermark into frequency domain, precisely the discrete wavelet transform (DWT). We applied our algorithm on satellite images of Tunisian center. The experiments show satisfying results. In addition, our algorithm showed an important resistance facing different attacks, notably the compression (JEPG, JPEG2000), the filtering, the histogram-s manipulation and geometric distortions such as rotation, cropping, scaling.Keywords: Digital data watermarking, Spatial Database, Satellite images, Discrete Wavelets Transform (DWT).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16779825 A Study of Adaptive Fault Detection Method for GNSS Applications
Authors: Je Young Lee, Hee Sung Kim, Kwang Ho Choi, Joonhoo Lim, Sebum Chun, Hyung Keun Lee
Abstract:
This study is purposed to develop an efficient fault detection method for Global Navigation Satellite Systems (GNSS) applications based on adaptive noise covariance estimation. Due to the dependence on radio frequency signals, GNSS measurements are dominated by systematic errors in receiver’s operating environment. In the proposed method, the pseudorange and carrier-phase measurement noise covariances are obtained at time propagations and measurement updates in process of Carrier-Smoothed Code (CSC) filtering, respectively. The test statistics for fault detection are generated by the estimated measurement noise covariances. To evaluate the fault detection capability, intentional faults were added to the filed-collected measurements. The experiment result shows that the proposed method is efficient in detecting unhealthy measurements and improves GNSS positioning accuracy against fault occurrences.
Keywords: Adaptive estimation, fault detection, GNSS, residual.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25559824 A Stochastic Analytic Hierarchy Process Based Weighting Model for Sustainability Measurement in an Organization
Authors: Faramarz Khosravi, Gokhan Izbirak
Abstract:
A weighted statistical stochastic based Analytical Hierarchy Process (AHP) model for modeling the potential barriers and enablers of sustainability for measuring and assessing the sustainability level is proposed. For context-dependent potential barriers and enablers, the proposed model takes the basis of the properties of the variables describing the sustainability functions and was developed into a realistic analytical model for the sustainable behavior of an organization. This thus serves as a means for measuring the sustainability of the organization. The main focus of this paper was the application of the AHP tool in a statistically-based model for measuring sustainability. Hence a strong weighted stochastic AHP based procedure was achieved. A case study scenario of a widely reported major Canadian electric utility was adopted to demonstrate the applicability of the developed model and comparatively examined its results with those of an equal-weighted model method. Variations in the sustainability of a company, as fluctuations, were figured out during the time. In the results obtained, sustainability index for successive years changed form 73.12%, 79.02%, 74.31%, 76.65%, 80.49%, 79.81%, 79.83% to more exact values 73.32%, 77.72%, 76.76%, 79.41%, 81.93%, 79.72%, and 80,45% according to priorities of factors that have found by expert views, respectively. By obtaining relatively necessary informative measurement indicators, the model can practically and effectively evaluate the sustainability extent of any organization and also to determine fluctuations in the organization over time.
Keywords: AHP, sustainability fluctuation, environmental indicators, performance measurement, environmental sustainability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9269823 Automatic Generation Control of Multi-Area Electric Energy Systems Using Modified GA
Authors: Gayadhar Panda, Sidhartha Panda, C. Ardil
Abstract:
A modified Genetic Algorithm (GA) based optimal selection of parameters for Automatic Generation Control (AGC) of multi-area electric energy systems is proposed in this paper. Simulations on multi-area reheat thermal system with and without consideration of nonlinearity like governor dead band followed by 1% step load perturbation is performed to exemplify the optimum parameter search. In this proposed method, a modified Genetic Algorithm is proposed where one point crossover with modification is employed. Positional dependency in respect of crossing site helps to maintain diversity of search point as well as exploitation of already known optimum value. This makes a trade-off between exploration and exploitation of search space to find global optimum in less number of generations. The proposed GA along with decomposition technique as developed has been used to obtain the optimum megawatt frequency control of multi-area electric energy systems. Time-domain simulations are conducted with trapezoidal integration along with decomposition technique. The superiority of the proposed method over existing one is verified from simulations and comparisons.
Keywords: Automatic Generation Control (AGC), Reheat, Proportional Integral (PI) controller, Dead Band, Genetic Algorithm(GA).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26589822 High Accuracy ESPRIT-TLS Technique for Wind Turbine Fault Discrimination
Authors: Saad Chakkor, Mostafa Baghouri, Abderrahmane Hajraoui
Abstract:
ESPRIT-TLS method appears a good choice for high resolution fault detection in induction machines. It has a very high effectiveness in the frequency and amplitude identification. Contrariwise, it presents a high computation complexity which affects its implementation in real time fault diagnosis. To avoid this problem, a Fast-ESPRIT algorithm that combined the IIR band-pass filtering technique, the decimation technique and the original ESPRIT-TLS method was employed to enhance extracting accurately frequencies and their magnitudes from the wind stator current with less computation cost. The proposed algorithm has been applied to verify the wind turbine machine need in the implementation of an online, fast, and proactive condition monitoring. This type of remote and periodic maintenance provides an acceptable machine lifetime, minimize its downtimes and maximize its productivity. The developed technique has evaluated by computer simulations under many fault scenarios. Study results prove the performance of Fast- ESPRIT offering rapid and high resolution harmonics recognizing with minimum computation time and less memory cost.
Keywords: Spectral Estimation, ESPRIT-TLS, Real Time, Diagnosis, Wind Turbine Faults, Band-Pass Filtering, Decimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22609821 Evaluating Hurst Parameters and Fractal Dimensions of Surveyed Dataset of Tailings Dam Embankment
Authors: I. Yakubu, Y. Y. Ziggah, C. Yeboah
Abstract:
In the mining environment, tailings dam embankment is among the hazards and risk areas. The tailings dam embankment could fail and result to damages to facilities, human injuries or even fatalities. Periodic monitoring of the dam embankment is needed to help assess the safety of the tailings dam embankment. Artificial intelligence techniques such as fractals can be used to analyse the stability of the monitored dataset from survey measurement techniques. In this paper, the fractal dimension (D) was determined using D = 2-H. The Hurst parameters (H) of each monitored prism were determined by using a time domain of rescaled range programming in MATLAB software. The fractal dimensions of each monitored prism were determined based on the values of H. The results reveal that the values of the determined H were all within the threshold of 0 ≤ H ≤ 1 m. The smaller the H, the bigger the fractal dimension is. Fractal dimension values ranging from 1.359 x 10-4 m to 1.8843 x 10-3 m were obtained from the monitored prisms on the based on the tailing dam embankment dataset used. The ranges of values obtained indicate that the tailings dam embankment is stable.Keywords: Hurst parameter, fractal dimension, tailings dam embankment, surveyed dataset.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7599820 Intelligent Face-Up CMP System Integrated with On-Line Optical Measurements
Authors: Sheng-Ming Huang, Nan-Chyuan Tsai, Chih-Che Lin, Chun-Chi Lin
Abstract:
An innovative design for intelligent Chemical Mechanical Polishing (CMP) system is proposed and verified by experiments in this report. On-line measurement and real-time feedback are integrated to eliminate the shortcomings of traditional approaches, e.g., the batch-to-batch discrepancy of required polishing time, over consumption of chemical slurry, and non-uniformity across the wafer. The major advantage of the proposed method is that the finish of local surface roughness can be consistent, no matter where the inner-ring region or outer-ring region is concerned. Secondly, it is able to eliminate the Edge effect. Conventionally, the interfacial induced stress near the wafer edge is generally much higher than that near the wafer center. At last, by using the proposed intelligent chemical mechanical polishing strategy, the cost of the entire machining cycle can be much reduced while the quality of the finished goods certainly upgraded.
Keywords: Chemical Mechanical Polishing, Active Magnetic Actuator, On-Line Measurement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17439819 Information Retrieval in Domain Specific Search Engine with Machine Learning Approaches
Authors: Shilpy Sharma
Abstract:
As the web continues to grow exponentially, the idea of crawling the entire web on a regular basis becomes less and less feasible, so the need to include information on specific domain, domain-specific search engines was proposed. As more information becomes available on the World Wide Web, it becomes more difficult to provide effective search tools for information access. Today, people access web information through two main kinds of search interfaces: Browsers (clicking and following hyperlinks) and Query Engines (queries in the form of a set of keywords showing the topic of interest) [2]. Better support is needed for expressing one's information need and returning high quality search results by web search tools. There appears to be a need for systems that do reasoning under uncertainty and are flexible enough to recover from the contradictions, inconsistencies, and irregularities that such reasoning involves. In a multi-view problem, the features of the domain can be partitioned into disjoint subsets (views) that are sufficient to learn the target concept. Semi-supervised, multi-view algorithms, which reduce the amount of labeled data required for learning, rely on the assumptions that the views are compatible and uncorrelated. This paper describes the use of semi-structured machine learning approach with Active learning for the “Domain Specific Search Engines". A domain-specific search engine is “An information access system that allows access to all the information on the web that is relevant to a particular domain. The proposed work shows that with the help of this approach relevant data can be extracted with the minimum queries fired by the user. It requires small number of labeled data and pool of unlabelled data on which the learning algorithm is applied to extract the required data.Keywords: Search engines; machine learning, Informationretrieval, Active logic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20839818 Evaluation of Handover Latency in Intra- Domain Mobility
Authors: Aisha Hassan Abdalla Hashim, Fauzana Ridzuan, Nazreen Rusli
Abstract:
Mobile IPv6 (MIPv6) describes how mobile node can change its point of attachment from one access router to another. As a demand for wireless mobile devices increases, many enhancements for macro-mobility (inter-domain) protocols have been proposed, designed and implemented in Mobile IPv6. Hierarchical Mobile IPv6 (HMIPv6) is one of them that is designed to reduce the amount of signaling required and to improve handover speed for mobile connections. This is achieved by introducing a new network entity called Mobility Anchor Point (MAP). This report presents a comparative study of the Hierarchical Mobility IPv6 and Mobile IPv6 protocols and we have narrowed down the scope to micro-mobility (intra-domain). The architecture and operation of each protocol is studied and they are evaluated based on the Quality of Service (QoS) parameter; handover latency. The simulation was carried out by using the Network Simulator-2. The outcome from this simulation has been discussed. From the results, it shows that, HMIPv6 performs best under intra-domain mobility compared to MIPv6. The MIPv6 suffers large handover latency. As enhancement we proposed to HMIPv6 to locate the MAP to be in the middle of the domain with respect to all Access Routers. That gives approximately same distance between MAP and Mobile Node (MN) regardless of the new location of MN, and possible shorter distance. This will reduce the delay since the distance is shorter. As a future work performance analysis is to be carried for the proposed HMIPv6 and compared to HMIPv6.
Keywords: Intra-domain mobility, HMIPv6, Handover Latency, proposed HMIPv6.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14039817 Globally Convergent Edge-preserving Reconstruction with Contour-line Smoothing
Authors: Marc C. Robini, Pierre-Jean Viverge, Yuemin Zhu, Jianhua Luo
Abstract:
The standard approach to image reconstruction is to stabilize the problem by including an edge-preserving roughness penalty in addition to faithfulness to the data. However, this methodology produces noisy object boundaries and creates a staircase effect. The existing attempts to favor the formation of smooth contour lines take the edge field explicitly into account; they either are computationally expensive or produce disappointing results. In this paper, we propose to incorporate the smoothness of the edge field in an implicit way by means of an additional penalty term defined in the wavelet domain. We also derive an efficient half-quadratic algorithm to solve the resulting optimization problem, including the case when the data fidelity term is non-quadratic and the cost function is nonconvex. Numerical experiments show that our technique preserves edge sharpness while smoothing contour lines; it produces visually pleasing reconstructions which are quantitatively better than those obtained without wavelet-domain constraints.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13459816 Denoising and Compression in Wavelet Domainvia Projection on to Approximation Coefficients
Authors: Mario Mastriani
Abstract:
We describe a new filtering approach in the wavelet domain for image denoising and compression, based on the projections of details subbands coefficients (resultants of the splitting procedure, typical in wavelet domain) onto the approximation subband coefficients (much less noisy). The new algorithm is called Projection Onto Approximation Coefficients (POAC). As a result of this approach, only the approximation subband coefficients and three scalars are stored and/or transmitted to the channel. Besides, with the elimination of the details subbands coefficients, we obtain a bigger compression rate. Experimental results demonstrate that our approach compares favorably to more typical methods of denoising and compression in wavelet domain.
Keywords: Compression, denoising, projections, wavelets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16169815 Motion Detection Techniques Using Optical Flow
Authors: A. A. Shafie, Fadhlan Hafiz, M. H. Ali
Abstract:
Motion detection is very important in image processing. One way of detecting motion is using optical flow. Optical flow cannot be computed locally, since only one independent measurement is available from the image sequence at a point, while the flow velocity has two components. A second constraint is needed. The method used for finding the optical flow in this project is assuming that the apparent velocity of the brightness pattern varies smoothly almost everywhere in the image. This technique is later used in developing software for motion detection which has the capability to carry out four types of motion detection. The motion detection software presented in this project also can highlight motion region, count motion level as well as counting object numbers. Many objects such as vehicles and human from video streams can be recognized by applying optical flow technique.Keywords: Background modeling, Motion detection, Optical flow, Velocity smoothness constant, motion trajectories.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 53869814 Maximum Norm Analysis of a Nonmatching Grids Method for Nonlinear Elliptic Boundary Value Problem −Δu = f(u)
Authors: Abida Harbi
Abstract:
We provide a maximum norm analysis of a finite element Schwarz alternating method for a nonlinear elliptic boundary value problem of the form -Δu = f(u), on two overlapping sub domains with non matching grids. We consider a domain which is the union of two overlapping sub domains where each sub domain has its own independently generated grid. The two meshes being mutually independent on the overlap region, a triangle belonging to one triangulation does not necessarily belong to the other one. Under a Lipschitz assumption on the nonlinearity, we establish, on each sub domain, an optimal L∞ error estimate between the discrete Schwarz sequence and the exact solution of the boundary value problem.Keywords: Error estimates, Finite elements, Nonlinear PDEs, Schwarz method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27569813 The Study of Super Hydrophobic Surfaces Using High Speed Shadowgraphy
Authors: D. Jasikova, M. Kotek, V. Kopecky
Abstract:
The aim of this article is the measurement of the basic characteristic of superhydrophobic surfaces using high speed shadowgraphy. Here we describe the novel patented system for the industrial production of superhydrophobic surfaces. These surfaces were investigated with two optically based measurement methods: impinging drop and inclined wall. The results of the visualization and analysis help to state the suitable sample with superhydrophobic properties for mathematic simulation.
Keywords: Antipearl effect, contact angle, hydrophobic, impinging drop, inclined wall, measurement, plasma, shadowgraphy, superhydrophobic surface.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24589812 Changes in EEG and HRV during Event-Related Attention
Authors: Sun K. Yoo, Chung K. Lee
Abstract:
Determination of attentional status is important because working performance and an unexpected accident is highly related with the attention. The autonomic nervous and the central nervous systems can reflect the changes in person’s attentional status. Reduced number of suitable pysiological parameters among autonomic and central nervous systems related signal parameters will be critical in optimum design of attentional devices. In this paper, we analyze the EEG (Electroencephalography) and HRV (Heart Rate Variability) signals to demonstrate the effective relation with brain signal and cardiovascular signal during event-related attention, which will be later used in selecting the minimum set of attentional parameters. Time and frequency domain parameters from HRV signal and frequency domain parameters from EEG signal are used as input to the optimum feature parameters selector.
Keywords: EEG, HRV, attentional status.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2790