Search results for: underwater noise
525 Optimization of Process Parameters in Wire Electrical Discharge Machining of Inconel X-750 for Dimensional Deviation Using Taguchi Technique
Authors: Mandeep Kumar, Hari Singh
Abstract:
The effective optimization of machining process parameters affects dramatically the cost and production time of machined components as well as the quality of the final products. This paper presents the optimization aspects of a Wire Electrical Discharge Machining operation using Inconel X-750 as work material. The objective considered in this study is minimization of the dimensional deviation. Six input process parameters of WEDM namely spark gap voltage, pulse-on time, pulse-off time, wire feed rate, peak current and wire tension, were chosen as variables to study the process performance. Taguchi's design of experiments methodology has been used for planning and designing the experiments. The analysis of variance was carried out for raw data as well as for signal to noise ratio. Four input parameters and one two-factor interaction have been found to be statistically significant for their effects on the response of interest. The confirmation experiments were also performed for validating the predicted results.Keywords: ANOVA, DOE, inconel, machining, optimization
Procedia PDF Downloads 204524 Process Data-Driven Representation of Abnormalities for Efficient Process Control
Authors: Hyun-Woo Cho
Abstract:
Unexpected operational events or abnormalities of industrial processes have a serious impact on the quality of final product of interest. In terms of statistical process control, fault detection and diagnosis of processes is one of the essential tasks needed to run the process safely. In this work, nonlinear representation of process measurement data is presented and evaluated using a simulation process. The effect of using different representation methods on the diagnosis performance is tested in terms of computational efficiency and data handling. The results have shown that the nonlinear representation technique produced more reliable diagnosis results and outperforms linear methods. The use of data filtering step improved computational speed and diagnosis performance for test data sets. The presented scheme is different from existing ones in that it attempts to extract the fault pattern in the reduced space, not in the original process variable space. Thus this scheme helps to reduce the sensitivity of empirical models to noise.Keywords: fault diagnosis, nonlinear technique, process data, reduced spaces
Procedia PDF Downloads 247523 Cooperative Jamming for Implantable Medical Device Security
Authors: Kim Lytle, Tim Talty, Alan Michaels, Jeff Reed
Abstract:
Implantable medical devices (IMDs) are medically necessary devices embedded in the human body that monitor chronic disorders or automatically deliver therapies. Most IMDs have wireless capabilities that allow them to share data with an offboard programming device to help medical providers monitor the patient’s health while giving the patient more insight into their condition. However, serious security concerns have arisen as researchers demonstrated these devices could be hacked to obtain sensitive information or harm the patient. Cooperative jamming can be used to prevent privileged information leaks by maintaining an adequate signal-to-noise ratio at the intended receiver while minimizing signal power elsewhere. This paper uses ray tracing to demonstrate how a low number of friendly nodes abiding by Bluetooth Low Energy (BLE) transmission regulations can enhance IMD communication security in an office environment, which in turn may inform how companies and individuals can protect their proprietary and personal information.Keywords: implantable biomedical devices, communication system security, array signal processing, ray tracing
Procedia PDF Downloads 114522 Sensor Fault-Tolerant Model Predictive Control for Linear Parameter Varying Systems
Authors: Yushuai Wang, Feng Xu, Junbo Tan, Xueqian Wang, Bin Liang
Abstract:
In this paper, a sensor fault-tolerant control (FTC) scheme using robust model predictive control (RMPC) and set theoretic fault detection and isolation (FDI) is extended to linear parameter varying (LPV) systems. First, a group of set-valued observers are designed for passive fault detection (FD) and the observer gains are obtained through minimizing the size of invariant set of state estimation-error dynamics. Second, an input set for fault isolation (FI) is designed offline through set theory for actively isolating faults after FD. Third, an RMPC controller based on state estimation for LPV systems is designed to control the system in the presence of disturbance and measurement noise and tolerate faults. Besides, an FTC algorithm is proposed to maintain the plant operate in the corresponding mode when the fault occurs. Finally, a numerical example is used to show the effectiveness of the proposed results.Keywords: fault detection, linear parameter varying, model predictive control, set theory
Procedia PDF Downloads 252521 Prediction of the Thermal Parameters of a High-Temperature Metallurgical Reactor Using Inverse Heat Transfer
Authors: Mohamed Hafid, Marcel Lacroix
Abstract:
This study presents an inverse analysis for predicting the thermal conductivities and the heat flux of a high-temperature metallurgical reactor simultaneously. Once these thermal parameters are predicted, the time-varying thickness of the protective phase-change bank that covers the inside surface of the brick walls of a metallurgical reactor can be calculated. The enthalpy method is used to solve the melting/solidification process of the protective bank. The inverse model rests on the Levenberg-Marquardt Method (LMM) combined with the Broyden method (BM). A statistical analysis for the thermal parameter estimation is carried out. The effect of the position of the temperature sensors, total number of measurements and measurement noise on the accuracy of inverse predictions is investigated. Recommendations are made concerning the location of temperature sensors.Keywords: inverse heat transfer, phase change, metallurgical reactor, Levenberg–Marquardt method, Broyden method, bank thickness
Procedia PDF Downloads 334520 Analysis of Financial Time Series by Using Ornstein-Uhlenbeck Type Models
Authors: Md Al Masum Bhuiyan, Maria C. Mariani, Osei K. Tweneboah
Abstract:
In the present work, we develop a technique for estimating the volatility of financial time series by using stochastic differential equation. Taking the daily closing prices from developed and emergent stock markets as the basis, we argue that the incorporation of stochastic volatility into the time-varying parameter estimation significantly improves the forecasting performance via Maximum Likelihood Estimation. While using the technique, we see the long-memory behavior of data sets and one-step-ahead-predicted log-volatility with ±2 standard errors despite the variation of the observed noise from a Normal mixture distribution, because the financial data studied is not fully Gaussian. Also, the Ornstein-Uhlenbeck process followed in this work simulates well the financial time series, which aligns our estimation algorithm with large data sets due to the fact that this algorithm has good convergence properties.Keywords: financial time series, maximum likelihood estimation, Ornstein-Uhlenbeck type models, stochastic volatility model
Procedia PDF Downloads 242519 Ottoman Archaeology in Kostence (Constanta, Romania): A Locality on the Periphery of the Ottoman World
Authors: Margareta Simina Stanc, Aurel Mototolea, Tiberiu Potarniche
Abstract:
The city of Constanta (former Köstence) is located in the Dobrogea region, on the west shore of the Black Sea. Between 1420-1878, Dobrogea was a possession of the Ottoman Empire. Archaeological researches starting with the second half of the 20th century revealed various traces of the Ottoman period in this region. Between 2016-2018, preventive archaeological research conducted in the perimeter of the old Ottoman city of Köstence led to the discovery of structures of habitation as well as of numerous artifacts of the Ottoman period (pottery, coins, buckles, etc.). This study uses the analysis of these new discoveries to complete the picture of daily life in the Ottoman period. In 2017, in the peninsular area of Constanta, preventive archaeological research began at a point in the former Ottoman area. In the range between the current ironing level and the -1.5m depth, the Ottoman period materials appeared constantly. It is worth noting the structure of a large building that has been repaired at least once but could not be fully investigated. In parallel to this wall, there was arranged a transversally arranged brick-lined drainage channel. The drainage channel is poured into a tank (hazna), filled with various vintage materials, but mainly gilded ceramics and iron objects. This type of hazna is commonly found in Constanta for the pre-modern and modern period due to the lack of a sewage system in the peninsular area. A similar structure, probably fountain, was discovered in 2016 in another part of the old city. An interesting piece is that of a cup (probably) Persians and a bowl belonging to Kütahya style, both of the 17th century, proof of commercial routes passing through Constanta during that period and indirectly confirming the documentary testimonies of the time. Also, can be mentioned the discovery, in the year 2016, on the occasion of underwater research carried out by specialists of the department of the Constanta Museum, at a depth of 15 meters, a Turkish oil lamp (17th - the beginning of the 18th century), among other objects of a sunken ship. The archaeological pieces, in a fragmentary or integral state, found in research campaigns 2016-2018, are undergoing processing or restoration, leaving out all the available information, and establishing exact analogies. These discoveries bring new data to the knowledge of daily life during the Ottoman administration in the former Köstence, a locality on the periphery of the Islamic world.Keywords: habitation, material culture, Ottoman administration, Ottoman archaeology, periphery
Procedia PDF Downloads 131518 Sparse Coding Based Classification of Electrocardiography Signals Using Data-Driven Complete Dictionary Learning
Authors: Fuad Noman, Sh-Hussain Salleh, Chee-Ming Ting, Hadri Hussain, Syed Rasul
Abstract:
In this paper, a data-driven dictionary approach is proposed for the automatic detection and classification of cardiovascular abnormalities. Electrocardiography (ECG) signal is represented by the trained complete dictionaries that contain prototypes or atoms to avoid the limitations of pre-defined dictionaries. The data-driven trained dictionaries simply take the ECG signal as input rather than extracting features to study the set of parameters that yield the most descriptive dictionary. The approach inherently learns the complicated morphological changes in ECG waveform, which is then used to improve the classification. The classification performance was evaluated with ECG data under two different preprocessing environments. In the first category, QT-database is baseline drift corrected with notch filter and it filters the 60 Hz power line noise. In the second category, the data are further filtered using fast moving average smoother. The experimental results on QT database confirm that our proposed algorithm shows a classification accuracy of 92%.Keywords: electrocardiogram, dictionary learning, sparse coding, classification
Procedia PDF Downloads 386517 Nexus Between Library and Information Science Education Training and Practice in Nigeria: A Critical Assessment of the Synergy
Authors: Adebayo Emmanuel Layi
Abstract:
Library and Information Science Education is about six (6) decades old in Nigeria. The first Library School was established in 1962 at the University of Ibadan, and since then, several institutions have been running the programme under various certifications, providing the manpower needs of professionals for libraries. As at June 2023, Nigeria has close to a thousand (1000) tertiary institutions and all needing the services of librarians. Apart from the tertiary institutions, several libraries exit in various establishments, both government, private and non-governmental organisations. These has underscored the enormous need for trained librarians for the libraries in these places. The Nexus between LIS Education training and Practice is like a puzzle of egg and chick, which one came first and against this background, this paper examined the roles of the colonial masters in educational development in Africa and vis-à-vis the influence of great library educators such as Melvil Dewey and other educators and the journey through Nigeria institutions. Despite the sound footing of LIS Education, Noise which seems to be a major obstacle on the practice as well as mending the broken link were all examined in the paper. Strategies and the way forward for overall development are suggested.Keywords: nexus, education, training, synergy
Procedia PDF Downloads 93516 Subspace Rotation Algorithm for Implementing Restricted Hopfield Network as an Auto-Associative Memory
Authors: Ci Lin, Tet Yeap, Iluju Kiringa
Abstract:
This paper introduces the subspace rotation algorithm (SRA) to train the Restricted Hopfield Network (RHN) as an auto-associative memory. Subspace rotation algorithm is a gradient-free subspace tracking approach based on the singular value decomposition (SVD). In comparison with Backpropagation Through Time (BPTT) on training RHN, it is observed that SRA could always converge to the optimal solution and BPTT could not achieve the same performance when the model becomes complex, and the number of patterns is large. The AUTS case study showed that the RHN model trained by SRA could achieve a better structure of attraction basin with larger radius(in general) than the Hopfield Network(HNN) model trained by Hebbian learning rule. Through learning 10000 patterns from MNIST dataset with RHN models with different number of hidden nodes, it is observed that an several components could be adjusted to achieve a balance between recovery accuracy and noise resistance.Keywords: hopfield neural network, restricted hopfield network, subspace rotation algorithm, hebbian learning rule
Procedia PDF Downloads 117515 A Soft Switching PWM DC-DC Boost Converter with Increased Efficiency by Using ZVT-ZCT Techniques
Authors: Yakup Sahin, Naim Suleyman Ting, Ismail Aksoy
Abstract:
In this paper, an improved active snubber cell is proposed on account of soft switching (SS) family of pulse width modulation (PWM) DC-DC converters. The improved snubber cell provides zero-voltage transition (ZVT) turn on and zero-current transition (ZCT) turn off for main switch. The snubber cell decreases EMI noise and operates with SS in a wide range of line and load voltages. Besides, all of the semiconductor devices in the converter operate with SS. There is no additional voltage and current stress on the main devices. Additionally, extra voltage stress does not occur on the auxiliary switch and its current stress is acceptable value. The improved converter has a low cost and simple structure. The theoretical analysis of converter is clarified and the operating states are given in detail. The experimental results of converter are obtained by prototype of 500 W and 100 kHz. It is observed that the experimental results and theoretical analysis of converter are suitable with each other perfectly.Keywords: active snubber cells, DC-DC converters, zero-voltage transition, zero-current transition
Procedia PDF Downloads 1020514 Phenomenon of Raveling Distress on the Flexible Pavements: An Overview
Authors: Syed Ali Shahbaz Shah
Abstract:
In the last few years, Bituminous Asphaltic roads are becoming popular day by day in the world. Plenty of research has been carried out to identify many advantages like safety, environmental effects, and comfort. Some other benefits are minimal noise and skid resistance enhancement. Besides the benefits of asphaltic roads, the permeable structure of the road also causes some distress, and raveling is one of the crucial defects. The main reason behind this distress is the failure of adhesion between bitumen mortar, specifically due to excessive load from heavy traffic. The main focus of this study is to identify the root cause and propose both the long-term and the short-term solutions of raveling on a specific road section depicting the overall road situation from the bridge of Kahuta road towards the intersection of the Islamabad express highway. The methodology adopted for this purpose is visual inspections in-situ. It was noted that there were chunks of debris on the road surface, which indicates that the asphalt binder is aged the most probably. Further laboratory testing would confirm that either asphalt binder is aged or inadequate compaction was adept during cold weather paving.Keywords: asphaltic roads, asphalt binder, distress, raveling
Procedia PDF Downloads 116513 A Novel NRIS Index to Evaluate Brain Activity in Prefrontal Regions While Listening to First and Second Languages for Long Time Periods
Authors: Kensho Takahashi, Ko Watanabe, Takashi Kaburagi, Hiroshi Tanaka, Kajiro Watanabe, Yosuke Kurihara
Abstract:
Near-infrared spectroscopy (NIRS) has been widely used as a non-invasive method to measure brain activity, but it is corrupted by baseline drift noise. Here we present a method to measure regional cerebral blood flow as a derivative of NIRS output. We investigate whether, when listening to languages, blood flow can reasonably localize and represent regional brain activity or not. The prefrontal blood flow distribution pattern when advanced second-language listeners listened to a second language (L2) was most similar to that when listening to their first language (L1) among the patterns of mean and standard deviation. In experiments with 25 healthy subjects, the maximum blood flow was localized to the left BA46 of advanced listeners. The blood flow presented is robust to baseline drift and stably localizes regional brain activity.Keywords: NIRS, oxy-hemoglobin, baseline drift, blood flow, working memory, BA46, first language, second language
Procedia PDF Downloads 559512 Analyzing On-Line Process Data for Industrial Production Quality Control
Authors: Hyun-Woo Cho
Abstract:
The monitoring of industrial production quality has to be implemented to alarm early warning for unusual operating conditions. Furthermore, identification of their assignable causes is necessary for a quality control purpose. For such tasks many multivariate statistical techniques have been applied and shown to be quite effective tools. This work presents a process data-based monitoring scheme for production processes. For more reliable results some additional steps of noise filtering and preprocessing are considered. It may lead to enhanced performance by eliminating unwanted variation of the data. The performance evaluation is executed using data sets from test processes. The proposed method is shown to provide reliable quality control results, and thus is more effective in quality monitoring in the example. For practical implementation of the method, an on-line data system must be available to gather historical and on-line data. Recently large amounts of data are collected on-line in most processes and implementation of the current scheme is feasible and does not give additional burdens to users.Keywords: detection, filtering, monitoring, process data
Procedia PDF Downloads 559511 Video Foreground Detection Based on Adaptive Mixture Gaussian Model for Video Surveillance Systems
Authors: M. A. Alavianmehr, A. Tashk, A. Sodagaran
Abstract:
Modeling background and moving objects are significant techniques for video surveillance and other video processing applications. This paper presents a foreground detection algorithm that is robust against illumination changes and noise based on adaptive mixture Gaussian model (GMM), and provides a novel and practical choice for intelligent video surveillance systems using static cameras. In the previous methods, the image of still objects (background image) is not significant. On the contrary, this method is based on forming a meticulous background image and exploiting it for separating moving objects from their background. The background image is specified either manually, by taking an image without vehicles, or is detected in real-time by forming a mathematical or exponential average of successive images. The proposed scheme can offer low image degradation. The simulation results demonstrate high degree of performance for the proposed method.Keywords: image processing, background models, video surveillance, foreground detection, Gaussian mixture model
Procedia PDF Downloads 516510 NFResNet: Multi-Scale and U-Shaped Networks for Deblurring
Authors: Tanish Mittal, Preyansh Agrawal, Esha Pahwa, Aarya Makwana
Abstract:
Multi-Scale and U-shaped Networks are widely used in various image restoration problems, including deblurring. Keeping in mind the wide range of applications, we present a comparison of these architectures and their effects on image deblurring. We also introduce a new block called as NFResblock. It consists of a Fast Fourier Transformation layer and a series of modified Non-Linear Activation Free Blocks. Based on these architectures and additions, we introduce NFResnet and NFResnet+, which are modified multi-scale and U-Net architectures, respectively. We also use three differ-ent loss functions to train these architectures: Charbonnier Loss, Edge Loss, and Frequency Reconstruction Loss. Extensive experiments on the Deep Video Deblurring dataset, along with ablation studies for each component, have been presented in this paper. The proposed architectures achieve a considerable increase in Peak Signal to Noise (PSNR) ratio and Structural Similarity Index (SSIM) value.Keywords: multi-scale, Unet, deblurring, FFT, resblock, NAF-block, nfresnet, charbonnier, edge, frequency reconstruction
Procedia PDF Downloads 136509 Developing an Advanced Algorithm Capable of Classifying News, Articles and Other Textual Documents Using Text Mining Techniques
Authors: R. B. Knudsen, O. T. Rasmussen, R. A. Alphinas
Abstract:
The reason for conducting this research is to develop an algorithm that is capable of classifying news articles from the automobile industry, according to the competitive actions that they entail, with the use of Text Mining (TM) methods. It is needed to test how to properly preprocess the data for this research by preparing pipelines which fits each algorithm the best. The pipelines are tested along with nine different classification algorithms in the realm of regression, support vector machines, and neural networks. Preliminary testing for identifying the optimal pipelines and algorithms resulted in the selection of two algorithms with two different pipelines. The two algorithms are Logistic Regression (LR) and Artificial Neural Network (ANN). These algorithms are optimized further, where several parameters of each algorithm are tested. The best result is achieved with the ANN. The final model yields an accuracy of 0.79, a precision of 0.80, a recall of 0.78, and an F1 score of 0.76. By removing three of the classes that created noise, the final algorithm is capable of reaching an accuracy of 94%.Keywords: Artificial Neural network, Competitive dynamics, Logistic Regression, Text classification, Text mining
Procedia PDF Downloads 121508 A Review on the Potential of Electric Vehicles in Reducing World CO2 Footprints
Authors: S. Alotaibi, S. Omer, Y. Su
Abstract:
The conventional Internal Combustion Engine (ICE) based vehicles are a threat to the environment as they account for a large proportion of the overall greenhouse gas (GHG) emissions in the world. Hence, it is required to replace these vehicles with more environment-friendly vehicles. Electric Vehicles (EVs) are promising technologies which offer both human comfort “noise, pollution” as well as reduced (or no) emissions of GHGs. In this paper, different types of EVs are reviewed and their advantages and disadvantages are identified. It is found that in terms of fuel economy, Plug-in Hybrid EVs (PHEVs) have the best fuel economy, followed by Hybrid EVs (HEVs) and ICE vehicles. Since Battery EVs (BEVs) do not use any fuel, their fuel economy is estimated as price per kilometer. Similarly, in terms of GHG emissions, BEVs are the most environmentally friendly since they do not result in any emissions while HEVs and PHEVs produce less emissions compared to the conventional ICE based vehicles. Fuel Cell EVs (FCEVs) are also zero-emission vehicles, but they have large costs associated with them. Finally, if the electricity is provided by using the renewable energy technologies through grid connection, then BEVs could be considered as zero emission vehicles.Keywords: electric vehicles, zero emission car, fuel economy, CO₂ footprint
Procedia PDF Downloads 147507 Distangling Biological Noise in Cellular Images with a Focus on Explainability
Authors: Manik Sharma, Ganapathy Krishnamurthi
Abstract:
The cost of some drugs and medical treatments has risen in recent years, that many patients are having to go without. A classification project could make researchers more efficient. One of the more surprising reasons behind the cost is how long it takes to bring new treatments to market. Despite improvements in technology and science, research and development continues to lag. In fact, finding new treatment takes, on average, more than 10 years and costs hundreds of millions of dollars. If successful, we could dramatically improve the industry's ability to model cellular images according to their relevant biology. In turn, greatly decreasing the cost of treatments and ensure these treatments get to patients faster. This work aims at solving a part of this problem by creating a cellular image classification model which can decipher the genetic perturbations in cell (occurring naturally or artificially). Another interesting question addressed is what makes the deep-learning model decide in a particular fashion, which can further help in demystifying the mechanism of action of certain perturbations and paves a way towards the explainability of the deep-learning model.Keywords: cellular images, genetic perturbations, deep-learning, explainability
Procedia PDF Downloads 112506 Design an Intelligent Fire Detection System Based on Neural Network and Particle Swarm Optimization
Authors: Majid Arvan, Peyman Beygi, Sina Rokhsati
Abstract:
In-time detection of fire in buildings is of great importance. Employing intelligent methods in data processing in fire detection systems leads to a significant reduction of fire damage at lowest cost. In this paper, the raw data obtained from the fire detection sensor networks in buildings is processed by using intelligent methods based on neural networks and the likelihood of fire happening is predicted. In order to enhance the quality of system, the noise in the sensor data is reduced by analyzing wavelets and applying SVD technique. Meanwhile, the proposed neural network is trained using particle swarm optimization (PSO). In the simulation work, the data is collected from sensor network inside the room and applied to the proposed network. Then the outputs are compared with conventional MLP network. The simulation results represent the superiority of the proposed method over the conventional one.Keywords: intelligent fire detection, neural network, particle swarm optimization, fire sensor network
Procedia PDF Downloads 380505 Influence of Inertial Forces of Large Bearings Utilized in Wind Energy Assemblies
Authors: S. Barabas, F. Sarbu, B. Barabas, A. Fota
Abstract:
Main objective of this paper is to establish a link between inertial forces of the bearings used in construction of wind power plant and its behavior. Using bearings with lower inertial forces has the immediate effect of decreasing inertia rotor system, with significant results in increased energy efficiency, due to decreased friction forces between rollers and raceways. The FEM analysis shows the appearance of uniform contact stress at the ends of the rollers, demonstrated the necessity of production of low mass bearings. Favorable results are expected in the economic field, by reducing material consumption and by increasing the durability of bearings. Using low mass bearings with hollow rollers instead of solid rollers has an impact on working temperature, on vibrations and noise which decrease. Implementation of types of hollow rollers of cylindrical tubular type, instead of expensive rollers with logarithmic profile, will bring significant inertial forces decrease with large benefits in behavior of wind power plant.Keywords: inertial forces, Von Mises stress, hollow rollers, wind turbine
Procedia PDF Downloads 354504 Indigenous Patch Clamp Technique: Design of Highly Sensitive Amplifier Circuit for Measuring and Monitoring of Real Time Ultra Low Ionic Current through Cellular Gates
Authors: Moez ul Hassan, Bushra Noman, Sarmad Hameed, Shahab Mehmood, Asma Bashir
Abstract:
The importance of Noble prize winning “Patch Clamp Technique” is well documented. However, Patch Clamp Technique is very expensive and hence hinders research in developing countries. In this paper, detection, processing and recording of ultra low current from induced cells by using transimpedence amplifier is described. The sensitivity of the proposed amplifier is in the range of femto amperes (fA). Capacitive-feedback is used with active load to obtain a 20MΩ transimpedance gain. The challenging task in designing includes achieving adequate performance in gain, noise immunity and stability. The circuit designed by the authors was able to measure current in the rangeof 300fA to 100pA. Adequate performance shown by the amplifier with different input current and outcome result was found to be within the acceptable error range. Results were recorded using LabVIEW 8.5®for further research.Keywords: drug discovery, ionic current, operational amplifier, patch clamp
Procedia PDF Downloads 519503 Digital Reconstruction of Museum's Statue Using 3D Scanner for Cultural Preservation in Indonesia
Authors: Ahmad Zaini, F. Muhammad Reza Hadafi, Surya Sumpeno, Muhtadin, Mochamad Hariadi
Abstract:
The lack of information about museum’s collection reduces the number of visits of museum. Museum’s revitalization is an urgent activity to increase the number of visits. The research's roadmap is building a web-based application that visualizes museum in the virtual form including museum's statue reconstruction in the form of 3D. This paper describes implementation of three-dimensional model reconstruction method based on light-strip pattern on the museum statue using 3D scanner. Noise removal, alignment, meshing and refinement model's processes is implemented to get a better 3D object reconstruction. Model’s texture derives from surface texture mapping between object's images with reconstructed 3D model. Accuracy test of dimension of the model is measured by calculating relative error of virtual model dimension compared against the original object. The result is realistic three-dimensional model textured with relative error around 4.3% to 5.8%.Keywords: 3D reconstruction, light pattern structure, texture mapping, museum
Procedia PDF Downloads 465502 Intelligent Process Data Mining for Monitoring for Fault-Free Operation of Industrial Processes
Authors: Hyun-Woo Cho
Abstract:
The real-time fault monitoring and diagnosis of large scale production processes is helpful and necessary in order to operate industrial process safely and efficiently producing good final product quality. Unusual and abnormal events of the process may have a serious impact on the process such as malfunctions or breakdowns. This work try to utilize process measurement data obtained in an on-line basis for the safe and some fault-free operation of industrial processes. To this end, this work evaluated the proposed intelligent process data monitoring framework based on a simulation process. The monitoring scheme extracts the fault pattern in the reduced space for the reliable data representation. Moreover, this work shows the results of using linear and nonlinear techniques for the monitoring purpose. It has shown that the nonlinear technique produced more reliable monitoring results and outperforms linear methods. The adoption of the qualitative monitoring model helps to reduce the sensitivity of the fault pattern to noise.Keywords: process data, data mining, process operation, real-time monitoring
Procedia PDF Downloads 640501 Surface Roughness of AlSi/10%AlN Metal Matrix Composite Material Using the Taguchi Method
Authors: Nurul Na'imy Wan, Mohamad Sazali Said, Jaharah Ab. Ghani, Mohd Asri Selamat
Abstract:
This paper presents the surface roughness of the Aluminium silicon alloy (AlSi) matrix composite which has been reinforced with aluminium nitride (AlN), with three types of carbide inserts. Experiments were conducted at various cutting speeds, feed rates, and depths of cut, according to the Taguchi method, using a standard orthogonal array L27 (34). The signal-to-noise (S/N) ratio and analysis of variance are applied to study the characteristic performance of machining parameters in measuring the surface roughness during the milling operation. The analysis of results, using the Taguchi method concluded that a combination of low feed rate, medium depth of cut, low cutting speed, and insert TiB2 give a better value of surface roughness. From Taguchi method, it was found that cutting speed of 230m/min, feed rate of 0.4 mm/tooth, depth of cut of 0.5mm and type of insert of TiB2 were the optimal machining parameters that gave the optimal value of surface roughness.Keywords: AlSi/AlN Metal Matrix Composite (MMC), surface roughness, Taguchi method
Procedia PDF Downloads 462500 Machine Learning Approach for Yield Prediction in Semiconductor Production
Authors: Heramb Somthankar, Anujoy Chakraborty
Abstract:
This paper presents a classification study on yield prediction in semiconductor production using machine learning approaches. A complicated semiconductor production process is generally monitored continuously by signals acquired from sensors and measurement sites. A monitoring system contains a variety of signals, all of which contain useful information, irrelevant information, and noise. In the case of each signal being considered a feature, "Feature Selection" is used to find the most relevant signals. The open-source UCI SECOM Dataset provides 1567 such samples, out of which 104 fail in quality assurance. Feature extraction and selection are performed on the dataset, and useful signals were considered for further study. Afterward, common machine learning algorithms were employed to predict whether the signal yields pass or fail. The most relevant algorithm is selected for prediction based on the accuracy and loss of the ML model.Keywords: deep learning, feature extraction, feature selection, machine learning classification algorithms, semiconductor production monitoring, signal processing, time-series analysis
Procedia PDF Downloads 109499 Blind Super-Resolution Reconstruction Based on PSF Estimation
Authors: Osama A. Omer, Amal Hamed
Abstract:
Successful blind image Super-Resolution algorithms require the exact estimation of the Point Spread Function (PSF). In the absence of any prior information about the imagery system and the true image; this estimation is normally done by trial and error experimentation until an acceptable restored image quality is obtained. Multi-frame blind Super-Resolution algorithms often have disadvantages of slow convergence and sensitiveness to complex noises. This paper presents a Super-Resolution image reconstruction algorithm based on estimation of the PSF that yields the optimum restored image quality. The estimation of PSF is performed by the knife-edge method and it is implemented by measuring spreading of the edges in the reproduced HR image itself during the reconstruction process. The proposed image reconstruction approach is using L1 norm minimization and robust regularization based on a bilateral prior to deal with different data and noise models. A series of experiment results show that the proposed method can outperform other previous work robustly and efficiently.Keywords: blind, PSF, super-resolution, knife-edge, blurring, bilateral, L1 norm
Procedia PDF Downloads 365498 Objective Evaluation on Medical Image Compression Using Wavelet Transformation
Authors: Amhimmid Mohammed Saffour, Mustafa Mohamed Abdullah
Abstract:
The use of computers for handling image data in the healthcare is growing. However, the amount of data produced by modern image generating techniques is vast. This data might be a problem from a storage point of view or when the data is sent over a network. This paper using wavelet transform technique for medical images compression. MATLAB program, are designed to evaluate medical images storage and transmission time problem at Sebha Medical Center Libya. In this paper, three different Computed Tomography images which are abdomen, brain and chest have been selected and compressed using wavelet transform. Objective evaluation has been performed to measure the quality of the compressed images. For this evaluation, the results show that the Peak Signal to Noise Ratio (PSNR) which indicates the quality of the compressed image is ranging from (25.89db to 34.35db for abdomen images, 23.26db to 33.3db for brain images and 25.5db to 36.11db for chest images. These values shows that the compression ratio is nearly to 30:1 is acceptable.Keywords: medical image, Matlab, image compression, wavelet's, objective evaluation
Procedia PDF Downloads 285497 Impact of Hard Limited Clipping Crest Factor Reduction Technique on Bit Error Rate in OFDM Based Systems
Authors: Theodore Grosch, Felipe Koji Godinho Hoshino
Abstract:
In wireless communications, 3GPP LTE is one of the solutions to meet the greater transmission data rate demand. One issue inherent to this technology is the PAPR (Peak-to-Average Power Ratio) of OFDM (Orthogonal Frequency Division Multiplexing) modulation. This high PAPR affects the efficiency of power amplifiers. One approach to mitigate this effect is the Crest Factor Reduction (CFR) technique. In this work, we simulate the impact of Hard Limited Clipping Crest Factor Reduction technique on BER (Bit Error Rate) in OFDM based Systems. In general, the results showed that CFR has more effects on higher digital modulation schemes, as expected. More importantly, we show the worst-case degradation due to CFR on QPSK, 16QAM, and 64QAM signals in a linear system. For example, hard clipping of 9 dB results in a 2 dB increase in signal to noise energy at a 1% BER for 64-QAM modulation.Keywords: bit error rate, crest factor reduction, OFDM, physical layer simulation
Procedia PDF Downloads 366496 Comparative Study of Different Enhancement Techniques for Computed Tomography Images
Authors: C. G. Jinimole, A. Harsha
Abstract:
One of the key problems facing in the analysis of Computed Tomography (CT) images is the poor contrast of the images. Image enhancement can be used to improve the visual clarity and quality of the images or to provide a better transformation representation for further processing. Contrast enhancement of images is one of the acceptable methods used for image enhancement in various applications in the medical field. This will be helpful to visualize and extract details of brain infarctions, tumors, and cancers from the CT image. This paper presents a comparison study of five contrast enhancement techniques suitable for the contrast enhancement of CT images. The types of techniques include Power Law Transformation, Logarithmic Transformation, Histogram Equalization, Contrast Stretching, and Laplacian Transformation. All these techniques are compared with each other to find out which enhancement provides better contrast of CT image. For the comparison of the techniques, the parameters Peak Signal to Noise Ratio (PSNR) and Mean Square Error (MSE) are used. Logarithmic Transformation provided the clearer and best quality image compared to all other techniques studied and has got the highest value of PSNR. Comparison concludes with better approach for its future research especially for mapping abnormalities from CT images resulting from Brain Injuries.Keywords: computed tomography, enhancement techniques, increasing contrast, PSNR and MSE
Procedia PDF Downloads 314