Search results for: uncertainty of processing times
1030 Speech Enhancement Using Kalman Filter in Communication
Authors: Eng. Alaa K. Satti Salih
Abstract:
Revolutions Applications such as telecommunications, hands-free communications, recording, etc. which need at least one microphone, the signal is usually infected by noise and echo. The important application is the speech enhancement, which is done to remove suppressed noises and echoes taken by a microphone, beside preferred speech. Accordingly, the microphone signal has to be cleaned using digital signal processing DSP tools before it is played out, transmitted, or stored. Engineers have so far tried different approaches to improving the speech by get back the desired speech signal from the noisy observations. Especially Mobile communication, so in this paper will do reconstruction of the speech signal, observed in additive background noise, using the Kalman filter technique to estimate the parameters of the Autoregressive Process (AR) in the state space model and the output speech signal obtained by the MATLAB. The accurate estimation by Kalman filter on speech would enhance and reduce the noise then compare and discuss the results between actual values and estimated values which produce the reconstructed signals.
Keywords: Autoregressive Process, Kalman filter, Matlab and Noise speech.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 40251029 Wavelet Feature Selection Approach for Heart Murmur Classification
Authors: G. Venkata Hari Prasad, P. Rajesh Kumar
Abstract:
Phonocardiography is important in appraisal of congenital heart disease and pulmonary hypertension as it reflects the duration of right ventricular systoles. The systolic murmur in patients with intra-cardiac shunt decreases as pulmonary hypertension develops and may eventually disappear completely as the pulmonary pressure reaches systemic level. Phonocardiography and auscultation are non-invasive, low-cost, and accurate methods to assess heart disease. In this work an objective signal processing tool to extract information from phonocardiography signal using Wavelet is proposed to classify the murmur as normal or abnormal. Since the feature vector is large, a Binary Particle Swarm Optimization (PSO) with mutation for feature selection is proposed. The extracted features improve the classification accuracy and were tested across various classifiers including Naïve Bayes, kNN, C4.5, and SVM.Keywords: Phonocardiography, Coiflet, Feature selection, Particle Swarm Optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24731028 Influence of Nanozeolite Particles on Improvement of Clayey Soil
Authors: A. Goodarzian, A. Ghasemipanah, R. Ziaie Moayed, H. Niroumand
Abstract:
The problem of soil stabilization has been one of the important issues in geotechnical engineering. Nowadays, nanomaterials have revolutionized many industries. In this research, improvement of the Kerman fine-grained soil by nanozeolite and nanobentonite additives separately has been investigated using Atterberg Limits and unconfined compression test. In unconfined compression test, the samples were prepared with 3, 5 and 7% nano additives, with 1, 7 and 28 days curing time with strain control method. Finally, the effect of different percentages of nanozeolite and nanobentonite on the geotechnical behavior and characteristics of Kerman fine-grained soil was investigated. The results showed that with increasing the amount of nanozeolite and also nanobentonite to fine-grained soil, the soil exhibits more compression strength. So that by adding 7% nanozeolite and nanobentonite with 1 day curing, the unconfined compression strength is 1.18 and 2.1 times higher than the unstabilized soil. In addition, the failure strain decreases in samples containing nanozeolite, whereas it increases in the presence of nanobentonite. Increasing the percentage of nanozeolite and nanobentonite also increased the elasticity modulus of soil.
Keywords: Nanozeolite particles, nanobentonite particles, clayey soil, unconfined compression stress, specific surface area, cation exchange capacity, Atterberg limits.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8741027 Improving the Performance of Deep Learning in Facial Emotion Recognition with Image Sharpening
Authors: Ksheeraj Sai Vepuri, Nada Attar
Abstract:
We as humans use words with accompanying visual and facial cues to communicate effectively. Classifying facial emotion using computer vision methodologies has been an active research area in the computer vision field. In this paper, we propose a simple method for facial expression recognition that enhances accuracy. We tested our method on the FER-2013 dataset that contains static images. Instead of using Histogram equalization to preprocess the dataset, we used Unsharp Mask to emphasize texture and details and sharpened the edges. We also used ImageDataGenerator from Keras library for data augmentation. Then we used Convolutional Neural Networks (CNN) model to classify the images into 7 different facial expressions, yielding an accuracy of 69.46% on the test set. Our results show that using image preprocessing such as the sharpening technique for a CNN model can improve the performance, even when the CNN model is relatively simple.Keywords: Facial expression recognition, image pre-processing, deep learning, CNN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5441026 Application of Mutual Information based Least dependent Component Analysis (MILCA) for Removal of Ocular Artifacts from Electroencephalogram
Authors: V Krishnaveni, S Jayaraman, K Ramadoss
Abstract:
The electrical potentials generated during eye movements and blinks are one of the main sources of artifacts in Electroencephalogram (EEG) recording and can propagate much across the scalp, masking and distorting brain signals. In recent times, signal separation algorithms are used widely for removing artifacts from the observed EEG data. In this paper, a recently introduced signal separation algorithm Mutual Information based Least dependent Component Analysis (MILCA) is employed to separate ocular artifacts from EEG. The aim of MILCA is to minimize the Mutual Information (MI) between the independent components (estimated sources) under a pure rotation. Performance of this algorithm is compared with eleven popular algorithms (Infomax, Extended Infomax, Fast ICA, SOBI, TDSEP, JADE, OGWE, MS-ICA, SHIBBS, Kernel-ICA, and RADICAL) for the actual independence and uniqueness of the estimated source components obtained for different sets of EEG data with ocular artifacts by using a reliable MI Estimator. Results show that MILCA is best in separating the ocular artifacts and EEG and is recommended for further analysis.
Keywords: Electroencephalogram, Ocular Artifacts (OA), Independent Component Analysis (ICA), Mutual Information (MI), Mutual Information based Least dependent Component Analysis(MILCA)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21931025 Performance Evaluation of Energy Efficient Communication Protocol for Mobile Ad Hoc Networks
Authors: Toshihiko Sasama, Kentaro Kishida, Kazunori Sugahara, Hiroshi Masuyama
Abstract:
A mobile ad hoc network is a network of mobile nodes without any notion of centralized administration. In such a network, each mobile node behaves not only as a host which runs applications but also as a router to forward packets on behalf of others. Clustering has been applied to routing protocols to achieve efficient communications. A CH network expresses the connected relationship among cluster-heads. This paper discusses the methods for constructing a CH network, and produces the following results: (1) The required running costs of 3 traditional methods for constructing a CH network are not so different from each other in the static circumstance, or in the dynamic circumstance. Their running costs in the static circumstance do not differ from their costs in the dynamic circumstance. Meanwhile, although the routing costs required for the above 3 methods are not so different in the static circumstance, the costs are considerably different from each other in the dynamic circumstance. Their routing costs in the static circumstance are also very different from their costs in the dynamic circumstance, and the former is one tenths of the latter. The routing cost in the dynamic circumstance is mostly the cost for re-routing. (2) On the strength of the above results, we discuss new 2 methods regarding whether they are tolerable or not in the dynamic circumstance, that is, whether the times of re-routing are small or not. These new methods are revised methods that are based on the traditional methods. We recommended the method which produces the smallest routing cost in the dynamic circumstance, therefore producing the smallest total cost.Keywords: cluster, mobile ad hoc network, re-routing cost, simulation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13501024 A Partially Accelerated Life Test Planning with Competing Risks and Linear Degradation Path under Tampered Failure Rate Model
Authors: Fariba Azizi, Firoozeh Haghighi, Viliam Makis
Abstract:
In this paper, we propose a method to model the relationship between failure time and degradation for a simple step stress test where underlying degradation path is linear and different causes of failure are possible. It is assumed that the intensity function depends only on the degradation value. No assumptions are made about the distribution of the failure times. A simple step-stress test is used to shorten failure time of products and a tampered failure rate (TFR) model is proposed to describe the effect of the changing stress on the intensities. We assume that some of the products that fail during the test have a cause of failure that is only known to belong to a certain subset of all possible failures. This case is known as masking. In the presence of masking, the maximum likelihood estimates (MLEs) of the model parameters are obtained through an expectation-maximization (EM) algorithm by treating the causes of failure as missing values. The effect of incomplete information on the estimation of parameters is studied through a Monte-Carlo simulation. Finally, a real example is analyzed to illustrate the application of the proposed methods.Keywords: Expectation-maximization (EM) algorithm, cause of failure, intensity, linear degradation path, masked data, reliability function.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10731023 Video Quality Assessment Methods: A Bird’s-Eye View
Authors: P. M. Arun Kumar, S. Chandramathi
Abstract:
The proliferation of multimedia technology and services in today’s world provide ample research scope in the frontiers of visual signal processing. Wide spread usage of video based applications in heterogeneous environment needs viable methods of Video Quality Assessment (VQA). The evaluation of video quality not only depends on high QoS requirements but also emphasis the need of novel term ‘QoE’ (Quality of Experience) that perceive video quality as user centric. This paper discusses two vital video quality assessment methods namely, subjective and objective assessment methods. The evolution of various video quality metrics, their classification models and applications are reviewed in this work. The Mean Opinion Score (MOS) based subjective measurements and algorithm based objective metrics are discussed and their challenges are outlined. Further, this paper explores the recent progress of VQA in emerging technologies such as mobile video and 3D video.
Keywords: 3D-Video, no reference metric, quality of experience, video quality assessment, video quality metrics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 40531022 Cost-Effective Design of Space Structures Joints: A Review
Authors: Mohammed I. Ali, Feng Fan, Peter N. Khakina, Ma H.H
Abstract:
In construction of any structure, the aesthetic and utility values should be considered in such a way as to make the structure cost-effective. Most structures are composed of elements and joints which are very critical in any skeletal space structure because they majorly determine the performance of the structure. In early times, most space structures were constructed using rigid joints which had the advantage of better performing structures as compared to pin-jointed structures but with the disadvantage of requiring all the construction work to be done on site. The discovery of semi-rigid joints now enables connections to be prefabricated and quickly assembled on site while maintaining good performance. In this paper, cost-effective is discussed basing on strength of connectors at the joints, buckling of joints and overall structure, and the effect of initial geometrical imperfections. Several existing joints are reviewed by classifying them into categories and discussing where they are most suited and how they perform structurally. Also, finite element modeling using ABAQUS is done to determine the buckling behavior. It is observed that some joints are more economical than others. The rise to span ratio and imperfections are also found to affect the buckling of the structures. Based on these, general principles that guide the design of cost-effective joints and structures are discussed.
Keywords: Buckling, Connectors, Joint stiffness, Eccentricity, Second moment of area, Semi-rigid joints.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 47761021 A Video-based Algorithm for Moving Objects Detection at Signalized Intersection
Authors: Juan Li, Chunfu Shao, Chunjiao Dong, Dan Zhao, Yinhong Liu
Abstract:
Mixed-traffic (e.g., pedestrians, bicycles, and vehicles) data at an intersection is one of the essential factors for intersection design and traffic control. However, some data such as pedestrian volume cannot be directly collected by common detectors (e.g. inductive loop, sonar and microwave sensors). In this paper, a video based detection algorithm is proposed for mixed-traffic data collection at intersections using surveillance cameras. The algorithm is derived from Gaussian Mixture Model (GMM), and uses a mergence time adjustment scheme to improve the traditional algorithm. Real-world video data were selected to test the algorithm. The results show that the proposed algorithm has the faster processing speed and more accuracy than the traditional algorithm. This indicates that the improved algorithm can be applied to detect mixed-traffic at signalized intersection, even when conflicts occur.Keywords: detection, intersection, mixed traffic, moving objects.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20331020 Classification of Computer Generated Images from Photographic Images Using Convolutional Neural Networks
Authors: Chaitanya Chawla, Divya Panwar, Gurneesh Singh Anand, M. P. S Bhatia
Abstract:
This paper presents a deep-learning mechanism for classifying computer generated images and photographic images. The proposed method accounts for a convolutional layer capable of automatically learning correlation between neighbouring pixels. In the current form, Convolutional Neural Network (CNN) will learn features based on an image's content instead of the structural features of the image. The layer is particularly designed to subdue an image's content and robustly learn the sensor pattern noise features (usually inherited from image processing in a camera) as well as the statistical properties of images. The paper was assessed on latest natural and computer generated images, and it was concluded that it performs better than the current state of the art methods.Keywords: Image forensics, computer graphics, classification, deep learning, convolutional neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11751019 Performance Analysis of Parallel Client-Server Model Versus Parallel Mobile Agent Model
Authors: K. B. Manwade, G. A. Patil
Abstract:
Mobile agent has motivated the creation of a new methodology for parallel computing. We introduce a methodology for the creation of parallel applications on the network. The proposed Mobile-Agent parallel processing framework uses multiple Javamobile Agents. Each mobile agent can travel to the specified machine in the network to perform its tasks. We also introduce the concept of master agent, which is Java object capable of implementing a particular task of the target application. Master agent is dynamically assigns the task to mobile agents. We have developed and tested a prototype application: Mobile Agent Based Parallel Computing. Boosted by the inherited benefits of using Java and Mobile Agents, our proposed methodology breaks the barriers between the environments, and could potentially exploit in a parallel manner all the available computational resources on the network. This paper elaborates performance issues of a mobile agent for parallel computing.Keywords: Parallel Computing, Mobile Agent.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16531018 Effects of Pressure and Temperature on the Extraction of Benzyl Isothiocyanate by Supercritical Fluids from Tropaeolum majus L. Leaves
Authors: Espinoza S. Clara, Gamarra Q. Flor, Marianela F. Ramos Quispe S. Miguel, Flores R. Omar
Abstract:
Tropaeolum majus L. is a native plant to South and Central America, used since ancient times by our ancestors to combat different diseases. Glucotropaeolonin is one of its main components, which when hydrolyzed, forms benzyl isothiocyanate (BIT) that promotes cellular apoptosis (programmed cell death in cancer cells). Therefore, the present research aims to evaluate the effect of the pressure and temperature of BIT extraction by supercritical CO2 from Tropaeolum majus L. The extraction was carried out in a supercritical fluid extractor equipment Speed SFE BASIC Brand: Poly science, the leaves of Tropaeolum majus L. were ground for one hour and lyophilized until obtaining a humidity of 6%. The extraction with supercritical CO2 was carried out with pressures of 200 bar and 300 bar, temperatures of 50°C, 60°C and 70°C, obtained by the conjugation of these six treatments. BIT was identified by thin layer chromatography using 98% BIT as the standard, and as the mobile phase hexane: dichloromethane (4:2). Subsequently, BIT quantification was performed by high performance liquid chromatography (HPLC). The highest yield of oleoresin by supercritical CO2 extraction was obtained pressure 300 bar and temperature at 60°C; and the higher content of BIT at pressure 200 bar and 70°C for 30 minutes to obtain 113.615 ± 0.03 mg BIT/100 g dry matter was obtained.
Keywords: Tropaeolum majus L., supercritical fluids, benzyl isothiocyanate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8791017 The Effect of the COVID-19 Pandemic on Foreign Students Studying in Hungary
Authors: Anita Kéri
Abstract:
Satisfying foreign student needs has been in the center of research interest in the past several years. Higher education institutions have been exploring factors influencing foreign student satisfaction to stay competitive on the educational market. Even though foreign student satisfaction and loyalty are topics investigated deeply in the literature, the academic years of 2020 and 2021 have revealed challenges never experienced before. With the COVID-19 pandemic, new factors have emerged that might influence foreign student satisfaction and loyalty in higher education. The aim of the current research is to shed lights on what factors influence foreign student satisfaction and loyalty in the post-pandemic educational era, and to reveal if the effects of factors influencing satisfaction and loyalty have changed compared to previous findings. Initial results show that students are less willing to participate in online surveys during and after the pandemic. The return rate of the survey instrument is below 5%. Results also reveal that there is a slight difference in what factors students deem important during pandemic times regarding their satisfaction and loyalty. The results of the current study help us determine what factors higher education institutions need to consider, when planning the future service affordances for their foreign students, that might influence their satisfaction and loyalty.
Keywords: COVID-19, foreign students, loyalty, pandemic, satisfaction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3191016 Laboratory Experimentation for Supporting Collaborative Working in Engineering Education over the Internet
Authors: S. Odeh, E. Abdelghani
Abstract:
Collaborative working environments for distance education can be considered as a more generic form of contemporary remote labs. At present, the majority of existing real laboratories are not constructed to allow the involved participants to collaborate in real time. To make this revolutionary learning environment possible we must allow the different users to carry out an experiment simultaneously. In recent times, multi-user environments are successfully applied in many applications such as air traffic control systems, team-oriented military systems, chat-text tools, multi-player games etc. Thus, understanding the ideas and techniques behind these systems could be of great importance in the contribution of ideas to our e-learning environment for collaborative working. In this investigation, collaborative working environments from theoretical and practical perspectives are considered in order to build an effective collaborative real laboratory, which allows two students or more to conduct remote experiments at the same time as a team. In order to achieve this goal, we have implemented distributed system architecture, enabling students to obtain an automated help by either a human tutor or a rule-based e-tutor.Keywords: Collaboration environment, e-tutor, multi-user environments, socio-technical system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14831015 Processor Scheduling on Parallel Computers
Authors: Mohammad S. Laghari, Gulzar A. Khuwaja
Abstract:
Many problems in computer vision and image processing present potential for parallel implementations through one of the three major paradigms of geometric parallelism, algorithmic parallelism and processor farming. Static process scheduling techniques are used successfully to exploit geometric and algorithmic parallelism, while dynamic process scheduling is better suited to dealing with the independent processes inherent in the process farming paradigm. This paper considers the application of parallel or multi-computers to a class of problems exhibiting spatial data characteristic of the geometric paradigm. However, by using processor farming paradigm, a dynamic scheduling technique is developed to suit the MIMD structure of the multi-computers. A hybrid scheme of scheduling is also developed and compared with the other schemes. The specific problem chosen for the investigation is the Hough transform for line detection.Keywords: Hough transforms, parallel computer, parallel paradigms, scheduling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16501014 GPU Implementation for Solving in Compressible Two-Phase Flows
Authors: Sheng-Hsiu Kuo, Pao-Hsiung Chiu, Reui-Kuo Lin, Yan-Ting Lin
Abstract:
A one-step conservative level set method, combined with a global mass correction method, is developed in this study to simulate the incompressible two-phase flows. The present framework do not need to solve the conservative level set scheme at two separated steps, and the global mass can be exactly conserved. The present method is then more efficient than two-step conservative level set scheme. The dispersion-relation-preserving schemes are utilized for the advection terms. The pressure Poisson equation solver is applied to GPU computation using the pCDR library developed by National Center for High-Performance Computing, Taiwan. The SMP parallelization is used to accelerate the rest of calculations. Three benchmark problems were done for the performance evaluation. Good agreements with the referenced solutions are demonstrated for all the investigated problems.
Keywords: Conservative level set method, two-phase flow, dispersion-relation-preserving, graphics processing unit (GPU), multi-threading.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20651013 Optical Flow Based Moving Object Detection and Tracking for Traffic Surveillance
Authors: Sepehr Aslani, Homayoun Mahdavi-Nasab
Abstract:
Automated motion detection and tracking is a challenging task in traffic surveillance. In this paper, a system is developed to gather useful information from stationary cameras for detecting moving objects in digital videos. The moving detection and tracking system is developed based on optical flow estimation together with application and combination of various relevant computer vision and image processing techniques to enhance the process. To remove noises, median filter is used and the unwanted objects are removed by applying thresholding algorithms in morphological operations. Also the object type restrictions are set using blob analysis. The results show that the proposed system successfully detects and tracks moving objects in urban videos.
Keywords: Optical flow estimation, moving object detection, tracking, morphological operation, blob analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 101561012 The Role of Paper in the Copy Identification of Safavid Era Shahnamehs of Tabriz Doctrine
Authors: Ashrafosadat Mousavi Lar, Elahe Moravej
Abstract:
To investigate and explain the history of each copy, we must refer to its past because it highlights parts of the civilization of people among which this copy has been codified. In this paper, eight Ferdowsi’s Shahnameh of Safavid era of Tabriz doctrine available in Iranian libraries and museums are studied. Undoubtedly, it can be said that Ferdowsi’s Shahnameh is one of the most important books that has been transcribed many times in different eras because it explains the Iranian champions’ prowess and it includes the history of Iran from Pishdadian to Sasanian dynasty. In addition, it has been attractive for governors and artists. The research methodology of this article is based on the analytical-descriptive arguments. The research hypothesis is based on papers used in Shahnameh writing in Safavid era of Tabriz doctrine were mostly Isfahanian papers existed. At that time, Isfahanian paper was unique in terms of quality, clarity, flatness of the sheets, volume, shape, softness and elegance, strength, and smoothness. This paper was mostly used to prepare the courtier and exquisite copies. This shows that the prepared copies in Safavid era of Tabriz doctrine were very important because the artists and people who ordered and were out of the court have ordered Isfahanian paper for writing their books.
Keywords: Shahnameh, Safavid era, Tabriz doctrine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7801011 Zero Voltage Switched Full Bridge Converters for the Battery Charger of Electric Vehicle
Authors: Rizwan Ullah, Abdar Ali, Zahid Ullah
Abstract:
This paper illustrates the study of three isolated zero voltage switched (ZVS) PWM full bridge (FB) converters to charge the high voltage battery in the charger of electric vehicle (EV). EV battery chargers have several challenges such as high efficiency, high reliability, low cost, isolation, and high power density. The cost of magnetic and filter components in the battery charger is reduced when switching frequency is increased. The increase in the switching frequency increases switching losses. ZVS is used to reduce switching losses and to operate the converter in the battery charger at high frequency. The performance of each of the three converters is evaluated on the basis of ZVS range, dead times of the switches, conduction losses of switches, circulating current stress, circulating energy, duty cycle loss, and efficiency. The limitations and merits of each PWM FB converter are reviewed. The converter with broader ZVS range, high efficiency and low switch stresses is selected for battery charger applications in EV.Keywords: Electric vehicle, PWM FB converter, zero voltage switching, circulating energy, duty cycle loss, battery charger.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24291010 Incremental Learning of Independent Topic Analysis
Authors: Takahiro Nishigaki, Katsumi Nitta, Takashi Onoda
Abstract:
In this paper, we present a method of applying Independent Topic Analysis (ITA) to increasing the number of document data. The number of document data has been increasing since the spread of the Internet. ITA was presented as one method to analyze the document data. ITA is a method for extracting the independent topics from the document data by using the Independent Component Analysis (ICA). ICA is a technique in the signal processing; however, it is difficult to apply the ITA to increasing number of document data. Because ITA must use the all document data so temporal and spatial cost is very high. Therefore, we present Incremental ITA which extracts the independent topics from increasing number of document data. Incremental ITA is a method of updating the independent topics when the document data is added after extracted the independent topics from a just previous the data. In addition, Incremental ITA updates the independent topics when the document data is added. And we show the result applied Incremental ITA to benchmark datasets.Keywords: Text mining, topic extraction, independent, incremental, independent component analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10591009 Single Frame Supercompression of Still Images,Video, High Definition TV and Digital Cinema
Authors: Mario Mastriani
Abstract:
Super-resolution is nowadays used for a high-resolution image produced from several low-resolution noisy frames. In this work, we consider the problem of high-quality interpolation of a single noise-free image. Such images may come from different sources, i.e., they may be frames of videos, individual pictures, etc. On the other hand, in the encoder we apply a downsampling via bidimen-sional interpolation of each frame, and in the decoder we apply a upsampling by which we restore the original size of the image. If the compression ratio is very high, then we use a convolutive mask that restores the edges, eliminating the blur. Finally, both, the encoder and the complete decoder are implemented on General-Purpose computation on Graphics Processing Units (GPGPU) cards. In fact, the mentioned mask is coded inside texture memory of a GPGPU.Keywords: General-Purpose computation on Graphics ProcessingUnits, Image Compression, Interpolation, Super-resolution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19991008 Development and Optimization of Automated Dry-Wafer Separation
Authors: Tim Giesen, Christian Fischmann, Fabian Böttinger, Alexander Ehm, Alexander Verl
Abstract:
In a state-of-the-art industrial production line of photovoltaic products the handling and automation processes are of particular importance and implication. While processing a fully functional crystalline solar cell an as-cut photovoltaic wafer is subject to numerous repeated handling steps. With respect to stronger requirements in productivity and decreasing rejections due to defects the mechanical stress on the thin wafers has to be reduced to a minimum as the fragility increases by decreasing wafer thicknesses. In relation to the increasing wafer fragility, researches at the Fraunhofer Institutes IPA and CSP showed a negative correlation between multiple handling processes and the wafer integrity. Recent work therefore focused on the analysis and optimization of the dry wafer stack separation process with compressed air. The achievement of a wafer sensitive process capability and a high production throughput rate is the basic motivation in this research.Keywords: Automation, Photovoltaic Manufacturing, Thin Wafer, Material Handling
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16711007 Method Development and Validation for the Determination of Cefixime in Pure and Commercial Dosage Forms by Specrophotometry
Authors: S. N. H. Azmi, B. Iqbal, J. K. Al Mamari, K. A. Al Hattali, W. N. Al Hadhrami
Abstract:
A simple, accurate and precise direct spectrophotometric method has been developed for the determination of cefixime in tablets and capsules. The method is based on the reaction of cefixime with a mixture of potassium iodide and potassium iodate to form yellow coloured product in ethanol-distilled water medium at room temperature which absorbed maximally at 352 nm. The factors affecting the reaction product were carefully studied and optimized. The validation parameters based on International Conference on Harmonisation (ICH, USA) guidelines were followed. The effect of common excipients used as additives has been tested and the tolerance limit was calculated for the determination of cefixime. Beer’s law is obeyed in the concentration range of 4 – 24 ug mL-1 with apparent molar absorptivity of 1.52 × 104 L mol-1cm-1 and Sandell’s sensitivity of 0.033 ug/cm2/ 0.001 absorbance unit. The limits of detection and quantitation for the proposed method are 0.32 and 1.06 ug mL-1, respectively. The proposed method has been successfully applied for the determination of cefixime in pharmaceutical formulations. The results obtained by the proposed method were statistically compared with the reference method using t- and F- values and found no significant difference between the two methods. The proposed method can be used as an alternate method for routine quality control analysis of cefixime in pharmaceutical formulations.
Keywords: Spectrophotometry, cefixime, validation, pharmaceutical formulations.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31631006 Simulating Economic Order Quantity and Reorder Point Policy for a Repairable Items Inventory System
Authors: Mojahid F. Saeed Osman
Abstract:
Repairable items inventory system is a management tool used to incorporate all information concerning inventory levels and movements for repaired and new items. This paper presents development of an effective simulation model for managing the inventory of repairable items for a production system where production lines send their faulty items to a repair shop considering the stochastic failure behavior and repair times. The developed model imitates the process of handling the on-hand inventory of repaired items and the replenishment of the inventory of new items using Economic Order Quantity and Reorder Point ordering policy in a flexible and risk-free environment. We demonstrate the appropriateness and effectiveness of the proposed simulation model using an illustrative case problem. The developed simulation model can be used as a reliable tool for estimating a healthy on-hand inventory of new and repaired items, backordered items, and downtime due to unavailability of repaired items, and validating and examining Economic Order Quantity and Reorder Point ordering policy, which would further be compared with other ordering strategies as future work.
Keywords: Inventory system, repairable items, simulation, maintenance, economic order quantity, reorder point.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6761005 Impact of Process Parameters on Tensile Strength of Fused Deposition Modeling Printed Crisscross Poylactic Acid
Authors: Shilpesh R. Rajpurohit, Harshit K. Dave
Abstract:
Additive manufacturing gains the popularity in recent times, due to its capability to create prototype as well functional as end use product directly from CAD data without any specific requirement of tooling. Fused deposition modeling (FDM) is one of the widely used additive manufacturing techniques that are used to create functional end use part of polymer that is comparable with the injection-molded parts. FDM printed part has an application in various fields such as automobile, aerospace, medical, electronic, etc. However, application of FDM part is greatly affected by poor mechanical properties. Proper selection of the process parameter could enhance the mechanical performance of the printed part. In the present study, experimental investigation has been carried out to study the behavior of the mechanical performance of the printed part with respect to process variables. Three process variables viz. raster angle, raster width and layer height have been varied to understand its effect on tensile strength. Further, effect of process variables on fractured surface has been also investigated.
Keywords: 3D printing, fused deposition modeling, layer height, raster angle, raster width, tensile strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16601004 Hydrogen and Biofuel Production from 2-Propanol Over Ru/Al2O3 Catalyst in Supercritical Water
Authors: Ekin Kıpçak, Yağmur Karakuş, Mesut Akgün
Abstract:
Hydrogen is an important chemical in many industries and it is expected to become one of the major fuels for energy generation in the future. Unfortunately, hydrogen does not exist in its elemental form in nature and therefore has to be produced from hydrocarbons, hydrogen-containing compounds or water.
Above its critical point (374.8oC and 22.1MPa), water has lower density and viscosity, and a higher heat capacity than those of ambient water. Mass transfer in supercritical water (SCW) is enhanced due to its increased diffusivity and transport ability. The reduced dielectric constant makes supercritical water a better solvent for organic compounds and gases. Hence, due to the aforementioned desirable properties, there is a growing interest toward studies regarding the gasification of organic matter containing biomass or model biomass solutions in supercritical water.
In this study, hydrogen and biofuel production by the catalytic gasification of 2-Propanol in supercritical conditions of water was investigated. Ru/Al2O3 was the catalyst used in the gasification reactions. All of the experiments were performed under a constant pressure of 25 MPa. The effects of five reaction temperatures (400, 450, 500, 550 and 600oC) and five reaction times (10, 15, 20, 25 and 30 s) on the gasification yield and flammable component content were investigated.
Keywords: 2-Propanol, Gasification, Ru/Al2O3, Supercritical water.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21311003 Digital Watermarking Based on Visual Cryptography and Histogram
Authors: R. Rama Kishore, Sunesh
Abstract:
Nowadays, robust and secure watermarking algorithm and its optimization have been need of the hour. A watermarking algorithm is presented to achieve the copy right protection of the owner based on visual cryptography, histogram shape property and entropy. In this, both host image and watermark are preprocessed. Host image is preprocessed by using Butterworth filter, and watermark is with visual cryptography. Applying visual cryptography on water mark generates two shares. One share is used for embedding the watermark, and the other one is used for solving any dispute with the aid of trusted authority. Usage of histogram shape makes the process more robust against geometric and signal processing attacks. The combination of visual cryptography, Butterworth filter, histogram, and entropy can make the algorithm more robust, imperceptible, and copy right protection of the owner.
Keywords: Butterworth filter, digital watermarking, histogram, visual cryptography.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16781002 Treatment of the Modern Management Mechanism of the Debris Flow Processes Expected in the Mletiskhevi
Authors: G. Chakhaia, S. Gogilava, L. Tsulukidze, Z. Laoshvili, I. Khubulava, S. Bosikashvili, T. Gugushvili
Abstract:
The work reviewed and evaluated various genesis debris flow phenomena recently formatted in the Mletiskhevi, accordingly it revealed necessity of treatment modern debris flow against measures. Based on this, it is proposed the debris flow against truncated semi cone shape construction, which elements are contained in the car’s secondary tires. its constituent elements (sections), due to the possibilities of amortization and geometric shapes is effective and sustainable towards debris flow hitting force. The construction is economical, because after crossing the debris flows in the river bed, the riverbed is not cleanable, also the elements of the building are resource saving. For assessment of influence of cohesive debris flow at the construction and evaluation of the construction effectiveness have been implemented calculation in the specific assumptions with approved methodology. According to the calculation, it was established that after passing debris flow in the debris flow construction (in 3 row case) its hitting force reduces 3 times, that causes reduce of debris flow speed and kinetic energy, as well as sedimentation on a certain section of water drain in the lower part of the construction. Based on the analysis and report on the debris flow against construction, it can be said that construction is effective, inexpensive, technically relatively easy-to-reach measure, that’s why its implementation is prospective.
Keywords: Construction, debris flow, sections, theoretical calculation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4001001 Prediction of Tool and Nozzle Flow Behavior in Ultrasonic Machining Process
Authors: Vinod Kumar, Jatinder Kumar
Abstract:
The use of hard and brittle material has become increasingly more extensive in recent years. Therefore processing of these materials for the parts fabrication has become a challenging problem. However, it is time-consuming to machine the hard brittle materials with the traditional metal-cutting technique that uses abrasive wheels. In addition, the tool would suffer excessive wear as well. However, if ultrasonic energy is applied to the machining process and coupled with the use of hard abrasive grits, hard and brittle materials can be effectively machined. Ultrasonic machining process is mostly used for the brittle materials. The present research work has developed models using finite element approach to predict the mechanical stresses sand strains produced in the tool during ultrasonic machining process. Also the flow behavior of abrasive slurry coming out of the nozzle has been studied for simulation using ANSYS CFX module. The different abrasives of different grit sizes have been used for the experimentation work.Keywords: Stress, MRR, Flow, Ultrasonic Machining
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2810