Search results for: Multimodality images
280 Investigating the Effect of Velocity Inlet and Carrying Fluid on the Flow inside Coronary Artery
Authors: Mohammadreza Nezamirad, Nasim Sabetpour, Azadeh Yazdi, Amirmasoud Hamedi
Abstract:
In this study OpenFOAM 4.4.2 was used to investigate flow inside the coronary artery of the heart. This step is the first step of our future project, which is to include conjugate heat transfer of the heart with three main coronary arteries. Three different velocities were used as inlet boundary conditions to see the effect of velocity increase on velocity, pressure, and wall shear of the coronary artery. Also, three different fluids, namely the University of Wisconsin solution, gelatin, and blood was used to investigate the effect of different fluids on flow inside the coronary artery. A code based on Reynolds Stress Navier Stokes (RANS) equations was written and implemented with the real boundary condition that was calculated based on MRI images. In order to improve the accuracy of the current numerical scheme, hex dominant mesh is utilized. When the inlet velocity increases to 0.5 m/s, velocity, wall shear stress, and pressure increase at the narrower parts.
Keywords: CFD, heart, simulation, OpenFOAM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 457279 Segmentation and Recognition of Handwritten Numeric Chains
Authors: Salim Ouchtati, Bedda Mouldi, Abderrazak Lachouri
Abstract:
In this paper we present an off line system for the recognition of the handwritten numeric chains. Our work is divided in two big parts. The first part is the realization of a recognition system of the isolated handwritten digits. In this case the study is based mainly on the evaluation of neural network performances, trained with the gradient back propagation algorithm. The used parameters to form the input vector of the neural network are extracted on the binary images of the digits by several methods: the distribution sequence, the Barr features and the centred moments of the different projections and profiles. The second part is the extension of our system for the reading of the handwritten numeric chains constituted of a variable number of digits. The vertical projection is used to segment the numeric chain at isolated digits and every digit (or segment) will be presented separately to the entry of the system achieved in the first part (recognition system of the isolated handwritten digits). The result of the recognition of the numeric chain will be displayed at the exit of the global system.Keywords: Optical Characters Recognition, Neural networks, Barr features, Image processing, Pattern Recognition, Featuresextraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1433278 Automatic Vehicle Identification by Plate Recognition
Authors: Serkan Ozbay, Ergun Ercelebi
Abstract:
Automatic Vehicle Identification (AVI) has many applications in traffic systems (highway electronic toll collection, red light violation enforcement, border and customs checkpoints, etc.). License Plate Recognition is an effective form of AVI systems. In this study, a smart and simple algorithm is presented for vehicle-s license plate recognition system. The proposed algorithm consists of three major parts: Extraction of plate region, segmentation of characters and recognition of plate characters. For extracting the plate region, edge detection algorithms and smearing algorithms are used. In segmentation part, smearing algorithms, filtering and some morphological algorithms are used. And finally statistical based template matching is used for recognition of plate characters. The performance of the proposed algorithm has been tested on real images. Based on the experimental results, we noted that our algorithm shows superior performance in car license plate recognition.Keywords: Character recognizer, license plate recognition, plate region extraction, segmentation, smearing, template matching.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7586277 Mining Image Features in an Automatic Two-Dimensional Shape Recognition System
Authors: R. A. Salam, M.A. Rodrigues
Abstract:
The number of features required to represent an image can be very huge. Using all available features to recognize objects can suffer from curse dimensionality. Feature selection and extraction is the pre-processing step of image mining. Main issues in analyzing images is the effective identification of features and another one is extracting them. The mining problem that has been focused is the grouping of features for different shapes. Experiments have been conducted by using shape outline as the features. Shape outline readings are put through normalization and dimensionality reduction process using an eigenvector based method to produce a new set of readings. After this pre-processing step data will be grouped through their shapes. Through statistical analysis, these readings together with peak measures a robust classification and recognition process is achieved. Tests showed that the suggested methods are able to automatically recognize objects through their shapes. Finally, experiments also demonstrate the system invariance to rotation, translation, scale, reflection and to a small degree of distortion.Keywords: Image mining, feature selection, shape recognition, peak measures.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1458276 Application of LSB Based Steganographic Technique for 8-bit Color Images
Authors: Mamta Juneja, Parvinder S. Sandhu, Ekta Walia
Abstract:
Steganography is the process of hiding one file inside another such that others can neither identify the meaning of the embedded object, nor even recognize its existence. Current trends favor using digital image files as the cover file to hide another digital file that contains the secret message or information. One of the most common methods of implementation is Least Significant Bit Insertion, in which the least significant bit of every byte is altered to form the bit-string representing the embedded file. Altering the LSB will only cause minor changes in color, and thus is usually not noticeable to the human eye. While this technique works well for 24-bit color image files, steganography has not been as successful when using an 8-bit color image file, due to limitations in color variations and the use of a colormap. This paper presents the results of research investigating the combination of image compression and steganography. The technique developed starts with a 24-bit color bitmap file, then compresses the file by organizing and optimizing an 8-bit colormap. After the process of compression, a text message is hidden in the final, compressed image. Results indicate that the final technique has potential of being useful in the steganographic world.
Keywords: Compression, Colormap, Encryption, Steganographyand LSB Insertion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3001275 Information Security in E-Learning through Identification of Humans
Authors: Hassan Haleh, Zohreh Nasiri, Parisa Farahpour
Abstract:
During recent years, the traditional learning approaches have undergone fundamental changes due to the emergence of new technologies such as multimedia, hypermedia and telecommunication. E-learning is a modern world phenomenon that has come into existence in the information age and in a knowledgebased society. E-learning has developed significantly within a short period of time. Thus it is of a great significant to secure information, allow a confident access and prevent unauthorized accesses. Making use of individuals- physiologic or behavioral (biometric) properties is a confident method to make the information secure. Among the biometrics, fingerprint is more acceptable and most countries use it as an efficient methods of identification. This article provides a new method to compare the fingerprint comparison by pattern recognition and image processing techniques. To verify fingerprint, the shortest distance method is used together with perceptronic multilayer neural network functioning based on minutiae. This method is highly accurate in the extraction of minutiae and it accelerates comparisons due to elimination of false minutiae and is more reliable compared with methods that merely use directional images.Keywords: Fingerprint, minutiae, extraction of properties, multilayer neural network
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1649274 Synthesis and Characterization of Silver/Polylactide Nanocomposites
Authors: Kamyar Shameli, Mansor Bin Ahmad, Wan Md Zin Wan Yunus, Nor Azowa Ibrahim, Maryam Jokar, Majid Darroudi
Abstract:
Silver/polylactide nanocomposites (Ag/PLA-NCs) were synthesized via chemical reduction method in diphase solvent. Silver nitrate and sodium borohydride were used as a silver precursor and reducing agent in the polylactide (PLA). The properties of Ag/PLA-NCs were studied as a function of the weight percentages of silver nanoparticles (8, 16 and 32 wt% of Ag-NPs) relative to the weight of PLA. The Ag/PLA-NCs were characterized by Xray diffraction (XRD), transmission electron microscopy (TEM), electro-optical microscopy (EOM), UV-visible spectroscopy (UV-vis) and Fourier transform infrared spectroscopy (FT-IR). XRD patterns confirmed that Ag-NPs crystallographic planes were face centered cubic (fcc) type. TEM images showed that mean diameters of Ag-NPs were 3.30, 3.80 and 4.80 nm. Electro-optical microscopy revealed excellent dispersion and interaction between Ag-NPs and PLA films. The generation of silver nanoparticles was confirmed from the UVvisible spectra. FT-IR spectra showed that there were no significant differences between PLA and Ag/PLA-NCs films. The synthesized Ag/PLA-NCs were stable in organic solution over a long period of time without sign of precipitation.Keywords: Nanocomposites, Polylactide, Silver Nanoparticles, Sodium Borohydride, Transmission Electron Microscopy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3555273 The Potential of Roof Top Rain Water Harvesting as a Water Resource in Jordan: Featuring Two Application Case Studies
Authors: Zain M. Al-Houri, Oday K. Abu-Hadba, Khaled A. Hamdan
Abstract:
Roof top rainwater harvesting (RWH) has been carried out worldwide to provide an inexpensive source of water for many people. This research aims at evaluating the potential of roof top rain water harvesting as a resource in Jordan. For the purpose of this work, two case studies at Al-Jubiha and Shafa-Badran districts in Amman city were selected. All existing rooftops in both districts were identified by digitizing 2012 satellite images of the two districts using Google earth and ArcGIS tools. Rational method was used to estimate the potential volume of rainwater that can be harvested from the digitized rooftops. Results indicated that 1.17 and 0.526 MCM/yr can be harvested in Al-Jubiha and Shafa-Badran districts, respectively. This study should increase the attention to the importance of implementing RWH technique in Jordanian residences as a viable alternative for ensuring a continued source of non-potable water.
Keywords: Amman districts, ArcGIS, Rational method, Roof top rain water harvesting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3304272 Review of the Software Used for 3D Volumetric Reconstruction of the Liver
Authors: P. Strakos, M. Jaros, T. Karasek, T. Kozubek, P. Vavra, T. Jonszta
Abstract:
In medical imaging, segmentation of different areas of human body like bones, organs, tissues, etc. is an important issue. Image segmentation allows isolating the object of interest for further processing that can lead for example to 3D model reconstruction of whole organs. Difficulty of this procedure varies from trivial for bones to quite difficult for organs like liver. The liver is being considered as one of the most difficult human body organ to segment. It is mainly for its complexity, shape versatility and proximity of other organs and tissues. Due to this facts usually substantial user effort has to be applied to obtain satisfactory results of the image segmentation. Process of image segmentation then deteriorates from automatic or semi-automatic to fairly manual one. In this paper, overview of selected available software applications that can handle semi-automatic image segmentation with further 3D volume reconstruction of human liver is presented. The applications are being evaluated based on the segmentation results of several consecutive DICOM images covering the abdominal area of the human body.
Keywords: Image segmentation, semi-automatic, software, 3D volumetric reconstruction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4469271 Conceptualization of the Attractive Work Environment and Organizational Activity for Humans in Future Deep Mines
Authors: M. A. Sanda, B. Johansson, J. Johansson
Abstract:
The purpose of this paper is to conceptualize a futureoriented human work environment and organizational activity in deep mines that entails a vision of good and safe workplace. Futureoriented technological challenges and mental images required for modern work organization design were appraised. It is argued that an intelligent-deep-mine covering the entire value chain, including environmental issues and with work organization that supports good working and social conditions towards increased human productivity could be designed. With such intelligent system and work organization in place, the mining industry could be seen as a place where cooperation, skills development and gender equality are key components. By this perspective, both the youth and women might view mining activity as an attractive job and the work environment as a safe, and this could go a long way in breaking the unequal gender balance that exists in most mines today. Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1653270 Enhanced Performance for Support Vector Machines as Multiclass Classifiers in Steel Surface Defect Detection
Authors: Ehsan Amid, Sina Rezaei Aghdam, Hamidreza Amindavar
Abstract:
Steel surface defect detection is essentially one of pattern recognition problems. Support Vector Machines (SVMs) are known as one of the most proper classifiers in this application. In this paper, we introduce a more accurate classification method by using SVMs as our final classifier of the inspection system. In this scheme, multiclass classification task is performed based on the "one-againstone" method and different kernels are utilized for each pair of the classes in multiclass classification of the different defects. In the proposed system, a decision tree is employed in the first stage for two-class classification of the steel surfaces to "defect" and "non-defect", in order to decrease the time complexity. Based on the experimental results, generated from over one thousand images, the proposed multiclass classification scheme is more accurate than the conventional methods and the overall system yields a sufficient performance which can meet the requirements in steel manufacturing.Keywords: Steel Surface Defect Detection, Support Vector Machines, Kernel Methods.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1916269 Investigation of the Possibility to Prepare Supervised Classification Map of Gully Erosion by RS and GIS
Authors: Ali Mohammadi Torkashvand, Hamid Reza Alipour
Abstract:
This study investigates the possibility providing gully erosion map by the supervised classification of satellite images (ETM+) in two mountainous and plain land types. These land types were the part of Varamin plain, Tehran province, and Roodbar subbasin, Guilan province, as plain and mountain land types, respectively. The position of 652 and 124 ground control points were recorded by GPS respectively in mountain and plain land types. Soil gully erosion, land uses or plant covers were investigated in these points. Regarding ground control points and auxiliary points, training points of gully erosion and other surface features were introduced to software (Ilwis 3.3 Academic). The supervised classified map of gully erosion was prepared by maximum likelihood method and then, overall accuracy of this map was computed. Results showed that the possibility supervised classification of gully erosion isn-t possible, although it need more studies for results generalization to other mountainous regions. Also, with increasing land uses and other surface features in plain physiography, it decreases the classification of accuracy.Keywords: Supervised classification, Gully erosion, Map.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1827268 Improved Rare Species Identification Using Focal Loss Based Deep Learning Models
Authors: Chad Goldsworthy, B. Rajeswari Matam
Abstract:
The use of deep learning for species identification in camera trap images has revolutionised our ability to study, conserve and monitor species in a highly efficient and unobtrusive manner, with state-of-the-art models achieving accuracies surpassing the accuracy of manual human classification. The high imbalance of camera trap datasets, however, results in poor accuracies for minority (rare or endangered) species due to their relative insignificance to the overall model accuracy. This paper investigates the use of Focal Loss, in comparison to the traditional Cross Entropy Loss function, to improve the identification of minority species in the “255 Bird Species” dataset from Kaggle. The results show that, although Focal Loss slightly decreased the accuracy of the majority species, it was able to increase the F1-score by 0.06 and improve the identification of the bottom two, five and ten (minority) species by 37.5%, 15.7% and 10.8%, respectively, as well as resulting in an improved overall accuracy of 2.96%.
Keywords: Convolutional neural networks, data imbalance, deep learning, focal loss, species classification, wildlife conservation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1419267 Development of Mobile Application Social Guidance and Counseling for Junior High School
Authors: Suyoto, Tri Prasetyaningrum
Abstract:
At this paper, we will present the development of mobile application Social Guidance and Counseling (GC) that called “m-NingBK: Social GC”. The application is used for GC services that run on mobile devices. The application is designed specifically for Junior High School student. The methods are a combination of interactive multimedia approaches and educational psychology. Therefore, the design process is carried out three processes, which are digitizing of material social GC services, visualizing wisely and making interactive. This method is intended to make students not only hear and see but also "do" the virtual. There are five components used in multimedia applications "m-NingBK: Social GC" i.e. text, images / graphics, audio / sound, animation and video. Four menus provided by this application is the potential self, social, Expert System and about. The application is built using the Java programming language. This application was tested using a Smartphone with Android Operating System. Based on the test, people give rating: 16.7% excellent, 61.1% good, 19.4% adequate, and 2.8% poor.
Keywords: Expert Systems, Guidance and Counseling, mobile application, multimedia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2919266 Fragile Watermarking for Color Images Using Thresholding Technique
Authors: Kuo-Cheng Liu
Abstract:
In this paper, we propose ablock-wise watermarking scheme for color image authentication to resist malicious tampering of digital media. The thresholding technique is incorporated into the scheme such that the tampered region of the color image can be recovered with high quality while the proofing result is obtained. The watermark for each block consists of its dual authentication data and the corresponding feature information. The feature information for recovery iscomputed bythe thresholding technique. In the proofing process, we propose a dual-option parity check method to proof the validity of image blocks. In the recovery process, the feature information of each block embedded into the color image is rebuilt for high quality recovery. The simulation results show that the proposed watermarking scheme can effectively proof the tempered region with high detection rate and can recover the tempered region with high quality.
Keywords: thresholding technique, tamper proofing, tamper recovery
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1632265 Quantum-Like Approach for Deriving a Theory Describing the Concept of Interpretation
Authors: Yehuda Roth
Abstract:
In quantum theory, a system’s time evolution is predictable unless an observer performs measurement, as the measurement process can randomize the system. This randomness appears when the measuring device does not accurately describe the measured item, i.e., when the states characterizing the measuring device appear as a superposition of those being measured. When such a mismatch occurs, the measured data randomly collapse into a single eigenstate of the measuring device. This scenario resembles the interpretation process in which the observer does not experience an objective reality but interprets it based on preliminary descriptions initially ingrained into his/her mind. This distinction is the motivation for the present study in which the collapse scenario is regarded as part of the interpretation process of the observer. By adopting the formalism of the quantum theory, we present a complete mathematical approach that describes the interpretation process. We demonstrate this process by applying the proposed interpretation formalism to the ambiguous image "My wife and mother-in-law" to identify whether a woman in the picture is young or old.
Keywords: Interpretation, ambiguous images, data reception, state matching, classification, determination.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 188264 An Efficient Pixel Based Cervical Disc Localization
Authors: J. Preetha, S. Selvarajan
Abstract:
When neck pain is associated with pain, numbness, or weakness in the arm, shoulder, or hand, further investigation is needed as these are symptoms indicating pressure on one or more nerve roots. Evaluation necessitates a neurologic examination and imaging using an MRI/CT scan. A degenerating disc loses some thickness and is less flexible, causing inter-vertebrae space to narrow. A radiologist diagnoses an Intervertebral Disc Degeneration (IDD) by localizing every inter-vertebral disc and identifying the pathology in a disc based on its geometry and appearance. Accurate localizing is necessary to diagnose IDD pathology. But, the underlying image signal is ambiguous: a disc’s intensity overlaps the spinal nerve fibres. Even the structure changes from case to case, with possible spinal column bending (scoliosis). The inter-vertebral disc pathology’s quantitative assessment needs accurate localization of the cervical region discs. In this work, the efficacy of multilevel set segmentation model, to segment cervical discs is investigated. The segmented images are annotated using a simple distance matrix.Keywords: Intervertebral Disc Degeneration (IDD), Cervical Disc Localization, multilevel set segmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1861263 Feature Vector Fusion for Image Based Human Age Estimation
Authors: D. Karthikeyan, G. Balakrishnan
Abstract:
Human faces, as important visual signals, express a significant amount of nonverbal info for usage in human-to-human communication. Age, specifically, is more significant among these properties. Human age estimation using facial image analysis as an automated method which has numerous potential real‐world applications. In this paper, an automated age estimation framework is presented. Support Vector Regression (SVR) strategy is utilized to investigate age prediction. This paper depicts a feature extraction taking into account Gray Level Co-occurrence Matrix (GLCM), which can be utilized for robust face recognition framework. It applies GLCM operation to remove the face's features images and Active Appearance Models (AAMs) to assess the human age based on image. A fused feature technique and SVR with GA optimization are proposed to lessen the error in age estimation.
Keywords: Support vector regression, feature extraction, gray level co-occurrence matrix, active appearance models.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1314262 A Stereo Image Processing System for Visually Impaired
Authors: G. Balakrishnan, G. Sainarayanan, R. Nagarajan, Sazali Yaacob
Abstract:
This paper presents a review on vision aided systems and proposes an approach for visual rehabilitation using stereo vision technology. The proposed system utilizes stereo vision, image processing methodology and a sonification procedure to support blind navigation. The developed system includes a wearable computer, stereo cameras as vision sensor and stereo earphones, all moulded in a helmet. The image of the scene infront of visually handicapped is captured by the vision sensors. The captured images are processed to enhance the important features in the scene in front, for navigation assistance. The image processing is designed as model of human vision by identifying the obstacles and their depth information. The processed image is mapped on to musical stereo sound for the blind-s understanding of the scene infront. The developed method has been tested in the indoor and outdoor environments and the proposed image processing methodology is found to be effective for object identification.Keywords: Blind navigation, stereo vision, image processing, object preference, music tones.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4115261 An Improved Illumination Normalization based on Anisotropic Smoothing for Face Recognition
Authors: Sanghoon Kim, Sun-Tae Chung, Souhwan Jung, Seongwon Cho
Abstract:
Robust face recognition under various illumination environments is very difficult and needs to be accomplished for successful commercialization. In this paper, we propose an improved illumination normalization method for face recognition. Illumination normalization algorithm based on anisotropic smoothing is well known to be effective among illumination normalization methods but deteriorates the intensity contrast of the original image, and incurs less sharp edges. The proposed method in this paper improves the previous anisotropic smoothing-based illumination normalization method so that it increases the intensity contrast and enhances the edges while diminishing the effect of illumination variations. Due to the result of these improvements, face images preprocessed by the proposed illumination normalization method becomes to have more distinctive feature vectors (Gabor feature vectors) for face recognition. Through experiments of face recognition based on Gabor feature vector similarity, the effectiveness of the proposed illumination normalization method is verified.Keywords: Illumination Normalization, Face Recognition, Anisotropic smoothing, Gabor feature vector.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1549260 OCR for Script Identification of Hindi (Devnagari) Numerals using Feature Sub Selection by Means of End-Point with Neuro-Memetic Model
Authors: Banashree N. P., R. Vasanta
Abstract:
Recognition of Indian languages scripts is challenging problems. In Optical Character Recognition [OCR], a character or symbol to be recognized can be machine printed or handwritten characters/numerals. There are several approaches that deal with problem of recognition of numerals/character depending on the type of feature extracted and different way of extracting them. This paper proposes a recognition scheme for handwritten Hindi (devnagiri) numerals; most admired one in Indian subcontinent. Our work focused on a technique in feature extraction i.e. global based approach using end-points information, which is extracted from images of isolated numerals. These feature vectors are fed to neuro-memetic model [18] that has been trained to recognize a Hindi numeral. The archetype of system has been tested on varieties of image of numerals. . In proposed scheme data sets are fed to neuro-memetic algorithm, which identifies the rule with highest fitness value of nearly 100 % & template associates with this rule is nothing but identified numerals. Experimentation result shows that recognition rate is 92-97 % compared to other models.Keywords: OCR, Global Feature, End-Points, Neuro-Memetic model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1759259 Create and Design Visual Presentation to Promote Thai Cuisine
Authors: Supaporn Wimonchailerk
Abstract:
This research aims to study how to design and create the media to promote Thai cuisine. The study used qualitative research methods by using in-depth interview 3 key informants who have experienced in the production of food or cooking shows in television programs with an aspect of acknowledging Thai foods. The results showed that visual presentation is divided into four categories. First, the light meals should be presented in details via the close-up camera with lighting to make the food look more delicious. Then the curry presentation should be arranged a clear and crisp light focus on a colorful curry paste. Besides the vision of hot steam floating from the plate and a view of curry spread on steamed rice can call great attentions. Third, delivering good appearances of the fried or spicy foods, the images must allow the audiences to see the shine of the coat covering the texture of the food and the colorful of the ingredients. Fourth, the presentation of sweets is recommended to focus on details of food design, composition, and layout.
Keywords: Media production, television, promote, Thai cuisine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 963258 Mechanical Characterization of Mango Peel Flour and Biopolypropylene Composites Compatibilized with PP-g-IA
Authors: J. Gomez-Caturla, L. Quiles-Carrillo, J. Ivorra-Martinez, D. Garcia-Garcia, R. Balart
Abstract:
The present work reports on the development of wood plastic composites based on biopolypropylene (BioPP) and mango peel flour (MPF) by extrusion and injection molding processes. PP-g-IA and dicumyl peroxide (DCP) have been used as a compatibilizer and as a free radical initiator for reactive extrusion, respectively. Mechanical and morphological properties have been characterized in order to study the compatibility of the blends. The obtained results showed that DCP and PP-g-IA improved the stiffness of BioPP in terms of elastic modulus. Moreover, they positively increased the tensile strength and elongation at break of the blends in comparison with the sample that only had BioPP and MPF on its composition, improving the affinity between both compounds. DCP and PP-g-IA even seem to have certain synergy, which was corroborated through Field Emission Scanning Electron Microscopy (FESEM) analysis. Images showed that the MPF particles had greater adhesion to the polymer matrix when PP-g-IA and DCP were added. This effect was more intense when both elements were added, observing an almost inexistent gap between MPF particles and the BioPP matrix.
Keywords: Biopolypropylene, compatibilization, mango peel flour, wood plastic composite.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 396257 Synthesis, Characterization and Coating of the Zinc Oxide Nanoparticles on Cotton Fabric by Mechanical Thermo-Fixation Techniques to Impart Antimicrobial Activity
Authors: Imana Shahrin Tania, Mohammad Ali
Abstract:
The present study reports the synthesis, characterization and application of nano-sized zinc-oxide (ZnO) particles on a cotton fabric surface. The aim of the investigations is to impart the antimicrobial activity on textile cloth. Nanoparticle is synthesized by wet chemical method from zinc sulphate and sodium hydroxide. SEM (scanning electron micrograph) images are taken to demonstrate the surface morphology of nanoparticles. XRD analysis is done to determine the crystal size of the nanoparticle. With the conformation of nanoformation, the cotton woven fabric is treated with ZnO nanoparticle by mechanical thermo-fixation (pad-dry-cure) technique. To increase the wash durability of nano treated fabric, an acrylic binder is used as a fixing agent. The treated fabric shows up to 90% bacterial reduction for S. aureus (Staphylococcus aureus) and 87% for E. coli (Escherichia coli) which is appreciable for bacteria protective clothing.Keywords: Nanoparticle, zinc oxide, cotton fabric, antibacterial activity, binder.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 585256 Towards Integrating Statistical Color Features for Human Skin Detection
Authors: Mohd Zamri Osman, Mohd Aizaini Maarof, Mohd Foad Rohani
Abstract:
Human skin detection recognized as the primary step in most of the applications such as face detection, illicit image filtering, hand recognition and video surveillance. The performance of any skin detection applications greatly relies on the two components: feature extraction and classification method. Skin color is the most vital information used for skin detection purpose. However, color feature alone sometimes could not handle images with having same color distribution with skin color. A color feature of pixel-based does not eliminate the skin-like color due to the intensity of skin and skin-like color fall under the same distribution. Hence, the statistical color analysis will be exploited such mean and standard deviation as an additional feature to increase the reliability of skin detector. In this paper, we studied the effectiveness of statistical color feature for human skin detection. Furthermore, the paper analyzed the integrated color and texture using eight classifiers with three color spaces of RGB, YCbCr, and HSV. The experimental results show that the integrating statistical feature using Random Forest classifier achieved a significant performance with an F1-score 0.969.
Keywords: Color space, neural network, random forest, skin detection, statistical feature.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1953255 Feature Extraction of Dorsal Hand Vein Pattern Using a Fast Modified PCA Algorithm Based On Cholesky Decomposition and Lanczos Technique
Authors: Maleika Heenaye- Mamode Khan , Naushad Mamode Khan, Raja K.Subramanian
Abstract:
Dorsal hand vein pattern is an emerging biometric which is attracting the attention of researchers, of late. Research is being carried out on existing techniques in the hope of improving them or finding more efficient ones. In this work, Principle Component Analysis (PCA) , which is a successful method, originally applied on face biometric is being modified using Cholesky decomposition and Lanczos algorithm to extract the dorsal hand vein features. This modified technique decreases the number of computation and hence decreases the processing time. The eigenveins were successfully computed and projected onto the vein space. The system was tested on a database of 200 images and using a threshold value of 0.9 to obtain the False Acceptance Rate (FAR) and False Rejection Rate (FRR). This modified algorithm is desirable when developing biometric security system since it significantly decreases the matching time.
Keywords: Dorsal hand vein pattern, PCA, Cholesky decomposition, Lanczos algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1837254 Sparsity-Based Unsupervised Unmixing of Hyperspectral Imaging Data Using Basis Pursuit
Authors: Ahmed Elrewainy
Abstract:
Mixing in the hyperspectral imaging occurs due to the low spatial resolutions of the used cameras. The existing pure materials “endmembers” in the scene share the spectra pixels with different amounts called “abundances”. Unmixing of the data cube is an important task to know the present endmembers in the cube for the analysis of these images. Unsupervised unmixing is done with no information about the given data cube. Sparsity is one of the recent approaches used in the source recovery or unmixing techniques. The l1-norm optimization problem “basis pursuit” could be used as a sparsity-based approach to solve this unmixing problem where the endmembers is assumed to be sparse in an appropriate domain known as dictionary. This optimization problem is solved using proximal method “iterative thresholding”. The l1-norm basis pursuit optimization problem as a sparsity-based unmixing technique was used to unmix real and synthetic hyperspectral data cubes.
Keywords: Basis pursuit, blind source separation, hyperspectral imaging, spectral unmixing, wavelets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 837253 Mining Association Rules from Unstructured Documents
Authors: Hany Mahgoub
Abstract:
This paper presents a system for discovering association rules from collections of unstructured documents called EART (Extract Association Rules from Text). The EART system treats texts only not images or figures. EART discovers association rules amongst keywords labeling the collection of textual documents. The main characteristic of EART is that the system integrates XML technology (to transform unstructured documents into structured documents) with Information Retrieval scheme (TF-IDF) and Data Mining technique for association rules extraction. EART depends on word feature to extract association rules. It consists of four phases: structure phase, index phase, text mining phase and visualization phase. Our work depends on the analysis of the keywords in the extracted association rules through the co-occurrence of the keywords in one sentence in the original text and the existing of the keywords in one sentence without co-occurrence. Experiments applied on a collection of scientific documents selected from MEDLINE that are related to the outbreak of H5N1 avian influenza virus.Keywords: Association rules, information retrieval, knowledgediscovery in text, text mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2442252 Object Detection Based on Plane Segmentation and Features Matching for a Service Robot
Authors: António J. R. Neves, Rui Garcia, Paulo Dias, Alina Trifan
Abstract:
With the aging of the world population and the continuous growth in technology, service robots are more and more explored nowadays as alternatives to healthcare givers or personal assistants for the elderly or disabled people. Any service robot should be capable of interacting with the human companion, receive commands, navigate through the environment, either known or unknown, and recognize objects. This paper proposes an approach for object recognition based on the use of depth information and color images for a service robot. We present a study on two of the most used methods for object detection, where 3D data is used to detect the position of objects to classify that are found on horizontal surfaces. Since most of the objects of interest accessible for service robots are on these surfaces, the proposed 3D segmentation reduces the processing time and simplifies the scene for object recognition. The first approach for object recognition is based on color histograms, while the second is based on the use of the SIFT and SURF feature descriptors. We present comparative experimental results obtained with a real service robot.Keywords: Service Robot, Object Recognition, 3D Sensors, Plane Segmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1674251 Scatterer Density in Nonlinear Diffusion for Speckle Reduction in Ultrasound Imaging: The Isotropic Case
Authors: Ahmed Badawi
Abstract:
This paper proposes a method for speckle reduction in medical ultrasound imaging while preserving the edges with the added advantages of adaptive noise filtering and speed. A nonlinear image diffusion method that incorporates local image parameter, namely, scatterer density in addition to gradient, to weight the nonlinear diffusion process, is proposed. The method was tested for the isotropic case with a contrast detail phantom and varieties of clinical ultrasound images, and then compared to linear and some other diffusion enhancement methods. Different diffusion parameters were tested and tuned to best reduce speckle noise and preserve edges. The method showed superior performance measured both quantitatively and qualitatively when incorporating scatterer density into the diffusivity function. The proposed filter can be used as a preprocessing step for ultrasound image enhancement before applying automatic segmentation, automatic volumetric calculations, or 3D ultrasound volume rendering.Keywords: Ultrasound imaging, Nonlinear isotropic diffusion, Speckle noise, Scattering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1950