Search results for: image dictionary creation
3933 Optimization Based Extreme Learning Machine for Watermarking of an Image in DWT Domain
Authors: RAM PAL SINGH, VIKASH CHAUDHARY, MONIKA VERMA
Abstract:
In this paper, we proposed the implementation of optimization based Extreme Learning Machine (ELM) for watermarking of B-channel of color image in discrete wavelet transform (DWT) domain. ELM, a regularization algorithm, works based on generalized single-hidden-layer feed-forward neural networks (SLFNs). However, hidden layer parameters, generally called feature mapping in context of ELM need not to be tuned every time. This paper shows the embedding and extraction processes of watermark with the help of ELM and results are compared with already used machine learning models for watermarking.Here, a cover image is divide into suitable numbers of non-overlapping blocks of required size and DWT is applied to each block to be transformed in low frequency sub-band domain. Basically, ELM gives a unified leaning platform with a feature mapping, that is, mapping between hidden layer and output layer of SLFNs, is tried for watermark embedding and extraction purpose in a cover image. Although ELM has widespread application right from binary classification, multiclass classification to regression and function estimation etc. Unlike SVM based algorithm which achieve suboptimal solution with high computational complexity, ELM can provide better generalization performance results with very small complexity. Efficacy of optimization method based ELM algorithm is measured by using quantitative and qualitative parameters on a watermarked image even though image is subjected to different types of geometrical and conventional attacks.Keywords: BER, DWT, extreme leaning machine (ELM), PSNR
Procedia PDF Downloads 3113932 Engineering of E-Learning Content Creation: Case Study for African Countries
Authors: María-Dolores Afonso-Suárez, Nayra Pumar-Carreras, Juan Ruiz-Alzola
Abstract:
This research addresses the use of an e-Learning creation methodology for learning objects. Throughout the process, indicators are being gathered, to determine if it responds to the main objectives of an engineering discipline. These parameters will also indicate if it is necessary to review the creation cycle and readjust any phase. Within the project developed for this study, apart from the use of structured methods, there has been a central objective: the establishment of a learning atmosphere. A place where all the professionals involved are able to collaborate, plan, solve problems and determine guides to follow in order to develop creative and innovative solutions. It has been outlined as a blended learning program with an assessment plan that proposes face to face lessons, coaching, collaboration, multimedia and web based learning objects as well as support resources. The project has been drawn as a long term task, the pilot teaching actions designed provide the preliminary results object of study. This methodology is been used in the creation of learning content for the African countries of Senegal, Mauritania and Cape Verde. It has been developed within the framework of the MACbioIDi, an Interreg European project for the International cooperation and development. The educational area of this project is focused in the training and advice of professionals of the medicine as well as engineers in the use of applications of medical imaging technology, specifically the 3DSlicer application and the Open Anatomy Browser.Keywords: teaching contents engineering, e-learning, blended learning, international cooperation, 3dslicer, open anatomy browser
Procedia PDF Downloads 1723931 Wealth Creation and its Externalities: Evaluating Economic Growth and Corporate Social Responsibility
Authors: Zhikang Rong
Abstract:
The 4th industrial revolution has introduced technologies like interconnectivity, machine learning, and real-time big data analytics that improve operations and business efficiency. This paper examines how these advancements have led to a concentration of wealth, specifically among the top 1%, and investigates whether this wealth provides value to society. Through analyzing impacts on employment, productivity, supply-demand dynamics, and potential externalities, it is shown that successful businesspeople, by enhancing productivity and creating jobs, contribute positively to long-term economic growth. Additionally, externalities such as environmental degradation are managed by social entrepreneurship and government policies.Keywords: wealth creation, employment, productivity, social entrepreneurship
Procedia PDF Downloads 283930 Intergenerational Succession within Family Businesses: The Role of Sharing and Creation Knowledge
Authors: Wissal Ben Arfi, Jean-Michel Sahut
Abstract:
The purpose of this paper is to provide a deeper understanding of the succession process from a knowledge management perspective. By doing that, succession process in family businesses, as an environment for creating and sharing knowledge, was explored. Design/Methodology/Approach: To support our reasoning, we collected qualitative data through 16 in-depth interviews conducted with all decision makers involved in the family businesses succession process in France. These open-ended responses were subsequently exposed to thematic discourse analysis. Findings: Central to this exhibit is the nature and magnitude of knowledge creation and sharing among the actors within the family succession context and how can tacit knowledge sharing facilitate the succession process. We also identified factors that inhibit down the knowledge creation and sharing processes. The sharing and creation of knowledge among members of a family business appear to be a complex process that must be part of a strategy for change. This implies that it requests trust and takes a certain amount of time because it requires organizational change and a clear and coherent strategic vision that is accepted and assimilated by all the members. Professional and leadership skills are of particular importance in knowledge sharing and creation processes. In most cases, tacit knowledge is crucial when it is shared and accumulated collectively. Our findings reveal that managers should find ways of implementing knowledge sharing and creation processes while acknowledging the succession process within family firms. This study highlights the importance of generating knowledge strategies in order to enhance the performance and the success of intergenerational succession. The empirical outcomes contribute to enrich the field of succession management process and enhance the role of knowledge in shaping family performance and longevity. To a large extent, the lessons learned from the study of succession processes in family-owned businesses are that when there is a deliberate effort to introduce a knowledge-based approach, this action becomes a seminal event in the life of the organization. Originality/Value: The paper contributes to the deep understanding of interactions among actors by examining the knowledge creation and sharing processes since current researches in family succession focused on aspects such as personal development of potential, intra-family succession intention, decision-making processes in family businesses. Besides, as succession is one of the key factors that determine the longevity and the performance of family businesses, it also contributes to literature by examining how tacit knowledge is transferred, shared and created in family businesses and how this can facilitate the intergenerational succession process.Keywords: family-owned businesses, succession process, knowledge, performance
Procedia PDF Downloads 2083929 Classification of Hyperspectral Image Using Mathematical Morphological Operator-Based Distance Metric
Authors: Geetika Barman, B. S. Daya Sagar
Abstract:
In this article, we proposed a pixel-wise classification of hyperspectral images using a mathematical morphology operator-based distance metric called “dilation distance” and “erosion distance”. This method involves measuring the spatial distance between the spectral features of a hyperspectral image across the bands. The key concept of the proposed approach is that the “dilation distance” is the maximum distance a pixel can be moved without changing its classification, whereas the “erosion distance” is the maximum distance that a pixel can be moved before changing its classification. The spectral signature of the hyperspectral image carries unique class information and shape for each class. This article demonstrates how easily the dilation and erosion distance can measure spatial distance compared to other approaches. This property is used to calculate the spatial distance between hyperspectral image feature vectors across the bands. The dissimilarity matrix is then constructed using both measures extracted from the feature spaces. The measured distance metric is used to distinguish between the spectral features of various classes and precisely distinguish between each class. This is illustrated using both toy data and real datasets. Furthermore, we investigated the role of flat vs. non-flat structuring elements in capturing the spatial features of each class in the hyperspectral image. In order to validate, we compared the proposed approach to other existing methods and demonstrated empirically that mathematical operator-based distance metric classification provided competitive results and outperformed some of them.Keywords: dilation distance, erosion distance, hyperspectral image classification, mathematical morphology
Procedia PDF Downloads 863928 Multiple Images Stitching Based on Gradually Changing Matrix
Authors: Shangdong Zhu, Yunzhou Zhang, Jie Zhang, Hang Hu, Yazhou Zhang
Abstract:
Image stitching is a very important branch in the field of computer vision, especially for panoramic map. In order to eliminate shape distortion, a novel stitching method is proposed based on gradually changing matrix when images are horizontal. For images captured horizontally, this paper assumes that there is only translational operation in image stitching. By analyzing each parameter of the homography matrix, the global homography matrix is gradually transferred to translation matrix so as to eliminate the effects of scaling, rotation, etc. in the image transformation. This paper adopts matrix approximation to get the minimum value of the energy function so that the shape distortion at those regions corresponding to the homography can be minimized. The proposed method can avoid multiple horizontal images stitching failure caused by accumulated shape distortion. At the same time, it can be combined with As-Projective-As-Possible algorithm to ensure precise alignment of overlapping area.Keywords: image stitching, gradually changing matrix, horizontal direction, matrix approximation, homography matrix
Procedia PDF Downloads 3193927 Algorithm for Path Recognition in-between Tree Rows for Agricultural Wheeled-Mobile Robots
Authors: Anderson Rocha, Pedro Miguel de Figueiredo Dinis Oliveira Gaspar
Abstract:
Machine vision has been widely used in recent years in agriculture, as a tool to promote the automation of processes and increase the levels of productivity. The aim of this work is the development of a path recognition algorithm based on image processing to guide a terrestrial robot in-between tree rows. The proposed algorithm was developed using the software MATLAB, and it uses several image processing operations, such as threshold detection, morphological erosion, histogram equalization and the Hough transform, to find edge lines along tree rows on an image and to create a path to be followed by a mobile robot. To develop the algorithm, a set of images of different types of orchards was used, which made possible the construction of a method capable of identifying paths between trees of different heights and aspects. The algorithm was evaluated using several images with different characteristics of quality and the results showed that the proposed method can successfully detect a path in different types of environments.Keywords: agricultural mobile robot, image processing, path recognition, hough transform
Procedia PDF Downloads 1463926 Automating Test Activities: Test Cases Creation, Test Execution, and Test Reporting with Multiple Test Automation Tools
Authors: Loke Mun Sei
Abstract:
Software testing has become a mandatory process in assuring the software product quality. Hence, test management is needed in order to manage the test activities conducted in the software test life cycle. This paper discusses on the challenges faced in the software test life cycle, and how the test processes and test activities, mainly on test cases creation, test execution, and test reporting is being managed and automated using several test automation tools, i.e. Jira, Robot Framework, and Jenkins.Keywords: test automation tools, test case, test execution, test reporting
Procedia PDF Downloads 5833925 Video Stabilization Using Feature Point Matching
Authors: Shamsundar Kulkarni
Abstract:
Video capturing by non-professionals will lead to unanticipated effects. Such as image distortion, image blurring etc. Hence, many researchers study such drawbacks to enhance the quality of videos. In this paper, an algorithm is proposed to stabilize jittery videos .A stable output video will be attained without the effect of jitter which is caused due to shaking of handheld camera during video recording. Firstly, salient points from each frame from the input video are identified and processed followed by optimizing and stabilize the video. Optimization includes the quality of the video stabilization. This method has shown good result in terms of stabilization and it discarded distortion from the output videos recorded in different circumstances.Keywords: video stabilization, point feature matching, salient points, image quality measurement
Procedia PDF Downloads 3133924 Experimental Characterization of Composite Material with Non Contacting Methods
Authors: Nikolaos Papadakis, Constantinos Condaxakis, Konstantinos Savvakis
Abstract:
The aim of this paper is to determine the elastic properties (elastic modulus and Poisson ratio) of a composite material based on noncontacting imaging methods. More specifically, the significantly reduced cost of digital cameras has given the opportunity of the high reliability of low-cost strain measurement. The open source platform Ncorr is used in this paper which utilizes the method of digital image correlation (DIC). The use of digital image correlation in measuring strain uses random speckle preparation on the surface of the gauge area, image acquisition, and postprocessing the image correlation to obtain displacement and strain field on surface under study. This study discusses technical issues relating to the quality of results to be obtained are discussed. [0]8 fabric glass/epoxy composites specimens were prepared and tested at different orientations 0[o], 30[o], 45[o], 60[o], 90[o]. Each test was recorded with the camera at a constant frame rate and constant lighting conditions. The recorded images were processed through the use of the image processing software. The parameters of the test are reported. The strain map output which is obtained through strain measurement using Ncorr is validated by a) comparing the elastic properties with expected values from Classical laminate theory, b) through finite element analysis.Keywords: composites, Ncorr, strain map, videoextensometry
Procedia PDF Downloads 1443923 Accountability Mechanisms of Leaders and Its Impact on Performance and Value Creation: Comparative Analysis (France, Germany, United Kingdom)
Authors: Bahram Soltani, Louai Ghazieh
Abstract:
The responsibility has a big importance further to the financial crisis and the various pressures, which companies face their duties. The main objective of this study is to explain the variation of mechanisms of the responsibility of the manager in the company among the advanced capitalist economies. Then we study the impact of these mechanisms on the performance and the value creation in European companies. To reach our goal, we established a final sample composed on average of 284 French, British and German companies quoted in stock exchanges with 2272 annual reports examined during the period from 2005 to 2012. We examined at first the link of causalities between the determining-mechanisms bound to the company such as the characteristics of the board of directors, the composition of the shareholding and the ethics of the company on one side and the profitability of the company on the other side. The results show that the smooth running of the board of directors and its specialist committees are very important determinants of the responsibility of the managers who impact positively the performance and the value creation in the company. Furthermore, our results confirm that the presence of a solid ethical environment within the company will be effective to increase the probability that the managers realize ethical choices in the organizational decision-making. At the second time, we studied the impact of the determining mechanisms bound to the function and to the profile of manager to know its relational links, his remuneration, his training, his age and his experiences about the performance and the value creation in the company. Our results highlight the existence of a negative relation between the relational links of the manager, his very high remuneration and the general profitability of the company. This study is a contribution to the literature on the determining mechanisms of company director's responsibility (Accountability). It establishes an empirical and comparative analysis between three influential countries of Europe, to know France, the United Kingdom and Germany.Keywords: leaders, company’s performance, accountability mechanisms, corporate governance, value creation of firm, financial crisis
Procedia PDF Downloads 3783922 Large Neural Networks Learning From Scratch With Very Few Data and Without Explicit Regularization
Authors: Christoph Linse, Thomas Martinetz
Abstract:
Recent findings have shown that Neural Networks generalize also in over-parametrized regimes with zero training error. This is surprising, since it is completely against traditional machine learning wisdom. In our empirical study we fortify these findings in the domain of fine-grained image classification. We show that very large Convolutional Neural Networks with millions of weights do learn with only a handful of training samples and without image augmentation, explicit regularization or pretraining. We train the architectures ResNet018, ResNet101 and VGG19 on subsets of the difficult benchmark datasets Caltech101, CUB_200_2011, FGVCAircraft, Flowers102 and StanfordCars with 100 classes and more, perform a comprehensive comparative study and draw implications for the practical application of CNNs. Finally, we show that VGG19 with 140 million weights learns to distinguish airplanes and motorbikes with up to 95% accuracy using only 20 training samples per class.Keywords: convolutional neural networks, fine-grained image classification, generalization, image recognition, over-parameterized, small data sets
Procedia PDF Downloads 883921 Computer Countenanced Diagnosis of Skin Nodule Detection and Histogram Augmentation: Extracting System for Skin Cancer
Authors: S. Zith Dey Babu, S. Kour, S. Verma, C. Verma, V. Pathania, A. Agrawal, V. Chaudhary, A. Manoj Puthur, R. Goyal, A. Pal, T. Danti Dey, A. Kumar, K. Wadhwa, O. Ved
Abstract:
Background: Skin cancer is now is the buzzing button in the field of medical science. The cyst's pandemic is drastically calibrating the body and well-being of the global village. Methods: The extracted image of the skin tumor cannot be used in one way for diagnosis. The stored image contains anarchies like the center. This approach will locate the forepart of an extracted appearance of skin. Partitioning image models has been presented to sort out the disturbance in the picture. Results: After completing partitioning, feature extraction has been formed by using genetic algorithm and finally, classification can be performed between the trained and test data to evaluate a large scale of an image that helps the doctors for the right prediction. To bring the improvisation of the existing system, we have set our objectives with an analysis. The efficiency of the natural selection process and the enriching histogram is essential in that respect. To reduce the false-positive rate or output, GA is performed with its accuracy. Conclusions: The objective of this task is to bring improvisation of effectiveness. GA is accomplishing its task with perfection to bring down the invalid-positive rate or outcome. The paper's mergeable portion conflicts with the composition of deep learning and medical image processing, which provides superior accuracy. Proportional types of handling create the reusability without any errors.Keywords: computer-aided system, detection, image segmentation, morphology
Procedia PDF Downloads 1503920 Prosperous Digital Image Watermarking Approach by Using DCT-DWT
Authors: Prabhakar C. Dhavale, Meenakshi M. Pawar
Abstract:
In this paper, everyday tons of data is embedded on digital media or distributed over the internet. The data is so distributed that it can easily be replicated without error, putting the rights of their owners at risk. Even when encrypted for distribution, data can easily be decrypted and copied. One way to discourage illegal duplication is to insert information known as watermark, into potentially valuable data in such a way that it is impossible to separate the watermark from the data. These challenges motivated researchers to carry out intense research in the field of watermarking. A watermark is a form, image or text that is impressed onto paper, which provides evidence of its authenticity. Digital watermarking is an extension of the same concept. There are two types of watermarks visible watermark and invisible watermark. In this project, we have concentrated on implementing watermark in image. The main consideration for any watermarking scheme is its robustness to various attacksKeywords: watermarking, digital, DCT-DWT, security
Procedia PDF Downloads 4223919 PET Image Resolution Enhancement
Authors: Krzysztof Malczewski
Abstract:
PET is widely applied scanning procedure in medical imaging based research. It delivers measurements of functioning in distinct areas of the human brain while the patient is comfortable, conscious and alert. This article presents the new compression sensing based super-resolution algorithm for improving the image resolution in clinical Positron Emission Tomography (PET) scanners. The issue of motion artifacts is well known in Positron Emission Tomography (PET) studies as its side effect. The PET images are being acquired over a limited period of time. As the patients cannot hold breath during the PET data gathering, spatial blurring and motion artefacts are the usual result. These may lead to wrong diagnosis. It is shown that the presented approach improves PET spatial resolution in cases when Compressed Sensing (CS) sequences are used. Compressed Sensing (CS) aims at signal and images reconstructing from significantly fewer measurements than were traditionally thought necessary. The application of CS to PET has the potential for significant scan time reductions, with visible benefits for patients and health care economics. In this study the goal is to combine super-resolution image enhancement algorithm with CS framework to achieve high resolution PET output. Both methods emphasize on maximizing image sparsity on known sparse transform domain and minimizing fidelity.Keywords: PET, super-resolution, image reconstruction, pattern recognition
Procedia PDF Downloads 3713918 Image Segmentation Using Active Contours Based on Anisotropic Diffusion
Authors: Shafiullah Soomro
Abstract:
Active contour is one of the image segmentation techniques and its goal is to capture required object boundaries within an image. In this paper, we propose a novel image segmentation method by using an active contour method based on anisotropic diffusion feature enhancement technique. The traditional active contour methods use only pixel information to perform segmentation, which produces inaccurate results when an image has some noise or complex background. We use Perona and Malik diffusion scheme for feature enhancement, which sharpens the object boundaries and blurs the background variations. Our main contribution is the formulation of a new SPF (signed pressure force) function, which uses global intensity information across the regions. By minimizing an energy function using partial differential framework the proposed method captures semantically meaningful boundaries instead of catching uninterested regions. Finally, we use a Gaussian kernel which eliminates the problem of reinitialization in level set function. We use several synthetic and real images from different modalities to validate the performance of the proposed method. In the experimental section, we have found the proposed method performance is better qualitatively and quantitatively and yield results with higher accuracy compared to other state-of-the-art methods.Keywords: active contours, anisotropic diffusion, level-set, partial differential equations
Procedia PDF Downloads 1603917 Mutual Information Based Image Registration of Satellite Images Using PSO-GA Hybrid Algorithm
Authors: Dipti Patra, Guguloth Uma, Smita Pradhan
Abstract:
Registration is a fundamental task in image processing. It is used to transform different sets of data into one coordinate system, where data are acquired from different times, different viewing angles, and/or different sensors. The registration geometrically aligns two images (the reference and target images). Registration techniques are used in satellite images and it is important in order to be able to compare or integrate the data obtained from these different measurements. In this work, mutual information is considered as a similarity metric for registration of satellite images. The transformation is assumed to be a rigid transformation. An attempt has been made here to optimize the transformation function. The proposed image registration technique hybrid PSO-GA incorporates the notion of Particle Swarm Optimization and Genetic Algorithm and is used for finding the best optimum values of transformation parameters. The performance comparision obtained with the experiments on satellite images found that the proposed hybrid PSO-GA algorithm outperforms the other algorithms in terms of mutual information and registration accuracy.Keywords: image registration, genetic algorithm, particle swarm optimization, hybrid PSO-GA algorithm and mutual information
Procedia PDF Downloads 4073916 Robust Medical Image Watermarking Using Frequency Domain and Least Significant Bits Algorithms
Authors: Volkan Kaya, Ersin Elbasi
Abstract:
Watermarking and stenography are getting importance recently because of copyright protection and authentication. In watermarking we embed stamp, logo, noise or image to multimedia elements such as image, video, audio, animation and text. There are several works have been done in watermarking for different purposes. In this research work, we used watermarking techniques to embed patient information into the medical magnetic resonance (MR) images. There are two methods have been used; frequency domain (Digital Wavelet Transform-DWT, Digital Cosine Transform-DCT, and Digital Fourier Transform-DFT) and spatial domain (Least Significant Bits-LSB) domain. Experimental results show that embedding in frequency domains resist against one type of attacks, and embedding in spatial domain is resist against another group of attacks. Peak Signal Noise Ratio (PSNR) and Similarity Ratio (SR) values are two measurement values for testing. These two values give very promising result for information hiding in medical MR images.Keywords: watermarking, medical image, frequency domain, least significant bits, security
Procedia PDF Downloads 2873915 Film Therapy on Adolescent Body Image: A Pilot Study
Authors: Sonia David, Uma Warrier
Abstract:
Background: Film therapy is the use of commercial or non-commercial films to enhance healing for therapeutic purposes. Objectives: The mixed-method study aims to evaluate the effect of film-based counseling on body image dissatisfaction among adolescents to precisely ascertain the cause of the alteration in body image dissatisfaction due to the said intervention. Method: The one group pre-test post-test research design study using inferential statistics and thematic analysis is based on a pre-test post-test design conducted on 44 school-going adolescents between 13 and 17. The Body Shape Questionnaire (BSQ- 34) was used as a pre-test and post-test measure. The film-based counseling intervention model was used through individual counseling sessions. The analysis involved paired sample t-test used to examine the data quantitatively, and thematic analysis was used to evaluate qualitative data. Findings: The results indicated that there is a significant difference between the pre-test and post-test means. Since t(44)= 9.042 is significant at a 99% confidence level, it is ascertained that film-based counseling intervention reduces body image dissatisfaction. The five distinct themes from the thematic analysis are “acceptance, awareness, empowered to change, empathy, and reflective.” Novelty: The paper originally contributes to the repertoire of research on film therapy as a successful counseling intervention for addressing the challenges of body image dissatisfaction. This study also opens avenues for considering alteration of teaching pedagogy to include video-based learning in various subjects.Keywords: body image dissatisfaction, adolescents, film-based counselling, film therapy, acceptance and commitment therapy
Procedia PDF Downloads 2943914 Experience Modularization for New Value of Evanescent Cultural Communities: Developing Creative Tourism Services in Bangkok
Authors: Wuttigrai Ngamsirijit
Abstract:
Creative tourism is an ongoing development in many countries as an attempt to moving away from serial reproduction of culture and reviving the culture. Despite, in the destinations with diverse and potential cultural resources, creating new tourism services can be vague. This paper presents how tourism experiences are modularized and consolidated in order to form new creative tourism service offerings in evanescent cultural communities of Bangkok, Thailand. The benefits from data mining in accommodating value co-creation are discussed, and implication of experience modularization to national creative tourism policy is addressed.Keywords: co-creation, creative tourism, new service design, experience modularization
Procedia PDF Downloads 3663913 Implementation of Achterbahn-128 for Images Encryption and Decryption
Authors: Aissa Belmeguenai, Khaled Mansouri
Abstract:
In this work, an efficient implementation of Achterbahn-128 for images encryption and decryption was introduced. The implementation for this simulated project is written by MATLAB.7.5. At first two different original images are used for validate the proposed design. Then our developed program was used to transform the original images data into image digits file. Finally, we used our implemented program to encrypt and decrypt images data. Several tests are done for proving the design performance including visual tests and security analysis; we discuss the security analysis of the proposed image encryption scheme including some important ones like key sensitivity analysis, key space analysis, and statistical attacks.Keywords: Achterbahn-128, stream cipher, image encryption, security analysis
Procedia PDF Downloads 5323912 A Study on Real-Time Fluorescence-Photoacoustic Imaging System for Mouse Thrombosis Monitoring
Authors: Sang Hun Park, Moung Young Lee, Su Min Yu, Hyun Sang Jo, Ji Hyeon Kim, Chul Gyu Song
Abstract:
A near-infrared light source used as a light source in the fluorescence imaging system is suitable for use in real-time during the operation since it has no interference in surgical vision. However, fluorescence images do not have depth information. In this paper, we configured the device with the research on molecular imaging systems for monitoring thrombus imaging using fluorescence and photoacoustic. Fluorescence imaging was performed using a phantom experiment in order to search the exact location, and the Photoacoustic image was in order to detect the depth. Fluorescence image obtained when evaluated through current phantom experiments when the concentration of the contrast agent is 25μg / ml, it was confirmed that it looked sharper. The phantom experiment is has shown the possibility with the fluorescence image and photoacoustic image using an indocyanine green contrast agent. For early diagnosis of cardiovascular diseases, more active research with the fusion of different molecular imaging devices is required.Keywords: fluorescence, photoacoustic, indocyanine green, carotid artery
Procedia PDF Downloads 6013911 Reactive and Concurrency-Based Image Resource Management Module for iOS Applications
Authors: Shubham V. Kamdi
Abstract:
This paper aims to serve as an introduction to image resource caching techniques for iOS mobile applications. It will explain how developers can break down multiple image-downloading tasks concurrently using state-of-the-art iOS frameworks, namely Swift Concurrency and Combine. The paper will explain how developers can leverage SwiftUI to develop reactive view components and use declarative coding patterns. Developers will learn to bypass built-in image caching systems by curating the procedure to implement a swift-based LRU cache system. The paper will provide a full architectural overview of a system, helping readers understand how mobile applications are designed professionally. It will cover technical discussion, helping readers understand the low-level details of threads and how they can switch between them, as well as the significance of the main and background threads for requesting HTTP services via mobile applications.Keywords: main thread, background thread, reactive view components, declarative coding
Procedia PDF Downloads 253910 A Nonlocal Means Algorithm for Poisson Denoising Based on Information Geometry
Authors: Dongxu Chen, Yipeng Li
Abstract:
This paper presents an information geometry NonlocalMeans(NLM) algorithm for Poisson denoising. NLM estimates a noise-free pixel as a weighted average of image pixels, where each pixel is weighted according to the similarity between image patches in Euclidean space. In this work, every pixel is a Poisson distribution locally estimated by Maximum Likelihood (ML), all distributions consist of a statistical manifold. A NLM denoising algorithm is conducted on the statistical manifold where Fisher information matrix can be used for computing distribution geodesics referenced as the similarity between patches. This approach was demonstrated to be competitive with related state-of-the-art methods.Keywords: image denoising, Poisson noise, information geometry, nonlocal-means
Procedia PDF Downloads 2853909 A Study of Non-Coplanar Imaging Technique in INER Prototype Tomosynthesis System
Authors: Chia-Yu Lin, Yu-Hsiang Shen, Cing-Ciao Ke, Chia-Hao Chang, Fan-Pin Tseng, Yu-Ching Ni, Sheng-Pin Tseng
Abstract:
Tomosynthesis is an imaging system that generates a 3D image by scanning in a limited angular range. It could provide more depth information than traditional 2D X-ray single projection. Radiation dose in tomosynthesis is less than computed tomography (CT). Because of limited angular range scanning, there are many properties depending on scanning direction. Therefore, non-coplanar imaging technique was developed to improve image quality in traditional tomosynthesis. The purpose of this study was to establish the non-coplanar imaging technique of tomosynthesis system and evaluate this technique by the reconstructed image. INER prototype tomosynthesis system contains an X-ray tube, a flat panel detector, and a motion machine. This system could move X-ray tube in multiple directions during the acquisition. In this study, we investigated three different imaging techniques that were 2D X-ray single projection, traditional tomosynthesis, and non-coplanar tomosynthesis. An anthropopathic chest phantom was used to evaluate the image quality. It contained three different size lesions (3 mm, 5 mm and, 8 mm diameter). The traditional tomosynthesis acquired 61 projections over a 30 degrees angular range in one scanning direction. The non-coplanar tomosynthesis acquired 62 projections over 30 degrees angular range in two scanning directions. A 3D image was reconstructed by iterative image reconstruction algorithm (ML-EM). Our qualitative method was to evaluate artifacts in tomosynthesis reconstructed image. The quantitative method was used to calculate a peak-to-valley ratio (PVR) that means the intensity ratio of the lesion to the background. We used PVRs to evaluate the contrast of lesions. The qualitative results showed that in the reconstructed image of non-coplanar scanning, anatomic structures of chest and lesions could be identified clearly and no significant artifacts of scanning direction dependent could be discovered. In 2D X-ray single projection, anatomic structures overlapped and lesions could not be discovered. In traditional tomosynthesis image, anatomic structures and lesions could be identified clearly, but there were many artifacts of scanning direction dependent. The quantitative results of PVRs show that there were no significant differences between non-coplanar tomosynthesis and traditional tomosynthesis. The PVRs of the non-coplanar technique were slightly higher than traditional technique in 5 mm and 8 mm lesions. In non-coplanar tomosynthesis, artifacts of scanning direction dependent could be reduced and PVRs of lesions were not decreased. The reconstructed image was more isotropic uniformity in non-coplanar tomosynthesis than in traditional tomosynthesis. In the future, scan strategy and scan time will be the challenges of non-coplanar imaging technique.Keywords: image reconstruction, non-coplanar imaging technique, tomosynthesis, X-ray imaging
Procedia PDF Downloads 3663908 Non-Local Simultaneous Sparse Unmixing for Hyperspectral Data
Authors: Fanqiang Kong, Chending Bian
Abstract:
Sparse unmixing is a promising approach in a semisupervised fashion by assuming that the observed pixels of a hyperspectral image can be expressed in the form of linear combination of only a few pure spectral signatures (end members) in an available spectral library. However, the sparse unmixing problem still remains a great challenge at finding the optimal subset of endmembers for the observed data from a large standard spectral library, without considering the spatial information. Under such circumstances, a sparse unmixing algorithm termed as non-local simultaneous sparse unmixing (NLSSU) is presented. In NLSSU, the non-local simultaneous sparse representation method for endmember selection of sparse unmixing, is used to finding the optimal subset of endmembers for the similar image patch set in the hyperspectral image. And then, the non-local means method, as a regularizer for abundance estimation of sparse unmixing, is used to exploit the abundance image non-local self-similarity. Experimental results on both simulated and real data demonstrate that NLSSU outperforms the other algorithms, with a better spectral unmixing accuracy.Keywords: hyperspectral unmixing, simultaneous sparse representation, sparse regression, non-local means
Procedia PDF Downloads 2453907 Theoretical Approaches to Graphic and Formal Generation from Evolutionary Genetics
Authors: Luz Estrada
Abstract:
The currents of evolutionary materialistic thought have argued that knowledge about an object is not obtained through the abstractive method. That is, the object cannot come to be understood if founded upon itself, nor does it take place by the encounter between form and matter. According to this affirmation, the research presented here identified as a problematic situation the absence of comprehension of the formal creation as a generative operation. This has been referred to as a recurrent lack in the production of objects and corresponds to the need to conceive the configurative process from the reality of its genesis. In this case, it is of interest to explore ways of creation that consider the object as if it were a living organism, as well as responding to the object’s experience as embodied in the designer since it unfolds its genesis simultaneously to the ways of existence of those who are involved in the generative experience.Keywords: architecture, theoretical graphics, evolutionary genetics, formal perception
Procedia PDF Downloads 1173906 6D Posture Estimation of Road Vehicles from Color Images
Authors: Yoshimoto Kurihara, Tad Gonsalves
Abstract:
Currently, in the field of object posture estimation, there is research on estimating the position and angle of an object by storing a 3D model of the object to be estimated in advance in a computer and matching it with the model. However, in this research, we have succeeded in creating a module that is much simpler, smaller in scale, and faster in operation. Our 6D pose estimation model consists of two different networks – a classification network and a regression network. From a single RGB image, the trained model estimates the class of the object in the image, the coordinates of the object, and its rotation angle in 3D space. In addition, we compared the estimation accuracy of each camera position, i.e., the angle from which the object was captured. The highest accuracy was recorded when the camera position was 75°, the accuracy of the classification was about 87.3%, and that of regression was about 98.9%.Keywords: 6D posture estimation, image recognition, deep learning, AlexNet
Procedia PDF Downloads 1553905 Learning at Workplace: Competences and Contexts in Sensory Evaluation
Authors: Ulriikka Savela-Huovinen, Hanni Muukkonen, Auli Toom
Abstract:
The development of workplace as a learning environment has been emphasized in research field of workplace learning. The prior literature on sensory performance emphasized the individual’s competences as assessor, while the competences in the collaborative interactional and knowledge creation practices as workplace learning method are not often mentioned. In the present study aims to find out what kinds of competences and contexts are central when assessor conducts food sensory evaluation in authentic professional context. The aim was to answer the following questions: first, what kinds of competences does sensory evaluation require according to assessors? And second, what kinds of contexts for sensory evaluation do assessors report? Altogether thirteen assessors from three Finnish food companies were interviewed by using semi-structural thematic interviews to map practices and development intentions as well as to explicate already established practices. The qualitative data were analyzed by following the principles of abductive and inductive content analysis. Analysis phases were combined and their results were considered together as a cross-analysis. When evaluated independently required competences were perception, knowledge of specific domains and methods and cognitive skills e.g. memory. Altogether, 42% of analysis units described individual evaluation contexts, 53% of analysis units described collaborative interactional contexts, and 5% of analysis units described collaborative knowledge creation contexts. Related to collaboration, analysis reviewed learning, sharing and reviewing both external and in-house consumer feedback, developing methods to moderate small-panel evaluation and developing product vocabulary collectively between the assessors. Knowledge creation contexts individualized from daily practices especially in cases product defects were sought and discussed. The study findings contribute to the explanation that sensory assessors learn extensively from one another in the collaborative interactional and knowledge creation context. Assessors learning and abilities to work collaboratively in the interactional and knowledge creation contexts need to be ensured in the development of the expertise.Keywords: assessor, collaboration, competences, contexts, learning and practices, sensory evaluation
Procedia PDF Downloads 2373904 The Mediating Effect of Destination Image on Intention to Use a Tourism App
Authors: Arej Alhemimah
Abstract:
This study investigates the influence of tourists’ perceptions of destination image on their intention to use a tourism app. It examines the roles played by tourists’ perceptions of app/website usability, information quality, and risk in shaping tourism destination image and, subsequently, their app use intention. Using an online questionnaire, the study surveyed 194 international tourists in Saudi Arabia. Results were analysed using PLS-SEM. All the proposed hypotheses were supported and significant. Perceived risk had the strongest influence, followed by the influence of tourists’ perceptions of information quality, then app usability. Additionally, perceived risk was found to have a strong effect on the application use intention. The study makes a significant contribution to the tourism website/application literature; its implications provide practical insights and recommendations for destination marketers and managers to improve their online and social media presence in terms of enhancing e-platform usability, quality of provided information, and most importantly, to create a destination strategy to manage tourists’ risk perceptions.Keywords: destination image, perceived risk, use intention, tourism app, information quality
Procedia PDF Downloads 83