Search results for: digital image correlation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8507

Search results for: digital image correlation

8267 CT Medical Images Denoising Based on New Wavelet Thresholding Compared with Curvelet and Contourlet

Authors: Amir Moslemi, Amir movafeghi, Shahab Moradi

Abstract:

One of the most important challenging factors in medical images is nominated as noise.Image denoising refers to the improvement of a digital medical image that has been infected by Additive White Gaussian Noise (AWGN). The digital medical image or video can be affected by different types of noises. They are impulse noise, Poisson noise and AWGN. Computed tomography (CT) images are subjected to low quality due to the noise. The quality of CT images is dependent on the absorbed dose to patients directly in such a way that increase in absorbed radiation, consequently absorbed dose to patients (ADP), enhances the CT images quality. In this manner, noise reduction techniques on the purpose of images quality enhancement exposing no excess radiation to patients is one the challenging problems for CT images processing. In this work, noise reduction in CT images was performed using two different directional 2 dimensional (2D) transformations; i.e., Curvelet and Contourlet and Discrete wavelet transform(DWT) thresholding methods of BayesShrink and AdaptShrink, compared to each other and we proposed a new threshold in wavelet domain for not only noise reduction but also edge retaining, consequently the proposed method retains the modified coefficients significantly that result in good visual quality. Data evaluations were accomplished by using two criterions; namely, peak signal to noise ratio (PSNR) and Structure similarity (Ssim).

Keywords: computed tomography (CT), noise reduction, curve-let, contour-let, signal to noise peak-peak ratio (PSNR), structure similarity (Ssim), absorbed dose to patient (ADP)

Procedia PDF Downloads 413
8266 The Democratization of 3D Capturing: An Application Investigating Google Tango Potentials

Authors: Carlo Bianchini, Lorenzo Catena

Abstract:

The appearance of 3D scanners and then, more recently, of image-based systems that generate point clouds directly from common digital images have deeply affected the survey process in terms of both capturing and 2D/3D modelling. In this context, low cost and mobile systems are increasingly playing a key role and actually paving the way to the democratization of what in the past was the realm of few specialized technicians and expensive equipment. The application of Google Tango on the ancient church of Santa Maria delle Vigne in Pratica di Mare – Rome presented in this paper is one of these examples.

Keywords: the architectural survey, augmented/mixed/virtual reality, Google Tango project, image-based 3D capturing

Procedia PDF Downloads 118
8265 Smartphone Photography in Urban China

Authors: Wen Zhang

Abstract:

The smartphone plays a significant role in media convergence, and smartphone photography is reconstructing the way we communicate and think. This article aims to explore the smartphone photography practices of urban Chinese smartphone users and images produced by smartphones from a techno-cultural perspective. The analysis consists of two types of data: One is a semi-structured interview of 21 participants, and the other consists of the images created by the participants. The findings are organised in two parts. The first part summarises the current tendencies of capturing, editing, sharing and archiving digital images via smartphones. The second part shows that food and selfie/anti-selfie are the preferred subjects of smartphone photographic images from a technical and multi-purpose perspective and demonstrates that screenshots and image texts are new genres of non-photographic images that are frequently made by smartphones, which contributes to improving operational efficiency, disseminating information and sharing knowledge. The analyses illustrate the positive impacts between smartphones and photography enthusiasm and practices based on the diffusion of innovation theory, which also makes us rethink the value of photographs and the practice of ‘photographic seeing’ from the screen itself.

Keywords: digital photography, image-text, media convergence, photographic- seeing, selfie/anti-selfie, smartphone, technological innovation

Procedia PDF Downloads 322
8264 Digital Activism and the Individual: A Utilitarian Perspective

Authors: Tania Mitra

Abstract:

Digital Activism or Cyber Activism uses digital media as a means to disseminate information and mobilize masses towards a specific goal. When digital activism was first born in the early 1990s, it was primarily used by groups of organized political activists. However, with the advent of social media, online activism has filtered down to the individual- one who does not necessarily belong to or identify with an agenda, group, or political party. A large part of digital activism today stems from the individual’s notion of what is right and wrong. This gives rise to a discourse around descriptive ethics and the implications of the independent digital activist. Although digital activism has paved the way for and bolstered support for causes like the MeToo Movement and Black Lives Matter, the lack of a unified, organized body has led to counterintuitive progressions and suspicions regarding the movements. The paper introduces the ideas of 'clout' culture, click baits, and clicktivism (the phenomenon where activism is reduced to a blind following of the online trends), to discuss the impacts of exclusive digital activism. By using Jeremy Bentham's utilitarian approach to ethics, that places emphasis on the best possible outcome for a society, the paper will show how individual online activism reaching for a larger, more common end can sometimes lead to an undermining of that end, not only in the online space but also how it manifests in the real world.

Keywords: digital activism, ethics, independent digital activist, utilitarianism

Procedia PDF Downloads 92
8263 Body Image Dissatifaction with and Personal Behavioral Control in Obese Patients Who are Attending to Treatment

Authors: Mariela Gonzalez, Zoraide Lugli, Eleonora Vivas, Rosana Guzmán

Abstract:

The objective was to determine the predictive capacity of self-efficacy perceived for weight control, locus of weight control and skills of weight self-management in the dissatisfaction of the body image in obese people who attend treatment. Sectional study conducted in the city of Maracay, Venezuela, with 243 obese who attend to treatment, 173 of the feminine gender and 70 of the male, with ages ranging between 18 and 57 years old. The sample body mass index ranged between 29.39 and 44.14. The following instruments were used: The Body Shape Questionnaire (BSQ), the inventory of body weight self-regulation, The Inventory of self-efficacy in the regulation of body weight and the Inventory of the Locus of weight control. Calculating the descriptive statistics and of central tendency, coefficients of correlation and multiple regression; it was found that a low ‘perceived Self-efficacy in the weight control’ and a high ‘Locus of external control’, predict the dissatisfaction with body image in obese who attend treatment. The findings are a first approximation to give an account of the importance of the personal control variables in the study of the psychological grief on the overweight individual.

Keywords: dissatisfaction with body image, obese people, personal control, psychological variables

Procedia PDF Downloads 401
8262 A Four-Step Ortho-Rectification Procedure for Geo-Referencing Video Streams from a Low-Cost UAV

Authors: B. O. Olawale, C. R. Chatwin, R. C. D. Young, P. M. Birch, F. O. Faithpraise, A. O. Olukiran

Abstract:

Ortho-rectification is the process of geometrically correcting an aerial image such that the scale is uniform. The ortho-image formed from the process is corrected for lens distortion, topographic relief, and camera tilt. This can be used to measure true distances, because it is an accurate representation of the Earth’s surface. Ortho-rectification and geo-referencing are essential to pin point the exact location of targets in video imagery acquired at the UAV platform. This can only be achieved by comparing such video imagery with an existing digital map. However, it is only when the image is ortho-rectified with the same co-ordinate system as an existing map that such a comparison is possible. The video image sequences from the UAV platform must be geo-registered, that is, each video frame must carry the necessary camera information before performing the ortho-rectification process. Each rectified image frame can then be mosaicked together to form a seamless image map covering the selected area. This can then be used for comparison with an existing map for geo-referencing. In this paper, we present a four-step ortho-rectification procedure for real-time geo-referencing of video data from a low-cost UAV equipped with multi-sensor system. The basic procedures for the real-time ortho-rectification are: (1) Decompilation of video stream into individual frames; (2) Finding of interior camera orientation parameters; (3) Finding the relative exterior orientation parameters for each video frames with respect to each other; (4) Finding the absolute exterior orientation parameters, using self-calibration adjustment with the aid of a mathematical model. Each ortho-rectified video frame is then mosaicked together to produce a 2-D planimetric mapping, which can be compared with a well referenced existing digital map for the purpose of georeferencing and aerial surveillance. A test field located in Abuja, Nigeria was used for testing our method. Fifteen minutes video and telemetry data were collected using the UAV and the data collected were processed using the four-step ortho-rectification procedure. The results demonstrated that the geometric measurement of the control field from ortho-images are more reliable than those from original perspective photographs when used to pin point the exact location of targets on the video imagery acquired by the UAV. The 2-D planimetric accuracy when compared with the 6 control points measured by a GPS receiver is between 3 to 5 meters.

Keywords: geo-referencing, ortho-rectification, video frame, self-calibration

Procedia PDF Downloads 454
8261 Development of a Tesla Music Coil from Signal Processing

Authors: Samaniego Campoverde José Enrique, Rosero Muñoz Jorge Enrique, Luzcando Narea Lorena Elizabeth

Abstract:

This paper presents a practical and theoretical model for the operation of the Tesla coil using digital signal processing. The research is based on the analysis of ten scientific papers exploring the development and operation of the Tesla coil. Starting from the Testa coil, several modifications were carried out on the Tesla coil, with the aim of amplifying the digital signal by making use of digital signal processing. To achieve this, an amplifier with a transistor and digital filters provided by MATLAB software were used, which were chosen according to the characteristics of the signals in question.

Keywords: tesla coil, digital signal process, equalizer, graphical environment

Procedia PDF Downloads 72
8260 Determination of Johnson-Cook Material and Failure Model Constants for High Tensile Strength Tendon Steel in Post-Tensioned Concrete Members

Authors: I. Gkolfinopoulos, N. Chijiwa

Abstract:

To evaluate the remaining capacity in concrete tensioned members, it is important to accurately estimate damage in precast concrete tendons. In this research Johnson-Cook model and damage parameters of high-strength steel material were calculated by static and dynamic uniaxial tensile tests. Replication of experimental results was achieved through finite element analysis for both single 8-noded three-dimensional element as well as the full-scale dob-bone shaped model and relevant model parameters are proposed. Finally, simulation results in terms of strain and deformation were verified using digital image correlation analysis.

Keywords: DIC analysis, Johnson-Cook, quasi-static, dynamic, rupture, tendon

Procedia PDF Downloads 107
8259 Morphometry of Female Reproductive Tract in Small Ruminants Using Ultrasonography

Authors: R. Jannat, N. S. Juyena, F. Y. Bari, M. N. Islam

Abstract:

Understanding anatomy of female reproductive organs is very much important to identify any variation in disease condition. Therefore, this study was conducted to determine the morphometry of female reproductive tract in small ruminant using ultrasonography. The reproductive tracts of 2l does and 20 ewes were collected, and both gross and ultrasonographic image measurements were performed to study morphometry of cervix, body of uterus, horn of uterus and ovary. Water bath ultrasonography technique was used with trans-abdominal linear probe for image measurements. Results revealed significant (P<0.001) variation among gross and image measurements of cervix, body of uterus and ovaries in does whereas, significant (P<0.001) variation existed between gross and image measurements of ovaries diameter in ewes. Gross measurements were proportionately higher than image measurements in both species. The mean length and width were found higher in right ovaries than those of left ovaries. In addition, the diameter of right ovaries was higher than those of left ovaries in both species. Pearson's correlation revealed a positive relation between two measurements. Moreover, it was found that echogenicity varied with reproductive organs. This is a model study. This study may help to identify female reproductive structures by trans-abdominal ultrasonography.

Keywords: female reproductive tract, morphometry, small ruminants, ultrasonography

Procedia PDF Downloads 237
8258 Digital Innovation and Business Transformation

Authors: Bisola Stella Sonde

Abstract:

Digital innovation has emerged as a pivotal driver of business transformation in the contemporary landscape. This case study research explores the dynamic interplay between digital innovation and the profound metamorphosis of businesses across industries. It delves into the multifaceted dimensions of digital innovation, elucidating its impact on organizational structures, customer experiences, and operational paradigms. The study investigates real-world instances of businesses harnessing digital technologies to enhance their competitiveness, agility, and sustainability. It scrutinizes the strategic adoption of digital platforms, data analytics, artificial intelligence, and emerging technologies as catalysts for transformative change. The cases encompass a diverse spectrum of industries, spanning from traditional enterprises to disruptive startups, offering insights into the universal relevance of digital innovation. Moreover, the research scrutinizes the challenges and opportunities posed by the digital era, shedding light on the intricacies of managing cultural shifts, data privacy, and cybersecurity concerns in the pursuit of innovation. It unveils the strategies that organizations employ to adapt, thrive, and lead in the era of digital disruption. In summary, this case study research underscores the imperative of embracing digital innovation as a cornerstone of business transformation. It offers a comprehensive exploration of the contemporary digital landscape, offering valuable lessons for organizations striving to navigate the ever-evolving terrain of the digital age.

Keywords: business transformation, digital innovation, emerging technologies, organizational structures

Procedia PDF Downloads 24
8257 Automatic Classification Using Dynamic Fuzzy C Means Algorithm and Mathematical Morphology: Application in 3D MRI Image

Authors: Abdelkhalek Bakkari

Abstract:

Image segmentation is a critical step in image processing and pattern recognition. In this paper, we proposed a new robust automatic image classification based on a dynamic fuzzy c-means algorithm and mathematical morphology. The proposed segmentation algorithm (DFCM_MM) has been applied to MR perfusion images. The obtained results show the validity and robustness of the proposed approach.

Keywords: segmentation, classification, dynamic, fuzzy c-means, MR image

Procedia PDF Downloads 437
8256 Detect Circles in Image: Using Statistical Image Analysis

Authors: Fathi M. O. Hamed, Salma F. Elkofhaifee

Abstract:

The aim of this work is to detect geometrical shape objects in an image. In this paper, the object is considered to be as a circle shape. The identification requires find three characteristics, which are number, size, and location of the object. To achieve the goal of this work, this paper presents an algorithm that combines from some of statistical approaches and image analysis techniques. This algorithm has been implemented to arrive at the major objectives in this paper. The algorithm has been evaluated by using simulated data, and yields good results, and then it has been applied to real data.

Keywords: image processing, median filter, projection, scale-space, segmentation, threshold

Procedia PDF Downloads 396
8255 Medical Image Compression by Region of Interest Based on DT-CWT Using Run-length Coding and Huffman Coding

Authors: Ali Seddiki, Mohamed Djebbouri, Driss Guerchi

Abstract:

Medical imaging produces human body pictures in digital form. Since these imaging techniques produce prohibitive amounts of data, compression is necessary for storage and communication purposes. In some areas in medicine, it may be sufficient to maintain high image quality only in region of interest (ROI). This paper discusses a contribution to quality purpose compression in the region of interest of scintigraphic images based on dual tree complex wavelet transform (DT-CWT) using Run-Length coding (RLE) and Huffman coding (HC).

Keywords: DT-CWT, region of interest, run length coding, Scintigraphic images

Procedia PDF Downloads 247
8254 Adaptive Dehazing Using Fusion Strategy

Authors: M. Ramesh Kanthan, S. Naga Nandini Sujatha

Abstract:

The goal of haze removal algorithms is to enhance and recover details of scene from foggy image. In enhancement the proposed method focus into two main categories: (i) image enhancement based on Adaptive contrast Histogram equalization, and (ii) image edge strengthened Gradient model. Many circumstances accurate haze removal algorithms are needed. The de-fog feature works through a complex algorithm which first determines the fog destiny of the scene, then analyses the obscured image before applying contrast and sharpness adjustments to the video in real-time to produce image the fusion strategy is driven by the intrinsic properties of the original image and is highly dependent on the choice of the inputs and the weights. Then the output haze free image has reconstructed using fusion methodology. In order to increase the accuracy, interpolation method has used in the output reconstruction. A promising retrieval performance is achieved especially in particular examples.

Keywords: single image, fusion, dehazing, multi-scale fusion, per-pixel, weight map

Procedia PDF Downloads 432
8253 Quick Similarity Measurement of Binary Images via Probabilistic Pixel Mapping

Authors: Adnan A. Y. Mustafa

Abstract:

In this paper we present a quick technique to measure the similarity between binary images. The technique is based on a probabilistic mapping approach and is fast because only a minute percentage of the image pixels need to be compared to measure the similarity, and not the whole image. We exploit the power of the Probabilistic Matching Model for Binary Images (PMMBI) to arrive at an estimate of the similarity. We show that the estimate is a good approximation of the actual value, and the quality of the estimate can be improved further with increased image mappings. Furthermore, the technique is image size invariant; the similarity between big images can be measured as fast as that for small images. Examples of trials conducted on real images are presented.

Keywords: big images, binary images, image matching, image similarity

Procedia PDF Downloads 156
8252 Investigation of the Relationship between Digital Game Playing, Internet Addiction and Perceived Stress Levels in University Students

Authors: Sevim Ugur, Cemile Kutmec Yilmaz, Omer Us, Sevdenur Koksaldi

Abstract:

Aim: This study aims to investigate the effect of digital game playing and Internet addiction on perceived stress levels in university students. Method: The descriptive study was conducted through face-to-face interview method with a total of 364 university students studying at Aksaray University between November 15 and December 30, 2017. The research data were collected using personal information form, a questionnaire to determine the characteristics of playing digital game, the Internet addiction scale and the perceived stress scale. In the evaluation of the data, Mann-Whitney U test was used for two-group comparison of the sample with non-normal distribution, Kruskal-Wallis H-test was used in the comparison of more than two groups, and the Spearman correlation test was used to determine the relationship between Internet addiction and the perceived stress level. Results: It was determined that the mean age of the students participated in the study was 20.13 ± 1.7 years, 67.6% was female, 35.7% was sophomore, and 62.1% had an income 500 TL or less. It was found that 83.5% of the students use the Internet every day and 70.6% uses the Internet for 5 hours or less per day. Of the students, 12.4% prefers digital games instead of spending time outdoors, 8% plays a game as the first activity in leisure time, 12.4% plays all day, 15.7% feels anger when he/she is prevented from playing, 14.8% prefers playing games to get away from his/her problems, 23.4% had his/her school achievement affected negatively because of game playing, and 8% argues with family members due to the time spent for gaming. Students who play games on the computer for a long time were found to feel back pain (30.8%), headache (28.6%), insomnia (26.9%), dryness and pain in the eyes (26.6%), pain in the wrist (21.2%), feeling excessive tension and anger (16.2%), humpback (12.9), vision loss (9.6%) and pain in the wrist and fingers (7.4%). In our study, students' Internet addiction scale mean score was found to be 45.47 ± 16.1 and mean perceived stress scale score was 28.56 ± 2.7. A significant and negative correlation (p=0.037) was found between the total score of the Internet addiction scale and the total score of the perceived stress scale (r=-0.110). Conclusion: It was found in the study that Internet addiction and perceived stress of the students were at a moderate level and that there was a negative correlation between Internet addiction and perceived stress levels. Internet addiction was found to increase with the increasing perceived stress levels of students, and students were found to have health problems such as back pain, dryness in the eyes, pain, insomnia, headache, and humpback. Therefore, it is recommended to inform students about different coping methods other than spending time on the Internet to cope with the stress they perceive.

Keywords: digital game, internet addiction, student, stress level

Procedia PDF Downloads 250
8251 Design of a Graphical User Interface for Data Preprocessing and Image Segmentation Process in 2D MRI Images

Authors: Enver Kucukkulahli, Pakize Erdogmus, Kemal Polat

Abstract:

The 2D image segmentation is a significant process in finding a suitable region in medical images such as MRI, PET, CT etc. In this study, we have focused on 2D MRI images for image segmentation process. We have designed a GUI (graphical user interface) written in MATLABTM for 2D MRI images. In this program, there are two different interfaces including data pre-processing and image clustering or segmentation. In the data pre-processing section, there are median filter, average filter, unsharp mask filter, Wiener filter, and custom filter (a filter that is designed by user in MATLAB). As for the image clustering, there are seven different image segmentations for 2D MR images. These image segmentation algorithms are as follows: PSO (particle swarm optimization), GA (genetic algorithm), Lloyds algorithm, k-means, the combination of Lloyds and k-means, mean shift clustering, and finally BBO (Biogeography Based Optimization). To find the suitable cluster number in 2D MRI, we have designed the histogram based cluster estimation method and then applied to these numbers to image segmentation algorithms to cluster an image automatically. Also, we have selected the best hybrid method for each 2D MR images thanks to this GUI software.

Keywords: image segmentation, clustering, GUI, 2D MRI

Procedia PDF Downloads 347
8250 Grid Pattern Recognition and Suppression in Computed Radiographic Images

Authors: Igor Belykh

Abstract:

Anti-scatter grids used in radiographic imaging for the contrast enhancement leave specific artifacts. Those artifacts may be visible or may cause Moiré effect when a digital image is resized on a diagnostic monitor. In this paper, we propose an automated grid artifacts detection and suppression algorithm which is still an actual problem. Grid artifacts detection is based on statistical approach in spatial domain. Grid artifacts suppression is based on Kaiser bandstop filter transfer function design and application avoiding ringing artifacts. Experimental results are discussed and concluded with description of advantages over existing approaches.

Keywords: grid, computed radiography, pattern recognition, image processing, filtering

Procedia PDF Downloads 250
8249 Modeling Visual Memorability Assessment with Autoencoders Reveals Characteristics of Memorable Images

Authors: Elham Bagheri, Yalda Mohsenzadeh

Abstract:

Image memorability refers to the phenomenon where certain images are more likely to be remembered by humans than others. It is a quantifiable and intrinsic attribute of an image. Understanding how visual perception and memory interact is important in both cognitive science and artificial intelligence. It reveals the complex processes that support human cognition and helps to improve machine learning algorithms by mimicking the brain's efficient data processing and storage mechanisms. To explore the computational underpinnings of image memorability, this study examines the relationship between an image's reconstruction error, distinctiveness in latent space, and its memorability score. A trained autoencoder is used to replicate human-like memorability assessment inspired by the visual memory game employed in memorability estimations. This study leverages a VGG-based autoencoder that is pre-trained on the vast ImageNet dataset, enabling it to recognize patterns and features that are common to a wide and diverse range of images. An empirical analysis is conducted using the MemCat dataset, which includes 10,000 images from five broad categories: animals, sports, food, landscapes, and vehicles, along with their corresponding memorability scores. The memorability score assigned to each image represents the probability of that image being remembered by participants after a single exposure. The autoencoder is finetuned for one epoch with a batch size of one, attempting to create a scenario similar to human memorability experiments where memorability is quantified by the likelihood of an image being remembered after being seen only once. The reconstruction error, which is quantified as the difference between the original and reconstructed images, serves as a measure of how well the autoencoder has learned to represent the data. The reconstruction error of each image, the error reduction, and its distinctiveness in latent space are calculated and correlated with the memorability score. Distinctiveness is measured as the Euclidean distance between each image's latent representation and its nearest neighbor within the autoencoder's latent space. Different structural and perceptual loss functions are considered to quantify the reconstruction error. The results indicate that there is a strong correlation between the reconstruction error and the distinctiveness of images and their memorability scores. This suggests that images with more unique distinct features that challenge the autoencoder's compressive capacities are inherently more memorable. There is also a negative correlation between the reduction in reconstruction error compared to the autoencoder pre-trained on ImageNet, which suggests that highly memorable images are harder to reconstruct, probably due to having features that are more difficult to learn by the autoencoder. These insights suggest a new pathway for evaluating image memorability, which could potentially impact industries reliant on visual content and mark a step forward in merging the fields of artificial intelligence and cognitive science. The current research opens avenues for utilizing neural representations as instruments for understanding and predicting visual memory.

Keywords: autoencoder, computational vision, image memorability, image reconstruction, memory retention, reconstruction error, visual perception

Procedia PDF Downloads 39
8248 Optimal Image Representation for Linear Canonical Transform Multiplexing

Authors: Navdeep Goel, Salvador Gabarda

Abstract:

Digital images are widely used in computer applications. To store or transmit the uncompressed images requires considerable storage capacity and transmission bandwidth. Image compression is a means to perform transmission or storage of visual data in the most economical way. This paper explains about how images can be encoded to be transmitted in a multiplexing time-frequency domain channel. Multiplexing involves packing signals together whose representations are compact in the working domain. In order to optimize transmission resources each 4x4 pixel block of the image is transformed by a suitable polynomial approximation, into a minimal number of coefficients. Less than 4*4 coefficients in one block spares a significant amount of transmitted information, but some information is lost. Different approximations for image transformation have been evaluated as polynomial representation (Vandermonde matrix), least squares + gradient descent, 1-D Chebyshev polynomials, 2-D Chebyshev polynomials or singular value decomposition (SVD). Results have been compared in terms of nominal compression rate (NCR), compression ratio (CR) and peak signal-to-noise ratio (PSNR) in order to minimize the error function defined as the difference between the original pixel gray levels and the approximated polynomial output. Polynomial coefficients have been later encoded and handled for generating chirps in a target rate of about two chirps per 4*4 pixel block and then submitted to a transmission multiplexing operation in the time-frequency domain.

Keywords: chirp signals, image multiplexing, image transformation, linear canonical transform, polynomial approximation

Procedia PDF Downloads 385
8247 Medical Image Compression Based on Region of Interest: A Review

Authors: Sudeepti Dayal, Neelesh Gupta

Abstract:

In terms of transmission, bigger the size of any image, longer the time the channel takes for transmission. It is understood that the bandwidth of the channel is fixed. Therefore, if the size of an image is reduced, a larger number of data or images can be transmitted over the channel. Compression is the technique used to reduce the size of an image. In terms of storage, compression reduces the file size which it occupies on the disk. Any image is based on two parameters, region of interest and non-region of interest. There are several algorithms of compression that compress the data more economically. In this paper we have reviewed region of interest and non-region of interest based compression techniques and the algorithms which compress the image most efficiently.

Keywords: compression ratio, region of interest, DCT, DWT

Procedia PDF Downloads 342
8246 Image Enhancement of Histological Slides by Using Nonlinear Transfer Function

Authors: D. Suman, B. Nikitha, J. Sarvani, V. Archana

Abstract:

Histological slides provide clinical diagnostic information about the subjects from the ancient times. Even with the advent of high resolution imaging cameras the image tend to have some background noise which makes the analysis complex. A study of the histological slides is done by using a nonlinear transfer function based image enhancement method. The method processes the raw, color images acquired from the biological microscope, which, in general, is associated with background noise. The images usually appearing blurred does not convey the intended information. In this regard, an enhancement method is proposed and implemented on 50 histological slides of human tissue by using nonlinear transfer function method. The histological image is converted into HSV color image. The luminance value of the image is enhanced (V component) because change in the H and S components could change the color balance between HSV components. The HSV image is divided into smaller blocks for carrying out the dynamic range compression by using a linear transformation function. Each pixel in the block is enhanced based on the contrast of the center pixel and its neighborhood. After the processing the V component, the HSV image is transformed into a colour image. The study has shown improvement of the characteristics of the image so that the significant details of the histological images were improved.

Keywords: HSV space, histology, enhancement, image

Procedia PDF Downloads 300
8245 Similarity Based Retrieval in Case Based Reasoning for Analysis of Medical Images

Authors: M. Dasgupta, S. Banerjee

Abstract:

Content Based Image Retrieval (CBIR) coupled with Case Based Reasoning (CBR) is a paradigm that is becoming increasingly popular in the diagnosis and therapy planning of medical ailments utilizing the digital content of medical images. This paper presents a survey of some of the promising approaches used in the detection of abnormalities in retina images as well in mammographic screening and detection of regions of interest in MRI scans of the brain. We also describe our proposed algorithm to detect hard exudates in fundus images of the retina of Diabetic Retinopathy patients.

Keywords: case based reasoning, exudates, retina image, similarity based retrieval

Procedia PDF Downloads 314
8244 Lifelong Learning and Digital Literacies in Language Learning

Authors: Selma Karabinar

Abstract:

Lifelong learning can be described as a system where learning takes place for a person over the course of a lifespan and comprises formal, non-formal and informal learning to achieve the maximum possible improvement in personal, social, and vocational life. 21st century is marked with the digital technologies and people need to learn and adapt to new literacies as part of their lifelong learning. Our current knowledge gap brings to mind several questions: Do people with digital mindsets have different assumptions about affordances of digital technologies? How do digital mindsets lead language learners use digital technologies within and beyond classrooms? Does digital literacies have different significance for the learners? The presentation is based on a study attempted to answer these questions and show the relationship between lifelong learning and digital literacies. The study was conducted with learners of English language at a state university in Istanbul. The quantitative data in terms of participants' lifelong learning perception was collected through a lifelong learning scale from 150 students. Then 5 students with high and 5 with low lifelong learning perception were interviewed. They were questioned about their personal sense of agency in lifelong learning and how they use digital technologies in their language learning. Therefore, the qualitative data was analyzed in terms of their knowledge about digital literacies and actual use of it in their personal and educational life. The results of the study suggest why teaching new literacies are important for lifelong learning and also suggests implications for language teachers' education and language pedagogy.

Keywords: digital mindsets, language learning, lifelong learning, new literacies

Procedia PDF Downloads 340
8243 An Efficient Encryption Scheme Using DWT and Arnold Transforms

Authors: Ali Abdrhman M. Ukasha

Abstract:

Data security needed in data transmission, storage, and communication to ensure the security. The color image is decomposed into red, green, and blue channels. The blue and green channels are compressed using 3-levels discrete wavelet transform. The Arnold transform uses to changes the locations of red image channel pixels as image scrambling process. Then all these channels are encrypted separately using a key image that has same original size and is generating using private keys and modulo operations. Performing the X-OR and modulo operations between the encrypted channels images for image pixel values change purpose. The extracted contours of color image recovery can be obtained with accepted level of distortion using Canny edge detector. Experiments have demonstrated that proposed algorithm can fully encrypt 2D color image and completely reconstructed without any distortion. It has shown that the color image can be protected with a higher security level. The presented method has easy hardware implementation and suitable for multimedia protection in real time applications such as wireless networks and mobile phone services.

Keywords: color image, wavelet transform, edge detector, Arnold transform, lossy image encryption

Procedia PDF Downloads 447
8242 A Closer Look on Economic and Fiscal Incentives for Digital TV Industry

Authors: Yunita Anwar, Maya Safira Dewi

Abstract:

With the increasing importance on digital TV industry, there must be several incentives given to support the growth of the industry. Prior research have found mixed findings of economic and fiscal incentives to economic growth, which means these incentives do not necessarily boost the economic growth while providing support to a particular industry. Focusing on a setting of digital TV transition in Indonesia, this research will conduct document analysis to analyze incentives have been given in other country and incentives currently available in Indonesia. Our results recommend that VAT exemption and local tax incentives could be considered to be added to the incentives list available for digital TV industry.

Keywords: Digital TV transition, Economic Incentives, Fiscal Incentives, Policy.

Procedia PDF Downloads 288
8241 Digital Literacy Landscape of Islamic Boarding Schools in Indonesia

Authors: Zainuddin Abuhamid Muhammad Ghozali, Andrew Whitworth

Abstract:

Islamic boarding school or pesantren is a distinctive education institution in Indonesia focusing on religious teachings. Its stance in restricting access to the internet raises a question about its students’ development of digital literacy. Inspired by Luckin’s ecology of resource model, this study aims to map out the digital literacy situation of the institution based on the availability of learning resources, such as digital facilities, digital accessibility, and digital competence. This study was carried out through a survey method involving 50 teachers from pesantrens across the nation. The result shows that pesantrens have provided students with digital facilities at a moderate level, yet the accessibility to using them is still limited. They also incorporated digital competencies into their curriculum, with an emphasis on digital ethics. The study also identifies different patterns of pesantrens’ behavior based on types and educational levels, where certain school types and educational levels tend to give a stricter policy compared to others or vice versa. The restriction of digital resources in pesantren indicated that they had done a filtration process to design their learning environment. The filtration was mainly motivated by sociocultural factors, where they drew concern for the negative impact of the internet. Notably, this restriction also contributes to students’ poor development of digital literacy.

Keywords: digital literacy, ecology of resources, Indonesia, Islamic boarding school

Procedia PDF Downloads 38
8240 Digital Recording System Identification Based on Audio File

Authors: Michel Kulhandjian, Dimitris A. Pados

Abstract:

The objective of this work is to develop a theoretical framework for reliable digital recording system identification from digital audio files alone, for forensic purposes. A digital recording system consists of a microphone and a digital sound processing card. We view the cascade as a system of unknown transfer function. We expect same manufacturer and model microphone-sound card combinations to have very similar/near identical transfer functions, bar any unique manufacturing defect. Input voice (or other) signals are modeled as non-stationary processes. The technical problem under consideration becomes blind deconvolution with non-stationary inputs as it manifests itself in the specific application of digital audio recording equipment classification.

Keywords: blind system identification, audio fingerprinting, blind deconvolution, blind dereverberation

Procedia PDF Downloads 276
8239 Filling the Gap of Extraction of Digital Evidence from Emerging Platforms Without Forensics Tools

Authors: Yi Anson Lam, Siu Ming Yiu, Kam Pui Chow

Abstract:

Digital evidence has been tendering to courts at an exponential rate in recent years. As an industrial practice, most digital evidence is extracted and preserved using specialized and well-accepted forensics tools. On the other hand, the advancement in technologies enables the creation of quite a few emerging platforms such as Telegram, Signal etc. Existing (well-accepted) forensics tools were not designed to extract evidence from these emerging platforms. While new forensics tools require a significant amount of time and effort to be developed and verified, this paper tries to address how to fill this gap using quick-fix alternative methods for digital evidence collection (e.g., based on APIs provided by Apps) and discuss issues related to the admissibility of this evidence to courts with support from international courts’ stance and the circumstances of accepting digital evidence using these proposed alternatives.

Keywords: extraction, digital evidence, laws, investigation

Procedia PDF Downloads 35
8238 Counting People Utilizing Space-Time Imagery

Authors: Ahmed Elmarhomy, K. Terada

Abstract:

An automated method for counting passerby has been proposed using virtual-vertical measurement lines. Space-time image is representing the human regions which are treated using the segmentation process. Different color space has been used to perform the template matching. A proper template matching has been achieved to determine direction and speed of passing people. Distinguish one or two passersby has been investigated using a correlation between passerby speed and the human-pixel area. Finally, the effectiveness of the presented method has been experimentally verified.

Keywords: counting people, measurement line, space-time image, segmentation, template matching

Procedia PDF Downloads 420