Search results for: Low resolution digital encoder
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1374

Search results for: Low resolution digital encoder

444 Estimation of Load Impedance in Presence of Harmonics

Authors: Khaled M. EL-Naggar

Abstract:

This paper presents a fast and efficient on-line technique for estimating impedance of unbalanced loads in power systems. The proposed technique is an application of a discrete timedynamic filter based on stochastic estimation theory which is suitable for estimating parameters in noisy environment. The algorithm uses sets of digital samples of the distorted voltage and current waveforms of the non-linear load to estimate the harmonic contents of these two signal. The non-linear load impedance is then calculated from these contents. The method is tested using practical data. Results are reported and compared with those obtained using the conventional least error squares technique. In addition to the very accurate results obtained, the method can detect and reject bad measurements. This can be considered as a very important advantage over the conventional static estimation methods such as the least error square method.

Keywords: Estimation, identification, Harmonics, Dynamic Filter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2063
443 Research on Maintenance Design Method based Virtual Maintenance

Authors: Yunbin Yang, Liangli He, Fengjun Wang

Abstract:

The essentiality of maintenance assessment and maintenance optimization in design stage is analyzed, and the existent problems of conventional maintenance design method are illuminated. MDMVM (Maintenance Design Method based Virtual Maintenance) is illuminated, and the process of MDMVM established, and the MDMVM architecture is given out. The key techniques of MDMVM are analyzed, and include maintenance design based KBE (Knowledge Based Engineering) and virtual maintenance based physically attribute. According to physical property, physically based modeling, visual object movement control, the simulation of operation force and maintenance sequence planning method are emphatically illuminated. Maintenance design system based virtual maintenance is established in foundation of maintenance design method.

Keywords: Digital mock-up, virtual maintenance, knowledge engineering, maintenance sequence planning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1365
442 High Capacity Reversible Watermarking through Interpolated Error Shifting

Authors: Hae-Yeoun Lee

Abstract:

Reversible watermarking that not only protects the copyright but also preserve the original quality of the digital content have been intensively studied. In particular, the demand for reversible watermarking has increased. In this paper, we propose a reversible watermarking scheme based on interpolation-error shifting and error pre-compensation. The intensity of a pixel is interpolated from the intensities of neighboring pixels, and the difference histogram between the interpolated and the original intensities is obtained and modified to embed the watermark message. By restoring the difference histogram, the embedded watermark is extracted and the original image is recovered by compensating for the interpolation error. The overflow and underflow are prevented by error pre-compensation. To show the performance of the method, the proposed algorithm is compared with other methods using various test images.

Keywords: Reversible watermarking, High capacity, High quality, Interpolated error shifting, Error pre-compensation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2221
441 Low Cost Chip Set Selection Algorithm for Multi-way Partitioning of Digital System

Authors: Jae Young Park, Soongyu Kwon, Kyu Han Kim, Hyeong Geon Lee, Jong Tae Kim

Abstract:

This paper considers the problem of finding low cost chip set for a minimum cost partitioning of a large logic circuits. Chip sets are selected from a given library. Each chip in the library has a different price, area, and I/O pin. We propose a low cost chip set selection algorithm. Inputs to the algorithm are a netlist and a chip information in the library. Output is a list of chip sets satisfied with area and maximum partitioning number and it is sorted by cost. The algorithm finds the sorted list of chip sets from minimum cost to maximum cost. We used MCNC benchmark circuits for experiments. The experimental results show that all of chip sets found satisfy the multiple partitioning constraints.

Keywords: lowest cost chip set, MCNC benchmark, multi-way partitioning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1503
440 Scintigraphic Image Coding of Region of Interest Based On SPIHT Algorithm Using Global Thresholding and Huffman Coding

Authors: A. Seddiki, M. Djebbouri, D. Guerchi

Abstract:

Medical imaging produces human body pictures in digital form. Since these imaging techniques produce prohibitive amounts of data, compression is necessary for storage and communication purposes. Many current compression schemes provide a very high compression rate but with considerable loss of quality. On the other hand, in some areas in medicine, it may be sufficient to maintain high image quality only in region of interest (ROI). This paper discusses a contribution to the lossless compression in the region of interest of Scintigraphic images based on SPIHT algorithm and global transform thresholding using Huffman coding.

Keywords: Global Thresholding Transform, Huffman Coding, Region of Interest, SPIHT Coding, Scintigraphic images.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1979
439 A Comparison of Experimental Data with Monte Carlo Calculations for Optimisation of the Sourceto- Detector Distance in Determining the Efficiency of a LaBr3:Ce (5%) Detector

Authors: H. Aldousari, T. Buchacher, N. M. Spyrou

Abstract:

Cerium-doped lanthanum bromide LaBr3:Ce(5%) crystals are considered to be one of the most advanced scintillator materials used in PET scanning, combining a high light yield, fast decay time and excellent energy resolution. Apart from the correct choice of scintillator, it is also important to optimise the detector geometry, not least in terms of source-to-detector distance in order to obtain reliable measurements and efficiency. In this study a commercially available 25 mm x 25 mm BrilLanCeTM 380 LaBr3: Ce (5%) detector was characterised in terms of its efficiency at varying source-to-detector distances. Gamma-ray spectra of 22Na, 60Co, and 137Cs were separately acquired at distances of 5, 10, 15, and 20cm. As a result of the change in solid angle subtended by the detector, the geometric efficiency reduced in efficiency with increasing distance. High efficiencies at low distances can cause pulse pile-up when subsequent photons are detected before previously detected events have decayed. To reduce this systematic error the source-to-detector distance should be balanced between efficiency and pulse pile-up suppression as otherwise pile-up corrections would need to be necessary at short distances. In addition to the experimental measurements Monte Carlo simulations have been carried out for the same setup, allowing a comparison of results. The advantages and disadvantages of each approach have been highlighted.

Keywords: BrilLanCeTM380 LaBr3:Ce(5%), Coincidence summing, GATE simulation, Geometric efficiency

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1891
438 Image Classification and Accuracy Assessment Using the Confusion Matrix, Contingency Matrix, and Kappa Coefficient

Authors: F. F. Howard, C. B. Boye, I. Yakubu, J. S. Y. Kuma

Abstract:

One of the ways that could be used for the production of land use and land cover maps by a procedure known as image classification is the use of the remote sensing technique. Numerous elements ought to be taken into consideration, including the availability of highly satisfactory Landsat imagery, secondary data and a precise classification process. The goal of this study was to classify and map the land use and land cover of the study area using remote sensing and Geospatial Information System (GIS) analysis. The classification was done using Landsat 8 satellite images acquired in December 2020 covering the study area. The Landsat image was downloaded from the USGS. The Landsat image with 30 m resolution was geo-referenced to the WGS_84 datum and Universal Transverse Mercator (UTM) Zone 30N coordinate projection system. A radiometric correction was applied to the image to reduce the noise in the image. This study consists of two sections: the Land Use/Land Cover (LULC) and Accuracy Assessments using the confusion and contingency matrix and the Kappa coefficient. The LULC classifications were vegetation (agriculture) (67.87%), water bodies (0.01%), mining areas (5.24%), forest (26.02%), and settlement (0.88%). The overall accuracy of 97.87% and the kappa coefficient (K) of 97.3% were obtained for the confusion matrix. While an overall accuracy of 95.7% and a Kappa coefficient of 0.947 were obtained for the contingency matrix, the kappa coefficients were rated as substantial; hence, the classified image is fit for further research.

Keywords: Confusion Matrix, contingency matrix, kappa coefficient, land used/ land cover, accuracy assessment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 252
437 Particle Image Velocimetry for Measuring Water Flow Velocity

Authors: King Kuok Kuok, Po Chan Chiu

Abstract:

Floods are natural phenomena, which may turn into disasters causing widespread damage, health problems and even deaths. Nowadays, floods had become more serious and more frequent due to climatic changes. During flooding, discharge measurement still can be taken by standing on the bridge across the river using portable measurement instrument. However, it is too dangerous to get near to the river especially during high flood. Therefore, this study employs Particle Image Velocimetry (PIV) as a tool to measure the surface flow velocity. PIV is a image processing technique to track the movement of water from one point to another. The PIV codes are developed using Matlab. In this study, 18 ping pong balls were scattered over the surface of the drain and images were taken with a digital SLR camera. The images obtained were analyzed using the PIV code. Results show that PIV is able to produce the flow velocity through analyzing the series of images captured.

Keywords: Particle Image Velocimetry, flow velocity, surface flow.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2845
436 Proposing a Conceptual Model of Customer Knowledge Management: A Study of CKM Tools in British Dotcoms

Authors: Mehdi Shami Zanjani, Roshanak Rouzbehani, Hosein Dabbagh

Abstract:

Although current competitive challenges induced by today-s digital economy place their main emphasis on organizational knowledge, customer knowledge has been overlooked. On the other hand, the business community has finally begun to realize the important role customer knowledge can play in the organizational boundaries of the corporate arena. As a result, there is an emerging market for the tools and utilities whose objective is to provide the intelligence for knowledge sharing between the businesses and their customers. In this paper, we present a conceptual model of customer knowledge management by identifying and analyzing the existing tools in the market. The focus will be upon the emerging British dotcom industry whose customer based B2C behavior has been an influential part of the knowledge based intelligence tools in existence today.

Keywords: Customer knowledge, customer knowledge management, knowledge management, B2C E-commerce.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3392
435 An Integrated Framework for Engaging Stakeholders in the Circular Economy Processes Using Building Information Modeling and Virtual Reality

Authors: Erisasadat Sahebzamani, Núria Forcada, Francisco Lendinez

Abstract:

Global climate change has become increasingly problematic over the past few decades. The construction industry has contributed to greenhouse gas emissions in recent decades. Considering these issues and the high demand for materials in the construction industry, Circular Economy (CE) is considered necessary to keep materials in the loop and extend their useful lives. By providing tangible benefits, Construction 4.0 facilitates the adoption of CE by reducing waste, updating standard work, sharing knowledge, and increasing transparency and stability. This study aims to present a framework for integrating CE and digital tools like Building Information Modeling (BIM) and Virtual Reality (VR) to examine the impact on the construction industry based on stakeholders' perspectives.

Keywords: Circular Economy, Building Information Modeling, Virtual Reality, stakeholder engagement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 267
434 Objective Evaluation of Mathematical Morphology Edge Detection on Computed Tomography (CT) Images

Authors: Emhimed Saffor, Abdelkader Salama

Abstract:

In this paper problem of edge detection in digital images is considered. Edge detection based on morphological operators was applied on two sets (brain & chest) ct images. Three methods of edge detection by applying line morphological filters with multi structures in different directions have been used. 3x3 filter for first method, 5x5 filter for second method, and 7x7 filter for third method. We had applied this algorithm on (13 images) under MATLAB program environment. In order to evaluate the performance of the above mentioned edge detection algorithms, standard deviation (SD) and peak signal to noise ratio (PSNR) were used for justification for all different ct images. The objective method and the comparison of different methods of edge detection,  shows that high values of both standard deviation and PSNR values of edge detection images were obtained. 

Keywords: Medical images, Matlab, Edge detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2638
433 A Normalization-based Robust Watermarking Scheme Using Zernike Moments

Authors: Say Wei Foo, Qi Dong

Abstract:

Digital watermarking has become an important technique for copyright protection but its robustness against attacks remains a major problem. In this paper, we propose a normalizationbased robust image watermarking scheme. In the proposed scheme, original host image is first normalized to a standard form. Zernike transform is then applied to the normalized image to calculate Zernike moments. Dither modulation is adopted to quantize the magnitudes of Zernike moments according to the watermark bit stream. The watermark extracting method is a blind method. Security analysis and false alarm analysis are then performed. The quality degradation of watermarked image caused by the embedded watermark is visually transparent. Experimental results show that the proposed scheme has very high robustness against various image processing operations and geometric attacks.

Keywords: Image watermarking, Image normalization, Zernike moments, Robustness.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1755
432 Digital Hypertexts vs. Traditional Books: An Inquiry into Non-Linearity

Authors: Federica Fornaciari

Abstract:

The current study begins with an awareness that today-s media environment is characterized by technological development and a new way of reading caused by the introduction of the Internet. The researcher conducted a meta analysis framed within Technological Determinism to investigate the process of hypertext reading, its differences from linear reading and the effects such differences can have on people-s ways of mentally structuring their world. The relationship between literacy and the comprehension achieved by reading hypertexts is also investigated. The results show hypertexts are not always user friendly. People experience hyperlinks as interruptions that distract their attention generating comprehension and disorientation. On one hand hypertextual jumping reading generates interruptions that finally make people lose their concentration. On the other hand hypertexts fascinate people who would rather read a document in such a format even though the outcome is often frustrating and affects their ability to elaborate and retain information.

Keywords: Hypertext reading, Internet, non-linearity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1491
431 Automatic Segmentation of Dermoscopy Images Using Histogram Thresholding on Optimal Color Channels

Authors: Rahil Garnavi, Mohammad Aldeen, M. Emre Celebi, Alauddin Bhuiyan, Constantinos Dolianitis, George Varigos

Abstract:

Automatic segmentation of skin lesions is the first step towards development of a computer-aided diagnosis of melanoma. Although numerous segmentation methods have been developed, few studies have focused on determining the most discriminative and effective color space for melanoma application. This paper proposes a novel automatic segmentation algorithm using color space analysis and clustering-based histogram thresholding, which is able to determine the optimal color channel for segmentation of skin lesions. To demonstrate the validity of the algorithm, it is tested on a set of 30 high resolution dermoscopy images and a comprehensive evaluation of the results is provided, where borders manually drawn by four dermatologists, are compared to automated borders detected by the proposed algorithm. The evaluation is carried out by applying three previously used metrics of accuracy, sensitivity, and specificity and a new metric of similarity. Through ROC analysis and ranking the metrics, it is shown that the best results are obtained with the X and XoYoR color channels which results in an accuracy of approximately 97%. The proposed method is also compared with two state-ofthe- art skin lesion segmentation methods, which demonstrates the effectiveness and superiority of the proposed segmentation method.

Keywords: Border detection, Color space analysis, Dermoscopy, Histogram thresholding, Melanoma, Segmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2085
430 Lagrangian Flow Skeletons Captured in the Wake of a Swimming Nematode C. elegans Using an Immersed Boundary Fluid-Structure Interaction Approach

Authors: Arash Taheri

Abstract:

In this paper, Lagrangian coherent structure (LCS) concept is applied to wake flows generated in the up/down-stream of a swimming nematode C. elegans in an intermediate Re number range, i.e., 250-1200. It materializes Lagrangian hidden structures depicting flow transport barriers. To pursue the goals, nematode swimming in a quiescent fluid flow environment is numerically simulated by a two-way fluid-structure interaction (FSI) approach with the aid of immersed boundary method (IBM). In this regard, incompressible Navier-Stokes equations, fully-coupled with Lagrangian deformation equations for the immersed body, are solved using IB2d code. For all simulations, nematode’s body is modeled with a parametrized spring-fiber built-in case available in the computational code. Reverse von-Kármán vortex street formation and vortex shedding characteristics are studied and discussed in details via LCS approach, including grid resolution, integration time and Reynolds number effects. Results unveil presence of different flow regions with distinct fluid particle fates in the swimming animal’s wake and formation of so-called ‘mushroom-shaped’ structures in attracting LCS identities.

Keywords: Lagrangian coherent structure, nematode swimming, fluid-structure interaction, immersed boundary method, bionics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 994
429 A Novel Interpolation Scheme and Apparatus to Extend DAC Usable Spectrum over Nyquist Frequency

Authors: Wang liguo, Wang zongmin, Kong ying

Abstract:

A novel interpolation scheme to extend usable spectrum and upconvert in high performance D/A converters is addressed in this paper. By adjusting the pulse width of cycle and the production circuit of code, the expansion code is a null code or complementary code that is interpolation process. What the times and codes of interpolation decide DAC works in one of a normal mode or multi-mixer mode so that convert the input digital data signal into normal signal or a mixed analog signal having a mixer frequency that is higher than the data frequency. Simulation results show that the novel scheme and apparatus most extend the usable frequency spectrum into fifth to sixth Nyquist zone beyond conventional DACs.

Keywords: interpolation, upconversion, modulation, switching function, duty cycle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1507
428 Blind Identification and Equalization of CDMA Signals Using the Levenvberg-Marquardt Algorithm

Authors: Mohammed Boutalline, Imad Badi, Belaid Bouikhalene, Said Safi

Abstract:

In this paper we describe the Levenvberg-Marquardt (LM) algorithm for identification and equalization of CDMA signals received by an antenna array in communication channels. The synthesis explains the digital separation and equalization of signals after propagation through multipath generating intersymbol interference (ISI). Exploiting discrete data transmitted and three diversities induced at the reception, the problem can be composed by the Block Component Decomposition (BCD) of a tensor of order 3 which is a new tensor decomposition generalizing the PARAFAC decomposition. We optimize the BCD decomposition by Levenvberg-Marquardt method gives encouraging results compared to classical alternating least squares algorithm (ALS). In the equalization part, we use the Minimum Mean Square Error (MMSE) to perform the presented method. The simulation results using the LM algorithm are important.

Keywords: Identification and equalization, communication channel, Levenvberg-Marquardt, tensor decomposition

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1823
427 A Ring Segmented Bus Architecture for Globally Asynchronous Locally Synchronous System

Authors: Masafumi Kondo, Yoichiro Sato, Kazuyuki Tashiro, Tomoyuki Yokogawa, Michiyoshi Hayase

Abstract:

Recently, most digital systems are designed as GALS (Globally Asynchronous Locally Synchronous) systems. Several architectures have been proposed as bus architectures for a GALS system : shared bus, segmented bus, ring bus, and so on. In this study, we propose a ring segmented bus architecture which is a combination of segmented bus and ring bus architecture with the aim of throughput enhancement. In a segmented bus architecture, segments are connected in series. By connecting the segments at the end of the bus and constructing the ring bus, it becomes possible to allocate a channel of the bus bidirectionally. The bus channel is allocated to the shortest path between segments. We consider a metastable operation caused by asynchronous communication between segments and a burst transfer between segments. According to the result of simulation, it is shown that the GALS system designed by the proposed method has the desired operations.

Keywords: GALS systems bus architecture, segmented bus, ring bus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1138
426 A Comparative Study of Image Segmentation Algorithms

Authors: Mehdi Hosseinzadeh, Parisa Khoshvaght

Abstract:

In some applications, such as image recognition or compression, segmentation refers to the process of partitioning a digital image into multiple segments. Image segmentation is typically used to locate objects and boundaries (lines, curves, etc.) in images. Image segmentation is to classify or cluster an image into several parts (regions) according to the feature of image, for example, the pixel value or the frequency response. More precisely, image segmentation is the process of assigning a label to every pixel in an image such that pixels with the same label share certain visual characteristics. The result of image segmentation is a set of segments that collectively cover the entire image, or a set of contours extracted from the image. Several image segmentation algorithms were proposed to segment an image before recognition or compression. Up to now, many image segmentation algorithms exist and be extensively applied in science and daily life. According to their segmentation method, we can approximately categorize them into region-based segmentation, data clustering, and edge-base segmentation. In this paper, we give a study of several popular image segmentation algorithms that are available.

Keywords: Image Segmentation, hierarchical segmentation, partitional segmentation, density estimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2918
425 Key Based Text Watermarking of E-Text Documents in an Object Based Environment Using Z-Axis for Watermark Embedding

Authors: Mussarat Abdullah, Fazal Wahab

Abstract:

Data hiding into text documents itself involves pretty complexities due to the nature of text documents. A robust text watermarking scheme targeting an object based environment is presented in this research. The heart of the proposed solution describes the concept of watermarking an object based text document where each and every text string is entertained as a separate object having its own set of properties. Taking advantage of the z-ordering of objects watermark is applied with the z-axis letting zero fidelity disturbances to the text. Watermark sequence of bits generated against user key is hashed with selected properties of given document, to determine the bit sequence to embed. Bits are embedded along z-axis and the document has no fidelity issues when printed, scanned or photocopied.

Keywords: Digital Watermarking, Object Based Environment, Watermark, z-ordering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1687
424 Spatiotemporal Analysis of Visual Evoked Responses Using Dense EEG

Authors: Rima Hleiss, Elie Bitar, Mahmoud Hassan, Mohamad Khalil

Abstract:

A comprehensive study of object recognition in the human brain requires combining both spatial and temporal analysis of brain activity. Here, we are mainly interested in three issues: the time perception of visual objects, the ability of discrimination between two particular categories (objects vs. animals), and the possibility to identify a particular spatial representation of visual objects. Our experiment consisted of acquiring dense electroencephalographic (EEG) signals during a picture-naming task comprising a set of objects and animals’ images. These EEG responses were recorded from nine participants. In order to determine the time perception of the presented visual stimulus, we analyzed the Event Related Potentials (ERPs) derived from the recorded EEG signals. The analysis of these signals showed that the brain perceives animals and objects with different time instants. Concerning the discrimination of the two categories, the support vector machine (SVM) was applied on the instantaneous EEG (excellent temporal resolution: on the order of millisecond) to categorize the visual stimuli into two different classes. The spatial differences between the evoked responses of the two categories were also investigated. The results showed a variation of the neural activity with the properties of the visual input. Results showed also the existence of a spatial pattern of electrodes over particular regions of the scalp in correspondence to their responses to the visual inputs.

Keywords: Brain activity, dense EEG, evoked responses, spatiotemporal analysis, SVM, perception.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1071
423 Structural and Optical Properties ofInxAlyGa1-x-yN Quaternary Alloys

Authors: N. H. Abd Raof, H. Abu Hassan, S.K. Mohd Bakhori, S. S. Ng, Z. Hassan

Abstract:

Quaternary InxAlyGa1-x-yN semiconductors have attracted much research interest because the use of this quaternary offer the great flexibility in tailoring their band gap profile while maintaining their lattice-matching and structural integrity. The structural and optical properties of InxAlyGa1-x-yN alloys grown by molecular beam epitaxy (MBE) is presented. The structural quality of InxAlyGa1-x-yN layers was characterized using high-resolution X-ray diffraction (HRXRD). The results confirm that the InxAlyGa1-x-yN films had wurtzite structure and without phase separation. As the In composition increases, the Bragg angle of the (0002) InxAlyGa1-x-yN peak gradually decreases, indicating the increase in the lattice constant c of the alloys. FWHM of (0002) InxAlyGa1-x-yN decreases with increasing In composition from 0 to 0.04, that could indicate the decrease of quality of the samples due to point defects leading to non-uniformity of the epilayers. UV-VIS spectroscopy have been used to study the energy band gap of InxAlyGa1-x-yN. As the indium (In) compositions increases, the energy band gap decreases. However, for InxAlyGa1-x-yN with In composition of 0.1, the band gap shows a sudden increase in energy. This is probably due to local alloy compositional fluctuations in the epilayer. The bowing parameter which appears also to be very sensitive on In content is investigated and obtained b = 50.08 for quaternary InxAlyGa1-x-yN alloys. From photoluminescence (PL) measurement, green luminescence (GL) appears at PL spectrum of InxAlyGa1-x-yN, emitted for all x at ~530 nm and it become more pronounced as the In composition (x) increased, which is believed cause by gallium vacancies and related to isolated native defects.

Keywords: HRXRD, nitrides, PL, quaternary, UV-VIS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1572
422 An Efficient Biometric Cryptosystem using Autocorrelators

Authors: R. Bremananth, A. Chitra

Abstract:

Cryptography provides the secure manner of information transmission over the insecure channel. It authenticates messages based on the key but not on the user. It requires a lengthy key to encrypt and decrypt the sending and receiving the messages, respectively. But these keys can be guessed or cracked. Moreover, Maintaining and sharing lengthy, random keys in enciphering and deciphering process is the critical problem in the cryptography system. A new approach is described for generating a crypto key, which is acquired from a person-s iris pattern. In the biometric field, template created by the biometric algorithm can only be authenticated with the same person. Among the biometric templates, iris features can efficiently be distinguished with individuals and produces less false positives in the larger population. This type of iris code distribution provides merely less intra-class variability that aids the cryptosystem to confidently decrypt messages with an exact matching of iris pattern. In this proposed approach, the iris features are extracted using multi resolution wavelets. It produces 135-bit iris codes from each subject and is used for encrypting/decrypting the messages. The autocorrelators are used to recall original messages from the partially corrupted data produced by the decryption process. It intends to resolve the repudiation and key management problems. Results were analyzed in both conventional iris cryptography system (CIC) and non-repudiation iris cryptography system (NRIC). It shows that this new approach provides considerably high authentication in enciphering and deciphering processes.

Keywords: Autocorrelators, biometrics cryptography, irispatterns, wavelets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1527
421 Impact of Fixation Time on Subjective Video Quality Metric: a New Proposal for Lossy Compression Impairment Assessment

Authors: M. G. Albanesi, R. Amadeo

Abstract:

In this paper, a new approach for quality assessment tasks in lossy compressed digital video is proposed. The research activity is based on the visual fixation data recorded by an eye tracker. The method involved both a new paradigm for subjective quality evaluation and the subsequent statistical analysis to match subjective scores provided by the observer to the data obtained from the eye tracker experiments. The study brings improvements to the state of the art, as it solves some problems highlighted in literature. The experiments prove that data obtained from an eye tracker can be used to classify videos according to the level of impairment due to compression. The paper presents the methodology, the experimental results and their interpretation. Conclusions suggest that the eye tracker can be useful in quality assessment, if data are collected and analyzed in a proper way.

Keywords: eye tracker, video compression, video qualityassessment, visual attention

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1606
420 A Digital Media e-Learning Training Strategy for Healthcare Employees: Cost effective Distance Learning by Collaborative offline / online Engagement and Assessment

Authors: Lynn. J. MacFarlane. A

Abstract:

Within the healthcare system, training and continued professional development although essential, can be effected by cost and logistical restraints due to the nature of healthcare provision e.g employee shift patterns, access to expertise, cost factors in releasing staff to attend training etc. The use of multimedia technology for the development of e-learning applications is also a major cost consideration for healthcare management staff, and this type of media whether optical or on line requires careful planning in order to remain inclusive of all staff with potentially varied access to multimedia computing. This paper discusses a project in which the use of DVD authoring technology has been successfully implemented to meet the needs of distance learning and user considerations, and is based on film production techniques and reduced product turnaround deadlines.

Keywords: DVD, healthcare, distance learning, cost.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1529
419 A Robust Salient Region Extraction Based on Color and Texture Features

Authors: Mingxin Zhang, Zhaogan Lu, Junyi Shen

Abstract:

In current common research reports, salient regions are usually defined as those regions that could present the main meaningful or semantic contents. However, there are no uniform saliency metrics that could describe the saliency of implicit image regions. Most common metrics take those regions as salient regions, which have many abrupt changes or some unpredictable characteristics. But, this metric will fail to detect those salient useful regions with flat textures. In fact, according to human semantic perceptions, color and texture distinctions are the main characteristics that could distinct different regions. Thus, we present a novel saliency metric coupled with color and texture features, and its corresponding salient region extraction methods. In order to evaluate the corresponding saliency values of implicit regions in one image, three main colors and multi-resolution Gabor features are respectively used for color and texture features. For each region, its saliency value is actually to evaluate the total sum of its Euclidean distances for other regions in the color and texture spaces. A special synthesized image and several practical images with main salient regions are used to evaluate the performance of the proposed saliency metric and other several common metrics, i.e., scale saliency, wavelet transform modulus maxima point density, and important index based metrics. Experiment results verified that the proposed saliency metric could achieve more robust performance than those common saliency metrics.

Keywords: salient regions, color and texture features, image segmentation, saliency metric

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1567
418 Promoting Authenticity in Employer Brands to Address the Global-Local Problem in Complex Organisations: The Case of a Developing Country

Authors: Saud A. Taj

Abstract:

Employer branding is considered as a useful tool for addressing the global-local problem facing complex organisations that have operations scattered across the globe and face challenges of dealing with the local environment alongside. Despite being an established field of study within the Western developed world, there is little empirical evidence concerning the relevance of employer branding to global companies that operate in the under-developed economies. This paper fills this gap by gaining rich insight into the implementation of employer branding programs in a foreign multinational operating in Pakistan dealing with the global-local problem. The study is qualitative in nature and employs semistructured and focus group interviews with senior/middle managers and local frontline employees to deeply examine the phenomenon in case organisation. Findings suggest that authenticity is required in employer brands to enable them to respond to the local needs thereby leading to the resolution of the global-local problem. However, the role of signaling theory is key to the development of authentic employer brands as it stresses on the need to establish an efficient and effective signaling environment where in signals travel in both directions (from signal designers to receivers and backwards) and facilitate firms with the global-local problem. The paper also identifies future avenues of research for the employer branding field.

Keywords: Authenticity, Counter-signals, Employer Branding, Global-Local Problem, Signaling Theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1806
417 Space Vector Pulse Width Modulation Technique Based Design and Simulation of a Three-Phase Voltage Source Converter Systems

Authors: Farhan Beg

Abstract:

A Space Vector based Pulse Width Modulation control technique for the three-phase PWM converter is proposed in this paper. The proposed control scheme is based on a synchronous reference frame model. High performance and efficiency is obtained with regards to the DC bus voltage and the power factor considerations of the PWM rectifier thus leading to low losses. MATLAB/SIMULINK are used as a platform for the simulations and a SIMULINK model is presented in the paper. The results show that the proposed model demonstrates better performance and properties compared to the traditional SPWM method and the method improves the dynamic performance of the closed loop drastically. For the Space Vector based Pulse Width Modulation, Sine signal is the reference waveform and triangle waveform is the carrier waveform. When the value sine signal is large than triangle signal, the pulse will start produce to high. And then when the triangular signals higher than sine signal, the pulse will come to low. SPWM output will changed by changing the value of the modulation index and frequency used in this system to produce more pulse width. The more pulse width produced, the output voltage will have lower harmonics contents and the resolution increase.

Keywords: Power Factor, SVPWM, PWM rectifier, SPWM.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4023
416 A Robust Image Watermarking Scheme using Image Moment Normalization

Authors: Latha Parameswaran, K. Anbumani

Abstract:

Multimedia security is an incredibly significant area of concern. A number of papers on robust digital watermarking have been presented, but there are no standards that have been defined so far. Thus multimedia security is still a posing problem. The aim of this paper is to design a robust image-watermarking scheme, which can withstand a different set of attacks. The proposed scheme provides a robust solution integrating image moment normalization, content dependent watermark and discrete wavelet transformation. Moment normalization is useful to recover the watermark even in case of geometrical attacks. Content dependent watermarks are a powerful means of authentication as the data is watermarked with its own features. Discrete wavelet transforms have been used as they describe image features in a better manner. The proposed scheme finds its place in validating identification cards and financial instruments.

Keywords: Watermarking, moments, wavelets, content-based, benchmarking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1546
415 A Tutorial on Dynamic Simulation of DC Motor and Implementation of Kalman Filter on a Floating Point DSP

Authors: Padmakumar S., Vivek Agarwal, Kallol Roy

Abstract:

With the advent of inexpensive 32 bit floating point digital signal processor-s availability in market, many computationally intensive algorithms such as Kalman filter becomes feasible to implement in real time. Dynamic simulation of a self excited DC motor using second order state variable model and implementation of Kalman Filter in a floating point DSP TMS320C6713 is presented in this paper with an objective to introduce and implement such an algorithm, for beginners. A fractional hp DC motor is simulated in both Matlab® and DSP and the results are included. A step by step approach for simulation of DC motor in Matlab® and “C" routines in CC Studio® is also given. CC studio® project file details and environmental setting requirements are addressed. This tutorial can be used with 6713 DSK, which is based on floating point DSP and CC Studio either in hardware mode or in simulation mode.

Keywords: DC motor, DSP, Dynamic simulation, Kalman Filter

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3015