Search results for: multivariate statistical technique
3400 Optimization of Conditions for Xanthan Gum Production from Waste Date in Submerged Fermantation
Authors: S. Moshaf, Z. Hamidi-Esfahani, M. H. Azizi
Abstract:
Xanthan gum is one of the major commercial biopolymers. Due to its excellent rheological properties xanthan gum is used in many applications, mainly in food industry. Commercial production of xanthan gum uses glucose as the carbon substrate; consequently the price of xanthan production is high. One of the ways to decrease xanthan price, is using cheaper substrate like agricultural wastes. Iran is one of the biggest date producer countries. However approximately 50% of date production is wasted annually. The goal of this study is to produce xanthan gum from waste date using Xanthomonas campestris PTCC1473 by submerged fermentation. In this study the effect of three variables including phosphor and nitrogen amount and agitation rate in three levels using response surface methodology (RSM) has been studied. Results achieved from statistical analysis Design Expert 7.0.0 software showed that xanthan increased with increasing level of phosphor. Low level of nitrogen leaded to higher xanthan production. Xanthan amount, increasing agitation had positive influence. The statistical model identified the optimum conditions nitrogen amount=3.15g/l, phosphor amount=5.03 g/l and agitation=394.8 rpm for xanthan. To model validation, experiments in optimum conditions for xanthan gum were carried out. The mean of result for xanthan was 6.72±0.26. The result was closed to the predicted value by using RSM.Keywords: Optimization, RSM, Waste date, Xanthan gum, Xanthomonas Campestris
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26123399 QSAR Studies of Certain Novel Heterocycles Derived from Bis-1, 2, 4 Triazoles as Anti-Tumor Agents
Authors: Madhusudan Purohit, Stephen Philip, Bharathkumar Inturi
Abstract:
In this paper we report the quantitative structure activity relationship of novel bis-triazole derivatives for predicting the activity profile. The full model encompassed a dataset of 46 Bis- triazoles. Tripos Sybyl X 2.0 program was used to conduct CoMSIA QSAR modeling. The Partial Least-Squares (PLS) analysis method was used to conduct statistical analysis and to derive a QSAR model based on the field values of CoMSIA descriptor. The compounds were divided into test and training set. The compounds were evaluated by various CoMSIA parameters to predict the best QSAR model. An optimum numbers of components were first determined separately by cross-validation regression for CoMSIA model, which were then applied in the final analysis. A series of parameters were used for the study and the best fit model was obtained using donor, partition coefficient and steric parameters. The CoMSIA models demonstrated good statistical results with regression coefficient (r2) and the cross-validated coefficient (q2) of 0.575 and 0.830 respectively. The standard error for the predicted model was 0.16322. In the CoMSIA model, the steric descriptors make a marginally larger contribution than the electrostatic descriptors. The finding that the steric descriptor is the largest contributor for the CoMSIA QSAR models is consistent with the observation that more than half of the binding site area is occupied by steric regions.
Keywords: 3D QSAR, CoMSIA, Triazoles.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14823398 3D Object Model Reconstruction Based on Polywogs Wavelet Network Parametrization
Authors: Mohamed Othmani, Yassine Khlifi
Abstract:
This paper presents a technique for compact three dimensional (3D) object model reconstruction using wavelet networks. It consists to transform an input surface vertices into signals,and uses wavelet network parameters for signal approximations. To prove this, we use a wavelet network architecture founded on several mother wavelet families. POLYnomials WindOwed with Gaussians (POLYWOG) wavelet families are used to maximize the probability to select the best wavelets which ensure the good generalization of the network. To achieve a better reconstruction, the network is trained several iterations to optimize the wavelet network parameters until the error criterion is small enough. Experimental results will shown that our proposed technique can effectively reconstruct an irregular 3D object models when using the optimized wavelet network parameters. We will prove that an accurateness reconstruction depends on the best choice of the mother wavelets.Keywords: 3D object, optimization, parametrization, Polywog wavelets, reconstruction, wavelet networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15023397 Comparison of Fricative Vocal Tract Transfer Functions Derived using Two Different Segmentation Techniques
Authors: K. S. Subari, C. H. Shadle, A. Barney, R. I. Damper
Abstract:
The acoustic and articulatory properties of fricative speech sounds are being studied using magnetic resonance imaging (MRI) and acoustic recordings from a single subject. Area functions were derived from a complete set of axial and coronal MR slices using two different methods: the Mermelstein technique and the Blum transform. Area functions derived from the two techniques were shown to differ significantly in some cases. Such differences will lead to different acoustic predictions and it is important to know which is the more accurate. The vocal tract acoustic transfer function (VTTF) was derived from these area functions for each fricative and compared with measured speech signals for the same fricative and same subject. The VTTFs for /f/ in two vowel contexts and the corresponding acoustic spectra are derived here; the Blum transform appears to show a better match between prediction and measurement than the Mermelstein technique.
Keywords: Area functions, fricatives, vocal tract transferfunction, MRI, speech.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16533396 Using Business Intelligence Capabilities to Improve the Quality of Decision-Making: A Case Study of Mellat Bank
Authors: Jalal Haghighat Monfared, Zahra Akbari
Abstract:
Today, business executives need to have useful information to make better decisions. Banks have also been using information tools so that they can direct the decision-making process in order to achieve their desired goals by rapidly extracting information from sources with the help of business intelligence. The research seeks to investigate whether there is a relationship between the quality of decision making and the business intelligence capabilities of Mellat Bank. Each of the factors studied is divided into several components, and these and their relationships are measured by a questionnaire. The statistical population of this study consists of all managers and experts of Mellat Bank's General Departments (including 190 people) who use commercial intelligence reports. The sample size of this study was 123 randomly determined by statistical method. In this research, relevant statistical inference has been used for data analysis and hypothesis testing. In the first stage, using the Kolmogorov-Smirnov test, the normalization of the data was investigated and in the next stage, the construct validity of both variables and their resulting indexes were verified using confirmatory factor analysis. Finally, using the structural equation modeling and Pearson's correlation coefficient, the research hypotheses were tested. The results confirmed the existence of a positive relationship between decision quality and business intelligence capabilities in Mellat Bank. Among the various capabilities, including data quality, correlation with other systems, user access, flexibility and risk management support, the flexibility of the business intelligence system was the most correlated with the dependent variable of the present research. This shows that it is necessary for Mellat Bank to pay more attention to choose the required business intelligence systems with high flexibility in terms of the ability to submit custom formatted reports. Subsequently, the quality of data on business intelligence systems showed the strongest relationship with quality of decision making. Therefore, improving the quality of data, including the source of data internally or externally, the type of data in quantitative or qualitative terms, the credibility of the data and perceptions of who uses the business intelligence system, improves the quality of decision making in Mellat Bank.
Keywords: Business intelligence, business intelligence capability, decision making, decision quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13853395 Active Contours with Prior Corner Detection
Authors: U.A.A. Niroshika, Ravinda G.N. Meegama
Abstract:
Deformable active contours are widely used in computer vision and image processing applications for image segmentation, especially in biomedical image analysis. The active contour or “snake" deforms towards a target object by controlling the internal, image and constraint forces. However, if the contour initialized with a lesser number of control points, there is a high probability of surpassing the sharp corners of the object during deformation of the contour. In this paper, a new technique is proposed to construct the initial contour by incorporating prior knowledge of significant corners of the object detected using the Harris operator. This new reconstructed contour begins to deform, by attracting the snake towards the targeted object, without missing the corners. Experimental results with several synthetic images show the ability of the new technique to deal with sharp corners with a high accuracy than traditional methods.Keywords: Active Contours, Image Segmentation, Harris Operator, Snakes
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22823394 OHASD: The First On-Line Arabic Sentence Database Handwritten on Tablet PC
Authors: Randa I. M. Elanwar, Mohsen A. Rashwan, Samia A. Mashali
Abstract:
In this paper we present the first Arabic sentence dataset for on-line handwriting recognition written on tablet pc. The dataset is natural, simple and clear. Texts are sampled from daily newspapers. To collect naturally written handwriting, forms are dictated to writers. The current version of our dataset includes 154 paragraphs written by 48 writers. It contains more than 3800 words and more than 19,400 characters. Handwritten texts are mainly written by researchers from different research centers. In order to use this dataset in a recognition system word extraction is needed. In this paper a new word extraction technique based on the Arabic handwriting cursive nature is also presented. The technique is applied to this dataset and good results are obtained. The results can be considered as a bench mark for future research to be compared with.Keywords: Arabic, Handwriting recognition, on-line dataset.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20573393 A New Vision of Fractal Geometry with Triangulati on Algorithm
Authors: Yasser M. Abd El-Latif, Fatma S.Abousaleh, Daoud S. S.
Abstract:
L-system is a tool commonly used for modeling and simulating the growth of fractal plants. The aim of this paper is to join some problems of the computational geometry with the fractal geometry by using the L-system technique to generate fractal plant in 3D. L-system constructs the fractal structure by applying rewriting rules sequentially and this technique depends on recursion process with large number of iterations to get different shapes of 3D fractal plants. Instead, it was reiterated a specific number of iterations up to three iterations. The vertices generated from the last stage of the Lsystem rewriting process are used as input to the triangulation algorithm to construct the triangulation shape of these vertices. The resulting shapes can be used as covers for the architectural objects and in different computer graphics fields. The paper presents a gallery of triangulation forms which application in architecture creates an alternative for domes and other traditional types of roofs.
Keywords: Computational geometry, fractal geometry, L-system, triangulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19213392 Effect of Peak-to-Average Power Ratio Reduction on the Multicarrier Communication System Performance Parameters
Authors: Sanjay Singh, M Sathish Kumar, H. S Mruthyunjaya
Abstract:
Multicarrier transmission system such as Orthogonal Frequency Division Multiplexing (OFDM) is a promising technique for high bit rate transmission in wireless communication system. OFDM is a spectrally efficient modulation technique that can achieve high speed data transmission over multipath fading channels without the need for powerful equalization techniques. However the price paid for this high spectral efficiency and less intensive equalization is low power efficiency. OFDM signals are very sensitive to nonlinear effects due to the high Peak-to-Average Power Ratio (PAPR), which leads to the power inefficiency in the RF section of the transmitter. This paper investigates the effect of PAPR reduction on the performance parameter of multicarrier communication system. Performance parameters considered are power consumption of Power Amplifier (PA) and Digital-to-Analog Converter (DAC), power amplifier efficiency, SNR of DAC and BER performance of the system. From our analysis it is found that irrespective of PAPR reduction technique being employed, the power consumption of PA and DAC reduces and power amplifier efficiency increases due to reduction in PAPR. Moreover, it has been shown that for a given BER performance the requirement of Input-Backoff (IBO) reduces with reduction in PAPR.Keywords: BER, Crest Factor (CF), Digital-to-Analog Converter(DAC), Input-Backoff (IBO), Orthogonal Frequency Division Multiplexing(OFDM), Peak-to-Average Power Ratio (PAPR), PowerAmplifier efficiency, SNR
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32883391 A Hybrid Approach for Color Image Quantization Using K-means and Firefly Algorithms
Authors: Parisut Jitpakdee, Pakinee Aimmanee, Bunyarit Uyyanonvara
Abstract:
Color Image quantization (CQ) is an important problem in computer graphics, image and processing. The aim of quantization is to reduce colors in an image with minimum distortion. Clustering is a widely used technique for color quantization; all colors in an image are grouped to small clusters. In this paper, we proposed a new hybrid approach for color quantization using firefly algorithm (FA) and K-means algorithm. Firefly algorithm is a swarmbased algorithm that can be used for solving optimization problems. The proposed method can overcome the drawbacks of both algorithms such as the local optima converge problem in K-means and the early converge of firefly algorithm. Experiments on three commonly used images and the comparison results shows that the proposed algorithm surpasses both the base-line technique k-means clustering and original firefly algorithm.Keywords: Clustering, Color quantization, Firefly algorithm, Kmeans.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22213390 Statistical Relation between Vegetation Cover and Land Surface Temperature in Phnom Penh City
Authors: Gulam Mohiuddin, Jan-Peter Mund
Abstract:
This study assessed the correlation between Normalized Difference Vegetation Index (NDVI) and Land Surface Temperature (LST) in Phnom Penh City (Cambodia) from 2016 to 2020. Understanding the LST and NDVI can be helpful to understand the Urban Heat Island (UHI) scenario, and it can contribute to planning urban greening and combating the effects of UHI. The study used Landsat-8 images as the data for analysis. They have 100 m spatial resolution (per pixel) in the thermal band. The current study used an approach for the statistical analysis that considers every pixel from the study area instead of taking few sample points or analyzing descriptive statistics. Also, this study is examining the correlation between NDVI and LST with a spatially explicit approach. The study found a strong negative correlation between NDVI and LST (coefficient range -0.56 to -0.59), and this relationship is linear. This study showed a way to avoid the probable error from the sample-based approach in examining two spatial variables. The method is reproducible for a similar type of analysis on the correlation between spatial phenomena. The findings of this study will be used further to understand the causation behind LST change in that area triangulating LST, NDVI and land-use changes.
Keywords: Land Surface Temperature, NDVI, Normalized Difference Vegetation Index, remote sensing, methodological development.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4743389 Audio Watermarking Based on Compression-expansion Technique
Authors: Say Wei Foo, Qi Dong
Abstract:
A novel robust audio watermarking scheme is proposed in this paper. In the proposed scheme, the host audio signals are segmented into frames. Two consecutive frames are assessed if they are suitable to represent a watermark bit. If so, frequency transform is performed on these two frames. The compressionexpansion technique is adopted to generate distortion over the two frames. The distortion is used to represent one watermark bit. Psychoacoustic model is applied to calculate local auditory mask to ensure that the distortion is not audible. The watermarking schemes using mono and stereo audio signals are designed differently. The correlation-based detection method is used to detect the distortion and extract embedded watermark bits. The experimental results show that the quality degradation caused by the embedded watermarks is perceptually transparent and the proposed schemes are very robust against different types of attacks.Keywords: Audio watermarking, Compression-expansion, Stereo signals, Robustness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16463388 The Significance of the Radiography Technique in the Non-Destructive Evaluation of the Integrity and Reliability of Cast Interconnects
Authors: Keshav Pujeri, Pranesh Jain, Krutibas Panda
Abstract:
Significant changes in oil and gas drilling have emphasized the need to verify the integrity and reliability of drill stem components. Defects are inevitable in cast components, regardless of application; but if these defects go undetected, any severe defect could cause down-hole failure. One such defect is shrinkage porosity. Castings with lower level shrinkage porosity (CB levels 1 and 2) have scattered pores and do not occupy large volumes; so pressure testing and helium leak testing (HLT) are sufficient for qualifying the castings. However, castings with shrinkage porosity of CB level 3 and higher, behave erratically under pressure testing and HLT making these techniques insufficient for evaluating the castings- integrity. This paper presents a case study to highlight how the radiography technique is much more effective than pressure testing and HLT.Keywords: Casting Defects, Interconnects, Leak Check, Pressure Test, Radiography.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35153387 Estimation of Functional Response Model by Supervised Functional Principal Component Analysis
Authors: Hyon I. Paek, Sang Rim Kim, Hyon A. Ryu
Abstract:
In functional linear regression, one typical problem is to reduce dimension. Compared with multivariate linear regression, functional linear regression is regarded as an infinite-dimensional case, and the main task is to reduce dimensions of functional response and functional predictors. One common approach is to adapt functional principal component analysis (FPCA) on functional predictors and then use a few leading functional principal components (FPC) to predict the functional model. The leading FPCs estimated by the typical FPCA explain a major variation of the functional predictor, but these leading FPCs may not be mostly correlated with the functional response, so they may not be significant in the prediction for response. In this paper, we propose a supervised FPCA method for a functional response model with FPCs obtained by considering the correlation of the functional response. Our method would have a better prediction accuracy than the typical FPCA method.
Keywords: Supervised, functional principal component analysis, functional response, functional linear regression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 273386 Topographical Image Transference Compatibility Generated Through Moiré Technique Applying Parametrical Softwares of Computer Assisted Design
Authors: M. V. G. Silva, J. Gazzola, I. M. Dal Fabbro, A. C. L. Lino
Abstract:
Computer aided design accounts with the support of parametric software in the design of machine components as well as of any other pieces of interest. The complexities of the element under study sometimes offer certain difficulties to computer design, or ever might generate mistakes in the final body conception. Reverse engineering techniques are based on the transformation of already conceived body images into a matrix of points which can be visualized by the design software. The literature exhibits several techniques to obtain machine components dimensional fields, as contact instrument (MMC), calipers and optical methods as laser scanner, holograms as well as moiré methods. The objective of this research work was to analyze the moiré technique as instrument of reverse engineering, applied to bodies of nom complex geometry as simple solid figures, creating matrices of points. These matrices were forwarded to a parametric software named SolidWorks to generate the virtual object. Volume data obtained by mechanical means, i.e., by caliper, the volume obtained through the moiré method and the volume generated by the SolidWorks software were compared and found to be in close agreement. This research work suggests the application of phase shifting moiré methods as instrument of reverse engineering, serving also to support farm machinery element designs.Keywords: Reverse engineering, Moiré technique, three dimensional image generation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34593385 Thermal Behavior of a Ventilated Façade Using Perforated Ceramic Bricks
Authors: H. López-Moreno, A. Rodríguez-Sánchez, C. Viñas-Arrebola, C. Porras-Amores
Abstract:
The ventilated façade has great advantages when compared to traditional façades as it reduces the air conditioning thermal loads due to the stack effect induced by solar radiation in the air chamber. Optimizing energy consumption by using a ventilated façade can be used not only in newly built buildings but also it can be implemented in existing buildings, opening the field of implementation to energy building retrofitting works. In this sense, the following three prototypes of façade where designed, built and further analyzed in this research: non-ventilated façade (NVF); slightly ventilated façade (SLVF) and strongly ventilated façade (STVF). The construction characteristics of the three facades are based on the Spanish regulation of building construction “Technical Building Code”. The façades have been monitored by type-k thermocouples in a representative day of the summer season in Madrid (Spain). Moreover, an analysis of variance (ANOVA) with repeated measures, studying the thermal lag in the ventilated and no-ventilated façades has been designed. Results show that STVF façade presents higher levels of thermal inertia as the thermal lag reduces up to 17% (daily mean) compared to the non-ventilated façade. In addition, the statistical analysis proves that an increase of the ventilation holes size in STVF façades can improve the thermal lag significantly (p >0.05) when compared to the SLVF façade.Keywords: Energy efficiency, experimental study, statistical analysis, thermal behavior, ventilated façade.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 41173384 Modern Vibration Signal Processing Techniques for Vehicle Gearbox Fault Diagnosis
Authors: Mohamed El Morsy, Gabriela Achtenová
Abstract:
This paper presents modern vibration signalprocessing techniques for vehicle gearbox fault diagnosis, via the wavelet analysis and the Squared Envelope (SE) technique. The wavelet analysis is regarded as a powerful tool for the detection of sudden changes in non-stationary signals. The Squared Envelope (SE) technique has been extensively used for rolling bearing diagnostics. In the present work a scheme of using the Squared Envelope technique for early detection of gear tooth pit. The pitting defect is manufactured on the tooth side of a fifth speed gear on the intermediate shaft of a vehicle gearbox. The objective is to supplement the current techniques of gearbox fault diagnosis based on using the raw vibration and ordered signals. The test stand is equipped with three dynamometers; the input dynamometer serves as the internal combustion engine, the output dynamometers introduce the load on the flanges of output joint shafts. The gearbox used for experimental measurements is the type most commonly used in modern small to mid-sized passenger cars with transversely mounted powertrain and front wheel drive; a five-speed gearbox with final drive gear and front wheel differential. The results show that the approaches methods are effective for detecting and diagnosing localized gear faults in early stage under different operation conditions, and are more sensitive and robust than current gear diagnostic techniques.
Keywords: Wavelet analysis, Squared Envelope, gear faults.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25783383 Evaluation of Stormwater Quantity and Quality Control through Constructed Mini Wet Pond: A Case Study
Authors: Y. S. Liew, K. A. Puteh Ariffin, M. A. Mohd Nor
Abstract:
One of the Best Management Practices (BMPs) promoted in Urban Stormwater Management Manual for Malaysia (MSMA) published by the Department of Irrigation and Drainage (DID) in 2001 is through the construction of wet ponds in new development projects for water quantity and quality control. Therefore, this paper aims to demonstrate a case study on evaluation of a constructed mini wet pond located at Sekolah Rendah Kebangsaan Seksyen 2, Puchong, Selangor, Malaysia in both stormwater quantity and quality aspect particularly to reduce the peak discharge by temporary storing and gradual release of stormwater runoff from an outlet structure or other release mechanism. The evaluation technique will be using InfoWorks Collection System (CS) as the numerical modeling approach for water quantity aspect. Statistical test by comparing the correlation coefficient (R2), mean error (ME), mean absolute error (MAE) and root mean square error (RMSE) were used to evaluate the model in simulating the peak discharge changes. Results demonstrated that there will be a reduction in peak flow at 11 % to 15% and time to peak flow is slower by 5 minutes through a wet pond. For water quality aspect, a survey on biological indicator of water quality carried out depicts that the pond is within the range of rather clean to clean water with the score of 5.3. This study indicates that a constructed wet pond with wetland facilities is able to help in managing water quantity and stormwater generated pollution at source, towards achieving ecologically sustainable development in urban areas.
Keywords: Wet pond, Retention Facilities, Best Management Practices (BMP), Urban Stormwater Management Manual for Malaysia (MSMA).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25273382 Loop Back Connected Component Labeling Algorithm and Its Implementation in Detecting Face
Authors: A. Rakhmadi, M. S. M. Rahim, A. Bade, H. Haron, I. M. Amin
Abstract:
In this study, a Loop Back Algorithm for component connected labeling for detecting objects in a digital image is presented. The approach is using loop back connected component labeling algorithm that helps the system to distinguish the object detected according to their label. Deferent than whole window scanning technique, this technique reduces the searching time for locating the object by focusing on the suspected object based on certain features defined. In this study, the approach was also implemented for a face detection system. Face detection system is becoming interesting research since there are many devices or systems that require detecting the face for certain purposes. The input can be from still image or videos, therefore the sub process of this system has to be simple, efficient and accurate to give a good result.Keywords: Image processing, connected components labeling, face detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23023381 Computer Generated Hologram for SemiFragile Watermarking with Encrypted Images
Authors: G. Schirripa Spagnolo, M. De Santis
Abstract:
The protection of the contents of digital products is referred to as content authentication. In some applications, to be able to authenticate a digital product could be extremely essential. For example, if a digital product is used as a piece of evidence in the court, its integrity could mean life or death of the accused. Generally, the problem of content authentication can be solved using semifragile digital watermarking techniques. Recently many authors have proposed Computer Generated Hologram Watermarking (CGHWatermarking) techniques. Starting from these studies, in this paper a semi-fragile Computer Generated Hologram coding technique is proposed, which is able to detect malicious tampering while tolerating some incidental distortions. The proposed technique uses as watermark an encrypted image, and it is well suitable for digital image authentication.Keywords: Asymmetric cryptography, Semi-Fragile watermarking, Image authentication, Hologram watermark, Public- Key Cryptography, RSA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16103380 Investigating Determinants of Medical User Expectations from Hospital Information System
Authors: G. Gürsel, K. H. Gülkesen, N. Zayim, A. Arifoğlu, O. Saka
Abstract:
User satisfaction is one of the most used success indicators in the research of information system (IS). Literature shows user expectations have great influence on user satisfaction. Both expectation and satisfaction of users are important for Hospital Information Systems (HIS). Education, IS experience, age, attitude towards change, business title, sex and working unit of the hospital, are examined as the potential determinant of the medical users’ expectations. Data about medical user expectations are collected by the “Expectation Questionnaire” developed for this study. Expectation data are used for calculating the Expectation Meeting Ratio (EMR) with the evaluation framework also developed for this study. The internal consistencies of the answers to the questionnaire are measured by Cronbach´s Alpha coefficient. The multivariate analysis of medical user’s EMRs of HIS is performed by forward stepwise binary logistic regression analysis. Education and business title is appeared to be the determinants of expectations from HIS.Keywords: Evaluation, Fuzzy Logic, Hospital Information System, User Expectation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19483379 Body Mass Index and Dietary Habits among Nursing College Students Living in the University Residence in Kirkuk City, Iraq
Authors: Jenan Shakoor
Abstract:
Obesity prevalence is increasing worldwide. University life is a challenging period especially for students who have to leave their familiar surroundings and settle in a new environment. The current study aimed to assess the diet and exercise habits and their association with body mass index (BMI) among nursing college students living at Kirkuk University residence. This was a descriptive study. A non-probability (purposive) sample of 101 students living in Kirkuk University residence was recruited during the period from the 15th November 2015 to the 5th May 2016. A questionnaire was constructed for the purpose of the study which consisted of four parts: the demographic characteristics of the study sample, eating habits, eating at college and healthy habits. The data were collected by interviewing the study sample and the weight and height were measured by a trained researcher at the college. Descriptive statistical analysis was undertaken. Data were prepared, organized and entered into the computer file; the Statistical Package for Social Science (SPSS 20) was used for data analysis. A p value≤ 0.05 was accepted as statistical significant. A total of 63 (62.4%) of the sample were aged20-21with a mean age of 22.1 (SD±0.653). A third of the sample 38 (37.6%) were from level four at college, 67 (66.3%) were female and 46 45.5% of participants were from a middle socio-economic status. 14 (13.9%) of the study sample were overweight (BMI =25-29.9kg/m2) and 6 (5.9%) were obese (BMI≥30kg/m2) compared to 73 (72.3%) were of normal weight (BMI =18.5-24.9kg/m2). With regard to eating habits and exercise, 42 (41.6%) of the students rarely ate breakfast, 79 (78.2%) eat lunch at university residence, 77 (78.2%) of the students reported rarely doing exercise and 62 (61.4%) of them were sleeping for less than eight hours. No significant association was found between the variables age, sex, level of college and socio-economic status and BMI, while there was a significant association between eating lunch at university and BMI (p =0.03). No significant association was found between eating habits, healthy habits and BMI. The prevalence of overweight and obesity among the study sample was 19.8% with female students being more obese than males. Further studies are needed to identify BMI among residence students in other colleges and increasing the awareness of undergraduate students to healthy food habits.
Keywords: Body mass index, diet, obesity, university residence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12803378 Characterization of Fabricated A 384.1-MgO Based Metal Matrix Composite and Optimization of Tensile Strength using Taguchi Techniques
Authors: Nripjit, Anand K Tyagi, Nirmal Singh
Abstract:
The present work consecutively on synthesis and characterization of composites, Al/Al alloy A 384.1 as matrix in which the main ingredient as Al/Al-5% MgO alloy based metal matrix composite. As practical implications the low cost processing route for the fabrication of Al alloy A 384.1 and operational difficulties of presently available manufacturing processes based in liquid manipulation methods. As all new developments, complete understanding of the influence of processing variables upon the final quality of the product. And the composite is applied comprehensively to the acquaintance for achieving superiority of information concerning the specific heat measurement of a material through the aid of thermographs. Products are evaluated concerning relative particle size and mechanical behavior under tensile strength. Furthermore, Taguchi technique was employed to examine the experimental optimum results are achieved, owing to effectiveness of this approach.Keywords: MMC, Thermographs, Tensile strength, Taguchi technique, Optimal parameters
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16403377 Guided Wave Sensitivity for De-Bond Defects in Aluminum Skin-Honeycomb Core
Authors: A. Satour, F. Boubenider, R. Halimi, A. Badidibouda
Abstract:
Sandwich plates are finding an increasing range of application in the aircraft industry. The inspection of honeycomb composite structure by conventional ultrasonic technique is complex and very time consuming. The present study demonstrates a technique using guided Lamb waves at low frequencies to predict de-bond defects in aluminum skin-honeycomb core sandwich structure used in aeronautics. The numerical method was investigated for drawing the dispersion and displacement curves of ultrasonic Lamb wave propagated in Aluminum plate. An experimental study was carried out to check the theoretical prediction. The detection of unsticking between the skin and the core was tested by the two first modes for a low frequency. It was found that A0 mode is more sensitive to delamination defect compared to S0 mode.
Keywords: Damage detection, delamination, guided waves, Sandwich structure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23253376 Integrating Decision Tree and Spatial Cluster Analysis for Landslide Susceptibility Zonation
Authors: Chien-Min Chu, Bor-Wen Tsai, Kang-Tsung Chang
Abstract:
Landslide susceptibility map delineates the potential zones for landslide occurrence. Previous works have applied multivariate methods and neural networks for mapping landslide susceptibility. This study proposed a new approach to integrate decision tree model and spatial cluster statistic for assessing landslide susceptibility spatially. A total of 2057 landslide cells were digitized for developing the landslide decision tree model. The relationships of landslides and instability factors were explicitly represented by using tree graphs in the model. The local Getis-Ord statistics were used to cluster cells with high landslide probability. The analytic result from the local Getis-Ord statistics was classed to create a map of landslide susceptibility zones. The map was validated using new landslide data with 482 cells. Results of validation show an accuracy rate of 86.1% in predicting new landslide occurrence. This indicates that the proposed approach is useful for improving landslide susceptibility mapping.Keywords: Landslide susceptibility Zonation, Decision treemodel, Spatial cluster, Local Getis-Ord statistics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19443375 Price Quoting Method for Contract Manufacturer
Authors: S. Homrossukon, W. Parinyasart
Abstract:
This is an applied research to propose the method for price quotation for a contract electronics manufacturer. It has had a precise price quoting method but such method could not quickly provide a result as the customer required. This reduces the ability of company to compete in this kind of business. In this case, the cause of long time quotation process was analyzed. A lot of product features have been demanded by customer. By checking routine processes, it was found that high fraction of quoting time was used for production time estimating which has effected to the manufacturing or production cost. Then the historical data of products including types, number of components, assembling method, and their assembling time were used to analyze the key components affecting to production time. The price quoting model then was proposed. The implementation of proposed model was able to remarkably reduce quoting time with an acceptable required precision.Keywords: Price quoting, Contract manufacturer, Stepwise technique, Best subset technique.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 44323374 Generating Class-Based Test Cases for Interface Classes of Object-Oriented Gray-Box Frameworks
Authors: Jehad Al Dallal, Paul Sorenson
Abstract:
An application framework provides a reusable design and implementation for a family of software systems. Application developers extend the framework to build their particular applications using hooks. Hooks are the places identified to show how to use and customize the framework. Hooks define Framework Interface Classes (FICs) and their possible specifications, which helps in building reusable test cases for the implementations of these classes. In applications developed using gray-box frameworks, FICs inherit framework classes or use them without inheritance. In this paper, a test-case generation technique is extended to build test cases for FICs built for gray-box frameworks. A tool is developed to automate the introduced technique.Keywords: Class testing, object-oriented framework, reusable test case.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15183373 A Low Cost and High Quality Duty-Cycle Modulation Scheme and Applications
Authors: B. Lonla Moffo, J. Mbihi, L. Nneme Nneme
Abstract:
In this paper, a low cost duty-cycle modulation scheme is studied in depth and compared to the standard pulse width modulation technique. Using a mix of analytical reasoning and electronics simulation tools, it is shown that under the same operating conditions, most characteristics of the proposed duty-cycle modulation scheme are better than those provided by a standard pulse width modulation technique. The simulation results obtained when testing both modulation control policies on prototyping systems, indicate that the proposed duty-cycle modulation approach, appears to be a high quality control policy in a wide variety of application areas, including A/D and D/A conversion, signal transmission and switching control in power electronics.
Keywords: Duty-cycle Modulation, Operational amplifiers, Pulse width modulation, Power electronics, Signal processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27453372 Concept of Transforaminal Lumbar Interbody Fusion Cage Insertion Device
Authors: Sangram A. Sathe, Neha A. Madgulkar, Shruti S. Raut, S. P. Wadkar
Abstract:
Transforaminal lumbar interbody fusion (TLIF) surgeries have nowadays became popular for treatment of degenerated spinal disorders. The interbody fusion technique like TLIF maintains load bearing capacity of the spine and a suitable disc height. Currently many techniques have been introduced to cure Spondylolisthesis. This surgery provides greater rehabilitation of degenerative spines. While performing this TLIF surgery existing methods use guideway, which is a troublesome surgery technique as the use of two separate instruments is required to perform this surgery. This paper presents a concept which eliminates the use of guideway. This concept also eliminates problems that occur like reverting the cage. The concept discussed in this paper also gives high accuracy while performing surgery.Keywords: Degenerative disc diseases, pedicle screw, spine, spondylolisthesis, transforaminal lumbar interbody fusion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13793371 The Optimal Indirect Vector Controller Design via an Adaptive Tabu Search Algorithm
Authors: P. Sawatnatee, S. Udomsuk, K-N. Areerak, K-L. Areerak, A. Srikaew
Abstract:
The paper presents how to design the indirect vector control of three-phase induction motor drive systems using the artificial intelligence technique called the adaptive tabu search. The results from the simulation and the experiment show that the drive system with the controller designed from the proposed method can provide the best output speed response compared with those of the conventional method. The controller design using the proposed technique can be used to create the software package for engineers to achieve the optimal controller design of the induction motor speed control based on the indirect vector concept.
Keywords: Indirect Vector Control, Induction Motor, Adaptive Tabu Search, Control Design, Artificial Intelligence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1933