Search results for: and sample entropy.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1452

Search results for: and sample entropy.

1272 Thermo-Elastic Properties of Artificial Limestone Bricks with Wood Sawdust

Authors: Paki Turgut, Mehmet Gumuscu

Abstract:

In this study, artificial limestone brick samples are produced by using wood sawdust wastes (WSW) having different grades of sizes and limestone powder waste (LPW). The thermo-elastic properties of produced brick samples in various WSW amounts are investigated. At 30% WSW replacement with LPW in the brick sample the thermal conductivity value is effectively reduced and the reduction in the thermal conductivity value of brick sample at 30% WSW replacement with LPW is about 38.9% as compared with control sample. The energy conservation in buildings by using LPW and WSW in masonry brick material production having low thermal conductivity reduces energy requirements. A strong relationship is also found among the thermal conductivity, unit weight and ultrasonic pulse velocity values of brick samples produced. It shows a potential to be used for walls, wooden board substitute, alternative to the concrete blocks, ceiling panels, sound barrier panels, absorption materials etc.

Keywords: Limestone dust, masonry brick, thermo-elastic properties, wood sawdust.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2445
1271 A Modification of Wireless and Internet Technologies for Logistics- Analysis

Authors: Apiwat Sangnoree

Abstract:

This research is designed for helping a WAPbased mobile phone-s user in order to analyze of logistics in the traffic area by applying and designing the accessible processes from mobile user to server databases. The research-s design comprises Mysql 4.1.8-nt database system for being the server which there are three sub-databases, traffic light – times of intersections in periods of the day, distances on the road of area-blocks where are divided from the main sample-area and speeds of sample vehicles (motorcycle, personal car and truck) in periods of the day. For interconnections between the server and user, PHP is used to calculate distances and travelling times from the beginning point to destination, meanwhile XHTML applied for receiving, sending and displaying data from PHP to user-s mobile. In this research, the main sample-area is focused at the Huakwang-Ratchada-s area, Bangkok, Thailand where usually the congested point and 6.25 km2 surrounding area which are split into 25 blocks, 0.25 km2 for each. For simulating the results, the designed server-database and all communicating models of this research have been uploaded to www.utccengineering.com/m4tg and used the mobile phone which supports WAP 2.0 XHTML/HTML multimode browser for observing values and displayed pictures. According to simulated results, user can check the route-s pictures from the requiring point to destination along with analyzed consuming times when sample vehicles travel in various periods of the day.

Keywords: WAP, logistics, XHTML, internet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1398
1270 Using “Eckel” Model to Measure Income Smoothing Practices: The Case of French Companies

Authors: Feddaoui Amina

Abstract:

Income smoothing represents an attempt on the part of the company's management to reduce variations in earnings through the manipulation of the accounting principles. In this study, we aimed to measure income smoothing practices in a sample of 30 French joint stock companies during the period (2007-2009), we used Dummy variables method and “ECKEL” model to measure income smoothing practices and Binomial test accourding to SPSS program, to confirm or refute our hypothesis. This study concluded that there are no significant statistical indicators of income smoothing practices in the sample studied of French companies during the period (2007-2009), so the income series in the same sample studied of is characterized by stability and non-volatility without any intervention of management through accounting manipulation. However, this type of accounting manipulation should be taken into account and efforts should be made by control bodies to apply Eckel model and generalize its use at the global level.

Keywords: Income, smoothing, “Eckel”, French companies.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 946
1269 Coconut Shells as the Alternative Equipment for Foot Reflexology

Authors: Nichanant Sermsri, Chananchida Yuktirat

Abstract:

This research was the experimental research. Its purpose was to find out how coconut shells can be adapted to be equipment for foot and calf reflexology. The sample group was 58 female street vendors in Thewet Market, Bangkok, selected by selection criteria and voluntary. The data collecting tool was the Visual Analogue Scale. The massaging tool made from coconut shells was the key equipment for this research. The research team assessed the level of exhaustion and heart rate among sample group before and after the massage, then analyzed the data by mean, standard deviation and paired sample t-test.

We found out from the research that

  1. The level of exhaustion decreased 4.529 levels after the massage and the standard deviation was 1.6195. The heart rates went down 11.67 times/minute and the standard deviation was 6.742.
  2. The level of exhaustion and heart rate after the massage decreased with the statistically significance at 0.01

Keywords: Coconut Shells, Foot Massage, Foot Reflexology, Massaging Plate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1737
1268 Bitrate Reduction Using FMO for Video Streaming over Packet Networks

Authors: Le Thanh Ha, Hye-Soo Kim, Chun-Su Park, Seung-Won Jung, Sung-Jea Ko

Abstract:

Flexible macroblock ordering (FMO), adopted in the H.264 standard, allows to partition all macroblocks (MBs) in a frame into separate groups of MBs called Slice Groups (SGs). FMO can not only support error-resilience, but also control the size of video packets for different network types. However, it is well-known that the number of bits required for encoding the frame is increased by adopting FMO. In this paper, we propose a novel algorithm that can reduce the bitrate overhead caused by utilizing FMO. In the proposed algorithm, all MBs are grouped in SGs based on the similarity of the transform coefficients. Experimental results show that our algorithm can reduce the bitrate as compared with conventional FMO.

Keywords: Data Partition, Entropy Coding, Greedy Algorithm, H.264/AVC, Slice Group.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1260
1267 Physical-Chemical Surface Characterization of Lake Nasser Sediments

Authors: Yousra M. Zakaria Helmy, Edward H. Smith

Abstract:

Lake Nasser is one of the largest reservoirs in the world. Over 120 million metric tons of sediments are deposited in its dead storage zone every year. The main objective of the present work was to determine the physical and chemical characteristics of Lake Nasser sediments. The sample had a relatively low surface area of 2.9 m2/g which increased more than 3-fold upon chemical activation. The main chemical elements of the raw sediments were C, O and Si with some traces of Al, Fe and Ca. The organic functional groups for the tested sample included O-H, C=C, C-H and C-O, with indications of Si-O and other metal-C and/or metal-O bonds normally associated with clayey materials. Potentiometric titration of the sample in different ionic strength backgrounds revealed an alkaline material with very strong positive surface charge at pH values just a little less than the pH of zero charge which is ~9. Surface interactions of the sediments with the background electrolyte were significant. An advanced surface complexation model was able to capture these effects, employing a single-site approach to represent protolysis reactions in aqueous solution, and to determine the significant surface species in the pH range of environmental interest.

Keywords: Lake Nasser, sediments, surface characterization

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1865
1266 Quantitative Determination of Trace Elements in Some Oriental Herb Products

Authors: Nguyen Thi Kim Dzung, Pham Ngoc Khai, Rainer Ludwig

Abstract:

The quantitative determination of several trace elements (Cr, As, Se, Cd, Hg, Pb) existing as inorganic impurities in some oriental herb-products such as Lingzhi Mushroom capsules, Philamin powder, etc using ICP-MS has been studied. Various instrumental parameters such as power, gas flow rate, sample depth, as well as the concentration of nitric acid and thick background due to high concentration of possible interferences on the determination of these above-mentioned elements was investigated and the optimum working conditions of the sample measurement on ICP-MS (Agilent-7500a) were reported. Appropriate isotope internal standards were also used to improve the accuracy of mercury determination. Optimal parameters for sampling digestion were also investigated. The recovery of analytical procedure was examined by using a Certified Reference Material (IAEA-CRM 359). The recommended procedure was then applied for the quantitative determination of Cr, As, Se, Cd, Hg, Pb in Lingzhi Mushroom capsule, and Philamine powder samples. The reproducibility of sample measurement (average value between 94 and 102%) and the uncertainty of analytical data (less than 20%) are acceptable.

Keywords: Oriental herbal product, trace elements, ICP-MS, biochemistry, medical chemistry.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3666
1265 Comparative Analysis of Concentration in Insurance Markets in New EU Member States

Authors: T. Pavic Kramaric, M. Kitic

Abstract:

The purpose of this article is to analyze the market structure as well as the degree of concentration in insurance markets in new EU member states. The analysis was conducted using several most commonly used concentration indicators such as concentration ratio, Herfindahl-Hirschman index and entropy index. These indicators were calculated for the 2000-2010 period on the basis of total gross written premium as the most relevant indicator of market power in insurance markets. The results of the analysis showed that in all observed countries the level of concentration decreased, though with significantly different intensity. Yet, in some countries, the level of concentration remains very high.

Keywords: insurance market, concentration, new EU member states

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3386
1264 The Role and Effectiveness of Audit Committee in Corporate Governance of Credit Institutions

Authors: Tina Vuko, Marija Maretić, Marko Čular

Abstract:

The aim of this study is to analyze the role and effectiveness of internal mechanism (audit committee) of corporate governance on credit institutions performance in Croatia. Based on research objective, sample of 78 credit institutions listed on Zagreb Stock Exchange, from 2007 to 2012, has been collected and efficiency index of audit committee (EIAC) has been created. Based on the sample and created EIAC, conclusions are as follows: audit committees of credit institutions have medium efficiency, based on EIAC measurement; there is a significant difference in audit committee effectiveness, in observed period; there is no positive relationship between audit committee effectiveness and credit institution performance; there is a significant difference between level of audit committee effectiveness and audit firm type. Future research should contain increased number of elements in EIAC creation and increased sample, for all obligators who need to establish audit committee.

Keywords: Corporate Governance, Audit Committee, Financial Institutions, Efficiency Index of Audit Committee.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2565
1263 Estimation of Bayesian Sample Size for Binomial Proportions Using Areas P-tolerance with Lowest Posterior Loss

Authors: H. Bevrani, N. Najafi

Abstract:

This paper uses p-tolerance with the lowest posterior loss, quadratic loss function, average length criteria, average coverage criteria, and worst outcome criterion for computing of sample size to estimate proportion in Binomial probability function with Beta prior distribution. The proposed methodology is examined, and its effectiveness is shown.

Keywords: Bayesian inference, Beta-binomial Distribution, LPLcriteria, quadratic loss function.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1709
1262 A Modified AES Based Algorithm for Image Encryption

Authors: M. Zeghid, M. Machhout, L. Khriji, A. Baganne, R. Tourki

Abstract:

With the fast evolution of digital data exchange, security information becomes much important in data storage and transmission. Due to the increasing use of images in industrial process, it is essential to protect the confidential image data from unauthorized access. In this paper, we analyze the Advanced Encryption Standard (AES), and we add a key stream generator (A5/1, W7) to AES to ensure improving the encryption performance; mainly for images characterised by reduced entropy. The implementation of both techniques has been realized for experimental purposes. Detailed results in terms of security analysis and implementation are given. Comparative study with traditional encryption algorithms is shown the superiority of the modified algorithm.

Keywords: Cryptography, Encryption, Advanced EncryptionStandard (AES), ECB mode, statistical analysis, key streamgenerator.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4986
1261 Peer-to-Peer Epidemic Algorithms for Reliable Multicasting in Ad Hoc Networks

Authors: Zülküf Genç, Öznur Özkasap

Abstract:

Characteristics of ad hoc networks and even their existence depend on the nodes forming them. Thus, services and applications designed for ad hoc networks should adapt to this dynamic and distributed environment. In particular, multicast algorithms having reliability and scalability requirements should abstain from centralized approaches. We aspire to define a reliable and scalable multicast protocol for ad hoc networks. Our target is to utilize epidemic techniques for this purpose. In this paper, we present a brief survey of epidemic algorithms for reliable multicasting in ad hoc networks, and describe formulations and analytical results for simple epidemics. Then, P2P anti-entropy algorithm for content distribution and our prototype simulation model are described together with our initial results demonstrating the behavior of the algorithm.

Keywords: Ad hoc networks, epidemic, peer-to-peer, reliablemulticast.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1716
1260 Rheological Properties of Dough and Sensory Quality of Crackers with Dietary Fibers

Authors: Ljubica Dokić, Ivana Nikolić, Dragana Šoronja–Simović, Zita Šereš, Biljana Pajin, Nils Juul, Nikola Maravić

Abstract:

The possibility of application the dietary fibers in production of crackers was observed in this work, as well as their influence on rheological and textural properties on the dough for crackers and influence on sensory properties of obtained crackers. Three different dietary fibers, oat, potato and pea fibers, replaced 10% of wheat flour. Long fermentation process and baking test method were used for crackers production. The changes of dough for crackers were observed by rheological methods of determination the viscoelastic dough properties and by textural measurements. Sensory quality of obtained crackers was described using quantity descriptive method (QDA) by trained members of descriptive panel. Additional analysis of crackers surface was performed by videometer. Based on rheological determination, viscoelastic properties of dough for crackers were reduced by application of dietary fibers. Manipulation of dough with 10% of potato fiber was disabled, thus the recipe modification included increase in water content at 35%. Dough compliance to constant stress for samples with dietary fibers decreased, due to more rigid and stiffer dough consistency compared to control sample. Also, hardness of dough for these samples increased and dough extensibility decreased. Sensory properties of final products, crackers, were reduced compared to control sample. Application of dietary fibers affected mostly hardness, structure and crispness of the crackers. Observed crackers were low marked for flavor and taste, due to influence of fibers specific aroma. The sample with 10% of potato fibers and increased water content was the most adaptable to applied stresses and to production process. Also this sample was close to control sample without dietary fibers by evaluation of sensory properties and by results of videometer method.

Keywords: Crackers, dietary fibers, rheology, sensory properties.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2443
1259 Consumer Load Profile Determination with Entropy-Based K-Means Algorithm

Authors: Ioannis P. Panapakidis, Marios N. Moschakis

Abstract:

With the continuous increment of smart meter installations across the globe, the need for processing of the load data is evident. Clustering-based load profiling is built upon the utilization of unsupervised machine learning tools for the purpose of formulating the typical load curves or load profiles. The most commonly used algorithm in the load profiling literature is the K-means. While the algorithm has been successfully tested in a variety of applications, its drawback is the strong dependence in the initialization phase. This paper proposes a novel modified form of the K-means that addresses the aforementioned problem. Simulation results indicate the superiority of the proposed algorithm compared to the K-means.

Keywords: Clustering, load profiling, load modeling, machine learning, energy efficiency and quality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1144
1258 A Physical Theory of Information vs. a Mathematical Theory of Communication

Authors: Manouchehr Amiri

Abstract:

This article presents a general notion of physical bit information that is compatible with the basics of quantum mechanics and incorporates the Shannon entropy as a special case. This notion of physical information leads to the Binary Data Matrix model (BDM), which predicts the basic results of quantum mechanics, general relativity, and black hole thermodynamics. The compatibility of the model with holographic, information conservation, and Landauer’s principle is investigated. After deriving the “Bit Information principle” as a consequence of BDM, the fundamental equations of Planck, De Broglie, Bekenstein, and mass-energy equivalence are derived.

Keywords: Physical theory of information, binary data matrix model, Shannon information theory, bit information principle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 43
1257 Analysis of Creative City Indicators in Isfahan City, Iran

Authors: Reza Mokhtari Malek Abadi, Mohsen Saghaei, Fatima Iman

Abstract:

This paper investigates the indices of a creative city in Isfahan. Its main aim is to evaluate quantitative status of the creative city indices in Isfahan city, analyze the dispersion and distribution of these indices in Isfahan city. Concerning these, this study tries to analyze the creative city indices in fifteen area of Isfahan through secondary data, questionnaire, TOPSIS model, Shannon entropy and SPSS. Based on this, the fifteen areas of Isfahan city have been ranked with 12 factors of creative city indices. The results of studies show that fifteen areas of Isfahan city are not equally benefiting from creative indices and there is much difference between the areas of Isfahan city.

Keywords: Grading, creative city, creative city evaluation indicators, regional planning model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2635
1256 Experimental Investigation on Effect of the Zirconium + Magnesium Coating of the Piston and Valve of the Single-Cylinder Diesel Engine to the Engine Performance and Emission

Authors: Erdinç Vural, Bülent Özdalyan, Serkan Özel

Abstract:

The four-stroke single cylinder diesel engine has been used in this study, the pistons and valves of the engine have been stabilized, the aluminum oxide (Al2O3) in different ratios has been added in the power of zirconium (ZrO2) magnesium oxide (MgO), and has been coated with the plasma spray method. The pistons and valves of the combustion chamber of the engine are coated with 5 different (ZrO2 + MgO), (ZrO2 + MgO + 25% Al2O3), (ZrO2 + MgO + 50% Al2O3), (ZrO2 + MgO + 75% Al2O3), (Al2O3) sample. The material tests have been made for each of the coated engine parts with the scanning electron microscopy (SEM), energy dispersive X-ray spectroscopy (EDX) and X-ray diffraction (XRD) using Cu Kα radiation surface analysis methods. The engine tests have been repeated for each sample in any electric dynamometer in full power 1600 rpm, 2000 rpm, 2400 rpm and 2800 rpm engine speeds. The material analysis and engine tests have shown that the best performance has been performed with (ZrO2 + MgO + 50% Al2O3). Thus, there is no significant change in HC and Smoke emissions, but NOx emission is increased, as the engine improves power, torque, specific fuel consumption and CO emissions in the tests made with sample A3.

Keywords: Ceramic coating, material characterization, engine performance, exhaust emissions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1599
1255 Optimizing the Number of Bits/Stage in 10-Bit, 50Ms/Sec Pipelined A/D Converter Considering Area, Speed, Power and Linearity

Authors: P. Prasad Rao, K. Lal Kishore

Abstract:

Pipeline ADCs are becoming popular at high speeds and with high resolution. This paper discusses the options of number of bits/stage conversion techniques in pipelined ADCs and their effect on Area, Speed, Power Dissipation and Linearity. The basic building blocks like op-amp, Sample and Hold Circuit, sub converter, DAC, Residue Amplifier used in every stage is assumed to be identical. The sub converters use flash architectures. The design is implemented using 0.18

Keywords: 1.5 bits/stage, Conversion Frequency, Redundancy Switched Capacitor Sample and Hold Circuit

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1679
1254 Efficient Lossless Compression of Weather Radar Data

Authors: Wei-hua Ai, Wei Yan, Xiang Li

Abstract:

Data compression is used operationally to reduce bandwidth and storage requirements. An efficient method for achieving lossless weather radar data compression is presented. The characteristics of the data are taken into account and the optical linear prediction is used for the PPI images in the weather radar data in the proposed method. The next PPI image is identical to the current one and a dramatic reduction in source entropy is achieved by using the prediction algorithm. Some lossless compression methods are used to compress the predicted data. Experimental results show that for the weather radar data, the method proposed in this paper outperforms the other methods.

Keywords: Lossless compression, weather radar data, optical linear prediction, PPI image

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2204
1253 Enhance Performance of Secure Image Using Wavelet Compression

Authors: Goh Han Keat, Azman Samsudin Zurinahni Zainol

Abstract:

The increase popularity of multimedia application especially in image processing places a great demand on efficient data storage and transmission techniques. Network communication such as wireless network can easily be intercepted and cause of confidential information leaked. Unfortunately, conventional compression and encryption methods are too slow; it is impossible to carry out real time secure image processing. In this research, Embedded Zerotree Wavelet (EZW) encoder which specially designs for wavelet compression is examined. With this algorithm, three methods are proposed to reduce the processing time, space and security protection that will be secured enough to protect the data.

Keywords: Embedded Zerotree Wavelet (EZW), Imagecompression, Wavelet encoder, Entropy encoder, Encryption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1631
1252 An Estimating Parameter of the Mean in Normal Distribution by Maximum Likelihood, Bayes, and Markov Chain Monte Carlo Methods

Authors: Autcha Araveeporn

Abstract:

This paper is to compare the parameter estimation of the mean in normal distribution by Maximum Likelihood (ML), Bayes, and Markov Chain Monte Carlo (MCMC) methods. The ML estimator is estimated by the average of data, the Bayes method is considered from the prior distribution to estimate Bayes estimator, and MCMC estimator is approximated by Gibbs sampling from posterior distribution. These methods are also to estimate a parameter then the hypothesis testing is used to check a robustness of the estimators. Data are simulated from normal distribution with the true parameter of mean 2, and variance 4, 9, and 16 when the sample sizes is set as 10, 20, 30, and 50. From the results, it can be seen that the estimation of MLE, and MCMC are perceivably different from the true parameter when the sample size is 10 and 20 with variance 16. Furthermore, the Bayes estimator is estimated from the prior distribution when mean is 1, and variance is 12 which showed the significant difference in mean with variance 9 at the sample size 10 and 20.

Keywords: Bayes method, Markov Chain Monte Carlo method, Maximum Likelihood method, normal distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1380
1251 Implementation of RC5 Block Cipher Algorithm for Image Cryptosystems

Authors: Hossam El-din H. Ahmed, Hamdy M. Kalash, Osama S. Farag Allah

Abstract:

This paper examines the implementation of RC5 block cipher for digital images along with its detailed security analysis. A complete specification for the method of application of the RC5 block cipher to digital images is given. The security analysis of RC5 block cipher for digital images against entropy attack, bruteforce, statistical, and differential attacks is explored from strict cryptographic viewpoint. Experiments and results verify and prove that RC5 block cipher is highly secure for real-time image encryption from cryptographic viewpoint. Thorough experimental tests are carried out with detailed analysis, demonstrating the high security of RC5 block cipher algorithm.

Keywords: Image encryption, security analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3617
1250 Feature Level Fusion of Multimodal Images Using Haar Lifting Wavelet Transform

Authors: Sudipta Majumdar, Jayant Bharadwaj

Abstract:

This paper presents feature level image fusion using Haar lifting wavelet transform. Feature fused is edge and boundary information, which is obtained using wavelet transform modulus maxima criteria. Simulation results show the superiority of the result as entropy, gradient, standard deviation are increased for fused image as compared to input images. The proposed methods have the advantages of simplicity of implementation, fast algorithm, perfect reconstruction, and reduced computational complexity. (Computational cost of Haar wavelet is very small as compared to other lifting wavelets.)

Keywords: Lifting wavelet transform, wavelet transform modulus maxima.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2373
1249 Adjusted Ratio and Regression Type Estimators for Estimation of Population Mean when some Observations are missing

Authors: Nuanpan Nangsue

Abstract:

Ratio and regression type estimators have been used by previous authors to estimate a population mean for the principal variable from samples in which both auxiliary x and principal y variable data are available. However, missing data are a common problem in statistical analyses with real data. Ratio and regression type estimators have also been used for imputing values of missing y data. In this paper, six new ratio and regression type estimators are proposed for imputing values for any missing y data and estimating a population mean for y from samples with missing x and/or y data. A simulation study has been conducted to compare the six ratio and regression type estimators with a previous estimator of Rueda. Two population sizes N = 1,000 and 5,000 have been considered with sample sizes of 10% and 30% and with correlation coefficients between population variables X and Y of 0.5 and 0.8. In the simulations, 10 and 40 percent of sample y values and 10 and 40 percent of sample x values were randomly designated as missing. The new ratio and regression type estimators give similar mean absolute percentage errors that are smaller than the Rueda estimator for all cases. The new estimators give a large reduction in errors for the case of 40% missing y values and sampling fraction of 30%.

Keywords: Auxiliary variable, missing data, ratio and regression type estimators.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1686
1248 Presenting a Combinatorial Feature to Estimate Depth of Anesthesia

Authors: Toktam Zoughi, Reza Boostani

Abstract:

Determining depth of anesthesia is a challenging problem in the context of biomedical signal processing. Various methods have been suggested to determine a quantitative index as depth of anesthesia, but most of these methods suffer from high sensitivity during the surgery. A novel method based on energy scattering of samples in the wavelet domain is suggested to represent the basic content of electroencephalogram (EEG) signal. In this method, first EEG signal is decomposed into different sub-bands, then samples are squared and energy of samples sequence is constructed through each scale and time, which is normalized and finally entropy of the resulted sequences is suggested as a reliable index. Empirical Results showed that applying the proposed method to the EEG signals can classify the awake, moderate and deep anesthesia states similar to BIS.

Keywords: Depth of anesthesia, EEG, BIS, Wavelet transforms.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1794
1247 FPGA Implementation of a Vision-Based Blind Spot Warning System

Authors: Yu Ren Lin, Yu Hong Li

Abstract:

Vision-based intelligent vehicle applications often require large amounts of memory to handle video streaming and image processing, which in turn increases complexity of hardware and software. This paper presents an FPGA implement of a vision-based blind spot warning system. Using video frames, the information of the blind spot area turns into one-dimensional information. Analysis of the estimated entropy of image allows the detection of an object in time. This idea has been implemented in the XtremeDSP video starter kit. The blind spot warning system uses only 13% of its logic resources and 95k bits block memory, and its frame rate is over 30 frames per sec (fps).

Keywords: blind-spot area, image, FPGA

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1788
1246 Fuzzy Clustering of Categorical Attributes and its Use in Analyzing Cultural Data

Authors: George E. Tsekouras, Dimitris Papageorgiou, Sotiris Kotsiantis, Christos Kalloniatis, Panagiotis Pintelas

Abstract:

We develop a three-step fuzzy logic-based algorithm for clustering categorical attributes, and we apply it to analyze cultural data. In the first step the algorithm employs an entropy-based clustering scheme, which initializes the cluster centers. In the second step we apply the fuzzy c-modes algorithm to obtain a fuzzy partition of the data set, and the third step introduces a novel cluster validity index, which decides the final number of clusters.

Keywords: Categorical data, cultural data, fuzzy logic clustering, fuzzy c-modes, cluster validity index.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1662
1245 A Hybrid Feature Selection by Resampling, Chi squared and Consistency Evaluation Techniques

Authors: Amir-Massoud Bidgoli, Mehdi Naseri Parsa

Abstract:

In this paper a combined feature selection method is proposed which takes advantages of sample domain filtering, resampling and feature subset evaluation methods to reduce dimensions of huge datasets and select reliable features. This method utilizes both feature space and sample domain to improve the process of feature selection and uses a combination of Chi squared with Consistency attribute evaluation methods to seek reliable features. This method consists of two phases. The first phase filters and resamples the sample domain and the second phase adopts a hybrid procedure to find the optimal feature space by applying Chi squared, Consistency subset evaluation methods and genetic search. Experiments on various sized datasets from UCI Repository of Machine Learning databases show that the performance of five classifiers (Naïve Bayes, Logistic, Multilayer Perceptron, Best First Decision Tree and JRIP) improves simultaneously and the classification error for these classifiers decreases considerably. The experiments also show that this method outperforms other feature selection methods.

Keywords: feature selection, resampling, reliable features, Consistency Subset Evaluation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2538
1244 High-Temperature X-Ray Powder Diffraction of Secondary Gypsum

Authors: D. Gazdič, I. Hájková, M. Fridrichová

Abstract:

This paper involved the performance of a hightemperature X-Ray powder diffraction analysis (XRD) of a sample of chemical gypsum generated in the production of titanium white; this gypsum originates by neutralizing highly acidic water with limestone suspension. Specifically, it was gypsum formed in the first stage of neutralization when the resulting material contains, apart from gypsum, a number of waste products resulting from the decomposition of ilmenite by sulphuric acid. So it can be described as red titanogypsum. By conducting the experiment using XRD apparatus Bruker D8 Advance with a Cu anode (λkα=1.54184 Å) equipped with high-temperature chamber Anton Paar HTK 16, it was possible to identify clearly in the sample each phase transition in the system of CaSO4·xH2O.

Keywords: Anhydrite, Gypsum, Bassanite, Hematite, XRD, Powder, High-Temperature.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2028
1243 Classification of Prostate Cell Nuclei using Artificial Neural Network Methods

Authors: M. Sinecen, M. Makinacı

Abstract:

The purpose of this paper is to assess the value of neural networks for classification of cancer and noncancer prostate cells. Gauss Markov Random Fields, Fourier entropy and wavelet average deviation features are calculated from 80 noncancer and 80 cancer prostate cell nuclei. For classification, artificial neural network techniques which are multilayer perceptron, radial basis function and learning vector quantization are used. Two methods are utilized for multilayer perceptron. First method has single hidden layer and between 3-15 nodes, second method has two hidden layer and each layer has between 3-15 nodes. Overall classification rate of 86.88% is achieved.

Keywords: Artificial neural networks, texture classification, cancer diagnosis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1536