Search results for: nonlinear statistical methods
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5865

Search results for: nonlinear statistical methods

3555 A new Adaptive Approach for Histogram based Mouth Segmentation

Authors: Axel Panning, Robert Niese, Ayoub Al-Hamadi, Bernd Michaelis

Abstract:

The segmentation of mouth and lips is a fundamental problem in facial image analyisis. In this paper we propose a method for lip segmentation based on rg-color histogram. Statistical analysis shows, using the rg-color-space is optimal for this purpose of a pure color based segmentation. Initially a rough adaptive threshold selects a histogram region, that assures that all pixels in that region are skin pixels. Based on that pixels we build a gaussian model which represents the skin pixels distribution and is utilized to obtain a refined, optimal threshold. We are not incorporating shape or edge information. In experiments we show the performance of our lip pixel segmentation method compared to the ground truth of our dataset and a conventional watershed algorithm.

Keywords: Feature extraction, Segmentation, Image processing, Application

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1793
3554 Cross Signal Identification for PSG Applications

Authors: Carmen Grigoraş, Victor Grigoraş, Daniela Boişteanu

Abstract:

The standard investigational method for obstructive sleep apnea syndrome (OSAS) diagnosis is polysomnography (PSG), which consists of a simultaneous, usually overnight recording of multiple electro-physiological signals related to sleep and wakefulness. This is an expensive, encumbering and not a readily repeated protocol, and therefore there is need for simpler and easily implemented screening and detection techniques. Identification of apnea/hypopnea events in the screening recordings is the key factor for the diagnosis of OSAS. The analysis of a solely single-lead electrocardiographic (ECG) signal for OSAS diagnosis, which may be done with portable devices, at patient-s home, is the challenge of the last years. A novel artificial neural network (ANN) based approach for feature extraction and automatic identification of respiratory events in ECG signals is presented in this paper. A nonlinear principal component analysis (NLPCA) method was considered for feature extraction and support vector machine for classification/recognition. An alternative representation of the respiratory events by means of Kohonen type neural network is discussed. Our prospective study was based on OSAS patients of the Clinical Hospital of Pneumology from Iaşi, Romania, males and females, as well as on non-OSAS investigated human subjects. Our computed analysis includes a learning phase based on cross signal PSG annotation.

Keywords: Artificial neural networks, feature extraction, obstructive sleep apnea syndrome, pattern recognition, signalprocessing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1548
3553 Analytical and Experimental Methods of Design for Supersonic Two-Stage Ejectors

Authors: S. Daneshmand, C. Aghanajafi, A. Bahrami

Abstract:

In this paper the supersonic ejectors are experimentally and analytically studied. Ejector is a device that uses the energy of a fluid to move another fluid. This device works like a vacuum pump without usage of piston, rotor or any other moving component. An ejector contains an active nozzle, a passive nozzle, a mixing chamber and a diffuser. Since the fluid viscosity is large, and the flow is turbulent and three dimensional in the mixing chamber, the numerical methods consume long time and high cost to analyze the flow in ejectors. Therefore this paper presents a simple analytical method that is based on the precise governing equations in fluid mechanics. According to achieved analytical relations, a computer code has been prepared to analyze the flow in different components of the ejector. An experiment has been performed in supersonic regime 1.5Keywords: Ejector, Wind Tunnel, Supersonic, Diffuser, Machnumber, Mixing Chamber

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3221
3552 A Retrievable Genetic Algorithm for Efficient Solving of Sudoku Puzzles

Authors: Seyed Mehran Kazemi, Bahare Fatemi

Abstract:

Sudoku is a logic-based combinatorial puzzle game which is popular among people of different ages. Due to this popularity, computer softwares are being developed to generate and solve Sudoku puzzles with different levels of difficulty. Several methods and algorithms have been proposed and used in different softwares to efficiently solve Sudoku puzzles. Various search methods such as stochastic local search have been applied to this problem. Genetic Algorithm (GA) is one of the algorithms which have been applied to this problem in different forms and in several works in the literature. In these works, chromosomes with little or no information were considered and obtained results were not promising. In this paper, we propose a new way of applying GA to this problem which uses more-informed chromosomes than other works in the literature. We optimize the parameters of our GA using puzzles with different levels of difficulty. Then we use the optimized values of the parameters to solve various puzzles and compare our results to another GA-based method for solving Sudoku puzzles.

Keywords: Genetic algorithm, optimization, solving Sudoku puzzles, stochastic local search.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3781
3551 Peakwise Smoothing of Data Models using Wavelets

Authors: D Sudheer Reddy, N Gopal Reddy, P V Radhadevi, J Saibaba, Geeta Varadan

Abstract:

Smoothing or filtering of data is first preprocessing step for noise suppression in many applications involving data analysis. Moving average is the most popular method of smoothing the data, generalization of this led to the development of Savitzky-Golay filter. Many window smoothing methods were developed by convolving the data with different window functions for different applications; most widely used window functions are Gaussian or Kaiser. Function approximation of the data by polynomial regression or Fourier expansion or wavelet expansion also gives a smoothed data. Wavelets also smooth the data to great extent by thresholding the wavelet coefficients. Almost all smoothing methods destroys the peaks and flatten them when the support of the window is increased. In certain applications it is desirable to retain peaks while smoothing the data as much as possible. In this paper we present a methodology called as peak-wise smoothing that will smooth the data to any desired level without losing the major peak features.

Keywords: smoothing, moving average, peakwise smoothing, spatialdensity models, planar shape models, wavelets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1756
3550 Revisiting Domestication and Foreignisation Methods: Translating the Quran by the Hybrid Approach

Authors: Aladdin Al-Tarawneh

Abstract:

The Quran, as it is the sacred book of Islam and considered the literal word of God (Allah) in Arabic, is highly translated into many languages; however, the foreignising or the literal approach excessively stains the quality and discredits the final product in the eyes of its receptors. Such an approach fails to capture the intended meaning of the Quran and to communicate it in any language. Therefore, this study is conducted to propose a different approach that seeks involving other ones according to a hybrid model. Indeed, this study challenges the binary adherence that is highly used in Translation Studies (TS) in general and in the translation of the Quran in particular. Drawing on the genuine fact that the Quran can be communicated in any language in terms of meaning, and the translation is not sacred; this paper approaches the translation of the Quran by blending different methods like domestication or foreignisation in a systematic way, avoiding the binary choice made by many translators. To reach this aim, the paper has a conceptual part that seeks to elucidate and clarify the main methods employed in TS, and criticise and modify them to propose the new hybrid approach (the hybrid model) for translating the Quran – that is, the deductive method. To support and validate the outcome of the previous part, a comparative model is employed in order to highlight the differences between the suggested translation and other widely used ones – that is, the inductive method. By applying this methodology, the paper proves that there is a deficiency of communicating the original meaning of the Quran in light of the foreignising approach. In conclusion, the paper suggests producing a Quran translation has to take into account the adoption of many techniques to express the meaning of the Quran as understood in the original, and to offer this understanding in English in the most native-like manner to serve the intended target readers.

Keywords: Quran translation, hybrid approach, domestication, foreignisation, hybrid model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1204
3549 Effectiveness of Business Software Systems Development and Enhancement Projects versus Work Effort Estimation Methods

Authors: Beata Czarnacka-Chrobot

Abstract:

Execution of Business Software Systems (BSS) Development and Enhancement Projects (D&EP) is characterized by the exceptionally low effectiveness, leading to considerable financial losses. The general reason for low effectiveness of such projects is that they are inappropriately managed. One of the factors of proper BSS D&EP management is suitable (reliable and objective) method of project work effort estimation since this is what determines correct estimation of its major attributes: project cost and duration. BSS D&EP is usually considered to be accomplished effectively if product of a planned functionality is delivered without cost and time overrun. The goal of this paper is to prove that choosing approach to the BSS D&EP work effort estimation has a considerable influence on the effectiveness of such projects execution.

Keywords: Business software systems, development and enhancement projects, effectiveness, work effort estimation methods, software product size, software product functionality, project duration, project cost.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2092
3548 A Study of Relationship between Mountaineering Participation Motivation and Risk Perception

Authors: Yen-Chieh Wen, Ching-Hui Lin

Abstract:

The main purpose of this study is to analyze climbers involved in motivation and risk perception and analysis of the predictive ability of the risk perception "mountaineering" involved in motivation. This study used questionnaires, to have to climb the 3000m high mountain in Taiwan climbers object to carry out an investigation in order to non-random sampling, a total of 231 valid questionnaires were. After statistical analysis, the study found that: 1. Climbers the highest climbers involved in motivation "to enjoy the natural beauty of the fun. 2 climbers for climbers "risk perception" the highest: the natural environment of risk. 3. Climbers “seeking adventure stimulate", “competence achievement" motivation highly predictive of risk perception. Based on these findings, this study not only practices the recommendations of the outdoor leisure industry, and also related research proposals for future researchers.

Keywords: Mountaineering, motivation, risk perception, decision-making.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2867
3547 Integrating the Theory of Constraints and Six Sigma in Manufacturing Process Improvement

Authors: Kai Jin, Hyder Abdul-Razzak, Yousri Elkassabgi, Hong Zhou, Aaron Herrera

Abstract:

Six Sigma is a well known discipline that reduces variation using complex statistical tools and the DMAIC model. By integrating Goldratts-s Theory of Constraints, the Five Focusing Points and System Thinking tools, Six Sigma projects can be selected where it can cause more impact in the company. This research defines an integrated model of six sigma and constraint management that shows a step-by-step guide using the original methodologies from each discipline and is evaluated in a case study from the production line of a Automobile engine monoblock V8, resulting in an increase in the line capacity from 18.7 pieces per hour to 22.4 pieces per hour, a reduction of 60% of Work-In-Process and a variation decrease of 0.73%.

Keywords: Constraint Management, Manufacturing Process Improvement, Six Sigma, System Thinking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1775
3546 CMOS Solid-State Nanopore DNA System-Level Sequencing Techniques Enhancement

Authors: Syed Islam, Yiyun Huang, Sebastian Magierowski, Ebrahim Ghafar-Zadeh

Abstract:

This paper presents system level CMOS solid-state nanopore techniques enhancement for speedup next generation molecular recording and high throughput channels. This discussion also considers optimum number of base-pair (bp) measurements through channel as an important role to enhance potential read accuracy. Effective power consumption estimation offered suitable range of multi-channel configuration. Nanopore bp extraction model in statistical method could contribute higher read accuracy with longer read-length (200 < read-length). Nanopore ionic current switching with Time Multiplexing (TM) based multichannel readout system contributed hardware savings.

Keywords: DNA, Nanopore, Amplifier, ADC, Multichannel.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2938
3545 The Feedback Control for Distributed Systems

Authors: Kamil Aida-zade, C. Ardil

Abstract:

We study the problem of synthesis of lumped sources control for the objects with distributed parameters on the basis of continuous observation of phase state at given points of object. In the proposed approach the phase state space (phase space) is beforehand somehow partitioned at observable points into given subsets (zones). The synthesizing control actions therewith are taken from the class of piecewise constant functions. The current values of control actions are determined by the subset of phase space that contains the aggregate of current states of object at the observable points (in these states control actions take constant values). In the paper such synthesized control actions are called zone control actions. A technique to obtain optimal values of zone control actions with the use of smooth optimization methods is given. With this aim, the formulas of objective functional gradient in the space of zone control actions are obtained.

Keywords: Feedback control, distributed systems, smooth optimization methods, lumped control synthesis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 623
3544 Determining Full Stage Creep Properties from Miniature Specimen Creep Test

Authors: W. Sun, W. Wen, J. Lu, A. A. Becker

Abstract:

In this work, methods for determining creep properties which can be used to represent the full life until failure from miniature specimen creep tests based on analytical solutions are presented. Examples used to demonstrate the application of the methods include a miniature rectangular thin beam specimen creep test under three-point bending and a miniature two-material tensile specimen creep test subjected to a steady load. Mathematical expressions for deflection and creep strain rate of the two specimens were presented for the Kachanov-Rabotnov creep damage model. On this basis, an inverse procedure was developed which has potential applications for deriving the full life creep damage constitutive properties from a very small volume of material, in particular, for various microstructure constitutive  regions, e.g. within heat-affected zones of power plant pipe weldments. Further work on validation and improvement of the method is addressed.

Keywords: Creep damage property, analytical solutions, inverse approach, miniature specimen test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 789
3543 Text Summarization for Oil and Gas Drilling Topic

Authors: Y. Y. Chen, O. M. Foong, S. P. Yong, Kurniawan Iwan

Abstract:

Information sharing and gathering are important in the rapid advancement era of technology. The existence of WWW has caused rapid growth of information explosion. Readers are overloaded with too many lengthy text documents in which they are more interested in shorter versions. Oil and gas industry could not escape from this predicament. In this paper, we develop an Automated Text Summarization System known as AutoTextSumm to extract the salient points of oil and gas drilling articles by incorporating statistical approach, keywords identification, synonym words and sentence-s position. In this study, we have conducted interviews with Petroleum Engineering experts and English Language experts to identify the list of most commonly used keywords in the oil and gas drilling domain. The system performance of AutoTextSumm is evaluated using the formulae of precision, recall and F-score. Based on the experimental results, AutoTextSumm has produced satisfactory performance with F-score of 0.81.

Keywords: Keyword's probability, synonym sets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1738
3542 Rotor Side Speed Control Methods Using MATLAB/Simulink for Wound Induction Motor

Authors: Rajesh Kumar, Roopali Dogra, Puneet Aggarwal

Abstract:

In recent advancements in electric machine and drives, wound rotor motor is extensively used. The merit of using wound rotor induction motor is to control speed/torque characteristics by inserting external resistance. Wound rotor induction motor can be used in the cases such as (a) low inrush current, (b) load requiring high starting torque, (c) lower starting current is required, (d) loads having high inertia, and (e) gradual built up of torque. Examples include conveyers, cranes, pumps, elevators, and compressors. This paper includes speed control of wound induction motor using MATLAB/Simulink for rotor resistance and slip power recovery method. The characteristics of these speed control methods are hence analyzed.

Keywords: Wound rotor induction motor, MATLAB/Simulink, rotor resistance method, slip power recovery method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2425
3541 Real-Time Image Analysis of Capsule Endoscopy for Bleeding Discrimination in Embedded System Platform

Authors: Yong-Gyu Lee, Gilwon Yoon

Abstract:

Image processing for capsule endoscopy requires large memory and it takes hours for diagnosis since operation time is normally more than 8 hours. A real-time analysis algorithm of capsule images can be clinically very useful. It can differentiate abnormal tissue from health structure and provide with correlation information among the images. Bleeding is our interest in this regard and we propose a method of detecting frames with potential bleeding in real-time. Our detection algorithm is based on statistical analysis and the shapes of bleeding spots. We tested our algorithm with 30 cases of capsule endoscopy in the digestive track. Results were excellent where a sensitivity of 99% and a specificity of 97% were achieved in detecting the image frames with bleeding spots.

Keywords: bleeding, capsule endoscopy, image processing, real time analysis

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1887
3540 Tracking Control of a Linear Parabolic PDE with In-domain Point Actuators

Authors: Amir Badkoubeh, Guchuan Zhu

Abstract:

This paper addresses the problem of asymptotic tracking control of a linear parabolic partial differential equation with indomain point actuation. As the considered model is a non-standard partial differential equation, we firstly developed a map that allows transforming this problem into a standard boundary control problem to which existing infinite-dimensional system control methods can be applied. Then, a combination of energy multiplier and differential flatness methods is used to design an asymptotic tracking controller. This control scheme consists of stabilizing state-feedback derived from the energy multiplier method and feed-forward control based on the flatness property of the system. This approach represents a systematic procedure to design tracking control laws for a class of partial differential equations with in-domain point actuation. The applicability and system performance are assessed by simulation studies.

Keywords: Tracking Control, In-domain point actuation, PartialDifferential Equations.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2064
3539 Factors Affecting the Work Efficiency of Employees of Suan Sunandha Rajabhat University

Authors: Unnop Panpuang

Abstract:

The objectives of this project are to study on the work efficiency of the employees, sorted by their profiles, and to study on the relation between job attributes and work efficiency of employees of Suan Sunandha Rajabhat University. The samples used for this study are 292 employees. The statistics used in this study are frequencies, standard deviations, One-way ANOVA and Pearson’s correlation coefficient. Majority of respondent were male with an undergraduate degree, married and lives together. The average age of respondents was between 31-41 years old, married and the educational background are higher than bachelor’s degree. The job attribute is correlated to the work efficiency with the statistical significance level of.o1. This concurs with the predetermined hypothesis. The correlation between the two main factors is in the moderate level. All the categories of job attributes such as the variety of skills, job clarity, job importance, freedom to do work are considered separately.

Keywords: Employees, Job Attributes, Work Efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3659
3538 Slovenian Spatial Legislation over Time and Its Issues

Authors: Andreja Benko

Abstract:

Article presents a short overview of the architects’ profession over time with outlined work of the architectural theoreticians. In the continuation is described a former affiliation of Slovenia as well as the spatial planning documents that were in use until the Slovenia joint Yugoslavia (last part in 1919). This legislation from former Austro-Hungarian monarchy was valid almost until 1950 in some parts of Yugoslavia even longer. Upon that will be mentioned some valid Slovenian spatial documents which will be compared with the German legislation. Analyzed will be the number of architect and spatial planners in Slovenia and also their number upon certain region in Slovenia. Based on that will be given also the number from statistical office of Slovenia of the number of buildings between years 2007 and 2012, and described also the collapse of the major construction companies in Slovenia and consequences of that. At the end will be outlined the morality and ethics by spatial interventions and lack of the architectural law in Slovenia as well as the problematic of minimal collaboration between the Ministry of infrastructure and spatial planning with the profession.

Keywords: Architect, history, legislation, Slovenia.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1680
3537 Effect of Peak-to-Average Power Ratio Reduction on the Multicarrier Communication System Performance Parameters

Authors: Sanjay Singh, M Sathish Kumar, H. S Mruthyunjaya

Abstract:

Multicarrier transmission system such as Orthogonal Frequency Division Multiplexing (OFDM) is a promising technique for high bit rate transmission in wireless communication system. OFDM is a spectrally efficient modulation technique that can achieve high speed data transmission over multipath fading channels without the need for powerful equalization techniques. However the price paid for this high spectral efficiency and less intensive equalization is low power efficiency. OFDM signals are very sensitive to nonlinear effects due to the high Peak-to-Average Power Ratio (PAPR), which leads to the power inefficiency in the RF section of the transmitter. This paper investigates the effect of PAPR reduction on the performance parameter of multicarrier communication system. Performance parameters considered are power consumption of Power Amplifier (PA) and Digital-to-Analog Converter (DAC), power amplifier efficiency, SNR of DAC and BER performance of the system. From our analysis it is found that irrespective of PAPR reduction technique being employed, the power consumption of PA and DAC reduces and power amplifier efficiency increases due to reduction in PAPR. Moreover, it has been shown that for a given BER performance the requirement of Input-Backoff (IBO) reduces with reduction in PAPR.

Keywords: BER, Crest Factor (CF), Digital-to-Analog Converter(DAC), Input-Backoff (IBO), Orthogonal Frequency Division Multiplexing(OFDM), Peak-to-Average Power Ratio (PAPR), PowerAmplifier efficiency, SNR

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3294
3536 One-Class Support Vector Machines for Aerial Images Segmentation

Authors: Chih-Hung Wu, Chih-Chin Lai, Chun-Yen Chen, Yan-He Chen

Abstract:

Interpretation of aerial images is an important task in various applications. Image segmentation can be viewed as the essential step for extracting information from aerial images. Among many developed segmentation methods, the technique of clustering has been extensively investigated and used. However, determining the number of clusters in an image is inherently a difficult problem, especially when a priori information on the aerial image is unavailable. This study proposes a support vector machine approach for clustering aerial images. Three cluster validity indices, distance-based index, Davies-Bouldin index, and Xie-Beni index, are utilized as quantitative measures of the quality of clustering results. Comparisons on the effectiveness of these indices and various parameters settings on the proposed methods are conducted. Experimental results are provided to illustrate the feasibility of the proposed approach.

Keywords: Aerial imaging, image segmentation, machine learning, support vector machine, cluster validity index

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1945
3535 A Novel Prediction Method for Tag SNP Selection using Genetic Algorithm based on KNN

Authors: Li-Yeh Chuang, Yu-Jen Hou, Jr., Cheng-Hong Yang

Abstract:

Single nucleotide polymorphisms (SNPs) hold much promise as a basis for disease-gene association. However, research is limited by the cost of genotyping the tremendous number of SNPs. Therefore, it is important to identify a small subset of informative SNPs, the so-called tag SNPs. This subset consists of selected SNPs of the genotypes, and accurately represents the rest of the SNPs. Furthermore, an effective evaluation method is needed to evaluate prediction accuracy of a set of tag SNPs. In this paper, a genetic algorithm (GA) is applied to tag SNP problems, and the K-nearest neighbor (K-NN) serves as a prediction method of tag SNP selection. The experimental data used was taken from the HapMap project; it consists of genotype data rather than haplotype data. The proposed method consistently identified tag SNPs with considerably better prediction accuracy than methods from the literature. At the same time, the number of tag SNPs identified was smaller than the number of tag SNPs in the other methods. The run time of the proposed method was much shorter than the run time of the SVM/STSA method when the same accuracy was reached.

Keywords: Genetic Algorithm (GA), Genotype, Single nucleotide polymorphism (SNP), tag SNPs.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1775
3534 Feature Reduction of Nearest Neighbor Classifiers using Genetic Algorithm

Authors: M. Analoui, M. Fadavi Amiri

Abstract:

The design of a pattern classifier includes an attempt to select, among a set of possible features, a minimum subset of weakly correlated features that better discriminate the pattern classes. This is usually a difficult task in practice, normally requiring the application of heuristic knowledge about the specific problem domain. The selection and quality of the features representing each pattern have a considerable bearing on the success of subsequent pattern classification. Feature extraction is the process of deriving new features from the original features in order to reduce the cost of feature measurement, increase classifier efficiency, and allow higher classification accuracy. Many current feature extraction techniques involve linear transformations of the original pattern vectors to new vectors of lower dimensionality. While this is useful for data visualization and increasing classification efficiency, it does not necessarily reduce the number of features that must be measured since each new feature may be a linear combination of all of the features in the original pattern vector. In this paper a new approach is presented to feature extraction in which feature selection, feature extraction, and classifier training are performed simultaneously using a genetic algorithm. In this approach each feature value is first normalized by a linear equation, then scaled by the associated weight prior to training, testing, and classification. A knn classifier is used to evaluate each set of feature weights. The genetic algorithm optimizes a vector of feature weights, which are used to scale the individual features in the original pattern vectors in either a linear or a nonlinear fashion. By this approach, the number of features used in classifying can be finely reduced.

Keywords: Feature reduction, genetic algorithm, pattern classification, nearest neighbor rule classifiers (k-NNR).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1775
3533 Calcium Biochemical Indicators in a Group of Schoolchildren with Low Socioeconomic Status from Barranquilla, Colombia

Authors: Carmiña L. Vargas-Zapata, María A. Conde-Sarmiento, Maria Consuelo Maestre-Vargas

Abstract:

Calcium is an essential element for good growth and development of the organism, and its requirement is increased at school age. Low socio-economic populations of developing countries such as Colombia may have food deficiency of this mineral in schoolchildren that could be reflected in calcium biochemical indicators, bone alterations and anthropometric indicators. The objective of this investigation was to evaluate some calcium biochemical indicators in a group of schoolchildren of low socioeconomic level from Barranquilla city and to correlate with body mass index. 60 schoolchildren aged 7 to 15 years were selected from Jesus’s Heart Educational Institution in Barranquilla-Atlántico, apparently healthy, without suffering from infectious or gastrointestinal diseases, without habits of drinking alcohol or smoking another hallucinogenic substance and without taking supplementation with calcium in the last six months or another substance that compromises bone metabolism. The research was approved by the ethics committee at Universidad del Atlántico. The selected children were invited to donate a blood and urine sample in a fasting time of 12 hours, the serum was separated by centrifugation and frozen at ˗20 ℃ until analyzed and the same was done with the urine sample. On the day of the biological collections, the weight and height of the students were measured to determine the nutritional status by BMI using the WHO tables. Calcium concentrations in serum and urine (SCa, UCa), alkaline phosphatase activity total and of bone origin (SAPT, SBAP) and urinary creatinine (UCr) were determined by spectrophotometric methods using commercial kits. Osteocalcin and Cross-linked N-telopeptides of type I collagen (NTx-1) in serum were measured with an enzyme-linked inmunosorbent assay. For statistical analysis the Statgraphics software Centurium XVII was used. 63% (n = 38) and 37% (n = 22) of the participants were male and female, respectively. 78% (n = 47), 5% (n = 3) and 17% (n = 10) had a normal, malnutrition and high nutritional status, respectively. The averages of evaluated indicators levels were (mean ± SD): 9.50 ± 1.06 mg/dL for SCa; 181.3 ± 64.3 U/L for SAPT, 143.8 ± 73.9 U/L for SBAP; 9.0 ± 3.48 ng/mL for osteocalcin and 101.3 ± 12.8 ng/mL for NTx-1. UCa level was 12.8 ± 7.7 mg/dL that adjusted with creatinine ranged from 0.005 to 0.395 mg/mg. Considering serum calcium values, approximately 7% of school children were hypocalcemic, 16% hypercalcemic and 77% normocalcemic. The indicators evaluated did not correlate with the BMI. Low values ​​were observed in calcium urinary excretion and high in NTx-1, suggesting that mechanisms such as increase in renal retention of calcium and in bone remodeling may be contributing to calcium homeostasis.

Keywords: Calcium, calcium biochemical, indicators, school children, low socioeconomic status.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 549
3532 Target and Kaizen Costing

Authors: Alireza Azimi Sani , Mahdi Allahverdizadeh

Abstract:

increased competition and increased costs of designing made it important for the firms to identify the right products and the right methods for manufacturing the products. Firms should focus on customers and identify customer demands directly to design the right products. Several management methods and techniques that are currently available improve one or more functions or processes in an industry and do not take the complete product life cycle into consideration. On the other hand target costing is a method / philosophy that takes financial, manufacturing and customer aspects into consideration during designing phase and helps firms in making product design decisions to increase the profit / value of the company. It uses various techniques to identify customer demands, to decrease costs of manufacturing and finally to achieve strategic goals. Target Costing forms an integral part of total product design / redesign based on strategic plans.

Keywords: Target Costing, Target Cost Management, Cost Management, Activity Based Costing, New product design

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10179
3531 Geochemical Assessment of Metal Concentrations in Mangrove Sediments along Mumbai Coast, India

Authors: Lina Fernandes, G. N. Nayak, D. Ilangovan

Abstract:

Two short sediment cores collected from mangrove areas of Manori and Thane creeks along Mumbai coast were analysed for sediment composition and metals (Fe, Mn, Cu, Pb, Co, Ni, Zn, Cr and V). The statistical analysis of Pearson correlation matrix proved that there is a significant relationship between metal concentration and finer grain size in Manori creek while poor correlation was observed in Thane creek. Based on the enrichment factor, the present metal to background metal ratios clearly reflected maximum enrichment of Cu and Pb in Manori creek and Mn in Thane creek. Geoaccumulation index calculated indicate that the study area is unpolluted with respect to Fe, Mn, Co, Ni, Zn and Cr in both the cores while moderately polluted with Cu and Pb in Manori creek. Based on contamination degree, both the core sediments were found to be considerably contaminated with metals.

Keywords: Creek, Igeo, Mumbai, trace metals

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2657
3530 Assessment of Sediment Quality According To Heavy Metal Status in the West Port of Malaysia

Authors: B.Tavakoly Sany, A.H .Sulaiman, GH. Monazami, A. Salleh

Abstract:

Eight heavy metals (Cu, Cr, Zn, Hg, Pb, Cd, Ni and As) were analyzed in sediment samples in the dry and wet seasons from November 2009 to October 2010 in West Port of Peninsular Malaysia. The heavy metal concentrations (mg/kg dry weight) were ranged from 23.4 to 98.3 for Zn, 22.3 to 80 for Pb, 7.4 to 27.6 Cu, 0.244 to 3.53 for Cd, 7.2 to 22.2 for Ni, 20.2 to 162 for As, 0.11 to 0.409 for Hg and 11.5 to 61.5 for Cr. Metals concentrations in dry season were higher than the rainy season except in cupper and chromium. Analysis of variance with Statistical Analysis System (SAS) shows that the mean concentration of metals in the two seasons (α level=0.05) are not significantly different which shows that the metals were held firmly in the matrix of sediment. Also there are significant differences between control point station with other stations. According to the Interim Sediment Quality guidelines (ISQG), the metal concentrations are moderately polluted, except in arsenic which shows the highest level of pollution.

Keywords: Heavy metals, sediment quality guidelines, west port.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3807
3529 A Generator from Cascade Markov Model for Packet Loss and Subsequent Bit Error Description

Authors: Jaroslav Polec, Viliam Hirner, Michal Martinovič, Kvetoslava Kotuliaková

Abstract:

In this paper we present a novel error model for packet loss and subsequent error description. The proposed model simulates the error performance of wireless communication link. The model is designed as two independent Markov chains, where the first one is used for packet generation and the second one generates correctly and incorrectly transmitted bits for received packets from the first chain. The statistical analyses of real communication on the wireless link are used for determination of model-s parameters. Using the obtained parameters and the implementation of the generator, we collected generated traffic. The obtained results generated by proposed model are compared with the real data collection.

Keywords: Wireless channel, error model, Markov chain, Elliot model, Gilbert model, generator, IEEE 802.11.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2119
3528 The Relation between Social Capital and Trust with Social Network Analysis

Authors: Safak Baykal

Abstract:

The purpose of this study is analyzing the relationship between trust and social capital of people with using Social Network Analysis. In this study, two aspects of social capital will be focused: Bonding, homophilous social capital (BoSC), and Bridging, heterophilous social capital (BrSC). These two aspects diverge each other regarding to the social theories. The other concept of the study is Trust (Tr), namely interpersonal trust, willing to ascribe good intentions to and have confidence in the words and actions of other people. In this study, the sample group, 61 people, was selected from a private firm from the defense industry. The relation between BoSC/BrSC and Tr is shown by using Social Network Analysis (SNA) and statistical analysis with Likert type-questionnaire. The results of the analysis show the Cronbach’s alpha value is 0.756 and social capital values (BoSC/BrSC) is not correlated with Tr values of the people.

Keywords: Social capital, interpersonal trust, social network analysis (SNA).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2599
3527 Data Privacy and Safety with Large Language Models

Authors: Ashly Joseph, Jithu Paulose

Abstract:

Large language models (LLMs) have revolutionized natural language processing capabilities, enabling applications such as chatbots, dialogue agents, image, and video generators. Nevertheless, their trainings on extensive datasets comprising personal information poses notable privacy and safety hazards. This study examines methods for addressing these challenges, specifically focusing on approaches to enhance the security of LLM outputs, safeguard user privacy, and adhere to data protection rules. We explore several methods including post-processing detection algorithms, content filtering, reinforcement learning from human and AI inputs, and the difficulties in maintaining a balance between model safety and performance. The study also emphasizes the dangers of unintentional data leakage, privacy issues related to user prompts, and the possibility of data breaches. We highlight the significance of corporate data governance rules and optimal methods for engaging with chatbots. In addition, we analyze the development of data protection frameworks, evaluate the adherence of LLMs to General Data Protection Regulation (GDPR), and examine privacy legislation in academic and business policies. We demonstrate the difficulties and remedies involved in preserving data privacy and security in the age of sophisticated artificial intelligence by employing case studies and real-life instances. This article seeks to educate stakeholders on practical strategies for improving the security and privacy of LLMs, while also assuring their responsible and ethical implementation.

Keywords: Data privacy, large language models, artificial intelligence, machine learning, cybersecurity, general data protection regulation, data safety.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 157
3526 On the Parameter Optimization of Fuzzy Inference Systems

Authors: Erika Martinez Ramirez, Rene V. Mayorga

Abstract:

Nowadays, more engineering systems are using some kind of Artificial Intelligence (AI) for the development of their processes. Some well-known AI techniques include artificial neural nets, fuzzy inference systems, and neuro-fuzzy inference systems among others. Furthermore, many decision-making applications base their intelligent processes on Fuzzy Logic; due to the Fuzzy Inference Systems (FIS) capability to deal with problems that are based on user knowledge and experience. Also, knowing that users have a wide variety of distinctiveness, and generally, provide uncertain data, this information can be used and properly processed by a FIS. To properly consider uncertainty and inexact system input values, FIS normally use Membership Functions (MF) that represent a degree of user satisfaction on certain conditions and/or constraints. In order to define the parameters of the MFs, the knowledge from experts in the field is very important. This knowledge defines the MF shape to process the user inputs and through fuzzy reasoning and inference mechanisms, the FIS can provide an “appropriate" output. However an important issue immediately arises: How can it be assured that the obtained output is the optimum solution? How can it be guaranteed that each MF has an optimum shape? A viable solution to these questions is through the MFs parameter optimization. In this Paper a novel parameter optimization process is presented. The process for FIS parameter optimization consists of the five simple steps that can be easily realized off-line. Here the proposed process of FIS parameter optimization it is demonstrated by its implementation on an Intelligent Interface section dealing with the on-line customization / personalization of internet portals applied to E-commerce.

Keywords: Artificial Intelligence, Fuzzy Logic, Fuzzy InferenceSystems, Nonlinear Optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1991