Search results for: work effort estimation methods
8319 A Posterior Predictive Model-Based Control Chart for Monitoring Healthcare
Authors: Yi-Fan Lin, Peter P. Howley, Frank A. Tuyl
Abstract:
Quality measurement and reporting systems are used in healthcare internationally. In Australia, the Australian Council on Healthcare Standards records and reports hundreds of clinical indicators (CIs) nationally across the healthcare system. These CIs are measures of performance in the clinical setting, and are used as a screening tool to help assess whether a standard of care is being met. Existing analysis and reporting of these CIs incorporate Bayesian methods to address sampling variation; however, such assessments are retrospective in nature, reporting upon the previous six or twelve months of data. The use of Bayesian methods within statistical process control for monitoring systems is an important pursuit to support more timely decision-making. Our research has developed and assessed a new graphical monitoring tool, similar to a control chart, based on the beta-binomial posterior predictive (BBPP) distribution to facilitate the real-time assessment of health care organizational performance via CIs. The BBPP charts have been compared with the traditional Bernoulli CUSUM (BC) chart by simulation. The more traditional “central” and “highest posterior density” (HPD) interval approaches were each considered to define the limits, and the multiple charts were compared via in-control and out-of-control average run lengths (ARLs), assuming that the parameter representing the underlying CI rate (proportion of cases with an event of interest) required estimation. Preliminary results have identified that the BBPP chart with HPD-based control limits provides better out-of-control run length performance than the central interval-based and BC charts. Further, the BC chart’s performance may be improved by using Bayesian parameter estimation of the underlying CI rate.
Keywords: Average run length, Bernoulli CUSUM chart, beta binomial posterior predictive distribution, clinical indicator, health care organization, highest posterior density interval.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8788318 Estimating Development Time of Software Projects Using a Neuro Fuzzy Approach
Authors: Venus Marza, Amin Seyyedi, Luiz Fernando Capretz
Abstract:
Software estimation accuracy is among the greatest challenges for software developers. This study aimed at building and evaluating a neuro-fuzzy model to estimate software projects development time. The forty-one modules developed from ten programs were used as dataset. Our proposed approach is compared with fuzzy logic and neural network model and Results show that the value of MMRE (Mean of Magnitude of Relative Error) applying neuro-fuzzy was substantially lower than MMRE applying fuzzy logic and neural network.Keywords: Artificial Neural Network, Fuzzy Logic, Neuro-Fuzzy, Software Estimation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16628317 A Parametric Study of an Inverse Electrostatics Problem (IESP) Using Simulated Annealing, Hooke & Jeeves and Sequential Quadratic Programming in Conjunction with Finite Element and Boundary Element Methods
Authors: Ioannis N. Koukoulis, Clio G. Vossou, Christopher G. Provatidis
Abstract:
The aim of the current work is to present a comparison among three popular optimization methods in the inverse elastostatics problem (IESP) of flaw detection within a solid. In more details, the performance of a simulated annealing, a Hooke & Jeeves and a sequential quadratic programming algorithm was studied in the test case of one circular flaw in a plate solved by both the boundary element (BEM) and the finite element method (FEM). The proposed optimization methods use a cost function that utilizes the displacements of the static response. The methods were ranked according to the required number of iterations to converge and to their ability to locate the global optimum. Hence, a clear impression regarding the performance of the aforementioned algorithms in flaw identification problems was obtained. Furthermore, the coupling of BEM or FEM with these optimization methods was investigated in order to track differences in their performance.
Keywords: Elastostatic, inverse problem, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18758316 Protein Delivery from Polymeric Nanoparticles
Authors: G. Spada, E. Gavini, P. Giunchedi
Abstract:
Aim of this work was to compare the efficacy of two loading methods of proteins onto polymeric nanocarriers: adsorption and encapsulation methods. Preliminary studies of protein loading were done using Bovine Serum Albumin (BSA) as model protein. Nanocarriers were prepared starting from polylactic co-glycolic acid (PLGA) polymer; production methods used are two different variants of emulsion evaporation method. Nanoparticles obtained were analyzed in terms of dimensions by Dynamic Light Scattering and Loading Efficiency of BSA by Bradford Assay. Loaded nanoparticles were then submitted to in-vitro protein dissolution test in order to study the effect of the delivery system on the release rate of the protein.Keywords: Drug delivery, nanoparticles, PLGA, proteinadsorption, protein encapsulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25148315 HaskellFL: A Tool for Detecting Logical Errors in Haskell
Authors: Vanessa Vasconcelos, Mariza A. S. Bigonha
Abstract:
Understanding and using the functional paradigm is a challenge for many programmers. Looking for logical errors in code may take a lot of a developer’s time when a program grows in size. In order to facilitate both processes, this paper presents HaskellFL, a tool that uses fault localization techniques to locate a logical error in Haskell code. The Haskell subset used in this work is sufficiently expressive for those studying Functional Programming to get immediate help debugging their code and to answer questions about key concepts associated with the functional paradigm. HaskellFL was tested against Functional Programming assignments submitted by students enrolled at the Functional Programming class at the Federal University of Minas Gerais and against exercises from the Exercism Haskell track that are publicly available in GitHub. This work also evaluated the effectiveness of two fault localization techniques, Tarantula and Ochiai, in the Haskell context. Furthermore, the EXAM score was chosen to evaluate the tool’s effectiveness, and results showed that HaskellFL reduced the effort needed to locate an error for all tested scenarios. The results also showed that the Ochiai method was more effective than Tarantula.
Keywords: Debug, fault localization, functional programming, Haskell.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7318314 Robust Adaptive Observer Design for Lipschitz Class of Nonlinear Systems
Authors: M. Pourgholi, V.J.Majd
Abstract:
This paper addresses parameter and state estimation problem in the presence of the perturbation of observer gain bounded input disturbances for the Lipschitz systems that are linear in unknown parameters and nonlinear in states. A new nonlinear adaptive resilient observer is designed, and its stability conditions based on Lyapunov technique are derived. The gain for this observer is derived systematically using linear matrix inequality approach. A numerical example is provided in which the nonlinear terms depend on unmeasured states. The simulation results are presented to show the effectiveness of the proposed method.
Keywords: Adaptive observer, linear matrix inequality, nonlinear systems, nonlinear observer, resilient observer, robust estimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26158313 Zero Truncated Strict Arcsine Model
Authors: Y. N. Phang, E. F. Loh
Abstract:
The zero truncated model is usually used in modeling count data without zero. It is the opposite of zero inflated model. Zero truncated Poisson and zero truncated negative binomial models are discussed and used by some researchers in analyzing the abundance of rare species and hospital stay. Zero truncated models are used as the base in developing hurdle models. In this study, we developed a new model, the zero truncated strict arcsine model, which can be used as an alternative model in modeling count data without zero and with extra variation. Two simulated and one real life data sets are used and fitted into this developed model. The results show that the model provides a good fit to the data. Maximum likelihood estimation method is used in estimating the parameters.
Keywords: Hurdle models, maximum likelihood estimation method, positive count data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18578312 Discrete Polyphase Matched Filtering-based Soft Timing Estimation for Mobile Wireless Systems
Authors: Thomas O. Olwal, Michael A. van Wyk, Barend J. van Wyk
Abstract:
In this paper we present a soft timing phase estimation (STPE) method for wireless mobile receivers operating in low signal to noise ratios (SNRs). Discrete Polyphase Matched (DPM) filters, a Log-maximum a posterior probability (MAP) and/or a Soft-output Viterbi algorithm (SOVA) are combined to derive a new timing recovery (TR) scheme. We apply this scheme to wireless cellular communication system model that comprises of a raised cosine filter (RCF), a bit-interleaved turbo-coded multi-level modulation (BITMM) scheme and the channel is assumed to be memory-less. Furthermore, no clock signals are transmitted to the receiver contrary to the classical data aided (DA) models. This new model ensures that both the bandwidth and power of the communication system is conserved. However, the computational complexity of ideal turbo synchronization is increased by 50%. Several simulation tests on bit error rate (BER) and block error rate (BLER) versus low SNR reveal that the proposed iterative soft timing recovery (ISTR) scheme outperforms the conventional schemes.
Keywords: discrete polyphase matched filters, maximum likelihood estimators, soft timing phase estimation, wireless mobile systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16928311 Transformer Top-Oil Temperature Modeling and Simulation
Authors: T. C. B. N. Assunção, J. L. Silvino, P. Resende
Abstract:
The winding hot-spot temperature is one of the most critical parameters that affect the useful life of the power transformers. The winding hot-spot temperature can be calculated as function of the top-oil temperature that can estimated by using the ambient temperature and transformer loading measured data. This paper proposes the estimation of the top-oil temperature by using a method based on Least Squares Support Vector Machines approach. The estimated top-oil temperature is compared with measured data of a power transformer in operation. The results are also compared with methods based on the IEEE Standard C57.91-1995/2000 and Artificial Neural Networks. It is shown that the Least Squares Support Vector Machines approach presents better performance than the methods based in the IEEE Standard C57.91-1995/2000 and artificial neural networks.Keywords: Artificial Neural Networks, Hot-spot Temperature, Least Squares Support Vector, Top-oil Temperature.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24928310 A New Computational Tool for Noise Prediction of Rotating Surfaces (FACT)
Authors: Ana Vieira, Fernando Lau, João Pedro Mortágua, Luís Cruz, Rui Santos
Abstract:
The air transport impact on environment is more than ever a limitative obstacle to the aeronautical industry continuous growth. Over the last decades, considerable effort has been carried out in order to obtain quieter aircraft solutions, whether by changing the original design or investigating more silent maneuvers. The noise propagated by rotating surfaces is one of the most important sources of annoyance, being present in most aerial vehicles. Bearing this is mind, CEIIA developed a new computational chain for noise prediction with in-house software tools to obtain solutions in relatively short time without using excessive computer resources. This work is based on the new acoustic tool, which aims to predict the rotor noise generated during steady and maneuvering flight, making use of the flexibility of the C language and the advantages of GPU programming in terms of velocity. The acoustic tool is based in the Formulation 1A of Farassat, capable of predicting two important types of noise: the loading and thickness noise. The present work describes the most important features of the acoustic tool, presenting its most relevant results and framework analyses for helicopters and UAV quadrotors.
Keywords: Rotor noise, acoustic tool, GPU Programming, UAV noise.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20588309 Pedometer Development Utilizing an Accelerometer Sensor
Authors: Ling-Mei Wu, Jia-Shing Sheu, Wei-Cian Jheng, Ying-Tung Hsiao
Abstract:
This paper develops a pedometer with a three-axis acceleration sensor that can be placed with any angle. The proposed pedometer measures the number of steps while users walk, jog or run. It can be worn on users’ waistband or placed within pocket or backpack. The work address to improve on the general pedometers, which can only be used in a single direction or can only count of steps without the continuous exercise judgment mechanism. Finally, experimental results confirm the superior performance of the proposed pedometer.
Keywords: Accelerometer sensor, Angle estimation, Pedometer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 47048308 Effect of Social Media on Knowledge Work
Authors: Pekka Makkonen, Georgios Lampropoulos, Kerstin Siakas
Abstract:
This paper examines the impact of social media on knowledge work. It discloses and highlights which specific aspects, areas and tasks of knowledge work can be improved by the use of social media. Moreover, the study includes a survey about higher education students’ viewpoints in regard to the use of social media as a means to enhance knowledge work and knowledge sharing. The analysis has been conducted based both on empirical data and on discussions about the sources dealing with knowledge work and how it can be enhanced by using social media. The results show that social media can improve knowledge work, knowledge building and maintenance tasks in which communication, information sharing and collaboration play a vital role. Additionally, by using social media, personal, collaborative and supplementary work activities can be enhanced. Based on the results of the study, we suggest how knowledge work can be enhanced when using the contemporary information and communications technologies (ICTs) of the 21st century and recommend future directions towards improving knowledge work.
Keywords: Knowledge work, social media, social media services, improving work performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10888307 Evaluation of Total Cross Section of Photo-Ionization of Helium in Weak Field on Base of Trajectory Method
Authors: Alexander B. Bichkov, Valery V. Smirnov
Abstract:
Total cross section of helium atom photo-ionization by weak short pulse is calculated using the variant of trajectory method, developed in our earlier work. The method enables simple estimation of total ionization probability (or cross section) without integration of differential one.
Keywords: Evaluation of Photo-Ionization, Helium, Trajectory Method
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18838306 Object Recognition in Color Images by the Self Configuring System MEMORI
Authors: Michela Lecca
Abstract:
System MEMORI automatically detects and recognizes rotated and/or rescaled versions of the objects of a database within digital color images with cluttered background. This task is accomplished by means of a region grouping algorithm guided by heuristic rules, whose parameters concern some geometrical properties and the recognition score of the database objects. This paper focuses on the strategies implemented in MEMORI for the estimation of the heuristic rule parameters. This estimation, being automatic, makes the system a self configuring and highly user-friendly tool.Keywords: Automatic Object Recognition, Clustering, Contentbased Image Retrieval System, Image Segmentation, Region Adjacency Graph, Region Grouping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12028305 Inference of Stress-Strength Model for a Lomax Distribution
Abstract:
In this paper, the estimation of the stress-strength parameter R = P(Y < X), when X and Y are independent and both are Lomax distributions with the common scale parameters but different shape parameters is studied. The maximum likelihood estimator of R is derived. Assuming that the common scale parameter is known, the bayes estimator and exact confidence interval of R are discussed. Simulation study to investigate performance of the different proposed methods has been carried out.Keywords: Stress-Strength model; maximum likelihoodestimator; Bayes estimator; Lomax distribution
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17938304 Kalman-s Shrinkage for Wavelet-Based Despeckling of SAR Images
Authors: Mario Mastriani, Alberto E. Giraldez
Abstract:
In this paper, a new probability density function (pdf) is proposed to model the statistics of wavelet coefficients, and a simple Kalman-s filter is derived from the new pdf using Bayesian estimation theory. Specifically, we decompose the speckled image into wavelet subbands, we apply the Kalman-s filter to the high subbands, and reconstruct a despeckled image from the modified detail coefficients. Experimental results demonstrate that our method compares favorably to several other despeckling methods on test synthetic aperture radar (SAR) images.Keywords: Kalman's filter, shrinkage, speckle, wavelets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16068303 A Comparison of Marginal and Joint Generalized Quasi-likelihood Estimating Equations Based On the Com-Poisson GLM: Application to Car Breakdowns Data
Authors: N. Mamode Khan, V. Jowaheer
Abstract:
In this paper, we apply and compare two generalized estimating equation approaches to the analysis of car breakdowns data in Mauritius. Number of breakdowns experienced by a machinery is a highly under-dispersed count random variable and its value can be attributed to the factors related to the mechanical input and output of that machinery. Analyzing such under-dispersed count observation as a function of the explanatory factors has been a challenging problem. In this paper, we aim at estimating the effects of various factors on the number of breakdowns experienced by a passenger car based on a study performed in Mauritius over a year. We remark that the number of passenger car breakdowns is highly under-dispersed. These data are therefore modelled and analyzed using Com-Poisson regression model. We use the two types of quasi-likelihood estimation approaches to estimate the parameters of the model: marginal and joint generalized quasi-likelihood estimating equation approaches. Under-dispersion parameter is estimated to be around 2.14 justifying the appropriateness of Com-Poisson distribution in modelling underdispersed count responses recorded in this study.
Keywords: Breakdowns, under-dispersion, com-poisson, generalized linear model, marginal quasi-likelihood estimation, joint quasi-likelihood estimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14708302 Knowledge Impact on Measurement: A Conceptual Metric for Evaluating Performance Improvement (PI) at the Kuwait Institute for Scientific Research (KISR)
Authors: AlMatrouk H. S., Juszczak M. D.
Abstract:
Research and development R&D work involves enormous amount of work that has to do with data measurement and collection. This process evolves as new information is fed, new technologies are utilized, and eventually new knowledge is created by the stakeholders i.e., researchers, clients, and end-users. When new knowledge is created, procedures of R&D work should evolve and produce better results within improved research skills and improved methods of data measurements and collection. This measurement improvement should then be benchmarked against a metric that should be developed at the organization. In this paper, we are suggesting a conceptual metric for R&D work performance improvement (PI) at the Kuwait Institute for Scientific Research (KISR). This PI is to be measured against a set of variables in the suggested metric, which are more closely correlated to organizational output, as opposed to organizational norms. The paper also mentions and discusses knowledge creation and management as an addedvalue to R&D work and measurement improvement. The research methodology followed in this work is qualitative in nature, based on a survey that was distributed to researchers and interviews held with senior researchers at KISR. Research and analyses in this paper also include looking at and analyzing KISR-s literature.Keywords: Knowledge Creation, Performance Improvement (PI), Conceptual Metric, Knowledge Management (KM) addedvalue.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12538301 The Role of Meaningful Work in Transformational Leadership and Work Outcomes Relationship
Authors: Zainur Rahman
Abstract:
Meaningful work is the topic that will be discussed in this article, especially in changing period. It has an important role because by reaching meaningful work, it will drive to be positive in the workplace. Therefore, task performance will be increased and cynicism about organizational change (CAOC) will be reduced. Moreover, it is influenced by situational factor, which is transformational leadership. In this conceptual paper, the author discusses how the construct of meaningful work influenced by transformational leadership that will have impact on the follower’ work outcomes in the organizational change. It is proposed that the construct of meaningful work are susceptible with situational variable. Transformational leaders who are respectful on the process of humanizing the followers affect task performance and reduce CAOC in organizational change.
Keywords: Meaningful work, organizational change, task performance, and work outcomes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10148300 Kinetic Parameter Estimation from Thermogravimetry and Microscale Combustion Calorimetry
Authors: Rhoda Afriyie Mensah, Lin Jiang, Solomon Asante-Okyere, Xu Qiang, Cong Jin
Abstract:
Flammability analysis of extruded polystyrene (XPS) has become crucial due to its utilization as insulation material for energy efficient buildings. Using the Kissinger-Akahira-Sunose and Flynn-Wall-Ozawa methods, the degradation kinetics of two pure XPS from the local market, red and grey ones, were obtained from the results of thermogravity analysis (TG) and microscale combustion calorimetry (MCC) experiments performed under the same heating rates. From the experiments, it was discovered that red XPS released more heat than grey XPS and both materials showed two mass loss stages. Consequently, the kinetic parameters for red XPS were higher than grey XPS. A comparative evaluation of activation energies from MCC and TG showed an insignificant degree of deviation signifying an equivalent apparent activation energy from both methods. However, different activation energy profiles as a result of the different chemical pathways were presented when the dependencies of the activation energies on extent of conversion for TG and MCC were compared.
Keywords: Flammability, microscale combustion calorimetry, thermogravity analysis, thermal degradation, kinetic analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8848299 Electromagnetic Source Direction of Arrival Estimation via Virtual Antenna Array
Authors: Meiling Yang, Shuguo Xie, Yilong Zhu
Abstract:
Nowadays, due to diverse electric products and complex electromagnetic environment, the localization and troubleshooting of the electromagnetic radiation source is urgent and necessary especially on the condition of far field. However, based on the existing DOA positioning method, the system or devices are complex, bulky and expensive. To address this issue, this paper proposes a single antenna radiation source localization method. A single antenna moves to form a virtual antenna array combined with DOA and MUSIC algorithm to position accurately, meanwhile reducing the cost and simplify the equipment. As shown in the results of simulations and experiments, the virtual antenna array DOA estimation modeling is correct and its positioning is credible.
Keywords: Virtual antenna array, DOA, localization, far field.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9708298 High Resolution Methods Based On Rank Revealing Triangular Factorizations
Authors: M. Bouri, S. Bourennane
Abstract:
In this paper, we propose a novel method for subspace estimation used high resolution method without eigendecomposition where the sample Cross-Spectral Matrix (CSM) is replaced by upper triangular matrix obtained from LU factorization. This novel method decreases the computational complexity. The method relies on a recently published result on Rank-Revealing LU (RRLU) factorization. Simulation results demonstrates that the new algorithm outperform the Householder rank-revealing QR (RRQR) factorization method and the MUSIC in the low Signal to Noise Ratio (SNR) scenarios.
Keywords: Factorization, Localization, Matrix, Signalsubspace.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13608297 A New Approach to Signal Processing for DC-Electromagnetic Flowmeters
Authors: Michael Schukat
Abstract:
Electromagnetic flowmeters with DC excitation are used for a wide range of fluid measurement tasks, but are rarely found in dosing applications with short measurement cycles due to the achievable accuracy. This paper will identify a number of factors that influence the accuracy of this sensor type when used for short-term measurements. Based on these results a new signal-processing algorithm will be described that overcomes the identified problems to some extend. This new method allows principally a higher accuracy of electromagnetic flowmeters with DC excitation than traditional methods.
Keywords: Electromagnetic Flowmeter, Kalman Filter, ShortMeasurement Cycles, Signal Estimation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16148296 Ratio Type Estimators of the Population Mean Based on Ranked Set Sampling
Authors: Said Ali Al-Hadhrami
Abstract:
Ranked set sampling (RSS) was first suggested to increase the efficiency of the population mean. It has been shown that this method is highly beneficial to the estimation based on simple random sampling (SRS). There has been considerable development and many modifications were done on this method. When a concomitant variable is available, ratio estimation based on ranked set sampling was proposed. This ratio estimator is more efficient than that based on SRS. In this paper some ratio type estimators of the population mean based on RSS are suggested. These estimators are found to be more efficient than the estimators of similar form using simple random sample.
Keywords: Bias, Efficiency, Ranked Set Sampling, Ratio Type Estimator
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13748295 Speech Intelligibility Improvement Using Variable Level Decomposition DWT
Authors: Samba Raju, Chiluveru, Manoj Tripathy
Abstract:
Intelligibility is an essential characteristic of a speech signal, which is used to help in the understanding of information in speech signal. Background noise in the environment can deteriorate the intelligibility of a recorded speech. In this paper, we presented a simple variance subtracted - variable level discrete wavelet transform, which improve the intelligibility of speech. The proposed algorithm does not require an explicit estimation of noise, i.e., prior knowledge of the noise; hence, it is easy to implement, and it reduces the computational burden. The proposed algorithm decides a separate decomposition level for each frame based on signal dominant and dominant noise criteria. The performance of the proposed algorithm is evaluated with speech intelligibility measure (STOI), and results obtained are compared with Universal Discrete Wavelet Transform (DWT) thresholding and Minimum Mean Square Error (MMSE) methods. The experimental results revealed that the proposed scheme outperformed competing methodsKeywords: Discrete Wavelet Transform, speech intelligibility, STOI, standard deviation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6938294 Root Growth of Morus alba as Affected by Size of Cuttings and Polythene Low Tunnel
Authors: Irfan Ahmad, Tahir Siddiqui, Rashid Ahmad Khan, Tahir Munir Butt
Abstract:
An effort to find out the smaller size of cuttings for propagation of Morus alba was made in experimental area Department of Forestry, Range Management and Wildlife, University of Agriculture, Faisalabad, Pakistan. Different size of cuttings i.e. 2", 4", 6" and 8" were planted in polythene tubes of 3.5"x7". The effort was also made to compare the performance of cuttings in open air and in polythene low tunnel. Root length, number of root branches, root diameter and root fresh and dry weight were found maximum in two inches cuttings while minimum in four inches cuttings. Root growth was found maximum in open air as compared to under polythene sheet.
Keywords: cutting sizes Morus alba, Open air and polythene sheet, root growth
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30728293 Identification of Aircraft Gas Turbine Engines Temperature Condition
Authors: Pashayev A., Askerov D., C. Ardil, Sadiqov R., Abdullayev P.
Abstract:
Groundlessness of application probability-statistic methods are especially shown at an early stage of the aviation GTE technical condition diagnosing, when the volume of the information has property of the fuzzy, limitations, uncertainty and efficiency of application of new technology Soft computing at these diagnosing stages by using the fuzzy logic and neural networks methods. It is made training with high accuracy of multiple linear and nonlinear models (the regression equations) received on the statistical fuzzy data basis. At the information sufficiency it is offered to use recurrent algorithm of aviation GTE technical condition identification on measurements of input and output parameters of the multiple linear and nonlinear generalized models at presence of noise measured (the new recursive least squares method (LSM)). As application of the given technique the estimation of the new operating aviation engine D30KU-154 technical condition at height H=10600 m was made.
Keywords: Identification of a technical condition, aviation gasturbine engine, fuzzy logic and neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16608292 On a Way for Constructing Numerical Methods on the Joint of Multistep and Hybrid Methods
Authors: G.Mehdiyeva, M.Imanova, V.Ibrahimov
Abstract:
Taking into account that many problems of natural sciences and engineering are reduced to solving initial-value problem for ordinary differential equations, beginning from Newton, the scientists investigate approximate solution of ordinary differential equations. There are papers of different authors devoted to the solution of initial value problem for ODE. The Euler-s known method that was developed under the guidance of the famous scientists Adams, Runge and Kutta is the most popular one among these methods. Recently the scientists began to construct the methods preserving some properties of Adams and Runge-Kutta methods and called them hybrid methods. The constructions of such methods are investigated from the middle of the XX century. Here we investigate one generalization of multistep and hybrid methods and on their base we construct specific methods of accuracy order p = 5 and p = 6 for k = 1 ( k is the order of the difference method).Keywords: Multistep and hybrid methods, initial value problem, degree and stability of hybrid methods
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15988291 Identification of Aircraft Gas Turbine Engine's Temperature Condition
Authors: Pashayev A., Askerov D., C. Ardil, Sadiqov R., Abdullayev P.
Abstract:
Groundlessness of application probability-statistic methods are especially shown at an early stage of the aviation GTE technical condition diagnosing, when the volume of the information has property of the fuzzy, limitations, uncertainty and efficiency of application of new technology Soft computing at these diagnosing stages by using the fuzzy logic and neural networks methods. It is made training with high accuracy of multiple linear and nonlinear models (the regression equations) received on the statistical fuzzy data basis. At the information sufficiency it is offered to use recurrent algorithm of aviation GTE technical condition identification on measurements of input and output parameters of the multiple linear and nonlinear generalized models at presence of noise measured (the new recursive least squares method (LSM)). As application of the given technique the estimation of the new operating aviation engine D30KU-154 technical condition at height H=10600 m was made.
Keywords: Identification of a technical condition, aviation gasturbine engine, fuzzy logic and neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16738290 Comparison of Methods of Testing Composite Slabs
Authors: J. Holomek, R. Karásek, M. Bajer, J. Barnat
Abstract:
Composite steel-concrete slabs using thin-walled corrugated steel sheets with embossments represent a modern and effective combination of steel and concrete. However, the design of new types of sheeting is conditional on the execution of expensive and time-consuming laboratory testing. The effort to develop a cheaper and faster method has lead to many investigations all over the world. In our paper we compare the results from our experiments involving vacuum loading, four-point bending and small-scale shear tests.Keywords: Composite slab, embossment, four-point bending, small-scale test, steel sheet, thin-walled, vacuum loading
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1945