Search results for: uncertainty of processing times
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2819

Search results for: uncertainty of processing times

2759 A Review in Advanced Digital Signal Processing Systems

Authors: Roza Dastres, Mohsen Soori

Abstract:

Digital Signal Processing (DSP) is the use of digital processing systems by computers in order to perform a variety of signal processing operations. It is the mathematical manipulation of a digital signal's numerical values in order to increase quality as well as effects of signals. DSP can include linear or nonlinear operators in order to process and analyze the input signals. The nonlinear DSP processing is closely related to nonlinear system detection and can be implemented in time, frequency and space-time domains. Applications of the DSP can be presented as control systems, digital image processing, biomedical engineering, speech recognition systems, industrial engineering, health care systems, radar signal processing and telecommunication systems. In this study, advanced methods and different applications of DSP are reviewed in order to move forward the interesting research filed.

Keywords: Digital signal processing, advanced telecommunication, nonlinear signal processing, speech recognition systems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 943
2758 Role of Investment in the Course of Economic Growth in Pakistan

Authors: Maqbool Hussain Sial, Maaida Hussain Hashmi, Sofia Anwar

Abstract:

The present research was focused to investigate the role of investment in the course of economic growth with reference to Pakistan. The study analyzed the role of the public and private investment and impact of the political and macroeconomic uncertainty on economic growth of Pakistan by using the vector autoregressive approach (VAR). In long-run both public and private investment showed a positive impact on economic growth but the growth was largely driven by private investment as compared to public investment. Government consumption expenditure, economic uncertainty and political instability hampered the economic growth of Pakistan. In short-run the private investment positively influences the growth but there was negative and insignificant effect of the public investment and government consumption expenditure on the growth. There was a positive relationship found between economic uncertainty (proxy for inflation) and GDP in short run.

Keywords: Investment, Government Consumption, Growth, Co-integration, Pakistan.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2072
2757 Forecasting Models for Steel Demand Uncertainty Using Bayesian Methods

Authors: Watcharin Sangma, Onsiri Chanmuang, Pitsanu Tongkhow

Abstract:

 A forecasting model for steel demand uncertainty in Thailand is proposed. It consists of trend, autocorrelation, and outliers in a hierarchical Bayesian frame work. The proposed model uses a cumulative Weibull distribution function, latent first-order autocorrelation, and binary selection, to account for trend, time-varying autocorrelation, and outliers, respectively. The Gibbs sampling Markov Chain Monte Carlo (MCMC) is used for parameter estimation. The proposed model is applied to steel demand index data in Thailand. The root mean square error (RMSE), mean absolute percentage error (MAPE), and mean absolute error (MAE) criteria are used for model comparison. The study reveals that the proposed model is more appropriate than the exponential smoothing method.

Keywords: Forecasting model, Steel demand uncertainty, Hierarchical Bayesian framework, Exponential smoothing method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2491
2756 Method for Determining the Probing Points for Efficient Measurement of Freeform Surface

Authors: Yi Xu, Zexiang Li

Abstract:

In inspection and workpiece localization, sampling point data is an important issue. Since the devices for sampling only sample discrete points, not the completely surface, sampling size and location of the points will be taken into consideration. In this paper a method is presented for determining the sampled points size and location for achieving efficient sampling. Firstly, uncertainty analysis of the localization parameters is investigated. A localization uncertainty model is developed to predict the uncertainty of the localization process. Using this model the minimum size of the sampled points is predicted. Secondly, based on the algebra theory an eigenvalue-optimal optimization is proposed. Then a freeform surface is used in the simulation. The proposed optimization is implemented. The simulation result shows its effectivity.

Keywords: eigenvalue-optimal optimization, freeform surface inspection, sampling size and location, sampled points.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1191
2755 Dynamic Slope Scaling Procedure for Stochastic Integer Programming Problem

Authors: Takayuki Shiina

Abstract:

Mathematical programming has been applied to various problems. For many actual problems, the assumption that the parameters involved are deterministic known data is often unjustified. In such cases, these data contain uncertainty and are thus represented as random variables, since they represent information about the future. Decision-making under uncertainty involves potential risk. Stochastic programming is a commonly used method for optimization under uncertainty. A stochastic programming problem with recourse is referred to as a two-stage stochastic problem. In this study, we consider a stochastic programming problem with simple integer recourse in which the value of the recourse variable is restricted to a multiple of a nonnegative integer. The algorithm of a dynamic slope scaling procedure for solving this problem is developed by using a property of the expected recourse function. Numerical experiments demonstrate that the proposed algorithm is quite efficient. The stochastic programming model defined in this paper is quite useful for a variety of design and operational problems.

Keywords: stochastic programming problem with recourse, simple integer recourse, dynamic slope scaling procedure

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1567
2754 Formulating the Stochastic Finite Elements for Free Vibration Analysis of Plates with Variable Elastic Modulus

Authors: Mojtaba Aghamiri Esfahani, Mohammad Karkon, Seyed Majid Hosseini Nezhad, Reza Hosseini-Ara

Abstract:

In this study, the effect of uncertainty in elastic modulus of a plate on free vibration response is investigated. For this purpose, the elastic modulus of the plate is modeled as stochastic variable with normal distribution. Moreover, the distance autocorrelation function is used for stochastic field. Then, by applying the finite element method and Monte Carlo simulation, stochastic finite element relations are extracted. Finally, with a numerical test, the effect of uncertainty in the elastic modulus on free vibration response of a plate is studied. The results show that the effect of uncertainty in elastic modulus of the plate cannot play an important role on the free vibration response.

Keywords: Stochastic finite elements, plate bending, free vibration, Monte Carlo, Neumann expansion method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1645
2753 A Two-Stage Expert System for Diagnosis of Leukemia Based on Type-2 Fuzzy Logic

Authors: Ali Akbar Sadat Asl

Abstract:

Diagnosis and deciding about diseases in medical fields is facing innate uncertainty which can affect the whole process of treatment. This decision is made based on expert knowledge and the way in which an expert interprets the patient's condition, and the interpretation of the various experts from the patient's condition may be different. Fuzzy logic can provide mathematical modeling for many concepts, variables, and systems that are unclear and ambiguous and also it can provide a framework for reasoning, inference, control, and decision making in conditions of uncertainty. In systems with high uncertainty and high complexity, fuzzy logic is a suitable method for modeling. In this paper, we use type-2 fuzzy logic for uncertainty modeling that is in diagnosis of leukemia. The proposed system uses an indirect-direct approach and consists of two stages: In the first stage, the inference of blood test state is determined. In this step, we use an indirect approach where the rules are extracted automatically by implementing a clustering approach. In the second stage, signs of leukemia, duration of disease until its progress and the output of the first stage are combined and the final diagnosis of the system is obtained. In this stage, the system uses a direct approach and final diagnosis is determined by the expert. The obtained results show that the type-2 fuzzy expert system can diagnose leukemia with the average accuracy about 97%.

Keywords: Expert system, leukemia, medical diagnosis, type-2 fuzzy logic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 975
2752 A Robust Optimization Model for the Single-Depot Capacitated Location-Routing Problem

Authors: Abdolsalam Ghaderi

Abstract:

In this paper, the single-depot capacitated location-routing problem under uncertainty is presented. The problem aims to find the optimal location of a single depot and the routing of vehicles to serve the customers when the parameters may change under different circumstances. This problem has many applications, especially in the area of supply chain management and distribution systems. To get closer to real-world situations, travel time of vehicles, the fixed cost of vehicles usage and customers’ demand are considered as a source of uncertainty. A combined approach including robust optimization and stochastic programming was presented to deal with the uncertainty in the problem at hand. For this purpose, a mixed integer programming model is developed and a heuristic algorithm based on Variable Neighborhood Search(VNS) is presented to solve the model. Finally, the computational results are presented and future research directions are discussed.

Keywords: Location-routing problem, robust optimization, Stochastic Programming, variable neighborhood search.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 700
2751 Sustainable Development and Kish Island Environment Protection, using Wind Energy

Authors: Amir Gandomkar

Abstract:

Kish Islands in South of Iran is located in coastal water near Hormozgan Province. Based on the wind 3-hour statistics in Kish station, the mean annual windspeed in this Island is 8.6 knot (4.3 m/s). The maximum windspeed recorded in this stations 47 knot (23.5 m/s). In 45.7 percent of recorded times, windspeed has been Zero or less than 8 knot which is not suitable to use the wind energy. But in 54.3 percent of recorded times, windspeed has been more than 8 knot and suitable to use wind energy to run turbines. In 40.2 percent of recorded times, windspeed has been between 8 to 16 knot, in 13 percent of times between 16 to 24 knot and in 1 percent of times it has been higher than 24 knot. In this station, the direction of winds higher than 8 is west and wind direction in Kish station is stable in most times of the year.With regard to high – speed and stable direction winds during the year and also shallow coasts near this is land, it is possible to build offshore wind farms near Kish Island and utilize wind energy produce the electricity required in this Island during most of the year.

Keywords: Kish Island, Wind energy, Offshore wind farm, Windspeed, Wind direction

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1682
2750 Aircraft Selection Problem Using Decision Uncertainty Distance in Fuzzy Multiple Criteria Decision Making Analysis

Authors: C. Ardil

Abstract:

Aircraft have different capabilities and specifications according to the required strategic goals and objectives in operations. With various types on the market with different aircraft characteristics, it becomes difficult to select a suitable aircraft for certain operations and requirements. The entropy weighting method (EWM) is a useful, highly consistent, and reliable method for obtaining the weights of the criteria and is worth integrating with the decision uncertainty distance (DUD) method, which is more applicable and requires less computation than other methods. An illustrative example is presented to demonstrate the validity and usability of the proposed methodology. Comparing the ranking results matches the distance-based approach, which is the technique for order preference by similarity to ideal solution (TOPSIS) method, which shows the robustness of the entropy DUD hybrid method. Validity analysis shows that the proposed hybrid multiple criteria decision-making analysis (MCDMA) methodology is quantitatively stable and reliable.

Keywords: aircraft selection, decision uncertainty distance (DUD), multiple criteria decision making analysis, MCDMA, TOPSIS

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 460
2749 From Type-I to Type-II Fuzzy System Modeling for Diagnosis of Hepatitis

Authors: Shahabeddin Sotudian, M. H. Fazel Zarandi, I. B. Turksen

Abstract:

Hepatitis is one of the most common and dangerous diseases that affects humankind, and exposes millions of people to serious health risks every year. Diagnosis of Hepatitis has always been a challenge for physicians. This paper presents an effective method for diagnosis of hepatitis based on interval Type-II fuzzy. This proposed system includes three steps: pre-processing (feature selection), Type-I and Type-II fuzzy classification, and system evaluation. KNN-FD feature selection is used as the preprocessing step in order to exclude irrelevant features and to improve classification performance and efficiency in generating the classification model. In the fuzzy classification step, an “indirect approach” is used for fuzzy system modeling by implementing the exponential compactness and separation index for determining the number of rules in the fuzzy clustering approach. Therefore, we first proposed a Type-I fuzzy system that had an accuracy of approximately 90.9%. In the proposed system, the process of diagnosis faces vagueness and uncertainty in the final decision. Thus, the imprecise knowledge was managed by using interval Type-II fuzzy logic. The results that were obtained show that interval Type-II fuzzy has the ability to diagnose hepatitis with an average accuracy of 93.94%. The classification accuracy obtained is the highest one reached thus far. The aforementioned rate of accuracy demonstrates that the Type-II fuzzy system has a better performance in comparison to Type-I and indicates a higher capability of Type-II fuzzy system for modeling uncertainty.

Keywords: Hepatitis disease, medical diagnosis, type-I fuzzy logic, type-II fuzzy logic, feature selection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1588
2748 The "Project" Approach in Urban: A Response to Uncertainty

Authors: Nedjima Mouhoubi, Souad Sassi Boudemagh

Abstract:

In this paper, we will try to demonstrate the importance of the project approach in the urban to deal with uncertainty, the importance of the involvement of all stakeholders in the urban project process and that the absence of an actor can lead to project failure but also the importance of the urban project management. These points are handled through the following questions: Does the urban adhere to the theory of complexity? Does the project approach bring hope and solution to make urban planning "sustainable"? How converging visions of actors for the same project? Is the management of urban project the solution to support the urban project approach?

Keywords: Strategic planning, project, urban project stakeholders, management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1239
2747 A Robust Optimization Method for Service Quality Improvement in Health Care Systems under Budget Uncertainty

Authors: H. Ashrafi, S. Ebrahimi, H. Kamalzadeh

Abstract:

With the development of business competition, it is important for healthcare providers to improve their service qualities. In order to improve service quality of a clinic, four important dimensions are defined: tangibles, responsiveness, empathy, and reliability. Moreover, there are several service stages in hospitals such as financial screening and examination. One of the most challenging limitations for improving service quality is budget which impressively affects the service quality. In this paper, we present an approach to address budget uncertainty and provide guidelines for service resource allocation. In this paper, a service quality improvement approach is proposed which can be adopted to multistage service processes to improve service quality, while controlling the costs. A multi-objective function based on the importance of each area and dimension is defined to link operational variables to service quality dimensions. The results demonstrate that our approach is not ultra-conservative and it shows the actual condition very well. Moreover, it is shown that different strategies can affect the number of employees in different stages.

Keywords: Service quality assessment, healthcare resource allocation, robust optimization, budget uncertainty.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1015
2746 Lower Bound of Time Span Product for a General Class of Signals in Fractional Fourier Domain

Authors: Sukrit Shankar, Chetana Shanta Patsa, Jaydev Sharma

Abstract:

Fractional Fourier Transform is a generalization of the classical Fourier Transform which is often symbolized as the rotation in time- frequency plane. Similar to the product of time and frequency span which provides the Uncertainty Principle for the classical Fourier domain, there has not been till date an Uncertainty Principle for the Fractional Fourier domain for a generalized class of finite energy signals. Though the lower bound for the product of time and Fractional Fourier span is derived for the real signals, a tighter lower bound for a general class of signals is of practical importance, especially for the analysis of signals containing chirps. We hence formulate a mathematical derivation that gives the lower bound of time and Fractional Fourier span product. The relation proves to be utmost importance in taking the Fractional Fourier Transform with adaptive time and Fractional span resolutions for a varied class of complex signals.

Keywords: Fractional Fourier Transform, uncertainty principle, Fractional Fourier Span, amplitude, phase.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1150
2745 Applications of Entropy Measures in Field of Queuing Theory

Authors: R.K.Tuli

Abstract:

In the present communication, we have studied different variations in the entropy measures in the different states of queueing processes. In case of steady state queuing process, it has been shown that as the arrival rate increases, the uncertainty increases whereas in the case of non-steady birth-death process, it is shown that the uncertainty varies differently. In this pattern, it first increases and attains its maximum value and then with the passage of time, it decreases and attains its minimum value.

Keywords: Entropy, Birth-death process, M/G/1 system, G/M/1system, Steady state, Non-steady state

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1550
2744 A New Intelligent, Dynamic and Real Time Management System of Sewerage

Authors: R. Tlili Yaakoubi, H. Nakouri, O. Blanpain, S. Lallahem

Abstract:

The current tools for real time management of sewer systems are based on two software tools: the software of weather forecast and the software of hydraulic simulation. The use of the first ones is an important cause of imprecision and uncertainty, the use of the second requires temporal important steps of decision because of their need in times of calculation. This way of proceeding fact that the obtained results are generally different from those waited. The major idea of this project is to change the basic paradigm by approaching the problem by the "automatic" face rather than by that "hydrology". The objective is to make possible the realization of a large number of simulations at very short times (a few seconds) allowing to take place weather forecasts by using directly the real time meditative pluviometric data. The aim is to reach a system where the decision-making is realized from reliable data and where the correction of the error is permanent. A first model of control laws was realized and tested with different return-period rainfalls. The gains obtained in rejecting volume vary from 19 to 100 %. The development of a new algorithm was then used to optimize calculation time and thus to overcome the subsequent combinatorial problem in our first approach. Finally, this new algorithm was tested with 16- year-rainfall series. The obtained gains are 40 % of total volume rejected to the natural environment and of 65 % in the number of discharges.

Keywords: Automation, optimization, paradigm, RTC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1451
2743 A Prediction Method for Large-Size Event Occurrences in the Sandpile Model

Authors: S. Channgam, A. Sae-Tang, T. Termsaithong

Abstract:

In this research, the occurrences of large size events in various system sizes of the Bak-Tang-Wiesenfeld sandpile model are considered. The system sizes (square lattice) of model considered here are 25×25, 50×50, 75×75 and 100×100. The cross-correlation between the ratio of sites containing 3 grain time series and the large size event time series for these 4 system sizes are also analyzed. Moreover, a prediction method of the large-size event for the 50×50 system size is also introduced. Lastly, it can be shown that this prediction method provides a slightly higher efficiency than random predictions.

Keywords: Bak-Tang-Wiesenfeld sandpile model, avalanches, cross-correlation, prediction method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1130
2742 Pulsed Multi-Layered Image Filtering: A VLSI Implementation

Authors: Christian Mayr, Holger Eisenreich, Stephan Henker, René Schüffny

Abstract:

Image convolution similar to the receptive fields found in mammalian visual pathways has long been used in conventional image processing in the form of Gabor masks. However, no VLSI implementation of parallel, multi-layered pulsed processing has been brought forward which would emulate this property. We present a technical realization of such a pulsed image processing scheme. The discussed IC also serves as a general testbed for VLSI-based pulsed information processing, which is of interest especially with regard to the robustness of representing an analog signal in the phase or duration of a pulsed, quasi-digital signal, as well as the possibility of direct digital manipulation of such an analog signal. The network connectivity and processing properties are reconfigurable so as to allow adaptation to various processing tasks.

Keywords: Neural image processing, pulse computation application, pulsed Gabor convolution, VLSI pulse routing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1344
2741 The Computational Psycholinguistic Situational-Fuzzy Self-Controlled Brain and Mind System under Uncertainty

Authors: Ben Khayut, Lina Fabri, Maya Avikhana

Abstract:

The modern Artificial Narrow Intelligence (ANI) models cannot: a) independently, situationally, and continuously function without of human intelligence, used for retraining and reprogramming the ANI’s models, and b) think, understand, be conscious, and cognize under uncertainty and changing of the environmental objects. To eliminate these shortcomings and build a new generation of Artificial Intelligence systems, the paper proposes a Conception, Model, and Method of Computational Psycholinguistic Cognitive Situational-Fuzzy Self-Controlled Brain and Mind System (CPCSFSCBMSUU). This system uses a neural network as its computational memory, and activates functions of the perception, identification of real objects, fuzzy situational control, and forming images of these objects. These images and objects are used for modeling their psychological, linguistic, cognitive, and neural values of properties and features, the meanings of which are identified, interpreted, generated, and formed taking into account the identified subject area, using the data, information, knowledge, accumulated in the Memory. The functioning of the CPCSFSCBMSUU is carried out by its subsystems of the: fuzzy situational control of all processes, computational perception, identifying of reactions and actions, Psycholinguistic Cognitive Fuzzy Logical Inference, Decision Making, Reasoning, Systems Thinking, Planning, Awareness, Consciousness, Cognition, Intuition, and Wisdom. In doing so are performed analysis and processing of the psycholinguistic, subject, visual, signal, sound and other objects, accumulation and using the data, information and knowledge of the Memory, communication, and interaction with other computing systems, robots and humans in order of solving the joint tasks. To investigate the functional processes of the proposed system, the principles of situational control, fuzzy logic, psycholinguistics, informatics, and modern possibilities of data science were applied. The proposed self-controlled system of brain and mind is oriented on use as a plug-in in multilingual subject applications.

Keywords: Computational psycholinguistic cognitive brain and mind system, situational fuzzy control, uncertainty, AI.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 321
2740 Data Preprocessing for Supervised Leaning

Authors: S. B. Kotsiantis, D. Kanellopoulos, P. E. Pintelas

Abstract:

Many factors affect the success of Machine Learning (ML) on a given task. The representation and quality of the instance data is first and foremost. If there is much irrelevant and redundant information present or noisy and unreliable data, then knowledge discovery during the training phase is more difficult. It is well known that data preparation and filtering steps take considerable amount of processing time in ML problems. Data pre-processing includes data cleaning, normalization, transformation, feature extraction and selection, etc. The product of data pre-processing is the final training set. It would be nice if a single sequence of data pre-processing algorithms had the best performance for each data set but this is not happened. Thus, we present the most well know algorithms for each step of data pre-processing so that one achieves the best performance for their data set.

Keywords: Data mining, feature selection, data cleaning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5935
2739 Extended Deductive Databases with Uncertain Information

Authors: Daniel Stamate

Abstract:

The paper presents an approach for handling uncertain information in deductive databases using multivalued logics. Uncertainty means that database facts may be assigned logical values other than the conventional ones - true and false. The logical values represent various degrees of truth, which may be combined and propagated by applying the database rules. A corresponding multivalued database semantics is defined. We show that it extends successful conventional semantics as the well-founded semantics, and has a polynomial time data complexity.

Keywords: Reasoning under uncertainty, multivalued logics, deductive databases, logic programs, multivalued semantics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1304
2738 Characterization of HD-V2 Gafchromic Film for Measurement of Spatial Dose Distribution from Alpha Particle of 5.5 MeV

Authors: A. Aydarous, M. El Ghazaly

Abstract:

The purpose of this study was to investigate the response of the newly released Gafchromic HD-V2 film for alpha particle of 5.5 MeV. Gafchromic HD-V2 was exposed to alpha particles of energy 5 MeV from 241Am for different durations. Then the films were scanned with a flatbed scanner. The dose response curve up to 2200 Gy has been achieved. The film’s reproducibility and sensitivity were evaluated. The results obtained show that the net optical density increases almost exponentially with the increase in the exposure time, and it becomes saturated after prolonged exposure times. The red channel shows the highest sensitivity, with a value of 4 x 10-3 Gy-1 at netOD of 0.4. The inter-film reproducibility was measured and the relative uncertainty found was 1.7 %, 2.1 % and 2.3 % for grey, red and green channels, respectively.

Keywords: Alpha dosimetry, 241Am, Gafchromic film.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3100
2737 The Validity Range of LSDP Robust Controller by Exploiting the Gap Metric Theory

Authors: Ali Ameur Haj Salah, Tarek Garna, Hassani Messaoud

Abstract:

This paper attempts to define the validity domain of LSDP (Loop Shaping Design Procedure) controller system, by determining the suitable uncertainty region, so that linear system be stable. Indeed the LSDP controller cannot provide stability for any perturbed system. For this, we will use the gap metric tool that is introduced into the control literature for studying robustness properties of feedback systems with uncertainty. A 2nd order electric linear system example is given to define the validity domain of LSDP controller and effectiveness gap metric.

Keywords: LSDP, Gap metric, Robust Control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1469
2736 Classification of Radio Communication Signals using Fuzzy Logic

Authors: Zuzana Dideková, Beata Mikovičová

Abstract:

Characterization of radio communication signals aims at automatic recognition of different characteristics of radio signals in order to detect their modulation type, the central frequency, and the level. Our purpose is to apply techniques used in image processing in order to extract pertinent characteristics. To the single analysis, we add several rules for checking the consistency of hypotheses using fuzzy logic. This allows taking into account ambiguity and uncertainty that may remain after the extraction of individual characteristics. The aim is to improve the process of radio communications characterization.

Keywords: fuzzy classification, fuzzy inference system, radio communication signals, telecommunications

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1914
2735 Information System for Early Diabetic Retinopathy Diagnostics Based on Multiscale Texture Gradient Method

Authors: L. S. Godlevsky, N. V. Kresyun, V. P. Martsenyuk, K. S. Shakun, T. V. Tatarchuk, K. O. Prybolovets, L. F. Kalinichenko, M. Karpinski, T. Gancarczyk

Abstract:

Structures of eye bottom were extracted using multiscale texture gradient method and color characteristics of macular zone and vessels were verified in CIELAB scale. The difference of average values of L*, a* and b* coordinates of CIE (International Commision of Illumination) scale in patients with diabetes and healthy volunteers was compared. The average value of L* in diabetic patients exceeded such one in the group of practically healthy persons by 2.71 times (P < 0.05), while the value of a* index was reduced by 3.8 times when compared with control one (P < 0.05). b* index exceeded such one in the control group by 12.4 times (P < 0.05). The integrated index on color difference (ΔE) exceeded control value by 2.87 times (P < 0.05). More pronounced differences with ΔE were followed by a shorter period of MA appearance with a correlation level at -0.56 (P < 0.05). The specificity of diagnostics raised by 2.17 times (P < 0.05) and negative prognostic index exceeded such one determined with the expert method by 2.26 times (P < 0.05).

Keywords: Diabetic retinopathy, multiscale texture gradient, color spectrum analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 504
2734 Minimal Critical Sets of Inertias for Irreducible Zero-nonzero Patterns of Order 3

Authors: Ber-Lin Yu, Ting-Zhu Huang

Abstract:

If there exists a nonempty, proper subset S of the set of all (n + 1)(n + 2)/2 inertias such that S Ôèå i(A) is sufficient for any n × n zero-nonzero pattern A to be inertially arbitrary, then S is called a critical set of inertias for zero-nonzero patterns of order n. If no proper subset of S is a critical set, then S is called a minimal critical set of inertias. In [3], Kim, Olesky and Driessche identified all minimal critical sets of inertias for 2 × 2 zero-nonzero patterns. Identifying all minimal critical sets of inertias for n × n zero-nonzero patterns with n ≥ 3 is posed as an open question in [3]. In this paper, all minimal critical sets of inertias for 3 × 3 zero-nonzero patterns are identified. It is shown that the sets {(0, 0, 3), (3, 0, 0)}, {(0, 0, 3), (0, 3, 0)}, {(0, 0, 3), (0, 1, 2)}, {(0, 0, 3), (1, 0, 2)}, {(0, 0, 3), (2, 0, 1)} and {(0, 0, 3), (0, 2, 1)} are the only minimal critical sets of inertias for 3 × 3 irreducible zerononzero patterns.

Keywords: Permutation digraph, zero-nonzero pattern, irreducible pattern, critical set of inertias, inertially arbitrary.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1191
2733 Representing Data without Lost Compression Properties in Time Series: A Review

Authors: Nabilah Filzah Mohd Radzuan, Zalinda Othman, Azuraliza Abu Bakar, Abdul Razak Hamdan

Abstract:

Uncertain data is believed to be an important issue in building up a prediction model. The main objective in the time series uncertainty analysis is to formulate uncertain data in order to gain knowledge and fit low dimensional model prior to a prediction task. This paper discusses the performance of a number of techniques in dealing with uncertain data specifically those which solve uncertain data condition by minimizing the loss of compression properties.

Keywords: Compression properties, uncertainty, uncertain time series, mining technique, weather prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1573
2732 The Effect on Lead Times When Normalizing a Supply Chain Process

Authors: Bassam Istanbouli

Abstract:

Organizations are living in a very competitive and dynamic environment which is constantly changing. In order to achieve a high level of service, the products and processes of these organizations need to be flexible and evolvable. If the supply chains are not modular and well designed, changes can bring combinatorial effects to most areas of a company from its management, financial, documentation, logistics and its information structure. Applying the normalized system’s concept to segments of the supply chain may help in reducing those ripple effects, but it may also increase lead times. Lead times are important and can become a decisive element in gaining customers. Industries are always under the pressure in providing good quality products, at competitive prices, when and how the customer wants them. Most of the time, the customers want their orders now, if not yesterday. The above concept will be proven by examining lead times in a manufacturing example before and after applying normalized systems concept to that segment of the chain. We will then show that although we can minimize the combinatorial effects when changes occur, the lead times will be increased.

Keywords: Supply chain, lead time, normalization, modular.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 544
2731 Evolution of Performance Measurement Methods in Conditions of Uncertainty: The Implementation of Fuzzy Sets in Performance Measurement

Authors: E. A. Tkachenko, E. M. Rogova, V. V. Klimov

Abstract:

One of the basic issues of development management is connected with performance measurement as a prerequisite for identifying the achievement of development objectives. The aim of our research is to develop an improved model of assessing a company’s development results. The model should take into account the cyclical nature of development and the high degree of uncertainty in dealing with numerous management tasks. Our hypotheses may be formulated as follows: Hypothesis 1. The cycle of a company’s development may be studied from the standpoint of a project cycle. To do that, methods and tools of project analysis are to be used. Hypothesis 2. The problem of the uncertainty when justifying managerial decisions within the framework of a company’s development cycle can be solved through the use of the mathematical apparatus of fuzzy logic. The reasoned justification of the validity of the hypotheses made is given in the suggested article. The fuzzy logic toolkit applies to the case of technology shift within an enterprise. It is proven that some restrictions in performance measurement that are incurred to conventional methods could be eliminated by implementation of the fuzzy logic apparatus in performance measurement models.

Keywords: Fuzzy logic, fuzzy sets, performance measurement, project analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1033
2730 Capability Prediction of Machining Processes Based on Uncertainty Analysis

Authors: Hamed Afrasiab, Saeed Khodaygan

Abstract:

Prediction of machining process capability in the design stage plays a key role to reach the precision design and manufacturing of mechanical products. Inaccuracies in machining process lead to errors in position and orientation of machined features on the part, and strongly affect the process capability in the final quality of the product. In this paper, an efficient systematic approach is given to investigate the machining errors to predict the manufacturing errors of the parts and capability prediction of corresponding machining processes. A mathematical formulation of fixture locators modeling is presented to establish the relationship between the part errors and the related sources. Based on this method, the final machining errors of the part can be accurately estimated by relating them to the combined dimensional and geometric tolerances of the workpiece – fixture system. This method is developed for uncertainty analysis based on the Worst Case and statistical approaches. The application of the presented method is illustrated through presenting an example and the computational results are compared with the Monte Carlo simulation results.

Keywords: Process capability, machining error, dimensional and geometrical tolerances, uncertainty analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1178