Search results for: ANN approach
4731 Risk Management Approach for Lean, Agile, Resilient and Green Supply Chain
Authors: Benmoussa Rachid, Deguio Roland, Dubois Sebastien, Rasovska Ivana
Abstract:
Implementation of LARG (Lean, Agile, Resilient, Green) practices in the supply chain management is a complex task mainly because ecological, economical and operational goals are usually in conflict. To implement these LARG practices successfully, companies’ need relevant decision making tools allowing processes performance control and improvement strategies visibility. To contribute to this issue, this work tries to answer the following research question: How to master performance and anticipate problems in supply chain LARG practices implementation? To answer this question, a risk management approach (RMA) is adopted. Indeed, the proposed RMA aims basically to assess the ability of a supply chain, guided by “Lean, Green and Achievement” performance goals, to face “agility and resilience risk” factors. To proof its relevance, a logistics academic case study based on simulation is used to illustrate all its stages. It shows particularly how to build the “LARG risk map” which is the main output of this approach.
Keywords: Risk approach, lean supply chain, agile supply chain, green supply chain, resilient supply chain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18904730 A K-Means Based Clustering Approach for Finding Faulty Modules in Open Source Software Systems
Authors: Parvinder S. Sandhu, Jagdeep Singh, Vikas Gupta, Mandeep Kaur, Sonia Manhas, Ramandeep Sidhu
Abstract:
Prediction of fault-prone modules provides one way to support software quality engineering. Clustering is used to determine the intrinsic grouping in a set of unlabeled data. Among various clustering techniques available in literature K-Means clustering approach is most widely being used. This paper introduces K-Means based Clustering approach for software finding the fault proneness of the Object-Oriented systems. The contribution of this paper is that it has used Metric values of JEdit open source software for generation of the rules for the categorization of software modules in the categories of Faulty and non faulty modules and thereafter empirically validation is performed. The results are measured in terms of accuracy of prediction, probability of Detection and Probability of False Alarms.Keywords: K-Means, Software Fault, Classification, ObjectOriented Metrics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23034729 A Logic Approach to Database Dynamic Updating
Authors: Daniel Stamate
Abstract:
We introduce a logic-based framework for database updating under constraints. In our framework, the constraints are represented as an instantiated extended logic program. When performing an update, database consistency may be violated. We provide an approach of maintaining database consistency, and study the conditions under which the maintenance process is deterministic. We show that the complexity of the computations and decision problems presented in our framework is in each case polynomial time.Keywords: Databases, knowledge bases, constraints, updates, minimal change, consistency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13584728 Walsh-Hadamard Transform for Facial Feature Extraction in Face Recognition
Authors: M. Hassan, I. Osman, M. Yahia
Abstract:
This Paper proposes a new facial feature extraction approach, Wash-Hadamard Transform (WHT). This approach is based on correlation between local pixels of the face image. Its primary advantage is the simplicity of its computation. The paper compares the proposed approach, WHT, which was traditionally used in data compression with two other known approaches: the Principal Component Analysis (PCA) and the Discrete Cosine Transform (DCT) using the face database of Olivetti Research Laboratory (ORL). In spite of its simple computation, the proposed algorithm (WHT) gave very close results to those obtained by the PCA and DCT. This paper initiates the research into WHT and the family of frequency transforms and examines their suitability for feature extraction in face recognition applications.
Keywords: Face Recognition, Facial Feature Extraction, Principal Component Analysis, and Discrete Cosine Transform, Wash-Hadamard Transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25704727 A Rough Sets Approach for Relevant Internet/Web Online Searching
Authors: Erika Martinez Ramirez, Rene V. Mayorga
Abstract:
The internet is constantly expanding. Identifying web links of interest from web browsers requires users to visit each of the links listed, individually until a satisfactory link is found, therefore those users need to evaluate a considerable amount of links before finding their link of interest; this can be tedious and even unproductive. By incorporating web assistance, web users could be benefited from reduced time searching on relevant websites. In this paper, a rough set approach is presented, which facilitates classification of unlimited available e-vocabulary, to assist web users in reducing search times looking for relevant web sites. This approach includes two methods for identifying relevance data on web links based on the priority and percentage of relevance. As a result of these methods, a list of web sites is generated in priority sequence with an emphasis of the search criteria.Keywords: Web search, Web Mining, Rough Sets, Web Intelligence, Intelligent Portals, Relevance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15494726 Interactive Compromise Approach with Particle Swarm Optimization for Environmental/Economic Power Dispatch
Authors: Ming-Tang Tsai, Chih-Wei Yen
Abstract:
In this paper, an Interactive Compromise Approach with Particle Swarm Optimization(ICA-PSO) is presented to solve the Economic Emission Dispatch(EED) problem. The cost function and emission function are modeled as the nonsmooth functions, respectively. The bi-objective including both the minimization of cost and emission is formulated in this paper. ICA-PSO is proposed to solve EED problem for finding a better compromise solution. The solution methodology can offer a global or near-global solution for decision-making requirements. The effectiveness and efficiency of ICA-PSO are demonstrated by a sample test system. Test results can be shown that the proposed method provide a practical and flexible framework for power dispatch.Keywords: Interactive Compromise Approach, Emission Control, Economic Dispatch, Particle Swarm Optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14524725 Methods of Estimating the Equilibrium Real Effective Exchange Rate (REER)
Authors: Pavla Ruzickova, Petr Teply
Abstract:
There are many debates now regarding undervalued and overvalued currencies currently traded on the world financial market. This paper contributes to these debates from a theoretical point of view. We present the three most commonly used methods of estimating the equilibrium real effective exchange rate (REER): macroeconomic balance approach, external sustainability approach and equilibrium real effective exchange rate approach in the reduced form. Moreover, we discuss key concepts of the calculation of the real exchange rate (RER) based on applied explanatory variables: nominal exchange rates, terms of trade and tradable and non-tradable goods. Last but not least, we discuss the three main driving forces behind real exchange rates movements which include terms of trade, relative productivity growth and the interest rate differential.Keywords: real exchange rate, real effective exchange rate, foreign exchange, terms of trade
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24884724 Proxisch: An Optimization Approach of Large-Scale Unstable Proxy Servers Scheduling
Authors: Xiaoming Jiang, Jinqiao Shi, Qingfeng Tan, Wentao Zhang, Xuebin Wang, Muqian Chen
Abstract:
Nowadays, big companies such as Google, Microsoft, which have adequate proxy servers, have perfectly implemented their web crawlers for a certain website in parallel. But due to lack of expensive proxy servers, it is still a puzzle for researchers to crawl large amounts of information from a single website in parallel. In this case, it is a good choice for researchers to use free public proxy servers which are crawled from the Internet. In order to improve efficiency of web crawler, the following two issues should be considered primarily: (1) Tasks may fail owing to the instability of free proxy servers; (2) A proxy server will be blocked if it visits a single website frequently. In this paper, we propose Proxisch, an optimization approach of large-scale unstable proxy servers scheduling, which allow anyone with extremely low cost to run a web crawler efficiently. Proxisch is designed to work efficiently by making maximum use of reliable proxy servers. To solve second problem, it establishes a frequency control mechanism which can ensure the visiting frequency of any chosen proxy server below the website’s limit. The results show that our approach performs better than the other scheduling algorithms.Keywords: Proxy server, priority queue, optimization approach, distributed web crawling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28914723 A Hybrid Approach to Fault Detection and Diagnosis in a Diesel Fuel Hydrotreatment Process
Authors: Salvatore L., Pires B., Campos M. C. M., De Souza Jr M. B.
Abstract:
It is estimated that the total cost of abnormal conditions to US process industries is around $20 billion dollars in annual losses. The hydrotreatment (HDT) of diesel fuel in petroleum refineries is a conversion process that leads to high profitable economical returns. However, this is a difficult process to control because it is operated continuously, with high hydrogen pressures and it is also subject to disturbances in feed properties and catalyst performance. So, the automatic detection of fault and diagnosis plays an important role in this context. In this work, a hybrid approach based on neural networks together with a pos-processing classification algorithm is used to detect faults in a simulated HDT unit. Nine classes (8 faults and the normal operation) were correctly classified using the proposed approach in a maximum time of 5 minutes, based on on-line data process measurements.Keywords: Fault detection, hydrotreatment, hybrid systems, neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16474722 Relationship between Behavioral Inhibition/Approach System and Perceived Stress: With White Blood Cell in Multiple Sclerosis Patients
Authors: Amin Alvani
Abstract:
Multiple sclerosis (MS) is a chronic, often disabling disease in which the immune system attacks the myelin sheath of neurons in the central nervous system. The purpose of this study was to explore the correlation between the Behavioral Inhibition/Approach System (BIS-BAS) and Perceived Stress (PS), while controlling for White Blood Cell (WBC) count. 60 MS patients (36.7% male, 63.3% female; aged 15-65 years) participated in this study. They completed a demographic questionnaire, underwent a complete blood cell (CBC) test, filled out the Behavioral Activation and Behavioral Inhibition Scale (BIS-BAS), and responded to the Perceived Stress Questionnaire (PSS-14). The results indicated a significant relationship between the BAS-Reward Responsiveness (BAS-RR) subscale and PS, particularly in a subset of MS patients with increased WBC counts.
Keywords: Behavioral inhibition/approach system, multiple sclerosis, perceived stress, white blood cell.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 544721 DNA Nanowires: A Charge Transfer Approach
Authors: S. Behnia, S. Fathizadeh
Abstract:
Conductivity properties of DNA molecule is investigated in a simple, but chemically specific approach that is intimately related to the Su-Schrieffer-Heeger (SSH) model. This model is a tight-binding linear nanoscale chain. We have tried to study the electrical current flowing in DNA and investigated the characteristic I-V diagram. As a result, It is shown that there are the (quasi-) ohmic areas in I-V diagram. On the other hand, the regions with a negative differential resistance (NDR) are detectable in diagram.Keywords: Charge transfer in DNA, Chaos theory, Molecular electronics, Negative Differential resistance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18434720 A Mathematical Model Approach Regarding the Children’s Height Development with Fractional Calculus
Authors: Nisa Özge Önal, Kamil Karaçuha, Göksu Hazar Erdinç, Banu Bahar Karaçuha, Ertuğrul Karaçuha
Abstract:
The study aims to use a mathematical approach with the fractional calculus which is developed to have the ability to continuously analyze the factors related to the children’s height development. Until now, tracking the development of the child is getting more important and meaningful. Knowing and determining the factors related to the physical development of the child any desired time would provide better, reliable and accurate results for childcare. In this frame, 7 groups for height percentile curve (3th, 10th, 25th, 50th, 75th, 90th, and 97th) of Turkey are used. By using discrete height data of 0-18 years old children and the least squares method, a continuous curve is developed valid for any time interval. By doing so, in any desired instant, it is possible to find the percentage and location of the child in Percentage Chart. Here, with the help of the fractional calculus theory, a mathematical model is developed. The outcomes of the proposed approach are quite promising compared to the linear and the polynomial method. The approach also yields to predict the expected values of children in the sense of height.
Keywords: Children growth percentile, children physical development, fractional calculus, linear and polynomial model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8594719 State Feedback Controller Design via Takagi- Sugeno Fuzzy Model: LMI Approach
Authors: F. Khaber, K. Zehar, A. Hamzaoui
Abstract:
In this paper, we introduce a robust state feedback controller design using Linear Matrix Inequalities (LMIs) and guaranteed cost approach for Takagi-Sugeno fuzzy systems. The purpose on this work is to establish a systematic method to design controllers for a class of uncertain linear and non linear systems. Our approach utilizes a certain type of fuzzy systems that are based on Takagi-Sugeno (T-S) fuzzy models to approximate nonlinear systems. We use a robust control methodology to design controllers. This method not only guarantees stability, but also minimizes an upper bound on a linear quadratic performance measure. A simulation example is presented to show the effectiveness of this method.
Keywords: Takagi-Sugeno fuzzy model, state feedback, linear matrix inequalities, robust stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25004718 A Frequency Grouping Approach for Blind Deconvolution of Fairly Motionless Sources
Authors: E. S. Gower, T. Tsalaile, E. Rakgati, M. O. J. Hawksford
Abstract:
A frequency grouping approach for multi-channel instantaneous blind source separation (I-BSS) of convolutive mixtures is proposed for a lower net residual inter-symbol interference (ISI) and inter-channel interference (ICI) than the conventional short-time Fourier transform (STFT) approach. Starting in the time domain, STFTs are taken with overlapping windows to convert the convolutive mixing problem into frequency domain instantaneous mixing. Mixture samples at the same frequency but from different STFT windows are grouped together forming unique frequency groups. The individual frequency group vectors are input to the I-BSS algorithm of choice, from which the output samples are dispersed back to their respective STFT windows. After applying the inverse STFT, the resulting time domain signals are used to construct the complete source estimates via the weighted overlap-add method (WOLA). The proposed algorithm is tested for source deconvolution given two mixtures, and simulated along with the STFT approach to illustrate its superiority for fairly motionless sources.Keywords: Blind source separation, short-time Fouriertransform, weighted overlap-add method
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15264717 K-best Night Vision Devices by Multi-Criteria Mixed-Integer Optimization Modeling
Authors: Daniela I. Borissova, Ivan C. Mustakerov
Abstract:
The paper describes an approach for defining of k-best night vision devices based on multi-criteria mixed-integer optimization modeling. The parameters of night vision devices are considered as criteria that have to be optimized. Using different user preferences for the relative importance between parameters different choice of k-best devices can be defined. An ideal device with all of its parameters at their optimum is used to determine how far the particular device from the ideal one is. A procedure for evaluation of deviation between ideal solution and k-best solutions is presented. The applicability of the proposed approach is numerically illustrated using real night vision devices data. The proposed approach contributes to quality of decisions about choice of night vision devices by making the decision making process more certain, rational and efficient.
Keywords: K-best devices, mixed-integer model, multi-criteria problem, night vision devices.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18024716 A Probabilistic Optimization Approach for a Gas Processing Plant under Uncertain Feed Conditions and Product Requirements
Authors: G. Mesfin, M. Shuhaimi
Abstract:
This paper proposes a new optimization techniques for the optimization a gas processing plant uncertain feed and product flows. The problem is first formulated using a continuous linear deterministic approach. Subsequently, the single and joint chance constraint models for steady state process with timedependent uncertainties have been developed. The solution approach is based on converting the probabilistic problems into their equivalent deterministic form and solved at different confidence levels Case study for a real plant operation has been used to effectively implement the proposed model. The optimization results indicate that prior decision has to be made for in-operating plant under uncertain feed and product flows by satisfying all the constraints at 95% confidence level for single chance constrained and 85% confidence level for joint chance constrained optimizations cases.Keywords: Butane, Feed composition, LPG, Productspecification, Propane.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13964715 Autonomously Determining the Parameters for SVDD with RBF Kernel from a One-Class Training Set
Authors: Andreas Theissler, Ian Dear
Abstract:
The one-class support vector machine “support vector data description” (SVDD) is an ideal approach for anomaly or outlier detection. However, for the applicability of SVDD in real-world applications, the ease of use is crucial. The results of SVDD are massively determined by the choice of the regularisation parameter C and the kernel parameter of the widely used RBF kernel. While for two-class SVMs the parameters can be tuned using cross-validation based on the confusion matrix, for a one-class SVM this is not possible, because only true positives and false negatives can occur during training. This paper proposes an approach to find the optimal set of parameters for SVDD solely based on a training set from one class and without any user parameterisation. Results on artificial and real data sets are presented, underpinning the usefulness of the approach.
Keywords: Support vector data description, anomaly detection, one-class classification, parameter tuning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29344714 Unified Fusion Approach with Application to SLAM
Authors: Xinde Li, Xinhan Huang, Min Wang
Abstract:
In this paper, we propose the pre-processor based on the Evidence Supporting Measure of Similarity (ESMS) filter and also propose the unified fusion approach (UFA) based on the general fusion machine coupled with ESMS filter, which improve the correctness and precision of information fusion in any fields of application. Here we mainly apply the new approach to Simultaneous Localization And Mapping (SLAM) of Pioneer II mobile robots. A simulation experiment was performed, where an autonomous virtual mobile robot with sonar sensors evolves in a virtual world map with obstacles. By comparing the result of building map according to the general fusion machine (here DSmT-based fusing machine and PCR5-based conflict redistributor considereded) coupling with ESMS filter and without ESMS filter, it shows the benefit of the selection of the sources as a prerequisite for improvement of the information fusion, and also testifies the superiority of the UFA in dealing with SLAM.Keywords: DSmT, ESMS filter, SLAM, UFA
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13494713 Multivariate Output-Associative RVM for Multi-Dimensional Affect Predictions
Authors: Achut Manandhar, Kenneth D. Morton, Peter A. Torrione, Leslie M. Collins
Abstract:
The current trends in affect recognition research are to consider continuous observations from spontaneous natural interactions in people using multiple feature modalities, and to represent affect in terms of continuous dimensions, incorporate spatio-temporal correlation among affect dimensions, and provide fast affect predictions. These research efforts have been propelled by a growing effort to develop affect recognition system that can be implemented to enable seamless real-time human-computer interaction in a wide variety of applications. Motivated by these desired attributes of an affect recognition system, in this work a multi-dimensional affect prediction approach is proposed by integrating multivariate Relevance Vector Machine (MVRVM) with a recently developed Output-associative Relevance Vector Machine (OARVM) approach. The resulting approach can provide fast continuous affect predictions by jointly modeling the multiple affect dimensions and their correlations. Experiments on the RECOLA database show that the proposed approach performs competitively with the OARVM while providing faster predictions during testing.Keywords: Dimensional affect prediction, Output-associative RVM, Multivariate regression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16674712 Specialized Translation Teaching Strategies: A Corpus-Based Approach
Authors: Yingying Ding
Abstract:
This study presents a methodology of specialized translation with the objective of helping teachers to improve the strategies in teaching translation. In order to allow students to acquire skills to translate specialized texts, they need to become familiar with the semantic and syntactic features of source texts and target texts. The aim of our study is to use a corpus-based approach in the teaching of specialized translation between Chinese and Italian. This study proposes to construct a specialized Chinese - Italian comparable corpus that consists of 50 economic contracts from the domain of food. With the help of AntConc, we propose to compile a comparable corpus in for translation teaching purposes. This paper attempts to provide insight into how teachers could benefit from comparable corpus in the teaching of specialized translation from Italian into Chinese and through some examples of passive sentences how students could learn to apply different strategies for translating appropriately the voice.
Keywords: Corpus-based approach, translation teaching, specialized translation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12664711 Probe Selection for Pathway-Specific Microarray Probe Design Minimizing Melting Temperature Variance
Authors: Fabian Horn, Reinhard Guthke
Abstract:
In molecular biology, microarray technology is widely and successfully utilized to efficiently measure gene activity. If working with less studied organisms, methods to design custom-made microarray probes are available. One design criterion is to select probes with minimal melting temperature variances thus ensuring similar hybridization properties. If the microarray application focuses on the investigation of metabolic pathways, it is not necessary to cover the whole genome. It is more efficient to cover each metabolic pathway with a limited number of genes. Firstly, an approach is presented which minimizes the overall melting temperature variance of selected probes for all genes of interest. Secondly, the approach is extended to include the additional constraints of covering all pathways with a limited number of genes while minimizing the overall variance. The new optimization problem is solved by a bottom-up programming approach which reduces the complexity to make it computationally feasible. The new method is exemplary applied for the selection of microarray probes in order to cover all fungal secondary metabolite gene clusters for Aspergillus terreus.
Keywords: bottom-up approach, gene clusters, melting temperature, metabolic pathway, microarray probe design, probe selection
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15584710 Model-Based Software Regression Test Suite Reduction
Authors: Shiwei Deng, Yang Bao
Abstract:
In this paper, we present a model-based regression test suite reducing approach that uses EFSM model dependence analysis and probability-driven greedy algorithm to reduce software regression test suites. The approach automatically identifies the difference between the original model and the modified model as a set of elementary model modifications. The EFSM dependence analysis is performed for each elementary modification to reduce the regression test suite, and then the probability-driven greedy algorithm is adopted to select the minimum set of test cases from the reduced regression test suite that cover all interaction patterns. Our initial experience shows that the approach may significantly reduce the size of regression test suites.Keywords: Dependence analysis, EFSM model, greedy algorithm, regression test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19204709 A New Approach to Optimal Control Problem Constrained by Canonical Form
Authors: B. Farhadinia
Abstract:
In this article, it is considered a class of optimal control problems constrained by differential and integral constraints are called canonical form. A modified measure theoretical approach is introduced to solve this class of optimal control problems.Keywords: control problem, Canonical form, Measure theory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12014708 A Hybrid Ontology Based Approach for Ranking Documents
Authors: Sarah Motiee, Azadeh Nematzadeh, Mehrnoush Shamsfard
Abstract:
Increasing growth of information volume in the internet causes an increasing need to develop new (semi)automatic methods for retrieval of documents and ranking them according to their relevance to the user query. In this paper, after a brief review on ranking models, a new ontology based approach for ranking HTML documents is proposed and evaluated in various circumstances. Our approach is a combination of conceptual, statistical and linguistic methods. This combination reserves the precision of ranking without loosing the speed. Our approach exploits natural language processing techniques to extract phrases from documents and the query and doing stemming on words. Then an ontology based conceptual method will be used to annotate documents and expand the query. To expand a query the spread activation algorithm is improved so that the expansion can be done flexible and in various aspects. The annotated documents and the expanded query will be processed to compute the relevance degree exploiting statistical methods. The outstanding features of our approach are (1) combining conceptual, statistical and linguistic features of documents, (2) expanding the query with its related concepts before comparing to documents, (3) extracting and using both words and phrases to compute relevance degree, (4) improving the spread activation algorithm to do the expansion based on weighted combination of different conceptual relationships and (5) allowing variable document vector dimensions. A ranking system called ORank is developed to implement and test the proposed model. The test results will be included at the end of the paper.Keywords: Document ranking, Ontology, Spread activation algorithm, Annotation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16294707 Optimal Power Allocation to Diversity Branches of Cooperative MISO Sensor Networks
Authors: Rooholah Hasanizadeh, Saadan Zokaei
Abstract:
In the context of sensor networks, where every few dB saving counts, the novel node cooperation schemes are reviewed where MIMO techniques play a leading role. These methods could be treated as joint approach for designing physical layer of their communication scenarios. Then we analyzed the BER performance of transmission diversity schemes under a general fading channel model and proposed a power allocation strategy to the transmitting sensor nodes. This approach is then compared to an equal-power assignment method and its performance enhancement is verified by the simulation. Another key point of the contribution lies in the combination of optimal power allocation and sensor nodes- cooperation in a transmission diversity regime (MISO). Numerical results are given through figures to demonstrate the optimality and efficiency of proposed combined approach.Keywords: Optimal power allocation, cooperative MISO scheme, sensor networks, diversity branch.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14104706 Dynamic Variational Multiscale LES of Bluff Body Flows on Unstructured Grids
Authors: Carine Moussaed, Stephen Wornom, Bruno Koobus, Maria Vittoria Salvetti, Alain Dervieux,
Abstract:
The effects of dynamic subgrid scale (SGS) models are investigated in variational multiscale (VMS) LES simulations of bluff body flows. The spatial discretization is based on a mixed finite element/finite volume formulation on unstructured grids. In the VMS approach used in this work, the separation between the largest and the smallest resolved scales is obtained through a variational projection operator and a finite volume cell agglomeration. The dynamic version of Smagorinsky and WALE SGS models are used to account for the effects of the unresolved scales. In the VMS approach, these effects are only modeled in the smallest resolved scales. The dynamic VMS-LES approach is applied to the simulation of the flow around a circular cylinder at Reynolds numbers 3900 and 20000 and to the flow around a square cylinder at Reynolds numbers 22000 and 175000. It is observed as in previous studies that the dynamic SGS procedure has a smaller impact on the results within the VMS approach than in LES. But improvements are demonstrated for important feature like recirculating part of the flow. The global prediction is improved for a small computational extra cost.Keywords: variational multiscale LES, dynamic SGS model, unstructured grids, circular cylinder, square cylinder.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18234705 Hybrid Reliability-Similarity-Based Approach for Supervised Machine Learning
Authors: Walid Cherif
Abstract:
Data mining has, over recent years, seen big advances because of the spread of internet, which generates everyday a tremendous volume of data, and also the immense advances in technologies which facilitate the analysis of these data. In particular, classification techniques are a subdomain of Data Mining which determines in which group each data instance is related within a given dataset. It is used to classify data into different classes according to desired criteria. Generally, a classification technique is either statistical or machine learning. Each type of these techniques has its own limits. Nowadays, current data are becoming increasingly heterogeneous; consequently, current classification techniques are encountering many difficulties. This paper defines new measure functions to quantify the resemblance between instances and then combines them in a new approach which is different from actual algorithms by its reliability computations. Results of the proposed approach exceeded most common classification techniques with an f-measure exceeding 97% on the IRIS Dataset.
Keywords: Data mining, knowledge discovery, machine learning, similarity measurement, supervised classification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15264704 Classification and Resolving Urban Problems by Means of Fuzzy Approach
Authors: F. Habib, A. Shokoohi
Abstract:
Urban problems are problems of organized complexity. Thus, many models and scientific methods to resolve urban problems are failed. This study is concerned with proposing of a fuzzy system driven approach for classification and solving urban problems. The proposed study investigated mainly the selection of the inputs and outputs of urban systems for classification of urban problems. In this research, five categories of urban problems, respect to fuzzy system approach had been recognized: control, polytely, optimizing, open and decision making problems. Grounded Theory techniques were then applied to analyze the data and develop new solving method for each category. The findings indicate that the fuzzy system methods are powerful processes and analytic tools for helping planners to resolve urban complex problems. These tools can be successful where as others have failed because both incorporate or address uncertainty and risk; complexity and systems interacting with other systems.
Keywords: Classification, complexity, Fuzzy theory, urban problems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21114703 Defect-Based Urgency Index for Bridge Maintenance Ranking and Prioritization
Authors: Saleh Abu Dabous, Khaled Hamad, Rami Al-Ruzouq
Abstract:
Bridge condition assessment and rating provide essential information needed for bridge management. This paper reviews bridge inspection and condition rating practices and introduces a defect-based urgency index. The index is estimated at the element-level based on the extent and severity of the different defects typical to the bridge element. The urgency index approach has the following advantages: (1) It facilitates judgment submission, i.e. instead of rating the bridge element with a specific linguistic overall expression (which can be subjective and used differently by different people), the approach is based on assessing the defects; (2) It captures multiple defects that can be present within a deteriorated element; and (3) It reflects how critical the element is through quantifying critical defects and their severity. The approach can be further developed and validated. It is expected to be useful for practical purposes as an early-warning system for critical bridge elements.
Keywords: Condition rating, deterioration, inspection, maintenance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18894702 An Event Based Approach to Extract the Run Time Execution Path of BPEL Process for Monitoring QoS in the Cloud
Authors: Rima Grati, Khouloud Boukadi, Hanene Ben-Abdallah
Abstract:
Due to the dynamic nature of the Cloud, continuous monitoring of QoS requirements is necessary to manage the Cloud computing environment. The process of QoS monitoring and SLA violation detection consists of: collecting low and high level information pertinent to the service, analyzing the collected information, and taking corrective actions when SLA violations are detected. In this paper, we detail the architecture and the implementation of the first step of this process. More specifically, we propose an event-based approach to obtain run time information of services developed as BPEL processes. By catching particular events (i.e., the low level information), our approach recognizes the run-time execution path of a monitored service and uses the BPEL execution patterns to compute QoS of the composite service (i.e., the high level information).
Keywords: Monitoring of Web service composition, Cloud environment, Run-time extraction of execution path of BPEL.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1685