Search results for: M5 decision tree model
7507 Approximately Similarity Measurement of Web Sites Using Genetic Algorithms and Binary Trees
Authors: Doru Anastasiu Popescu, Dan Rădulescu
Abstract:
In this paper, we determine the similarity of two HTML web applications. We are going to use a genetic algorithm in order to determine the most significant web pages of each application (we are not going to use every web page of a site). Using these significant web pages, we will find the similarity value between the two applications. The algorithm is going to be efficient because we are going to use a reduced number of web pages for comparisons but it will return an approximate value of the similarity. The binary trees are used to keep the tags from the significant pages. The algorithm was implemented in Java language.
Keywords: Tag, HTML, web page, genetic algorithm, similarity value, binary tree.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13107506 Analysis of Model in Pregnant and Non-Pregnant Dengue Patients
Authors: R. Kongnuy, P. Pongsumpun
Abstract:
We used mathematical model to study the transmission of dengue disease. The model is developed in which the human population is separated into two populations, pregnant and non-pregnant humans. The dynamical analysis method is used for analyzing this modified model. Two equilibrium states are found and the conditions for stability of theses two equilibrium states are established. Numerical results are shown for each equilibrium state. The basic reproduction numbers are found and they are compared by using numerical simulations.Keywords: Basic reproductive number, dengue disease, equilibrium states, pregnancy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15937505 Measuring the Structural Similarity of Web-based Documents: A Novel Approach
Authors: Matthias Dehmer, Frank Emmert Streib, Alexander Mehler, Jürgen Kilian
Abstract:
Most known methods for measuring the structural similarity of document structures are based on, e.g., tag measures, path metrics and tree measures in terms of their DOM-Trees. Other methods measures the similarity in the framework of the well known vector space model. In contrast to these we present a new approach to measuring the structural similarity of web-based documents represented by so called generalized trees which are more general than DOM-Trees which represent only directed rooted trees.We will design a new similarity measure for graphs representing web-based hypertext structures. Our similarity measure is mainly based on a novel representation of a graph as strings of linear integers, whose components represent structural properties of the graph. The similarity of two graphs is then defined as the optimal alignment of the underlying property strings. In this paper we apply the well known technique of sequence alignments to solve a novel and challenging problem: Measuring the structural similarity of generalized trees. More precisely, we first transform our graphs considered as high dimensional objects in linear structures. Then we derive similarity values from the alignments of the property strings in order to measure the structural similarity of generalized trees. Hence, we transform a graph similarity problem to a string similarity problem. We demonstrate that our similarity measure captures important structural information by applying it to two different test sets consisting of graphs representing web-based documents.
Keywords: Graph similarity, hierarchical and directed graphs, hypertext, generalized trees, web structure mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25587504 ANN Based Model Development for Material Removal Rate in Dry Turning in Indian Context
Authors: Mangesh R. Phate, V. H. Tatwawadi
Abstract:
This paper is intended to develop an artificial neural network (ANN) based model of material removal rate (MRR) in the turning of ferrous and nonferrous material in a Indian small-scale industry. MRR of the formulated model was proved with the testing data and artificial neural network (ANN) model was developed for the analysis and prediction of the relationship between inputs and output parameters during the turning of ferrous and nonferrous materials. The input parameters of this model are operator, work-piece, cutting process, cutting tool, machine and the environment.
The ANN model consists of a three layered feedforward back propagation neural network. The network is trained with pairs of independent/dependent datasets generated when machining ferrous and nonferrous material. A very good performance of the neural network, in terms of contract with experimental data, was achieved. The model may be used for the testing and forecast of the complex relationship between dependent and the independent parameters in turning operations.
Keywords: Field data based model, Artificial neural network, Simulation, Convectional Turning, Material removal rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19707503 Parameters Extraction for Pseudomorphic HEMTs Using Genetic Algorithms
Authors: Mazhar B. Tayel, Amr H. Yassin
Abstract:
A proposed small-signal model parameters for a pseudomorphic high electron mobility transistor (PHEMT) is presented. Both extrinsic and intrinsic circuit elements of a smallsignal model are determined using genetic algorithm (GA) as a stochastic global search and optimization tool. The parameters extraction of the small-signal model is performed on 200-μm gate width AlGaAs/InGaAs PHEMT. The equivalent circuit elements for a proposed 18 elements model are determined directly from the measured S- parameters. The GA is used to extract the parameters of the proposed small-signal model from 0.5 up to 18 GHz.
Keywords: PHEMT, Genetic Algorithms, small signal modeling, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22627502 On-line Control of the Natural and Anthropogenic Safety in Krasnoyarsk Region
Authors: T. Penkova, A. Korobko, V. Nicheporchuk., L. Nozhenkova, A. Metus
Abstract:
This paper presents an approach of on-line control of the state of technosphere and environment objects based on the integration of Data Warehouse, OLAP and Expert systems technologies. It looks at the structure and content of data warehouse that provides consolidation and storage of monitoring data. There is a description of OLAP-models that provide a multidimensional analysis of monitoring data and dynamic analysis of principal parameters of controlled objects. The authors suggest some criteria of emergency risk assessment using expert knowledge about danger levels. It is demonstrated now some of the proposed solutions could be adopted in territorial decision making support systems. Operational control allows authorities to detect threat, prevent natural and anthropogenic emergencies and ensure a comprehensive safety of territory.Keywords: Decision making support systems, Emergency risk assessment, Natural and anthropogenic safety, On-line control, Territory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18917501 A BERT-Based Model for Financial Social Media Sentiment Analysis
Authors: Josiel Delgadillo, Johnson Kinyua, Charles Mutigwe
Abstract:
The purpose of sentiment analysis is to determine the sentiment strength (e.g., positive, negative, neutral) from a textual source for good decision-making. Natural Language Processing (NLP) in domains such as financial markets requires knowledge of domain ontology, and pre-trained language models, such as BERT, have made significant breakthroughs in various NLP tasks by training on large-scale un-labeled generic corpora such as Wikipedia. However, sentiment analysis is a strong domain-dependent task. The rapid growth of social media has given users a platform to share their experiences and views about products, services, and processes, including financial markets. StockTwits and Twitter are social networks that allow the public to express their sentiments in real time. Hence, leveraging the success of unsupervised pre-training and a large amount of financial text available on social media platforms could potentially benefit a wide range of financial applications. This work is focused on sentiment analysis using social media text on platforms such as StockTwits and Twitter. To meet this need, SkyBERT, a domain-specific language model pre-trained and fine-tuned on financial corpora, has been developed. The results show that SkyBERT outperforms current state-of-the-art models in financial sentiment analysis. Extensive experimental results demonstrate the effectiveness and robustness of SkyBERT.
Keywords: BERT, financial markets, Twitter, sentiment analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7177500 An Exploration of Cross-Cultural Behaviour: The Characteristics of Chinese Consumers’ Decision Making in Europe
Authors: Yongsheng Guo, Xiaoxian Zhu, Mandella Osei-Assibey Bonsu
Abstract:
This study explores the effects of national culture on consumer behaviour by identifying the characteristics of Chinese consumers’ decision making in Europe. It offers a better understanding of how cultural factors affect consumers’ behaviour, and how consumers make decisions in other nations with different culture. It adopted a grounded theory approach and conducted 24 in-depth interviews. Grounded theory models are developed to link the causal conditions, process, and consequences. Results reveal that some cultural factors including conservatism, emotionality, acquaintance community, long-term orientation and principles affect Chinese consumers when making purchase decisions in Europe. Most Chinese consumers plan and prepare their expenditure and stay in Europe as cultural learners, and purchase durable products or assets as investment, and share their experiences within a community. This study identified potential problems such as political and social environment, complex procedures, and restrictions. This study found that external factors influence internal factors and then internal characters determine consumer behaviour. This study proposes that cultural traits developed in convergence evolution through social selection and Chinese consumers persist most characters but adapt some perceptions and actions overtime in other countries. This study suggests that cultural marketing could be adopted by companies to reflect consumers’ preferences. Agencies, shops, and the authorities could take actions to reduce the complexity and restrictions.
Keywords: National culture, consumer behaviour, cultural marketing, decision making.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4787499 The Relationship between Business-model Innovation and Firm Value: A Dynamic Perspective
Authors: Yung C. Ho, Hui C. Fang, Ming J. Hsieh
Abstract:
When consistently innovative business-models can give companies a competitive advantage, longitudinal empirical research, which can reflect dynamic business-model changes, has yet to prove a definitive connection. This study consequently employs a dynamic perspective in conjunction with innovation theory to examine the relationship between the types of business-model innovation and firm value. This study tries to examine various types of business-model innovation in high-end and low-end technology industries such as HTC and the 7-Eleven chain stores with research periods of 14 years and 32 years, respectively. The empirical results suggest that adopting radical business-model innovation in addition to expanding new target markets can successfully lead to a competitive advantage. Sustained advanced technological competences and service/product innovation are the key successful factors in high-end and low-end technology industry business-models respectively. In sum up, the business-model innovation can yield a higher market value and financial value in high-end technology industries than low-end ones.Keywords: Business-model, Dynamic Perspective, Firm Value, Innovation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27447498 Manual Testing of Web Software Systems Supported by Direct Guidance of the Tester Based On Design Model
Authors: Karel Frajtak, Miroslav Bures, Ivan Jelinek
Abstract:
Software testing is important stage of software development cycle. Current testing process involves tester and electronic documents with test case scenarios. In this paper we focus on new approach to testing process using automated test case generation and tester guidance through the system based on the model of the system. Test case generation and model-based testing is not possible without proper system model. We aim on providing better feedback from the testing process thus eliminating the unnecessary paper work.
Keywords: Model based testing, test automation, test generating, tester support.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19597497 Evolution of Quality Function Deployment (QFD) via Fuzzy Concepts and Neural Networks
Authors: M. Haghighi, M. Zowghi, B. Zohouri
Abstract:
Quality Function Deployment (QFD) is an expounded, multi-step planning method for delivering commodity, services, and processes to customers, both external and internal to an organization. It is a way to convert between the diverse customer languages expressing demands (Voice of the Customer), and the organization-s languages expressing results that sate those demands. The policy is to establish one or more matrices that inter-relate producer and consumer reciprocal expectations. Due to its visual presence is called the “House of Quality" (HOQ). In this paper, we assumed HOQ in multi attribute decision making (MADM) pattern and through a proposed MADM method, rank technical specifications. Thereafter compute satisfaction degree of customer requirements and for it, we apply vagueness and uncertainty conditions in decision making by fuzzy set theory. This approach would propound supervised neural network (perceptron) for MADM problem solving.
Keywords: MADM, fuzzy set, QFD, supervised neural network (perceptron).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17427496 An Extension of Multi-Layer Perceptron Based on Layer-Topology
Authors: Jānis Zuters
Abstract:
There are a lot of extensions made to the classic model of multi-layer perceptron (MLP). A notable amount of them has been designed to hasten the learning process without considering the quality of generalization. The paper proposes a new MLP extension based on exploiting topology of the input layer of the network. Experimental results show the extended model to improve upon generalization capability in certain cases. The new model requires additional computational resources to compare to the classic model, nevertheless the loss in efficiency isn-t regarded to be significant.
Keywords: Learning algorithm, multi-layer perceptron, topology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15127495 Geometric and Material Nonlinear Analysis of Reinforced Concrete Structure Considering Soil-Structure Interaction
Authors: Mohamed M. El-Gendy, Ibrahim A. El-Arabi, Rafik W. Abdel-Missih, Omar A. Kandil
Abstract:
In the present research, a finite element model is presented to study the geometrical and material nonlinear behavior of reinforced concrete plane frames considering soil-structure interaction. The nonlinear behaviors of concrete and reinforcing steel are considered both in compression and tension up to failure. The model takes account also for the number, diameter, and distribution of rebar along every cross section. Soil behavior is taken into consideration using four different models; namely: linear-, nonlinear Winkler's model, and linear-, nonlinear continuum model. A computer program (NARC) is specially developed in order to perform the analysis. The results achieved by the present model show good agreement with both theoretical and experimental published literature. The nonlinear behavior of a rectangular frame resting on soft soil up to failure using the proposed model is introduced for demonstration.Keywords: Nonlinear analysis, Geometric nonlinearity, Material nonlinearity, Reinforced concrete, Finite element method, Soilstructure interaction, Winkler's soil model, Continuum soil model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26677494 A Methodological Test to Study the Concrete Workability with the Fractal Model
Authors: F. Achouri, K. Chouicha
Abstract:
The main parameters affecting the workability are the water content, particle size, and the total surface of the grains, as long as the mixing water begins by wetting the surface of the grains and then fills the voids between the grains to form entrapped water, the quantity of water remaining is called free water. The aim of this study is to undertake a fractal approach through the relationship between the concrete formulation parameters and workability. To develop this approach a series of concrete taken from the literature was investigated by varying formulation parameters such as G/S, the quantity of cement C and the quantity of water W. We also call another model as the model of water layer thickness and model of paste layer thickness to judge their relevance, hence the following results: the relevance of the water layer thickness model is considered as a relevant when there is a variation in the water quantity. The model of the paste layer thickness is only applicable if we considered that the paste is made with the grain value Dmax = 2.85: value from which we see a stability of the model.Keywords: Concrete, fractal method, paste layer thickness, water layer thickness, workability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16367493 Development of a Catchment Water Quality Model for Continuous Simulations of Pollutants Build-up and Wash-off
Authors: Iqbal Hossain, Dr. Monzur Imteaz, Dr. Shirley Gato-Trinidad, Prof. Abdallah Shanableh
Abstract:
Estimation of runoff water quality parameters is required to determine appropriate water quality management options. Various models are used to estimate runoff water quality parameters. However, most models provide event-based estimates of water quality parameters for specific sites. The work presented in this paper describes the development of a model that continuously simulates the accumulation and wash-off of water quality pollutants in a catchment. The model allows estimation of pollutants build-up during dry periods and pollutants wash-off during storm events. The model was developed by integrating two individual models; rainfall-runoff model, and catchment water quality model. The rainfall-runoff model is based on the time-area runoff estimation method. The model allows users to estimate the time of concentration using a range of established methods. The model also allows estimation of the continuing runoff losses using any of the available estimation methods (i.e., constant, linearly varying or exponentially varying). Pollutants build-up in a catchment was represented by one of three pre-defined functions; power, exponential, or saturation. Similarly, pollutants wash-off was represented by one of three different functions; power, rating-curve, or exponential. The developed runoff water quality model was set-up to simulate the build-up and wash-off of total suspended solids (TSS), total phosphorus (TP) and total nitrogen (TN). The application of the model was demonstrated using available runoff and TSS field data from road and roof surfaces in the Gold Coast, Australia. The model provided excellent representation of the field data demonstrating the simplicity yet effectiveness of the proposed model.
Keywords: Catchment, continuous pollutants build-up, pollutants wash-off, runoff, runoff water quality model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31367492 Approaches to Determining Optimal Asset Structure for a Commercial Bank
Authors: Svetlana Saksonova
Abstract:
Every commercial bank optimises its asset portfolio depending on the profitability of assets and chosen or imposed constraints. This paper proposes and applies a stylized model for optimising banks' asset and liability structure, reflecting profitability of different asset categories and their risks as well as costs associated with different liability categories and reserve requirements. The level of detail for asset and liability categories is chosen to create a suitably parsimonious model and to include the most important categories in the model. It is shown that the most appropriate optimisation criterion for the model is the maximisation of the ratio of net interest income to assets. The maximisation of this ratio is subject to several constraints. Some are accounting identities or dictated by legislative requirements; others vary depending on the market objectives for a particular bank. The model predicts variable amount of assets allocated to loan provision.Keywords: asset structure, commercial bank, model, optimisation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29697491 Dynamic Analyses for Passenger Volume of Domestic Airline and High Speed Rail
Authors: Shih-Ching Lo
Abstract:
Discrete choice model is the most used methodology for studying traveler-s mode choice and demand. However, to calibrate the discrete choice model needs to have plenty of questionnaire survey. In this study, an aggregative model is proposed. The historical data of passenger volumes for high speed rail and domestic civil aviation are employed to calibrate and validate the model. In this study, different models are compared so as to propose the best one. From the results, systematic equations forecast better than single equation do. Models with the external variable, which is oil price, are better than models based on closed system assumption.
Keywords: forecasting, passenger volume, dynamic competition model, external variable, oil price
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14637490 An Erosion-based Modeling of Abrasive Waterjet Turning
Authors: I. Zohourkari, M. Zohoor
Abstract:
In this paper, an erosion-based model for abrasive waterjet (AWJ) turning process is presented. By using modified Hashish erosion model, the volume of material removed by impacting of abrasive particles to surface of the rotating cylindrical specimen is estimated and radius reduction at each rotation is calculated. Different to previous works, the proposed model considers the continuous change in local impact angle due to change in workpiece diameter, axial traverse rate of the jet, the abrasive particle roundness and density. The accuracy of the proposed model is examined by experimental tests under various traverse rates. The final diameters estimated by the proposed model are in good accordance with experiments.Keywords: Abrasive, Erosion, impact, Particle, Waterjet, Turning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24237489 A Dual Model for Efficiency Evaluation Considering Time Lag Effect
Authors: Yan Shuang Zhang, Taehan Lee, Byung Ho Jeong
Abstract:
A DEA model can generally evaluate the performance using multiple inputs and outputs for the same period. However, it is hard to avoid the production lead time phenomenon some times, such as long-term project or marketing activity. A couple of models have been suggested to capture this time lag issue in the context of DEA. This paper develops a dual-MPO model to deal with time lag effect in evaluating efficiency. A numerical example is also given to show that the proposed model can be used to get efficiency and reference set of inefficient DMUs and to obtain projected target value of input attributes for inefficient DMUs to be efficient.
Keywords: DEA, efficiency, time lag, dual problem.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18097488 Dynamical Transmission Model of Chikungunya in Thailand
Authors: P. Pongsumpun
Abstract:
One of the important tropical diseases is Chikunkunya. This disease is transmitted between the human by the insect-borne virus, of the genus Alphavirus. It occurs in Africa, Asia and the Indian subcontinent. In Thailand, the incidences due to this disease are increasing every year. In this study, the transmission of this disease is studied through dynamical model analysis.Keywords: Chikunkunya, dynamical model, Endemic region, Routh-Hurwitz criteria.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13707487 Reference Model for the Implementation of an E-Commerce Solution in Peruvian SMEs in the Retail Sector
Authors: Julio Kauss, Miguel Cadillo, David Mauricio
Abstract:
E-commerce is a business model that allows companies to optimize the processes of buying, selling, transferring goods and exchanging services through computer networks or the Internet. In Peru, the electronic commerce is used infrequently. This situation is due, in part to the fact that there is no model that allows companies to implement an e-commerce solution, which means that most SMEs do not have adequate knowledge to adapt to electronic commerce. In this work, a reference model is proposed for the implementation of an e-commerce solution in Peruvian SMEs in the retail sector. It consists of five phases: Business Analysis, Business Modeling, Implementation, Post Implementation and Results. The present model was validated in a SME of the Peruvian retail sector through the implementation of an electronic commerce platform, through which the company increased its sales through the delivery channel by 10% in the first month of deployment. This result showed that the model is easy to implement, is economical and agile. In addition, it allowed the company to increase its business offer, adapt to e-commerce and improve customer loyalty.
Keywords: E-commerce, retail, SMEs, reference model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10507486 An Evaluation of Barriers to Implement Reverse Logistics: A Case Study of Indian Fastener Industry
Authors: D. Garg, S. Luthra, A. Haleem
Abstract:
Reverse logistics (RL) is supposed to be a systematic procedure that helps in improving the environmental hazards and maintain business sustainability for industries. Industries in Indian are now opting for adoption of RL techniques in business. But, RL practices are not popular in Indian industries because of many barriers for its successful implementation. Therefore, need arises to identify and evaluate the barriers to implement RL practices by taking an Indian industries perspective. Literature review approach and case study approach have been adapted to identify relevant barriers to implement RL practices. Further, Fuzzy Decision Making Trial and Evaluation Laboratory methodology has been brought into use for evaluating causal relationships among the barriers to implement RL practices. Seven barriers out of ten barriers have been categorized into the cause group and remaining into effect group. This research will help Indian industries to manage these barriers towards effective implementing RL practices.Keywords: Barriers, decision making trial and evaluation laboratory, fuzzy set theory, Indian industries, reverse logistics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21907485 Kinetic Façade Design Using 3D Scanning to Convert Physical Models into Digital Models
Authors: Do-Jin Jang, Sung-Ah Kim
Abstract:
In designing a kinetic façade, it is hard for the designer to make digital models due to its complex geometry with motion. This paper aims to present a methodology of converting a point cloud of a physical model into a single digital model with a certain topology and motion. The method uses a Microsoft Kinect sensor, and color markers were defined and applied to three paper folding-inspired designs. Although the resulted digital model cannot represent the whole folding range of the physical model, the method supports the designer to conduct a performance-oriented design process with the rough physical model in the reduced folding range.
Keywords: Design media, kinetic façades, tangible user interface, 3D scanning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14207484 Long Term Examination of the Profitability Estimation Focused on Benefits
Authors: Stephan Printz, Kristina Lahl, René Vossen, Sabina Jeschke
Abstract:
Strategic investment decisions are characterized by high innovation potential and long-term effects on the competitiveness of enterprises. Due to the uncertainty and risks involved in this complex decision making process, the need arises for well-structured support activities. A method that considers cost and the long-term added value is the cost-benefit effectiveness estimation. One of those methods is the “profitability estimation focused on benefits – PEFB”-method developed at the Institute of Management Cybernetics at RWTH Aachen University. The method copes with the challenges associated with strategic investment decisions by integrating long-term non-monetary aspects whilst also mapping the chronological sequence of an investment within the organization’s target system. Thus, this method is characterized as a holistic approach for the evaluation of costs and benefits of an investment. This participation-oriented method was applied to business environments in many workshops. The results of the workshops are a library of more than 96 cost aspects, as well as 122 benefit aspects. These aspects are preprocessed and comparatively analyzed with regards to their alignment to a series of risk levels. For the first time, an accumulation and a distribution of cost and benefit aspects regarding their impact and probability of occurrence are given. The results give evidence that the PEFB-method combines precise measures of financial accounting with the incorporation of benefits. Finally, the results constitute the basics for using information technology and data science for decision support when applying within the PEFB-method.Keywords: Cost-benefit analysis, multi-criteria decision, profitability estimation focused on benefits, risk and uncertainty analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15007483 A Novel Model for Simultaneously Minimising Costs and Risks in Just-in-Time Systems Using Multi-Backup Suppliers: Part 2- Results
Authors: Faraj El Dabee, Romeo Marian, Yousef Amer
Abstract:
This paper implements the inventory model developed in the first part of this paper in a simplified problem to simultaneously reduce costs and risks in JIT systems. This model is developed to ascertain an optimal ordering strategy for procuring raw materials by using regular multi-external and local backup suppliers to reduce the total cost of the products, and at the same time to reduce the risks arising from this cost reduction within production systems. A comparison between the cost of using the JIT system and using the proposed inventory model shows the superiority of the use of the inventory model.
Keywords: Lean manufacturing, Just-in-Time (JIT), production system, cost-risk reduction, inventory model, eternal supplier, local backup supplier.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15477482 Opportunistic Routing with Secure Coded Wireless Multicast Using MAS Approach
Authors: E. Golden Julie, S. Tamil Selvi, Y. Harold Robinson
Abstract:
Many Wireless Sensor Network (WSN) applications necessitate secure multicast services for the purpose of broadcasting delay sensitive data like video files and live telecast at fixed time-slot. This work provides a novel method to deal with end-to-end delay and drop rate of packets. Opportunistic Routing chooses a link based on the maximum probability of packet delivery ratio. Null Key Generation helps in authenticating packets to the receiver. Markov Decision Process based Adaptive Scheduling algorithm determines the time slot for packet transmission. Both theoretical analysis and simulation results show that the proposed protocol ensures better performance in terms of packet delivery ratio, average end-to-end delay and normalized routing overhead.
Keywords: Delay-sensitive data, Markovian Decision Process based Adaptive Scheduling, Opportunistic Routing, Digital Signature authentication.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19577481 A New History Based Method to Handle the Recurring Concept Shifts in Data Streams
Authors: Hossein Morshedlou, Ahmad Abdollahzade Barforoush
Abstract:
Recent developments in storage technology and networking architectures have made it possible for broad areas of applications to rely on data streams for quick response and accurate decision making. Data streams are generated from events of real world so existence of associations, which are among the occurrence of these events in real world, among concepts of data streams is logical. Extraction of these hidden associations can be useful for prediction of subsequent concepts in concept shifting data streams. In this paper we present a new method for learning association among concepts of data stream and prediction of what the next concept will be. Knowing the next concept, an informed update of data model will be possible. The results of conducted experiments show that the proposed method is proper for classification of concept shifting data streams.Keywords: Data Stream, Classification, Concept Shift, History.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12787480 On Speeding Up Support Vector Machines: Proximity Graphs Versus Random Sampling for Pre-Selection Condensation
Authors: Xiaohua Liu, Juan F. Beltran, Nishant Mohanchandra, Godfried T. Toussaint
Abstract:
Support vector machines (SVMs) are considered to be the best machine learning algorithms for minimizing the predictive probability of misclassification. However, their drawback is that for large data sets the computation of the optimal decision boundary is a time consuming function of the size of the training set. Hence several methods have been proposed to speed up the SVM algorithm. Here three methods used to speed up the computation of the SVM classifiers are compared experimentally using a musical genre classification problem. The simplest method pre-selects a random sample of the data before the application of the SVM algorithm. Two additional methods use proximity graphs to pre-select data that are near the decision boundary. One uses k-Nearest Neighbor graphs and the other Relative Neighborhood Graphs to accomplish the task.Keywords: Machine learning, data mining, support vector machines, proximity graphs, relative-neighborhood graphs, k-nearestneighbor graphs, random sampling, training data condensation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19197479 Nonlinear Mathematical Model of the Rotor Motion in a Thin Hydrodynamic Gap
Authors: Jaroslav Krutil, František Pochylý, Simona Fialová
Abstract:
The article presents two mathematical models of the interaction between a rotating shaft and an incompressible fluid. The mathematical model includes both the journal bearings and the axially traversed hydrodynamic sealing gaps of hydraulic machines. A method is shown for the identification of additional effects of the fluid acting on the rotor of the machine, both for a linear and a nonlinear model. The interaction is expressed by matrices of mass, stiffness and damping.Keywords: CFD modeling, hydrodynamic gap, matrices of mass, stiffness and damping, nonlinear mathematical model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18427478 Multinomial Dirichlet Gaussian Process Model for Classification of Multidimensional Data
Authors: Wanhyun Cho, Soonja Kang, Sangkyoon Kim, Soonyoung Park
Abstract:
We present probabilistic multinomial Dirichlet classification model for multidimensional data and Gaussian process priors. Here, we have considered efficient computational method that can be used to obtain the approximate posteriors for latent variables and parameters needed to define the multiclass Gaussian process classification model. We first investigated the process of inducing a posterior distribution for various parameters and latent function by using the variational Bayesian approximations and important sampling method, and next we derived a predictive distribution of latent function needed to classify new samples. The proposed model is applied to classify the synthetic multivariate dataset in order to verify the performance of our model. Experiment result shows that our model is more accurate than the other approximation methods.Keywords: Multinomial dirichlet classification model, Gaussian process priors, variational Bayesian approximation, Importance sampling, approximate posterior distribution, Marginal likelihood evidence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1615