Search results for: partially observable Markov decision processes
9807 Reliability Analysis for the Functioning of Complete and Low Capacity MLDB Systems in Piston Plants
Authors: Ramanpreet Kaur, Upasana Sharma
Abstract:
The purpose of this paper is to address the challenges facing the water supply for the Machine Learning Database (MLDB) system at the piston foundry plant. In the MLDB system, one main unit, i.e., robotic, is connected by two sub-units. The functioning of the system depends on the robotic and water supply. Lack of water supply causes system failure. The system operates at full capacity with the help of two sub-units. If one sub-unit fails, the system runs at a low capacity. Reliability modeling is performed using semi-Markov processes and regenerative point techniques. Several system effects such as mean time to system failure, availability at full capacity, availability at reduced capacity, busy period for repair and expected number of visits have been achieved. Benefits have been analyzed. The graphical study is designed for a specific case using programming in C++ and MS Excel.Keywords: MLDB system, robotic, semi-Markov process, regenerative point technique
Procedia PDF Downloads 1039806 Intrapreneurship Discovery: Standard Strategy to Boost Innovation inside Companies
Authors: Chiara Mansanta, Daniela Sani
Abstract:
This paper studies the concept of intrapreneurship discovery for innovation and technology development related to the manufacturing industries set up in the center of Italy, in Marche Region. The study underlined the key drivers of the innovation process and the main factors that influence innovation. Starting from a literature study on open innovation, this paper examines the role of human capital to support company’s development. The empirical part of the study is based on a survey to 151 manufacturing companies that represent the 34% of that universe at the regional level. The survey underlined the main KPI’s that influence companies in their decision processes; then tools for these decision processes are presented.Keywords: business model, decision making, intrapreneurship discovery, standard methodology
Procedia PDF Downloads 1759805 Time Pressure and Its Effect at Tactical Level of Disaster Management
Authors: Agoston Restas
Abstract:
Introduction: In case of managing disasters decision makers can face many times such a special situation where any pre-sign of the drastically change is missing therefore the improvised decision making can be required. The complexity, ambiguity, uncertainty or the volatility of the situation can require many times the improvisation as decision making. It can be taken at any level of the management (strategic, operational and tactical) but at tactical level the main reason of the improvisation is surely time pressure. It is certainly the biggest problem during the management. Methods: The author used different tools and methods to achieve his goals; one of them was the study of the relevant literature, the other one was his own experience as a firefighting manager. Other results come from two surveys that are referred to; one of them was an essay analysis, the second one was a word association test, specially created for the research. Results and discussion: This article proves that, in certain situations, the multi-criteria, evaluating decision-making processes simply cannot be used or only in a limited manner. However, it can be seen that managers, directors or commanders are many times in situations that simply cannot be ignored when making decisions which should be made in a short time. The functional background of decisions made in a short time, their mechanism, which is different from the conventional, was studied lately and this special decision procedure was given the name recognition-primed decision. In the article, author illustrates the limits of the possibilities of analytical decision-making, presents the general operating mechanism of recognition-primed decision-making, elaborates on its special model relevant to managers at tactical level, as well as explore and systemize the factors that facilitate (catalyze) the processes with an example with fire managers.Keywords: decision making, disaster managers, recognition primed decision, model for making decisions in emergencies
Procedia PDF Downloads 2599804 Regularized Euler Equations for Incompressible Two-Phase Flow Simulations
Authors: Teng Li, Kamran Mohseni
Abstract:
This paper presents an inviscid regularization technique for the incompressible two-phase flow simulations. This technique is known as observable method due to the understanding of observability that any feature smaller than the actual resolution (physical or numerical), i.e., the size of wire in hotwire anemometry or the grid size in numerical simulations, is not able to be captured or observed. Differ from most regularization techniques that applies on the numerical discretization, the observable method is employed at PDE level during the derivation of equations. Difficulties in the simulation and analysis of realistic fluid flow often result from discontinuities (or near-discontinuities) in the calculated fluid properties or state. Accurately capturing these discontinuities is especially crucial when simulating flows involving shocks, turbulence or sharp interfaces. Over the past several years, the properties of this new regularization technique have been investigated that show the capability of simultaneously regularizing shocks and turbulence. The observable method has been performed on the direct numerical simulations of shocks and turbulence where the discontinuities are successfully regularized and flow features are well captured. In the current paper, the observable method will be extended to two-phase interfacial flows. Multiphase flows share the similar features with shocks and turbulence that is the nonlinear irregularity caused by the nonlinear terms in the governing equations, namely, Euler equations. In the direct numerical simulation of two-phase flows, the interfaces are usually treated as the smooth transition of the properties from one fluid phase to the other. However, in high Reynolds number or low viscosity flows, the nonlinear terms will generate smaller scales which will sharpen the interface, causing discontinuities. Many numerical methods for two-phase flows fail at high Reynolds number case while some others depend on the numerical diffusion from spatial discretization. The observable method regularizes this nonlinear mechanism by filtering the convective terms and this process is inviscid. The filtering effect is controlled by an observable scale which is usually about a grid length. Single rising bubble and Rayleigh-Taylor instability are studied, in particular, to examine the performance of the observable method. A pseudo-spectral method is used for spatial discretization which will not introduce numerical diffusion, and a Total Variation Diminishing (TVD) Runge Kutta method is applied for time integration. The observable incompressible Euler equations are solved for these two problems. In rising bubble problem, the terminal velocity and shape of the bubble are particularly examined and compared with experiments and other numerical results. In the Rayleigh-Taylor instability, the shape of the interface are studied for different observable scale and the spike and bubble velocities, as well as positions (under a proper observable scale), are compared with other simulation results. The results indicate that this regularization technique can potentially regularize the sharp interface in the two-phase flow simulationsKeywords: Euler equations, incompressible flow simulation, inviscid regularization technique, two-phase flow
Procedia PDF Downloads 5029803 Performance Modeling and Availability Analysis of Yarn Dyeing System of a Textile Industry
Authors: P. C. Tewari, Rajiv Kumar, Dinesh Khanduja
Abstract:
This paper discusses the performance modeling and availability analysis of Yarn Dyeing System of a Textile Industry. The Textile Industry is a complex and repairable engineering system. Yarn Dyeing System of Textile Industry consists of five subsystems arranged in series configuration. For performance modeling and analysis of availability, a performance evaluating model has been developed with the help of mathematical formulation based on Markov-Birth-Death Process. The differential equations have been developed on the basis of Probabilistic Approach using a Transition Diagram. These equations have further been solved using normalizing condition in order to develop the steady state availability, a performance measure of the system concerned. The system performance has been further analyzed with the help of decision matrices. These matrices provide various availability levels for different combinations of failure and repair rates for various subsystems. The findings of this paper are, therefore, considered to be useful for the analysis of availability and determination of the best possible maintenance strategies which can be implemented in future to enhance the system performance.Keywords: performance modeling, markov process, steady state availability, availability analysis
Procedia PDF Downloads 3359802 Quantum Statistical Machine Learning and Quantum Time Series
Authors: Omar Alzeley, Sergey Utev
Abstract:
Minimizing a constrained multivariate function is the fundamental of Machine learning, and these algorithms are at the core of data mining and data visualization techniques. The decision function that maps input points to output points is based on the result of optimization. This optimization is the central of learning theory. One approach to complex systems where the dynamics of the system is inferred by a statistical analysis of the fluctuations in time of some associated observable is time series analysis. The purpose of this paper is a mathematical transition from the autoregressive model of classical time series to the matrix formalization of quantum theory. Firstly, we have proposed a quantum time series model (QTS). Although Hamiltonian technique becomes an established tool to detect a deterministic chaos, other approaches emerge. The quantum probabilistic technique is used to motivate the construction of our QTS model. The QTS model resembles the quantum dynamic model which was applied to financial data. Secondly, various statistical methods, including machine learning algorithms such as the Kalman filter algorithm, are applied to estimate and analyses the unknown parameters of the model. Finally, simulation techniques such as Markov chain Monte Carlo have been used to support our investigations. The proposed model has been examined by using real and simulated data. We establish the relation between quantum statistical machine and quantum time series via random matrix theory. It is interesting to note that the primary focus of the application of QTS in the field of quantum chaos was to find a model that explain chaotic behaviour. Maybe this model will reveal another insight into quantum chaos.Keywords: machine learning, simulation techniques, quantum probability, tensor product, time series
Procedia PDF Downloads 4699801 Fuzzy Analytic Hierarchy Process for Determination of Supply Chain Performance Evaluation Criteria
Authors: Ibrahim Cil, Onur Kurtcu, H. Ibrahim Demir, Furkan Yener, Yusuf. S. Turkan, Muharrem Unver, Ramazan Evren
Abstract:
Fuzzy AHP (Analytic Hierarchy Process) method is decision-making way at the end of integrating the current AHP method with fuzzy structure. In this study, the processes of production planning, inventory management and purchasing department of a system were analysed and were requested to decide the performance criteria of each area. At this point, the current work processes were analysed by various decision-makers and comparing each criteria by giving points according to 1-9 scale were completed. The criteria were listed in order to their weights by using Fuzzy AHP approach and top three performance criteria of each department were determined. After that, the performance criteria of supply chain consisting of three departments were asked to determine. The processes of each department were compared by decision-makers at the point of building the supply chain performance system and getting the performance criteria. According to the results, the criteria of performance system of supply chain by using Fuzzy AHP were determined for which will be used in the supply chain performance system in the future.Keywords: AHP, fuzzy, performance evaluation, supply chain
Procedia PDF Downloads 3469800 Considering Partially Developed Artifacts in Change Impact Analysis Implementation
Authors: Nazri Kama, Sufyan Basri, Roslina Ibrahim
Abstract:
It is important to manage the changes in the software to meet the evolving needs of the customer. Accepting too many changes causes delay in the completion and it incurs additional cost. One type of information that helps to make the decision is through change impact analysis. Current impact analysis approaches assume that all classes in the class artifact are completely developed and the class artifact is used as a source of analysis. However, these assumptions are impractical for impact analysis in the software development phase as some classes in the class artifact are still under development or partially developed that leads to inaccuracy. This paper presents a novel impact analysis approach to be used in the software development phase. The significant achievements of the approach are demonstrated through an extensive experimental validation using three case studies.Keywords: software development, impact analysis, traceability, static analysis.
Procedia PDF Downloads 6089799 Hidden Markov Movement Modelling with Irregular Data
Authors: Victoria Goodall, Paul Fatti, Norman Owen-Smith
Abstract:
Hidden Markov Models have become popular for the analysis of animal tracking data. These models are being used to model the movements of a variety of species in many areas around the world. A common assumption of the model is that the observations need to have regular time steps. In many ecological studies, this will not be the case. The objective of the research is to modify the movement model to allow for irregularly spaced locations and investigate the effect on the inferences which can be made about the latent states. A modification of the likelihood function to allow for these irregular spaced locations is investigated, without using interpolation or averaging the movement rate. The suitability of the modification is investigated using GPS tracking data for lion (Panthera leo) in South Africa, with many observations obtained during the night, and few observations during the day. Many nocturnal predator tracking studies are set up in this way, to obtain many locations at night when the animal is most active and is difficult to observe. Few observations are obtained during the day, when the animal is expected to rest and is potentially easier to observe. Modifying the likelihood function allows the popular Hidden Markov Model framework to be used to model these irregular spaced locations, making use of all the observed data.Keywords: hidden Markov Models, irregular observations, animal movement modelling, nocturnal predator
Procedia PDF Downloads 2449798 Discrete State Prediction Algorithm Design with Self Performance Enhancement Capacity
Authors: Smail Tigani, Mohamed Ouzzif
Abstract:
This work presents a discrete quantitative state prediction algorithm with intelligent behavior making it able to self-improve some performance aspects. The specificity of this algorithm is the capacity of self-rectification of the prediction strategy before the final decision. The auto-rectification mechanism is based on two parallel mathematical models. In one hand, the algorithm predicts the next state based on event transition matrix updated after each observation. In the other hand, the algorithm extracts its residues trend with a linear regression representing historical residues data-points in order to rectify the first decision if needs. For a normal distribution, the interactivity between the two models allows the algorithm to self-optimize its performance and then make better prediction. Designed key performance indicator, computed during a Monte Carlo simulation, shows the advantages of the proposed approach compared with traditional one.Keywords: discrete state, Markov Chains, linear regression, auto-adaptive systems, decision making, Monte Carlo Simulation
Procedia PDF Downloads 4989797 Setting Ground for Improvement of Knowledge Managament System in the Educational Organization
Authors: Mladen Djuric, Ivan Janicijevic, Sasa Lazarevic
Abstract:
One of the organizational issues is how to develop and shape decision making and knowledge management systems which will continually avoid traps of both paralyses by analyses“ and extinction by instinct“, the concepts that are a kind of tolerant limits anti-patterns which define what we can call decision making and knowledge management patterns control zone. This paper discusses potentials for development of a core base for recognizing, capturing, and analyzing anti-patterns in the educational organization, thus creating a space for improving decision making and knowledge management processes in education.Keywords: anti-patterns, decision making, education, knowledge management
Procedia PDF Downloads 6329796 A Prioritisation Guide for More Sustainable Manufacturing Processes
Authors: Cansu Kandemir, Marco Franchino
Abstract:
To attain sustainability goals, the manufacturing industries must assess and improve their processes, adopt the latest technologies, and ensure minimal environmental impact. Ongoing debates claim that the definition of sustainability and its assessment is vague. Companies struggle with understanding which processes they should prioritise and necessitate a methodology to aid decision-making. For that reason, our investigation focused on defining a prioritisation guide to help to manufacture engineers identify areas of a facility to prioritise de-carbonisation efforts based on existing sources of data. The authors at the University of Sheffield Advanced Manufacturing Research Centre (AMRC) worked with a range of major businesses, including Food and Drink (Moy Park), Automotive (Nissan), Aerospace and Defence (BAE, Meggitt, Leonardo, and GKN) and Technology (Accenture and Intellium AI). Collected information has been integrated into a prioritisation guide framework that helps process comparison and decision-making. The framework developed in this study aims to ensure that companies have guidance on where to focus their efforts whilst striving to fulfil their environmental and societal obligations.Keywords: decision making, sustainability, carbon emissions, manufacturing
Procedia PDF Downloads 619795 Speed Breaker/Pothole Detection Using Hidden Markov Models: A Deep Learning Approach
Authors: Surajit Chakrabarty, Piyush Chauhan, Subhasis Panda, Sujoy Bhattacharya
Abstract:
A large proportion of roads in India are not well maintained as per the laid down public safety guidelines leading to loss of direction control and fatal accidents. We propose a technique to detect speed breakers and potholes using mobile sensor data captured from multiple vehicles and provide a profile of the road. This would, in turn, help in monitoring roads and revolutionize digital maps. Incorporating randomness in the model formulation for detection of speed breakers and potholes is crucial due to substantial heterogeneity observed in data obtained using a mobile application from multiple vehicles driven by different drivers. This is accomplished with Hidden Markov Models, whose hidden state sequence is found for each time step given the observables sequence, and are then fed as input to LSTM network with peephole connections. A precision score of 0.96 and 0.63 is obtained for classifying bumps and potholes, respectively, a significant improvement from the machine learning based models. Further visualization of bumps/potholes is done by converting time series to images using Markov Transition Fields where a significant demarcation among bump/potholes is observed.Keywords: deep learning, hidden Markov model, pothole, speed breaker
Procedia PDF Downloads 1449794 Implicit and Explicit Mechanisms of Emotional Contagion
Authors: Andres Pinilla Palacios, Ricardo Tamayo
Abstract:
Emotional contagion is characterized as an automatic tendency to synchronize behaviors that facilitate emotional convergence among humans. It might thus play a pivotal role to understand the dynamics of key social interactions. However, a few research has investigated its potential mechanisms. We suggest two complementary but independent processes that may underlie emotional contagion. The efficient contagion hypothesis, based on fast and implicit bottom-up processes, modulated by familiarity and spread of activation in the emotional associative networks of memory. Secondly, the emotional contrast hypothesis, based on slow and explicit top-down processes guided by deliberated appraisal and hypothesis-testing. In order to assess these two hypotheses, an experiment with 39 participants was conducted. In the first phase, participants were induced (between-groups) to an emotional state (positive, neutral or negative) using a standardized video taken from the FilmStim database. In the second phase, participants classified and rated (within-subject) the emotional state of 15 faces (5 for each emotional state) taken from the POFA database. In the third phase, all participants were returned to a baseline emotional state using the same neutral video used in the first phase. In a fourth phase, participants classified and rated a new set of 15 faces. The accuracy in the identification and rating of emotions was partially explained by the efficient contagion hypothesis, but the speed with which these judgments were made was partially explained by the emotional contrast hypothesis. However, results are ambiguous, so a follow-up experiment is proposed in which emotional expressions and activation of the sympathetic system will be measured using EMG and EDA respectively.Keywords: electromyography, emotional contagion, emotional valence, identification of emotions, imitation
Procedia PDF Downloads 3169793 Innovation in Information Technology Services: Framework to Improve the Effectiveness and Efficiency of Information Technology Service Management Processes, Projects and Decision Support Management
Authors: Pablo Cardozo Herrera
Abstract:
In a dynamic market of Information Technology (IT) Service and with high quality demands and high performance requirements in decreasing costs, it is imperative that IT companies invest organizational effort in order to increase the effectiveness of their Information Technology Service Management (ITSM) processes through the improvement of ITSM project management and through solid support to the strategic decision-making process of IT directors. In this article, the author presents an analysis of common issues of IT companies around the world, with strategic needs of information unmet that provoke their ITSM processes and projects management that do not achieve the effectiveness and efficiency expected of their results. In response to the issues raised, the author proposes a framework consisting of an innovative theoretical framework model of ITSM management and a technological solution aligned to the Information Technology Infrastructure Library (ITIL) good practices guidance and ISO/IEC 20000-1 requirements. The article describes a research that proves the proposed framework is able to integrate, manage and coordinate in a holistic way, measurable and auditable, all ITSM processes and projects of IT organization and utilize the effectiveness assessment achieved for their strategic decision-making process increasing the process maturity level and improving the capacity of an efficient management.Keywords: innovation in IT services, ITSM processes, ITIL and ISO/IEC 20000-1, IT service management, IT service excellence
Procedia PDF Downloads 3979792 Copula Markov Switching Multifractal Models for Forecasting Value-at-Risk
Authors: Giriraj Achari, Malay Bhattacharyya
Abstract:
In this paper, the effectiveness of Copula Markov Switching Multifractal (MSM) models at forecasting Value-at-Risk of a two-stock portfolio is studied. The innovations are allowed to be drawn from distributions that can capture skewness and leptokurtosis, which are well documented empirical characteristics observed in financial returns. The candidate distributions considered for this purpose are Johnson-SU, Pearson Type-IV and α-Stable distributions. The two univariate marginal distributions are combined using the Student-t copula. The estimation of all parameters is performed by Maximum Likelihood Estimation. Finally, the models are compared in terms of accurate Value-at-Risk (VaR) forecasts using tests of unconditional coverage and independence. It is found that Copula-MSM-models with leptokurtic innovation distributions perform slightly better than Copula-MSM model with Normal innovations. Copula-MSM models, in general, produce better VaR forecasts as compared to traditional methods like Historical Simulation method, Variance-Covariance approach and Copula-Generalized Autoregressive Conditional Heteroscedasticity (Copula-GARCH) models.Keywords: Copula, Markov Switching, multifractal, value-at-risk
Procedia PDF Downloads 1659791 A Straightforward Approach for Determining the Weights of Decision Makers Based on Angle Cosine and Projection Method
Authors: Qiang Yang, Ping-An Du
Abstract:
Group decision making with multiple attribute has attracted intensive concern in the decision analysis area. This paper assumes that the contributions of all the decision makers (DMs) are not equal to the decision process based on different knowledge and experience in group setting. The aim of this paper is to develop a novel approach to determine weights of DMs in the group decision making problems. In this paper, the weights of DMs are determined in the group decision environment via angle cosine and projection method. First of all, the average decision of all individual decisions is defined as the ideal decision. After that, we define the weight of each decision maker (DM) by aggregating the angle cosine and projection between individual decision and ideal decision with associated direction indicator μ. By using the weights of DMs, all individual decisions are aggregated into a collective decision. Further, the preference order of alternatives is ranked in accordance with the overall row value of collective decision. Finally, an example in a chemical company is provided to illustrate the developed approach.Keywords: angel cosine, ideal decision, projection method, weights of decision makers
Procedia PDF Downloads 3789790 Complex Decision Rules in the Form of Decision Trees
Authors: Avinash S. Jagtap, Sharad D. Gore, Rajendra G. Gurao
Abstract:
Decision rules become more and more complex as the number of conditions increase. As a consequence, the complexity of the decision rule also influences the time complexity of computer implementation of such a rule. Consider, for example, a decision that depends on four conditions A, B, C and D. For simplicity, suppose each of these four conditions is binary. Even then the decision rule will consist of 16 lines, where each line will be of the form: If A and B and C and D, then action 1. If A and B and C but not D, then action 2 and so on. While executing this decision rule, each of the four conditions will be checked every time until all the four conditions in a line are satisfied. The minimum number of logical comparisons is 4 whereas the maximum number is 64. This paper proposes to present a complex decision rule in the form of a decision tree. A decision tree divides the cases into branches every time a condition is checked. In the form of a decision tree, every branching eliminates half of the cases that do not satisfy the related conditions. As a result, every branch of the decision tree involves only four logical comparisons and hence is significantly simpler than the corresponding complex decision rule. The conclusion of this paper is that every complex decision rule can be represented as a decision tree and the decision tree is mathematically equivalent but computationally much simpler than the original complex decision ruleKeywords: strategic, tactical, operational, adaptive, innovative
Procedia PDF Downloads 2879789 Selecting the Best Software Product Using Analytic Hierarchy Process and Fuzzy-Analytic Hierarchy Process Modules
Authors: Anas Hourani, Batool Ahmad
Abstract:
Software applications play an important role inside any institute. They are employed to manage all processes and store entities-related data in the computer. Therefore, choosing the right software product that meets institute requirements is not an easy decision in view of considering multiple criteria, different points of views, and many standards. As a case study, Mutah University, located in Jordan, is in essential need of customized software, and several companies presented their software products which are very similar in quality. In this regard, an analytic hierarchy process (AHP) and a fuzzy analytic hierarchy process (Fuzzy-AHP) models are proposed in this research to identify the most suitable and best-fit software product that meets the institute requirements. The results indicate that both modules are able to help the decision-makers to make a decision, especially in complex decision problems.Keywords: analytic hierarchy process, decision modeling, fuzzy analytic hierarchy process, software product
Procedia PDF Downloads 3929788 Quantifying Spatiotemporal Patterns of Past and Future Urbanization Trends in El Paso, Texas and Their Impact on Electricity Consumption
Authors: Joanne Moyer
Abstract:
El Paso, Texas is a southwest border city that has experienced continuous growth within the last 15-years. Understanding the urban growth trends and patterns using data from the National Land Cover Database (NLCD) and landscape metrics, provides a quantitative description of growth. Past urban growth provided a basis to predict 2031 future land-use for El Paso using the CA-Markov model. As a consequence of growth, an increase in demand of resources follows. Using panel data analysis, an understanding of the relation between landscape metrics and electricity consumption is further analyzed. The studies’ findings indicate that past growth focused within three districts within the City of El Paso. The landscape metrics suggest as the city has grown, fragmentation has decreased. Alternatively, the landscape metrics for the projected 2031 land-use indicates possible fragmentation within one of these districts. Panel data suggests electricity consumption and mean patch area landscape metric are positively correlated. The study provides local decision makers to make informed decisions for policies and urban planning to ensure a future sustainable community.Keywords: landscape metrics, CA-Markov, El Paso, Texas, panel data
Procedia PDF Downloads 1439787 Numerical Analysis of Multiplicity and Transition Phenomena in Natural Convection
Authors: Hadi Kafil, Ali Ecder
Abstract:
Heat transfer by natural convection in two-dimensional and three-dimensional axisymmetric enclosure fitted with partially heated vertical walls is investigated numerically. The range of Rayleigh number is varied from 10³ until convective flow becomes unstable. This research focuses on multiplicity and transition phenomena in natural convection and is based on a parametric analysis to study the onset of bifurcations. It is found that, even at low Rayleigh numbers, the flow undergoes a series of turning-point bifurcations which increase the rate of natural convention. On the other hand, by partially heating or cooling the walls, more effective results can be achieved for both heating and cooling applications, such as cooling of electronic devices and heating processes in solidification and crystal growth.Keywords: natural convection, partial heated, onset of bifurcation, Rayleigh number
Procedia PDF Downloads 3709786 Mean Field Model Interaction for Computer and Communication Systems: Modeling and Analysis of Wireless Sensor Networks
Authors: Irina A. Gudkova, Yousra Demigha
Abstract:
Scientific research is moving more and more towards the study of complex systems in several areas of economics, biology physics, and computer science. In this paper, we will work on complex systems in communication networks, Wireless Sensor Networks (WSN) that are considered as stochastic systems composed of interacting entities. The current advancements of the sensing in computing and communication systems is an investment ground for research in several tracks. A detailed presentation was made for the WSN, their use, modeling, different problems that can occur in their application and some solutions. The main goal of this work reintroduces the idea of mean field method since it is a powerful technique to solve this type of models especially systems that evolve according to a Continuous Time Markov Chain (CTMC). Modeling of a CTMC has been focused; we obtained a large system of interacting Continuous Time Markov Chain with population entities. The main idea was to work on one entity and replace the others with an average or effective interaction. In this context to make the solution easier, we consider a wireless sensor network as a multi-body problem and we reduce it to one body problem. The method was applied to a system of WSN modeled as a Markovian queue showing the results of the used technique.Keywords: Continuous-Time Markov Chain, Hidden Markov Chain, mean field method, Wireless sensor networks
Procedia PDF Downloads 1659785 Framework for the Modeling of the Supply Chain Collaborative Planning Process
Authors: D. Pérez, M. M. E. Alemany
Abstract:
In this work a Framework to model the Supply Chain (SC) Collaborative Planning (CP) Process is proposed, and particularly its Decisional view. The main Framework contributions with regards to previous related works are the following, 1) the consideration of not only the Decision view, the most important one due to the Process type, but other additional three views which are the Physical, Organisation and Information ones, closely related and complementing the Decision View, 2) the joint consideration of two interdependence types, the Temporal (among Decision Centres belonging to different Decision Levels) and Spatial (among Decision Centres belonging to the same Decision Level) to support the distributed Decision-Making process in SC where several decision Centres interact among them in a collaborative manner.Keywords: collaborative planning, decision view, distributed decision-making, framework
Procedia PDF Downloads 4679784 Data-Mining Approach to Analyzing Industrial Process Information for Real-Time Monitoring
Authors: Seung-Lock Seo
Abstract:
This work presents a data-mining empirical monitoring scheme for industrial processes with partially unbalanced data. Measurement data of good operations are relatively easy to gather, but in unusual special events or faults it is generally difficult to collect process information or almost impossible to analyze some noisy data of industrial processes. At this time some noise filtering techniques can be used to enhance process monitoring performance in a real-time basis. In addition, pre-processing of raw process data is helpful to eliminate unwanted variation of industrial process data. In this work, the performance of various monitoring schemes was tested and demonstrated for discrete batch process data. It showed that the monitoring performance was improved significantly in terms of monitoring success rate of given process faults.Keywords: data mining, process data, monitoring, safety, industrial processes
Procedia PDF Downloads 4019783 Identification and Selection of a Supply Chain Target Process for Re-Design
Authors: Jaime A. Palma-Mendoza
Abstract:
A supply chain consists of different processes and when conducting supply chain re-design is necessary to identify the relevant processes and select a target for re-design. A solution was developed which consists to identify first the relevant processes using the Supply Chain Operations Reference (SCOR) model, then to use Analytical Hierarchy Process (AHP) for target process selection. An application was conducted in an Airline MRO supply chain re-design project which shows this combination can clearly aid the identification of relevant supply chain processes and the selection of a target process for re-design.Keywords: decision support systems, multiple criteria analysis, supply chain management
Procedia PDF Downloads 4929782 Vulnerability Assessment of Healthcare Interdependent Critical Infrastructure Coloured Petri Net Model
Authors: N. Nivedita, S. Durbha
Abstract:
Critical Infrastructure (CI) consists of services and technological networks such as healthcare, transport, water supply, electricity supply, information technology etc. These systems are necessary for the well-being and to maintain effective functioning of society. Critical Infrastructures can be represented as nodes in a network where they are connected through a set of links depicting the logical relationship among them; these nodes are interdependent on each other and interact with each at other at various levels, such that the state of each infrastructure influences or is correlated to the state of another. Disruption in the service of one infrastructure nodes of the network during a disaster would lead to cascading and escalating disruptions across other infrastructures nodes in the network. The operation of Healthcare Infrastructure is one such Critical Infrastructure that depends upon a complex interdependent network of other Critical Infrastructure, and during disasters it is very vital for the Healthcare Infrastructure to be protected, accessible and prepared for a mass casualty. To reduce the consequences of a disaster on the Critical Infrastructure and to ensure a resilient Critical Health Infrastructure network, knowledge, understanding, modeling, and analyzing the inter-dependencies between the infrastructures is required. The paper would present inter-dependencies related to Healthcare Critical Infrastructure based on Hierarchical Coloured Petri Nets modeling approach, given a flood scenario as the disaster which would disrupt the infrastructure nodes. The model properties are being analyzed for the various state changes which occur when there is a disruption or damage to any of the Critical Infrastructure. The failure probabilities for the failure risk of interconnected systems are calculated by deriving a reachability graph, which is later mapped to a Markov chain. By analytically solving and analyzing the Markov chain, the overall vulnerability of the Healthcare CI HCPN model is demonstrated. The entire model would be integrated with Geographic information-based decision support system to visualize the dynamic behavior of the interdependency of the Healthcare and related CI network in a geographically based environment.Keywords: critical infrastructure interdependency, hierarchical coloured petrinet, healthcare critical infrastructure, Petri Nets, Markov chain
Procedia PDF Downloads 5299781 Reliability Analysis of a Fuel Supply System in Automobile Engine
Authors: Chitaranjan Sharma
Abstract:
The present paper deals with the analysis of a fuel supply system in an automobile engine of a four wheeler which is having both the option of fuel i.e. PETROL and CNG. Since CNG is cheaper than petrol so the priority is given to consume CNG as compared to petrol. An automatic switch is used to start petrol supply at the time of failure of CNG supply. Using regenerative point technique with Markov renewal process, the reliability characteristics which are useful to system designers are obtained.Keywords: reliability, redundancy, repair time, transition, probability, regenerative points, markov renewal, process
Procedia PDF Downloads 5519780 The Impact of Social Media on Urban E-planning: A Review of the Literature
Authors: Farnoosh Faal
Abstract:
The rapid growth of social media has brought significant changes to the field of urban e-planning. This study aims to review the existing literature on the impact of social media on urban e-planning processes. The study begins with a discussion of the evolution of social media and its role in urban e-planning. The review covers research on the use of social media for public engagement, citizen participation, stakeholder communication, decision-making, and monitoring and evaluation of urban e-planning initiatives. The findings suggest that social media has the potential to enhance public participation and improve decision-making in urban e-planning processes. Social media platforms such as Facebook, Twitter, and Instagram can provide a platform for citizens to engage with planners and policymakers, express their opinions, and provide feedback on planning proposals. Social media can also facilitate the collection and analysis of data, including real-time data, to inform urban e-planning decision-making. However, the literature also highlights some challenges associated with the use of social media in urban e-planning. These challenges include issues related to the representativeness of social media users, the quality of information obtained from social media, the potential for bias and manipulation of social media content, and the need for effective data management and analysis. The study concludes with recommendations for future research on the use of social media in urban e-planning. The recommendations include the need for further research on the impact of social media on equity and social justice in planning processes, the need for more research on effective strategies for engaging underrepresented groups, and the development of guidelines for the use of social media in urban e-planning processes. Overall, the study suggests that social media has the potential to transform urban e-planning processes but that careful consideration of the opportunities and challenges associated with its use is essential for effective and ethical planning practice.Keywords: social media, Urban e-planning, public participation, citizen engagement
Procedia PDF Downloads 2359779 Monetary Policy and Economic Growth in West African Business Cycles: Markov Switching Approach
Authors: Omolade Adeleke, Jonathan Olusegun Famoroti
Abstract:
This study empirically examined the monetary policy and economic growth in the classical cycles in 8 member countries of the West African Economic and Monetary Union (WAEMU), using the Markov switching model for the Two-phase Regime, covering the period 1980Q1 to 2020Q4. Our estimates suggest that these countries demonstrate to have similar business cycles, and the economies stay more in an expansion regime than a recession regime. The result further shows that the union has an average duration period of 3.1 and 15.9 quarters for contraction and expansion periods, respectively. The business cycle duration, on average, suggests 19 quarters, varying from country to country. Therefore, the formulation of policies that can enhance aggregate demand by member countries in the union is an antidote for recession and is necessary to drive the economy into equilibrium. Also, a low-interest rate and reduced inflation rate would ginger long-run economic growth.Keywords: monetary policy, business cycle, economic growth, Markov switching
Procedia PDF Downloads 759778 Deep Reinforcement Learning Approach for Trading Automation in The Stock Market
Authors: Taylan Kabbani, Ekrem Duman
Abstract:
The design of adaptive systems that take advantage of financial markets while reducing the risk can bring more stagnant wealth into the global market. However, most efforts made to generate successful deals in trading financial assets rely on Supervised Learning (SL), which suffered from various limitations. Deep Reinforcement Learning (DRL) offers to solve these drawbacks of SL approaches by combining the financial assets price "prediction" step and the "allocation" step of the portfolio in one unified process to produce fully autonomous systems capable of interacting with its environment to make optimal decisions through trial and error. In this paper, a continuous action space approach is adopted to give the trading agent the ability to gradually adjust the portfolio's positions with each time step (dynamically re-allocate investments), resulting in better agent-environment interaction and faster convergence of the learning process. In addition, the approach supports the managing of a portfolio with several assets instead of a single one. This work represents a novel DRL model to generate profitable trades in the stock market, effectively overcoming the limitations of supervised learning approaches. We formulate the trading problem, or what is referred to as The Agent Environment as Partially observed Markov Decision Process (POMDP) model, considering the constraints imposed by the stock market, such as liquidity and transaction costs. More specifically, we design an environment that simulates the real-world trading process by augmenting the state representation with ten different technical indicators and sentiment analysis of news articles for each stock. We then solve the formulated POMDP problem using the Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithm, which can learn policies in high-dimensional and continuous action spaces like those typically found in the stock market environment. From the point of view of stock market forecasting and the intelligent decision-making mechanism, this paper demonstrates the superiority of deep reinforcement learning in financial markets over other types of machine learning such as supervised learning and proves its credibility and advantages of strategic decision-making.Keywords: the stock market, deep reinforcement learning, MDP, twin delayed deep deterministic policy gradient, sentiment analysis, technical indicators, autonomous agent
Procedia PDF Downloads 178