Search results for: decisions under uncertainty
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 774

Search results for: decisions under uncertainty

174 A New Framework and a Model for Product Development with an Application in the Telecommunications Services Sector

Authors: Ghada A. El Khayat

Abstract:

This paper argues that a product development exercise involves in addition to the conventional stages, several decisions regarding other aspects. These aspects should be addressed simultaneously in order to develop a product that responds to the customer needs and that helps realize objectives of the stakeholders in terms of profitability, market share and the like. We present a framework that encompasses these different development dimensions. The framework shows that a product development methodology such as the Quality Function Deployment (QFD) is the basic tool which allows definition of the target specifications of a new product. Creativity is the first dimension that enables the development exercise to live and end successfully. A number of group processes need to be followed by the development team in order to ensure enough creativity and innovation. Secondly, packaging is considered to be an important extension of the product. Branding strategies, quality and standardization requirements, identification technologies, design technologies, production technologies and costing and pricing are also integral parts to the development exercise. These dimensions constitute the proposed framework. The paper also presents a mathematical model used to calculate the design targets based on the target costing principle. The framework is used to study a case of a new product development in the telecommunications services sector.

Keywords: Product Development Framework, Quality FunctionDeployment, Mathematical Models, Telecommunications.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1559
173 Automated Vehicle Traffic Control Tower: A Solution to Support the Next Level Automation

Authors: Xiaoyun Zhao, Rami Darwish, Anna Pernestål

Abstract:

Automated vehicles (AVs) have the potential to enhance road capacity, improving road safety and traffic efficiency. Research and development on AVs have been going on for many years. However, when the complicated traffic rules and real situations interacted, AVs fail to make decisions on contradicting situations, and are not able to have control in all conditions due to highly dynamic driving scenarios. This limits AVs’ usage and restricts the full potential benefits that they can bring. Furthermore, regulations, infrastructure development, and public acceptance cannot keep up at the same pace as technology breakthroughs. Facing these challenges, this paper proposes automated vehicle traffic control tower (AVTCT) acting as a safe, efficient and integrated solution for AV control. It introduces a concept of AVTCT for control, management, decision-making, communication and interaction with various aspects in transportation. With the prototype demonstrations and simulations, AVTCT has the potential to overcome the control challenges with AVs and can facilitate AV reaching their full potential. Possible functionalities, benefits as well as challenges of AVTCT are discussed, which set the foundation for the conceptual model, simulation and real application of AVTCT.

Keywords: Automated vehicle, connectivity and automation, intelligent transport system, traffic control, traffic safety.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1073
172 Generic Workload Management System Using Condor-Based Pilot Factory in PanDA Framework

Authors: Po-Hsiang Chiu, Torre Wenaus

Abstract:

In the current Grid environment, efficient workload management presents a significant challenge, for which there are exorbitant de facto standards encompassing resource discovery, brokerage, and data transfer, among others. In addition, the real-time resource status, essential for an optimal resource allocation strategy, is often not readily accessible. To address these issues and provide a cleaner abstraction of the Grid with the potential of generalizing into arbitrary resource-sharing environment, this paper proposes a new Condor-based pilot mechanism applied in the PanDA architecture, PanDA-PF WMS, with the goal of providing a more generic yet efficient resource allocating strategy. In this architecture, the PanDA server primarily acts as a repository of user jobs, responding to pilot requests from distributed, remote resources. Scheduling decisions are subsequently made according to the real-time resource information reported by pilots. Pilot Factory is a Condor-inspired solution for a scalable pilot dissemination and effectively functions as a resource provisioning mechanism through which the user-job server, PanDA, reaches out to the candidate resources only on demand.

Keywords: Condor, glidein, PanDA, Pilot, Pilot Factory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2099
171 Evaluating Portfolio Performance by Highlighting Network Property and the Sharpe Ratio in the Stock Market

Authors: Zahra Hatami, Hesham Ali, David Volkman

Abstract:

Selecting a portfolio for investing is a crucial decision for individuals and legal entities. In the last two decades, with economic globalization, a stream of financial innovations has rushed to the aid of financial institutions. The importance of selecting stocks for the portfolio is always a challenging task for investors. This study aims to create a financial network to identify optimal portfolios using network centralities metrics. This research presents a community detection technique of superior stocks that can be described as an optimal stock portfolio to be used by investors. By using the advantages of a network and its property in extracted communities, a group of stocks was selected for each of the various time periods. The performance of the optimal portfolios was compared to the famous index. Their Sharpe ratio was calculated in a timely manner to evaluate their profit for making decisions. The analysis shows that the selected potential portfolio from stocks with low centrality measurement can outperform the market; however, they have a lower Sharpe ratio than stocks with high centrality scores. In other words, stocks with low centralities could outperform the S&P500 yet have a lower Sharpe ratio than high central stocks.

Keywords: Portfolio management performance, network analysis, centrality measurements, Sharpe ratio.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 408
170 Ways for the Development of the Audit Quality Control System through the Analysis of Ongoing Problems, Experience and Challenges: Example of the Republic of Georgia

Authors: Levan Sabauri

Abstract:

Audit is an independent inspection of the financial statement of the audited person and expresses the opinion of an auditor on the reliability of this statement. The auditor’s activity (auditor’s service) is realized by auditing organizations, individual auditors in connection to conduction of an audit and rendering of audit accompanying services. The profession of auditor means a high level of responsibility for rendered service. Results of decisions made by information users depend on the quality of the auditor’s conclusion. Owners, investors, creditors, and society rely on the opinion of the auditor under the condition that inspection was conducted with good quality. Therefore, the existence of the well-functioning audit quality control system for the administering of the audit is an important issue. An efficient audit quality control system is a substantial challenge that many countries face worldwide, especially those states where these systems are being formed within the respective reform program. The presented article reflects on the best practices of the leading countries, the assumptions and recommendations for the financial accounting, reporting and audit; current reforms in Georgia are made based on this comparative analysis.

Keywords: Audit quality control, audit program, financial statement, perspective analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 696
169 Integrated Approaches to Enhance Aggregate Production Planning with Inventory Uncertainty Based On Improved Harmony Search Algorithm

Authors: P. Luangpaiboon, P. Aungkulanon

Abstract:

This work presents a multiple objective linear programming (MOLP) model based on the desirability function approach for solving the aggregate production planning (APP) decision problem upon Masud and Hwang-s model. The proposed model minimises total production costs, carrying or backordering costs and rates of change in labor levels. An industrial case demonstrates the feasibility of applying the proposed model to the APP problems with three scenarios of inventory levels. The proposed model yields an efficient compromise solution and the overall levels of DM satisfaction with the multiple combined response levels. There has been a trend to solve complex planning problems using various metaheuristics. Therefore, in this paper, the multi-objective APP problem is solved by hybrid metaheuristics of the hunting search (HuSIHSA) and firefly (FAIHSA) mechanisms on the improved harmony search algorithm. Results obtained from the solution of are then compared. It is observed that the FAIHSA can be used as a successful alternative solution mechanism for solving APP problems over three scenarios. Furthermore, the FAIHSA provides a systematic framework for facilitating the decision-making process, enabling a decision maker interactively to modify the desirability function approach and related model parameters until a good optimal solution is obtained with proper selection of control parameters when compared.

Keywords: Aggregate Production Planning, Desirability Function Approach, Improved Harmony Search Algorithm, Hunting Search Algorithm and Firefly Algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1927
168 Improving Spatiotemporal Change Detection: A High Level Fusion Approach for Discovering Uncertain Knowledge from Satellite Image Database

Authors: Wadii Boulila, Imed Riadh Farah, Karim Saheb Ettabaa, Basel Solaiman, Henda Ben Ghezala

Abstract:

This paper investigates the problem of tracking spa¬tiotemporal changes of a satellite image through the use of Knowledge Discovery in Database (KDD). The purpose of this study is to help a given user effectively discover interesting knowledge and then build prediction and decision models. Unfortunately, the KDD process for spatiotemporal data is always marked by several types of imperfections. In our paper, we take these imperfections into consideration in order to provide more accurate decisions. To achieve this objective, different KDD methods are used to discover knowledge in satellite image databases. Each method presents a different point of view of spatiotemporal evolution of a query model (which represents an extracted object from a satellite image). In order to combine these methods, we use the evidence fusion theory which considerably improves the spatiotemporal knowledge discovery process and increases our belief in the spatiotemporal model change. Experimental results of satellite images representing the region of Auckland in New Zealand depict the improvement in the overall change detection as compared to using classical methods.

Keywords: Knowledge discovery in satellite databases, knowledge fusion, data imperfection, data mining, spatiotemporal change detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1547
167 Psychopathic Disorders and Judges Sentencing: Can Neurosciences Change This Aggravating Factor in a Mitigating Factor?

Authors: Kevin Moustapha

Abstract:

Psychopathic disorders are taking an important part in judge sentencing, especially in Canada. First, we will see how this phenomenon can be illustrated by the high proportion of psychopath offenders incarcerated in North American prisons. Many decisions in Canadians courtrooms seem to point out that psychopathy is often used as a strong argument by the judges to preserve public safety. The fact that psychopathy is often associated with violence, recklessness and recidivism, could explain why many judges consider psychopathic disorders as an aggravating factor. Generally, the judge reasoning is based on Article 753 of Canadian Criminal Code related to dangerous offenders, which is used for individuals who show a pattern of repetitive and persistent aggressive behaviour. Then we will show how, with cognitive neurosciences, the psychopath’s situation in courtrooms would probably change. Cerebral imaging and news data provided by the neurosciences show that emotional and volitional functions in psychopath’s brains are impaired. Understanding these new issues could enable some judges to recognize psychopathic disorders as a mitigating factor. Finally, two important questions ought to be raised in this article: can exploring psychopaths ‘brains really change the judge sentencing in Canadian courtrooms? If yes, can judges consider psychopathy more as a mitigating factor than an aggravating factor?

Keywords: Criminal law, judges sentencing, neurosciences, psychopathy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2841
166 Analysis of the Effect of 1980 Transformation on the Foreign Trade of Turkey with Chow Test

Authors: Zeynep Karaçor, Savaş Erdoğan, Perihan Hazel Er

Abstract:

While import-substituting industrialization policy constitute the basis for the industrialization strategies of the 1960s and 1970s in Turkey, this policy was no longer sustainable by the 1980s. For this reason, export-oriented industrialization policy was adopted with the decisions taken on January 24, 1980. In other words, the post-1980 period, Turkey's economy has adopted outwardoriented industrialization strategy. In this study, it is aimed to analyze the effect of the change in economic structure on foreign trade with the transformation of foreign trade and industrialization policies in the post-1980 period. In this respect, in order to analyze the relationship between import, export and economic growth by using variables of the 1960-2011 period, Chow test was applied. In the analysis the reason for using Chow test is whether there is any difference in economic terms between import-substituting industrialization policy applied in the 1960-1980 period and the 1981-2011 period during which exportoriented industrialization policy was applied as a result of the structural transformation.

Keywords: Chow Test, Export-Oriented Industrialization Policy, Import-Substituting Industrialization Policy, Turkey.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1486
165 Sensor Network Based Emergency Response and Navigation Support Architecture

Authors: Dilusha Weeraddana, Ashanie Gunathillake, Samiru Gayan

Abstract:

In an emergency, combining Wireless Sensor Network's data with the knowledge gathered from various other information sources and navigation algorithms, could help safely guide people to a building exit while avoiding the risky areas. This paper presents an emergency response and navigation support architecture for data gathering, knowledge manipulation, and navigational support in an emergency situation. At normal state, the system monitors the environment. When an emergency event detects, the system sends messages to first responders and immediately identifies the risky areas from safe areas to establishing escape paths. The main functionalities of the system include, gathering data from a wireless sensor network which is deployed in a multi-story indoor environment, processing it with information available in a knowledge base, and sharing the decisions made, with first responders and people in the building. The proposed architecture will act to reduce risk of losing human lives by evacuating people much faster with least congestion in an emergency environment. 

Keywords: Emergency response, Firefighters, Navigation, Wireless sensor network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2005
164 Power Flow Tracing Based Reactive Power Ancillary Service (AS) in Restructured Power Market

Authors: M. Susithra, R. Gnanadass

Abstract:

Ancillary services are support services which are essential for humanizing and enhancing the reliability and security of the electric power system. Reactive power ancillary service is one of the important ancillary services in a restructured electricity market which determines the cost of supplying ancillary services and finding of how this cost would change with respect to operating decisions. This paper presents a new formation that can be used to minimize the Independent System Operator (ISO)’s total payment for reactive power ancillary service. The modified power flow tracing algorithm estimates the availability of reserve reactive power for ancillary service. In order to find optimum reactive power dispatch, Biogeography based optimization method (BPO) is proposed. Market Reactive Clearing Price (MRCP) is then estimated and it encourages generator companies (GENCOs) to participate in an ancillary service. Finally, optimal weighting factor and real time utilization factor of reactive power give the minimum ISO’s total payment. The effectiveness of proposed design is verified using IEEE 30 bus system.

Keywords: Biogeography based optimization method, Power flow tracing method, Reactive generation capability curve and Reactive power ancillary service.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3235
163 A Framework to Assess the Maturity of Customer Involvement in the Service Design of Product-Service Systems

Authors: Taghreed Abu-Salim

Abstract:

This paper develops and investigates a framework for the assessment of customer involvement in the service design process of result oriented product-service systems in order to improve the service offering in a business-to-business (B2B) context. The framework comprises five main criteria and fifteen sub-criteria that contribute to customer involvement in a hierarchy using a maturity grid to highlight the strengths and weaknesses for each criterion. To develop the customer involvement framework, an extensive literature review related to service design, result oriented product-service system (PSS) and customer involvement in service design was carried out. Key factors that significantly influence customer involvement from industry and literature were identified to develop the framework. A major contribution of the developed framework includes a hierarchy of appropriate criteria for assessing customer involvement in the service design process within results oriented PSS; the definition of four maturity levels which are suitable to describe the whole spectrum of customer involvement in the service design process; and finally, The paper concludes by enabling service providers to: take proactive decisions; screen and evaluate new services; improve perceived service quality; and provide barriers against imitation.

Keywords: Customer involvement, maturity grid, new service development, result oriented product-service system, service design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2159
162 An AHP-Delphi Multi-Criteria Usage Cases Model with Application to Citrogypsum Decisions, Case Study: Kimia Gharb Gostar Industries Company

Authors: Mohsen Pirdashti, Masoomeh Omidi, Hemmatollah Pidashti

Abstract:

Today, advantage of biotechnology especially in environmental issues compared to other technologies is irrefragable. Kimia Gharb Gostar Industries Company, as a largest producer of citric acid in Middle East, applies biotechnology for this goal. Citrogypsum is a by–product of citric acid production and it considered as a valid residuum of this company. At this paper summary of acid citric production and condition of Citrogypsum production in company were introduced in addition to defmition of Citrogypsum production and its applications in world. According to these information and evaluation of present conditions about Iran needing to Citrogypsum, the best priority was introduced and emphasized on strategy selection and proper programming for self-sufficiency. The Delphi technique was used to elicit expert opinions about criteria for evaluating the usages. The criteria identified by the experts were profitability, capacity of production, the degree of investment, marketable, production ease and time production. The Analytical Hierarchy Process (ARP) and Expert Choice software were used to compare the alternatives on the criteria derived from the Delphi process.

Keywords: Analytical Hierarchy Process, ARP, Delphi, Multi- criteria decision making, Citrogypsum

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2315
161 Characteristics of Football Spectators Using Second Screen

Authors: Florian Pfeffel, Christoph A. Kexel, Peter Kexel, Maria Ratz

Abstract:

The parallel usage of different media channels has increased recently owing to technological advances. Second Screen describes the use of a second device by television viewers to consume further content which is related to the program they are watching. This study analysed the characteristics of football spectators regarding their media consumption in relation to Second Screen usage while watching a football match on TV. The existing literature on Second Screen usage is still very limited, especially in the context of particular broadcasting settings such as sport or even more specific such as football matches. Therefore, the primary research objective was to reveal first insights into the user behaviour of football spectators regarding Second Screen services. The survey, which was conducted among German football supporters in 2015, revealed some characteristics such as the identification and involvement into the sports which are related to an increased use of Second Screen services. One important finding for football supporters was that at the time of a match they have a lower parallel media usage compared to other TV broadcastings. Nevertheless, if supporters used a second device while watching a match on TV, then they were using specific Second Screen services. This means they searched for more content related information. The findings on the habits and characteristics of people who are using Second Screen services are relevant for future developments in that area as well as for marketing decisions.

Keywords: Media consumption, second screen, sport marketing, user behaviour.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1786
160 The Role of Local Government Authorities in Managing the Pre-Hospital Emergency Medical Service (EMS) Systems in Thailand

Authors: Chanisada Choosuk, Napisporn Memongkol Runchana Sinthavalai, Fareeda Lambensah

Abstract:

The objective of this research is to explore the role of actors at the local level in managing the Pre-hospital Emergency Medical Service (EMS) system in Thailand. The research method was done through documentary research, individual interviews, and one forum conducted in each province. This paper uses the case of three provinces located in three regions in Thailand including; Ubon Ratchathani (North-eastern region), Lampang (Northern Region), and Songkhla (Southern Region). The result shows that, recently, the role of the local government in being the service provider for their local people is increasingly concerned. In identifying the key success factors towards the EMS system, it includes; (i) the local executives- vision and influence that the decisions made by them, for both PAO (Provincial Administration Organisation (PAO) and TAO (Tambon Administration Organisation), is vital to address the overall challenges in EMS development, (ii) the administrative system through reforming their working style create the flexibility in running the EMS task, (iii) the network-based management among different agencies at the local level leads to the better EMS practices, and (iv) the development in human resource is very vital in delivering the effective services.

Keywords: Local governments, Management, Emergency Medical Services (EMS), Thailand

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1623
159 Information Retrieval in Domain Specific Search Engine with Machine Learning Approaches

Authors: Shilpy Sharma

Abstract:

As the web continues to grow exponentially, the idea of crawling the entire web on a regular basis becomes less and less feasible, so the need to include information on specific domain, domain-specific search engines was proposed. As more information becomes available on the World Wide Web, it becomes more difficult to provide effective search tools for information access. Today, people access web information through two main kinds of search interfaces: Browsers (clicking and following hyperlinks) and Query Engines (queries in the form of a set of keywords showing the topic of interest) [2]. Better support is needed for expressing one's information need and returning high quality search results by web search tools. There appears to be a need for systems that do reasoning under uncertainty and are flexible enough to recover from the contradictions, inconsistencies, and irregularities that such reasoning involves. In a multi-view problem, the features of the domain can be partitioned into disjoint subsets (views) that are sufficient to learn the target concept. Semi-supervised, multi-view algorithms, which reduce the amount of labeled data required for learning, rely on the assumptions that the views are compatible and uncorrelated. This paper describes the use of semi-structured machine learning approach with Active learning for the “Domain Specific Search Engines". A domain-specific search engine is “An information access system that allows access to all the information on the web that is relevant to a particular domain. The proposed work shows that with the help of this approach relevant data can be extracted with the minimum queries fired by the user. It requires small number of labeled data and pool of unlabelled data on which the learning algorithm is applied to extract the required data.

Keywords: Search engines; machine learning, Informationretrieval, Active logic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2083
158 A Simulation-Optimization Approach to Control Production, Subcontracting and Maintenance Decisions for a Deteriorating Production System

Authors: Héctor Rivera-Gómez, Eva Selene Hernández-Gress, Oscar Montaño-Arango, Jose Ramon Corona-Armenta

Abstract:

This research studies the joint production, maintenance and subcontracting control policy for an unreliable deteriorating manufacturing system. Production activities are controlled by a derivation of the Hedging Point Policy, and given that the system is subject to deterioration, it reduces progressively its capacity to satisfy product demand. Multiple deterioration effects are considered, reflected mainly in the quality of the parts produced and the reliability of the machine. Subcontracting is available as support to satisfy product demand; also, overhaul maintenance can be conducted to reduce the effects of deterioration. The main objective of the research is to determine simultaneously the production, maintenance and subcontracting rate, which minimize the total, incurred cost. A stochastic dynamic programming model is developed and solved through a simulation-based approach composed of statistical analysis and optimization with the response surface methodology. The obtained results highlight the strong interactions between production, deterioration and quality, which justify the development of an integrated model. A numerical example and a sensitivity analysis are presented to validate our results.

Keywords: Deterioration, simulation, subcontracting, production planning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1901
157 Optimisation of Structural Design by Integrating Genetic Algorithms in the Building Information Modelling Environment

Authors: Tofigh Hamidavi, Sepehr Abrishami, Pasquale Ponterosso, David Begg

Abstract:

Structural design and analysis is an important and time-consuming process, particularly at the conceptual design stage. Decisions made at this stage can have an enormous effect on the entire project, as it becomes ever costlier and more difficult to alter the choices made early on in the construction process. Hence, optimisation of the early stages of structural design can provide important efficiencies in terms of cost and time. This paper suggests a structural design optimisation (SDO) framework in which Genetic Algorithms (GAs) may be used to semi-automate the production and optimisation of early structural design alternatives. This framework has the potential to leverage conceptual structural design innovation in Architecture, Engineering and Construction (AEC) projects. Moreover, this framework improves the collaboration between the architectural stage and the structural stage. It will be shown that this SDO framework can make this achievable by generating the structural model based on the extracted data from the architectural model. At the moment, the proposed SDO framework is in the process of validation, involving the distribution of an online questionnaire among structural engineers in the UK.

Keywords: Building Information Modelling, BIM, Genetic Algorithm, GA, architecture-engineering-construction, AEC, Optimisation, structure, design, population, generation, selection, mutation, crossover, offspring.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 821
156 Evidence Theory Enabled Quickest Change Detection Using Big Time-Series Data from Internet of Things

Authors: Hossein Jafari, Xiangfang Li, Lijun Qian, Alexander Aved, Timothy Kroecker

Abstract:

Traditionally in sensor networks and recently in the Internet of Things, numerous heterogeneous sensors are deployed in distributed manner to monitor a phenomenon that often can be model by an underlying stochastic process. The big time-series data collected by the sensors must be analyzed to detect change in the stochastic process as quickly as possible with tolerable false alarm rate. However, sensors may have different accuracy and sensitivity range, and they decay along time. As a result, the big time-series data collected by the sensors will contain uncertainties and sometimes they are conflicting. In this study, we present a framework to take advantage of Evidence Theory (a.k.a. Dempster-Shafer and Dezert-Smarandache Theories) capabilities of representing and managing uncertainty and conflict to fast change detection and effectively deal with complementary hypotheses. Specifically, Kullback-Leibler divergence is used as the similarity metric to calculate the distances between the estimated current distribution with the pre- and post-change distributions. Then mass functions are calculated and related combination rules are applied to combine the mass values among all sensors. Furthermore, we applied the method to estimate the minimum number of sensors needed to combine, so computational efficiency could be improved. Cumulative sum test is then applied on the ratio of pignistic probability to detect and declare the change for decision making purpose. Simulation results using both synthetic data and real data from experimental setup demonstrate the effectiveness of the presented schemes.

Keywords: CUSUM, evidence theory, KL divergence, quickest change detection, time series data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 994
155 Forecasting 24-Hour Ahead Electricity Load Using Time Series Models

Authors: Ramin Vafadary, Maryam Khanbaghi

Abstract:

Forecasting electricity load is important for various purposes like planning, operation and control. Forecasts can save operating and maintenance costs, increase the reliability of power supply and delivery systems, and correct decisions for future development. This paper compares various time series methods to forecast 24 hours ahead of electricity load. The methods considered are the Holt-Winters smoothing, SARIMA Modeling, LSTM Network, Fbprophet and Tensorflow probability. The performance of each method is evaluated by using the forecasting accuracy criteria namely, the Mean Absolute Error and Root Mean Square Error. The National Renewable Energy Laboratory (NREL) residential energy consumption data are used to train the models. The results of this study show that SARIMA model is superior to the others for 24 hours ahead forecasts. Furthermore, a Bagging technique is used to make the predictions more robust. The obtained results show that by Bagging multiple time-series forecasts we can improve the robustness of the models for 24 hour ahead electricity load forecasting.

Keywords: Bagging, Fbprophet, Holt-Winters, LSTM, Load Forecast, SARIMA, tensorflow probability, time series.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 483
154 Fuzzy Multi-Criteria Decision-Making Based on Ignatian Discernment Process

Authors: Pathinathan Theresanathan, Ajay Minj

Abstract:

Ignatian Discernment Process (IDP) is an intense decision-making tool to decide on life-issues. Decisions are influenced by various factors outside of the decision maker and inclination within. This paper develops IDP in the context of Fuzzy Multi-criteria Decision Making (FMCDM) process. Extended VIKOR method is a decision-making method which encompasses even conflict situations and accommodates weightage to various issues. Various aspects of IDP, namely three ways of decision making and tactics of inner desires, are observed, analyzed and articulated within the frame work of fuzzy rules. The decision-making situations are broadly categorized into two types. The issues outside of the decision maker influence the person. The inner feeling also plays vital role in coming to a conclusion. IDP integrates both the categories using Extended VIKOR method. Case studies are carried out and analyzed with FMCDM process. Finally, IDP is verified with an illustrative case study and results are interpreted. A confused person who could not come to a conclusion is able to take decision on a concrete way of life through IDP. The proposed IDP model recommends an integrated and committed approach to value-based decision making.

Keywords: Analytical hierarchy process, fuzzy multi-criteria decision making, Ignatian discernment process, Ignatian discernment, multi-criteria decision making, VIKOR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1273
153 Fuzzy Optimization in Metabolic Systems

Authors: Feng-Sheng Wang, Wu-Hsiung Wu, Kai-Cheng Hsu

Abstract:

The optimization of biological systems, which is a branch of metabolic engineering, has generated a lot of industrial and academic interest for a long time. In the last decade, metabolic engineering approaches based on mathematical optimizations have been used extensively for the analysis and manipulation of metabolic networks. In practical optimization of metabolic reaction networks, designers have to manage the nature of uncertainty resulting from qualitative characters of metabolic reactions, e.g., the possibility of enzyme effects. A deterministic approach does not give an adequate representation for metabolic reaction networks with uncertain characters. Fuzzy optimization formulations can be applied to cope with this problem. A fuzzy multi-objective optimization problem can be introduced for finding the optimal engineering interventions on metabolic network systems considering the resilience phenomenon and cell viability constraints. The accuracy of optimization results depends heavily on the development of essential kinetic models of metabolic networks. Kinetic models can quantitatively capture the experimentally observed regulation data of metabolic systems and are often used to find the optimal manipulation of external inputs. To address the issues of optimizing the regulatory structure of metabolic networks, it is necessary to consider qualitative effects, e.g., the resilience phenomena and cell viability constraints. Combining the qualitative and quantitative descriptions for metabolic networks makes it possible to design a viable strain and accurately predict the maximum possible flux rates of desired products. Considering the resilience phenomena in metabolic networks can improve the predictions of gene intervention and maximum synthesis rates in metabolic engineering. Two case studies will present in the conference to illustrate the phenomena.

Keywords: Fuzzy multi-objective optimization problem, kinetic model, metabolic engineering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2018
152 Spatio-Temporal Data Mining with Association Rules for Lake Van

Authors: T. Aydin, M. F. Alaeddinoglu

Abstract:

People, throughout the history, have made estimates and inferences about the future by using their past experiences. Developing information technologies and the improvements in the database management systems make it possible to extract useful information from knowledge in hand for the strategic decisions. Therefore, different methods have been developed. Data mining by association rules learning is one of such methods. Apriori algorithm, one of the well-known association rules learning algorithms, is not commonly used in spatio-temporal data sets. However, it is possible to embed time and space features into the data sets and make Apriori algorithm a suitable data mining technique for learning spatiotemporal association rules. Lake Van, the largest lake of Turkey, is a closed basin. This feature causes the volume of the lake to increase or decrease as a result of change in water amount it holds. In this study, evaporation, humidity, lake altitude, amount of rainfall and temperature parameters recorded in Lake Van region throughout the years are used by the Apriori algorithm and a spatio-temporal data mining application is developed to identify overflows and newlyformed soil regions (underflows) occurring in the coastal parts of Lake Van. Identifying possible reasons of overflows and underflows may be used to alert the experts to take precautions and make the necessary investments.

Keywords: Apriori algorithm, association rules, data mining, spatio-temporal data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1405
151 Stackelberg Security Game for Optimizing Security of Federated Internet of Things Platform Instances

Authors: Violeta Damjanovic-Behrendt

Abstract:

This paper presents an approach for optimal cyber security decisions to protect instances of a federated Internet of Things (IoT) platform in the cloud. The presented solution implements the repeated Stackelberg Security Game (SSG) and a model called Stochastic Human behaviour model with AttRactiveness and Probability weighting (SHARP). SHARP employs the Subjective Utility Quantal Response (SUQR) for formulating a subjective utility function, which is based on the evaluations of alternative solutions during decision-making. We augment the repeated SSG (including SHARP and SUQR) with a reinforced learning algorithm called Naïve Q-Learning. Naïve Q-Learning belongs to the category of active and model-free Machine Learning (ML) techniques in which the agent (either the defender or the attacker) attempts to find an optimal security solution. In this way, we combine GT and ML algorithms for discovering optimal cyber security policies. The proposed security optimization components will be validated in a collaborative cloud platform that is based on the Industrial Internet Reference Architecture (IIRA) and its recently published security model.

Keywords: Security, internet of things, cloud computing, Stackelberg security game, machine learning, Naïve Q-learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1644
150 Aerial Firefighting Aircraft Selection with Standard Fuzzy Sets using Multiple Criteria Group Decision Making Analysis

Authors: C. Ardil

Abstract:

Aircraft selection decisions can be challenging due to their multidimensional and interdisciplinary nature. They involve multiple stakeholders with conflicting objectives and numerous alternative options with uncertain outcomes. This study focuses on the analysis of aerial firefighting aircraft that can be chosen for the Air Fire Service to extinguish forest fires. To make such a selection, the characteristics of the fire zones must be considered, and the capability to manage the logistics involved in such operations, as well as the purchase and maintenance of the aircraft, must be determined. The selection of firefighting aircraft is particularly complex because they have longer fleet lives and require more demanding operation and maintenance than scheduled passenger air service. This paper aims to use the fuzzy proximity measure method to select the most appropriate aerial firefighting aircraft based on decision criteria using multiple attribute decision making analysis. Following fuzzy decision analysis, the most suitable aerial firefighting aircraft is ranked and determined for the Air Fire Service.

Keywords: Aerial firefighting aircraft selection, multiple criteria decision making, fuzzy sets, standard fuzzy sets, determinate fuzzy sets, indeterminate fuzzy sets, proximity measure method, Minkowski distance family function, Hausdorff distance function, MCDM, PMM, PMM-F

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 399
149 Soft Real-Time Fuzzy Task Scheduling for Multiprocessor Systems

Authors: Mahdi Hamzeh, Sied Mehdi Fakhraie, Caro Lucas

Abstract:

All practical real-time scheduling algorithms in multiprocessor systems present a trade-off between their computational complexity and performance. In real-time systems, tasks have to be performed correctly and timely. Finding minimal schedule in multiprocessor systems with real-time constraints is shown to be NP-hard. Although some optimal algorithms have been employed in uni-processor systems, they fail when they are applied in multiprocessor systems. The practical scheduling algorithms in real-time systems have not deterministic response time. Deterministic timing behavior is an important parameter for system robustness analysis. The intrinsic uncertainty in dynamic real-time systems increases the difficulties of scheduling problem. To alleviate these difficulties, we have proposed a fuzzy scheduling approach to arrange real-time periodic and non-periodic tasks in multiprocessor systems. Static and dynamic optimal scheduling algorithms fail with non-critical overload. In contrast, our approach balances task loads of the processors successfully while consider starvation prevention and fairness which cause higher priority tasks have higher running probability. A simulation is conducted to evaluate the performance of the proposed approach. Experimental results have shown that the proposed fuzzy scheduler creates feasible schedules for homogeneous and heterogeneous tasks. It also and considers tasks priorities which cause higher system utilization and lowers deadline miss time. According to the results, it performs very close to optimal schedule of uni-processor systems.

Keywords: Computational complexity, Deadline, Feasible scheduling, Fuzzy scheduling, Priority, Real-time multiprocessor systems, Robustness, System utilization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2129
148 A New Brazilian Friction-Resistant Low Alloy High Strength Steel – A Life Testing Approach

Authors: D. I. De Souza, G. P. Azevedo, R. Rocha

Abstract:

In this paper we will develop a sequential life test approach applied to a modified low alloy-high strength steel part used in highway overpasses in Brazil.We will consider two possible underlying sampling distributions: the Normal and theInverse Weibull models. The minimum life will be considered equal to zero. We will use the two underlying models to analyze a fatigue life test situation, comparing the results obtained from both.Since a major chemical component of this low alloy-high strength steel part has been changed, there is little information available about the possible values that the parameters of the corresponding Normal and Inverse Weibull underlying sampling distributions could have. To estimate the shape and the scale parameters of these two sampling models we will use a maximum likelihood approach for censored failure data. We will also develop a truncation mechanism for the Inverse Weibull and Normal models. We will provide rules to truncate a sequential life testing situation making one of the two possible decisions at the moment of truncation; that is, accept or reject the null hypothesis H0. An example will develop the proposed truncated sequential life testing approach for the Inverse Weibull and Normal models.

Keywords: Sequential life testing, normal and inverse Weibull models, maximum likelihood approach, truncation mechanism.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1429
147 Comparing Field Displacement History with Numerical Results to Estimate Geotechnical Parameters: Case Study of Arash-Esfandiar-Niayesh under Passing Tunnel, 2.5 Traffic Lane Tunnel, Tehran, Iran

Authors: A. Golshani, M. Gharizade Varnusefaderani, S. Majidian

Abstract:

Underground structures are of those structures that have uncertainty in design procedures. That is due to the complexity of soil condition around. Under passing tunnels are also such affected structures. Despite geotechnical site investigations, lots of uncertainties exist in soil properties due to unknown events. As results, it possibly causes conflicting settlements in numerical analysis with recorded values in the project. This paper aims to report a case study on a specific under passing tunnel constructed by New Austrian Tunnelling Method in Iran. The intended tunnel has an overburden of about 11.3m, the height of 12.2m and, the width of 14.4m with 2.5 traffic lane. The numerical modeling was developed by a 2D finite element program (PLAXIS Version 8). Comparing displacement histories at the ground surface during the entire installation of initial lining, the estimated surface settlement was about four times the field recorded one, which indicates that some local unknown events affect that value. Also, the displacement ratios were in a big difference between the numerical and field data. Consequently, running several numerical back analyses using laboratory and field tests data, the geotechnical parameters were accurately revised to match with the obtained monitoring data. Finally, it was found that usually the values of soil parameters are conservatively low-estimated up to 40 percent by typical engineering judgment. Additionally, it could be attributed to inappropriate constitutive models applied for the specific soil condition.

Keywords: NATM, surface displacement history, soil tests, monitoring data, numerical back-analysis, geotechnical parameters.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 799
146 Development of a Technology Assessment Model by Patents and Customers' Review Data

Authors: Kisik Song, Sungjoo Lee

Abstract:

Recent years have seen an increasing number of patent disputes due to excessive competition in the global market and a reduced technology life-cycle; this has increased the risk of investment in technology development. While many global companies have started developing a methodology to identify promising technologies and assess for decisions, the existing methodology still has some limitations. Post hoc assessments of the new technology are not being performed, especially to determine whether the suggested technologies turned out to be promising. For example, in existing quantitative patent analysis, a patent’s citation information has served as an important metric for quality assessment, but this analysis cannot be applied to recently registered patents because such information accumulates over time. Therefore, we propose a new technology assessment model that can replace citation information and positively affect technological development based on post hoc analysis of the patents for promising technologies. Additionally, we collect customer reviews on a target technology to extract keywords that show the customers’ needs, and we determine how many keywords are covered in the new technology. Finally, we construct a portfolio (based on a technology assessment from patent information) and a customer-based marketability assessment (based on review data), and we use them to visualize the characteristics of the new technologies.

Keywords: Technology assessment, patents, citation information, opinion mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 992
145 Recommended Practice for Experimental Evaluation of the Seepage Sensitivity Damage of Coalbed Methane Reservoirs

Authors: Hao Liu, Lihui Zheng, Chinedu J. Okere, Chao Wang, Xiangchun Wang, Peng Zhang

Abstract:

The coalbed methane (CBM) extraction industry (an unconventional energy source) has not established guidelines for experimental evaluation of sensitivity damage for coal samples. The existing experimental process of previous researches mainly followed the industry standard for conventional oil and gas reservoirs (CIS). However, the existing evaluation method ignores certain critical differences between CBM reservoirs and conventional reservoirs, which could inevitably result in an inaccurate evaluation of sensitivity damage and, eventually, poor decisions regarding the formulation of formation damage prevention measures. In this study, we propose improved experimental guidelines for evaluating seepage sensitivity damage of CBM reservoirs by leveraging on the shortcomings of the existing methods. The proposed method was established via a theoretical analysis of the main drawbacks of the existing methods and validated through comparative experiments. The results show that the proposed evaluation technique provided reliable experimental results that can better reflect actual reservoir conditions and correctly guide the future development of CBM reservoirs. This study is pioneering the research on the optimization of experimental parameters for efficient exploration and development of CBM reservoirs.

Keywords: Coalbed methane, formation damage, permeability, unconventional energy source.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 385