Search results for: decision based artificial neural network
30887 Optimal Bayesian Control of the Proportion of Defectives in a Manufacturing Process
Authors: Viliam Makis, Farnoosh Naderkhani, Leila Jafari
Abstract:
In this paper, we present a model and an algorithm for the calculation of the optimal control limit, average cost, sample size, and the sampling interval for an optimal Bayesian chart to control the proportion of defective items produced using a semi-Markov decision process approach. Traditional p-chart has been widely used for controlling the proportion of defectives in various kinds of production processes for many years. It is well known that traditional non-Bayesian charts are not optimal, but very few optimal Bayesian control charts have been developed in the literature, mostly considering finite horizon. The objective of this paper is to develop a fast computational algorithm to obtain the optimal parameters of a Bayesian p-chart. The decision problem is formulated in the partially observable framework and the developed algorithm is illustrated by a numerical example.Keywords: Bayesian control chart, semi-Markov decision process, quality control, partially observable process
Procedia PDF Downloads 31930886 Method for Requirements Analysis and Decision Making for Restructuring Projects in Factories
Authors: Rene Hellmuth
Abstract:
The requirements for the factory planning and the building concerned have changed in the last years. Factory planning has the task of designing products, plants, processes, organization, areas, and the building of a factory. Regular restructuring gains more importance in order to maintain the competitiveness of a factory. Restrictions regarding new areas, shorter life cycles of product and production technology as well as a VUCA (volatility, uncertainty, complexity and ambiguity) world cause more frequently occurring rebuilding measures within a factory. Restructuring of factories is the most common planning case today. Restructuring is more common than new construction, revitalization and dismantling of factories. The increasing importance of restructuring processes shows that the ability to change was and is a promising concept for the reaction of companies to permanently changing conditions. The factory building is the basis for most changes within a factory. If an adaptation of a construction project (factory) is necessary, the inventory documents must be checked and often time-consuming planning of the adaptation must take place to define the relevant components to be adapted, in order to be able to finally evaluate them. The different requirements of the planning participants from the disciplines of factory planning (production planner, logistics planner, automation planner) and industrial construction planning (architect, civil engineer) come together during reconstruction and must be structured. This raises the research question: Which requirements do the disciplines involved in the reconstruction planning place on a digital factory model? A subordinate research question is: How can model-based decision support be provided for a more efficient design of the conversion within a factory? Because of the high adaptation rate of factories and its building described above, a methodology for rescheduling factories based on the requirements engineering method from software development is conceived and designed for practical application in factory restructuring projects. The explorative research procedure according to Kubicek is applied. Explorative research is suitable if the practical usability of the research results has priority. Furthermore, it will be shown how to best use a digital factory model in practice. The focus will be on mobile applications to meet the needs of factory planners on site. An augmented reality (AR) application will be designed and created to provide decision support for planning variants. The aim is to contribute to a shortening of the planning process and model-based decision support for more efficient change management. This requires the application of a methodology that reduces the deficits of the existing approaches. The time and cost expenditure are represented in the AR tablet solution based on a building information model (BIM). Overall, the requirements of those involved in the planning process for a digital factory model in the case of restructuring within a factory are thus first determined in a structured manner. The results are then applied and transferred to a construction site solution based on augmented reality.Keywords: augmented reality, digital factory model, factory planning, restructuring
Procedia PDF Downloads 13430885 Community Empowerment: The Contribution of Network Urbanism on Urban Poverty Reduction
Authors: Lucia Antonela Mitidieri
Abstract:
This research analyzes the application of a model of settlements management based on networks of territorial integration that advocates planning as a cyclical and participatory process that engages early on with civic society, the private sector and the state. Through qualitative methods such as participant observation, interviews with snowball technique and an active research on territories, concrete results of community empowerment are obtained from the promotion of productive enterprises and community spaces of encounter and exchange. Studying the cultural and organizational dimensions of empowerment allows building indicators such as increase of capacities or community cohesion that can lead to support local governments in achieving sustainable urban development for a reduction of urban poverty.Keywords: community spaces, empowerment, network urbanism, participatory process
Procedia PDF Downloads 33130884 Location Privacy Preservation of Vehicle Data In Internet of Vehicles
Authors: Ying Ying Liu, Austin Cooke, Parimala Thulasiraman
Abstract:
Internet of Things (IoT) has attracted a recent spark in research on Internet of Vehicles (IoV). In this paper, we focus on one research area in IoV: preserving location privacy of vehicle data. We discuss existing location privacy preserving techniques and provide a scheme for evaluating these techniques under IoV traffic condition. We propose a different strategy in applying Differential Privacy using k-d tree data structure to preserve location privacy and experiment on real world Gowalla data set. We show that our strategy produces differentially private data, good preservation of utility by achieving similar regression accuracy to the original dataset on an LSTM (Long Term Short Term Memory) neural network traffic predictor.Keywords: differential privacy, internet of things, internet of vehicles, location privacy, privacy preservation scheme
Procedia PDF Downloads 18030883 Patronage Network and Ideological Manipulations in Translation of Literary Texts: A Case Study of George Orwell's “1984” in Persian Translation in the Period 1980 to 2015
Authors: Masoud Hassanzade Novin, Bahloul Salmani
Abstract:
The process of the translation is not merely the linguistic aspects. It is also considered in the cultural framework of both the source and target text cultures. The translation process and translated texts are confronted the new aspect in 20th century which is considered mostly in the patronage framework and ideological grillwork of the target language. To have these factors scrutinized in the process of the translation both micro-element factors and macro-element factors can be taken into consideration. For the purpose of this study through a qualitative type of research based on critical discourse analysis approach, the case study of the novel “1984” written by George Orwell was chosen as the corpus of the study to have the contrastive analysis by its Persian translated texts. Results of the study revealed some distortions embedded in the target texts which were overshadowed by ideological aspect and patronage network. The outcomes of the manipulated terms were different in various categories which revealed the manipulation aspects in the texts translated.Keywords: critical discourse analysis, ideology, patronage network, translated texts
Procedia PDF Downloads 32230882 IoT Based Information Processing and Computing
Authors: Mannan Ahmad Rasheed, Sawera Kanwal, Mansoor Ahmad Rasheed
Abstract:
The Internet of Things (IoT) has revolutionized the way we collect and process information, making it possible to gather data from a wide range of connected devices and sensors. This has led to the development of IoT-based information processing and computing systems that are capable of handling large amounts of data in real time. This paper provides a comprehensive overview of the current state of IoT-based information processing and computing, as well as the key challenges and gaps that need to be addressed. This paper discusses the potential benefits of IoT-based information processing and computing, such as improved efficiency, enhanced decision-making, and cost savings. Despite the numerous benefits of IoT-based information processing and computing, several challenges need to be addressed to realize the full potential of these systems. These challenges include security and privacy concerns, interoperability issues, scalability and reliability of IoT devices, and the need for standardization and regulation of IoT technologies. Moreover, this paper identifies several gaps in the current research related to IoT-based information processing and computing. One major gap is the lack of a comprehensive framework for designing and implementing IoT-based information processing and computing systems.Keywords: IoT, computing, information processing, Iot computing
Procedia PDF Downloads 18830881 Optimal Protection Coordination in Distribution Systems with Distributed Generations
Authors: Abdorreza Rabiee, Shahla Mohammad Hoseini Mirzaei
Abstract:
The advantages of distributed generations (DGs) based on renewable energy sources (RESs) leads to high penetration level of DGs in distribution network. With incorporation of DGs in distribution systems, the system reliability and security, as well as voltage profile, is improved. However, the protection of such systems is still challenging. In this paper, at first, the related papers are reviewed and then a practical scheme is proposed for coordination of OCRs in distribution system with DGs. The coordination problem is formulated as a nonlinear programming (NLP) optimization problem with the object function of minimizing total operating time of OCRs. The proposed method is studied based on a simple test system. The optimization problem is solved by General Algebraic Modeling System (GAMS) to calculate the optimal time dial setting (TDS) and also pickup current setting of OCRs. The results show the effectiveness of the proposed method and its applicability.Keywords: distributed generation, DG, distribution network, over current relay, OCR, protection coordination, pickup current, time dial setting, TDS
Procedia PDF Downloads 13930880 A Recommender System for Job Seekers to Show up Companies Based on Their Psychometric Preferences and Company Sentiment Scores
Authors: A. Ashraff
Abstract:
The increasing importance of the web as a medium for electronic and business transactions has served as a catalyst or rather a driving force for the introduction and implementation of recommender systems. Recommender Systems play a major role in processing and analyzing thousands of data rows or reviews and help humans make a purchase decision of a product or service. It also has the ability to predict whether a particular user would rate a product or service based on the user’s profile behavioral pattern. At present, Recommender Systems are being used extensively in every domain known to us. They are said to be ubiquitous. However, in the field of recruitment, it’s not being utilized exclusively. Recent statistics show an increase in staff turnover, which has negatively impacted the organization as well as the employee. The reasons being company culture, working flexibility (work from home opportunity), no learning advancements, and pay scale. Further investigations revealed that there are lacking guidance or support, which helps a job seeker find the company that will suit him best, and though there’s information available about companies, job seekers can’t read all the reviews by themselves and get an analytical decision. In this paper, we propose an approach to study the available review data on IT companies (score their reviews based on user review sentiments) and gather information on job seekers, which includes their Psychometric evaluations. Then presents the job seeker with useful information or rather outputs on which company is most suitable for the job seeker. The theoretical approach, Algorithmic approach and the importance of such a system will be discussed in this paper.Keywords: psychometric tests, recommender systems, sentiment analysis, hybrid recommender systems
Procedia PDF Downloads 10730879 Destination Decision Model for Cruising Taxis Based on Embedding Model
Authors: Kazuki Kamada, Haruka Yamashita
Abstract:
In Japan, taxi is one of the popular transportations and taxi industry is one of the big businesses. However, in recent years, there has been a difficult problem of reducing the number of taxi drivers. In the taxi business, mainly three passenger catching methods are applied. One style is "cruising" that drivers catches passengers while driving on a road. Second is "waiting" that waits passengers near by the places with many requirements for taxies such as entrances of hospitals, train stations. The third one is "dispatching" that is allocated based on the contact from the taxi company. Above all, the cruising taxi drivers need the experience and intuition for finding passengers, and it is difficult to decide "the destination for cruising". The strong recommendation system for the cruising taxies supports the new drivers to find passengers, and it can be the solution for the decreasing the number of drivers in the taxi industry. In this research, we propose a method of recommending a destination for cruising taxi drivers. On the other hand, as a machine learning technique, the embedding models that embed the high dimensional data to a low dimensional space is widely used for the data analysis, in order to represent the relationship of the meaning between the data clearly. Taxi drivers have their favorite courses based on their experiences, and the courses are different for each driver. We assume that the course of cruising taxies has meaning such as the course for finding business man passengers (go around the business area of the city of go to main stations) and course for finding traveler passengers (go around the sightseeing places or big hotels), and extract the meaning of their destinations. We analyze the cruising history data of taxis based on the embedding model and propose the recommendation system for passengers. Finally, we demonstrate the recommendation of destinations for cruising taxi drivers based on the real-world data analysis using proposing method.Keywords: taxi industry, decision making, recommendation system, embedding model
Procedia PDF Downloads 13830878 Automated Feature Extraction and Object-Based Detection from High-Resolution Aerial Photos Based on Machine Learning and Artificial Intelligence
Authors: Mohammed Al Sulaimani, Hamad Al Manhi
Abstract:
With the development of Remote Sensing technology, the resolution of optical Remote Sensing images has greatly improved, and images have become largely available. Numerous detectors have been developed for detecting different types of objects. In the past few years, Remote Sensing has benefited a lot from deep learning, particularly Deep Convolution Neural Networks (CNNs). Deep learning holds great promise to fulfill the challenging needs of Remote Sensing and solving various problems within different fields and applications. The use of Unmanned Aerial Systems in acquiring Aerial Photos has become highly used and preferred by most organizations to support their activities because of their high resolution and accuracy, which make the identification and detection of very small features much easier than Satellite Images. And this has opened an extreme era of Deep Learning in different applications not only in feature extraction and prediction but also in analysis. This work addresses the capacity of Machine Learning and Deep Learning in detecting and extracting Oil Leaks from Flowlines (Onshore) using High-Resolution Aerial Photos which have been acquired by UAS fixed with RGB Sensor to support early detection of these leaks and prevent the company from the leak’s losses and the most important thing environmental damage. Here, there are two different approaches and different methods of DL have been demonstrated. The first approach focuses on detecting the Oil Leaks from the RAW Aerial Photos (not processed) using a Deep Learning called Single Shoot Detector (SSD). The model draws bounding boxes around the leaks, and the results were extremely good. The second approach focuses on detecting the Oil Leaks from the Ortho-mosaiced Images (Georeferenced Images) by developing three Deep Learning Models using (MaskRCNN, U-Net and PSP-Net Classifier). Then, post-processing is performed to combine the results of these three Deep Learning Models to achieve a better detection result and improved accuracy. Although there is a relatively small amount of datasets available for training purposes, the Trained DL Models have shown good results in extracting the extent of the Oil Leaks and obtaining excellent and accurate detection.Keywords: GIS, remote sensing, oil leak detection, machine learning, aerial photos, unmanned aerial systems
Procedia PDF Downloads 3430877 The Impact of Artificial Intelligence on Higher Education in Latin America
Authors: Luis Rodrigo Valencia Perez, Francisco Flores Aguero, Gibran Aguilar Rangel
Abstract:
Artificial Intelligence (AI) is rapidly transforming diverse sectors, and higher education in Latin America is no exception. This article explores the impact of AI on higher education institutions in the region, highlighting the imperative need for well-trained teachers in emerging technologies and a cultural shift towards the adoption and efficient use of these tools. AI offers significant opportunities to improve learning personalization, optimize administrative processes, and promote more inclusive and accessible education. However, the effectiveness of its implementation depends largely on the preparation and willingness of teachers to integrate these technologies into their pedagogical practices. Furthermore, it is essential that Latin American countries develop and implement public policies that encourage the adoption of AI in the education sector, thus ensuring that institutions can compete globally. Policies should focus on the continuous training of educators, investment in technological infrastructure, and the creation of regulatory frameworks that promote innovation and the ethical use of AI. Only through a comprehensive and collaborative approach will it be possible to fully harness the potential of AI to transform higher education in Latin America, thereby boosting the region's development and competitiveness on the global stage.Keywords: artificial intelligence (AI), higher education, teacher training, public policies, latin america, global competitiveness
Procedia PDF Downloads 2830876 2D Convolutional Networks for Automatic Segmentation of Knee Cartilage in 3D MRI
Authors: Ananya Ananya, Karthik Rao
Abstract:
Accurate segmentation of knee cartilage in 3-D magnetic resonance (MR) images for quantitative assessment of volume is crucial for studying and diagnosing osteoarthritis (OA) of the knee, one of the major causes of disability in elderly people. Radiologists generally perform this task in slice-by-slice manner taking 15-20 minutes per 3D image, and lead to high inter and intra observer variability. Hence automatic methods for knee cartilage segmentation are desirable and are an active field of research. This paper presents design and experimental evaluation of 2D convolutional neural networks based fully automated methods for knee cartilage segmentation in 3D MRI. The architectures are validated based on 40 test images and 60 training images from SKI10 dataset. The proposed methods segment 2D slices one by one, which are then combined to give segmentation for whole 3D images. Proposed methods are modified versions of U-net and dilated convolutions, consisting of a single step that segments the given image to 5 labels: background, femoral cartilage, tibia cartilage, femoral bone and tibia bone; cartilages being the primary components of interest. U-net consists of a contracting path and an expanding path, to capture context and localization respectively. Dilated convolutions lead to an exponential expansion of receptive field with only a linear increase in a number of parameters. A combination of modified U-net and dilated convolutions has also been explored. These architectures segment one 3D image in 8 – 10 seconds giving average volumetric Dice Score Coefficients (DSC) of 0.950 - 0.962 for femoral cartilage and 0.951 - 0.966 for tibia cartilage, reference being the manual segmentation.Keywords: convolutional neural networks, dilated convolutions, 3 dimensional, fully automated, knee cartilage, MRI, segmentation, U-net
Procedia PDF Downloads 26130875 Orbiting Intelligence: A Comprehensive Survey of AI Applications and Advancements in Space Exploration
Authors: Somoshree Datta, Chithra A. V., Sandeep Nithyanandan, Smitha K. K.
Abstract:
Space exploration has always been at the forefront of technological innovation, pushing the boundaries of human knowledge and capabilities. In recent years, the integration of Artificial Intelligence (AI) has revolutionized the field, offering unprecedented opportunities to enhance the efficiency, autonomy and intelligence of space missions. This survey paper aims to provide a comprehensive overview of the multifaceted applications of AI in space exploration, exploring the evolution of this synergy and its impact on mission success, scientific discovery, and the future of space endeavors. Indian Space Research Organization (ISRO) has achieved great feats in the recent moon mission (Chandrayaan-3) and sun mission (Aditya L1) by using artificial intelligence to enhance moon navigation as well as help young scientists to study the Sun even before the launch by creating AI-generated image visualizations. Throughout this survey, we will review key advancements, challenges and prospects in the intersection of AI and space exploration. As humanity continues its quest to explore the cosmos, the integration of AI promises to unlock new frontiers, reshape mission architectures, and redefine our understanding of the universe. This survey aims to serve as a comprehensive resource for researchers, engineers and enthusiasts interested in the dynamic and evolving landscape of AI applications in space exploration.Keywords: artificial intelligence, space exploration, space missions, deep learning
Procedia PDF Downloads 3430874 Mitigating Denial of Service Attacks in Information Centric Networking
Authors: Bander Alzahrani
Abstract:
Information-centric networking (ICN) using architectures such as Publish-Subscribe Internet Routing Paradigm (PSIRP) is one of the promising candidates for a future Internet, has recently been under the spotlight by the research community to investigate the possibility of redesigning the current Internet architecture to solve many issues such as routing scalability, security, and quality of services issues.. The Bloom filter-based forwarding is a source-routing approach that is used in the PSIRP architecture. This mechanism is vulnerable to brute force attacks which may lead to denial-of-service (DoS) attacks. In this work, we present a new forwarding approach that keeps the advantages of Bloom filter-based forwarding while mitigates attacks on the forwarding mechanism. In practice, we introduce a special type of forwarding nodes called Edge-FW to be placed at the edge of the network. The role of these node is to add an extra security layer by validating and inspecting packets at the edge of the network against brute-force attacks and check whether the packet contains a legitimate forwarding identifier (FId) or not. We leverage Certificateless Aggregate Signature (CLAS) scheme with a small size of 64-bit which is used to sign the FId. Hence, this signature becomes bound to a specific FId. Therefore, malicious nodes that inject packets with random FIds will be easily detected and dropped at the Edge-FW node when the signature verification fails. Our preliminary security analysis suggests that with the proposed approach, the forwarding plane is able to resist attacks such as DoS with very high probability.Keywords: bloom filter, certificateless aggregate signature, denial-of-service, information centric network
Procedia PDF Downloads 19830873 Power Control in Solar Battery Charging Station Using Fuzzy Decision Support System
Authors: Krishnan Manickavasagam, Manikandan Shanmugam
Abstract:
Clean and abundant renewable energy sources (RES) such as solar energy is seen as the best solution to replace conventional energy source. Unpredictable power generation is a major issue in the penetration of solar energy, as power generated is governed by the irradiance received. Controlling the power generated from solar PV (SPV) panels to battery and load is a challenging task. In this paper, power flow control from SPV to load and energy storage device (ESD) is controlled by a fuzzy decision support system (FDSS) on the availability of solar irradiation. The results show that FDSS implemented with the energy management system (EMS) is capable of managing power within the area, and if excess power is available, then shared with the neighboring area.Keywords: renewable energy sources, fuzzy decision support system, solar photovoltaic, energy storage device, energy management system
Procedia PDF Downloads 10030872 Analysis of Performance Improvement Factors in Supply Chain Manufacturing Using Analytic Network Process and Kaizen
Authors: Juliza Hidayati, Yesie M. Sinuhaji, Sawarni Hasibuan
Abstract:
A company producing drinking water through many incompatibility issues that affect supply chain performance. The study was conducted to determine the factors that affect the performance of the supply chain and improve it. To obtain the dominant factors affecting the performance of the supply chain used Analytic Network Process, while to improve performance is done by using Kaizen. Factors affecting the performance of the supply chain to be a reference to identify the cause of the non-conformance. Results weighting using ANP indicates that the dominant factor affecting the level of performance is the precision of the number of shipments (15%), the ability of the fulfillment of the booking amount (12%), and the number of rejected products when signing (12%). Incompatibility of the factors that affect the performance of the supply chain are identified, so that found the root cause of the problem is most dominant. Based on the weight of Risk Priority Number (RPN) gained the most dominant root cause of the problem, namely the poorly maintained engine, the engine worked for three shifts, machine parts that are not contained in the plant. Improvements then performed using the Kaizen method of systematic and sustainable.Keywords: analytic network process, booking amount, risk priority number, supply chain performance
Procedia PDF Downloads 29430871 A Location-based Authentication and Key Management Scheme for Border Surveillance Wireless Sensor Networks
Authors: Walid Abdallah, Noureddine Boudriga
Abstract:
Wireless sensor networks have shown their effectiveness in the deployment of many critical applications especially in the military domain. Border surveillance is one of these applications where a set of wireless sensors are deployed along a country border line to detect illegal intrusion attempts to the national territory and report this to a control center to undergo the necessary measures. Regarding its nature, this wireless sensor network can be the target of many security attacks trying to compromise its normal operation. Particularly, in this application the deployment and location of sensor nodes are of great importance for detecting and tracking intruders. This paper proposes a location-based authentication and key distribution mechanism to secure wireless sensor networks intended for border surveillance where the key establishment is performed using elliptic curve cryptography and identity-based public key scheme. In this scheme, the public key of each sensor node will be authenticated by keys that depend on its position in the monitored area. Before establishing a pairwise key between two nodes, each one of them must verify the neighborhood location of the other node using a message authentication code (MAC) calculated on the corresponding public key and keys derived from encrypted beacon messages broadcast by anchor nodes. We show that our proposed public key authentication and key distribution scheme is more resilient to node capture and node replication attacks than currently available schemes. Also, the achievement of the key distribution between nodes in our scheme generates less communication overhead and hence increases network performances.Keywords: wireless sensor networks, border surveillance, security, key distribution, location-based
Procedia PDF Downloads 66030870 Data-driven Decision-Making in Digital Entrepreneurship
Authors: Abeba Nigussie Turi, Xiangming Samuel Li
Abstract:
Data-driven business models are more typical for established businesses than early-stage startups that strive to penetrate a market. This paper provided an extensive discussion on the principles of data analytics for early-stage digital entrepreneurial businesses. Here, we developed data-driven decision-making (DDDM) framework that applies to startups prone to multifaceted barriers in the form of poor data access, technical and financial constraints, to state some. The startup DDDM framework proposed in this paper is novel in its form encompassing startup data analytics enablers and metrics aligning with startups' business models ranging from customer-centric product development to servitization which is the future of modern digital entrepreneurship.Keywords: startup data analytics, data-driven decision-making, data acquisition, data generation, digital entrepreneurship
Procedia PDF Downloads 32930869 Long-Term Mechanical and Structural Properties of Metakaolin-Based Geopolymers
Authors: Lenka Matulova
Abstract:
Geopolymers are alumosilicate materials that have long been studied. Despite this fact, little is known about the long-term stability of geopolymer mechanical and structural properties, so crucial for their successful industrial application. To improve understanding, we investigated the effect of four different types of environments on the mechanical and structural properties of a metakaolin-based geopolymer (MK GP). The MK GP samples were stored in laboratory conditions (control samples), in water at 20 °C, in water at 80 °C, and outside exposed to the weather. Compressive and tensile strengths were measured after 28, 56, 90, and 360 days. In parallel, structural properties were analyzed using XRD, SEM, and mercury intrusion porosimetry. Whereas the mechanical properties of the samples in laboratory conditions and in 20 °C water were stable, the mechanical properties of the outdoor samples and the samples 80 °C water decreased noticeably after 360 days. Structural analyses were focused on changes in sample microstructure (developing microcrack network, porosity) and identifying zeolites, the presence of which would indicate detrimental processes in the structure that can change it from amorphous to crystalline. No zeolites were found during the 360-day period in MK GP samples, but the reduction in mechanical properties coincided with a developing network of microcracks and changes in pore size distribution.Keywords: geopolymer, long-term properties, mechanical properties, metakaolin, structural properties
Procedia PDF Downloads 24130868 Predicting Options Prices Using Machine Learning
Authors: Krishang Surapaneni
Abstract:
The goal of this project is to determine how to predict important aspects of options, including the ask price. We want to compare different machine learning models to learn the best model and the best hyperparameters for that model for this purpose and data set. Option pricing is a relatively new field, and it can be very complicated and intimidating, especially to inexperienced people, so we want to create a machine learning model that can predict important aspects of an option stock, which can aid in future research. We tested multiple different models and experimented with hyperparameter tuning, trying to find some of the best parameters for a machine-learning model. We tested three different models: a Random Forest Regressor, a linear regressor, and an MLP (multi-layer perceptron) regressor. The most important feature in this experiment is the ask price; this is what we were trying to predict. In the field of stock pricing prediction, there is a large potential for error, so we are unable to determine the accuracy of the models based on if they predict the pricing perfectly. Due to this factor, we determined the accuracy of the model by finding the average percentage difference between the predicted and actual values. We tested the accuracy of the machine learning models by comparing the actual results in the testing data and the predictions made by the models. The linear regression model performed worst, with an average percentage error of 17.46%. The MLP regressor had an average percentage error of 11.45%, and the random forest regressor had an average percentage error of 7.42%Keywords: finance, linear regression model, machine learning model, neural network, stock price
Procedia PDF Downloads 7630867 Artificial Intelligent Tax Simulator to Minimize Tax Liability for Multinational Corporations
Authors: Sean Goltz, Michael Mayo
Abstract:
The purpose of this research is to use Global-Regulation.com database of the world laws, focusing on tax treaties between countries, in order to create an AI-driven tax simulator that will run an AI agent through potential tax scenarios across countries. The AI agent goal is to identify the scenario that will result in minimum tax liability based on tax treaties between countries. The results will be visualized by a three dimensional matrix. This will be an online web application. Multinational corporations are running their business through multiple countries. These countries, in turn, have a tax treaty with many other countries to regulate the payment of taxes on income that is transferred between these countries. As a result, planning the best tax scenario across multiple countries and numerous tax treaties is almost impossible. This research propose to use Global-Regulation.com database of word laws in English (machine translated by Google and Microsoft API’s) in order to create a simulator that will include the information in the tax treaties. Once ready, an AI agent will be sent through the simulator to identify the scenario that will result in minimum tax liability. Identifying the best tax scenario across countries may save multinational corporations, like Google, billions of dollars annually. Given the nature of the raw data and the domain of taxes (i.e., numbers), this is a promising ground to employ artificial intelligence towards a practical and beneficial purpose.Keywords: taxation, law, multinational, corporation
Procedia PDF Downloads 19930866 Evaluating Multiple Diagnostic Tests: An Application to Cervical Intraepithelial Neoplasia
Authors: Areti Angeliki Veroniki, Sofia Tsokani, Evangelos Paraskevaidis, Dimitris Mavridis
Abstract:
The plethora of diagnostic test accuracy (DTA) studies has led to the increased use of systematic reviews and meta-analysis of DTA studies. Clinicians and healthcare professionals often consult DTA meta-analyses to make informed decisions regarding the optimum test to choose and use for a given setting. For example, the human papilloma virus (HPV) DNA, mRNA, and cytology can be used for the cervical intraepithelial neoplasia grade 2+ (CIN2+) diagnosis. But which test is the most accurate? Studies directly comparing test accuracy are not always available, and comparisons between multiple tests create a network of DTA studies that can be synthesized through a network meta-analysis of diagnostic tests (DTA-NMA). The aim is to summarize the DTA-NMA methods for at least three index tests presented in the methodological literature. We illustrate the application of the methods using a real data set for the comparative accuracy of HPV DNA, HPV mRNA, and cytology tests for cervical cancer. A search was conducted in PubMed, Web of Science, and Scopus from inception until the end of July 2019 to identify full-text research articles that describe a DTA-NMA method for three or more index tests. Since the joint classification of the results from one index against the results of another index test amongst those with the target condition and amongst those without the target condition are rarely reported in DTA studies, only methods requiring the 2x2 tables of the results of each index test against the reference standard were included. Studies of any design published in English were eligible for inclusion. Relevant unpublished material was also included. Ten relevant studies were finally included to evaluate their methodology. DTA-NMA methods that have been presented in the literature together with their advantages and disadvantages are described. In addition, using 37 studies for cervical cancer obtained from a published Cochrane review as a case study, an application of the identified DTA-NMA methods to determine the most promising test (in terms of sensitivity and specificity) for use as the best screening test to detect CIN2+ is presented. As a conclusion, different approaches for the comparative DTA meta-analysis of multiple tests may conclude to different results and hence may influence decision-making. Acknowledgment: This research is co-financed by Greece and the European Union (European Social Fund- ESF) through the Operational Programme «Human Resources Development, Education and Lifelong Learning 2014-2020» in the context of the project “Extension of Network Meta-Analysis for the Comparison of Diagnostic Tests ” (MIS 5047640).Keywords: colposcopy, diagnostic test, HPV, network meta-analysis
Procedia PDF Downloads 13930865 Decision Making Approach through Generalized Fuzzy Entropy Measure
Authors: H. D. Arora, Anjali Dhiman
Abstract:
Uncertainty is found everywhere and its understanding is central to decision making. Uncertainty emerges as one has less information than the total information required describing a system and its environment. Uncertainty and information are so closely associated that the information provided by an experiment for example, is equal to the amount of uncertainty removed. It may be pertinent to point out that uncertainty manifests itself in several forms and various kinds of uncertainties may arise from random fluctuations, incomplete information, imprecise perception, vagueness etc. For instance, one encounters uncertainty due to vagueness in communication through natural language. Uncertainty in this sense is represented by fuzziness resulting from imprecision of meaning of a concept expressed by linguistic terms. Fuzzy set concept provides an appropriate mathematical framework for dealing with the vagueness. Both information theory, proposed by Shannon (1948) and fuzzy set theory given by Zadeh (1965) plays an important role in human intelligence and various practical problems such as image segmentation, medical diagnosis etc. Numerous approaches and theories dealing with inaccuracy and uncertainty have been proposed by different researcher. In the present communication, we generalize fuzzy entropy proposed by De Luca and Termini (1972) corresponding to Shannon entropy(1948). Further, some of the basic properties of the proposed measure were examined. We also applied the proposed measure to the real life decision making problem.Keywords: entropy, fuzzy sets, fuzzy entropy, generalized fuzzy entropy, decision making
Procedia PDF Downloads 45030864 The Artificial Intelligence (AI) Impact on Project Management: A Destructive or Transformative Agent
Authors: Kwame Amoah
Abstract:
Artificial intelligence (AI) has the prospect of transforming project management, significantly improving efficiency and accuracy. By automating specific tasks with defined guidelines, AI can assist project managers in making better decisions and allocating resources efficiently, with possible risk mitigation. This study explores how AI is already impacting project management and likely future AI's impact on the field. The AI's reaction has been a divided opinion; while others picture it as a destroyer of jobs, some welcome it as an innovation advocate. Both sides agree that AI will be disruptive and revolutionize PM's functions. If current research is to go by, AI or some form will replace one-third of all learning graduate PM jobs by as early as 2030. A recent survey indicates AI spending will reach $97.9 billion by the end of 2023. Considering such a profound impact, the project management profession will also see a paradigm shift driven by AI. The study examines what the project management profession will look like in the next 5-10 years after this technological disruption. The research methods incorporate existing literature, develop trend analysis, and conduct structured interviews with project management stakeholders from North America to gauge the trend. PM professionals can harness the power of AI, ensuring a smooth transition and positive outcomes. AI adoption will maximize benefits, minimize adverse consequences, and uphold ethical standards, leading to improved project performance.Keywords: project management, disruptive teacnologies, project management function, AL applications, artificial intelligence
Procedia PDF Downloads 8330863 Proactive Approach to Innovation Management
Authors: Andrus Pedai, Igor Astrov
Abstract:
The focus of this paper is to compare common approaches for Systems of Innovation (SI) and identify proactive alternatives for driving the innovation. Proactive approaches will also consider short and medium term perspectives with developments in the field of Computer Technology and Artificial Intelligence. Concerning computer technology and large connected information systems, it is reasonable to predict that during current or the next century, intelligence and innovation will be separated from the constraints of human-driven management. After this happens, humans will no longer be driving the innovation and there is possibility that SI for new intelligent systems will set its own targets and exclude humans. Over long time scale, these developments could result in a scenario, which will lead to the development of larger, cross galactic (universal) proactive SI and Intelligence.Keywords: artificial intelligence, DARPA, Moore’s law, proactive innovation, singularity, systems of innovation
Procedia PDF Downloads 47930862 Bioengineering System for Prediction and Early Prenosological Diagnostics of Stomach Diseases Based on Energy Characteristics of Bioactive Points with Fuzzy Logic
Authors: Mahdi Alshamasin, Riad Al-Kasasbeh, Nikolay Korenevskiy
Abstract:
We apply mathematical models for the interaction of the internal and biologically active points of meridian structures. Amongst the diseases for which reflex diagnostics are effective are those of the stomach disease. It is shown that use of fuzzy logic decision-making yields good results for the prediction and early diagnosis of gastrointestinal tract diseases, depending on the reaction energy of biologically active points (acupuncture points). It is shown that good results for the prediction and early diagnosis of diseases from the reaction energy of biologically active points (acupuncture points) are obtained by using fuzzy logic decision-making.Keywords: acupuncture points, fuzzy logic, diagnostically important points (DIP), confidence factors, membership functions, stomach diseases
Procedia PDF Downloads 46730861 MegaProjects and the Governing Processes That Lead to Success and Failure: A Literature Review
Authors: Fangwei Zhu, Wei Tian, Linzhuo Wang, Miao Yu
Abstract:
Megaproject has long been a critical issue in project governance, for its low success rate and large impact on society. Although the extant literature on megaproject governance is vast, to our best knowledge, the lacking of a thorough literature review makes it hard for us to gain a holistic view on current scenario of megaproject governance. The study conducts a systematic literature review process to analyze the existing literatures on megaproject governance. The finding indicates that mega project governance needs to be handled at network level and forming a network level governance provides a holistic framework for governing megaproject towards sustainable development of the projects. Theoretical and practical implications, as well as future studies and limitations, were discussed.Keywords: megaproject, governance, literature review, network
Procedia PDF Downloads 20030860 Web Service Architectural Style Selection in Multi-Criteria Requirements
Authors: Ahmad Mohsin, Syda Fatima, Falak Nawaz, Aman Ullah Khan
Abstract:
Selection of an appropriate architectural style is vital to the success of target web service under development. The nature of architecture design and selection for service-oriented computing applications is quite different as compared to traditional software. Web Services have complex and rigorous architectural styles to choose. Due to this, selection for accurate architectural style for web services development has become a more complex decision to be made by architects. Architectural style selection is a multi-criteria decision and demands lots of experience in service oriented computing. Decision support systems are good solutions to simplify the selection process of a particular architectural style. Our research suggests a new approach using DSS for selection of architectural styles while developing a web service to cater FRs and NFRs. Our proposed DSS helps architects to select right web service architectural pattern according to the domain and non-functional requirements. In this paper, a rule base DSS has been developed using CLIPS (C Language Integrated Production System) to support decisions using multi-criteria requirements. This DSS takes architectural characteristics, domain requirements and software architect preferences for NFRs as input for different architectural styles in use today in service-oriented computing. Weighted sum model has been applied to prioritize quality attributes and domain requirements. Scores are calculated using multiple criterions to choose the final architecture style.Keywords: software architecture, web-service, rule-based, DSS, multi-criteria requirements, quality attributes
Procedia PDF Downloads 36430859 Application of Artificial Ground-Freezing to Construct a Passenger Interchange Tunnel for the Subway Line 14 in Paris, France
Authors: G. Lancellotta, G. Di Salvo, A. Rigazio, A. Davout, V. Pastore, G. Tonoli, A. Martin, P. Jullien, R. Jagow-Klaff, R. Wernecke
Abstract:
Artificial ground freezing (AGF) technique is a well-proven soil improvement approach used worldwide to construct shafts, tunnels and many other civil structures in difficult subsoil or ambient conditions. As part of the extension of Line 14 of the Paris subway, a passenger interchange tunnel between the new station at Porte de CI ichy and the new Tribunal the Grand Instance has been successfully constructed using this technique. The paper presents the successful application of AGF by Liquid Nitrogen and Brine implemented to provide structural stability and groundwater cut-off around the passenger interchange tunnel. The working conditions were considered to be rather challenging, due to the proximity of a hundred-year-old existing service tunnel of the Line 13, and subsoil conditions on site. Laboratory tests were carried out to determine the relevant soil parameters for hydro-thermal-mechanical aspects and to implement numerical analyses. Monitoring data were used in order to check and control the development and the efficiency of the freezing process as well as to back analyze the parameters assumed for the design, both during the freezing and thawing phases.Keywords: artificial ground freezing, brine method, case history, liquid nitrogen
Procedia PDF Downloads 22530858 A New Multi-Target, Multi-Agent Search and Rescue Path Planning Approach
Authors: Jean Berger, Nassirou Lo, Martin Noel
Abstract:
Perfectly suited for natural or man-made emergency and disaster management situations such as flood, earthquakes, tornadoes, or tsunami, multi-target search path planning for a team of rescue agents is known to be computationally hard, and most techniques developed so far come short to successfully estimate optimality gap. A novel mixed-integer linear programming (MIP) formulation is proposed to optimally solve the multi-target multi-agent discrete search and rescue (SAR) path planning problem. Aimed at maximizing cumulative probability of successful target detection, it captures anticipated feedback information associated with possible observation outcomes resulting from projected path execution, while modeling agent discrete actions over all possible moving directions. Problem modeling further takes advantage of network representation to encompass decision variables, expedite compact constraint specification, and lead to substantial problem-solving speed-up. The proposed MIP approach uses CPLEX optimization machinery, efficiently computing near-optimal solutions for practical size problems, while giving a robust upper bound obtained from Lagrangean integrality constraint relaxation. Should eventually a target be positively detected during plan execution, a new problem instance would simply be reformulated from the current state, and then solved over the next decision cycle. A computational experiment shows the feasibility and the value of the proposed approach.Keywords: search path planning, search and rescue, multi-agent, mixed-integer linear programming, optimization
Procedia PDF Downloads 371