Search results for: software development models
24374 Transition from Linear to Circular Business Models with Service Design Methodology
Authors: Minna-Maari Harmaala, Hanna Harilainen
Abstract:
Estimates of the economic value of transitioning to circular economy models vary but it has been estimated to represent $1 trillion worth of new business into the global economy. In Europe alone, estimates claim that adopting circular-economy principles could not only have environmental and social benefits but also generate a net economic benefit of €1.8 trillion by 2030. Proponents of a circular economy argue that it offers a major opportunity to increase resource productivity, decrease resource dependence and waste, and increase employment and growth. A circular system could improve competitiveness and unleash innovation. Yet, most companies are not capturing these opportunities and thus the even abundant circular opportunities remain uncaptured even though they would seem inherently profitable. Service design in broad terms relates to developing an existing or a new service or service concept with emphasis and focus on the customer experience from the onset of the development process. Service design may even mean starting from scratch and co-creating the service concept entirely with the help of customer involvement. Service design methodologies provide a structured way of incorporating customer understanding and involvement in the process of designing better services with better resonance to customer needs. A business model is a depiction of how the company creates, delivers, and captures value; i.e. how it organizes its business. The process of business model development and adjustment or modification is also called business model innovation. Innovating business models has become a part of business strategy. Our hypothesis is that in addition to linear models still being easier to adopt and often with lower threshold costs, companies lack an understanding of how circular models can be adopted into their business and how customers will be willing and ready to adopt the new circular business models. In our research, we use robust service design methodology to develop circular economy solutions with two case study companies. The aim of the process is to not only develop the service concepts and portfolio, but to demonstrate the willingness to adopt circular solutions exists in the customer base. In addition to service design, we employ business model innovation methods to develop, test, and validate the new circular business models further. The results clearly indicate that amongst the customer groups there are specific customer personas that are willing to adopt and in fact are expecting the companies to take a leading role in the transition towards a circular economy. At the same time, there is a group of indifferents, to whom the idea of circularity provides no added value. In addition, the case studies clearly show what changes adoption of circular economy principles brings to the existing business model and how they can be integrated.Keywords: business model innovation, circular economy, circular economy business models, service design
Procedia PDF Downloads 13524373 Predicting Loss of Containment in Surface Pipeline using Computational Fluid Dynamics and Supervised Machine Learning Model to Improve Process Safety in Oil and Gas Operations
Authors: Muhammmad Riandhy Anindika Yudhy, Harry Patria, Ramadhani Santoso
Abstract:
Loss of containment is the primary hazard that process safety management is concerned within the oil and gas industry. Escalation to more serious consequences all begins with the loss of containment, starting with oil and gas release from leakage or spillage from primary containment resulting in pool fire, jet fire and even explosion when reacted with various ignition sources in the operations. Therefore, the heart of process safety management is avoiding loss of containment and mitigating its impact through the implementation of safeguards. The most effective safeguard for the case is an early detection system to alert Operations to take action prior to a potential case of loss of containment. The detection system value increases when applied to a long surface pipeline that is naturally difficult to monitor at all times and is exposed to multiple causes of loss of containment, from natural corrosion to illegal tapping. Based on prior researches and studies, detecting loss of containment accurately in the surface pipeline is difficult. The trade-off between cost-effectiveness and high accuracy has been the main issue when selecting the traditional detection method. The current best-performing method, Real-Time Transient Model (RTTM), requires analysis of closely positioned pressure, flow and temperature (PVT) points in the pipeline to be accurate. Having multiple adjacent PVT sensors along the pipeline is expensive, hence generally not a viable alternative from an economic standpoint.A conceptual approach to combine mathematical modeling using computational fluid dynamics and a supervised machine learning model has shown promising results to predict leakage in the pipeline. Mathematical modeling is used to generate simulation data where this data is used to train the leak detection and localization models. Mathematical models and simulation software have also been shown to provide comparable results with experimental data with very high levels of accuracy. While the supervised machine learning model requires a large training dataset for the development of accurate models, mathematical modeling has been shown to be able to generate the required datasets to justify the application of data analytics for the development of model-based leak detection systems for petroleum pipelines. This paper presents a review of key leak detection strategies for oil and gas pipelines, with a specific focus on crude oil applications, and presents the opportunities for the use of data analytics tools and mathematical modeling for the development of robust real-time leak detection and localization system for surface pipelines. A case study is also presented.Keywords: pipeline, leakage, detection, AI
Procedia PDF Downloads 19124372 The Promotion Effects for a Supply Chain System with a Dominant Retailer
Authors: Tai-Yue Wang, Yi-Ho Chen
Abstract:
In this study, we investigate a two-echelon supply chain with two suppliers and three retailers among which one retailer dominates other retailers. A price competition demand function is used to model this dominant retailer, which is leading market. The promotion strategies and negotiation schemes are integrated to form decision-making models under different scenarios. These models are then formulated into different mathematical programming models. The decision variables such as promotional costs, retailer prices, wholesale price, and order quantity are included in these models. At last, the distributions of promotion costs under different cost allocation strategies are discussed. Finally, an empirical example used to validate our models. The results from this empirical example show that the profit model will create the largest profit for the supply chain but with different profit-sharing results. At the same time, the more risk a member can take, the more profits are distributed to that member in the utility model.Keywords: supply chain, price promotion, mathematical models, dominant retailer
Procedia PDF Downloads 40024371 A Review on Modeling and Optimization of Integration of Renewable Energy Resources (RER) for Minimum Energy Cost, Minimum CO₂ Emissions and Sustainable Development, in Recent Years
Authors: M. M. Wagh, V. V. Kulkarni
Abstract:
The rising economic activities, growing population and improving living standards of world have led to a steady growth in its appetite for quality and quantity of energy services. As the economy expands the electricity demand is going to grow further, increasing the challenges of the more generation and stresses on the utility grids. Appropriate energy model will help in proper utilization of the locally available renewable energy sources such as solar, wind, biomass, small hydro etc. to integrate in the available grid, reducing the investments in energy infrastructure. Further to these new technologies like smart grids, decentralized energy planning, energy management practices, energy efficiency are emerging. In this paper, the attempt has been made to study and review the recent energy planning models, energy forecasting models, and renewable energy integration models. In addition, various modeling techniques and tools are reviewed and discussed.Keywords: energy modeling, integration of renewable energy, energy modeling tools, energy modeling techniques
Procedia PDF Downloads 34524370 An Analysis of Innovative Cloud Model as Bridging the Gap between Physical and Virtualized Business Environments: The Customer Perspective
Authors: Asim Majeed, Rehan Bhana, Mak Sharma, Rebecca Goode, Nizam Bolia, Mike Lloyd-Williams
Abstract:
This study aims to investigate and explore the underlying causes of security concerns of customers emerged when WHSmith transformed its physical system to virtualized business model through NetSuite. NetSuite is essentially fully integrated software which helps transforming the physical system to virtualized business model. Modern organisations are moving away from traditional business models to cloud based models and consequently it is expected to have a better, secure and innovative environment for customers. The vital issue of the modern age race is the security when transforming virtualized through cloud based models and designers of interactive systems often misunderstand privacy and even often ignore it, thus causing concerns for users. The content analysis approach is being used to collect the qualitative data from 120 online bloggers including TRUSTPILOT. The results and finding provide useful new insights into the nature and form of security concerns of online users after they have used the WHSmith services offered online through their website. Findings have theoretical as well as practical implications for the successful adoption of cloud computing Business-to-Business model and similar systems.Keywords: innovation, virtualization, cloud computing, organizational flexibility
Procedia PDF Downloads 38424369 Application of the Global Optimization Techniques to the Optical Thin Film Design
Authors: D. Li
Abstract:
Optical thin films are used in a wide variety of optical components and there are many software tools programmed for advancing multilayer thin film design. The available software packages for designing the thin film structure may not provide optimum designs. Normally, almost all current software programs obtain their final designs either from optimizing a starting guess or by technique, which may or may not involve a pseudorandom process, that give different answers every time, depending upon the initial conditions. With the increasing power of personal computers, functional methods in optimization and synthesis of optical multilayer systems have been developed such as DGL Optimization, Simulated Annealing, Genetic Algorithms, Needle Optimization, Inductive Optimization and Flip-Flop Optimization. Among these, DGL Optimization has proved its efficiency in optical thin film designs. The application of the DGL optimization technique to the design of optical coating is presented. A DGL optimization technique is provided, and its main features are discussed. Guidelines on the application of the DGL optimization technique to various types of design problems are given. The innovative global optimization strategies used in a software tool, OnlyFilm, to optimize multilayer thin film designs through different filter designs are outlined. OnlyFilm is a powerful, versatile, and user-friendly thin film software on the market, which combines optimization and synthesis design capabilities with powerful analytical tools for optical thin film designers. It is also the only thin film design software that offers a true global optimization function.Keywords: optical coatings, optimization, design software, thin film design
Procedia PDF Downloads 31624368 Copper Price Prediction Model for Various Economic Situations
Authors: Haidy S. Ghali, Engy Serag, A. Samer Ezeldin
Abstract:
Copper is an essential raw material used in the construction industry. During the year 2021 and the first half of 2022, the global market suffered from a significant fluctuation in copper raw material prices due to the aftermath of both the COVID-19 pandemic and the Russia-Ukraine war, which exposed its consumers to an unexpected financial risk. Thereto, this paper aims to develop two ANN-LSTM price prediction models, using Python, that can forecast the average monthly copper prices traded in the London Metal Exchange; the first model is a multivariate model that forecasts the copper price of the next 1-month and the second is a univariate model that predicts the copper prices of the upcoming three months. Historical data of average monthly London Metal Exchange copper prices are collected from January 2009 till July 2022, and potential external factors are identified and employed in the multivariate model. These factors lie under three main categories: energy prices and economic indicators of the three major exporting countries of copper, depending on the data availability. Before developing the LSTM models, the collected external parameters are analyzed with respect to the copper prices using correlation and multicollinearity tests in R software; then, the parameters are further screened to select the parameters that influence the copper prices. Then, the two LSTM models are developed, and the dataset is divided into training, validation, and testing sets. The results show that the performance of the 3-Month prediction model is better than the 1-Month prediction model, but still, both models can act as predicting tools for diverse economic situations.Keywords: copper prices, prediction model, neural network, time series forecasting
Procedia PDF Downloads 11324367 Coupling Large Language Models with Disaster Knowledge Graphs for Intelligent Construction
Authors: Zhengrong Wu, Haibo Yang
Abstract:
In the context of escalating global climate change and environmental degradation, the complexity and frequency of natural disasters are continually increasing. Confronted with an abundance of information regarding natural disasters, traditional knowledge graph construction methods, which heavily rely on grammatical rules and prior knowledge, demonstrate suboptimal performance in processing complex, multi-source disaster information. This study, drawing upon past natural disaster reports, disaster-related literature in both English and Chinese, and data from various disaster monitoring stations, constructs question-answer templates based on large language models. Utilizing the P-Tune method, the ChatGLM2-6B model is fine-tuned, leading to the development of a disaster knowledge graph based on large language models. This serves as a knowledge database support for disaster emergency response.Keywords: large language model, knowledge graph, disaster, deep learning
Procedia PDF Downloads 5624366 Copula-Based Estimation of Direct and Indirect Effects in Path Analysis Models
Authors: Alam Ali, Ashok Kumar Pathak
Abstract:
Path analysis is a statistical technique used to evaluate the direct and indirect effects of variables in path models. One or more structural regression equations are used to estimate a series of parameters in path models to find the better fit of data. However, sometimes the assumptions of classical regression models, such as ordinary least squares (OLS), are violated by the nature of the data, resulting in insignificant direct and indirect effects of exogenous variables. This article aims to explore the effectiveness of a copula-based regression approach as an alternative to classical regression, specifically when variables are linked through an elliptical copula.Keywords: path analysis, copula-based regression models, direct and indirect effects, k-fold cross validation technique
Procedia PDF Downloads 4124365 Numerical Modeling of Waves and Currents by Using a Hydro-Sedimentary Model
Authors: Mustapha Kamel Mihoubi, Hocine Dahmani
Abstract:
Over recent years much progress has been achieved in the fields of numerical modeling shoreline processes: waves, currents, waves and current. However, there are still some problems in the existing models to link the on the first, the hydrodynamics of waves and currents and secondly, the sediment transport processes and due to the variability in time, space and interaction and the simultaneous action of wave-current near the shore. This paper is the establishment of a numerical modeling to forecast the sediment transport from development scenarios of harbor structure. It is established on the basis of a numerical simulation of a water-sediment model via a 2D model using a set of codes calculation MIKE 21-DHI software. This is to examine the effect of the sediment transport drivers following the dominant incident wave in the direction to pass input harbor work under different variants planning studies to find the technical and economic limitations to the sediment transport and protection of the harbor structure optimum solution.Keywords: swell, current, radiation, stress, mesh, mike21, sediment
Procedia PDF Downloads 46924364 Drugstore Control System Design and Realization Based on Programmable Logic Controller (PLC)
Authors: Muhammad Faheem Khakhi, Jian Yu Wang, Salman Muhammad, Muhammad Faisal Shabir
Abstract:
Population growth and Chinese two-child policy will boost pharmaceutical market, and it will continue to maintain the growth for a period of time in the future, the traditional pharmacy dispensary has been unable to meet the growing medical needs of the peoples. Under the strong support of the national policy, the automatic transformation of traditional pharmacies is the inclination of the Times, the new type of intelligent pharmacy system will continue to promote the development of the pharmaceutical industry. Under this background, based on PLC control, the paper proposed an intelligent storage and automatic drug delivery system; complete design of the lower computer's control system and the host computer's software system has been present. The system can be applied to dispensing work for Chinese herbal medicinal and Western medicines. Firstly, the essential of intelligent control system for pharmacy is discussed. After the analysis of the requirements, the overall scheme of the system design is presented. Secondly, introduces the software and hardware design of the lower computer's control system, including the selection of PLC and the selection of motion control system, the problem of the human-computer interaction module and the communication between PC and PLC solves, the program design and development of the PLC control system is completed. The design of the upper computer software management system is described in detail. By analyzing of E-R diagram, built the establish data, the communication protocol between systems is customize, C++ Builder is adopted to realize interface module, supply module, main control module, etc. The paper also gives the implementations of the multi-threaded system and communication method. Lastly, each module of the lower computer control system is tested. Then, after building a test environment, the function test of the upper computer software management system is completed. On this basis, the entire control system accepts the overall test.Keywords: automatic pharmacy, PLC, control system, management system, communication
Procedia PDF Downloads 31024363 Performance Evaluation of Using Genetic Programming Based Surrogate Models for Approximating Simulation Complex Geochemical Transport Processes
Authors: Hamed K. Esfahani, Bithin Datta
Abstract:
Transport of reactive chemical contaminant species in groundwater aquifers is a complex and highly non-linear physical and geochemical process especially for real life scenarios. Simulating this transport process involves solving complex nonlinear equations and generally requires huge computational time for a given aquifer study area. Development of optimal remediation strategies in aquifers may require repeated solution of such complex numerical simulation models. To overcome this computational limitation and improve the computational feasibility of large number of repeated simulations, Genetic Programming based trained surrogate models are developed to approximately simulate such complex transport processes. Transport process of acid mine drainage, a hazardous pollutant is first simulated using a numerical simulated model: HYDROGEOCHEM 5.0 for a contaminated aquifer in a historic mine site. Simulation model solution results for an illustrative contaminated aquifer site is then approximated by training and testing a Genetic Programming (GP) based surrogate model. Performance evaluation of the ensemble GP models as surrogate models for the reactive species transport in groundwater demonstrates the feasibility of its use and the associated computational advantages. The results show the efficiency and feasibility of using ensemble GP surrogate models as approximate simulators of complex hydrogeologic and geochemical processes in a contaminated groundwater aquifer incorporating uncertainties in historic mine site.Keywords: geochemical transport simulation, acid mine drainage, surrogate models, ensemble genetic programming, contaminated aquifers, mine sites
Procedia PDF Downloads 27624362 Evaluation of Energy Upgrade Measures and Connection of Renewable Energy Sources Using Software Tools: Case Study of an Academic Library Building in Larissa, Greece
Authors: Giwrgos S. Gkarmpounis, Aikaterini G. Rokkou, Marios N. Moschakis
Abstract:
Increased energy consumption in the academic buildings, creates the need to implement energy saving measures and to take advantage of the renewable energy sources to cover the electrical needs of those buildings. An Academic Library will be used as a case study. With the aid of RETScreen software that takes into account the energy consumptions and characteristics of the Library Building, it is proved that measures such as the replacement of fluorescent lights with led lights, the installation of outdoor shading, the replacement of the openings and Building Management System installation, provide a high level of energy savings. Moreover, given the available space of the building and the climatic data, the installation of a photovoltaic system of 100 kW can also cover a serious amount of the building energy consumption, unlike a wind system that seems uncompromising. Lastly, HOMER software is used to compare the use of a photovoltaic system against a wind system in order to verify the results that came up from the RETScreen software concerning the renewable energy sources.Keywords: building sector, energy saving measures, energy upgrading, homer software, renewable energy sources, RETScreen software
Procedia PDF Downloads 22924361 Factors of Influence in Software Process Improvement: An ISO/IEC 29110 for Very-Small Entities
Authors: N. Wongsai, R. Wetprasit, V. Siddoo
Abstract:
The recently introduced ISO/IEC 29110 standard Lifecycle profile for Very Small Entities (VSE) has been adopted and practiced in many small and medium software companies, including in Thailand’s software industry. Many Thai companies complete their software process improvement (SPI) initiative program and have been certified. There are, however, a number of participants fail to success. This study was concerned with the factors that influence the accomplishment of the standard implementation in various VSE characteristics. In order to achieve this goal, exploring and extracting critical factors from prior studies were carried out and then the obtained factors were validated by the standard experts. Data analysis of comments and recommendations was performed using a qualitative content analysis method. This paper presents the initial set of influence factors in both positive and negative impact the ISO/IEC 29110 implementation with an aim at helping such SPI practitioners with some considerations to manage appropriate adoption approach in order to achieve its implementation.Keywords: barriers, critical success factors, ISO/IEC 29110, Software Process Improvement, SPI, Very-Small Entity, VSE
Procedia PDF Downloads 31524360 Investigating the Relationship between Place Attachment and Sustainable Development of Urban Spaces
Authors: Hamid Reza Zeraatpisheh, Ali Akbar Heidari, Soleiman Mohammadi Doust
Abstract:
This study has examined the relationship between place attachment and sustainable development of urban spaces. To perform this, the components of place identity, emotional attachment, place attachment and social bonding which totally constitute the output of place attachment, by means of the standardized questionnaire measure place attachment in three domains of (cognitive) the place identity, (affective) emotional attachment and (behavioral) place attachment and social bonding. To measure sustainable development, three components of sustainable development, including society, economy and environment has been considered. The study is descriptive. The assessment instrument is the standard questionnaire of Safarnia which has been used to measure the variable of place attachment and to measure the variable of sustainable development, a questionnaire has been made by the researcher and been based on the combined theoretical framework. The statistical population of this research has been the city of Shiraz. The statistical sample has been Hafeziyeh. SPSS software has been used to analyze the data and examined the results of both descriptive and inferential statistics. In inferential statistics, Pearson correlation coefficient has been used to examine the hypotheses. In this study, the variable of place attachment is high and sustainable development is also in a high level. These results suggest a positive relationship between attachment to place and sustainable development.Keywords: place attachment, sustainable development, economy-society-environment, Hafez's tomb
Procedia PDF Downloads 70124359 Analysis of Diabetes Patients Using Pearson, Cost Optimization, Control Chart Methods
Authors: Devatha Kalyan Kumar, R. Poovarasan
Abstract:
In this paper, we have taken certain important factors and health parameters of diabetes patients especially among children by birth (pediatric congenital) where using the above three metrics methods we are going to assess the importance of each attributes in the dataset and thereby determining the most highly responsible and co-related attribute causing diabetics among young patients. We use cost optimization, control chart and Spearmen methodologies for the real-time application of finding the data efficiency in this diabetes dataset. The Spearmen methodology is the correlation methodologies used in software development process to identify the complexity between the various modules of the software. Identifying the complexity is important because if the complexity is higher, then there is a higher chance of occurrence of the risk in the software. With the use of control; chart mean, variance and standard deviation of data are calculated. With the use of Cost optimization model, we find to optimize the variables. Hence we choose the Spearmen, control chart and cost optimization methods to assess the data efficiency in diabetes datasets.Keywords: correlation, congenital diabetics, linear relationship, monotonic function, ranking samples, pediatric
Procedia PDF Downloads 25624358 A Collaborative Platform for Multilingual Ontology Development
Authors: Ahmed Tawfik, Fausto Giunchiglia, Vincenzo Maltese
Abstract:
Ontologies provide a common understanding of a specific domain of interest that can be communicated between people and used as background knowledge for automated reasoning in a wide range of applications. In this paper we address the design of multilingual ontologies following well-defined knowledge engineering methodologies with the support of novel collaborative development approaches. In particular, we present a collaborative platform which allows ontologies to be developed incrementally in multiple languages. This is made possible via an appropriate mapping between language independent concepts and one lexicalization per language (or a lexical gap in case such lexicalization does not exist). The collaborative platform has been designed to support the development of the Universal Knowledge Core, a multilingual ontology currently in English, Italian, Chinese, Mongolian, Hindi, and Bangladeshi. Its design follows a workflow-based development methodology that models resources as a set of collaborative objects and assigns customizable workflows to build and maintain each collaborative object in a community driven manner, with extensive support of modern web 2.0 social and collaborative features.Keywords: knowledge diversity, knowledge representation, ontology, development
Procedia PDF Downloads 39224357 Optimal Cropping Pattern in an Irrigation Project: A Hybrid Model of Artificial Neural Network and Modified Simplex Algorithm
Authors: Safayat Ali Shaikh
Abstract:
Software has been developed for optimal cropping pattern in an irrigation project considering land constraint, water availability constraint and pick up flow constraint using modified Simplex Algorithm. Artificial Neural Network Models (ANN) have been developed to predict rainfall. AR (1) model used to generate 1000 years rainfall data to train the ANN. Simulation has been done with expected rainfall data. Eight number crops and three types of soil class have been considered for optimization model. Area under each crop and each soil class have been quantified using Modified Simplex Algorithm to get optimum net return. Efficacy of the software has been tested using data of large irrigation project in India.Keywords: artificial neural network, large irrigation project, modified simplex algorithm, optimal cropping pattern
Procedia PDF Downloads 20324356 Predicting Success and Failure in Drug Development Using Text Analysis
Authors: Zhi Hao Chow, Cian Mulligan, Jack Walsh, Antonio Garzon Vico, Dimitar Krastev
Abstract:
Drug development is resource-intensive, time-consuming, and increasingly expensive with each developmental stage. The success rates of drug development are also relatively low, and the resources committed are wasted with each failed candidate. As such, a reliable method of predicting the success of drug development is in demand. The hypothesis was that some examples of failed drug candidates are pushed through developmental pipelines based on false confidence and may possess common linguistic features identifiable through sentiment analysis. Here, the concept of using text analysis to discover such features in research publications and investor reports as predictors of success was explored. R studios were used to perform text mining and lexicon-based sentiment analysis to identify affective phrases and determine their frequency in each document, then using SPSS to determine the relationship between our defined variables and the accuracy of predicting outcomes. A total of 161 publications were collected and categorised into 4 groups: (i) Cancer treatment, (ii) Neurodegenerative disease treatment, (iii) Vaccines, and (iv) Others (containing all other drugs that do not fit into the 3 categories). Text analysis was then performed on each document using 2 separate datasets (BING and AFINN) in R within the category of drugs to determine the frequency of positive or negative phrases in each document. A relative positivity and negativity value were then calculated by dividing the frequency of phrases with the word count of each document. Regression analysis was then performed with SPSS statistical software on each dataset (values from using BING or AFINN dataset during text analysis) using a random selection of 61 documents to construct a model. The remaining documents were then used to determine the predictive power of the models. Model constructed from BING predicts the outcome of drug performance in clinical trials with an overall percentage of 65.3%. AFINN model had a lower accuracy at predicting outcomes compared to the BING model at 62.5% but was not effective at predicting the failure of drugs in clinical trials. Overall, the study did not show significant efficacy of the model at predicting outcomes of drugs in development. Many improvements may need to be made to later iterations of the model to sufficiently increase the accuracy.Keywords: data analysis, drug development, sentiment analysis, text-mining
Procedia PDF Downloads 15724355 Application of Axiomatic Design in Industrial Control and Automation Software
Authors: Aydin Homay, Mario de Sousa, Martin Wollschlaeger
Abstract:
Axiomatic design is a system design methodology that systematically analyses the transformation of customer needs into functional requirements, design parameters, and process variables. This approach aims to create high-quality product or system designs by adhering to specific design principles or axioms, namely, the independence and information axiom. The application of axiomatic design in the design of industrial control and automation software systems could be challenging due to the high flexibility exposed by the software system and the coupling enforced by the hardware part. This paper aims to present how to use axiomatic design for designing industrial control and automation software systems and how to satisfy the independence axiom within these tightly coupled systems.Keywords: axiomatic design, decoupling, uncoupling, automation
Procedia PDF Downloads 5024354 Strategic Tools for Entrepreneurship: Model Proposal for Manufacturing Companies
Authors: Chiara Mansanta, Daniela Sani
Abstract:
The present paper presents the further development of the application of a standard methodology to boost innovation inside real case studies of manufacturing companies. The proposed methodology provides a viable solution for manufacturing companies that have to evaluate new business ideas. The study underlined the concept of entrepreneurship and how a manager can use it to promote innovation inside their companies. Starting from a literature study on entrepreneurship, this paper examines the role of the manager in supporting a company’s development. The empirical part of the study is based on two manufacturing companies that used the proposed methodology to favour entrepreneurship through an alternative approach. The research demonstrated the need for companies to have a structured and well-defined methodology to achieve their goals. The purpose of this article is to understand the significance of business models inside companies and explore how they affect business strategy and innovation management. The idea is to use business models to support entrepreneurs in their decision-making processes, reducing risks and avoiding errors.Keywords: entrepreneurship, manufacturing companies, solution validation, strategic management
Procedia PDF Downloads 9524353 Business Process Mashup
Authors: Fethia Zenak, Salima Benbernou, Linda Zaoui
Abstract:
Recently, many companies are based on process development from scratch to achieve their business goals. The process development is not trivial and the main objective of enterprise managing processes is to decrease the software development time. Several concepts have been proposed in the field of business process-based reused development, known as BP Mashup. This concept consists of reusing existing business processes which have been modeled in order to respond to a particular goal. To meet user process requirements, our contribution is to mix parts of processes as 'processes fragments' components to build a new process (i.e. process mashup). The main idea of our paper is to offer graphical framework tool for both creating and running processes mashup. Allow users to perform a mixture of fragments, using a simple interface with set of graphical mixture operators based on a proposed formal model. A process mashup and mixture behavior are described within a new specification of a high-level language, language for process mashup (BPML).Keywords: business process, mashup, fragments, bp mashup
Procedia PDF Downloads 63524352 Analysis on Prediction Models of TBM Performance and Selection of Optimal Input Parameters
Authors: Hang Lo Lee, Ki Il Song, Hee Hwan Ryu
Abstract:
An accurate prediction of TBM(Tunnel Boring Machine) performance is very difficult for reliable estimation of the construction period and cost in preconstruction stage. For this purpose, the aim of this study is to analyze the evaluation process of various prediction models published since 2000 for TBM performance, and to select the optimal input parameters for the prediction model. A classification system of TBM performance prediction model and applied methodology are proposed in this research. Input and output parameters applied for prediction models are also represented. Based on these results, a statistical analysis is performed using the collected data from shield TBM tunnel in South Korea. By performing a simple regression and residual analysis utilizinFg statistical program, R, the optimal input parameters are selected. These results are expected to be used for development of prediction model of TBM performance.Keywords: TBM performance prediction model, classification system, simple regression analysis, residual analysis, optimal input parameters
Procedia PDF Downloads 30924351 Characterizing the Rectification Process for Designing Scoliosis Braces: Towards Digital Brace Design
Authors: Inigo Sanz-Pena, Shanika Arachchi, Dilani Dhammika, Sanjaya Mallikarachchi, Jeewantha S. Bandula, Alison H. McGregor, Nicolas Newell
Abstract:
The use of orthotic braces for adolescent idiopathic scoliosis (AIS) patients is the most common non-surgical treatment to prevent deformity progression. The traditional method to create an orthotic brace involves casting the patient’s torso to obtain a representative geometry, which is then rectified by an orthotist to the desired geometry of the brace. Recent improvements in 3D scanning technologies, rectification software, CNC, and additive manufacturing processes have given the possibility to compliment, or in some cases, replace manual methods with digital approaches. However, the rectification process remains dependent on the orthotist’s skills. Therefore, the rectification process needs to be carefully characterized to ensure that braces designed through a digital workflow are as efficient as those created using a manual process. The aim of this study is to compare 3D scans of patients with AIS against 3D scans of both pre- and post-rectified casts that have been manually shaped by an orthotist. Six AIS patients were recruited from the Ragama Rehabilitation Clinic, Colombo, Sri Lanka. All patients were between 10 and 15 years old, were skeletally immature (Risser grade 0-3), and had Cobb angles between 20-45°. Seven spherical markers were placed at key anatomical locations on each patient’s torso and on the pre- and post-rectified molds so that distances could be reliably measured. 3D scans were obtained of 1) the patient’s torso and pelvis, 2) the patient’s pre-rectification plaster mold, and 3) the patient’s post-rectification plaster mold using a Structure Sensor Mark II 3D scanner (Occipital Inc., USA). 3D stick body models were created for each scan to represent the distances between anatomical landmarks. The 3D stick models were used to analyze the changes in position and orientation of the anatomical landmarks between scans using Blender open-source software. 3D Surface deviation maps represented volume differences between the scans using CloudCompare open-source software. The 3D stick body models showed changes in the position and orientation of thorax anatomical landmarks between the patient and the post-rectification scans for all patients. Anatomical landmark position and volume differences were seen between 3D scans of the patient’s torsos and the pre-rectified molds. Between the pre- and post-rectified molds, material removal was consistently seen on the anterior side of the thorax and the lateral areas below the ribcage. Volume differences were seen in areas where the orthotist planned to place pressure pads (usually at the trochanter on the side to which the lumbar curve was tilted (trochanter pad), at the lumbar apical vertebra (lumbar pad), on the rib connected to the apical vertebrae at the mid-axillary line (thoracic pad), and on the ribs corresponding to the upper thoracic vertebra (axillary extension pad)). The rectification process requires the skill and experience of an orthotist; however, this study demonstrates that the brace shape, location, and volume of material removed from the pre-rectification mold can be characterized and quantified. Results from this study can be fed into software that can accelerate the brace design process and make steps towards the automated digital rectification process.Keywords: additive manufacturing, orthotics, scoliosis brace design, sculpting software, spinal deformity
Procedia PDF Downloads 14524350 The Use of Thermal Infrared Wavelengths to Determine the Volcanic Soils
Authors: Levent Basayigit, Mert Dedeoglu, Fadime Ozogul
Abstract:
In this study, an application was carried out to determine the Volcanic Soils by using remote sensing. The study area was located on the Golcuk formation in Isparta-Turkey. The thermal bands of Landsat 7 image were used for processing. The implementation of the climate model that was based on the water index was used in ERDAS Imagine software together with pixel based image classification. Soil Moisture Index (SMI) was modeled by using the surface temperature (Ts) which was obtained from thermal bands and vegetation index (NDVI) derived from Landsat 7. Surface moisture values were grouped and classified by using scoring system. Thematic layers were compared together with the field studies. Consequently, different moisture levels for volcanic soils were indicator for determination and separation. Those thermal wavelengths are preferable bands for separation of volcanic soils using moisture and temperature models.Keywords: Landsat 7, soil moisture index, temperature models, volcanic soils
Procedia PDF Downloads 30524349 Piping Fragility Composed of Different Materials by Using OpenSees Software
Authors: Woo Young Jung, Min Ho Kwon, Bu Seog Ju
Abstract:
A failure of the non-structural component can cause significant damages in critical facilities such as nuclear power plants and hospitals. Historically, it was reported that the damage from the leakage of sprinkler systems, resulted in the shutdown of hospitals for several weeks by the 1971 San Fernando and 1994 North Ridge earthquakes. In most cases, water leakages were observed at the cross joints, sprinkler heads, and T-joint connections in piping systems during and after the seismic events. Hence, the primary objective of this study was to understand the seismic performance of T-joint connections and to develop an analytical Finite Element (FE) model for the T-joint systems of 2-inch fire protection piping system in hospitals subjected to seismic ground motions. In order to evaluate the FE models of the piping systems using OpenSees, two types of materials were used: 1) Steel 02 materials and 2) Pinching 4 materials. Results of the current study revealed that the nonlinear moment-rotation FE models for the threaded T-joint reconciled well with the experimental results in both FE material models. However, the system-level fragility determined from multiple nonlinear time history analyses at the threaded T-joint was slightly different. The system-level fragility at the T-joint, determined by Pinching 4 material was more conservative than that of using Steel 02 material in the piping system.Keywords: fragility, t-joint, piping, leakage, sprinkler
Procedia PDF Downloads 30324348 Data-Centric Anomaly Detection with Diffusion Models
Authors: Sheldon Liu, Gordon Wang, Lei Liu, Xuefeng Liu
Abstract:
Anomaly detection, also referred to as one-class classification, plays a crucial role in identifying product images that deviate from the expected distribution. This study introduces Data-centric Anomaly Detection with Diffusion Models (DCADDM), presenting a systematic strategy for data collection and further diversifying the data with image generation via diffusion models. The algorithm addresses data collection challenges in real-world scenarios and points toward data augmentation with the integration of generative AI capabilities. The paper explores the generation of normal images using diffusion models. The experiments demonstrate that with 30% of the original normal image size, modeling in an unsupervised setting with state-of-the-art approaches can achieve equivalent performances. With the addition of generated images via diffusion models (10% equivalence of the original dataset size), the proposed algorithm achieves better or equivalent anomaly localization performance.Keywords: diffusion models, anomaly detection, data-centric, generative AI
Procedia PDF Downloads 8224347 National System of Innovation in Zambia: Towards Socioeconomic Development
Authors: Ephraim Daka, Maxim Kotsemir
Abstract:
The National system Innovation (NSI) have recently proliferated as a vehicle for addressing poverty and national competitiveness in the developing countries. While several governments in Sub-Saharan Africa have adopted the developed countries’ models of innovation to local conditions, the Zambian case is rather unique. This study highlights conceptual and socioeconomic challenges directed to the performances of the NSI. The paper analyses science and technology strategies with the inclusion of “innovation” and its effect towards improving socioeconomic elements. The authors reviewed STI policy and national strategy documents, followed by interviews compared to economical regional and national data sets. The NSI and its related to inter-linkages and support mechanism to socioeconomic development were explored.Keywords: national system of innovation, socioeconomics, development, Zambia
Procedia PDF Downloads 22324346 Open Source Software in Higher Education: Oman SQU Case Study
Authors: Amal S. Al-Badi, Ali H. Al-Badi
Abstract:
Many organizations are opting to adopt Open Source Software (OSS) as it is the current trend to rely on each other rather than on companies (Software vendors). It is a clear shift from organizations to individuals, the concept being to rely on collective participation rather than companies/vendors. The main objectives of this research are 1) to identify the current level of OSS usage in Sultan Qaboos University; 2) to identify the potential benefits of using OSS in educational institutes; 3) to identify the OSS applications that are most likely to be used within an educational institute; 4) to identify the existing and potential barriers to the successful adoption of OSS in education. To achieve these objectives a two-stage research method was conducted. First a rigorous literature review of previously published material was performed (interpretive/descriptive approach), and then a set of interviews were conducted with the IT professionals at Sultan Qaboos University in Oman in order to explore the extent and nature of their usage of OSS.Keywords: open source software, social software, e-learning 2.0, Web 2.0, connectivism, personal learning environment (PLE), OpenCourseWare
Procedia PDF Downloads 30524345 Copula Markov Switching Multifractal Models for Forecasting Value-at-Risk
Authors: Giriraj Achari, Malay Bhattacharyya
Abstract:
In this paper, the effectiveness of Copula Markov Switching Multifractal (MSM) models at forecasting Value-at-Risk of a two-stock portfolio is studied. The innovations are allowed to be drawn from distributions that can capture skewness and leptokurtosis, which are well documented empirical characteristics observed in financial returns. The candidate distributions considered for this purpose are Johnson-SU, Pearson Type-IV and α-Stable distributions. The two univariate marginal distributions are combined using the Student-t copula. The estimation of all parameters is performed by Maximum Likelihood Estimation. Finally, the models are compared in terms of accurate Value-at-Risk (VaR) forecasts using tests of unconditional coverage and independence. It is found that Copula-MSM-models with leptokurtic innovation distributions perform slightly better than Copula-MSM model with Normal innovations. Copula-MSM models, in general, produce better VaR forecasts as compared to traditional methods like Historical Simulation method, Variance-Covariance approach and Copula-Generalized Autoregressive Conditional Heteroscedasticity (Copula-GARCH) models.Keywords: Copula, Markov Switching, multifractal, value-at-risk
Procedia PDF Downloads 165