Search results for: support vector machine (SVM)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9767

Search results for: support vector machine (SVM)

8867 Automatic Aggregation and Embedding of Microservices for Optimized Deployments

Authors: Pablo Chico De Guzman, Cesar Sanchez

Abstract:

Microservices are a software development methodology in which applications are built by composing a set of independently deploy-able, small, modular services. Each service runs a unique process and it gets instantiated and deployed in one or more machines (we assume that different microservices are deployed into different machines). Microservices are becoming the de facto standard for developing distributed cloud applications due to their reduced release cycles. In principle, the responsibility of a microservice can be as simple as implementing a single function, which can lead to the following issues: - Resource fragmentation due to the virtual machine boundary. - Poor communication performance between microservices. Two composition techniques can be used to optimize resource fragmentation and communication performance: aggregation and embedding of microservices. Aggregation allows the deployment of a set of microservices on the same machine using a proxy server. Aggregation helps to reduce resource fragmentation, and is particularly useful when the aggregated services have a similar scalability behavior. Embedding deals with communication performance by deploying on the same virtual machine those microservices that require a communication channel (localhost bandwidth is reported to be about 40 times faster than cloud vendor local networks and it offers better reliability). Embedding can also reduce dependencies on load balancer services since the communication takes place on a single virtual machine. For example, assume that microservice A has two instances, a1 and a2, and it communicates with microservice B, which also has two instances, b1 and b2. One embedding can deploy a1 and b1 on machine m1, and a2 and b2 are deployed on a different machine m2. This deployment configuration allows each pair (a1-b1), (a2-b2) to communicate using the localhost interface without the need of a load balancer between microservices A and B. Aggregation and embedding techniques are complex since different microservices might have incompatible runtime dependencies which forbid them from being installed on the same machine. There is also a security concern since the attack surface between microservices can be larger. Luckily, container technology allows to run several processes on the same machine in an isolated manner, solving the incompatibility of running dependencies and the previous security concern, thus greatly simplifying aggregation/embedding implementations by just deploying a microservice container on the same machine as the aggregated/embedded microservice container. Therefore, a wide variety of deployment configurations can be described by combining aggregation and embedding to create an efficient and robust microservice architecture. This paper presents a formal method that receives a declarative definition of a microservice architecture and proposes different optimized deployment configurations by aggregating/embedding microservices. The first prototype is based on i2kit, a deployment tool also submitted to ICWS 2018. The proposed prototype optimizes the following parameters: network/system performance, resource usage, resource costs and failure tolerance.

Keywords: aggregation, deployment, embedding, resource allocation

Procedia PDF Downloads 192
8866 Business Intelligent to a Decision Support Tool for Green Entrepreneurship: Meso and Macro Regions

Authors: Anishur Rahman, Maria Areias, Diogo Simões, Ana Figeuiredo, Filipa Figueiredo, João Nunes

Abstract:

The circular economy (CE) has gained increased awareness among academics, businesses, and decision-makers as it stimulates resource circularity in the production and consumption systems. A large epistemological study has explored the principles of CE, but scant attention eagerly focused on analysing how CE is evaluated, consented to, and enforced using economic metabolism data and business intelligent framework. Economic metabolism involves the ongoing exchange of materials and energy within and across socio-economic systems and requires the assessment of vast amounts of data to provide quantitative analysis related to effective resource management. Limited concern, the present work has focused on the regional flows pilot region from Portugal. By addressing this gap, this study aims to promote eco-innovation and sustainability in the regions of Intermunicipal Communities Região de Coimbra, Viseu Dão Lafões and Beiras e Serra da Estrela, using this data to find precise synergies in terms of material flows and give companies a competitive advantage in form of valuable waste destinations, access to new resources and new markets, cost reduction and risk sharing benefits. In our work, emphasis on applying artificial intelligence (AI) and, more specifically, on implementing state-of-the-art deep learning algorithms is placed, contributing to construction a business intelligent approach. With the emergence of new approaches generally highlighted under the sub-heading of AI and machine learning (ML), the methods for statistical analysis of complex and uncertain production systems are facing significant changes. Therefore, various definitions of AI and its differences from traditional statistics are presented, and furthermore, ML is introduced to identify its place in data science and the differences in topics such as big data analytics and in production problems that using AI and ML are identified. A lifecycle-based approach is then taken to analyse the use of different methods in each phase to identify the most useful technologies and unifying attributes of AI in manufacturing. Most of macroeconomic metabolisms models are mainly direct to contexts of large metropolis, neglecting rural territories, so within this project, a dynamic decision support model coupled with artificial intelligence tools and information platforms will be developed, focused on the reality of these transition zones between the rural and urban. Thus, a real decision support tool is under development, which will surpass the scientific developments carried out to date and will allow to overcome imitations related to the availability and reliability of data.

Keywords: circular economy, artificial intelligence, economic metabolisms, machine learning

Procedia PDF Downloads 58
8865 Grating Scale Thermal Expansion Error Compensation for Large Machine Tools Based on Multiple Temperature Detection

Authors: Wenlong Feng, Zhenchun Du, Jianguo Yang

Abstract:

To decrease the grating scale thermal expansion error, a novel method which based on multiple temperature detections is proposed. Several temperature sensors are installed on the grating scale and the temperatures of these sensors are recorded. The temperatures of every point on the grating scale are calculated by interpolating between adjacent sensors. According to the thermal expansion principle, the grating scale thermal expansion error model can be established by doing the integral for the variations of position and temperature. A novel compensation method is proposed in this paper. By applying the established error model, the grating scale thermal expansion error is decreased by 90% compared with no compensation. The residual positioning error of the grating scale is less than 15um/10m and the accuracy of the machine tool is significant improved.

Keywords: thermal expansion error of grating scale, error compensation, machine tools, integral method

Procedia PDF Downloads 354
8864 Developing a Rational Database Management System (RDBMS) Supporting Product Life Cycle Appications

Authors: Yusri Yusof, Chen Wong Keong

Abstract:

This paper presents the implementation details of a Relational Database Management System of a STEP-technology product model repository. It is able support the implementation of any EXPRESS language schema, although it has been primarily implemented to support mechanical product life cycle applications. This database support the input of STEP part 21 file format from CAD in geometrical and topological data format and support a range of queries for mechanical product life cycle applications. This proposed relational database management system uses entity-to-table method (R1) rather than type-to-table method (R4). The two mapping methods have their own strengths and drawbacks.

Keywords: RDBMS, CAD, ISO 10303, part-21 file

Procedia PDF Downloads 525
8863 Aggregate Supply Response of Some Livestock Commodities in Algeria: Cointegration- Vector Error Correction Model Approach

Authors: Amine M. Benmehaia, Amine Oulmane

Abstract:

The supply response of agricultural commodities to changes in price incentives is an important issue for the success of any policy reform in the agricultural sector. This study aims to quantify the responsiveness of producers of some livestock commodities to price incentives in Algerian context. Time series analysis is used on annual data for a period of 52 years (1966-2018). Both co-integration and vector error correction model (VECM) are used through the Nerlove model of partial adjustment. The study attempts to determine the long-run and short-run relationships along with the magnitudes of disequilibria in the selected commodities. Results show that the short-run price elasticities are low in cow and sheep meat sectors (8.7 and 8% respectively), while their respective long-run elasticities are 16.5 and 10.5, whereas eggs and milk have very high short-run price elasticities (82 and 90% respectively) with long-run elasticities of 40 and 46 respectively. The error correction coefficient, reflecting the speed of adjustment towards the long-run equilibrium, is statistically significant and have the expected negative sign. Its estimates are 12.7 for cow meat, 33.5 for sheep meat, 46.7 for eggs and 8.4 for milk. It seems that cow meat and milk producers have a weak feedback of about 12.7% and 8.4% respectively of the previous year's disequilibrium from the long-run price elasticity, whereas sheep meat and eggs producers adjust to correct long run disequilibrium with a high speed of adjustment (33.5% and 46.7 % respectively). The implication of this is that much more in-depth research is needed to identify those factors that affect agricultural supply and to describe the effect of factors that shift supply in response to price incentives. This could provide valuable information for government in the use of appropriate policy measures.

Keywords: Algeria, cointegration, livestock, supply response, vector error correction model

Procedia PDF Downloads 126
8862 Assessment of Ultra-High Cycle Fatigue Behavior of EN-GJL-250 Cast Iron Using Ultrasonic Fatigue Testing Machine

Authors: Saeedeh Bakhtiari, Johannes Depessemier, Stijn Hertelé, Wim De Waele

Abstract:

High cycle fatigue comprising up to 107 load cycles has been the subject of many studies, and the behavior of many materials was recorded adequately in this regime. However, many applications involve larger numbers of load cycles during the lifetime of machine components. In this ultra-high cycle regime, other failure mechanisms play, and the concept of a fatigue endurance limit (assumed for materials such as steel) is often an oversimplification of reality. When machine component design demands a high geometrical complexity, cast iron grades become interesting candidate materials. Grey cast iron is known for its low cost, high compressive strength, and good damping properties. However, the ultra-high cycle fatigue behavior of cast iron is poorly documented. The current work focuses on the ultra-high cycle fatigue behavior of EN-GJL-250 (GG25) grey cast iron by developing an ultrasonic (20 kHz) fatigue testing system. Moreover, the testing machine is instrumented to measure the temperature and the displacement of  the specimen, and to control the temperature. The high resonance frequency allowed to assess the  behavior of the cast iron of interest within a matter of days for ultra-high numbers of cycles, and repeat the tests to quantify the natural scatter in fatigue resistance.

Keywords: GG25, cast iron, ultra-high cycle fatigue, ultrasonic test

Procedia PDF Downloads 160
8861 Machine Learning for Targeting of Conditional Cash Transfers: Improving the Effectiveness of Proxy Means Tests to Identify Future School Dropouts and the Poor

Authors: Cristian Crespo

Abstract:

Conditional cash transfers (CCTs) have been targeted towards the poor. Thus, their targeting assessments check whether these schemes have been allocated to low-income households or individuals. However, CCTs have more than one goal and target group. An additional goal of CCTs is to increase school enrolment. Hence, students at risk of dropping out of school also are a target group. This paper analyses whether one of the most common targeting mechanisms of CCTs, a proxy means test (PMT), is suitable to identify the poor and future school dropouts. The PMT is compared with alternative approaches that use the outputs of a predictive model of school dropout. This model was built using machine learning algorithms and rich administrative datasets from Chile. The paper shows that using machine learning outputs in conjunction with the PMT increases targeting effectiveness by identifying more students who are either poor or future dropouts. This joint targeting approach increases effectiveness in different scenarios except when the social valuation of the two target groups largely differs. In these cases, the most likely optimal approach is to solely adopt the targeting mechanism designed to find the highly valued group.

Keywords: conditional cash transfers, machine learning, poverty, proxy means tests, school dropout prediction, targeting

Procedia PDF Downloads 190
8860 Computational Intelligence and Machine Learning for Urban Drainage Infrastructure Asset Management

Authors: Thewodros K. Geberemariam

Abstract:

The rapid physical expansion of urbanization coupled with aging infrastructure presents a unique decision and management challenges for many big city municipalities. Cities must therefore upgrade and maintain the existing aging urban drainage infrastructure systems to keep up with the demands. Given the overall contribution of assets to municipal revenue and the importance of infrastructure to the success of a livable city, many municipalities are currently looking for a robust and smart urban drainage infrastructure asset management solution that combines management, financial, engineering and technical practices. This robust decision-making shall rely on sound, complete, current and relevant data that enables asset valuation, impairment testing, lifecycle modeling, and forecasting across the multiple asset portfolios. On this paper, predictive computational intelligence (CI) and multi-class machine learning (ML) coupled with online, offline, and historical record data that are collected from an array of multi-parameter sensors are used for the extraction of different operational and non-conforming patterns hidden in structured and unstructured data to determine and produce actionable insight on the current and future states of the network. This paper aims to improve the strategic decision-making process by identifying all possible alternatives; evaluate the risk of each alternative, and choose the alternative most likely to attain the required goal in a cost-effective manner using historical and near real-time urban drainage infrastructure data for urban drainage infrastructures assets that have previously not benefited from computational intelligence and machine learning advancements.

Keywords: computational intelligence, machine learning, urban drainage infrastructure, machine learning, classification, prediction, asset management space

Procedia PDF Downloads 145
8859 Design of Fuzzy Logic Based Global Power System Stabilizer for Dynamic Stability Enhancement in Multi-Machine Power System

Authors: N. P. Patidar, J. Earnest, Laxmikant Nagar, Akshay Sharma

Abstract:

This paper describes the diligence of a new input signal based fuzzy power system stabilizer in multi-machine power system. Instead of conventional input pairs like speed deviation (∆ω) and derivative of speed deviation i.e. acceleration (∆ω ̇) or speed deviation and accelerating power deviation of each machine, in this paper, deviation of active power through the tie line colligating two areas is used as one of the inputs to the fuzzy logic controller in concurrence with the speed deviation. Fuzzy Logic has the features of simple concept, easy effectuation, and computationally efficient. The advantage of this input is that, the same signal can be fed to each of the fuzzy logic controller connected with each machine. The simulated system comprises of two fully symmetrical areas coupled together by two 230 kV lines. Each area is equipped with two superposable generators rated 20 kV/900MVA and area-1 is exporting 413 MW to area-2. The effectiveness of the proposed control scheme has been assessed by performing small signal stability assessment and transient stability assessment. The proposed control scheme has been compared with a conventional PSS. Digital simulation is used to demonstrate the performance of fuzzy logic controller.

Keywords: Power System Stabilizer (PSS), small signal stability, inter-area oscillation, fuzzy logic controller, membership function, rule base

Procedia PDF Downloads 515
8858 An Intrusion Detection Systems Based on K-Means, K-Medoids and Support Vector Clustering Using Ensemble

Authors: A. Mohammadpour, Ebrahim Najafi Kajabad, Ghazale Ipakchi

Abstract:

Presently, computer networks’ security rise in importance and many studies have also been conducted in this field. By the penetration of the internet networks in different fields, many things need to be done to provide a secure industrial and non-industrial network. Fire walls, appropriate Intrusion Detection Systems (IDS), encryption protocols for information sending and receiving, and use of authentication certificated are among things, which should be considered for system security. The aim of the present study is to use the outcome of several algorithms, which cause decline in IDS errors, in the way that improves system security and prevents additional overload to the system. Finally, regarding the obtained result we can also detect the amount and percentage of more sub attacks. By running the proposed system, which is based on the use of multi-algorithmic outcome and comparing that by the proposed single algorithmic methods, we observed a 78.64% result in attack detection that is improved by 3.14% than the proposed algorithms.

Keywords: intrusion detection systems, clustering, k-means, k-medoids, SV clustering, ensemble

Procedia PDF Downloads 210
8857 Procedure Model for Data-Driven Decision Support Regarding the Integration of Renewable Energies into Industrial Energy Management

Authors: M. Graus, K. Westhoff, X. Xu

Abstract:

The climate change causes a change in all aspects of society. While the expansion of renewable energies proceeds, industry could not be convinced based on general studies about the potential of demand side management to reinforce smart grid considerations in their operational business. In this article, a procedure model for a case-specific data-driven decision support for industrial energy management based on a holistic data analytics approach is presented. The model is executed on the example of the strategic decision problem, to integrate the aspect of renewable energies into industrial energy management. This question is induced due to considerations of changing the electricity contract model from a standard rate to volatile energy prices corresponding to the energy spot market which is increasingly more affected by renewable energies. The procedure model corresponds to a data analytics process consisting on a data model, analysis, simulation and optimization step. This procedure will help to quantify the potentials of sustainable production concepts based on the data from a factory. The model is validated with data from a printer in analogy to a simple production machine. The overall goal is to establish smart grid principles for industry via the transformation from knowledge-driven to data-driven decisions within manufacturing companies.

Keywords: data analytics, green production, industrial energy management, optimization, renewable energies, simulation

Procedia PDF Downloads 428
8856 Malaria Outbreak Facilitated by Appearance of Vector-Breeding Sites after Heavy Rainfall and Inadequate Preventive Measures: Nwoya District, Uganda, March–May 2018

Authors: Godfrey Nsereko, Daniel Kadobera, Denis Okethwangu, Joyce Nguna, Alex Riolexus Ario

Abstract:

Background: Malaria is a leading cause of morbidity and mortality in Uganda. In April 2018, malaria cases surged in Nwoya District, northern Uganda, exceeding the action thresholds. We investigated to assess the outbreak’s magnitude, identify transmission risk factors, and recommend evidence-based control measures. Methods: We defined a malaria case as onset of fever in a resident of Nwoya District with a positive Rapid Diagnostic Test or microscopy for malaria P. falciparum from 1 February to 22 May 2018. We reviewed medical records in all health facilities of affected sub-counties to find cases. In a case-control study, we compared exposure risk factors between 107 case-persons and 107 asymptomatic controls matched by age and village. We conducted entomological assessment on vector-density and behavior. Results: We identified 3,879 case-persons (attack rate [AR]=6.5%) and 2 deaths (case-fatality rate=5.2/10,000). Females (AR=8.1%) were more affected than males (AR=4.7%). Of all age groups, the 5-18 year age group (AR=8.4%) was most affected. Heavy rain started on 4 March; a propagated outbreak began during the week of 2 April. In the case-control study, 55% (59/107) of case-patients and 18% (19/107) of controls had stagnant water around households for several days following rainfall (ORM-H=5.6, 95%CI=3.0-11); 25% (27/107) of case-patients and 51% (55/107) of controls wore long-sleeve cloths during evening hours (ORM-H=0.30, 95%CI=0.20-0.60); 29% (31/107) of case-patients and 15% (16/107) of controls did not sleep under a long-lasting insecticide-treated net (LLIN) (ORM-H=2.3, 95%CI=1.1-4.9); 37% (40/107) of case-patients and 52% (56/107) of controls had ≥1 LLIN per 2 household members (ORM-H=0.54, 95%CI=0.30-0.97). Entomological assessment indicated active breeding sites; Anopheles gambiae sensu lato species were the predominant vector. Conclusion: Increased vector breeding sites after heavy rainfall, together with inadequate malaria preventive measures caused this outbreak. We recommended increasing coverage for LLINs and larviciding breeding sites.

Keywords: malaria outbreak, Plasmodium falciparum, global health security, Uganda

Procedia PDF Downloads 218
8855 Machine Learning in Patent Law: How Genetic Breeding Algorithms Challenge Modern Patent Law Regimes

Authors: Stefan Papastefanou

Abstract:

Artificial intelligence (AI) is an interdisciplinary field of computer science with the aim of creating intelligent machine behavior. Early approaches to AI have been configured to operate in very constrained environments where the behavior of the AI system was previously determined by formal rules. Knowledge was presented as a set of rules that allowed the AI system to determine the results for specific problems; as a structure of if-else rules that could be traversed to find a solution to a particular problem or question. However, such rule-based systems typically have not been able to generalize beyond the knowledge provided. All over the world and especially in IT-heavy industries such as the United States, the European Union, Singapore, and China, machine learning has developed to be an immense asset, and its applications are becoming more and more significant. It has to be examined how such products of machine learning models can and should be protected by IP law and for the purpose of this paper patent law specifically, since it is the IP law regime closest to technical inventions and computing methods in technical applications. Genetic breeding models are currently less popular than recursive neural network method and deep learning, but this approach can be more easily described by referring to the evolution of natural organisms, and with increasing computational power; the genetic breeding method as a subset of the evolutionary algorithms models is expected to be regaining popularity. The research method focuses on patentability (according to the world’s most significant patent law regimes such as China, Singapore, the European Union, and the United States) of AI inventions and machine learning. Questions of the technical nature of the problem to be solved, the inventive step as such, and the question of the state of the art and the associated obviousness of the solution arise in the current patenting processes. Most importantly, and the key focus of this paper is the problem of patenting inventions that themselves are developed through machine learning. The inventor of a patent application must be a natural person or a group of persons according to the current legal situation in most patent law regimes. In order to be considered an 'inventor', a person must actually have developed part of the inventive concept. The mere application of machine learning or an AI algorithm to a particular problem should not be construed as the algorithm that contributes to a part of the inventive concept. However, when machine learning or the AI algorithm has contributed to a part of the inventive concept, there is currently a lack of clarity regarding the ownership of artificially created inventions. Since not only all European patent law regimes but also the Chinese and Singaporean patent law approaches include identical terms, this paper ultimately offers a comparative analysis of the most relevant patent law regimes.

Keywords: algorithms, inventor, genetic breeding models, machine learning, patentability

Procedia PDF Downloads 101
8854 Introducing Design Principles for Clinical Decision Support Systems

Authors: Luca Martignoni

Abstract:

The increasing usage of clinical decision support systems in healthcare and the demand for software that enables doctors to take informed decisions is changing everyday clinical practice. However, as technology advances not only are the benefits of technology growing, but so are the potential risks. A growing danger is the doctors’ over-reliance on the proposed decision of the clinical decision support system, leading towards deskilling and rash decisions by doctors. In that regard, identifying doctors' requirements for software and developing approaches to prevent technological over-reliance is of utmost importance. In this paper, we report the results of a design science research study, focusing on the requirements and design principles of ultrasound software. We conducted a total of 15 interviews with experts about poten-tial ultrasound software functions. Subsequently, we developed meta-requirements and design principles to design future clinical decision support systems efficiently and as free from the occur-rence of technological over-reliance as possible.

Keywords: clinical decision support systems, technological over-reliance, design principles, design science research

Procedia PDF Downloads 89
8853 Understanding the Qualitative Nature of Product Reviews by Integrating Text Processing Algorithm and Usability Feature Extraction

Authors: Cherry Yieng Siang Ling, Joong Hee Lee, Myung Hwan Yun

Abstract:

The quality of a product to be usable has become the basic requirement in consumer’s perspective while failing the requirement ends up the customer from not using the product. Identifying usability issues from analyzing quantitative and qualitative data collected from usability testing and evaluation activities aids in the process of product design, yet the lack of studies and researches regarding analysis methodologies in qualitative text data of usability field inhibits the potential of these data for more useful applications. While the possibility of analyzing qualitative text data found with the rapid development of data analysis studies such as natural language processing field in understanding human language in computer, and machine learning field in providing predictive model and clustering tool. Therefore, this research aims to study the application capability of text processing algorithm in analysis of qualitative text data collected from usability activities. This research utilized datasets collected from LG neckband headset usability experiment in which the datasets consist of headset survey text data, subject’s data and product physical data. In the analysis procedure, which integrated with the text-processing algorithm, the process includes training of comments onto vector space, labeling them with the subject and product physical feature data, and clustering to validate the result of comment vector clustering. The result shows 'volume and music control button' as the usability feature that matches best with the cluster of comment vectors where centroid comments of a cluster emphasized more on button positions, while centroid comments of the other cluster emphasized more on button interface issues. When volume and music control buttons are designed separately, the participant experienced less confusion, and thus, the comments mentioned only about the buttons' positions. While in the situation where the volume and music control buttons are designed as a single button, the participants experienced interface issues regarding the buttons such as operating methods of functions and confusion of functions' buttons. The relevance of the cluster centroid comments with the extracted feature explained the capability of text processing algorithms in analyzing qualitative text data from usability testing and evaluations.

Keywords: usability, qualitative data, text-processing algorithm, natural language processing

Procedia PDF Downloads 274
8852 Predicting the Quality of Life on the Basis of Perceived Social Support among Patients with Coronary Artery Bypass Graft

Authors: Azadeh Yaraghchi, Reza Bagherian Sararoodi, Niknaz Salehi Moghadam, Mohammad Hossein Mandegar, Adis Kraskian Mujembari, Omid Rezaei

Abstract:

Background: Quality of life is one of the most important consequences of disease in psychosomatic disorders. Many psychological factors are considered in predicting quality of life in patients with coronary artery bypass graft (CABG). The present study was aimed to determine the relationship between perceived social support and quality of life in patients with coronary artery bypass graft (CABG). Methods: The population included 82 patients who had undergone CABG from October 2014 to May 2015 in four different hospitals in Tehran. The patients were evaluated with Multi-dimension scale of perceived social support (MSPSS) and after three months follow up were evaluated by Short-Form quality of life questionnaire (SF-36). The obtained data were analyzed through Pearson correlation test and multiple variable regression models. Findings: A relationship between perceived social support and quality of life in patients with CABG was observed (r=0.374, p<0.01). The results showed that 22.4% of variation in quality of life is predicted by perceived social support components (p<0.01, R2 =0.224). Conclusion: Based on the results, perceived social support is one of the predictors of quality of life in patients with coronary artery bypass graft. Accordingly, these results can be useful in conceiving proactive policies, detecting high risk patients and planning for psychological interventions.

Keywords: coronary artery bypass graft, perceived social support, psychological factors, quality of life

Procedia PDF Downloads 353
8851 Determination of Water Pollution and Water Quality with Decision Trees

Authors: Çiğdem Bakır, Mecit Yüzkat

Abstract:

With the increasing emphasis on water quality worldwide, the search for and expanding the market for new and intelligent monitoring systems has increased. The current method is the laboratory process, where samples are taken from bodies of water, and tests are carried out in laboratories. This method is time-consuming, a waste of manpower, and uneconomical. To solve this problem, we used machine learning methods to detect water pollution in our study. We created decision trees with the Orange3 software we used in our study and tried to determine all the factors that cause water pollution. An automatic prediction model based on water quality was developed by taking many model inputs such as water temperature, pH, transparency, conductivity, dissolved oxygen, and ammonia nitrogen with machine learning methods. The proposed approach consists of three stages: preprocessing of the data used, feature detection, and classification. We tried to determine the success of our study with different accuracy metrics and the results. We presented it comparatively. In addition, we achieved approximately 98% success with the decision tree.

Keywords: decision tree, water quality, water pollution, machine learning

Procedia PDF Downloads 73
8850 Examining Resilience, Social Supports, and Self-Esteem as Predictors of the Quality of Life of ODAPUS (Orang Dengan Lupus)

Authors: Yulmaida Amir, Fahrul Rozi, Insany C. Kamil, Fanny Aryani

Abstract:

ODAPUS (Orang dengan Lupus) is an Indonesian term for people with Lupus, a chronic autoimmune disease in which immune system of the body becomes hyperactive and attacks normal tissue. The number of ODAPUS indicate an increase in Indonesia, thereby helping to improve their quality of life to be important to help their recovery. This study aims to examine the effect of resilience, self-esteem, and social support on the quality of life of women who had been diagnosed as having Lupus. Data were collected from 64 ODAPUS in Indonesia, using the World Health Organization Quality of Life (WHOQOL), Resilience Scale from Wagnil and Young (1993), self-esteem scale (developed from Coopersmith’s theory), and Social Support Questioner from Northouse (1988). Regression data analysis showed that resilience, social support, and self-esteem predict the quality of life of the ODAPUS simultaneously. If the variable was analysed individually, self-esteem did not significantly contribute to the quality of life. Resilience contributed most significantly to the quality of life, followed by social support. Of five sources of social supports included in the research, support from family members (parents and brother/sisters) has the most significant contribution to the quality of life, followed by support from spouse, and from friends. Interestingly, social support from medical personnel (medical doctors and nurses) had not a significant contribution to the quality of life of ODAPUS. As a conclusion, this research showed that the ability of ODAPUS to cope with difficulty in life, and support from family members, spouse, and friends were the significant predictors for their quality of life.

Keywords: quality of life, resilience, self-esteem, social supports

Procedia PDF Downloads 154
8849 Hybrid Fermentation System for Improvement of Ergosterol Biosynthesis

Authors: Alexandra Tucaliuc, Alexandra C. Blaga, Anca I. Galaction, Lenuta Kloetzer, Dan Cascaval

Abstract:

Ergosterol (ergosta-5,7,22-trien-3β-ol), also known as provitamin D2, is the precursor of vitamin D2 (ergocalciferol), because it is converted under UV radiation to this vitamin. The natural sources of ergosterol are mainly the yeasts (Saccharomyces sp., Candida sp.), but it can be also found in fungus (Claviceps sp.) or plants (orchids). In the yeasts cells, ergosterol is accumulated in membranes, especially in free form in the plasma membrane, but also as esters with fatty acids in membrane lipids. The chemical synthesis of ergosterol does not represent an efficient method for its production, in these circumstances, the most attractive alternative for producing ergosterol at larger-scale remains the aerobic fermentation using S. cerevisiae on glucose or by-products from agriculture of food industry as substrates, in batch or fed-batch operating systems. The aim of this work is to analyze comparatively the influence of aeration efficiency on ergosterol production by S. cerevisiae in batch and fed-batch fermentations, by considering different levels of mixing intensity, aeration rate, and n-dodecane concentration. The effects of the studied factors are quantitatively described by means of the mathematical correlations proposed for each of the two fermentation systems, valid both for the absence and presence of oxygen-vector inside the broth. The experiments were carried out in a laboratory stirred bioreactor, provided with computer-controlled and recorded parameters. n-Dodecane was used as oxygen-vector and the ergosterol content inside the yeasts cells has been considered at the fermentation moment related to the maximum concentration of ergosterol, 9 hrs for batch process and 20 hrs for fed-batch one. Ergosterol biosynthesis is strongly dependent on the dissolved oxygen concentration. The hydrocarbon concentration exhibits a significant influence on ergosterol production mainly by accelerating the oxygen transfer rate. Regardless of n-dodecane addition, by maintaining the glucose concentration at a constant level in the fed-batch process, the amount of ergosterol accumulated into the yeasts cells has been almost tripled. In the presence of hydrocarbon, the ergosterol concentration increased by over 50%. The value of oxygen-vector concentration corresponding to the maximum level of ergosterol depends mainly on biomass concentration, due to its negative influences on broth viscosity and interfacial phenomena of air bubbles blockage through the adsorption of hydrocarbon droplets–yeast cells associations. Therefore, for the batch process, the maximum ergosterol amount was reached for 5% vol. n-dodecane, while for the fed-batch process for 10% vol. hydrocarbon.

Keywords: bioreactors, ergosterol, fermentation, oxygen-vector

Procedia PDF Downloads 170
8848 Leadership's Controlling via Complexity Investigation in Crisis Scenarios

Authors: Jiří Barta, Oldřich Svoboda, Jiří F. Urbánek

Abstract:

In this paper will be discussed two coin´s sides of crisis scenarios dynamics. On the one's side is negative role of subsidiary scenario branches in its compactness weakening by means unduly chaotic atomizing, having many interactive feedbacks cases, increasing a value of a complexity here. This negative role reflects the complexity of use cases, weakening leader compliancy, which brings something as a ´readiness for controlling capabilities provision´. Leader´s dissatisfaction has zero compliancy, but factual it is a ´crossbar´ (interface in fact) between planning and executing use cases. On the other side of this coin, an advantage of rich scenarios embranchment is possible to see in a support of response awareness, readiness, preparedness, adaptability, creativity and flexibility. Here rich scenarios embranchment contributes to the steadiness and resistance of scenario mission actors. These all will be presented in live power-points ´Blazons´, modelled via DYVELOP (Dynamic Vector Logistics of Processes) on the Conference.

Keywords: leadership, controlling, complexity, DYVELOP, scenarios

Procedia PDF Downloads 391
8847 Effects of Training on Self-Efficacy, Competence, and Target Complaints of Dementia Family Support Program Facilitators

Authors: Myonghwa Park, Eun Jeong Choi

Abstract:

Persons with dementia living at home have complex caregiving demands, which can be significant sources of stress for the family caregivers. Thus, the dementia family support program facilitators struggle to provide various health and social services, facing diverse challenges. The purpose of this study was to research the effects of training program for the dementia family support program facilitators on self-efficacy, competence, and target complaints concerning operating their program. We created a training program with systematic contents, which was composed of 10 sessions and we provided the program for the facilitators. The participants were 32 people at 28 community dementia support centers who manage dementia family support programs and they completed quantitative and qualitative self-report questionnaire before and after participating in the training program. For analyzing the data, descriptive statistics were used and with a paired t-test, pretest and posttest scores of self-efficacy, competence, and target complaints were analyzed. We used Statistical Package for the Social Sciences (SPSS) statistics (Version 21) to analyze the data. The average age of the participants was 39.6 years old and the 84.4% of participants were nurses. There were statistically meaningful increases in facilitators’ self-efficacy scores (t = -4.45, p < .001) and competence scores (t = -2.133, p = 0.041) after participating in training program and operating their own dementia family support program. Also, the facilitators’ difficulties in conducting their dementia family support program were decreased which was assessed with target complaints. Especially, the facilitators’ lack of dementia expertise and experience was decreased statistically significantly (t = 3.520, p = 0.002). Findings provided evidence of the benefits of the training program for facilitators to enhance managing dementia family support program by improving the facilitators’ self-efficacy and competence and decreasing their difficulties regarding operating their program.

Keywords: competence, dementia, facilitator, family, self-efficacy, training

Procedia PDF Downloads 202
8846 Reliability Analysis of a Life Support System in a Public Aquarium

Authors: Mehmet Savsar

Abstract:

Complex Life Support Systems (LSS) are used in all large commercial and public aquariums in order to keep the fish alive. Reliabilities of individual equipment, as well as the complete system, are extremely important and critical since the life and safety of important fish depend on these life support systems. Failure of some critical device or equipment, which do not have redundancy, results in negative consequences and affects life support as a whole. In this paper, we have considered a life support system in a large public aquarium in Kuwait Scientific Center and presented a procedure and analysis to show how the reliability of such systems can be estimated by using appropriate tools and collected data. We have also proposed possible improvements for systems reliability. In particular, addition of parallel components and spare parts are considered and the numbers of spare parts needed for each component to achieve a required reliability during specified lead time are calculated. The results show that significant improvements in system reliability can be achieved by operating some LSS components in parallel and having certain numbers of spares available in the spare parts inventories. The procedures and the results presented in this paper are expected to be useful for aquarium engineers and maintenance managers dealing with LSS.

Keywords: life support systems, aquariums, reliability, failures, availability, spare parts

Procedia PDF Downloads 271
8845 Characteristics of an Impact on Reading Comprehension of Elementary School Students

Authors: Judith Hanke

Abstract:

Due to the rise of students with reading difficulties, a digital reading support was developed. The digital reading support focuses on reading comprehension of elementary school students. It consists of literary texts and reading exercises with diagnostics. To analyze the use of the reading packages an intervention study took place in 2023. For the methodology, an ABA-design was selected for the intervention study to examine the reading packages. The study was expedited from April 2023 until July 2023 and collected quantitative data of individuals, groups, and classes. It consisted of a survey group (N = 58) and a control group (N = 53). The pretest was conducted before the reading support intervention. The students of the survey group received reading support on their ability level to aid the individual student’s needs. At the beginning of the study characteristics of the students were collected. The characteristics included gender, age, repetition of a class, spoken language at home, German as a second language, and special support needs such as dyslexia; right after the intervention, the posttest was examined. At least three weeks after the intervention, the follow-up testing was administered. A standardized reading comprehension test was used for the three test times. The test consists of three subtests: word comprehension, sentence comprehension, and text comprehension. The focus of this paper is to determine which characteristics have an impact on reading comprehension of elementary school students. The students’ characteristics were correlated with the three test times through a Pearson correlation. The main findings are that age, repetition of a class, spoken language at home, German as a second language have an effect on reading comprehension. Interestingly gender and special support needs did not have a significant effect on the reading comprehension of the students. The significance of the study is to determine which characteristics have an impact on reading comprehension and then to assess how reading support can be modified to support the diverse students.

Keywords: class repetition, reading comprehension, reading support, second language, spoken language at home

Procedia PDF Downloads 14
8844 Advancing Urban Sustainability through Data-Driven Machine Learning Solutions

Authors: Nasim Eslamirad, Mahdi Rasoulinezhad, Francesco De Luca, Sadok Ben Yahia, Kimmo Sakari Lylykangas, Francesco Pilla

Abstract:

With the ongoing urbanization, cities face increasing environmental challenges impacting human well-being. To tackle these issues, data-driven approaches in urban analysis have gained prominence, leveraging urban data to promote sustainability. Integrating Machine Learning techniques enables researchers to analyze and predict complex environmental phenomena like Urban Heat Island occurrences in urban areas. This paper demonstrates the implementation of data-driven approach and interpretable Machine Learning algorithms with interpretability techniques to conduct comprehensive data analyses for sustainable urban design. The developed framework and algorithms are demonstrated for Tallinn, Estonia to develop sustainable urban strategies to mitigate urban heat waves. Geospatial data, preprocessed and labeled with UHI levels, are used to train various ML models, with Logistic Regression emerging as the best-performing model based on evaluation metrics to derive a mathematical equation representing the area with UHI or without UHI effects, providing insights into UHI occurrences based on buildings and urban features. The derived formula highlights the importance of building volume, height, area, and shape length to create an urban environment with UHI impact. The data-driven approach and derived equation inform mitigation strategies and sustainable urban development in Tallinn and offer valuable guidance for other locations with varying climates.

Keywords: data-driven approach, machine learning transparent models, interpretable machine learning models, urban heat island effect

Procedia PDF Downloads 19
8843 Automatic Lead Qualification with Opinion Mining in Customer Relationship Management Projects

Authors: Victor Radich, Tania Basso, Regina Moraes

Abstract:

Lead qualification is one of the main procedures in Customer Relationship Management (CRM) projects. Its main goal is to identify potential consumers who have the ideal characteristics to establish a profitable and long-term relationship with a certain organization. Social networks can be an important source of data for identifying and qualifying leads since interest in specific products or services can be identified from the users’ expressed feelings of (dis)satisfaction. In this context, this work proposes the use of machine learning techniques and sentiment analysis as an extra step in the lead qualification process in order to improve it. In addition to machine learning models, sentiment analysis or opinion mining can be used to understand the evaluation that the user makes of a particular service, product, or brand. The results obtained so far have shown that it is possible to extract data from social networks and combine the techniques for a more complete classification.

Keywords: lead qualification, sentiment analysis, opinion mining, machine learning, CRM, lead scoring

Procedia PDF Downloads 66
8842 Natural Language Processing for the Classification of Social Media Posts in Post-Disaster Management

Authors: Ezgi Şendil

Abstract:

Information extracted from social media has received great attention since it has become an effective alternative for collecting people’s opinions and emotions based on specific experiences in a faster and easier way. The paper aims to put data in a meaningful way to analyze users’ posts and get a result in terms of the experiences and opinions of the users during and after natural disasters. The posts collected from Reddit are classified into nine different categories, including injured/dead people, infrastructure and utility damage, missing/found people, donation needs/offers, caution/advice, and emotional support, identified by using labelled Twitter data and four different machine learning (ML) classifiers.

Keywords: disaster, NLP, postdisaster management, sentiment analysis

Procedia PDF Downloads 61
8841 Copyright Clearance for Artificial Intelligence Training Data: Challenges and Solutions

Authors: Erva Akin

Abstract:

– The use of copyrighted material for machine learning purposes is a challenging issue in the field of artificial intelligence (AI). While machine learning algorithms require large amounts of data to train and improve their accuracy and creativity, the use of copyrighted material without permission from the authors may infringe on their intellectual property rights. In order to overcome copyright legal hurdle against the data sharing, access and re-use of data, the use of copyrighted material for machine learning purposes may be considered permissible under certain circumstances. For example, if the copyright holder has given permission to use the data through a licensing agreement, then the use for machine learning purposes may be lawful. It is also argued that copying for non-expressive purposes that do not involve conveying expressive elements to the public, such as automated data extraction, should not be seen as infringing. The focus of such ‘copy-reliant technologies’ is on understanding language rules, styles, and syntax and no creative ideas are being used. However, the non-expressive use defense is within the framework of the fair use doctrine, which allows the use of copyrighted material for research or educational purposes. The questions arise because the fair use doctrine is not available in EU law, instead, the InfoSoc Directive provides for a rigid system of exclusive rights with a list of exceptions and limitations. One could only argue that non-expressive uses of copyrighted material for machine learning purposes do not constitute a ‘reproduction’ in the first place. Nevertheless, the use of machine learning with copyrighted material is difficult because EU copyright law applies to the mere use of the works. Two solutions can be proposed to address the problem of copyright clearance for AI training data. The first is to introduce a broad exception for text and data mining, either mandatorily or for commercial and scientific purposes, or to permit the reproduction of works for non-expressive purposes. The second is that copyright laws should permit the reproduction of works for non-expressive purposes, which opens the door to discussions regarding the transposition of the fair use principle from the US into EU law. Both solutions aim to provide more space for AI developers to operate and encourage greater freedom, which could lead to more rapid innovation in the field. The Data Governance Act presents a significant opportunity to advance these debates. Finally, issues concerning the balance of general public interests and legitimate private interests in machine learning training data must be addressed. In my opinion, it is crucial that robot-creation output should fall into the public domain. Machines depend on human creativity, innovation, and expression. To encourage technological advancement and innovation, freedom of expression and business operation must be prioritised.

Keywords: artificial intelligence, copyright, data governance, machine learning

Procedia PDF Downloads 75
8840 Tracing Back the Bot Master

Authors: Sneha Leslie

Abstract:

The current situation in the cyber world is that crimes performed by Botnets are increasing and the masterminds (botmaster) are not detectable easily. The botmaster in the botnet compromises the legitimate host machines in the network and make them bots or zombies to initiate the cyber-attacks. This paper will focus on the live detection of the botmaster in the network by using the strong framework 'metasploit', when distributed denial of service (DDOS) attack is performed by the botnet. The affected victim machine will be continuously monitoring its incoming packets. Once the victim machine gets to know about the excessive count of packets from any IP, that particular IP is noted and details of the noted systems are gathered. Using the vulnerabilities present in the zombie machines (already compromised by botmaster), the victim machine will compromise them. By gaining access to the compromised systems, applications are run remotely. By analyzing the incoming packets of the zombies, the victim comes to know the address of the botmaster. This is an effective and a simple system where no specific features of communication protocol are considered.

Keywords: bonet, DDoS attack, network security, detection system, metasploit framework

Procedia PDF Downloads 245
8839 New Machine Learning Optimization Approach Based on Input Variables Disposition Applied for Time Series Prediction

Authors: Hervice Roméo Fogno Fotsoa, Germaine Djuidje Kenmoe, Claude Vidal Aloyem Kazé

Abstract:

One of the main applications of machine learning is the prediction of time series. But a more accurate prediction requires a more optimal model of machine learning. Several optimization techniques have been developed, but without considering the input variables disposition of the system. Thus, this work aims to present a new machine learning architecture optimization technique based on their optimal input variables disposition. The validations are done on the prediction of wind time series, using data collected in Cameroon. The number of possible dispositions with four input variables is determined, i.e., twenty-four. Each of the dispositions is used to perform the prediction, with the main criteria being the training and prediction performances. The results obtained from a static architecture and a dynamic architecture of neural networks have shown that these performances are a function of the input variable's disposition, and this is in a different way from the architectures. This analysis revealed that it is necessary to take into account the input variable's disposition for the development of a more optimal neural network model. Thus, a new neural network training algorithm is proposed by introducing the search for the optimal input variables disposition in the traditional back-propagation algorithm. The results of the application of this new optimization approach on the two single neural network architectures are compared with the previously obtained results step by step. Moreover, this proposed approach is validated in a collaborative optimization method with a single objective optimization technique, i.e., genetic algorithm back-propagation neural networks. From these comparisons, it is concluded that each proposed model outperforms its traditional model in terms of training and prediction performance of time series. Thus the proposed optimization approach can be useful in improving the accuracy of time series forecasts. This proves that the proposed optimization approach can be useful in improving the accuracy of time series prediction based on machine learning.

Keywords: input variable disposition, machine learning, optimization, performance, time series prediction

Procedia PDF Downloads 91
8838 Propylene Self-Metathesis to Ethylene and Butene over WOx/SiO2, Effect of Nano-Sized Extra Supports (SiO2 and TiO2)

Authors: Adisak Guntida

Abstract:

Propylene self-metathesis to ethylene and butene was studied over WOx/SiO2 catalysts at 450 °C and atmospheric pressure. The WOx/SiO2 catalysts were prepared by incipient wetness impregnation of ammonium metatungstate aqueous solution. It was found that, adding nano-sized extra supports (SiO2 and TiO2) by physical mixing with the WOx/SiO2 enhanced propylene conversion. The UV-Vis and FT-Raman results revealed that WOx could migrate from the original silica support to the extra support, leading to a better dispersion of WOx. The ICP-OES results also indicate that WOx existed on the extra support. Coke formation was investigated on the catalysts after 10 h time-on-stream by TPO. However, adding nano-sized extra supports led to higher coke formation which may be related to acidity as characterized by NH3-TPD.

Keywords: extra support, nanomaterial, propylene self-metathesis, tungsten oxide

Procedia PDF Downloads 239