Search results for: cluster model approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 26806

Search results for: cluster model approach

24136 Using Bidirectional Encoder Representations from Transformers to Extract Topic-Independent Sentiment Features for Social Media Bot Detection

Authors: Maryam Heidari, James H. Jones Jr.

Abstract:

Millions of online posts about different topics and products are shared on popular social media platforms. One use of this content is to provide crowd-sourced information about a specific topic, event or product. However, this use raises an important question: what percentage of information available through these services is trustworthy? In particular, might some of this information be generated by a machine, i.e., a bot, instead of a human? Bots can be, and often are, purposely designed to generate enough volume to skew an apparent trend or position on a topic, yet the consumer of such content cannot easily distinguish a bot post from a human post. In this paper, we introduce a model for social media bot detection which uses Bidirectional Encoder Representations from Transformers (Google Bert) for sentiment classification of tweets to identify topic-independent features. Our use of a Natural Language Processing approach to derive topic-independent features for our new bot detection model distinguishes this work from previous bot detection models. We achieve 94\% accuracy classifying the contents of data as generated by a bot or a human, where the most accurate prior work achieved accuracy of 92\%.

Keywords: bot detection, natural language processing, neural network, social media

Procedia PDF Downloads 103
24135 Modelling of Cavity Growth in Underground Coal Gasification

Authors: Preeti Aghalayam, Jay Shah

Abstract:

Underground coal gasification (UCG) is the in-situ gasification of unmineable coals to produce syngas. In UCG, gasifying agents are injected into the coal seam, and a reactive cavity is formed due to coal consumption. The cavity formed is typically hemispherical, and this report consists of the MATLAB model of the UCG cavity to predict the composition of the output gases. There are seven radial and two time-variant ODEs. A MATLAB solver (ode15s) is used to solve the radial ODEs from the above equations. Two for-loops are implemented in the model, i.e., one for time variations and another for radial variation. In the time loop, the radial odes are solved using the MATLAB solver. The radial loop is nested inside the time loop, and the density odes are numerically solved using the Euler method. The model is validated by comparing it with the literature results of laboratory-scale experiments. The model predicts the radial and time variation of the product gases inside the cavity.

Keywords: gasification agent, MATLAB model, syngas, underground coal gasification (UCG)

Procedia PDF Downloads 187
24134 Prediction Compressive Strength of Self-Compacting Concrete Containing Fly Ash Using Fuzzy Logic Inference System

Authors: Belalia Douma Omar, Bakhta Boukhatem, Mohamed Ghrici

Abstract:

Self-compacting concrete (SCC) developed in Japan in the late 80s has enabled the construction industry to reduce demand on the resources, improve the work condition and also reduce the impact of environment by elimination of the need for compaction. Fuzzy logic (FL) approaches has recently been used to model some of the human activities in many areas of civil engineering applications. Especially from these systems in the model experimental studies, very good results have been obtained. In the present study, a model for predicting compressive strength of SCC containing various proportions of fly ash, as partial replacement of cement has been developed by using Adaptive Neuro-Fuzzy Inference System (ANFIS). For the purpose of building this model, a database of experimental data were gathered from the literature and used for training and testing the model. The used data as the inputs of fuzzy logic models are arranged in a format of five parameters that cover the total binder content, fly ash replacement percentage, water content, super plasticizer and age of specimens. The training and testing results in the fuzzy logic model have shown a strong potential for predicting the compressive strength of SCC containing fly ash in the considered range.

Keywords: self-compacting concrete, fly ash, strength prediction, fuzzy logic

Procedia PDF Downloads 315
24133 Toward Concerned Leadership: A Novel Conceptual Model to Raise the Well-Being of Employees and the Leaderful Practice of Organizations

Authors: Robert McGrath, Zara Qureshi

Abstract:

A innovative leadership philosophy that is proposed herein is distinctly more humane than most leadership approaches Concerned Leadership. The central idea to this approach is to consider the whole person that comes to work; their professional skills and talents, as well as any personal, emotional challenges that could be affecting productivity and effectiveness at work. This paper explores Concerned Leadership as an integration of the two conceptual models areas examined in this paper –(1) leaderful organizations and practices, as well as (2) organizational culture, and defines leadership in the context of Mental Health and Wellness in the workplace. Leaderful organizations calls for organizations to implement leaderful practice. Leaderful practice is when leadership responsibility and decision-making is shared across all team members and levels, versus only delegated to top management as commonly seen. A healthy culture thrives off key aspects such as acceptance, employee pride, equal opportunity, and strong company leadership. Concerned Leadership is characterized by five main components: Self-Concern, Leaderful Practice, Human Touch, Belonging, and Compassion. As scholars and practitioners conceptualize leadership in practice, the present model seeks to uphold the dignity of each organizational member, thereby having the potential to transform workplaces and support all members.

Keywords: leadership, mental health, reflective practice, organizational culture

Procedia PDF Downloads 64
24132 A Study on the Application of Machine Learning and Deep Learning Techniques for Skin Cancer Detection

Authors: Hritwik Ghosh, Irfan Sadiq Rahat, Sachi Nandan Mohanty, J. V. R. Ravindra

Abstract:

In the rapidly evolving landscape of medical diagnostics, the early detection and accurate classification of skin cancer remain paramount for effective treatment outcomes. This research delves into the transformative potential of Artificial Intelligence (AI), specifically Deep Learning (DL), as a tool for discerning and categorizing various skin conditions. Utilizing a diverse dataset of 3,000 images representing nine distinct skin conditions, we confront the inherent challenge of class imbalance. This imbalance, where conditions like melanomas are over-represented, is addressed by incorporating class weights during the model training phase, ensuring an equitable representation of all conditions in the learning process. Our pioneering approach introduces a hybrid model, amalgamating the strengths of two renowned Convolutional Neural Networks (CNNs), VGG16 and ResNet50. These networks, pre-trained on the ImageNet dataset, are adept at extracting intricate features from images. By synergizing these models, our research aims to capture a holistic set of features, thereby bolstering classification performance. Preliminary findings underscore the hybrid model's superiority over individual models, showcasing its prowess in feature extraction and classification. Moreover, the research emphasizes the significance of rigorous data pre-processing, including image resizing, color normalization, and segmentation, in ensuring data quality and model reliability. In essence, this study illuminates the promising role of AI and DL in revolutionizing skin cancer diagnostics, offering insights into its potential applications in broader medical domains.

Keywords: artificial intelligence, machine learning, deep learning, skin cancer, dermatology, convolutional neural networks, image classification, computer vision, healthcare technology, cancer detection, medical imaging

Procedia PDF Downloads 62
24131 A Mathematical Description of a Growing Cell Colony Based on the Mechanical Bidomain Model

Authors: Debabrata Auddya, Bradley J. Roth

Abstract:

The mechanical bidomain model is used to describe a colony of cells growing on a substrate. Analytical expressions are derived for the intracellular and extracellular displacements. Mechanotransduction events are driven by the difference between the displacements in the two spaces, corresponding to the force acting on integrins. The equation for the displacement consists of two terms: one proportional to the radius that is the same in the intracellular and extracellular spaces (the monodomain term) and one that is proportional to a modified Bessel function that is responsible for mechanotransduction (the bidomain term). The model predicts that mechanotransduction occurs within a few length constants of the colony’s edge, and an expression for the length constant contains the intracellular and extracellular shear moduli and the spring constant of the integrins coupling the two spaces. The model predictions are qualitatively consistent with experiments on human embryonic stem cell colonies, in which differentiation is localized near the edge.

Keywords: cell colony, integrin, mechanical bidomain model, stem cell, stress-strain, traction force

Procedia PDF Downloads 227
24130 Simple Assessments to Demystify Complementary Feeding: Leveraging a Successful Literacy Initiative Assessment Approach in Gujarat, India

Authors: Smriti Pahwa, Karishma Vats, Aditi Macwan, Jija Dutt, Sumukhi Vaid

Abstract:

Age approporiate complementary feeding has been stressed upon for sound young child nutrition and appropriate growth. National Infant and Young Child Feeding guidelines, policies and programs indicate cognizance of the issue taken by the country’s government, policy makers and technical experts. However, it is important that ordinary people, the caregivers of young children too understand the importance of appropriate feeding. For this, an interface might be required where ordinary people could participate in assessing the gaps in IYCF as a first step to take subsequent action. In this context an attempt was made to extrapolate a citizen led learning level survey that has been involving around 25000 ordinary citizens to reach out to 600,000 children annually for over a decade in India. Based on this philosophy of involving ordinary people in simple assessments to produce understandable actionable evidence, a rapid diet assessment tool was developed and collected from caregivers of 90 < 3year children from two urban clusters in Ahmedabad and Baroda, Gujarat. Target sample for pilot was selected after cluster census. Around half the mothers reported that they had not yet introduced water or other fluids to their < 6 month babies. However, about a third were already feeding them food other than mother’s milk. Although complementary feeding was initiated in almost all (95%) children more than 6 months old, frequency was suboptimal in 60%; in 80% cases no measure was taken to either improve energy or nutrient density; only 33% were fed protective foods; Green Leafy Vegetables consumption was negligible (1.4%). Anganwadi food was not consumed. By engaging ordinary people to generate evidence and understand the gaps, such assessments have the potential to be used to generate useful evidence for action at scale as well as locally.

Keywords: citizen led, grass root engagement, IYCF (Infant and Young Child Feeding), rapid diet assessment, under nutrition

Procedia PDF Downloads 155
24129 Design an Assessment Model of Research and Development Capabilities with the New Product Development Approach: A Case Study of Iran Khodro Company

Authors: Hamid Hanifi, Adel Azar, Alireza Booshehri

Abstract:

In order to know about the capability level of R & D units in automotive industry, it is essential that organizations always compare themselves with standard level and higher than themselves so that to be improved continuously. In this research, with respect to the importance of this issue, we have tried to present an assessment model for R & D capabilities having reviewed on new products development in automotive industry of Iran. Iran Khodro Company was selected for the case study. To this purpose, first, having a review on the literature, about 200 indicators effective in R & D capabilities and new products development were extracted. Then, of these numbers, 29 indicators which were more important were selected by industry and academia experts and the questionnaire was distributed among statistical population. Statistical population was consisted of 410 individuals in Iran Khodro Company. We used the 410 questionnaires for exploratory factor analysis and then used the data of 308 questionnaires from the same population randomly for confirmatory factor analysis. The results of exploratory factor analysis led to categorization of dimensions in 9 secondary dimensions. Naming the dimensions was done according to a literature review and the professors’ opinion. Using structural equation modeling and AMOS software, confirmatory factor analysis was conducted and ultimate model with 9 secondary dimensions was confirmed. Meanwhile, 9 secondary dimensions of this research are as follows: 1) Research and design capability, 2) Customer and market capability, 3) Technology capability, 4) Financial resources capability, 5) Organizational chart, 6) Intellectual capital capability, 7) NPD process capability, 8) Managerial capability and 9) Strategy capability.

Keywords: research and development, new products development, structural equations, exploratory factor analysis, confirmatory factor analysis

Procedia PDF Downloads 322
24128 RAPD Analysis of the Genetic Polymorphism in the Collection of Rye Cultivars

Authors: L. Petrovičová, Ž. Balážová, Z. Gálová, M. Wójcik-Jagła, M. Rapacz

Abstract:

In the present study, RAPD-PCR was used to assess genetic diversity of the rye including landrances and new rye cultivars coming from Central Europe and the Union of Soviet Socialist Republics (SUN). Five arbitrary random primers were used to determine RAPD polymorphism in the set of 38 rye genotypes. These primers amplified altogether 43 different DNA fragments with an average number of 8.6 fragments per genotypes. The number of fragments ranged from 7 (RLZ 8, RLZ 9 and RLZ 10) to 12 (RLZ 6). DI and PIC values of all RAPD markers were higher than 0.8 that generally means high level of polymorphism detected between rye genotypes. The dendrogram based on hierarchical cluster analysis using UPGMA algorithm was prepared. The cultivars were grouped into two main clusters. In this experiment, RAPD proved to be a rapid, reliable and practicable method for revealing of polymorphism in the rye cultivars.

Keywords: genetic diversity, polymorphism, RAPD markers, Secale cereale L.

Procedia PDF Downloads 428
24127 A PROMETHEE-BELIEF Approach for Multi-Criteria Decision Making Problems with Incomplete Information

Authors: H. Moalla, A. Frikha

Abstract:

Multi-criteria decision aid methods consider decision problems where numerous alternatives are evaluated on several criteria. These methods are used to deal with perfect information. However, in practice, it is obvious that this information requirement is too much strict. In fact, the imperfect data provided by more or less reliable decision makers usually affect decision results since any decision is closely linked to the quality and availability of information. In this paper, a PROMETHEE-BELIEF approach is proposed to help multi-criteria decisions based on incomplete information. This approach solves problems with incomplete decision matrix and unknown weights within PROMETHEE method. On the base of belief function theory, our approach first determines the distributions of belief masses based on PROMETHEE’s net flows and then calculates weights. Subsequently, it aggregates the distribution masses associated to each criterion using Murphy’s modified combination rule in order to infer a global belief structure. The final action ranking is obtained via pignistic probability transformation. A case study of real-world application concerning the location of a waste treatment center from healthcare activities with infectious risk in the center of Tunisia is studied to illustrate the detailed process of the BELIEF-PROMETHEE approach.

Keywords: belief function theory, incomplete information, multiple criteria analysis, PROMETHEE method

Procedia PDF Downloads 149
24126 Interoperability Maturity Models for Consideration When Using School Management Systems in South Africa: A Scoping Review

Authors: Keneilwe Maremi, Marlien Herselman, Adele Botha

Abstract:

The main purpose and focus of this paper are to determine the Interoperability Maturity Models to consider when using School Management Systems (SMS). The importance of this is to inform and help schools with knowing which Interoperability Maturity Model is best suited for their SMS. To address the purpose, this paper will apply a scoping review to ensure that all aspects are provided. The scoping review will include papers written from 2012-2019 and a comparison of the different types of Interoperability Maturity Models will be discussed in detail, which includes the background information, the levels of interoperability, and area for consideration in each Maturity Model. The literature was obtained from the following databases: IEEE Xplore and Scopus, the following search engines were used: Harzings, and Google Scholar. The topic of the paper was used as a search term for the literature and the term ‘Interoperability Maturity Models’ was used as a keyword. The data were analyzed in terms of the definition of Interoperability, Interoperability Maturity Models, and levels of interoperability. The results provide a table that shows the focus area of concern for each Maturity Model (based on the scoping review where only 24 papers were found to be best suited for the paper out of 740 publications initially identified in the field). This resulted in the most discussed Interoperability Maturity Model for consideration (Information Systems Interoperability Maturity Model (ISIMM) and Organizational Interoperability Maturity Model for C2 (OIM)).

Keywords: interoperability, interoperability maturity model, school management system, scoping review

Procedia PDF Downloads 183
24125 Co-integration for Soft Commodities with Non-Constant Volatility

Authors: E. Channol, O. Collet, N. Kostyuchyk, T. Mesbah, Quoc Hoang Long Nguyen

Abstract:

In this paper, a pricing model is proposed for co-integrated commodities extending Larsson model. The futures formulae have been derived and tests have been performed with non-constant volatility. The model has been applied to energy commodities (gas, CO2, energy) and soft commodities (corn, wheat). Results show that non-constant volatility leads to more accurate short term prices, which provides better evaluation of value-at-risk and more generally improve the risk management.

Keywords: co-integration, soft commodities, risk management, value-at-risk

Procedia PDF Downloads 533
24124 Developing a GIS-Based Tool for the Management of Fats, Oils, and Grease (FOG): A Case Study of Thames Water Wastewater Catchment

Authors: Thomas D. Collin, Rachel Cunningham, Bruce Jefferson, Raffaella Villa

Abstract:

Fats, oils and grease (FOG) are by-products of food preparation and cooking processes. FOG enters wastewater systems through a variety of sources such as households, food service establishments, and industrial food facilities. Over time, if no source control is in place, FOG builds up on pipe walls, leading to blockages, and potentially to sewer overflows which are a major risk to the Environment and Human Health. UK water utilities spend millions of pounds annually trying to control FOG. Despite UK legislation specifying that discharge of such material is against the law, it is often complicated for water companies to identify and prosecute offenders. Hence, it leads to uncertainties regarding the attitude to take in terms of FOG management. Research is needed to seize the full potential of implementing current practices. The aim of this research was to undertake a comprehensive study to document the extent of FOG problems in sewer lines and reinforce existing knowledge. Data were collected to develop a model estimating quantities of FOG available for recovery within Thames Water wastewater catchments. Geographical Information System (GIS) software was used in conjunction to integrate data with a geographical component. FOG was responsible for at least 1/3 of sewer blockages in Thames Water waste area. A waste-based approach was developed through an extensive review to estimate the potential for FOG collection and recovery. Three main sources were identified: residential, commercial and industrial. Commercial properties were identified as one of the major FOG producers. The total potential FOG generated was estimated for the 354 wastewater catchments. Additionally, raw and settled sewage were sampled and analysed for FOG (as hexane extractable material) monthly at 20 sewage treatment works (STW) for three years. A good correlation was found with the sampled FOG and population equivalent (PE). On average, a difference of 43.03% was found between the estimated FOG (waste-based approach) and sampled FOG (raw sewage sampling). It was suggested that the approach undertaken could overestimate the FOG available, the sampling could only capture a fraction of FOG arriving at STW, and/or the difference could account for FOG accumulating in sewer lines. Furthermore, it was estimated that on average FOG could contribute up to 12.99% of the primary sludge removed. The model was further used to investigate the relationship between estimated FOG and number of blockages. The higher the FOG potential, the higher the number of FOG-related blockages is. The GIS-based tool was used to identify critical areas (i.e. high FOG potential and high number of FOG blockages). As reported in the literature, FOG was one of the main causes of sewer blockages. By identifying critical areas (i.e. high FOG potential and high number of FOG blockages) the model further explored the potential for source-control in terms of ‘sewer relief’ and waste recovery. Hence, it helped targeting where benefits from implementation of management strategies could be the highest. However, FOG is still likely to persist throughout the networks, and further research is needed to assess downstream impacts (i.e. at STW).

Keywords: fat, FOG, GIS, grease, oil, sewer blockages, sewer networks

Procedia PDF Downloads 192
24123 Modeling Sustainable Truck Rental Operations Using Closed-Loop Supply Chain Network

Authors: Khaled S. Abdallah, Abdel-Aziz M. Mohamed

Abstract:

Moving industries consume numerous resources and dispose masses of used packaging materials. Proper sorting, recycling and disposing the packaging materials is necessary to avoid a sever pollution disaster. This research paper presents a conceptual model to propose sustainable truck rental operations instead of the regular one. An optimization model was developed to select the locations of truck rental centers, collection sites, maintenance and repair sites, and identify the rental fees to be charged for all routes that maximize the total closed supply chain profits. Fixed costs of vehicle purchasing, costs of constructing collection centers and repair centers, as well as the fixed costs paid to use disposal and recycling centers are considered. Operating costs include the truck maintenance, repair costs as well as the cost of recycling and disposing the packing materials, and the costs of relocating the truck are presented in the model. A mixed integer model is developed followed by a simulation model to examine the factors affecting the operation of the model.

Keywords: modeling, truck rental, supply chains management.

Procedia PDF Downloads 215
24122 Analyzing Conflict Text; ‘Akunyili Memo: State of the Nation’: an Approach from CDA

Authors: Nengi A. H. Ejiobih

Abstract:

Conflict is one of the defining features of human societies. Often, the use or misuse of language in interaction is the genesis of conflict. As such, it is expected that when people use language they do so in socially determined ways and with almost predictable social effects. The objective of this paper was to examine the interest at work as manifested in language choice and collocations in conflict discourse. It also scrutinized the implications of linguistic features in conflict discourse as it concerns ideology and power relations in political discourse in Nigeria. The methodology used for this paper is an approach from Critical discourse analysis because of its multidisciplinary model of analysis, linguistic features and its implications were analysed. The datum used is a text from the Sunday Sun Newspaper in Nigeria, West Africa titled Akunyili Memo: State of the Nation. Some of the findings include; different ideologies are inherent in conflict discourse, there is the presence of power relations being produced, exercised, maintained and produced throughout the discourse and the use of pronouns in conflict discourse is valuable because it is used to initiate and maintain relationships in social context. This paper has provided evidence that, taking into consideration the nature of the social actions and the way these activities are translated into languages, the meanings people convey by their words are identified by their immediate social, political and historical conditions.

Keywords: conflicts, discourse, language, linguistic features, social context

Procedia PDF Downloads 458
24121 Developing an Out-of-Distribution Generalization Model Selection Framework through Impurity and Randomness Measurements and a Bias Index

Authors: Todd Zhou, Mikhail Yurochkin

Abstract:

Out-of-distribution (OOD) detection is receiving increasing amounts of attention in the machine learning research community, boosted by recent technologies, such as autonomous driving and image processing. This newly-burgeoning field has called for the need for more effective and efficient methods for out-of-distribution generalization methods. Without accessing the label information, deploying machine learning models to out-of-distribution domains becomes extremely challenging since it is impossible to evaluate model performance on unseen domains. To tackle this out-of-distribution detection difficulty, we designed a model selection pipeline algorithm and developed a model selection framework with different impurity and randomness measurements to evaluate and choose the best-performing models for out-of-distribution data. By exploring different randomness scores based on predicted probabilities, we adopted the out-of-distribution entropy and developed a custom-designed score, ”CombinedScore,” as the evaluation criterion. This proposed score was created by adding labeled source information into the judging space of the uncertainty entropy score using harmonic mean. Furthermore, the prediction bias was explored through the equality of opportunity violation measurement. We also improved machine learning model performance through model calibration. The effectiveness of the framework with the proposed evaluation criteria was validated on the Folktables American Community Survey (ACS) datasets.

Keywords: model selection, domain generalization, model fairness, randomness measurements, bias index

Procedia PDF Downloads 111
24120 An Improved Dynamic Window Approach with Environment Awareness for Local Obstacle Avoidance of Mobile Robots

Authors: Baoshan Wei, Shuai Han, Xing Zhang

Abstract:

Local obstacle avoidance is critical for mobile robot navigation. It is a challenging task to ensure path optimality and safety in cluttered environments. We proposed an Environment Aware Dynamic Window Approach in this paper to cope with the issue. The method integrates environment characterization into Dynamic Window Approach (DWA). Two strategies are proposed in order to achieve the integration. The local goal strategy guides the robot to move through openings before approaching the final goal, which solves the local minima problem in DWA. The adaptive control strategy endows the robot to adjust its state according to the environment, which addresses path safety compared with DWA. Besides, the evaluation shows that the path generated from the proposed algorithm is safer and smoother compared with state-of-the-art algorithms.

Keywords: adaptive control, dynamic window approach, environment aware, local obstacle avoidance, mobile robots

Procedia PDF Downloads 136
24119 Model Based Fault Diagnostic Approach for Limit Switches

Authors: Zafar Mahmood, Surayya Naz, Nazir Shah Khattak

Abstract:

The degree of freedom relates to our capability to observe or model the energy paths within the system. Higher the number of energy paths being modeled leaves to us a higher degree of freedom, but increasing the time and modeling complexity rendering it useless for today’s world’s need for minimum time to market. Since the number of residuals that can be uniquely isolated are dependent on the number of independent outputs of the system, increasing the number of sensors required. The examples of discrete position sensors that may be used to form an array include limit switches, Hall effect sensors, optical sensors, magnetic sensors, etc. Their mechanical design can usually be tailored to fit in the transitional path of an STME in a variety of mechanical configurations. The case studies into multi-sensor system were carried out and actual data from sensors is used to test this generic framework. It is being investigated, how the proper modeling of limit switches as timing sensors, could lead to unified and neutral residual space while keeping the implementation cost reasonably low.

Keywords: low-cost limit sensors, fault diagnostics, Single Throw Mechanical Equipment (STME), parameter estimation, parity-space

Procedia PDF Downloads 593
24118 Evaluation of Biochemical Oxygen Demand and Dissolved Oxygen for Thames River by Using Stream Water Quality Model

Authors: Ghassan Al-Dulaimi

Abstract:

This paper studied the biochemical parameter (BOD5) and (DO) for the Thames River (Canada-Ontario). Water samples have been collected from Thames River along different points between Chatham to Woodstock and were analysed for various water quality parameters during the low flow season (April). The study involves the application of the stream water quality model QUAL2K model to simulate and predict the dissolved oxygen (DO) and biochemical oxygen demand (BOD5) profiles for Thames River in a stretch of 251 kilometers. The model output showed that DO in the entire river was within the limit of not less than 4 mg/L. For Carbonaceous Biochemical Oxygen Demand CBOD, the entire river may be divided into two main reaches; the first one is extended from Chatham City (0 km) to London (150 km) and has a CBOD concentration of 2 mg/L, and the second reach has CBOD range (2–4) mg/L in which begins from London city and extend to near Woodstock city (73km).

Keywords: biochemical oxygen demand, dissolved oxygen, Thames river, QUAL2K model

Procedia PDF Downloads 78
24117 Glorification Trap in Combating Human Trafficking in Indonesia: An Application of Three-Dimensional Model of Anti-Trafficking Policy

Authors: M. Kosandi, V. Susanti, N. I. Subono, E. Kartini

Abstract:

This paper discusses the risk of glorification trap in combating human trafficking, as it is shown in the case of Indonesia. Based on a research on Indonesian combat against trafficking in 2017-2018, this paper shows the tendency of misinterpretation and misapplication of the Indonesian anti-trafficking law into misusing the law for glorification, to create an image of certain extent of achievement in combating human trafficking. The objective of this paper is to explain the persistent occurrence of human trafficking crimes despite the significant progress of anti-trafficking efforts of Indonesian government. The research was conducted in 2017-2018 by qualitative approach through observation, depth interviews, discourse analysis, and document study, applying the three-dimensional model for analyzing human trafficking in the source country. This paper argues that the drive for glorification of achievement in the combat against trafficking has trapped Indonesian government in the loop of misinterpretation, misapplication, and misuse of the anti-trafficking law. In return, the so-called crime against humanity remains high and tends to increase in Indonesia.

Keywords: human trafficking, anti-trafficking policy, transnational crime, source country, glorification trap

Procedia PDF Downloads 153
24116 Diabetes Diagnosis Model Using Rough Set and K- Nearest Neighbor Classifier

Authors: Usiobaifo Agharese Rosemary, Osaseri Roseline Oghogho

Abstract:

Diabetes is a complex group of disease with a variety of causes; it is a disorder of the body metabolism in the digestion of carbohydrates food. The application of machine learning in the field of medical diagnosis has been the focus of many researchers and the use of recognition and classification model as a decision support tools has help the medical expert in diagnosis of diseases. Considering the large volume of medical data which require special techniques, experience, and high diagnostic skill in the diagnosis of diseases, the application of an artificial intelligent system to assist medical personnel in order to enhance their efficiency and accuracy in diagnosis will be an invaluable tool. In this study will propose a diabetes diagnosis model using rough set and K-nearest Neighbor classifier algorithm. The system consists of two modules: the feature extraction module and predictor module, rough data set is used to preprocess the attributes while K-nearest neighbor classifier is used to classify the given data. The dataset used for this model was taken for University of Benin Teaching Hospital (UBTH) database. Half of the data was used in the training while the other half was used in testing the system. The proposed model was able to achieve over 80% accuracy.

Keywords: classifier algorithm, diabetes, diagnostic model, machine learning

Procedia PDF Downloads 323
24115 The Use of Boosted Multivariate Trees in Medical Decision-Making for Repeated Measurements

Authors: Ebru Turgal, Beyza Doganay Erdogan

Abstract:

Machine learning aims to model the relationship between the response and features. Medical decision-making researchers would like to make decisions about patients’ course and treatment, by examining the repeated measurements over time. Boosting approach is now being used in machine learning area for these aims as an influential tool. The aim of this study is to show the usage of multivariate tree boosting in this field. The main reason for utilizing this approach in the field of decision-making is the ease solutions of complex relationships. To show how multivariate tree boosting method can be used to identify important features and feature-time interaction, we used the data, which was collected retrospectively from Ankara University Chest Diseases Department records. Dataset includes repeated PF ratio measurements. The follow-up time is planned for 120 hours. A set of different models is tested. In conclusion, main idea of classification with weighed combination of classifiers is a reliable method which was shown with simulations several times. Furthermore, time varying variables will be taken into consideration within this concept and it could be possible to make accurate decisions about regression and survival problems.

Keywords: boosted multivariate trees, longitudinal data, multivariate regression tree, panel data

Procedia PDF Downloads 190
24114 Applying Sliding Autonomy for a Human-Robot Team on USARSim

Authors: Fang Tang, Jacob Longazo

Abstract:

This paper describes a sliding autonomy approach for coordinating a team of robots to assist the human operator to accomplish tasks while adapting to new or unexpected situations by requesting help from the human operator. While sliding autonomy has been well studied in the context of controlling a single robot. Much work needs to be done to apply sliding autonomy to a multi-robot team, especially human-robot team. Our approach aims at a hierarchical sliding control structure, with components that support human-robot collaboration. We validated our approach in the USARSim simulation and demonstrated that the human-robot team's overall performance can be improved under the sliding autonomy control.

Keywords: sliding autonomy, multi-robot team, human-robot collaboration, USARSim

Procedia PDF Downloads 528
24113 Validation Study of Radial Aircraft Engine Model

Authors: Lukasz Grabowski, Tytus Tulwin, Michal Geca, P. Karpinski

Abstract:

This paper presents the radial aircraft engine model which has been created in AVL Boost software. This model is a one-dimensional physical model of the engine, which enables us to investigate the impact of an ignition system design on engine performance (power, torque, fuel consumption). In addition, this model allows research under variable environmental conditions to reflect varied flight conditions (altitude, humidity, cruising speed). Before the simulation research the identifying parameters and validating of model were studied. In order to verify the feasibility to take off power of gasoline radial aircraft engine model, some validation study was carried out. The first stage of the identification was completed with reference to the technical documentation provided by manufacturer of engine and the experiments on the test stand of the real engine. The second stage involved a comparison of simulation results with the results of the engine stand tests performed on a WSK ’PZL-Kalisz’. The engine was loaded by a propeller in a special test bench. Identifying the model parameters referred to a comparison of the test results to the simulation in terms of: pressure behind the throttles, pressure in the inlet pipe, and time course for pressure in the first inlet pipe, power, and specific fuel consumption. Accordingly, the required coefficients and error of simulation calculation relative to the real-object experiments were determined. Obtained the time course for pressure and its value is compatible with the experimental results. Additionally the engine power and specific fuel consumption tends to be significantly compatible with the bench tests. The mapping error does not exceed 1.5%, which verifies positively the model of combustion and allows us to predict engine performance if the process of combustion will be modified. The next conducted tests verified completely model. The maximum mapping error for the pressure behind the throttles and the inlet pipe pressure is 4 %, which proves the model of the inlet duct in the engine with the charging compressor to be correct.

Keywords: 1D-model, aircraft engine, performance, validation

Procedia PDF Downloads 322
24112 Evaluation of Low-Reducible Sinter in Blast Furnace Technology by Mathematical Model Developed at Centre ENET, VSB: Technical University of Ostrava

Authors: S. Jursová, P. Pustějovská, S. Brožová, J. Bilík

Abstract:

The paper deals with possibilities of interpretation of iron ore reducibility tests. It presents a mathematical model developed at Centre ENET, VŠB–Technical University of Ostrava, Czech Republic for an evaluation of metallurgical material of blast furnace feedstock such as iron ore, sinter or pellets. According to the data from the test, the model predicts its usage in blast furnace technology and its effects on production parameters of shaft aggregate. At the beginning, the paper sums up the general concept and experience in mathematical modelling of iron ore reduction. It presents basic equation for the calculation and the main parts of the developed model. In the experimental part, there is an example of usage of the mathematical model. The paper describes the usage of data for some predictive calculation. There are presented material, method of carried test of iron ore reducibility. Then there are graphically interpreted effects of used material on carbon consumption, rate of direct reduction and the whole reduction process.

Keywords: blast furnace technology, iron ore reduction, mathematical model, prediction of iron ore reduction

Procedia PDF Downloads 659
24111 Early Diagnosis of Myocardial Ischemia Based on Support Vector Machine and Gaussian Mixture Model by Using Features of ECG Recordings

Authors: Merve Begum Terzi, Orhan Arikan, Adnan Abaci, Mustafa Candemir

Abstract:

Acute myocardial infarction is a major cause of death in the world. Therefore, its fast and reliable diagnosis is a major clinical need. ECG is the most important diagnostic methodology which is used to make decisions about the management of the cardiovascular diseases. In patients with acute myocardial ischemia, temporary chest pains together with changes in ST segment and T wave of ECG occur shortly before the start of myocardial infarction. In this study, a technique which detects changes in ST/T sections of ECG is developed for the early diagnosis of acute myocardial ischemia. For this purpose, a database of real ECG recordings that contains a set of records from 75 patients presenting symptoms of chest pain who underwent elective percutaneous coronary intervention (PCI) is constituted. 12-lead ECG’s of the patients were recorded before and during the PCI procedure. Two ECG epochs, which are the pre-inflation ECG which is acquired before any catheter insertion and the occlusion ECG which is acquired during balloon inflation, are analyzed for each patient. By using pre-inflation and occlusion recordings, ECG features that are critical in the detection of acute myocardial ischemia are identified and the most discriminative features for the detection of acute myocardial ischemia are extracted. A classification technique based on support vector machine (SVM) approach operating with linear and radial basis function (RBF) kernels to detect ischemic events by using ST-T derived joint features from non-ischemic and ischemic states of the patients is developed. The dataset is randomly divided into training and testing sets and the training set is used to optimize SVM hyperparameters by using grid-search method and 10fold cross-validation. SVMs are designed specifically for each patient by tuning the kernel parameters in order to obtain the optimal classification performance results. As a result of implementing the developed classification technique to real ECG recordings, it is shown that the proposed technique provides highly reliable detections of the anomalies in ECG signals. Furthermore, to develop a detection technique that can be used in the absence of ECG recording obtained during healthy stage, the detection of acute myocardial ischemia based on ECG recordings of the patients obtained during ischemia is also investigated. For this purpose, a Gaussian mixture model (GMM) is used to represent the joint pdf of the most discriminating ECG features of myocardial ischemia. Then, a Neyman-Pearson type of approach is developed to provide detection of outliers that would correspond to acute myocardial ischemia. Neyman – Pearson decision strategy is used by computing the average log likelihood values of ECG segments and comparing them with a range of different threshold values. For different discrimination threshold values and number of ECG segments, probability of detection and probability of false alarm values are computed, and the corresponding ROC curves are obtained. The results indicate that increasing number of ECG segments provide higher performance for GMM based classification. Moreover, the comparison between the performances of SVM and GMM based classification showed that SVM provides higher classification performance results over ECG recordings of considerable number of patients.

Keywords: ECG classification, Gaussian mixture model, Neyman–Pearson approach, support vector machine

Procedia PDF Downloads 145
24110 A Model for Operating Rooms Scheduling

Authors: Jose Francisco Ferreira Ribeiro, Alexandre Bevilacqua Leoneti, Andre Lucirton Costa

Abstract:

This paper presents a mathematical model in binary variables 0/1 to make the assignment of surgical procedures to the operating rooms in a hospital. The proposed mathematical model is based on the generalized assignment problem, which maximizes the sum of preferences for the use of the operating rooms by doctors, respecting the time available in each room. The corresponding program was written in Visual Basic of Microsoft Excel, and tested to schedule surgeries at St. Lydia Hospital in Ribeirao Preto, Brazil.

Keywords: generalized assignment problem, logistics, optimization, scheduling

Procedia PDF Downloads 271
24109 A Study of a Diachronic Relationship between Two Weak Inflection Classes in Norwegian, with Emphasis on Unexpected Productivity

Authors: Emilija Tribocka

Abstract:

This contribution presents parts of an ongoing study of a diachronic relationship between two weak verb classes in Norwegian, the a-class (cf. the paradigm of ‘throw’: kasta – kastar – kasta – kasta) and the e-class (cf. the paradigm of ‘buy’: kjøpa – kjøper – kjøpte – kjøpt). The study investigates inflection class shifts between the two classes with Old Norse, the ancestor of Modern Norwegian, as a starting point. Examination of inflection in 38 verbs in four chosen dialect areas (106 places of attestations) demonstrates that the shifts from the a-class to the e-class are widespread to varying degrees in three out of four investigated areas and are more common than the shifts in the opposite direction. The diachronic productivity of the e-class is unexpected for several reasons. There is general agreement that type frequency is an important factor influencing productivity. The a-class (53% of all weak verbs) was more type frequent in Old Norse than the e-class (42% of all weak verbs). Thus, given the type frequency, the expansion of the e-class is unexpected. Furthermore, in the ‘core’ areas of expanded e-class inflection, the shifts disregard phonological principles creating forms with uncomfortable consonant clusters, e.g., fiskte instead of fiska, the preterit of fiska ‘fish’. Later on, these forms may be contracted, i.e., fiskte > fiste. In this contribution, two factors influencing the shifts are presented: phonological form and token frequency. Verbs with the stem ending in a consonant cluster, particularly when the cluster ends in -t, hardly ever shift to the e-class. As a matter of fact, verbs with this structure belonging to the e-class in Old Norse shift to the a-class in Modern Norwegian, e.g., ON e-class verb skipta ‘change’ shifts to the a-class. This shift occurs as a result of the lack of morpho-phonological transparency between the stem and the preterit suffix of the e-class, -te. As there is a phonological fusion between the stem ending in -t and the suffix beginning in -t, the transparent a-class inflection is chosen. Token frequency plays an important role in the shifts, too, in some dialects. In one of the investigated areas, the most token frequent verbs of the ON e-class remain in the e-class (e.g., høyra ‘hear’, leva ‘live’, kjøpa ‘buy’), while less frequent verbs may shift to the a-class. Furthermore, the results indicate that the shift from the a-class to the e-class occurs in some of the most token frequent verbs of the ON a-class in this area, e.g., lika ‘like’, lova ‘promise’, svara ‘answer’. The latter is unexpected as frequent items tend to remain stable. This study presents a case of unexpected productivity, demonstrating that minor patterns can grow and outdo major patterns. Thus, type frequency is not the only factor that determines productivity. The study addresses the role of phonological form and token frequency in the spread of inflection patterns.

Keywords: inflection class, productivity, token frequency, phonological form

Procedia PDF Downloads 46
24108 Building an Absurdist Approach to the Philosophy of Science: Combining Camus and Feyerabend

Authors: Robert Herold

Abstract:

This project aims to begin building out a new approach within the philosophy of science that is based around a combination of insights from Albert Camus and Paul Feyerabend. This approach is one that will be labeled an absurdist approach as it uses, for its foundation, the philosophy of the absurd as discussed by Camus. While Camus didn’t directly discuss the philosophy of science, nor did he offer his own views on the subject in any substantial way, that doesn’t mean that his work doesn’t have applications within the philosophy of science. In fact, as is argued throughout the piece, much of the work done by Paul Feyerabend stems from a similar metaphysical and epistemological foundation as Camus. This foundation is the notion of the absurd and the inability of us as humans to reach some sort of objective truth. In modern times both Camus and Feyerabend have been largely pushed to the wayside, though Feyerabend has undoubtedly received the most unfair treatment of the two, and this is something that serves to act more as a hindrance than anything else. Much of the claims and arguments made by both Camus and Feyerabend have not been truly refuted and have simply been pushed aside by pointing to supposed contradictions or inconsistencies. However, while it would be a monumental task to attempt to discuss all of this past work, perhaps it might be better to move beyond both Camus and Feyerabend and chart a new path. This is the overall goal of this paper. This research will demonstrate that not only are the philosophies of Camus and Feyerabend surprisingly similar and able to mesh well together, they also are able to form into something that is truly more than the sum of its parts. While the task of actually building out an approach is a monumental undertaking, the plan is to use this project as a jumping-off point. As such, this paper will start by examining some of the main claims made by both Camus and Feyerabend. Once this is done, then begin weaving them together and demonstrating where the links between the philosophies of both are. Then this study will end by building out the very begging foundations of the absurdist approach to the philosophy of science.

Keywords: philosophy, philosophy of science, albert camus, paul feyerabend

Procedia PDF Downloads 230
24107 Intelligent Indoor Localization Using WLAN Fingerprinting

Authors: Gideon C. Joseph

Abstract:

The ability to localize mobile devices is quite important, as some applications may require location information of these devices to operate or deliver better services to the users. Although there are several ways of acquiring location data of mobile devices, the WLAN fingerprinting approach has been considered in this work. This approach uses the Received Signal Strength Indicator (RSSI) measurement as a function of the position of the mobile device. RSSI is a quantitative technique of describing the radio frequency power carried by a signal. RSSI may be used to determine RF link quality and is very useful in dense traffic scenarios where interference is of major concern, for example, indoor environments. This research aims to design a system that can predict the location of a mobile device, when supplied with the mobile’s RSSIs. The developed system takes as input the RSSIs relating to the mobile device, and outputs parameters that describe the location of the device such as the longitude, latitude, floor, and building. The relationship between the Received Signal Strengths (RSSs) of mobile devices and their corresponding locations is meant to be modelled; hence, subsequent locations of mobile devices can be predicted using the developed model. It is obvious that describing mathematical relationships between the RSSIs measurements and localization parameters is one option to modelling the problem, but the complexity of such an approach is a serious turn-off. In contrast, we propose an intelligent system that can learn the mapping of such RSSIs measurements to the localization parameters to be predicted. The system is capable of upgrading its performance as more experiential knowledge is acquired. The most appealing consideration to using such a system for this task is that complicated mathematical analysis and theoretical frameworks are excluded or not needed; the intelligent system on its own learns the underlying relationship in the supplied data (RSSI levels) that corresponds to the localization parameters. These localization parameters to be predicted are of two different tasks: Longitude and latitude of mobile devices are real values (regression problem), while the floor and building of the mobile devices are of integer values or categorical (classification problem). This research work presents artificial neural network based intelligent systems to model the relationship between the RSSIs predictors and the mobile device localization parameters. The designed systems were trained and validated on the collected WLAN fingerprint database. The trained networks were then tested with another supplied database to obtain the performance of trained systems on achieved Mean Absolute Error (MAE) and error rates for the regression and classification tasks involved therein.

Keywords: indoor localization, WLAN fingerprinting, neural networks, classification, regression

Procedia PDF Downloads 328