Search results for: q-rung orthopair fuzzy sets
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1071

Search results for: q-rung orthopair fuzzy sets

531 Hardware Implementation on Field Programmable Gate Array of Two-Stage Algorithm for Rough Set Reduct Generation

Authors: Tomasz Grzes, Maciej Kopczynski, Jaroslaw Stepaniuk

Abstract:

The rough sets theory developed by Prof. Z. Pawlak is one of the tools that can be used in the intelligent systems for data analysis and processing. Banking, medicine, image recognition and security are among the possible fields of utilization. In all these fields, the amount of the collected data is increasing quickly, but with the increase of the data, the computation speed becomes the critical factor. Data reduction is one of the solutions to this problem. Removing the redundancy in the rough sets can be achieved with the reduct. A lot of algorithms of generating the reduct were developed, but most of them are only software implementations, therefore have many limitations. Microprocessor uses the fixed word length, consumes a lot of time for either fetching as well as processing of the instruction and data; consequently, the software based implementations are relatively slow. Hardware systems don’t have these limitations and can process the data faster than a software. Reduct is the subset of the decision attributes that provides the discernibility of the objects. For the given decision table there can be more than one reduct. Core is the set of all indispensable condition attributes. None of its elements can be removed without affecting the classification power of all condition attributes. Moreover, every reduct consists of all the attributes from the core. In this paper, the hardware implementation of the two-stage greedy algorithm to find the one reduct is presented. The decision table is used as an input. Output of the algorithm is the superreduct which is the reduct with some additional removable attributes. First stage of the algorithm is calculating the core using the discernibility matrix. Second stage is generating the superreduct by enriching the core with the most common attributes, i.e., attributes that are more frequent in the decision table. Described above algorithm has two disadvantages: i) generating the superreduct instead of reduct, ii) additional first stage may be unnecessary if the core is empty. But for the systems focused on the fast computation of the reduct the first disadvantage is not the key problem. The core calculation can be achieved with a combinational logic block, and thus add respectively little time to the whole process. Algorithm presented in this paper was implemented in Field Programmable Gate Array (FPGA) as a digital device consisting of blocks that process the data in a single step. Calculating the core is done by the comparators connected to the block called 'singleton detector', which detects if the input word contains only single 'one'. Calculating the number of occurrences of the attribute is performed in the combinational block made up of the cascade of the adders. The superreduct generation process is iterative and thus needs the sequential circuit for controlling the calculations. For the research purpose, the algorithm was also implemented in C language and run on a PC. The times of execution of the reduct calculation in a hardware and software were considered. Results show increase in the speed of data processing.

Keywords: data reduction, digital systems design, field programmable gate array (FPGA), reduct, rough set

Procedia PDF Downloads 205
530 Variations in the Frequency-Magnitude Distribution with Depth in Kalabsha Area, Aswan, South Egypt

Authors: Ezzat Mohamed El-Amin

Abstract:

Mapping the earthquake-size distribution in various tectonic regimes on a local to regional scale reveals statistically significant variations in the range of at least 0.4 to 2.0 for the b-value in the frequency-magnitude distribution. We map the earthquake frequency–magnitude distribution (b value) as a function of depth in the Reservoir Triggered Seismicity (RTS) region in Kalabsha region, in south Egypt. About 1680 well-located events recorded during 1981–2014 in the Kalabsha region are selected for the analysis. The earthquake data sets are separated in 5 km zones from 0 to 25 km depth. The result shows a systematic decrease in b value up to 12 km followed by an increase. The increase in b value is interpreted to be caused by the presence of fluids. We also investigate the spatial distribution of b value with depth. Significant variations in the b value are detected, with b ranging from b 0.7 to 1.19. Low b value areas at 5 km depth indicate localized high stresses which are favorable for future rupture.

Keywords: seismicity, frequency-magnitude, b-value, earthquake

Procedia PDF Downloads 547
529 Undirected Endo-Cayley Digraphs of Cyclic Groups of Order Primes

Authors: Chanon Promsakon, Sayan Panma

Abstract:

Let S be a finite semigroup, A a subset of S and f an endomorphism on S. The endo-Cayley digraph of a semigroup S corresponding to a connecting set A and an endomorphism f, denoted by endo − Cayf (S, A) is a digraph whose vertex set is S and a vertex u is adjacent to a vertex v if and only if v = f(u)a for some a ∈ A. A digraph D is called undirected if any edge uv in D, there exists an edge vu in D. We consider the undirectedness of an endo-Cayley of a cyclic group of order prime, Zp. In this work, we investigate conditions for connecting sets and endomorphisms to make endo-Cayley digraphs of cyclic groups of order primes be undirected. Moreover, we give some conditions for an undirected endo-Cayley of cycle group of any order.

Keywords: endo-Cayley graph, undirected digraphs, cyclic groups, endomorphism

Procedia PDF Downloads 330
528 A Multi-Agent Intelligent System for Monitoring Health Conditions of Elderly People

Authors: Ayman M. Mansour

Abstract:

In this paper, we propose a multi-agent intelligent system that is used for monitoring the health conditions of elderly people. Monitoring the health condition of elderly people is a complex problem that involves different medical units and requires continuous monitoring. Such expert system is highly needed in rural areas because of inadequate number of available specialized physicians or nurses. Such monitoring must have autonomous interactions between these medical units in order to be effective. A multi-agent system is formed by a community of agents that exchange information and proactively help one another to achieve the goal of elderly monitoring. The agents in the developed system are equipped with intelligent decision maker that arms them with the rule-based reasoning capability that can assist the physicians in making decisions regarding the medical condition of elderly people.

Keywords: fuzzy logic, inference system, monitoring system, multi-agent system

Procedia PDF Downloads 584
527 Solving Flowshop Scheduling Problems with Ant Colony Optimization Heuristic

Authors: Arshad Mehmood Ch, Riaz Ahmad, Imran Ali Ch, Waqas Durrani

Abstract:

This study deals with the application of Ant Colony Optimization (ACO) approach to solve no-wait flowshop scheduling problem (NW-FSSP). ACO algorithm so developed has been coded on Matlab computer application. The paper covers detailed steps to apply ACO and focuses on judging the strength of ACO in relation to other solution techniques previously applied to solve no-wait flowshop problem. The general purpose approach was able to find reasonably accurate solutions for almost all the problems under consideration and was able to handle a fairly large spectrum of problems with far reduced CPU effort. Careful scrutiny of the results reveals that the algorithm presented results better than other approaches like Genetic algorithm and Tabu Search heuristics etc; earlier applied to solve NW-FSSP data sets.

Keywords: no-wait, flowshop, scheduling, ant colony optimization (ACO), makespan

Procedia PDF Downloads 421
526 Performance Analysis and Optimization for Diagonal Sparse Matrix-Vector Multiplication on Machine Learning Unit

Authors: Qiuyu Dai, Haochong Zhang, Xiangrong Liu

Abstract:

Diagonal sparse matrix-vector multiplication is a well-studied topic in the fields of scientific computing and big data processing. However, when diagonal sparse matrices are stored in DIA format, there can be a significant number of padded zero elements and scattered points, which can lead to a degradation in the performance of the current DIA kernel. This can also lead to excessive consumption of computational and memory resources. In order to address these issues, the authors propose the DIA-Adaptive scheme and its kernel, which leverages the parallel instruction sets on MLU. The researchers analyze the effect of allocating a varying number of threads, clusters, and hardware architectures on the performance of SpMV using different formats. The experimental results indicate that the proposed DIA-Adaptive scheme performs well and offers excellent parallelism.

Keywords: adaptive method, DIA, diagonal sparse matrices, MLU, sparse matrix-vector multiplication

Procedia PDF Downloads 109
525 Evolving Credit Scoring Models using Genetic Programming and Language Integrated Query Expression Trees

Authors: Alexandru-Ion Marinescu

Abstract:

There exist a plethora of methods in the scientific literature which tackle the well-established task of credit score evaluation. In its most abstract form, a credit scoring algorithm takes as input several credit applicant properties, such as age, marital status, employment status, loan duration, etc. and must output a binary response variable (i.e. “GOOD” or “BAD”) stating whether the client is susceptible to payment return delays. Data imbalance is a common occurrence among financial institution databases, with the majority being classified as “GOOD” clients (clients that respect the loan return calendar) alongside a small percentage of “BAD” clients. But it is the “BAD” clients we are interested in since accurately predicting their behavior is crucial in preventing unwanted loss for loan providers. We add to this whole context the constraint that the algorithm must yield an actual, tractable mathematical formula, which is friendlier towards financial analysts. To this end, we have turned to genetic algorithms and genetic programming, aiming to evolve actual mathematical expressions using specially tailored mutation and crossover operators. As far as data representation is concerned, we employ a very flexible mechanism – LINQ expression trees, readily available in the C# programming language, enabling us to construct executable pieces of code at runtime. As the title implies, they model trees, with intermediate nodes being operators (addition, subtraction, multiplication, division) or mathematical functions (sin, cos, abs, round, etc.) and leaf nodes storing either constants or variables. There is a one-to-one correspondence between the client properties and the formula variables. The mutation and crossover operators work on a flattened version of the tree, obtained via a pre-order traversal. A consequence of our chosen technique is that we can identify and discard client properties which do not take part in the final score evaluation, effectively acting as a dimensionality reduction scheme. We compare ourselves with state of the art approaches, such as support vector machines, Bayesian networks, and extreme learning machines, to name a few. The data sets we benchmark against amount to a total of 8, of which we mention the well-known Australian credit and German credit data sets, and the performance indicators are the following: percentage correctly classified, area under curve, partial Gini index, H-measure, Brier score and Kolmogorov-Smirnov statistic, respectively. Finally, we obtain encouraging results, which, although placing us in the lower half of the hierarchy, drive us to further refine the algorithm.

Keywords: expression trees, financial credit scoring, genetic algorithm, genetic programming, symbolic evolution

Procedia PDF Downloads 104
524 Students' Perception of Using Dental E-Models in an Inquiry-Based Curriculum

Authors: Yanqi Yang, Chongshan Liao, Cheuk Hin Ho, Susan Bridges

Abstract:

Aim: To investigate student’s perceptions of using e-models in an inquiry-based curriculum. Approach: 52 second-year dental students completed a pre- and post-test questionnaire relating to their perceptions of e-models and their use in inquiry-based learning. The pre-test occurred prior to any learning with e-models. The follow-up survey was conducted after one year's experience of using e-models. Results: There was no significant difference between the two sets of questionnaires regarding student’s perceptions of the usefulness of e-models and their willingness to use e-models in future inquiry-based learning. Most of the students preferred using both plaster models and e-models in tandem. Conclusion: Students did not change their attitude towards e-models and most of them agreed or were neutral that e-models are useful in inquiry-based learning. Whilst recognizing the utility of 3D models for learning, student's preference for combining these with solid models has implications for the development of haptic sensibility in an operative discipline.

Keywords: e-models, inquiry-based curriculum, education, questionnaire

Procedia PDF Downloads 411
523 Range Suitability Model for Livestock Grazing in Taleghan Rangelands

Authors: Hossein Arzani, Masoud Jafari Shalamzari, Z. Arzani

Abstract:

This paper follows FAO model of suitability analysis. Influential factors affecting extensive grazing were determined and converted into a model. Taleghan rangelands were examined for common types of grazing animals as an example. Advantages and limitations were elicited. All range ecosystems’ components affect range suitability but due to the time and money restrictions, the most important and feasible elements were investigated. From which three sub-models including water accessibility, forage production and erosion sensitivity were considered. Suitable areas in four levels of suitability were calculated using GIS. This suitability modeling approach was adopted due to its simplicity and the minimal time that is required for transforming and analyzing the data sets. Managers could be benefited from the model to devise the measures more wisely to cope with the limitations and enhance the rangelands health and condition.

Keywords: range suitability, land-use, extensive grazing, modeling, land evaluation

Procedia PDF Downloads 330
522 Generation of Knowlege with Self-Learning Methods for Ophthalmic Data

Authors: Klaus Peter Scherer, Daniel Knöll, Constantin Rieder

Abstract:

Problem and Purpose: Intelligent systems are available and helpful to support the human being decision process, especially when complex surgical eye interventions are necessary and must be performed. Normally, such a decision support system consists of a knowledge-based module, which is responsible for the real assistance power, given by an explanation and logical reasoning processes. The interview based acquisition and generation of the complex knowledge itself is very crucial, because there are different correlations between the complex parameters. So, in this project (semi)automated self-learning methods are researched and developed for an enhancement of the quality of such a decision support system. Methods: For ophthalmic data sets of real patients in a hospital, advanced data mining procedures seem to be very helpful. Especially subgroup analysis methods are developed, extended and used to analyze and find out the correlations and conditional dependencies between the structured patient data. After finding causal dependencies, a ranking must be performed for the generation of rule-based representations. For this, anonymous patient data are transformed into a special machine language format. The imported data are used as input for algorithms of conditioned probability methods to calculate the parameter distributions concerning a special given goal parameter. Results: In the field of knowledge discovery advanced methods and applications could be performed to produce operation and patient related correlations. So, new knowledge was generated by finding causal relations between the operational equipment, the medical instances and patient specific history by a dependency ranking process. After transformation in association rules logically based representations were available for the clinical experts to evaluate the new knowledge. The structured data sets take account of about 80 parameters as special characteristic features per patient. For different extended patient groups (100, 300, 500), as well one target value as well multi-target values were set for the subgroup analysis. So the newly generated hypotheses could be interpreted regarding the dependency or independency of patient number. Conclusions: The aim and the advantage of such a semi-automatically self-learning process are the extensions of the knowledge base by finding new parameter correlations. The discovered knowledge is transformed into association rules and serves as rule-based representation of the knowledge in the knowledge base. Even more, than one goal parameter of interest can be considered by the semi-automated learning process. With ranking procedures, the most strong premises and also conjunctive associated conditions can be found to conclude the interested goal parameter. So the knowledge, hidden in structured tables or lists can be extracted as rule-based representation. This is a real assistance power for the communication with the clinical experts.

Keywords: an expert system, knowledge-based support, ophthalmic decision support, self-learning methods

Procedia PDF Downloads 240
521 Plagiarism Detection for Flowchart and Figures in Texts

Authors: Ahmadu Maidorawa, Idrissa Djibo, Muhammad Tella

Abstract:

This paper presents a method for detecting flow chart and figure plagiarism based on shape of image processing and multimedia retrieval. The method managed to retrieve flowcharts with ranked similarity according to different matching sets. Plagiarism detection is well known phenomenon in the academic arena. Copying other people is considered as serious offense that needs to be checked. There are many plagiarism detection systems such as turn-it-in that has been developed to provide these checks. Most, if not all, discard the figures and charts before checking for plagiarism. Discarding the figures and charts result in look holes that people can take advantage. That means people can plagiarize figures and charts easily without the current plagiarism systems detecting it. There are very few papers which talks about flowcharts plagiarism detection. Therefore, there is a need to develop a system that will detect plagiarism in figures and charts.

Keywords: flowchart, multimedia retrieval, figures similarity, image comparison, figure retrieval

Procedia PDF Downloads 447
520 Mosaic Augmentation: Insights and Limitations

Authors: Olivia A. Kjorlien, Maryam Asghari, Farshid Alizadeh-Shabdiz

Abstract:

The goal of this paper is to investigate the impact of mosaic augmentation on the performance of object detection solutions. To carry out the study, YOLOv4 and YOLOv4-Tiny models have been selected, which are popular, advanced object detection models. These models are also representatives of two classes of complex and simple models. The study also has been carried out on two categories of objects, simple and complex. For this study, YOLOv4 and YOLOv4 Tiny are trained with and without mosaic augmentation for two sets of objects. While mosaic augmentation improves the performance of simple object detection, it deteriorates the performance of complex object detection, specifically having the largest negative impact on the false positive rate in a complex object detection case.

Keywords: accuracy, false positives, mosaic augmentation, object detection, YOLOV4, YOLOV4-Tiny

Procedia PDF Downloads 113
519 A Design for Supply Chain Model by Integrated Evaluation of Design Value and Supply Chain Cost

Authors: Yuan-Jye Tseng, Jia-Shu Li

Abstract:

To design a product with the given product requirement and design objective, there can be alternative ways to propose the detailed design specifications of the product. In the design modeling stage, alternative design cases with detailed specifications can be modeled to fulfill the product requirement and design objective. Therefore, in the design evaluation stage, it is required to perform an evaluation of the alternative design cases for deciding the final design. The purpose of this research is to develop a product evaluation model for evaluating the alternative design cases by integrated evaluating the criteria of functional design, Kansei design, and design for supply chain. The criteria in the functional design group include primary function, expansion function, improved function, and new function. The criteria in the Kansei group include geometric shape, dimension, surface finish, and layout. The criteria in the design for supply chain group include material, manufacturing process, assembly, and supply chain operation. From the point of view of value and cost, the criteria in the functional design group and Kansei design group represent the design value of the product. The criteria in the design for supply chain group represent the supply chain and manufacturing cost of the product. It is required to evaluate the design value and the supply chain cost to determine the final design. For the purpose of evaluating the criteria in the three criteria groups, a fuzzy analytic network process (FANP) method is presented to evaluate a weighted index by calculating the total relational values among the three groups. A method using the technique for order preference by similarity to ideal solution (TOPSIS) is used to compare and rank the design alternative cases according to the weighted index using the total relational values of the criteria. The final decision of a design case can be determined by using the ordered ranking. For example, the design case with the top ranking can be selected as the final design case. Based on the criteria in the evaluation, the design objective can be achieved with a combined and weighted effect of the design value and manufacturing cost. An example product is demonstrated and illustrated in the presentation. It shows that the design evaluation model is useful for integrated evaluation of functional design, Kansei design, and design for supply chain to determine the best design case and achieve the design objective.

Keywords: design for supply chain, design evaluation, functional design, Kansei design, fuzzy analytic network process, technique for order preference by similarity to ideal solution

Procedia PDF Downloads 304
518 A Double Acceptance Sampling Plan for Truncated Life Test Having Exponentiated Transmuted Weibull Distribution

Authors: A. D. Abdellatif, A. N. Ahmed, M. E. Abdelaziz

Abstract:

The main purpose of this paper is to design a double acceptance sampling plan under the time truncated life test when the product lifetime follows an exponentiated transmuted Weibull distribution. Here, the motive is to meet both the consumer’s risk and producer’s risk simultaneously at the specified quality levels, while the termination time is specified. A comparison between the results of the double and single acceptance sampling plans is conducted. We demonstrate the applicability of our results to real data sets.

Keywords: double sampling plan, single sampling plan, producer’s risk, consumer’s risk, exponentiated transmuted weibull distribution, time truncated experiment, single, double, Marshal-Olkin

Procedia PDF Downloads 472
517 Polymer in Electronic Waste: An Analysis

Authors: Anis A. Ansari, Aftab A. Ansari

Abstract:

Electronic waste is inundating the traditional solid-waste-disposal facilities, which are inadequately designed to handle and manage such type of new wastes. Since electronic waste contains mostly hazardous and even toxic materials, the seriousness of its effects on human health and the environment cannot be ignored in present scenario. Waste from the electronic industry is increasing exponentially day by day. From the last 20 years, we are continuously generating huge quantities of e-waste such as obsolete computers and other discarded electronic components, mainly due to evolution of newer technologies as a result of constant efforts in research and development in this sector. Polymers, one of the major constituents in almost every electronic waste, such as computers, printers, electronic equipment, entertainment devices, mobile phones, television sets etc., are if properly recycled can create a new business opportunity. This would not only create potential market for polymers to improve economy but also the priceless land used as dumping sites of electronic waste, can be utilized for other productive purposes.

Keywords: polymer recycling, electronic waste, hazardous materials, electronic components

Procedia PDF Downloads 455
516 Effect of Magnetic Field on Mixed Convection Boundary Layer Flow over an Exponentially Shrinking Vertical Sheet with Suction

Authors: S. S. P. M. Isa, N. M. Arifin, R. Nazar, N. Bachok, F. M. Ali, I. Pop

Abstract:

A theoretical study has been presented to describe the boundary layer flow and heat transfer on an exponentially shrinking sheet with a variable wall temperature and suction, in the presence of magnetic field. The governing nonlinear partial differential equations are converted into ordinary differential equations by similarity transformation, which are then solved numerically using the shooting method. Results for the skin friction coefficient, local Nusselt number, velocity profiles as well as temperature profiles are presented through graphs and tables for several sets of values of the parameters. The effects of the governing parameters on the flow and heat transfer characteristics are thoroughly examined.

Keywords: exponentially shrinking sheet, magnetic field, mixed convection, suction

Procedia PDF Downloads 318
515 Comparison of ANFIS Update Methods Using Genetic Algorithm, Particle Swarm Optimization, and Artificial Bee Colony

Authors: Michael R. Phangtriastu, Herriyandi Herriyandi, Diaz D. Santika

Abstract:

This paper presents a comparison of the implementation of metaheuristic algorithms to train the antecedent parameters and consequence parameters in the adaptive network-based fuzzy inference system (ANFIS). The algorithms compared are genetic algorithm (GA), particle swarm optimization (PSO), and artificial bee colony (ABC). The objective of this paper is to benchmark well-known metaheuristic algorithms. The algorithms are applied to several data set with different nature. The combinations of the algorithms' parameters are tested. In all algorithms, a different number of populations are tested. In PSO, combinations of velocity are tested. In ABC, a different number of limit abandonment are tested. Experiments find out that ABC is more reliable than other algorithms, ABC manages to get better mean square error (MSE) than other algorithms in all data set.

Keywords: ANFIS, artificial bee colony, genetic algorithm, metaheuristic algorithm, particle swarm optimization

Procedia PDF Downloads 335
514 Polynomially Adjusted Bivariate Density Estimates Based on the Saddlepoint Approximation

Authors: S. B. Provost, Susan Sheng

Abstract:

An alternative bivariate density estimation methodology is introduced in this presentation. The proposed approach involves estimating the density function associated with the marginal distribution of each of the two variables by means of the saddlepoint approximation technique and applying a bivariate polynomial adjustment to the product of these density estimates. Since the saddlepoint approximation is utilized in the context of density estimation, such estimates are determined from empirical cumulant-generating functions. In the univariate case, the saddlepoint density estimate is itself adjusted by a polynomial. Given a set of observations, the coefficients of the polynomial adjustments are obtained from the sample moments. Several illustrative applications of the proposed methodology shall be presented. Since this approach relies essentially on a determinate number of sample moments, it is particularly well suited for modeling massive data sets.

Keywords: density estimation, empirical cumulant-generating function, moments, saddlepoint approximation

Procedia PDF Downloads 264
513 Effect of Personality Traits on Classification of Political Orientation

Authors: Vesile Evrim, Aliyu Awwal

Abstract:

Today as in the other domains, there are an enormous number of political transcripts available in the Web which is waiting to be mined and used for various purposes such as statistics and recommendations. Therefore, automatically determining the political orientation on these transcripts becomes crucial. The methodologies used by machine learning algorithms to do the automatic classification are based on different features such as Linguistic. Considering the ideology differences between Liberals and Conservatives, in this paper, the effect of Personality Traits on political orientation classification is studied. This is done by considering the correlation between LIWC features and the BIG Five Personality Traits. Several experiments are conducted on Convote U.S. Congressional-Speech dataset with seven benchmark classification algorithms. The different methodologies are applied on selecting different feature sets that constituted by 8 to 64 varying number of features. While Neuroticism is obtained to be the most differentiating personality trait on classification of political polarity, when its top 10 representative features are combined with several classification algorithms, it outperformed the results presented in previous research.

Keywords: politics, personality traits, LIWC, machine learning

Procedia PDF Downloads 478
512 The Prospective Assessment of Zero-Energy Dwellings

Authors: Jovana Dj. Jovanovic, Svetlana M. Stevovic

Abstract:

The highest priority of so called, projected passive houses is to meet the appropriate energy demand. Every single material and layer which is injected into a dwelling has a certain energy quantity stored. The passive houses include optimized insulation levels with minimal thermal bridges, minimum of air leakage through the building, utilization of passive solar and internal gains, and good circulation of air which leans on mechanical ventilation system. The focus of this paper is on passive house features, benefits and targets, their feasibility and energy demands which are set up during each project. Numerous passive house-standards outline the very significant role of zero-energy dwellings towards the modern label of sustainable development. It is clear that the performance of both built and existing housing stock must be addressed if the population across the world sets out the energy objectives. This scientific article examines passive house features since the many passive house cases are launched.

Keywords: benefits, energy demands, passive houses, sustainable development

Procedia PDF Downloads 317
511 Decentralized Control of Interconnected Systems with Non-Linear Unknown Interconnections

Authors: Haci Mehmet Guzey, Levent Acar

Abstract:

In this paper, a novel decentralized controller is developed for linear systems with nonlinear unknown interconnections. A model linear decoupled system is assigned for each system. By using the difference actual and model state dynamics, the problem is formulated as inverse problem. Then, the interconnected dynamics are approximated by using Galerkin’s expansion method for inverse problems. Two different sets of orthogonal basis functions are utilized to approximate the interconnected dynamics. Approximated interconnections are utilized in the controller to cancel the interconnections and decouple the systems. Subsequently, the interconnected systems behave as a collection of decoupled systems.

Keywords: decentralized control, inverse problems, large scale systems, nonlinear interconnections, basis functions, system identification

Procedia PDF Downloads 522
510 Use of Socially Assistive Robots in Early Rehabilitation to Promote Mobility for Infants with Motor Delays

Authors: Elena Kokkoni, Prasanna Kannappan, Ashkan Zehfroosh, Effrosyni Mavroudi, Kristina Strother-Garcia, James C. Galloway, Jeffrey Heinz, Rene Vidal, Herbert G. Tanner

Abstract:

Early immobility affects the motor, cognitive, and social development. Current pediatric rehabilitation lacks the technology that will provide the dosage needed to promote mobility for young children at risk. The addition of socially assistive robots in early interventions may help increase the mobility dosage. The aim of this study is to examine the feasibility of an early intervention paradigm where non-walking infants experience independent mobility while socially interacting with robots. A dynamic environment is developed where both the child and the robot interact and learn from each other. The environment involves: 1) a range of physical activities that are goal-oriented, age-appropriate, and ability-matched for the child to perform, 2) the automatic functions that perceive the child’s actions through novel activity recognition algorithms, and decide appropriate actions for the robot, and 3) a networked visual data acquisition system that enables real-time assessment and provides the means to connect child behavior with robot decision-making in real-time. The environment was tested by bringing a two-year old boy with Down syndrome for eight sessions. The child presented delays throughout his motor development with the current being on the acquisition of walking. During the sessions, the child performed physical activities that required complex motor actions (e.g. climbing an inclined platform and/or staircase). During these activities, a (wheeled or humanoid) robot was either performing the action or was at its end point 'signaling' for interaction. From these sessions, information was gathered to develop algorithms to automate the perception of activities which the robot bases its actions on. A Markov Decision Process (MDP) is used to model the intentions of the child. A 'smoothing' technique is used to help identify the model’s parameters which are a critical step when dealing with small data sets such in this paradigm. The child engaged in all activities and socially interacted with the robot across sessions. With time, the child’s mobility was increased, and the frequency and duration of complex and independent motor actions were also increased (e.g. taking independent steps). Simulation results on the combination of the MDP and smoothing support the use of this model in human-robot interaction. Smoothing facilitates learning MDP parameters from small data sets. This paradigm is feasible and provides an insight on how social interaction may elicit mobility actions suggesting a new early intervention paradigm for very young children with motor disabilities. Acknowledgment: This work has been supported by NIH under grant #5R01HD87133.

Keywords: activity recognition, human-robot interaction, machine learning, pediatric rehabilitation

Procedia PDF Downloads 275
509 The Modality of Multivariate Skew Normal Mixture

Authors: Bader Alruwaili, Surajit Ray

Abstract:

Finite mixtures are a flexible and powerful tool that can be used for univariate and multivariate distributions, and a wide range of research analysis has been conducted based on the multivariate normal mixture and multivariate of a t-mixture. Determining the number of modes is an important activity that, in turn, allows one to determine the number of homogeneous groups in a population. Our work currently being carried out relates to the study of the modality of the skew normal distribution in the univariate and multivariate cases. For the skew normal distribution, the aims are associated with studying the modality of the skew normal distribution and providing the ridgeline, the ridgeline elevation function, the $\Pi$ function, and the curvature function, and this will be conducive to an exploration of the number and location of mode when mixing the two components of skew normal distribution. The subsequent objective is to apply these results to the application of real world data sets, such as flow cytometry data.

Keywords: mode, modality, multivariate skew normal, finite mixture, number of mode

Procedia PDF Downloads 472
508 A New Approach for Improving Accuracy of Multi Label Stream Data

Authors: Kunal Shah, Swati Patel

Abstract:

Many real world problems involve data which can be considered as multi-label data streams. Efficient methods exist for multi-label classification in non streaming scenarios. However, learning in evolving streaming scenarios is more challenging, as the learners must be able to adapt to change using limited time and memory. Classification is used to predict class of unseen instance as accurate as possible. Multi label classification is a variant of single label classification where set of labels associated with single instance. Multi label classification is used by modern applications, such as text classification, functional genomics, image classification, music categorization etc. This paper introduces the task of multi-label classification, methods for multi-label classification and evolution measure for multi-label classification. Also, comparative analysis of multi label classification methods on the basis of theoretical study, and then on the basis of simulation was done on various data sets.

Keywords: binary relevance, concept drift, data stream mining, MLSC, multiple window with buffer

Procedia PDF Downloads 568
507 Analyzing the Characteristics and Shifting Patterns of Creative Hubs in Bandung

Authors: Fajar Ajie Setiawan, Ratu Azima Mayangsari, Bunga Aprilia

Abstract:

The emergence of creative hubs around the world, including in Bandung, was primarily driven by the needs of collaborative-innovative spaces for creative industry activities such as the Maker Movement and the Coworking Movement. These activities pose challenges for identification and formulation of sets of indicators for modeling creative hubs in Bandung to help stakeholders in formulating strategies. This study intends to identify their characteristics. This research was conducted using a qualitative approach comparing three concepts of creative hub categorization and integrating them into a single instrument to analyze 12 selected creative hubs. Our results showed three new functions of creative hubs in Bandung: (1) cultural, (2) retail business, and (3) community network. Results also suggest that creative hubs in Bandung are commonly established for networking and community activities. Another result shows that there was a shifting pattern of creative hubs before the 2000s and after the 2000s, which also creates a hybrid group of creative hubs.

Keywords: creative industry, creative hubs, Ngariung, Bandung

Procedia PDF Downloads 155
506 Further Analysis of Global Robust Stability of Neural Networks with Multiple Time Delays

Authors: Sabri Arik

Abstract:

In this paper, we study the global asymptotic robust stability of delayed neural networks with norm-bounded uncertainties. By employing the Lyapunov stability theory and Homeomorphic mapping theorem, we derive some new types of sufficient conditions ensuring the existence, uniqueness and global asymptotic stability of the equilibrium point for the class of neural networks with discrete time delays under parameter uncertainties and with respect to continuous and slopebounded activation functions. An important aspect of our results is their low computational complexity as the reported results can be verified by checking some properties symmetric matrices associated with the uncertainty sets of network parameters. The obtained results are shown to be generalization of some of the previously published corresponding results. Some comparative numerical examples are also constructed to compare our results with some closely related existing literature results.

Keywords: neural networks, delayed systems, lyapunov functionals, stability analysis

Procedia PDF Downloads 510
505 Bayesian Hidden Markov Modelling of Blood Type Distribution for COVID-19 Cases Using Poisson Distribution

Authors: Johnson Joseph Kwabina Arhinful, Owusu-Ansah Emmanuel Degraft Johnson, Okyere Gabrial Asare, Adebanji Atinuke Olusola

Abstract:

This paper proposes a model to describe the blood types distribution of new Coronavirus (COVID-19) cases using the Bayesian Poisson - Hidden Markov Model (BP-HMM). With the help of the Gibbs sampler algorithm, using OpenBugs, the study first identifies the number of hidden states fitting European (EU) and African (AF) data sets of COVID-19 cases by blood type frequency. The study then compares the state-dependent mean of infection within and across the two geographical areas. The study findings show that the number of hidden states and infection rates within and across the two geographical areas differ according to blood type.

Keywords: BP-HMM, COVID-19, blood types, GIBBS sampler

Procedia PDF Downloads 114
504 Evaluation of Sensor Pattern Noise Estimators for Source Camera Identification

Authors: Benjamin Anderson-Sackaney, Amr Abdel-Dayem

Abstract:

This paper presents a comprehensive survey of recent source camera identification (SCI) systems. Then, the performance of various sensor pattern noise (SPN) estimators was experimentally assessed, under common photo response non-uniformity (PRNU) frameworks. The experiments used 1350 natural and 900 flat-field images, captured by 18 individual cameras. 12 different experiments, grouped into three sets, were conducted. The results were analyzed using the receiver operator characteristic (ROC) curves. The experimental results demonstrated that combining the basic SPN estimator with a wavelet-based filtering scheme provides promising results. However, the phase SPN estimator fits better with both patch-based (BM3D) and anisotropic diffusion (AD) filtering schemes.

Keywords: sensor pattern noise, source camera identification, photo response non-uniformity, anisotropic diffusion, peak to correlation energy ratio

Procedia PDF Downloads 424
503 The Tourism Management: The Case of Kingdom of Cambodia

Authors: Chanpen Meenakorn

Abstract:

The purpose of this study are (1) development plan and management strategy of Virachey Natioanl Park, (2) to study stakeholders’ perception on tourism development for sustainable tourism planning and management. The data was collected through 28 sets of questionnaires with the total population of international visitors who were interested in Ecotourism in northeast Cambodia and traveled to Virachey National Park. The SPSS programme was used to analyze the level of visitors’ satisfaction and perception on tourism development. The results of the study indicated that moderate potentiality to be developed as tourist attraction for sustainable tourism development in the park. The components with moderate potential are physical condition, management, activities and process of natural and cultural tourism, and organization and participation of the local community. The study also found that most local communities satisfy with tourism development in the park as well as in their community.

Keywords: Kingdom of Cambodia, stakeholders’ perception, tourism management, Virachey National Park

Procedia PDF Downloads 349
502 Event Extraction, Analysis, and Event Linking

Authors: Anam Alam, Rahim Jamaluddin Kanji

Abstract:

With the rapid growth of event in everywhere, event extraction has now become an important matter to retrieve the information from the unstructured data. One of the challenging problems is to extract the event from it. An event is an observable occurrence of interaction among entities. The paper investigates the effectiveness of event extraction capabilities of three software tools that are Wandora, Nitro and SPSS. We performed standard text mining techniques of these tools on the data sets of (i) Afghan War Diaries (AWD collection), (ii) MUC4 and (iii) WebKB. Information retrieval measures such as precision and recall which are computed under extensive set of experiments for Event Extraction. The experimental study analyzes the difference between events extracted by the software and human. This approach helps to construct an algorithm that will be applied for different machine learning methods.

Keywords: event extraction, Wandora, nitro, SPSS, event analysis, extraction method, AFG, Afghan War Diaries, MUC4, 4 universities, dataset, algorithm, precision, recall, evaluation

Procedia PDF Downloads 572