Search results for: quadratic assignment problem
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7477

Search results for: quadratic assignment problem

7057 Uncertain Time-Cost Trade off Problems of Construction Projects Using Fuzzy Set Theory

Authors: V. S. S. Kumar, B. Vikram

Abstract:

The development of effective decision support tools that adopted in the construction industry is vital in the world we live in today, since it can lead to substantial cost reduction and efficient resource consumption. Solving the time-cost trade off problems and its related variants is at the heart of scientific research for optimizing construction planning problems. In general, the classical optimization techniques have difficulties in dealing with TCT problems. One of the main reasons of their failure is that they can easily be entrapped in local minima. This paper presents an investigation on the application of meta-heuristic techniques to two particular variants of the time-cost trade of analysis, the time-cost trade off problem (TCT), and time-cost trade off optimization problem (TCO). In first problem, the total project cost should be minimized, and in the second problem, the total project cost and total project duration should be minimized simultaneously. Finally it is expected that, the optimization models developed in this paper will contribute significantly for efficient planning and management of construction project.

Keywords: fuzzy sets, uncertainty, optimization, time cost trade off problems

Procedia PDF Downloads 356
7056 E–Learning System in Virtual Learning Environment to Develop Problem Solving Ability and Team Learning for Learners in Higher Education

Authors: Noawanit Songkram

Abstract:

This paper is a report on the findings of a study conducted on e–learning system in virtual learning environment to develop problem solving ability and team learning for learners in higher education. The methodology of this study was R&D research. The subjects were 18 undergraduate students in Faculty of Education, Chulalongkorn University in the academic year of 2013. The research instruments were a problem solving ability assessment, a team learning evaluation form, and an attitude questionnaire. The data was statistically analyzed using mean, standard deviation, one way repeated measure ANOVA and t–test. The research findings discovered the e –learning system in virtual learning environment to develop problem solving ability and team learning for learners in higher education consisted of five components:(1) online collaborative tools, (2) active learning activities, (3) creative thinking, (4) knowledge sharing process, (5) evaluation and nine processes which were (1) preparing in group working, (2) identifying interested topic, (3) analysing interested topic, (4) collecting data, (5) concluding idea (6) proposing idea, (7) creating workings, (8) workings evaluation, (9) sharing knowledge from empirical experience.

Keywords: e-learning system, problem solving ability, team leaning, virtual learning environment

Procedia PDF Downloads 438
7055 Demonstration of Land Use Changes Simulation Using Urban Climate Model

Authors: Barbara Vojvodikova, Katerina Jupova, Iva Ticha

Abstract:

Cities in their historical evolution have always adapted their internal structure to the needs of society (for example protective city walls during classicism era lost their defense function, became unnecessary, were demolished and gave space for new features such as roads, museums or parks). Today it is necessary to modify the internal structure of the city in order to minimize the impact of climate changes on the environment of the population. This article discusses the results of the Urban Climate model owned by VITO, which was carried out as part of a project from the European Union's Horizon grant agreement No 730004 Pan-European Urban Climate Services Climate-Fit city. The use of the model was aimed at changes in land use and land cover in cities related to urban heat islands (UHI). The task of the application was to evaluate possible land use change scenarios in connection with city requirements and ideas. Two pilot areas in the Czech Republic were selected. One is Ostrava and the other Hodonín. The paper provides a demonstration of the application of the model for various possible future development scenarios. It contains an assessment of the suitability or inappropriateness of scenarios of future development depending on the temperature increase. Cities that are preparing to reconstruct the public space are interested in eliminating proposals that would lead to an increase in temperature stress as early as in the assignment phase. If they have evaluation on the unsuitability of some type of design, they can limit it into the proposal phases. Therefore, especially in the application of models on Local level - in 1 m spatial resolution, it was necessary to show which type of proposals would create a significant temperature island in its implementation. Such a type of proposal is considered unsuitable. The model shows that the building itself can create a shady place and thus contribute to the reduction of the UHI. If it sensitively approaches the protection of existing greenery, this new construction may not pose a significant problem. More massive interventions leading to the reduction of existing greenery create a new heat island space.

Keywords: climate model, heat islands, Hodonin, land use changes, Ostrava

Procedia PDF Downloads 143
7054 Legal Issues of Collecting and Processing Big Health Data in the Light of European Regulation 679/2016

Authors: Ioannis Iglezakis, Theodoros D. Trokanas, Panagiota Kiortsi

Abstract:

This paper aims to explore major legal issues arising from the collection and processing of Health Big Data in the light of the new European secondary legislation for the protection of personal data of natural persons, placing emphasis on the General Data Protection Regulation 679/2016. Whether Big Health Data can be characterised as ‘personal data’ or not is really the crux of the matter. The legal ambiguity is compounded by the fact that, even though the processing of Big Health Data is premised on the de-identification of the data subject, the possibility of a combination of Big Health Data with other data circulating freely on the web or from other data files cannot be excluded. Another key point is that the application of some provisions of GPDR to Big Health Data may both absolve the data controller of his legal obligations and deprive the data subject of his rights (e.g., the right to be informed), ultimately undermining the fundamental right to the protection of personal data of natural persons. Moreover, data subject’s rights (e.g., the right not to be subject to a decision based solely on automated processing) are heavily impacted by the use of AI, algorithms, and technologies that reclaim health data for further use, resulting in sometimes ambiguous results that have a substantial impact on individuals. On the other hand, as the COVID-19 pandemic has revealed, Big Data analytics can offer crucial sources of information. In this respect, this paper identifies and systematises the legal provisions concerned, offering interpretative solutions that tackle dangers concerning data subject’s rights while embracing the opportunities that Big Health Data has to offer. In addition, particular attention is attached to the scope of ‘consent’ as a legal basis in the collection and processing of Big Health Data, as the application of data analytics in Big Health Data signals the construction of new data and subject’s profiles. Finally, the paper addresses the knotty problem of role assignment (i.e., distinguishing between controller and processor/joint controllers and joint processors) in an era of extensive Big Health data sharing. The findings are the fruit of a current research project conducted by a three-member research team at the Faculty of Law of the Aristotle University of Thessaloniki and funded by the Greek Ministry of Education and Religious Affairs.

Keywords: big health data, data subject rights, GDPR, pandemic

Procedia PDF Downloads 129
7053 Efficient Reconstruction of DNA Distance Matrices Using an Inverse Problem Approach

Authors: Boris Melnikov, Ye Zhang, Dmitrii Chaikovskii

Abstract:

We continue to consider one of the cybernetic methods in computational biology related to the study of DNA chains. Namely, we are considering the problem of reconstructing the not fully filled distance matrix of DNA chains. When applied in a programming context, it is revealed that with a modern computer of average capabilities, creating even a small-sized distance matrix for mitochondrial DNA sequences is quite time-consuming with standard algorithms. As the size of the matrix grows larger, the computational effort required increases significantly, potentially spanning several weeks to months of non-stop computer processing. Hence, calculating the distance matrix on conventional computers is hardly feasible, and supercomputers are usually not available. Therefore, we started publishing our variants of the algorithms for calculating the distance between two DNA chains; then, we published algorithms for restoring partially filled matrices, i.e., the inverse problem of matrix processing. In this paper, we propose an algorithm for restoring the distance matrix for DNA chains, and the primary focus is on enhancing the algorithms that shape the greedy function within the branches and boundaries method framework.

Keywords: DNA chains, distance matrix, optimization problem, restoring algorithm, greedy algorithm, heuristics

Procedia PDF Downloads 118
7052 Evaluation of the Socio-Economic Impact of Marine Debris in Coastal Nigeria

Authors: Chibuzo Okoye Daniels, Gillian Glegg, Lynda Rodwell

Abstract:

Marine debris from fishing nets to medical equipment to food packaging that play major roles in boosting the economy and protecting human health is now more than an environmental problem that can be solved by legislation, law enforcement and technical solutions. It has also been identified as a cultural problem that can only be addressed by identifying instruments that can be used to change human attitudes and behaviors. This may be through management approaches, education and involvement of all sectors/interests, including the public. To contribute to the sustainable development of coastal Nigeria, two case study areas (Ikoyi and Victoria Islands of Lagos State) were used to evaluate the socio-economic impacts of marine debris problem in coastal Nigeria. The following methods were used: (1) semi-structured interviews with key stakeholders and businesses on beaches, waterfronts and waterways within the study areas and (2) observational study of beaches, waterfronts and waterways within the study areas. The results of the study have shown that marine debris is a cultural and multi-sectoral problem that poses great threat not only to the environmental sustainability of the study areas but also to the wellbeing of its citizens and the economy of coastal Nigeria. Current solid waste and marine debris management practices are inefficient due to inadequate knowledge of how to tackle the problem. To ensure environmental sustainability in coastal Nigeria and avoid waste of scarce financial resources, adequate, appropriate and cost effective solutions to the marine debris problem need to be identified and effectively transferred for implementation in the study areas.

Keywords: sustainability, coastal Nigeria, study areas, aquaculture

Procedia PDF Downloads 554
7051 Labor Productivity in the Construction Industry: Factors Influencing the Spanish Construction Labor Productivity

Authors: G. Robles, A. Stifi, José L. Ponz-Tienda, S. Gentes

Abstract:

This research paper aims to identify, analyze and rank factors affecting labor productivity in Spain with respect to their relative importance. Using a selected set of 35 factors, a structured questionnaire survey was utilized as the method to collect data from companies. Target population is comprised by a random representative sample of practitioners related with the Spanish construction industry. Findings reveal the top five ranked factors are as follows: (1) shortage or late supply of materials; (2) clarity of the drawings and project documents; (3) clear and daily task assignment; (4) tools or equipment shortages; (5) level of skill and experience of laborers. Additionally, this research also pretends to provide simple and comprehensive recommendations so that they could be implemented by construction managers for an effective management of construction labor forces.

Keywords: construction management, factors, improvement, labor productivity, lean construction

Procedia PDF Downloads 292
7050 Optimization of Biodiesel Production from Sunflower Oil Using Central Composite Design

Authors: Pascal Mwenge, Jefrey Pilusa, Tumisang Seodigeng

Abstract:

The current study investigated the effect of catalyst ratio and methanol to oil ratio on biodiesel production by using central composite design. Biodiesel was produced by transesterification using sodium hydroxide as a homogeneous catalyst, a laboratory scale reactor consisting of flat bottom flask mounts with a reflux condenser, and a heating plate was used to produce biodiesel. Key parameters, including time, temperature, and mixing rate was kept constant at 60 minutes, 60 oC and 600 RPM, respectively. From the results obtained, it was observed that the biodiesel yield depends on catalyst ratio and methanol to oil ratio. The highest yield of 50.65% was obtained at catalyst ratio of 0.5 wt.% and methanol to oil mole ratio 10.5. The analysis of variances of biodiesel yield showed the R Squared value of 0.8387. A quadratic mathematical model was developed to predict the biodiesel yield in the specified parameters ranges.

Keywords: ANOVA, biodiesel, catalyst, transesterification, central composite design

Procedia PDF Downloads 152
7049 The Classification Performance in Parametric and Nonparametric Discriminant Analysis for a Class- Unbalanced Data of Diabetes Risk Groups

Authors: Lily Ingsrisawang, Tasanee Nacharoen

Abstract:

Introduction: The problems of unbalanced data sets generally appear in real world applications. Due to unequal class distribution, many research papers found that the performance of existing classifier tends to be biased towards the majority class. The k -nearest neighbors’ nonparametric discriminant analysis is one method that was proposed for classifying unbalanced classes with good performance. Hence, the methods of discriminant analysis are of interest to us in investigating misclassification error rates for class-imbalanced data of three diabetes risk groups. Objective: The purpose of this study was to compare the classification performance between parametric discriminant analysis and nonparametric discriminant analysis in a three-class classification application of class-imbalanced data of diabetes risk groups. Methods: Data from a healthy project for 599 staffs in a government hospital in Bangkok were obtained for the classification problem. The staffs were diagnosed into one of three diabetes risk groups: non-risk (90%), risk (5%), and diabetic (5%). The original data along with the variables; diabetes risk group, age, gender, cholesterol, and BMI was analyzed and bootstrapped up to 50 and 100 samples, 599 observations per sample, for additional estimation of misclassification error rate. Each data set was explored for the departure of multivariate normality and the equality of covariance matrices of the three risk groups. Both the original data and the bootstrap samples show non-normality and unequal covariance matrices. The parametric linear discriminant function, quadratic discriminant function, and the nonparametric k-nearest neighbors’ discriminant function were performed over 50 and 100 bootstrap samples and applied to the original data. In finding the optimal classification rule, the choices of prior probabilities were set up for both equal proportions (0.33: 0.33: 0.33) and unequal proportions with three choices of (0.90:0.05:0.05), (0.80: 0.10: 0.10) or (0.70, 0.15, 0.15). Results: The results from 50 and 100 bootstrap samples indicated that the k-nearest neighbors approach when k = 3 or k = 4 and the prior probabilities of {non-risk:risk:diabetic} as {0.90:0.05:0.05} or {0.80:0.10:0.10} gave the smallest error rate of misclassification. Conclusion: The k-nearest neighbors approach would be suggested for classifying a three-class-imbalanced data of diabetes risk groups.

Keywords: error rate, bootstrap, diabetes risk groups, k-nearest neighbors

Procedia PDF Downloads 434
7048 A Survey of Discrete Facility Location Problems

Authors: Z. Ulukan, E. Demircioğlu,

Abstract:

Facility location is a complex real-world problem which needs a strategic management decision. This paper provides a general review on studies, efforts and developments in Facility Location Problems which are classical optimization problems having a wide-spread applications in various areas such as transportation, distribution, production, supply chain decisions and telecommunication. Our goal is not to review all variants of different studies in FLPs or to describe very detailed computational techniques and solution approaches, but rather to provide a broad overview of major location problems that have been studied, indicating how they are formulated and what are proposed by researchers to tackle the problem. A brief, elucidative table based on a grouping according to “General Problem Type” and “Methods Proposed” used in the studies is also presented at the end of the work.

Keywords: discrete location problems, exact methods, heuristic algorithms, single source capacitated facility location problems

Procedia PDF Downloads 471
7047 Optimisation of Extraction of Phenolic Compounds in Algerian Lavandula multifida, Algeria, NW

Authors: Mustapha Mahmoud Dif, Fouzia Benali-Toumi, Mohamed Benyahia, Sofiane Bouazza, Abbes Dellal, Slimane Baha

Abstract:

L. multifida is applied to treat rheumatism and cold and has hypoglycemic and anti-inflammatory properties. The present study is to optimize the extraction of phenolic compounds in Algerian Lavandula multifida. The influences of parameters including temperature (decoction and maceration) and extraction time (15min to 45 min) on the flavonoids concentration are studied. The optimal conditions are determined and the quadratic response surfaces draw from the mathematical models. Total phenols were evaluated using Folin sicaltieu methods, total flavonoids were estimated using the Tri chloral aluminum method. The maximum concentration extracted, for total flavonoids, equal to 0.043 mg/g was achieved with decoction and extraction time of 41.55 min. However, for total phenol compounds highest concentration of 0.218 mg/g, is obtained with 45 min at 49.99°C.

Keywords: L multifidi, phenolic content, optimization, time, temperature

Procedia PDF Downloads 420
7046 Decomposition-Based Pricing Technique for Solving Large-Scale Mixed IP

Authors: M. Babul Hasan

Abstract:

Management sciences (MS), big group of companies and industries or government policies (GP) is affiliated with a huge number of decision ingredients and complicated restrictions. Every factor in MS, every product in Industries or decision in GP is not always bankable in practice. After formulating these models there arises large-scale mixed integer programming (MIP) problem. In this paper, we developed decomposition-based pricing procedure to filter the unnecessary decision ingredients from MIP where the variables in huge number will be abated and the complicacy of restrictions will be elementary. A real life numerical example has been illustrated to demonstrate the methods. We develop the computer techniques for these methods by using a mathematical programming language (AMPL).

Keywords: Lagrangian relaxation, decomposition, sub-problem, master-problem, pricing, mixed IP, AMPL

Procedia PDF Downloads 505
7045 A Dual-Mode Infinite Horizon Predictive Control Algorithm for Load Tracking in PUSPATI TRIGA Reactor

Authors: Mohd Sabri Minhat, Nurul Adilla Mohd Subha

Abstract:

The PUSPATI TRIGA Reactor (RTP), Malaysia reached its first criticality on June 28, 1982, with power capacity 1MW thermal. The Feedback Control Algorithm (FCA) which is conventional Proportional-Integral (PI) controller, was used for present power control method to control fission process in RTP. It is important to ensure the core power always stable and follows load tracking within acceptable steady-state error and minimum settling time to reach steady-state power. At this time, the system could be considered not well-posed with power tracking performance. However, there is still potential to improve current performance by developing next generation of a novel design nuclear core power control. In this paper, the dual-mode predictions which are proposed in modelling Optimal Model Predictive Control (OMPC), is presented in a state-space model to control the core power. The model for core power control was based on mathematical models of the reactor core, OMPC, and control rods selection algorithm. The mathematical models of the reactor core were based on neutronic models, thermal hydraulic models, and reactivity models. The dual-mode prediction in OMPC for transient and terminal modes was based on the implementation of a Linear Quadratic Regulator (LQR) in designing the core power control. The combination of dual-mode prediction and Lyapunov which deal with summations in cost function over an infinite horizon is intended to eliminate some of the fundamental weaknesses related to MPC. This paper shows the behaviour of OMPC to deal with tracking, regulation problem, disturbance rejection and caters for parameter uncertainty. The comparison of both tracking and regulating performance is analysed between the conventional controller and OMPC by numerical simulations. In conclusion, the proposed OMPC has shown significant performance in load tracking and regulating core power for nuclear reactor with guarantee stabilising in the closed-loop.

Keywords: core power control, dual-mode prediction, load tracking, optimal model predictive control

Procedia PDF Downloads 162
7044 The Boundary Element Method in Excel for Teaching Vector Calculus and Simulation

Authors: Stephen Kirkup

Abstract:

This paper discusses the implementation of the boundary element method (BEM) on an Excel spreadsheet and how it can be used in teaching vector calculus and simulation. There are two separate spreadheets, within which Laplace equation is solved by the BEM in two dimensions (LIBEM2) and axisymmetric three dimensions (LBEMA). The main algorithms are implemented in the associated programming language within Excel, Visual Basic for Applications (VBA). The BEM only requires a boundary mesh and hence it is a relatively accessible method. The BEM in the open spreadsheet environment is demonstrated as being useful as an aid to teaching and learning. The application of the BEM implemented on a spreadsheet for educational purposes in introductory vector calculus and simulation is explored. The development of assignment work is discussed, and sample results from student work are given. The spreadsheets were found to be useful tools in developing the students’ understanding of vector calculus and in simulating heat conduction.

Keywords: boundary element method, Laplace’s equation, vector calculus, simulation, education

Procedia PDF Downloads 163
7043 Recursive Parametric Identification of a Doubly Fed Induction Generator-Based Wind Turbine

Authors: A. El Kachani, E. Chakir, A. Ait Laachir, A. Niaaniaa, J. Zerouaoui

Abstract:

This document presents an adaptive controller based on recursive parametric identification applied to a wind turbine based on the doubly-fed induction machine (DFIG), to compensate the faults and guarantee efficient of the DFIG. The proposed adaptive controller is based on the recursive least square algorithm which considers that the best estimator for the vector parameter is the vector x minimizing a quadratic criterion. Furthermore, this method can improve the rapidity and precision of the controller based on a model. The proposed controller is validated via simulation on a 5.5 kW DFIG-based wind turbine. The results obtained seem to be good. In addition, they show the advantages of an adaptive controller based on recursive least square algorithm.

Keywords: adaptive controller, recursive least squares algorithm, wind turbine, doubly fed induction generator

Procedia PDF Downloads 288
7042 Implementation of Project-Based Learning with Peer Assessment in Large Classes under Consideration of Faculty’s Scare Resources

Authors: Margit Kastner

Abstract:

To overcome the negative consequences associated with large class sizes and to support students in developing the necessary competences (e.g., critical thinking, problem-solving, or team-work skills) a marketing course has been redesigned by implementing project-based learning with peer assessment (PBL&PA). This means that students can voluntarily take advantage of this supplementary offer and explore -in addition to attending the lecture where clicker questions are asked- a real-world problem, find a solution, and assess the results of peers while working in small collaborative groups. In order to handle this with little further effort, the process is technically supported by the university’s e-learning system in such a way that students upload their solution in form of an assignment which is then automatically distributed to peer groups who have to assess the work of three other groups. Finally, students’ work is graded automatically considering both, students’ contribution to the project and the conformity of the peer assessment. The purpose of this study is to evaluate students’ perception of PBL&PA using an online-questionnaire to collect the data. More specifically, it aims to discover students’ motivations for (not) working on a project and the benefits and problems students encounter. In addition to the survey, students’ performance was analyzed by comparing the final grades of those who participated in PBL&PA with those who did not participate. Among the 260 students who filled out the questionnaire, 47% participated in PBL&PA. Besides extrinsic motivations (bonus credits), students’ participation was often motivated by learning and social benefits. Reasons for not working on a project were connected to students’ organization and management of their studies (e.g., time constraints, no/wrong information) and teamwork concerns (e.g., missing engagement of peers, prior negative experiences). In addition, high workload and insufficient extrinsic motivation (bonus credits) were mentioned. With regards to benefits and problems students encountered during the project, students provided more positive than negative comments. Positive aspects most often stated were learning and social benefits while negative ones were mainly attached to the technical implementation. Interestingly, bonus credits were hardly named as a positive aspect meaning that intrinsic motivations have become more important when working on the project. Team aspects generated mixed feelings. In addition, students who voluntarily participated in PBL&PA were, in general, more active and utilized further course offers such as clicker questions. Examining students’ performance at the final exam revealed that students without participating in any of the offered active learning tasks performed poorest in the exam while students who used all activities were best. In conclusion, the goals of the implementation were met in terms of students’ perceived benefits and the positive impact on students’ exam performance. Since the comparison of the automatic grading with faculty grading showed valid results, it is possible to rely only on automatic grading in the future. That way, the additional workload for faculty will be within limits. Thus, the implementation of project-based learning with peer assessment can be recommended for large classes.

Keywords: automated grading, large classes, peer assessment, project-based learning

Procedia PDF Downloads 165
7041 Machine Learning Facing Behavioral Noise Problem in an Imbalanced Data Using One Side Behavioral Noise Reduction: Application to a Fraud Detection

Authors: Salma El Hajjami, Jamal Malki, Alain Bouju, Mohammed Berrada

Abstract:

With the expansion of machine learning and data mining in the context of Big Data analytics, the common problem that affects data is class imbalance. It refers to an imbalanced distribution of instances belonging to each class. This problem is present in many real world applications such as fraud detection, network intrusion detection, medical diagnostics, etc. In these cases, data instances labeled negatively are significantly more numerous than the instances labeled positively. When this difference is too large, the learning system may face difficulty when tackling this problem, since it is initially designed to work in relatively balanced class distribution scenarios. Another important problem, which usually accompanies these imbalanced data, is the overlapping instances between the two classes. It is commonly referred to as noise or overlapping data. In this article, we propose an approach called: One Side Behavioral Noise Reduction (OSBNR). This approach presents a way to deal with the problem of class imbalance in the presence of a high noise level. OSBNR is based on two steps. Firstly, a cluster analysis is applied to groups similar instances from the minority class into several behavior clusters. Secondly, we select and eliminate the instances of the majority class, considered as behavioral noise, which overlap with behavior clusters of the minority class. The results of experiments carried out on a representative public dataset confirm that the proposed approach is efficient for the treatment of class imbalances in the presence of noise.

Keywords: machine learning, imbalanced data, data mining, big data

Procedia PDF Downloads 130
7040 Some Codes for Variants in Graphs

Authors: Sofia Ait Bouazza

Abstract:

We consider the problem of finding a minimum identifying code in a graph. This problem was initially introduced in 1998 and has been since fundamentally connected to a wide range of applications (fault diagnosis, location detection …). Suppose we have a building into which we need to place fire alarms. Suppose each alarm is designed so that it can detect any fire that starts either in the room in which it is located or in any room that shares a doorway with the room. We want to detect any fire that may occur or use the alarms which are sounding to not only to not only detect any fire but be able to tell exactly where the fire is located in the building. For reasons of cost, we want to use as few alarms as necessary. The first problem involves finding a minimum domination set of a graph. If the alarms are three state alarms capable of distinguishing between a fire in the same room as the alarm and a fire in an adjacent room, we are trying to find a minimum locating domination set. If the alarms are two state alarms that can only sound if there is a fire somewhere nearby, we are looking for a differentiating domination set of a graph. These three areas are the subject of much active research; we primarily focus on the third problem. An identifying code of a graph G is a dominating set C such that every vertex x of G is distinguished from other vertices by the set of vertices in C that are at distance at most r≥1 from x. When only vertices out of the code are asked to be identified, we get the related concept of a locating dominating set. The problem of finding an identifying code (resp a locating dominating code) of minimum size is a NP-hard problem, even when the input graph belongs to a number of specific graph classes. Therefore, we study this problem in some restricted classes of undirected graphs like split graph, line graph and path in a directed graph. Then we present some results on the identifying code by giving an exact value of upper total locating domination and a total 2-identifying code in directed and undirected graph. Moreover we determine exact values of locating dominating code and edge identifying code of thin headless spider and locating dominating code of complete suns.

Keywords: identiying codes, locating dominating set, split graphs, thin headless spider

Procedia PDF Downloads 478
7039 Image Classification with Localization Using Convolutional Neural Networks

Authors: Bhuyain Mobarok Hossain

Abstract:

Image classification and localization research is currently an important strategy in the field of computer vision. The evolution and advancement of deep learning and convolutional neural networks (CNN) have greatly improved the capabilities of object detection and image-based classification. Target detection is important to research in the field of computer vision, especially in video surveillance systems. To solve this problem, we will be applying a convolutional neural network of multiple scales at multiple locations in the image in one sliding window. Most translation networks move away from the bounding box around the area of interest. In contrast to this architecture, we consider the problem to be a classification problem where each pixel of the image is a separate section. Image classification is the method of predicting an individual category or specifying by a shoal of data points. Image classification is a part of the classification problem, including any labels throughout the image. The image can be classified as a day or night shot. Or, likewise, images of cars and motorbikes will be automatically placed in their collection. The deep learning of image classification generally includes convolutional layers; the invention of it is referred to as a convolutional neural network (CNN).

Keywords: image classification, object detection, localization, particle filter

Procedia PDF Downloads 305
7038 Discrete Group Search Optimizer for the Travelling Salesman Problem

Authors: Raed Alnajjar, Mohd Zakree, Ahmad Nazri

Abstract:

In this study, we apply Discrete Group Search Optimizer (DGSO) for solving Traveling Salesman Problem (TSP). The DGSO is a nature inspired optimization algorithm that imitates the animal behavior, especially animal searching behavior. The proposed DGSO uses a vector representation and some discrete operators, such as destruction, construction, differential evolution, swap and insert. The TSP is a well-known hard combinatorial optimization problem, which seeks to find the shortest path among numbers of cities. The performance of the proposed DGSO is evaluated and tested on benchmark instances which listed in LIBTSP dataset. The experimental results show that the performance of the proposed DGSO is comparable with the other methods in the state of the art for some instances. The results show that DGSO outperform Ant Colony System (ACS) in some instances whilst outperform other metaheuristic in most instances. In addition to that, the new results obtained a number of optimal solutions and some best known results. DGSO was able to obtain feasible and good quality solution across all dataset.

Keywords: discrete group search optimizer (DGSO); Travelling salesman problem (TSP); Variable neighborhood search(VNS)

Procedia PDF Downloads 324
7037 The Effect of Problem-Based Mobile-Assisted Tasks on Spoken Intelligibility of English as a Foreign Language Learners

Authors: Loghman Ansarian, Teoh Mei Lin

Abstract:

In an attempt to increase oral proficiency of Iranian EFL learners, the researchers compared the effect of problem-based mobile-assisted language learning with the conventional language learning approach (Communicative Language Teaching) in Iran. The experimental group (n=37) went through PBL instruction and the control group (n=33) went through conventional instruction. The results of quantitative data analysis after 26 sessions of treatment revealed that PBL could positively affect participants' knowledge of grammar, vocabulary, spoken fluency, and pronunciation; however, in terms of task achievement, no significant effect was found. This study can have pedagogical implications for language teachers, and material developers.

Keywords: problem-based learning, spoken intelligibility, Iranian EFL context, cognitive learning

Procedia PDF Downloads 175
7036 Variation Theory and Mixed Instructional Approaches: Advancing Conceptual Understanding in Geometry

Authors: Belete Abebaw, Mulugeta Atinafu, Awoke Shishigu

Abstract:

The study aimed to examine students’ problem-solving skills through mixed instruction (variation theory based Geogerba assisted problem-solving instructional approaches). A total of 125 students divided into 4 intact groups participated in the study. The study employed a quasi-experimental research design. Three intact groups were randomly assigned as a treatment group, while one group was taken as a comparison group. Each of the groups took a specific instructional approach, while the comparison group proceeded as usual without any changes to the instructional process for all sessions. Both pre and post problem-solving tests were administered to all groups. To analyze the data and examine the differences (if any) in each group, ANCOVA and Paired samples t-tests were employed. There was a significant mean difference between students pre-test and post-test in their conceptual understanding of each treatment group. Furthermore, the mixed treatment had a large mean difference. It was recommended that teachers give attention to using variation theory-based geometry problem-solving approaches for students’ better understanding. Administrators should emphasize launching Geogebra software through IT labs in schools, and government officials should appreciate the implementation of technology in schools.

Keywords: conceptual understanding, Geogebra, learning geometry, problem solving approaches, variation theory

Procedia PDF Downloads 25
7035 AHP and TOPSIS Methods for Supplier Selection Problem in Medical Devices Company

Authors: Sevde D. Karayel, Ediz Atmaca

Abstract:

Supplier selection subject is vital because of development competitiveness and performance of firms which have right, rapid and with low cost procurement. Considering the fact that competition between firms is no longer on their supply chains, hence it is very clear that performance of the firms’ not only depend on their own success but also success of all departments in supply chain. For this purpose, firms want to work with suppliers which are cost effective, flexible in terms of demand and high quality level for customer satisfaction. However, diversification and redundancy of their expectations from suppliers, supplier selection problems need to be solved as a hard problem. In this study, supplier selection problem is discussed for critical piece, which is using almost all production of products in and has troubles with lead time from supplier, in a firm that produces medical devices. Analyzing policy in the current situation of the firm in the supplier selection indicates that supplier selection is made based on the purchasing department experience and other authorized persons’ general judgments. Because selection do not make based on the analytical methods, it is caused disruptions in production, lateness and extra cost. To solve the problem, AHP and TOPSIS which are multi-criteria decision making techniques, which are effective, easy to implement and can analyze many criteria simultaneously, are used to make a selection among alternative suppliers.

Keywords: AHP-TOPSIS methods, multi-criteria decision making, supplier selection problem, supply chain management

Procedia PDF Downloads 264
7034 Finite Eigenstrains in Nonlinear Elastic Solid Wedges

Authors: Ashkan Golgoon, Souhayl Sadik, Arash Yavari

Abstract:

Eigenstrains in nonlinear solids are created due to anelastic effects such as non-uniform temperature distributions, growth, remodeling, and defects. Eigenstrains understanding is indispensable, as they can generate residual stresses and strongly affect the overall response of solids. Here, we study the residual stress and deformation fields of an incompressible isotropic infinite wedge with a circumferentially-symmetric distribution of finite eigenstrains. We construct a material manifold, whose Riemannian metric explicitly depends on the eigenstrain distribution, thereby we turn the problem into a classical nonlinear elasticity problem, where we find an embedding of the Riemannian material manifold into the ambient Euclidean space. In particular, we find exact solutions for the residual stress and deformation fields of a neo-Hookean wedge having a symmetric inclusion with finite radial and circumferential eigenstrains. Moreover, we numerically solve a similar problem when a symmetric Mooney-Rivlin inhomogeneity with finite eigenstrains is placed in a neo-Hookean wedge. Generalization of the eigenstrain problem to other geometries are also discussed.

Keywords: finite eigenstrains, geometric mechanics, inclusion, inhomogeneity, nonlinear elasticity

Procedia PDF Downloads 255
7033 Secure Bio Semantic Computing Scheme

Authors: Hiroshi Yamaguchi, Phillip C. Y. Sheu, Ryo Fujita, Shigeo Tsujii

Abstract:

In this paper, the secure BioSemantic Scheme is presented to bridge biological/biomedical research problems and computational solutions via semantic computing. Due to the diversity of problems in various research fields, the semantic capability description language (SCDL) plays and important role as a common language and generic form for problem formalization. SCDL is expected the essential for future semantic and logical computing in Biosemantic field. We show several example to Biomedical problems in this paper. Moreover, in the coming age of cloud computing, the security problem is considered to be crucial issue and we presented a practical scheme to cope with this problem.

Keywords: biomedical applications, private information retrieval (PIR), semantic capability description language (SCDL), semantic computing

Procedia PDF Downloads 390
7032 Elastohydrodynamic Lubrication Study Using Discontinuous Finite Volume Method

Authors: Prawal Sinha, Peeyush Singh, Pravir Dutt

Abstract:

Problems in elastohydrodynamic lubrication have attracted a lot of attention in the last few decades. Solving a two-dimensional problem has always been a big challenge. In this paper, a new discontinuous finite volume method (DVM) for two-dimensional point contact Elastohydrodynamic Lubrication (EHL) problem has been developed and analyzed. A complete algorithm has been presented for solving such a problem. The method presented is robust and easily parallelized in MPI architecture. GMRES technique is implemented to solve the matrix obtained after the formulation. A new approach is followed in which discontinuous piecewise polynomials are used for the trail functions. It is natural to assume that the advantages of using discontinuous functions in finite element methods should also apply to finite volume methods. The nature of the discontinuity of the trail function is such that the elements in the corresponding dual partition have the smallest support as compared with the Classical finite volume methods. Film thickness calculation is done using singular quadrature approach. Results obtained have been presented graphically and discussed. This method is well suited for solving EHL point contact problem and can probably be used as commercial software.

Keywords: elastohydrodynamic, lubrication, discontinuous finite volume method, GMRES technique

Procedia PDF Downloads 257
7031 Asymptotic Expansion of Double Oscillatory Integrals: Contribution of Non Stationary Critical Points of the Second Kind

Authors: Abdallah Benaissa

Abstract:

In this paper, we consider the problem of asymptotics of double oscillatory integrals in the case of critical points of the second kind, the order of contact between the boundary and a level curve of the phase being even, the situation when the order of contact is odd will be studied in other occasions. Complete asymptotic expansions will be derived and the coefficient of the leading term will be computed in terms of the original data of the problem. A multitude of people have studied this problem using a variety of methods, but only in a special case when the order of contact is minimal: the more cited papers are a paper of Jones and Kline and an other one of Chako. These integrals are encountered in many areas of science, especially in problems of diffraction of optics.

Keywords: asymptotic expansion, double oscillatory integral, critical point of the second kind, optics diffraction

Procedia PDF Downloads 350
7030 Singularity Theory in Yakam Matrix by Multiparameter Bifurcation Interfacial in Coupled Problem

Authors: Leonard Kabeya Mukeba Yakasham

Abstract:

The theoretical machinery from singularity theory introduced by Glolubitsky, Stewart, and Schaeffer, to study equivariant bifurcation problem is completed and expanded wile generalized to the multiparameter context. In this setting the finite deterinancy theorem or normal forms, the stability of equivariant bifurcation problem, and the structural stability of universal unfolding are discussed. With Yakam Matrix the solutions are limited for some partial differential equations stochastic nonlinear of the open questions in singularity artificial intelligence for future.

Keywords: equivariant bifurcation, symmetry singularity, equivariant jets and transversality; normal forms, universal unfolding instability, structural stability, artificial intelligence, pdens, yakam matrix

Procedia PDF Downloads 25
7029 Stochastic Control of Decentralized Singularly Perturbed Systems

Authors: Walid S. Alfuhaid, Saud A. Alghamdi, John M. Watkins, M. Edwin Sawan

Abstract:

Designing a controller for stochastic decentralized interconnected large scale systems usually involves a high degree of complexity and computation ability. Noise, observability, and controllability of all system states, connectivity, and channel bandwidth are other constraints to design procedures for distributed large scale systems. The quasi-steady state model investigated in this paper is a reduced order model of the original system using singular perturbation techniques. This paper results in an optimal control synthesis to design an observer based feedback controller by standard stochastic control theory techniques using Linear Quadratic Gaussian (LQG) approach and Kalman filter design with less complexity and computation requirements. Numerical example is given at the end to demonstrate the efficiency of the proposed method.

Keywords: decentralized, optimal control, output, singular perturb

Procedia PDF Downloads 370
7028 Modeling and Optimal Control of Hybrid Unmanned Aerial Vehicles with Wind Disturbance

Authors: Sunsoo Kim, Niladri Das, Raktim Bhattacharya

Abstract:

This paper addresses modeling and control of a six-degree-of-freedom unmanned aerial vehicle capable of vertical take-off and landing in the presence of wind disturbances. We design a hybrid vehicle that combines the benefits of both the fixed-wing and the rotary-wing UAVs. A non-linear model for the hybrid vehicle is rapidly built, combining rigid body dynamics, aerodynamics of wing, and dynamics of the motor and propeller. Further, we design a H₂ optimal controller to make the UAV robust to wind disturbances. We compare its results against that of proportional-integral-derivative and linear-quadratic regulator based control. Our proposed controller results in better performance in terms of root mean squared errors and time responses during two scenarios: hover and level- flight.

Keywords: hybrid UAVs, VTOL, aircraft modeling, H2 optimal control, wind disturbances

Procedia PDF Downloads 156