Search results for: ATM location problem
8754 Upon One Smoothing Problem in Project Management
Authors: Dimitri Golenko-Ginzburg
Abstract:
A CPM network project with deterministic activity durations, in which activities require homogenous resources with fixed capacities, is considered. The problem is to determine the optimal schedule of starting times for all network activities within their maximal allowable limits (in order not to exceed the network's critical time) to minimize the maximum required resources for the project at any point in time. In case when a non-critical activity may start only at discrete moments with the pregiven time span, the problem becomes NP-complete and an optimal solution may be obtained via a look-over algorithm. For the case when a look-over requires much computational time an approximate algorithm is suggested. The algorithm's performance ratio, i.e., the relative accuracy error, is determined. Experimentation has been undertaken to verify the suggested algorithm.Keywords: resource smoothing problem, CPM network, lookover algorithm, lexicographical order, approximate algorithm, accuracy estimate
Procedia PDF Downloads 3028753 Survey Paper on Graph Coloring Problem and Its Application
Authors: Prateek Chharia, Biswa Bhusan Ghosh
Abstract:
Graph coloring is one of the prominent concepts in graph coloring. It can be defined as a coloring of the various regions of the graph such that all the constraints are fulfilled. In this paper various graphs coloring approaches like greedy coloring, Heuristic search for maximum independent set and graph coloring using edge table is described. Graph coloring can be used in various real time applications like student time tabling generation, Sudoku as a graph coloring problem, GSM phone network.Keywords: graph coloring, greedy coloring, heuristic search, edge table, sudoku as a graph coloring problem
Procedia PDF Downloads 5398752 Structural Damage Detection via Incomplete Model Data Using Output Data Only
Authors: Ahmed Noor Al-qayyim, Barlas Özden Çağlayan
Abstract:
Structural failure is caused mainly by damage that often occurs on structures. Many researchers focus on obtaining very efficient tools to detect the damage in structures in the early state. In the past decades, a subject that has received considerable attention in literature is the damage detection as determined by variations in the dynamic characteristics or response of structures. This study presents a new damage identification technique. The technique detects the damage location for the incomplete structure system using output data only. The method indicates the damage based on the free vibration test data by using “Two Points - Condensation (TPC) technique”. This method creates a set of matrices by reducing the structural system to two degrees of freedom systems. The current stiffness matrices are obtained from optimization of the equation of motion using the measured test data. The current stiffness matrices are compared with original (undamaged) stiffness matrices. High percentage changes in matrices’ coefficients lead to the location of the damage. TPC technique is applied to the experimental data of a simply supported steel beam model structure after inducing thickness change in one element. Where two cases are considered, the method detects the damage and determines its location accurately in both cases. In addition, the results illustrate that these changes in stiffness matrix can be a useful tool for continuous monitoring of structural safety using ambient vibration data. Furthermore, its efficiency proves that this technique can also be used for big structures.Keywords: damage detection, optimization, signals processing, structural health monitoring, two points–condensation
Procedia PDF Downloads 3658751 A General Variable Neighborhood Search Algorithm to Minimize Makespan of the Distributed Permutation Flowshop Scheduling Problem
Authors: G. M. Komaki, S. Mobin, E. Teymourian, S. Sheikh
Abstract:
This paper addresses minimizing the makespan of the distributed permutation flow shop scheduling problem. In this problem, there are several parallel identical factories or flowshops each with series of similar machines. Each job should be allocated to one of the factories and all of the operations of the jobs should be performed in the allocated factory. This problem has recently gained attention and due to NP-Hard nature of the problem, metaheuristic algorithms have been proposed to tackle it. Majority of the proposed algorithms require large computational time which is the main drawback. In this study, a general variable neighborhood search algorithm (GVNS) is proposed where several time-saving schemes have been incorporated into it. Also, the GVNS uses the sophisticated method to change the shaking procedure or perturbation depending on the progress of the incumbent solution to prevent stagnation of the search. The performance of the proposed algorithm is compared to the state-of-the-art algorithms based on standard benchmark instances.Keywords: distributed permutation flow shop, scheduling, makespan, general variable neighborhood search algorithm
Procedia PDF Downloads 3548750 Simulation Model of Induction Heating in COMSOL Multiphysics
Authors: K. Djellabi, M. E. H. Latreche
Abstract:
The induction heating phenomenon depends on various factors, making the problem highly nonlinear. The mathematical analysis of this problem in most cases is very difficult and it is reduced to simple cases. Another knowledge of induction heating systems is generated in production environments, but these trial-error procedures are long and expensive. The numerical models of induction heating problem are another approach to reduce abovementioned drawbacks. This paper deals with the simulation model of induction heating problem. The simulation model of induction heating system in COMSOL Multiphysics is created. In this work we present results of numerical simulations of induction heating process in pieces of cylindrical shapes, in an inductor with four coils. The modeling of the inducting heating process was made with the software COMSOL Multiphysics Version 4.2a, for the study we present the temperature charts.Keywords: induction heating, electromagnetic field, inductor, numerical simulation, finite element
Procedia PDF Downloads 3168749 Identification of Impact Load and Partial System Parameters Using 1D-CNN
Authors: Xuewen Yu, Danhui Dan
Abstract:
The identification of impact load and some hard-to-obtain system parameters is crucial for the activities of analysis, validation, and evaluation in the engineering field. This paper proposes a method that utilizes neural networks based on 1D-CNN to identify the impact load and partial system parameters from measured responses. To this end, forward computations are conducted to provide datasets consisting of the triples (parameter θ, input u, output y). Then neural networks are trained to learn the mapping from input to output, fu|{θ} : y → u, as well as from input and output to parameter, fθ : (u, y) → θ. Afterward, feeding the trained neural networks the measured output response, the input impact load and system parameter can be calculated, respectively. The method is tested on two simulated examples and shows sound accuracy in estimating the impact load (waveform and location) and system parameters.Keywords: convolutional neural network, impact load identification, system parameter identification, inverse problem
Procedia PDF Downloads 1228748 Relation of Black Carbon Aerosols and Atmospheric Boundary Layer Height during Wet Removal Processes over a Semi Urban Location
Authors: M. Ashok Williams, T. V. Lakshmi Kumar
Abstract:
The life cycle of Black carbon aerosols depends on their physical removal processes from the atmosphere during the precipitation events. Black Carbon (BC) mass concentration has been analysed during rainy and non-rainy days of Northeast (NE) Monsoon months of the years 2015 and 2017 over a semi-urban environment near Chennai (12.81 N, 80.03 E), located on the east coast of India. BC, measured using an Aethalometer (AE-31) has been related to the atmospheric boundary layer height (BLH) obtained from the ERA Interim Reanalysis data during rainy and non-rainy days on monthly mean basis to understand the wet removal of BC over the study location. The study reveals that boundary layer height has a profound effect on the BC concentration on rainy days and non rainy days. It is found that the BC concentration in the night time is lower on rainy days compared to non rainy days owing to wash out on rainy days and the boundary layer height remaining nearly the same on rainy and non rainy days. On the other hand, in the daytime, it is found that the BC concentration remains nearly the same on rainy and non rainy days whereas the boundary layer height is lower on rainy days compared to non rainy days. This reveals that in daytime, lower boundary layer heights compensate for the wet removal effect on BC concentration on rainy days. A quantitative relation is found between the product of BC and BLH during rainy and non-rainy days which indicates the extent of redistribution of BC during non-rainy days when compared to the rainy days. Further work on the wet removal processes of the BC is in progress considering the individual rain events and other related parameters like wind speed.Keywords: black carbon aerosols, atmospheric boundary layer, scavenging processes, tropical coastal location
Procedia PDF Downloads 1528747 Solutions of Fuzzy Transportation Problem Using Best Candidates Method and Different Ranking Techniques
Authors: M. S. Annie Christi
Abstract:
Transportation Problem (TP) is based on supply and demand of commodities transported from one source to the different destinations. Usual methods for finding solution of TPs are North-West Corner Rule, Least Cost Method Vogel’s Approximation Method etc. The transportation costs tend to vary at each time. We can use fuzzy numbers which would give solution according to this situation. In this study the Best Candidate Method (BCM) is applied. For ranking Centroid Ranking Technique (CRT) and Robust Ranking Technique have been adopted to transform the fuzzy TP and the above methods are applied to EDWARDS Vacuum Company, Crawley, in West Sussex in the United Kingdom. A Comparative study is also given. We see that the transportation cost can be minimized by the application of CRT under BCM.Keywords: best candidate method, centroid ranking technique, fuzzy transportation problem, robust ranking technique, transportation problem
Procedia PDF Downloads 2948746 Site Analysis’ Importance as a Valid Factor in Building Design
Authors: Mekwa Eme, Anya chukwuma
Abstract:
The act of evaluating a particular site physically and socially in order to create a good design solution that will address the physical and interior environment of the location is known as architectural site analysis. This essay will describe site analysis as a useful design component. According to the introduction and supporting research, site evaluation and analysis are crucial to good design in terms of topography, orientation, site size, accessibility, rainfall, wind direction, and times of sunrise and sunset. Methodology: Both quantitative and qualitative analyses are used in this paper. The primary and secondary types of data collection are as follows. This information was gathered via the case study approach, already published literature, journals, the internet, a local poll, oral interviews, inquiries, and in-person interviews. The purpose of this is to clarify the benefits of site analysis for the design process and its implications for the working or building stage. Results: Each site's criteria are unique in terms of things like soil, plants, trees, accessibility, topography, and security. This will make it easier for the architect and environmentalist to decide on the idea, shape, and supporting structures of the design. It is crucial because before any design work is done, the nature of the target location will be determined through site visits and research. The location, contours, site features, and accessibility are just a few of the topics included in this site study. In order for students and working architects to understand the nature of the site they will be working on, site analysis is a key component of architectural education. The building's orientation, the site's circulation, and the sustainability of the site may all be determined with thorough research of the site's features.Keywords: analysis, climate, statistics, design
Procedia PDF Downloads 2498745 A Study on Shear Field Test Method in Timber Shear Modulus Determination Using Stereo Vision System
Authors: Niaz Gharavi, Hexin Zhang
Abstract:
In the structural timber design, the shear modulus of the timber beam is an important factor that needs to be determined accurately. According to BS EN 408, shear modulus can be determined using torsion test or shear field test method. Although torsion test creates pure shear status in the beam, it does not represent the real-life situation when the beam is in the service. On the other hand, shear field test method creates similar loading situation as in reality. The latter method is based on shear distortion measurement of the beam at the zone with the constant transverse load in the standardized four-point bending test as indicated in BS EN 408. Current testing practice code advised using two metallic arms act as an instrument to measure the diagonal displacement of the constructing square. Timber is not a homogenous material, but a heterogeneous and this characteristic makes timber to undergo a non-uniform deformation. Therefore, the dimensions and the location of the constructing square in the area with the constant transverse force might alter the shear modulus determination. This study aimed to investigate the impact of the shape, size, and location of the square in the shear field test method. A binocular stereo vision system was developed to capture the 3D displacement of a grid of target points. This approach is an accurate and non-contact method to extract the 3D coordination of targeted object using two cameras. Two group of three glue laminated beams were produced and tested by the mean of four-point bending test according to BS EN 408. Group one constructed using two materials, laminated bamboo lumber and structurally graded C24 timber and group two consisted only structurally graded C24 timber. Analysis of Variance (ANOVA) was performed on the acquired data to evaluate the significance of size and location of the square in the determination of shear modulus of the beam. The results have shown that the size of the square is an affecting factor in shear modulus determination. However, the location of the square in the area with the constant shear force does not affect the shear modulus.Keywords: shear field test method, BS EN 408, timber shear modulus, photogrammetry approach
Procedia PDF Downloads 2118744 Rodriguez Diego, Del Valle Martin, Hargreaves Matias, Riveros Jose Luis
Authors: Nathainail Bashir, Neil Anderson
Abstract:
The objective of this study site was to investigate the current state of the practice with regards to karst detection methods and recommend the best method and pattern of arrays to acquire the desire results. Proper site investigation in karst prone regions is extremely valuable in determining the location of possible voids. Two geophysical techniques were employed: multichannel analysis of surface waves (MASW) and electric resistivity tomography (ERT).The MASW data was acquired at each test location using different array lengths and different array orientations (to increase the probability of getting interpretable data in karst terrain). The ERT data were acquired using a dipole-dipole array consisting of 168 electrodes. The MASW data was interpreted (re: estimated depth to physical top of rock) and used to constrain and verify the interpretation of the ERT data. The ERT data indicates poorer quality MASW data were acquired in areas where there was significant local variation in the depth to top of rock.Keywords: dipole-dipole, ERT, Karst terrains, MASW
Procedia PDF Downloads 3158743 Distributed Coverage Control by Robot Networks in Unknown Environments Using a Modified EM Algorithm
Authors: Mohammadhosein Hasanbeig, Lacra Pavel
Abstract:
In this paper, we study a distributed control algorithm for the problem of unknown area coverage by a network of robots. The coverage objective is to locate a set of targets in the area and to minimize the robots’ energy consumption. The robots have no prior knowledge about the location and also about the number of the targets in the area. One efficient approach that can be used to relax the robots’ lack of knowledge is to incorporate an auxiliary learning algorithm into the control scheme. A learning algorithm actually allows the robots to explore and study the unknown environment and to eventually overcome their lack of knowledge. The control algorithm itself is modeled based on game theory where the network of the robots use their collective information to play a non-cooperative potential game. The algorithm is tested via simulations to verify its performance and adaptability.Keywords: distributed control, game theory, multi-agent learning, reinforcement learning
Procedia PDF Downloads 4578742 Robust Optimisation Model and Simulation-Particle Swarm Optimisation Approach for Vehicle Routing Problem with Stochastic Demands
Authors: Mohanad Al-Behadili, Djamila Ouelhadj
Abstract:
In this paper, a specific type of vehicle routing problem under stochastic demand (SVRP) is considered. This problem is of great importance because it models for many of the real world vehicle routing applications. This paper used a robust optimisation model to solve the problem along with the novel Simulation-Particle Swarm Optimisation (Sim-PSO) approach. The proposed Sim-PSO approach is based on the hybridization of the Monte Carlo simulation technique with the PSO algorithm. A comparative study between the proposed model and the Sim-PSO approach against other solution methods in the literature has been given in this paper. This comparison including the Analysis of Variance (ANOVA) to show the ability of the model and solution method in solving the complicated SVRP. The experimental results show that the proposed model and Sim-PSO approach has a significant impact on the obtained solution by providing better quality solutions comparing with well-known algorithms in the literature.Keywords: stochastic vehicle routing problem, robust optimisation model, Monte Carlo simulation, particle swarm optimisation
Procedia PDF Downloads 2778741 An Attempt at the Multi-Criterion Classification of Small Towns
Authors: Jerzy Banski
Abstract:
The basic aim of this study is to discuss and assess different classifications and research approaches to small towns that take their social and economic functions into account, as well as relations with surrounding areas. The subject literature typically includes three types of approaches to the classification of small towns: 1) the structural, 2) the location-related, and 3) the mixed. The structural approach allows for the grouping of towns from the point of view of the social, cultural and economic functions they discharge. The location-related approach draws on the idea of there being a continuum between the center and the periphery. A mixed classification making simultaneous use of the different approaches to research brings the most information to bear in regard to categories of the urban locality. Bearing in mind the approaches to classification, it is possible to propose a synthetic method for classifying small towns that takes account of economic structure, location and the relationship between the towns and their surroundings. In the case of economic structure, the small centers may be divided into two basic groups – those featuring a multi-branch structure and those that are specialized economically. A second element of the classification reflects the locations of urban centers. Two basic types can be identified – the small town within the range of impact of a large agglomeration, or else the town outside such areas, which is to say located peripherally. The third component of the classification arises out of small towns’ relations with their surroundings. In consequence, it is possible to indicate 8 types of small-town: from local centers enjoying good accessibility and a multi-branch economic structure to peripheral supra-local centers characterised by a specialized economic structure.Keywords: small towns, classification, functional structure, localization
Procedia PDF Downloads 1828740 Comparing Xbar Charts: Conventional versus Reweighted Robust Estimation Methods for Univariate Data Sets
Authors: Ece Cigdem Mutlu, Burak Alakent
Abstract:
Maintaining the quality of manufactured products at a desired level depends on the stability of process dispersion and location parameters and detection of perturbations in these parameters as promptly as possible. Shewhart control chart is the most widely used technique in statistical process monitoring to monitor the quality of products and control process mean and variability. In the application of Xbar control charts, sample standard deviation and sample mean are known to be the most efficient conventional estimators in determining process dispersion and location parameters, respectively, based on the assumption of independent and normally distributed datasets. On the other hand, there is no guarantee that the real-world data would be normally distributed. In the cases of estimated process parameters from Phase I data clouded with outliers, efficiency of traditional estimators is significantly reduced, and performance of Xbar charts are undesirably low, e.g. occasional outliers in the rational subgroups in Phase I data set may considerably affect the sample mean and standard deviation, resulting a serious delay in detection of inferior products in Phase II. For more efficient application of control charts, it is required to use robust estimators against contaminations, which may exist in Phase I. In the current study, we present a simple approach to construct robust Xbar control charts using average distance to the median, Qn-estimator of scale, M-estimator of scale with logistic psi-function in the estimation of process dispersion parameter, and Harrell-Davis qth quantile estimator, Hodge-Lehmann estimator and M-estimator of location with Huber psi-function and logistic psi-function in the estimation of process location parameter. Phase I efficiency of proposed estimators and Phase II performance of Xbar charts constructed from these estimators are compared with the conventional mean and standard deviation statistics both under normality and against diffuse-localized and symmetric-asymmetric contaminations using 50,000 Monte Carlo simulations on MATLAB. Consequently, it is found that robust estimators yield parameter estimates with higher efficiency against all types of contaminations, and Xbar charts constructed using robust estimators have higher power in detecting disturbances, compared to conventional methods. Additionally, utilizing individuals charts to screen outlier subgroups and employing different combination of dispersion and location estimators on subgroups and individual observations are found to improve the performance of Xbar charts.Keywords: average run length, M-estimators, quality control, robust estimators
Procedia PDF Downloads 1908739 A Coordinate-Based Heuristic Route Search Algorithm for Delivery Truck Routing Problem
Authors: Ahmed Tarek, Ahmed Alveed
Abstract:
Vehicle routing problem is a well-known re-search avenue in computing. Modern vehicle routing is more focused with the GPS-based coordinate system, as the state-of-the-art vehicle, and trucking systems are equipped with digital navigation. In this paper, a new two dimensional coordinate-based algorithm for addressing the vehicle routing problem for a supply chain network is proposed and explored, and the algorithm is compared with other available, and recently devised heuristics. For the algorithms discussed, which includes the pro-posed coordinate-based search heuristic as well, the advantages and the disadvantages associated with the heuristics are explored. The proposed algorithm is studied from the stand point of a small supermarket chain delivery network that supplies to its stores in four different states around the East Coast area, and is trying to optimize its trucking delivery cost. Minimizing the delivery cost for the supply network of a supermarket chain is important to ensure its business success.Keywords: coordinate-based optimal routing, Hamiltonian Circuit, heuristic algorithm, traveling salesman problem, vehicle routing problem
Procedia PDF Downloads 1478738 Use of Cyber-Physical Devices for the Implementation of Virtual and Augmented Realities in Bridge Construction
Authors: Muhammmad Fawad
Abstract:
The bridge construction industry has been revolutionized by the applications of Virtual Reality (VR) and Augmented Reality (AR). In this article, the author has focused on the field applications of digital technologies in structural, especially in bridge engineering. This research analyzed the use of VR/AR for the assessment of bridge concepts. For this purpose, the author has used Cyber-Physical Devices, i.e., Oculus Quest (OQ) for the implementation of VR, Trimble Microsoft HoloLens (THL), and Trimble Site Vision (TSV) for the implementation of AR/MR by visualizing the models of bridge planned to be constructed in Poland. The visualization of the models in Extended Reality (XR) is based on the development of BIM models of the bridge, which are further uploaded to the platforms required to implement these models in XR. This research helped to implement the models in MR so a bridge with a 1:1 scale at the exact location was placed, and authorities were presented with the possibility to visualize the exact scale and location of the bridge before its construction.Keywords: augmented reality, virtual reality, HoloLens, BIM, bridges
Procedia PDF Downloads 1228737 Optimal Relaxation Parameters for Obtaining Efficient Iterative Methods for the Solution of Electromagnetic Scattering Problems
Authors: Nadaniela Egidi, Pierluigi Maponi
Abstract:
The approximate solution of a time-harmonic electromagnetic scattering problem for inhomogeneous media is required in several application contexts, and its two-dimensional formulation is a Fredholm integral equation of the second kind. This integral equation provides a formulation for the direct scattering problem, but it has to be solved several times also in the numerical solution of the corresponding inverse scattering problem. The discretization of this Fredholm equation produces large and dense linear systems that are usually solved by iterative methods. In order to improve the efficiency of these iterative methods, we use the Symmetric SOR preconditioning, and we propose an algorithm for the evaluation of the associated relaxation parameter. We show the efficiency of the proposed algorithm by several numerical experiments, where we use two Krylov subspace methods, i.e., Bi-CGSTAB and GMRES.Keywords: Fredholm integral equation, iterative method, preconditioning, scattering problem
Procedia PDF Downloads 1038736 Benders Decomposition Approach to Solve the Hybrid Flow Shop Scheduling Problem
Authors: Ebrahim Asadi-Gangraj
Abstract:
Hybrid flow shop scheduling problem (HFS) contains sequencing in a flow shop where, at any stage, there exist one or more related or unrelated parallel machines. This production system is a common manufacturing environment in many real industries, such as the steel manufacturing, ceramic tile manufacturing, and car assembly industries. In this research, a mixed integer linear programming (MILP) model is presented for the hybrid flow shop scheduling problem, in which, the objective consists of minimizing the maximum completion time (makespan). For this purpose, a Benders Decomposition (BD) method is developed to solve the research problem. The proposed approach is tested on some test problems, small to moderate scale. The experimental results show that the Benders decomposition approach can solve the hybrid flow shop scheduling problem in a reasonable time, especially for small and moderate-size test problems.Keywords: hybrid flow shop, mixed integer linear programming, Benders decomposition, makespan
Procedia PDF Downloads 1888735 Calpoly Autonomous Transportation Experience: Software for Driverless Vehicle Operating on Campus
Authors: F. Tang, S. Boskovich, A. Raheja, Z. Aliyazicioglu, S. Bhandari, N. Tsuchiya
Abstract:
Calpoly Autonomous Transportation Experience (CATE) is a driverless vehicle that we are developing to provide safe, accessible, and efficient transportation of passengers throughout the Cal Poly Pomona campus for events such as orientation tours. Unlike the other self-driving vehicles that are usually developed to operate with other vehicles and reside only on the road networks, CATE will operate exclusively on walk-paths of the campus (potentially narrow passages) with pedestrians traveling from multiple locations. Safety becomes paramount as CATE operates within the same environment as pedestrians. As driverless vehicles assume greater roles in today’s transportation, this project will contribute to autonomous driving with pedestrian traffic in a highly dynamic environment. The CATE project requires significant interdisciplinary work. Researchers from mechanical engineering, electrical engineering and computer science are working together to attack the problem from different perspectives (hardware, software and system). In this abstract, we describe the software aspects of the project, with a focus on the requirements and the major components. CATE shall provide a GUI interface for the average user to interact with the car and access its available functionalities, such as selecting a destination from any origin on campus. We have developed an interface that provides an aerial view of the campus map, the current car location, routes, and the goal location. Users can interact with CATE through audio or manual inputs. CATE shall plan routes from the origin to the selected destination for the vehicle to travel. We will use an existing aerial map for the campus and convert it to a spatial graph configuration where the vertices represent the landmarks and edges represent paths that the car should follow with some designated behaviors (such as stay on the right side of the lane or follow an edge). Graph search algorithms such as A* will be implemented as the default path planning algorithm. D* Lite will be explored to efficiently recompute the path when there are any changes to the map. CATE shall avoid any static obstacles and walking pedestrians within some safe distance. Unlike traveling along traditional roadways, CATE’s route directly coexists with pedestrians. To ensure the safety of the pedestrians, we will use sensor fusion techniques that combine data from both lidar and stereo vision for obstacle avoidance while also allowing CATE to operate along its intended route. We will also build prediction models for pedestrian traffic patterns. CATE shall improve its location and work under a GPS-denied situation. CATE relies on its GPS to give its current location, which has a precision of a few meters. We have implemented an Unscented Kalman Filter (UKF) that allows the fusion of data from multiple sensors (such as GPS, IMU, odometry) in order to increase the confidence of localization. We also noticed that GPS signals can easily get degraded or blocked on campus due to high-rise buildings or trees. UKF can also help here to generate a better state estimate. In summary, CATE will provide on-campus transportation experience that coexists with dynamic pedestrian traffic. In future work, we will extend it to multi-vehicle scenarios.Keywords: driverless vehicle, path planning, sensor fusion, state estimate
Procedia PDF Downloads 1448734 A Versatile Algorithm to Propose Optimized Solutions to the Dengue Disease Problem
Authors: Fernando L. P. Santos, Luiz G. Lyra, Helenice O. Florentino, Daniela R. Cantane
Abstract:
Dengue is a febrile infectious disease caused by a virus of the family Flaviridae. It is transmitted by the bite of mosquitoes, usually of the genus Aedes aegypti. It occurs in tropical and subtropical areas of the world. This disease has been a major public health problem worldwide, especially in tropical countries such as Brazil, and its incidence has increased in recent years. Dengue is a subject of intense research. Efficient forms of mosquito control must be considered. In this work, the mono-objective optimal control problem was solved for analysing the dengue disease problem. Chemical and biological controls were considered in the mathematical aspect. This model describes the dynamics of mosquitoes in water and winged phases. We applied the genetic algorithms (GA) to obtain optimal strategies for the control of dengue. Numerical simulations have been performed to verify the versatility and the applicability of this algorithm. On the basis of the present results we may recommend the GA to solve optimal control problem with a large region of feasibility.Keywords: genetic algorithm, dengue, Aedes aegypti, biological control, chemical control
Procedia PDF Downloads 3498733 Probabilistic Approach to the Spatial Identification of the Environmental Sources behind Mortality Rates in Europe
Authors: Alina Svechkina, Boris A. Portnov
Abstract:
In line with a rapid increase in pollution sources and enforcement of stricter air pollution regulation, which lowers pollution levels, it becomes more difficult to identify actual risk sources behind the observed morbidity patterns, and new approaches are required to identify potential risks and take preventive actions. In the present study, we discuss a probabilistic approach to the spatial identification of a priori unidentified environmental health hazards. The underlying assumption behind the tested approach is that the observed adverse health patterns (morbidity, mortality) can become a source of information on the geographic location of environmental risk factors that stand behind them. Using this approach, we analyzed sources of environmental exposure using data on mortality rates available for the year 2015 for NUTS 3 (Nomenclature of Territorial Units for Statistics) subdivisions of the European Union. We identified several areas in the southwestern part of Europe as primary risk sources for the observed mortality patterns. Multivariate regressions, controlled by geographical location, climate conditions, GDP (gross domestic product) per capita, dependency ratios, population density, and the level of road freight revealed that mortality rates decline as a function of distance from the identified hazard location. We recommend the proposed approach an exploratory analysis tool for initial investigation of regional patterns of population morbidity patterns and factors behind it.Keywords: mortality, environmental hazards, air pollution, distance decay gradient, multi regression analysis, Europe, NUTS3
Procedia PDF Downloads 1678732 Data Collection with Bounded-Sized Messages in Wireless Sensor Networks
Authors: Min Kyung An
Abstract:
In this paper, we study the data collection problem in Wireless Sensor Networks (WSNs) adopting the two interference models: The graph model and the more realistic physical interference model known as Signal-to-Interference-Noise-Ratio (SINR). The main issue of the problem is to compute schedules with the minimum number of timeslots, that is, to compute the minimum latency schedules, such that data from every node can be collected without any collision or interference to a sink node. While existing works studied the problem with unit-sized and unbounded-sized message models, we investigate the problem with the bounded-sized message model, and introduce a constant factor approximation algorithm. To the best known of our knowledge, our result is the first result of the data collection problem with bounded-sized model in both interference models.Keywords: data collection, collision-free, interference-free, physical interference model, SINR, approximation, bounded-sized message model, wireless sensor networks
Procedia PDF Downloads 2218731 Demonstration of Logical Inconsistency in the Discussion of the Problem of Evil
Authors: Mohammad Soltani Renani
Abstract:
The problem of evil is one of the heated battlegrounds of the idea of theism and its critics. Since time immemorial and in various philosophical schools and religions, the belief in an Omniscient, Omnipotent, and Absolutely Good God has been considered inconsistent with the existence of the evil in the universe. The theist thinkers have generally adopted one of the following four ways for answering this problem: denial of the existence of evil or considering it to be relative, privation theory of evil, attribution of evil to something other than God, and depiction of an alternative picture of God. Defense or criticism of these alternative answers have given rise to an extensive and unending dispute. However, evaluation of the presupposition and context upon/in which a question is raised precedes offering an answer to it. This point in the discussion of the problem of evil is of paramount importance for both parties, i.e., questioners and answerers, that the attributes of knowledge, power, love, good-will, among others, can be supposed to be infinite only in the essence of the attributed and the domain of potentiality but what can be realized in the domain of actuality is always finite. Therefore, infinite nature of Divine Attributes and realization of evil belong to two spheres. Divine Attributes are infinite (absolute) in Divine Essence, but when they are created, each one becomes bounded by the other. This boundedness is a result of the state of being surrounded of the attributes by each other in finite world of possibility. Evil also appears in this limited world. This inconsistency leads to the collapse of the problem of evil from within: the place of infinity of the Divine Attributes, in the words of Muslim mystics, lies in the Holiest Manifestation [Feyze Aqdas] while evil emerges in the Holy Manifestation where the Divine Attributes become bounded by each other. This idea is neither a new answer to the problem of evil nor a defense of theism; rather it reveals a logical inconsistency in the discussion of the problem of evil.Keywords: problem of evil, infinity of divine attributes, boundedness of divine attributes, holiest manifestation, holy manifestation
Procedia PDF Downloads 1468730 The Solution of the Direct Problem of Electrical Prospecting with Direct Current Under Conditions of Ground Surface Relief
Authors: Balgaisha Mukanova, Tolkyn Mirgalikyzy
Abstract:
Theory of interpretation of electromagnetic fields studied in the electrical prospecting with direct current is mainly developed for the case of a horizontal surface observation. However in practice we often have to work in difficult terrain surface. Conducting interpretation without the influence of topography can cause non-existent anomalies on sections. This raises the problem of studying the impact of different shapes of ground surface relief on the results of electrical prospecting's research. This research examines the numerical solutions of the direct problem of electrical prospecting for two-dimensional and three-dimensional media, taking into account the terrain. The problem is solved using the method of integral equations. The density of secondary currents on the relief surface is obtained.Keywords: ground surface relief, method of integral equations, numerical method, electromagnetic
Procedia PDF Downloads 3638729 A Novel Approach of Power Transformer Diagnostic Using 3D FEM Parametrical Model
Authors: M. Brandt, A. Peniak, J. Makarovič, P. Rafajdus
Abstract:
This paper deals with a novel approach of power transformers diagnostics. This approach identifies the exact location and the range of a fault in the transformer and helps to reduce operation costs related to handling of the faulty transformer, its disassembly and repair. The advantage of the approach is a possibility to simulate healthy transformer and also all faults, which can occur in transformer during its operation without its disassembling, which is very expensive in practice. The approach is based on creating frequency dependent impedance of the transformer by sweep frequency response analysis measurements and by 3D FE parametrical modeling of the fault in the transformer. The parameters of the 3D FE model are the position and the range of the axial short circuit. Then, by comparing the frequency dependent impedances of the parametrical models with the measured ones, the location and the range of the fault is identified. The approach was tested on a real transformer and showed high coincidence between the real fault and the simulated one.Keywords: transformer, parametrical model of transformer, fault, sweep frequency response analysis, finite element method
Procedia PDF Downloads 4828728 A Practical Protection Method for Parallel Transmission-Lines Based on the Fault Travelling-Waves
Authors: Mohammad Reza Ebrahimi
Abstract:
In new restructured power systems, swift fault detection is very important. The parallel transmission-lines are vastly used in this kind of power systems because of high amount of energy transferring. In this paper, a method based on the comparison of two schemes, i.e., i) maximum magnitude of travelling-wave (TW) energy ii) the instants of maximum energy occurrence at the circuits of parallel transmission-line is proposed. Using the travelling-wave of fault in order to faulted line identification this method has noticeable operation time. Moreover, the algorithm can cover for identification of faults as external or internal faults. For an internal fault, the exact location of the fault can be estimated confidently. A lot of simulations have been done with PSCAD/EMTDC to verify the performance of the proposed algorithm.Keywords: travelling-wave, maximum energy, parallel transmission-line, fault location
Procedia PDF Downloads 1858727 Problem Solving in Mathematics Education: A Case Study of Nigerian Secondary School Mathematics Teachers’ Conceptions in Relation to Classroom Instruction
Authors: Carol Okigbo
Abstract:
Mathematical problem solving has long been accorded an important place in mathematics curricula at every education level in both advanced and emerging economies. Its classroom approaches have varied, such as teaching for problem-solving, teaching about problem-solving, and teaching mathematics through problem-solving. It requires engaging in tasks for which the solution methods are not eminent, making sense of problems and persevering in solving them by exhibiting processes, strategies, appropriate attitude, and adequate exposure. Teachers play important roles in helping students acquire competency in problem-solving; thus, they are expected to be good problem-solvers and have proper conceptions of problem-solving. Studies show that teachers’ conceptions influence their decisions about what to teach and how to teach. Therefore, how teachers view their roles in teaching problem-solving will depend on their pedagogical conceptions of problem-solving. If teaching problem-solving is a major component of secondary school mathematics instruction, as recommended by researchers and mathematics educators, then it is necessary to establish teachers’ conceptions, what they do, and how they approach problem-solving. This study is designed to determine secondary school teachers’ conceptions regarding mathematical problem solving, its current situation, how teachers’ conceptions relate to their demographics, as well as the interaction patterns in the mathematics classroom. There have been many studies of mathematics problem solving, some of which addressed teachers’ conceptions using single-method approaches, thereby presenting only limited views of this important phenomenon. To address the problem more holistically, this study adopted an integrated mixed methods approach which involved a quantitative survey, qualitative analysis of open-ended responses, and ethnographic observations of teachers in class. Data for the analysis came from a random sample of 327 secondary school mathematics teachers in two Nigerian states - Anambra State and Enugu State who completed a 45-item questionnaire. Ten of the items elicited demographic information, 11 items were open-ended questions, and 25 items were Likert-type questions. Of the 327 teachers who responded to the questionnaires, 37 were randomly selected and observed in their classes. Data analysis using ANOVA, t-tests, chi-square tests, and open coding showed that the teachers had different conceptions about problem-solving, which fall into three main themes: practice on exercises and word application problems, a process of solving mathematical problems, and a way of teaching mathematics. Teachers reported that no period is set aside for problem-solving; typically, teachers solve problems on the board, teach problem-solving strategies, and allow students time to struggle with problems on their own. The result shows a significant difference between male and female teachers’ conception of problems solving, a significant relationship among teachers’ conceptions and academic qualifications, and teachers who have spent ten years or more teaching mathematics were significantly different from the group with seven to nine years of experience in terms of their conceptions of problem-solving.Keywords: conceptions, education, mathematics, problem solving, teacher
Procedia PDF Downloads 768726 Optimizing Approach for Sifting Process to Solve a Common Type of Empirical Mode Decomposition Mode Mixing
Authors: Saad Al-Baddai, Karema Al-Subari, Elmar Lang, Bernd Ludwig
Abstract:
Empirical mode decomposition (EMD), a new data-driven of time-series decomposition, has the advantage of supposing that a time series is non-linear or non-stationary, as is implicitly achieved in Fourier decomposition. However, the EMD suffers of mode mixing problem in some cases. The aim of this paper is to present a solution for a common type of signals causing of EMD mode mixing problem, in case a signal suffers of an intermittency. By an artificial example, the solution shows superior performance in terms of cope EMD mode mixing problem comparing with the conventional EMD and Ensemble Empirical Mode decomposition (EEMD). Furthermore, the over-sifting problem is also completely avoided; and computation load is reduced roughly six times compared with EEMD, an ensemble number of 50.Keywords: empirical mode decomposition (EMD), mode mixing, sifting process, over-sifting
Procedia PDF Downloads 3938725 Location Detection of Vehicular Accident Using Global Navigation Satellite Systems/Inertial Measurement Units Navigator
Authors: Neda Navidi, Rene Jr. Landry
Abstract:
Vehicle tracking and accident recognizing are considered by many industries like insurance and vehicle rental companies. The main goal of this paper is to detect the location of a car accident by combining different methods. The methods, which are considered in this paper, are Global Navigation Satellite Systems/Inertial Measurement Units (GNSS/IMU)-based navigation and vehicle accident detection algorithms. They are expressed by a set of raw measurements, which are obtained from a designed integrator black box using GNSS and inertial sensors. Another concern of this paper is the definition of accident detection algorithm based on its jerk to identify the position of that accident. In fact, the results convinced us that, even in GNSS blockage areas, the position of the accident could be detected by GNSS/INS integration with 50% improvement compared to GNSS stand alone.Keywords: driver behavior monitoring, integration, IMU, GNSS, monitoring, tracking
Procedia PDF Downloads 234