Search results for: facility location selection problem
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11586

Search results for: facility location selection problem

10566 Image Reconstruction Method Based on L0 Norm

Authors: Jianhong Xiang, Hao Xiang, Linyu Wang

Abstract:

Compressed sensing (CS) has a wide range of applications in sparse signal reconstruction. Aiming at the problems of low recovery accuracy and long reconstruction time of existing reconstruction algorithms in medical imaging, this paper proposes a corrected smoothing L0 algorithm based on compressed sensing (CSL0). First, an approximate hyperbolic tangent function (AHTF) that is more similar to the L0 norm is proposed to approximate the L0 norm. Secondly, in view of the "sawtooth phenomenon" in the steepest descent method and the problem of sensitivity to the initial value selection in the modified Newton method, the use of the steepest descent method and the modified Newton method are jointly optimized to improve the reconstruction accuracy. Finally, the CSL0 algorithm is simulated on various images. The results show that the algorithm proposed in this paper improves the reconstruction accuracy of the test image by 0-0. 98dB.

Keywords: smoothed L0, compressed sensing, image processing, sparse reconstruction

Procedia PDF Downloads 111
10565 Soybean Based Farming System Assessment in Pasuruan East Java Indonesia

Authors: Mohammad Saeri, Noor Rizkiyah, Kambang Vetrani Asie, Titin Apung Atikah

Abstract:

The study aims to assess efficient specific-location soybean farming technology assembly by assisting the farmers in applying the suggested technology. Superimposed trial was conducted to know NPK fertilizer effect toward soybean growth and yield and soybean improved variety test for the dissemination of improved variety. The assessment was conducted at the farmers group of Sumber Rejeki, Kepulungan Village, Gempol Sub-district, Pasuruan Regency as the soybean central at Pasuruan area. The number of farmers involved in the study was 38 people with 25 ha soybean area. This study was held from July to October 2012.  The recommended technology package agreed at the socialization time and used in this research were: using Argomulyo variety seeds of 40 kg/ha, planting by drilling, planting by distance of 40x10 cm, deciding the seeds amount of 2-3 seeds per hole, and giving fertilization based on recommendation of East Java AIAT of 50 kg Urea, 100 kg SP-36 and 50 kg KCl.  Farmers around the research location were used as control group. Assessment on soybean farming system was considered effective because it could increase the production up to 38%. The farming analysis showed that the result collaborator farmers gained were positively higher than non-collaborator farmers with RC ratio of 2.03 and 1.54, respectively. Argomulyo variety has the prospect to be developed due to the high yield of about 2 tons/ha and the larger seeds. The NPK fertilization test at the soybean plants showed that the fertilization had minor effect on the yield.

Keywords: farming system, soybean, variety, location specific

Procedia PDF Downloads 169
10564 Promoting Authenticity in Employer Brands to Address the Global-Local Problem in Complex Organisations: The Case of a Developing Country

Authors: Saud Al Taj

Abstract:

Employer branding is considered as a useful tool for addressing the global-local problem facing complex organisations that have operations scattered across the globe and face challenges of dealing with the local environment alongside. Despite being an established field of study within the Western developed world, there is little empirical evidence concerning the relevance of employer branding to global companies that operate in the under-developed economies. This paper fills this gap by gaining rich insight into the implementation of employer branding programs in a foreign multinational operating in Pakistan dealing with the global-local problem. The study is qualitative in nature and employs semi-structured and focus group interviews with senior/middle managers and local frontline employees to deeply examine the phenomenon in case organisation. Findings suggest that authenticity is required in employer brands to enable them to respond to the local needs thereby leading to the resolution of the global-local problem. However, the role of signaling theory is key to the development of authentic employer brands as it stresses on the need to establish an efficient and effective signaling environment wherein signals travel in both directions (from signal designers to receivers and backwards) and facilitate firms with the global-local problem. The paper also identifies future avenues of research for the employer branding field.

Keywords: authenticity, counter-signals, employer branding, global-local problem, signaling theory

Procedia PDF Downloads 360
10563 Development and Modeling of a Geographic Information System Solar Flux in Adrar, Algeria

Authors: D. Benatiallah, A. Benatiallah, K. Bouchouicha, A. Harouz

Abstract:

The development and operation of renewable energy known an important development in the world with significant growth potential. Estimate the solar radiation on terrestrial geographic locality is of extreme importance, firstly to choose the appropriate site where to place solar systems (solar power plants for electricity generation, for example) and also for the design and performance analysis of any system using solar energy. In addition, solar radiation measurements are limited to a few areas only in Algeria. Thus, we use theoretical approaches to assess the solar radiation on a given location. The Adrar region is one of the most favorable sites for solar energy use with a medium flow that exceeds 7 kWh / m2 / d and saddle of over 3500 hours per year. Our goal in this work focuses on the creation of a data bank for the given data in the energy field of the Adrar region for the period of the year and the month then the integration of these data into a geographic Information System (GIS) to estimate the solar flux on a location on the map.

Keywords: Adrar, flow, GIS, deposit potential

Procedia PDF Downloads 365
10562 The Relationship between Confidence, Accuracy, and Decision Making in a Mobile Review Program

Authors: Carla Van De Sande, Jana Vandenberg

Abstract:

Just like physical skills, cognitive skills grow rusty over time unless they are regularly used and practiced, so academic breaks can have negative consequences on student learning and success. The Keeping in School Shape (KiSS) program is an engaging, accessible, and cost-effective intervention that harnesses the benefits of retrieval practice by using technology to help students maintain proficiency over breaks from school by delivering a daily review problem via text message or email. A growth mindset is promoted through feedback messages encouraging students to try again if they get a problem wrong and to take on a challenging problem if they get a problem correct. This paper reports on the relationship between confidence, accuracy, and decision-making during the implementation of the KiSS Program at a large university during winter break for students enrolled in an engineering introductory Calculus course sequence.

Keywords: growth mindset, learning loss, on-the-go learning, retrieval practice

Procedia PDF Downloads 202
10561 Interval Bilevel Linear Fractional Programming

Authors: F. Hamidi, N. Amiri, H. Mishmast Nehi

Abstract:

The Bilevel Programming (BP) model has been presented for a decision making process that consists of two decision makers in a hierarchical structure. In fact, BP is a model for a static two person game (the leader player in the upper level and the follower player in the lower level) wherein each player tries to optimize his/her personal objective function under dependent constraints; this game is sequential and non-cooperative. The decision making variables are divided between the two players and one’s choice affects the other’s benefit and choices. In other words, BP consists of two nested optimization problems with two objective functions (upper and lower) where the constraint region of the upper level problem is implicitly determined by the lower level problem. In real cases, the coefficients of an optimization problem may not be precise, i.e. they may be interval. In this paper we develop an algorithm for solving interval bilevel linear fractional programming problems. That is to say, bilevel problems in which both objective functions are linear fractional, the coefficients are interval and the common constraint region is a polyhedron. From the original problem, the best and the worst bilevel linear fractional problems have been derived and then, using the extended Charnes and Cooper transformation, each fractional problem can be reduced to a linear problem. Then we can find the best and the worst optimal values of the leader objective function by two algorithms.

Keywords: best and worst optimal solutions, bilevel programming, fractional, interval coefficients

Procedia PDF Downloads 439
10560 Non-Stationary Stochastic Optimization of an Oscillating Water Column

Authors: María L. Jalón, Feargal Brennan

Abstract:

A non-stationary stochastic optimization methodology is applied to an OWC (oscillating water column) to find the design that maximizes the wave energy extraction. Different temporal cycles are considered to represent the long-term variability of the wave climate at the site in the optimization problem. The results of the non-stationary stochastic optimization problem are compared against those obtained by a stationary stochastic optimization problem. The comparative analysis reveals that the proposed non-stationary optimization provides designs with a better fit to reality. However, the stationarity assumption can be adequate when looking at averaged system response.

Keywords: non-stationary stochastic optimization, oscillating water, temporal variability, wave energy

Procedia PDF Downloads 364
10559 Non-Dominated Sorting Genetic Algorithm (NSGA-II) for the Redistricting Problem in Mexico

Authors: Antonin Ponsich, Eric Alfredo Rincon Garcia, Roman Anselmo Mora Gutierrez, Miguel Angel Gutierrez Andrade, Sergio Gerardo De Los Cobos Silva, Pedro Lara Velzquez

Abstract:

The electoral zone design problem consists in redrawing the boundaries of legislative districts for electoral purposes in such a way that federal or state requirements are fulfilled. In Mexico, this process has been historically carried out by the National Electoral Institute (INE), by optimizing an integer nonlinear programming model, in which population equality and compactness of the designed districts are considered as two conflicting objective functions, while contiguity is included as a hard constraint. The solution technique used by the INE is a Simulated Annealing (SA) based algorithm, which handles the multi-objective nature of the problem through an aggregation function. The present work represents the first intent to apply a classical Multi-Objective Evolutionary Algorithm (MOEA), the second version of the Non-dominated Sorting Genetic Algorithm (NSGA-II), to this hard combinatorial problem. First results show that, when compared with the SA algorithm, the NSGA-II obtains promising results. The MOEA manages to produce well-distributed solutions over a wide-spread front, even though some convergence troubles for some instances constitute an issue, which should be corrected in future adaptations of MOEAs to the redistricting problem.

Keywords: multi-objective optimization, NSGA-II, redistricting, zone design problem

Procedia PDF Downloads 363
10558 Confidence Envelopes for Parametric Model Selection Inference and Post-Model Selection Inference

Authors: I. M. L. Nadeesha Jayaweera, Adao Alex Trindade

Abstract:

In choosing a candidate model in likelihood-based modeling via an information criterion, the practitioner is often faced with the difficult task of deciding just how far up the ranked list to look. Motivated by this pragmatic necessity, we construct an uncertainty band for a generalized (model selection) information criterion (GIC), defined as a criterion for which the limit in probability is identical to that of the normalized log-likelihood. This includes common special cases such as AIC & BIC. The method starts from the asymptotic normality of the GIC for the joint distribution of the candidate models in an independent and identically distributed (IID) data framework and proceeds by deriving the (asymptotically) exact distribution of the minimum. The calculation of an upper quantile for its distribution then involves the computation of multivariate Gaussian integrals, which is amenable to efficient implementation via the R package "mvtnorm". The performance of the methodology is tested on simulated data by checking the coverage probability of nominal upper quantiles and compared to the bootstrap. Both methods give coverages close to nominal for large samples, but the bootstrap is two orders of magnitude slower. The methodology is subsequently extended to two other commonly used model structures: regression and time series. In the regression case, we derive the corresponding asymptotically exact distribution of the minimum GIC invoking Lindeberg-Feller type conditions for triangular arrays and are thus able to similarly calculate upper quantiles for its distribution via multivariate Gaussian integration. The bootstrap once again provides a default competing procedure, and we find that similar comparison performance metrics hold as for the IID case. The time series case is complicated by far more intricate asymptotic regime for the joint distribution of the model GIC statistics. Under a Gaussian likelihood, the default in most packages, one needs to derive the limiting distribution of a normalized quadratic form for a realization from a stationary series. Under conditions on the process satisfied by ARMA models, a multivariate normal limit is once again achieved. The bootstrap can, however, be employed for its computation, whence we are once again in the multivariate Gaussian integration paradigm for upper quantile evaluation. Comparisons of this bootstrap-aided semi-exact method with the full-blown bootstrap once again reveal a similar performance but faster computation speeds. One of the most difficult problems in contemporary statistical methodological research is to be able to account for the extra variability introduced by model selection uncertainty, the so-called post-model selection inference (PMSI). We explore ways in which the GIC uncertainty band can be inverted to make inferences on the parameters. This is being attempted in the IID case by pivoting the CDF of the asymptotically exact distribution of the minimum GIC. For inference one parameter at a time and a small number of candidate models, this works well, whence the attained PMSI confidence intervals are wider than the MLE-based Wald, as expected.

Keywords: model selection inference, generalized information criteria, post model selection, Asymptotic Theory

Procedia PDF Downloads 82
10557 An Ant Colony Optimization Approach for the Pollution Routing Problem

Authors: P. Parthiban, Sonu Rajak, N. Kannan, R. Dhanalakshmi

Abstract:

This paper deals with the Vehicle Routing Problem (VRP) with environmental considerations which is called Pollution Routing Problem (PRP). The objective is to minimize the operational and environmental costs. It consists of routing a number of vehicles to serve a set of customers, and determining fuel consumption, driver wages and their speed on each route segment, while respecting the capacity constraints and time windows. In this context, we presented an Ant Colony Optimization (ACO) approach, combined with a Speed Optimization Algorithm (SOA) to solve the PRP. The proposed solution method consists of two stages. Stage one is to solve a Vehicle Routing Problem with Time Window (VRPTW) using ACO and in the second stage a SOA is run on the resulting VRPTW solutions. Given a vehicle route, the SOA consists of finding the optimal speed on each arc of the route in order to minimize an objective function comprising fuel consumption costs and driver wages. The proposed algorithm tested on benchmark problem, the preliminary results show that the proposed algorithm is able to provide good solutions.

Keywords: ant colony optimization, CO2 emissions, combinatorial optimization, speed optimization, vehicle routing

Procedia PDF Downloads 317
10556 Method for Selecting and Prioritising Smart Services in Manufacturing Companies

Authors: Till Gramberg, Max Kellner, Erwin Gross

Abstract:

This paper presents a comprehensive investigation into the topic of smart services and IIoT-Platforms, focusing on their selection and prioritization in manufacturing organizations. First, a literature review is conducted to provide a basic understanding of the current state of research in the area of smart services. Based on discussed and established definitions, a definition approach for this paper is developed. In addition, value propositions for smart services are identified based on the literature and expert interviews. Furthermore, the general requirements for the provision of smart services are presented. Subsequently, existing approaches for the selection and development of smart services are identified and described. In order to determine the requirements for the selection of smart services, expert opinions from successful companies that have already implemented smart services are collected through semi-structured interviews. Based on the results, criteria for the evaluation of existing methods are derived. The existing methods are then evaluated according to the identified criteria. Furthermore, a novel method for the selection of smart services in manufacturing companies is developed, taking into account the identified criteria and the existing approaches. The developed concept for the method is verified in expert interviews. The method includes a collection of relevant smart services identified in the literature. The actual relevance of the use cases in the industrial environment was validated in an online survey. The required data and sensors are assigned to the smart service use cases. The value proposition of the use cases is evaluated in an expert workshop using different indicators. Based on this, a comparison is made between the identified value proposition and the required data, leading to a prioritization process. The prioritization process follows an established procedure for evaluating technical decision-making processes. In addition to the technical requirements, the prioritization process includes other evaluation criteria such as the economic benefit, the conformity of the new service offering with the company strategy, or the customer retention enabled by the smart service. Finally, the method is applied and validated in an industrial environment. The results of these experiments are critically reflected upon and an outlook on future developments in the area of smart services is given. This research contributes to a deeper understanding of the selection and prioritization process as well as the technical considerations associated with smart service implementation in manufacturing organizations. The proposed method serves as a valuable guide for decision makers, helping them to effectively select the most appropriate smart services for their specific organizational needs.

Keywords: smart services, IIoT, industrie 4.0, IIoT-platform, big data

Procedia PDF Downloads 77
10555 A Hybrid Algorithm Based on Greedy Randomized Adaptive Search Procedure and Chemical Reaction Optimization for the Vehicle Routing Problem with Hard Time Windows

Authors: Imen Boudali, Marwa Ragmoun

Abstract:

The Vehicle Routing Problem with Hard Time Windows (VRPHTW) is a basic distribution management problem that models many real-world problems. The objective of the problem is to deliver a set of customers with known demands on minimum-cost vehicle routes while satisfying vehicle capacity and hard time windows for customers. In this paper, we propose to deal with our optimization problem by using a new hybrid stochastic algorithm based on two metaheuristics: Chemical Reaction Optimization (CRO) and Greedy Randomized Adaptive Search Procedure (GRASP). The first method is inspired by the natural process of chemical reactions enabling the transformation of unstable substances with excessive energy to stable ones. During this process, the molecules interact with each other through a series of elementary reactions to reach minimum energy for their existence. This property is embedded in CRO to solve the VRPHTW. In order to enhance the population diversity throughout the search process, we integrated the GRASP in our method. Simulation results on the base of Solomon’s benchmark instances show the very satisfactory performances of the proposed approach.

Keywords: Benchmark Problems, Combinatorial Optimization, Vehicle Routing Problem with Hard Time Windows, Meta-heuristics, Hybridization, GRASP, CRO

Procedia PDF Downloads 408
10554 Solving the Economic Load Dispatch Problem Using Differential Evolution

Authors: Alaa Sheta

Abstract:

Economic Load Dispatch (ELD) is one of the vital optimization problems in power system planning. Solving the ELD problems mean finding the best mixture of power unit outputs of all members of the power system network such that the total fuel cost is minimized while sustaining operation requirements limits satisfied across the entire dispatch phases. Many optimization techniques were proposed to solve this problem. A famous one is the Quadratic Programming (QP). QP is a very simple and fast method but it still suffer many problem as gradient methods that might trapped at local minimum solutions and cannot handle complex nonlinear functions. Numbers of metaheuristic algorithms were used to solve this problem such as Genetic Algorithms (GAs) and Particle Swarm Optimization (PSO). In this paper, another meta-heuristic search algorithm named Differential Evolution (DE) is used to solve the ELD problem in power systems planning. The practicality of the proposed DE based algorithm is verified for three and six power generator system test cases. The gained results are compared to existing results based on QP, GAs and PSO. The developed results show that differential evolution is superior in obtaining a combination of power loads that fulfill the problem constraints and minimize the total fuel cost. DE found to be fast in converging to the optimal power generation loads and capable of handling the non-linearity of ELD problem. The proposed DE solution is able to minimize the cost of generated power, minimize the total power loss in the transmission and maximize the reliability of the power provided to the customers.

Keywords: economic load dispatch, power systems, optimization, differential evolution

Procedia PDF Downloads 277
10553 Reverse Logistics Network Optimization for E-Commerce

Authors: Albert W. K. Tan

Abstract:

This research consolidates a comprehensive array of publications from peer-reviewed journals, case studies, and seminar reports focused on reverse logistics and network design. By synthesizing this secondary knowledge, our objective is to identify and articulate key decision factors crucial to reverse logistics network design for e-commerce. Through this exploration, we aim to present a refined mathematical model that offers valuable insights for companies seeking to optimize their reverse logistics operations. The primary goal of this research endeavor is to develop a comprehensive framework tailored to advising organizations and companies on crafting effective networks for their reverse logistics operations, thereby facilitating the achievement of their organizational goals. This involves a thorough examination of various network configurations, weighing their advantages and disadvantages to ensure alignment with specific business objectives. The key objectives of this research include: (i) Identifying pivotal factors pertinent to network design decisions within the realm of reverse logistics across diverse supply chains. (ii) Formulating a structured framework designed to offer informed recommendations for sound network design decisions applicable to relevant industries and scenarios. (iii) Propose a mathematical model to optimize its reverse logistics network. A conceptual framework for designing a reverse logistics network has been developed through a combination of insights from the literature review and information gathered from company websites. This framework encompasses four key stages in the selection of reverse logistics operations modes: (1) Collection, (2) Sorting and testing, (3) Processing, and (4) Storage. Key factors to consider in reverse logistics network design: I) Centralized vs. decentralized processing: Centralized processing, a long-standing practice in reverse logistics, has recently gained greater attention from manufacturing companies. In this system, all products within the reverse logistics pipeline are brought to a central facility for sorting, processing, and subsequent shipment to their next destinations. Centralization offers the advantage of efficiently managing the reverse logistics flow, potentially leading to increased revenues from returned items. Moreover, it aids in determining the most appropriate reverse channel for handling returns. On the contrary, a decentralized system is more suitable when products are returned directly from consumers to retailers. In this scenario, individual sales outlets serve as gatekeepers for processing returns. Considerations encompass the product lifecycle, product value and cost, return volume, and the geographic distribution of returns. II) In-house vs. third-party logistics providers: The decision between insourcing and outsourcing in reverse logistics network design is pivotal. In insourcing, a company handles the entire reverse logistics process, including material reuse. In contrast, outsourcing involves third-party providers taking on various aspects of reverse logistics. Companies may choose outsourcing due to resource constraints or lack of expertise, with the extent of outsourcing varying based on factors such as personnel skills and cost considerations. Based on the conceptual framework, the authors have constructed a mathematical model that optimizes reverse logistics network design decisions. The model will consider key factors identified in the framework, such as transportation costs, facility capacities, and lead times. The authors have employed mixed LP to find the optimal solutions that minimize costs while meeting organizational objectives.

Keywords: reverse logistics, supply chain management, optimization, e-commerce

Procedia PDF Downloads 31
10552 Production Plan and Technological Variants Optimization by Goal Programming Methods

Authors: Tunjo Perić, Franjo Bratić

Abstract:

In this paper the goal programming methodology for solving multiple objective problem of the technological variants and production plan optimization has been applied. The optimization criteria are determined and the multiple objective linear programming model for solving a problem of the technological variants and production plan optimization is formed and solved. Then the obtained results are analysed. The obtained results point out to the possibility of efficient application of the goal programming methodology in solving the problem of the technological variants and production plan optimization. The paper points out on the advantages of the application of the goal programming methodolohy compare to the Surrogat Worth Trade-off method in solving this problem.

Keywords: goal programming, multi objective programming, production plan, SWT method, technological variants

Procedia PDF Downloads 370
10551 A Parallel Algorithm for Solving the PFSP on the Grid

Authors: Samia Kouki

Abstract:

Solving NP-hard combinatorial optimization problems by exact search methods, such as Branch-and-Bound, may degenerate to complete enumeration. For that reason, exact approaches limit us to solve only small or moderate size problem instances, due to the exponential increase in CPU time when problem size increases. One of the most promising ways to reduce significantly the computational burden of sequential versions of Branch-and-Bound is to design parallel versions of these algorithms which employ several processors. This paper describes a parallel Branch-and-Bound algorithm called GALB for solving the classical permutation flowshop scheduling problem as well as its implementation on a Grid computing infrastructure. The experimental study of our distributed parallel algorithm gives promising results and shows clearly the benefit of the parallel paradigm to solve large-scale instances in moderate CPU time.

Keywords: grid computing, permutation flow shop problem, branch and bound, load balancing

Procedia PDF Downloads 280
10550 A Framework for Automated Nuclear Waste Classification

Authors: Seonaid Hume, Gordon Dobie, Graeme West

Abstract:

Detecting and localizing radioactive sources is a necessity for safe and secure decommissioning of nuclear facilities. An important aspect for the management of the sort-and-segregation process is establishing the spatial distributions and quantities of the waste radionuclides, their type, corresponding activity, and ultimately classification for disposal. The data received from surveys directly informs decommissioning plans, on-site incident management strategies, the approach needed for a new cell, as well as protecting the workforce and the public. Manual classification of nuclear waste from a nuclear cell is time-consuming, expensive, and requires significant expertise to make the classification judgment call. Also, in-cell decommissioning is still in its relative infancy, and few techniques are well-developed. As with any repetitive and routine tasks, there is the opportunity to improve the task of classifying nuclear waste using autonomous systems. Hence, this paper proposes a new framework for the automatic classification of nuclear waste. This framework consists of five main stages; 3D spatial mapping and object detection, object classification, radiological mapping, source localisation based on gathered evidence and finally, waste classification. The first stage of the framework, 3D visual mapping, involves object detection from point cloud data. A review of related applications in other industries is provided, and recommendations for approaches for waste classification are made. Object detection focusses initially on cylindrical objects since pipework is significant in nuclear cells and indeed any industrial site. The approach can be extended to other commonly occurring primitives such as spheres and cubes. This is in preparation of stage two, characterizing the point cloud data and estimating the dimensions, material, degradation, and mass of the objects detected in order to feature match them to an inventory of possible items found in that nuclear cell. Many items in nuclear cells are one-offs, have limited or poor drawings available, or have been modified since installation, and have complex interiors, which often and inadvertently pose difficulties when accessing certain zones and identifying waste remotely. Hence, this may require expert input to feature match objects. The third stage, radiological mapping, is similar in order to facilitate the characterization of the nuclear cell in terms of radiation fields, including the type of radiation, activity, and location within the nuclear cell. The fourth stage of the framework takes the visual map for stage 1, the object characterization from stage 2, and radiation map from stage 3 and fuses them together, providing a more detailed scene of the nuclear cell by identifying the location of radioactive materials in three dimensions. The last stage involves combining the evidence from the fused data sets to reveal the classification of the waste in Bq/kg, thus enabling better decision making and monitoring for in-cell decommissioning. The presentation of the framework is supported by representative case study data drawn from an application in decommissioning from a UK nuclear facility. This framework utilises recent advancements of the detection and mapping capabilities of complex radiation fields in three dimensions to make the process of classifying nuclear waste faster, more reliable, cost-effective and safer.

Keywords: nuclear decommissioning, radiation detection, object detection, waste classification

Procedia PDF Downloads 195
10549 Location Choice of Firms in an Unequal Length Streets Model: Game Theory Approach as an Extension of the Spoke Model

Authors: Kiumars Shahbazi, Salah Salimian, Abdolrahim Hashemi Dizaj

Abstract:

Locating is one of the key elements in success and survival of industrial centers and has great impact on cost reduction of establishment and launching of various economic activities. In this study, streets with unequal length model have been used that is the classic extension of Spoke model; however with unlimited number of streets with uneven lengths. The results showed that the spoke model is a special case of streets with unequal length model. According to the results of this study, if the strategy of enterprises and firms is to select both price and location, there would be no balance in the game. Furthermore, increased length of streets leads to increased profit of enterprises and with increased number of streets, the enterprises choose locations that are far from center (the maximum differentiation), and the enterprises' output will decrease. Moreover, the enterprise production rate will incline toward zero when the number of streets goes to infinity, and complete competition outcome will be achieved.

Keywords: locating, Nash equilibrium, streets with unequal length model, streets with unequal length model

Procedia PDF Downloads 196
10548 Introduction of Microbial Symbiosis in Genus of Tridacna and Kiwaidae with Insights into Aquaculture

Authors: Jincao Guo

Abstract:

Aquaculture plays a significant role in the diet of people in many regions. However, problems such as bioaccumulation have risen with the rapidly growing industry due to a lack of control in the feeding process, which brings uncertainty to the quality of the products. The paper tackles the problem by introducing the symbiosis of the Giant Clam (Tridacna) with photosynthetic algae and Yeti Crab (Kiwaidae) with chemosynthetic bacteria in molecular and developmental details. By combing the knowledge gained from the two models and past studies, innovative ideas such as using mass selection methods to domesticate and farm those symbiotic species, as well as improvements for the current farming methods, such as introducing algae feeding, are discussed. Further studies are needed, but experiments are worth conducting since it increases the variety of choices for consumers and can potentially improve the quality and efficiency of aquaculture.

Keywords: the giant clam Tridacna, yeti crab Kiwaidae, autotroph microbes, microbial symbiosis, aquaculture, bivalves, crustaceans, mollusk, photosynthesis, chemosynthesis

Procedia PDF Downloads 66
10547 An Ensemble Learning Method for Applying Particle Swarm Optimization Algorithms to Systems Engineering Problems

Authors: Ken Hampshire, Thomas Mazzuchi, Shahram Sarkani

Abstract:

As a subset of metaheuristics, nature-inspired optimization algorithms such as particle swarm optimization (PSO) have shown promise both in solving intractable problems and in their extensibility to novel problem formulations due to their general approach requiring few assumptions. Unfortunately, single instantiations of algorithms require detailed tuning of parameters and cannot be proven to be best suited to a particular illustrative problem on account of the “no free lunch” (NFL) theorem. Using these algorithms in real-world problems requires exquisite knowledge of the many techniques and is not conducive to reconciling the various approaches to given classes of problems. This research aims to present a unified view of PSO-based approaches from the perspective of relevant systems engineering problems, with the express purpose of then eliciting the best solution for any problem formulation in an ensemble learning bucket of models approach. The central hypothesis of the research is that extending the PSO algorithms found in the literature to real-world optimization problems requires a general ensemble-based method for all problem formulations but a specific implementation and solution for any instance. The main results are a problem-based literature survey and a general method to find more globally optimal solutions for any systems engineering optimization problem.

Keywords: particle swarm optimization, nature-inspired optimization, metaheuristics, systems engineering, ensemble learning

Procedia PDF Downloads 90
10546 Evaluation of Current Methods in Modelling and Analysis of Track with Jointed Rails

Authors: Hossein Askarinejad, Manicka Dhanasekar

Abstract:

In railway tracks, two adjacent rails are either welded or connected using bolted jointbars. In recent years the number of bolted rail joints is reduced by introduction of longer rail sections and by welding the rails at location of some joints. However, significant number of bolted rail joints remains in railways around the world as they are required to allow for rail thermal expansion or to provide electrical insulation in some sections of track. Regardless of the quality and integrity of the jointbar and bolt connections, the bending stiffness of jointbars is much lower than the rail generating large deflections under the train wheels. In addition, the gap or surface discontinuity on the rail running surface leads to generation of high wheel-rail impact force at the joint gap. These fundamental weaknesses have caused high rate of failure in track components at location of rail joints resulting in significant economic and safety issues in railways. The mechanical behavior of railway track at location of joints has not been fully understood due to various structural and material complexities. Although there have been some improvements in the methods for analysis of track at jointed rails in recent years, there are still uncertainties concerning the accuracy and reliability of the current methods. In this paper the current methods in analysis of track with a rail joint are critically evaluated and the new advances and recent research outcomes in this area are discussed. This research is part of a large granted project on rail joints which was defined by Cooperative Research Centre (CRC) for Rail Innovation with supports from Australian Rail Track Corporation (ARTC) and Queensland Rail (QR).

Keywords: jointed rails, railway mechanics, track dynamics, wheel-rail interaction

Procedia PDF Downloads 346
10545 A Study of General Attacks on Elliptic Curve Discrete Logarithm Problem over Prime Field and Binary Field

Authors: Tun Myat Aung, Ni Ni Hla

Abstract:

This paper begins by describing basic properties of finite field and elliptic curve cryptography over prime field and binary field. Then we discuss the discrete logarithm problem for elliptic curves and its properties. We study the general common attacks on elliptic curve discrete logarithm problem such as the Baby Step, Giant Step method, Pollard’s rho method and Pohlig-Hellman method, and describe in detail experiments of these attacks over prime field and binary field. The paper finishes by describing expected running time of the attacks and suggesting strong elliptic curves that are not susceptible to these attacks.c

Keywords: discrete logarithm problem, general attacks, elliptic curve, prime field, binary field

Procedia PDF Downloads 224
10544 Smart Water Main Inspection and Condition Assessment Using a Systematic Approach for Pipes Selection

Authors: Reza Moslemi, Sebastien Perrier

Abstract:

Water infrastructure deterioration can result in increased operational costs owing to increased repair needs and non-revenue water and consequently cause a reduced level of service and customer service satisfaction. Various water main condition assessment technologies have been introduced to the market in order to evaluate the level of pipe deterioration and to develop appropriate asset management and pipe renewal plans. One of the challenges for any condition assessment and inspection program is to determine the percentage of the water network and the combination of pipe segments to be inspected in order to obtain a meaningful representation of the status of the entire water network with a desirable level of accuracy. Traditionally, condition assessment has been conducted by selecting pipes based on age or location. However, this may not necessarily offer the best approach, and it is believed that by using a smart sampling methodology, a better and more reliable estimate of the condition of a water network can be achieved. This research investigates three different sampling methodologies, including random, stratified, and systematic. It is demonstrated that selecting pipes based on the proposed clustering and sampling scheme can considerably improve the ability of the inspected subset to represent the condition of a wider network. With a smart sampling methodology, a smaller data sample can provide the same insight as a larger sample. This methodology offers increased efficiency and cost savings for condition assessment processes and projects.

Keywords: condition assessment, pipe degradation, sampling, water main

Procedia PDF Downloads 146
10543 Lateral Capacity of Helical-Pile Groups Subjected to Bearing Combined Loads

Authors: Hesham Hamdy Abdelmohsen, Ahmed Shawky Abdul Azizb, Mona Fawzy Aldaghma

Abstract:

Helical piles have earned considerable attention as an effective deep foundation alternative due to their rapid installation process and their dual purpose in compression and tension. These piles find common uses as foundations for structures like solar panels, wind turbines, offshore platforms, and some kinds of retaining walls. These structures usually transfer different combinations of loads to their helical-pile foundations in the form of axial and lateral loads. Extensive research has been conducted to investigate and understand the behavior of these piles under the influence of either axial or lateral loads. However, the impacts of loading patterns that may act on the helical piles as combinations of axial compression and lateral loads still need more efforts of research work. This paper presents the results of an experimental (Lab tests) and numerical (PLAXIS-3D) study performed on vertical helical-pile groups under the action of combined loads as axial compression (bearing loads), acting successively with lateral (horizontal) loads. The study aims to clarify the effects of key factors, like helix location and direction of lateral load, on the lateral capacity of helical-pile groups and, consequently, on group efficiency. Besides the variation of helix location and lateral load direction, three patterns of successive bearing combined loads were considered, in which the axial vertical compression load was either zero, V1 or V2, whereas the lateral horizontal loads were varied under each vertical compression load. The study concluded that the lateral capacity of the helical-pile group is significantly affected by helix location within the length of the pile shaft. The optimal lateral performance is achieved with helices at a depth ratio of H/L = 0.4. Furthermore, groups of rectangular plan distribution exhibit greater lateral capacity if subjected to lateral horizontal load in the direction of its long axis. Additionally, the research emphasizes that the presence of vertical compression loading can enhance the lateral capacity of the group. This enhancement depends on the value of the vertical compression load, lateral load direction, and helix location, which highlights the complex interaction effect of these factors on the efficiency of helical-pile groups.

Keywords: helical piles, experimental, numerical, lateral loading, group efficiency

Procedia PDF Downloads 17
10542 Hybridized Simulated Annealing with Chemical Reaction Optimization for Solving to Sequence Alignment Problem

Authors: Ernesto Linan, Linda Cruz, Lucero Becerra

Abstract:

In this paper, a new hybridized algorithm based on Chemical Reaction Optimization and Simulated Annealing is proposed to solve the alignment sequence Problem. The Chemical Reaction Optimization is a population-based meta-heuristic algorithm based on the principles of a chemical reaction. Simulated Annealing is applied to solve a large number of combinatorial optimization problems of general-purpose. In this paper, we propose hybridization between Chemical Reaction Optimization algorithm and Simulated Annealing in order to solve the Sequence Alignment Problem. An initial population of molecules is defined at beginning of the proposed algorithm, where each molecule represents a sequence alignment problem. In order to simulate inter-molecule collisions, the process of Chemical Reaction is placed inside the Metropolis Cycle at certain values of temperature. Inside this cycle, change of molecules is done due to collisions; some molecules are accepted by applying Boltzmann probability. The results with the hybrid scheme are better than the results obtained separately.

Keywords: chemical reaction optimization, sequence alignment problem, simulated annealing algorithm, metaheuristics

Procedia PDF Downloads 208
10541 Location Management in Wireless Sensor Networks with Mobility

Authors: Amrita Anil Agashe, Sumant Tapas, Ajay Verma Yogesh Sonavane, Sourabh Yeravar

Abstract:

Due to advancement in MEMS technology today wireless sensors network has gained a lot of importance. The wide range of its applications includes environmental and habitat monitoring, object localization, target tracking, security surveillance etc. Wireless sensor networks consist of tiny sensor devices called as motes. The constrained computation power, battery power, storage capacity and communication bandwidth of the tiny motes pose challenging problems in the design and deployment of such systems. In this paper, we propose a ubiquitous framework for Real-Time Tracking, Sensing and Management System using IITH motes. Also, we explain the algorithm that we have developed for location management in wireless sensor networks with the aspect of mobility. Our developed framework and algorithm can be used to detect emergency events and safety threats and provides warning signals to handle the emergency.

Keywords: mobility management, motes, multihop, wireless sensor networks

Procedia PDF Downloads 411
10540 Occupational Health Hazards of Itinerant Waste Buyers (IWBs) in Kathmandu, Nepal

Authors: Ashish Khanal, Suja Giri

Abstract:

The scrap collection work is associated with multiple health hazards. Cut and scratches during collection and transportation of scraps are common. IWBs purchase the scraps mainly papers, cartoons, glass bottles and metals from the households. This study was conducted in Kathmandu, the capital city of Nepal. The location was chosen because Kathmandu is the biggest city of Nepal with highest number of IWBs. The research used a case study strategy to examine the occupational health hazards of IWBs. The only mode of collecting and transporting of scraps in Kathmandu is the bicycle. They have to do this regular work even during the scorching sun and chilled winter. The musculoskeletal and gastrointestinal disorders are the common health problem shared by IWBs in Kathmandu, Nepal. Despite of these problems, IWBs don’t take it seriously and rarely goes for the health check-up. There is need of personal protective equipment and guidance for safety of IWBs. IWBs need to wear closed shoes and use gloves to avoid cuts during the collection and transportation of the recyclables.

Keywords: itinerant waste buyers, Kathmandu, occupational health, scrap

Procedia PDF Downloads 167
10539 Voltage and Frequency Regulation Using the Third-Party Mid-Size Battery

Authors: Roghieh A. Biroon, Zoleikha Abdollahi

Abstract:

The recent growth of renewables, e.g., solar panels, batteries, and electric vehicles (EVs) in residential and small commercial sectors, has potential impacts on the stability and operation of power grids. Considering approximately 50 percent share of the residential and the commercial sectors in the electricity demand market, the significance of these impacts, and the necessity of addressing them are more highlighted. Utilities and power system operators should manage the renewable electricity sources integration with power systems in such a way to extract the most possible advantages for the power systems. The most common effect of high penetration level of the renewables is the reverse power flow in the distribution feeders when the customers generate more power than their needs. The reverse power flow causes voltage rise and thermal issues in the power grids. To overcome the voltage rise issues in the distribution system, several techniques have been proposed including reducing transformers short circuit resistance and feeder impedance, installing autotransformers/voltage regulators along the line, absorbing the reactive power by distributed generators (DGs), and limiting the PV and battery sizes. In this study, we consider a medium-scale battery energy storage to manage the power energy and address the aforementioned issues on voltage deviation and power loss increase. We propose an optimization algorithm to find the optimum size and location for the battery. The optimization for the battery location and size is so that the battery maintains the feeder voltage deviation and power loss at a certain desired level. Moreover, the proposed optimization algorithm controls the charging/discharging profile of the battery to absorb the negative power flow from residential and commercial customers in the feeder during the peak time and sell the power back to the system during the off-peak time. The proposed battery regulates the voltage problem in the distribution system while it also can play frequency regulation role in islanded microgrids. This battery can be regulated and controlled by the utilities or a third-party ancillary service provider for the utilities to reduce the power system loss and regulate the distribution feeder voltage and frequency in standard level.

Keywords: ancillary services, battery, distribution system and optimization

Procedia PDF Downloads 126
10538 Water Leakage Detection System of Pipe Line using Radial Basis Function Neural Network

Authors: A. Ejah Umraeni Salam, M. Tola, M. Selintung, F. Maricar

Abstract:

Clean water is an essential and fundamental human need. Therefore, its supply must be assured by maintaining the quality, quantity and water pressure. However the fact is, on its distribution system, leakage happens and becomes a common world issue. One of the technical causes of the leakage is a leaking pipe. The purpose of the research is how to use the Radial Basis Function Neural (RBFNN) model to detect the location and the magnitude of the pipeline leakage rapidly and efficiently. In this study the RBFNN are trained and tested on data from EPANET hydraulic modeling system. Method of Radial Basis Function Neural Network is proved capable to detect location and magnitude of pipeline leakage with of the accuracy of the prediction results based on the value of RMSE (Root Meant Square Error), comparison prediction and actual measurement approaches 0.000049 for the whole pipeline system.

Keywords: radial basis function neural network, leakage pipeline, EPANET, RMSE

Procedia PDF Downloads 352
10537 Maternal Review: Challenges Experienced by Midwives in Malawi

Authors: Mercy D. Chirwa, Juliet Nyasulu, Lebisti Modiba, Makombo Ganga-Limando

Abstract:

Maternal death review is an initiative that provides a deeper understanding of the causes and circumstances sorounding of maternal deaths in Malawi and globally. Midwives are frontline members of the healthcare team and have stories about what pregnant women go through as such they are better placed to contribute to these reviews. Despite midwives’ participation as members of the facility-based maternal death review team, maternal deaths continues to occur. A lot has been documented around processes involved in maternal review, however, not much has been written around challenges experienced by midwives in maternal death review. This study explored the challenges faced by midwives in the implementation of maternal death reviews in the context of the healthcare system in Malawi. Methodology: This was a qualitative exploratory study design. Focus group discussions and individual face-to-face interviews were used to collect data in the study. A total of 40 midwives, who met the inclusion criteria, participated in the study. Data was analysed manually using a thematic content procedure. Findings: The four major challenges identified were: knowledge and skill gaps; lack of leadership and accountability; lack of institutional political will and inconsistency in conducting FBMDR, impeding midwives’ effective contribution to the implementation of maternal death review. The practical solutions and recommendations that emerged were: need-based knowledge and skills updates, supportive leadership, effective and efficient interdisciplinary work ethics, and sustained availability of material and human resources. Conclusion: Midwives have the highest potential to contribute to the reduction of maternal deaths. Practice development strategies are required to improve their practice in all the areas they are challenged with.

Keywords: facility-based maternal death review, maternal deaths, midwife, midwife challenges

Procedia PDF Downloads 62