Search results for: discontinuous cost function
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10403

Search results for: discontinuous cost function

10403 Mixed Integer Programing for Multi-Tier Rebate with Discontinuous Cost Function

Authors: Y. Long, L. Liu, K. V. Branin

Abstract:

One challenge faced by procurement decision-maker during the acquisition process is how to compare similar products from different suppliers and allocate orders among different products or services. This work focuses on allocating orders among multiple suppliers considering rebate. The objective function is to minimize the total acquisition cost including purchasing cost and rebate benefit. Rebate benefit is complex and difficult to estimate at the ordering step. Rebate rules vary for different suppliers and usually change over time. In this work, we developed a system to collect the rebate policies, standardized the rebate policies and developed two-stage optimization models for ordering allocation. Rebate policy with multi-tiers is considered in modeling. The discontinuous cost function of rebate benefit is formulated for different scenarios. A piecewise linear function is used to approximate the discontinuous cost function of rebate benefit. And a Mixed Integer Programing (MIP) model is built for order allocation problem with multi-tier rebate. A case study is presented and it shows that our optimization model can reduce the total acquisition cost by considering rebate rules.

Keywords: discontinuous cost function, mixed integer programming, optimization, procurement, rebate

Procedia PDF Downloads 237
10402 Elastohydrodynamic Lubrication Study Using Discontinuous Finite Volume Method

Authors: Prawal Sinha, Peeyush Singh, Pravir Dutt

Abstract:

Problems in elastohydrodynamic lubrication have attracted a lot of attention in the last few decades. Solving a two-dimensional problem has always been a big challenge. In this paper, a new discontinuous finite volume method (DVM) for two-dimensional point contact Elastohydrodynamic Lubrication (EHL) problem has been developed and analyzed. A complete algorithm has been presented for solving such a problem. The method presented is robust and easily parallelized in MPI architecture. GMRES technique is implemented to solve the matrix obtained after the formulation. A new approach is followed in which discontinuous piecewise polynomials are used for the trail functions. It is natural to assume that the advantages of using discontinuous functions in finite element methods should also apply to finite volume methods. The nature of the discontinuity of the trail function is such that the elements in the corresponding dual partition have the smallest support as compared with the Classical finite volume methods. Film thickness calculation is done using singular quadrature approach. Results obtained have been presented graphically and discussed. This method is well suited for solving EHL point contact problem and can probably be used as commercial software.

Keywords: elastohydrodynamic, lubrication, discontinuous finite volume method, GMRES technique

Procedia PDF Downloads 236
10401 Discontinuous Galerkin Method for Higher-Order Ordinary Differential Equations

Authors: Helmi Temimi

Abstract:

In this paper, we study the super-convergence properties of the discontinuous Galerkin (DG) method applied to one-dimensional mth-order ordinary differential equations without introducing auxiliary variables. We found that nth−derivative of the DG solution exhibits an optimal O (hp+1−n) convergence rates in the L2-norm when p-degree piecewise polynomials with p≥1 are used. We further found that the odd-derivatives and the even derivatives are super convergent, respectively, at the upwind and downwind endpoints.

Keywords: discontinuous, galerkin, superconvergence, higherorder, error, estimates

Procedia PDF Downloads 453
10400 Estimation of Fuel Cost Function Characteristics Using Cuckoo Search

Authors: M. R. Al-Rashidi, K. M. El-Naggar, M. F. Al-Hajri

Abstract:

The fuel cost function describes the electric power generation-cost relationship in thermal plants, hence, it sheds light on economical aspects of power industry. Different models have been proposed to describe this relationship with the quadratic function model being the most popular one. Parameters of second order fuel cost function are estimated in this paper using cuckoo search algorithm. It is a new population based meta-heuristic optimization technique that has been used in this study primarily as an accurate estimation tool. Its main features are flexibility, simplicity, and effectiveness when compared to other estimation techniques. The parameter estimation problem is formulated as an optimization one with the goal being minimizing the error associated with the estimated parameters. A case study is considered in this paper to illustrate cuckoo search promising potential as a valuable estimation and optimization technique.

Keywords: cuckoo search, parameters estimation, fuel cost function, economic dispatch

Procedia PDF Downloads 553
10399 Network Functions Virtualization-Based Virtual Routing Function Deployment under Network Delay Constraints

Authors: Kenichiro Hida, Shin-Ichi Kuribayashi

Abstract:

NFV-based network implements a variety of network functions with software on general-purpose servers, and this allows the network operator to select any capabilities and locations of network functions without any physical constraints. In this paper, we evaluate the influence of the maximum tolerable network delay on the virtual routing function deployment guidelines which the authors proposed previously. Our evaluation results have revealed the following: (1) the more the maximum tolerable network delay condition becomes severe, the more the number of areas where the route selection function is installed increases and the total network cost increases, (2) the higher the routing function cost relative to the circuit bandwidth cost, the increase ratio of total network cost becomes larger according to the maximum tolerable network delay condition.

Keywords: NFV (Network Functions Virtualization), resource allocation, virtual routing function, minimum total network cost

Procedia PDF Downloads 221
10398 Unmanned Aerial Vehicle Landing Based on Ultra-Wideband Localization System and Optimal Strategy for Searching Optimal Landing Point

Authors: Meng Wu

Abstract:

Unmanned aerial vehicle (UAV) landing technology is a common task that is required to be fulfilled by fly robots. In this paper, the crazyflie2.0 is located by ultra-wideband (UWB) localization system that contains 4 UWB anchors. Another UWB anchor is introduced and installed on a stationary platform. One cost function is designed to find the minimum distance between crazyflie2.0 and the anchor installed on the stationary platform. The coordinates of the anchor are unknown in advance, and the goal of the cost function is to define the location of the anchor, which can be considered as an optimal landing point. When the cost function reaches the minimum value, the corresponding coordinates of the UWB anchor fixed on the stationary platform can be calculated and defined as the landing point. The simulation shows the effectiveness of the method in this paper.

Keywords: UAV landing, UWB localization system, UWB anchor, cost function, stationary platform

Procedia PDF Downloads 48
10397 Value Engineering and Its Impact on Drainage Design Optimization for Penang International Airport Expansion

Authors: R.M. Asyraf, A. Norazah, S.M. Khairuddin, B. Noraziah

Abstract:

Designing a system at present requires a vital, challenging task; to ensure the design philosophy is maintained in economical ways. This paper perceived the value engineering (VE) approach applied in infrastructure works, namely stormwater drainage. This method is adopted in line as consultants have completed the detailed design. Function Analysis System Technique (FAST) diagram and VE job plan, information, function analysis, creative judgement, development, and recommendation phase are used to scrutinize the initial design of stormwater drainage. An estimated cost reduction using the VE approach of 2% over the initial proposal was obtained. This cost reduction is obtained from the design optimization of the drainage foundation and structural system, where the pile design and drainage base structure are optimized. Likewise, the design of the on-site detention tank (OSD) pump was revised and contribute to the cost reduction obtained. This case study shows that the VE approach can be an important tool in optimizing the design to reduce costs.

Keywords: value engineering, function analysis system technique, stormwater drainage, cost reduction

Procedia PDF Downloads 124
10396 Evidence on Scale Economies in National Bank of Pakistan

Authors: Sohail Zafar, Sardar Javaid Iqbal Khan

Abstract:

We use a parametric approach within a translog cost function framework to estimate the economies of scale in National Bank of Pakistan from 1997 to 2013. The results indicate significant economies of scale throughout the sample at aggregates and disaggregates taking in account size subject to stipulation ownership. The factor markets often produce scale inefficiencies in the banking of developing countries like Pakistan such inefficiencies are common due to distortion in factor markets leading to the use of inappropriate factor proportions. The findings suggest that National Bank of Pakistan diversify their asset portfolios that it has cost advantage, therefore, expansion in size should be encouraged under current technology because it appears to be cost effective. In addition, our findings support the implementation of universal banking model in Pakistan.

Keywords: scale economies, cost function, disaggregates, aggregates

Procedia PDF Downloads 303
10395 A Versatile Data Processing Package for Ground-Based Synthetic Aperture Radar Deformation Monitoring

Authors: Zheng Wang, Zhenhong Li, Jon Mills

Abstract:

Ground-based synthetic aperture radar (GBSAR) represents a powerful remote sensing tool for deformation monitoring towards various geohazards, e.g. landslides, mudflows, avalanches, infrastructure failures, and the subsidence of residential areas. Unlike spaceborne SAR with a fixed revisit period, GBSAR data can be acquired with an adjustable temporal resolution through either continuous or discontinuous operation. However, challenges arise from processing high temporal-resolution continuous GBSAR data, including the extreme cost of computational random-access-memory (RAM), the delay of displacement maps, and the loss of temporal evolution. Moreover, repositioning errors between discontinuous campaigns impede the accurate measurement of surface displacements. Therefore, a versatile package with two complete chains is developed in this study in order to process both continuous and discontinuous GBSAR data and address the aforementioned issues. The first chain is based on a small-baseline subset concept and it processes continuous GBSAR images unit by unit. Images within a window form a basic unit. By taking this strategy, the RAM requirement is reduced to only one unit of images and the chain can theoretically process an infinite number of images. The evolution of surface displacements can be detected as it keeps temporarily-coherent pixels which are present only in some certain units but not in the whole observation period. The chain supports real-time processing of the continuous data and the delay of creating displacement maps can be shortened without waiting for the entire dataset. The other chain aims to measure deformation between discontinuous campaigns. Temporal averaging is carried out on a stack of images in a single campaign in order to improve the signal-to-noise ratio of discontinuous data and minimise the loss of coherence. The temporal-averaged images are then processed by a particular interferometry procedure integrated with advanced interferometric SAR algorithms such as robust coherence estimation, non-local filtering, and selection of partially-coherent pixels. Experiments are conducted using both synthetic and real-world GBSAR data. Displacement time series at the level of a few sub-millimetres are achieved in several applications (e.g. a coastal cliff, a sand dune, a bridge, and a residential area), indicating the feasibility of the developed GBSAR data processing package for deformation monitoring of a wide range of scientific and practical applications.

Keywords: ground-based synthetic aperture radar, interferometry, small baseline subset algorithm, deformation monitoring

Procedia PDF Downloads 138
10394 Cost Effectiveness and Performance Study of Perpetual Pavement Using ABAQUS

Authors: Mansour Fakhri, Monire Zokaei

Abstract:

Where there are many demolitions on conventional asphalt pavements, heavy costs are paid to repair and reconstruct the pavement roads annually. Recently some research has been done in order to increase the pavement life. Perpetual pavement is regarded as one of them which can improve the pavement life and minimize the maintenance activity and cost. In this research, ABAQUS which is a finite element software is implemented for analyzing and simulation of perpetual pavement. Viscoelastic model of material is used and loading wheel is considered to be dynamic. Effect of different parameters on pavement function has been considered. Because of high primary cost these pavements are not widely used. In this regard, life cost analysis was also carried out to compare perpetual pavement to conventional asphalt concrete pavement. It was concluded that although the initial cost of perpetual pavement is higher than that of conventional asphalt pavement, life cycle cost analysis during 50 years of service life showed that the performance of this pavement is better and the whole life cost of that is less.

Keywords: ABAQUS, lifecycle cost analysis, mechanistic empirical, perpetual pavement

Procedia PDF Downloads 361
10393 Production Line Layout Planning Based on Complexity Measurement

Authors: Guoliang Fan, Aiping Li, Nan Xie, Liyun Xu, Xuemei Liu

Abstract:

Mass customization production increases the difficulty of the production line layout planning. The material distribution process for variety of parts is very complex, which greatly increases the cost of material handling and logistics. In response to this problem, this paper presents an approach of production line layout planning based on complexity measurement. Firstly, by analyzing the influencing factors of equipment layout, the complexity model of production line is established by using information entropy theory. Then, the cost of the part logistics is derived considering different variety of parts. Furthermore, the function of optimization including two objectives of the lowest cost, and the least configuration complexity is built. Finally, the validity of the function is verified in a case study. The results show that the proposed approach may find the layout scheme with the lowest logistics cost and the least complexity. Optimized production line layout planning can effectively improve production efficiency and equipment utilization with lowest cost and complexity.

Keywords: production line, layout planning, complexity measurement, optimization, mass customization

Procedia PDF Downloads 363
10392 An Experimental Study on the Positive Streamer Leader Propagation under Slow Front Impulse Voltages in a 10m Rod-Plane Air Gap

Authors: Wahab Ali Shah, Junjia He

Abstract:

In this work, we performed a large-scale investigation into leader development in a 10 m rod-plane gap under a long front positive impulse. To describe the leader propagation under slow front impulse voltages, we recorded the leader propagation with a high-speed charge coupled device (CCD) camera. It is important to figure out this phenomenon to deepen our understanding of leader discharge. The observation results showed that the leader mechanism is a very complex physical phenomenon; it could be categorized into two types of leader process, namely, continuous and the discontinuous leader streamer-leader propagation. Furthermore, we studied the continuous leader development parameters, including two-dimensional (2-D) leader length, injected charge, and final jump stage, as well as leader velocity for rod–plane configuration. We observed that the discontinuous leader makes an important contribution to the appearance of channel re-illuminations of the positive leader. The comparative study shows better results in terms of standard switch impulse and long front positive impulse. Finally, the results are presented with a view toward improving our understanding of propagation mechanisms related to restrike phenomena, which are rarely reported. To clarify the above doubts under long front cases, we carried out extensive experiments in this study.

Keywords: continuous and discontinuous leader, high-speed photographs, long air gap, positive long front impulse, restrike phenomena

Procedia PDF Downloads 148
10391 Online Robust Model Predictive Control for Linear Fractional Transformation Systems Using Linear Matrix Inequalities

Authors: Peyman Sindareh Esfahani, Jeffery Kurt Pieper

Abstract:

In this paper, the problem of robust model predictive control (MPC) for discrete-time linear systems in linear fractional transformation form with structured uncertainty and norm-bounded disturbance is investigated. The problem of minimization of the cost function for MPC design is converted to minimization of the worst case of the cost function. Then, this problem is reduced to minimization of an upper bound of the cost function subject to a terminal inequality satisfying the l2-norm of the closed loop system. The characteristic of the linear fractional transformation system is taken into account, and by using some mathematical tools, the robust predictive controller design problem is turned into a linear matrix inequality minimization problem. Afterwards, a formulation which includes an integrator to improve the performance of the proposed robust model predictive controller in steady state condition is studied. The validity of the approaches is illustrated through a robust control benchmark problem.

Keywords: linear fractional transformation, linear matrix inequality, robust model predictive control, state feedback control

Procedia PDF Downloads 373
10390 A Study of the Effects of Temperatures and Optimum pH on the Specific Methane Production of Perennial Ryegrass during Anaerobic Digestion Process under a Discontinuous Daily Feeding Condition

Authors: Uchenna Egwu, Paul Jonathan Sallis

Abstract:

Perennial ryegrass is an abundant renewable lignocellulosic biofuel feedstock for biomethane production through anaerobic digestion (AD). In this study, six anaerobic continuously stirred tank reactors (CSTRs) were set up in three pairs. Each pair of the CSTRs was then used to study the effects of operating temperatures – psychrophilic, mesophilic, and thermophilic, and optimum pH on the specific methane production (SMP) of the ryegrass during AD under discontinuous daily feeding conditions. The reactors were fed at an organic loading rate (OLR) ranging from 1-1.5 kgVS.L⁻¹d⁻¹ and hydraulic residence time, HRT=20 days for 140 days. The pH of the digesters was maintained at the range of 6.8-7.2 using 1 M NH₄HCO₃ solution, but this was replaced with biomass ash-extracts from day 105-140. The results obtained showed that the mean SMP of ryegrass measured between HRT 3 and 4 were 318.4, 425.4 and 335 N L CH₄ kg⁻¹VS.d⁻¹ for the psychrophilic (25 ± 2°C), mesophilic (40 ± 1°C) and thermophilic (60 ± 1°C) temperatures respectively. It was also observed that the buffering ability of the reactors increased with operating temperature, probably due to an increase in the solubility of ammonium bicarbonate (NH₄HCO₃) with temperature. The reactors also achieved a mean VS destruction of 61.9, 68.5 and 63.5%, respectively, which signifies that the mesophilic reactors achieved the highest specific methane production (SMP), while the psychrophilic reactors achieved the lowest. None of the reactors attained steady-state condition due to the discontinuous daily feeding times, and therefore, such feeding practice may not be the most effective for maximum biogas production over long periods of time. The addition of NH₄HCO₃ as supplement provided a good buffering condition in these AD digesters, but the digesters failed in the long run due to inhibition from the accumulation of free ammonia, which later led to decrease in pH, acidification, and souring of the digesters. However, the addition of biomass ash extracts was shown to potentially revive failed AD reactors by providing an adequate buffering and essential trace nutrient supplements necessary for optimal bacterial growth.

Keywords: anaerobic digestion, discontinuous feeding, perennial ryegrass, specific methane production, supplements, temperature

Procedia PDF Downloads 112
10389 Inventory Policy Above Country Level for Cooperating Countries for Vaccines

Authors: Aysun Pınarbaşı, Béla Vizvári

Abstract:

The countries are the units that procure the vaccines during the COVID-19 pandemic. The delivered quantities are huge. The countries must bear the inventory holding cost according to the variation of stock quantities. This cost depends on the speed of the vaccination in the country. This speed is time-dependent. The vaccinated portion of the population can be approximated by the cumulative distribution function of the Cauchy distribution. A model is provided for determining the minimal-cost inventory policy, and its optimality conditions are provided. The model is solved for 20 countries for different numbers of procurements. The results reveal the individual behavior of each country. We provide an inventory policy for the pandemic period for the countries. This paper presents a deterministic model for vaccines with a demand rate variable over time for the countries. It is aimed to provide an analytical model to deal with the minimization of holding cost and develop inventory policies regarding this aim to be used for a variety of perishable products such as vaccines. The saturation process is introduced, and an approximation of the vaccination curve of the countries has been discussed. According to this aspect, a deterministic model for inventory policy has been developed.

Keywords: covid-19, vaccination, inventory policy, bounded total demand, inventory holding cost, cauchy distribution, sigmoid function

Procedia PDF Downloads 55
10388 Infrastructure Sharing Synergies: Optimal Capacity Oversizing and Pricing

Authors: Robin Molinier

Abstract:

Industrial symbiosis (I.S) deals with both substitution synergies (exchange of waste materials, fatal energy and utilities as resources for production) and infrastructure/service sharing synergies. The latter is based on the intensification of use of an asset and thus requires to balance capital costs increments with snowball effects (network externalities) for its implementation. Initial investors must specify ex-ante arrangements (cost sharing and pricing schedule) to commit toward investments in capacities and transactions. Our model investigate the decision of 2 actors trying to choose cooperatively a level of infrastructure capacity oversizing to set a plug-and-play offer to a potential entrant whose capacity requirement is randomly distributed while satisficing their own requirements. Capacity cost exhibits sub-additive property so that there is room for profitable overcapacity setting in the first period. The entrant’s willingness-to-pay for the access to the infrastructure is dependent upon its standalone cost and the capacity gap that it must complete in case the available capacity is insufficient ex-post (the complement cost). Since initial capacity choices are driven by ex-ante (expected) yield extractible from the entrant we derive the expected complement cost function which helps us defining the investors’ objective function. We first show that this curve is decreasing and convex in the capacity increments and that it is shaped by the distribution function of the potential entrant’s requirements. We then derive the general form of solutions and solve the model for uniform and triangular distributions. Depending on requirements volumes and cost assumptions different equilibria occurs. We finally analyze the effect of a per-unit subsidy a public actor would apply to foster such sharing synergies.

Keywords: capacity, cooperation, industrial symbiosis, pricing

Procedia PDF Downloads 186
10387 Stochastic Frontier Application for Evaluating Cost Inefficiencies in Organic Saffron

Authors: Pawan Kumar Sharma, Sudhakar Dwivedi, R. K. Arora

Abstract:

Saffron is one of the most precious spices grown on the earth and is cultivated in a very limited area in few countries of the world. It has also been grown as a niche crop in Kishtwar district of Jammu region of Jammu and Kashmir State of India. This paper attempts to examine the presence of cost inefficiencies in saffron production and the associated socio-economic characteristics of saffron growers in the mentioned area. Although the numbers of inputs used in cultivation of saffron were limited, still cost inefficiencies were present in its production. The net present value (NPV), internal rate of return (IRR) and profitability index (PI) of investment in five years of saffron production were INR 1120803, 95.67 % and 3.52 respectively. The estimated coefficients of saffron stochastic cost function for saffron bulbs, human labour, animal labour, manure and saffron output were positive. The saffron growers having non-farm income were more cost inefficient as compared to farmers who did not have sources of income other than farming by 0.04 %. The maximum value of cost efficiency for saffron grower was 1.69 with mean value of 1.12. The majority of farmers have low cost inefficiencies, as the highest frequency of occurrence of the predicted cost efficiency was below 1.06.

Keywords: saffron, internal rate of return, cost efficiency, stochastic frontier model

Procedia PDF Downloads 128
10386 Licensing in a Hotelling Model with Quadratic Transportation Costs

Authors: Fehmi Bouguezzi

Abstract:

This paper studies optimal licensing regimes in a linear Hotelling model where firms are located at the end points of the city and where the transportation cost is not linear but quadratic. We study for that a more general cost function and we try to compare the findings with the results of the linear cost. We find the same optimal licensing regimes. A per unit royalty is optimal when innovation is not drastic and no licensing is better when innovation is drastic. We also find that no licensing is always better than fixed fee licensing.

Keywords: Hotelling model, technology transfer, patent licensing, quadratic transportation cost

Procedia PDF Downloads 327
10385 A Runge Kutta Discontinuous Galerkin Method for Lagrangian Compressible Euler Equations in Two-Dimensions

Authors: Xijun Yu, Zhenzhen Li, Zupeng Jia

Abstract:

This paper presents a new cell-centered Lagrangian scheme for two-dimensional compressible flow. The new scheme uses a semi-Lagrangian form of the Euler equations. The system of equations is discretized by Discontinuous Galerkin (DG) method using the Taylor basis in Eulerian space. The vertex velocities and the numerical fluxes through the cell interfaces are computed consistently by a nodal solver. The mesh moves with the fluid flow. The time marching is implemented by a class of the Runge-Kutta (RK) methods. A WENO reconstruction is used as a limiter for the RKDG method. The scheme is conservative for the mass, momentum and total energy. The scheme maintains second-order accuracy and has free parameters. Results of some numerical tests are presented to demonstrate the accuracy and the robustness of the scheme.

Keywords: cell-centered Lagrangian scheme, compressible Euler equations, RKDG method

Procedia PDF Downloads 523
10384 Reliability-Based Life-Cycle Cost Model for Engineering Systems

Authors: Reza Lotfalian, Sudarshan Martins, Peter Radziszewski

Abstract:

The effect of reliability on life-cycle cost, including initial and maintenance cost of a system is studied. The failure probability of a component is used to calculate the average maintenance cost during the operation cycle of the component. The standard deviation of the life-cycle cost is also calculated as an error measure for the average life-cycle cost. As a numerical example, the model is used to study the average life cycle cost of an electric motor.

Keywords: initial cost, life-cycle cost, maintenance cost, reliability

Procedia PDF Downloads 572
10383 Coupling Random Demand and Route Selection in the Transportation Network Design Problem

Authors: Shabnam Najafi, Metin Turkay

Abstract:

Network design problem (NDP) is used to determine the set of optimal values for certain pre-specified decision variables such as capacity expansion of nodes and links by optimizing various system performance measures including safety, congestion, and accessibility. The designed transportation network should improve objective functions defined for the system by considering the route choice behaviors of network users at the same time. The NDP studies mostly investigated the random demand and route selection constraints separately due to computational challenges. In this work, we consider both random demand and route selection constraints simultaneously. This work presents a nonlinear stochastic model for land use and road network design problem to address the development of different functional zones in urban areas by considering both cost function and air pollution. This model minimizes cost function and air pollution simultaneously with random demand and stochastic route selection constraint that aims to optimize network performance via road capacity expansion. The Bureau of Public Roads (BPR) link impedance function is used to determine the travel time function in each link. We consider a city with origin and destination nodes which can be residential or employment or both. There are set of existing paths between origin-destination (O-D) pairs. Case of increasing employed population is analyzed to determine amount of roads and origin zones simultaneously. Minimizing travel and expansion cost of routes and origin zones in one side and minimizing CO emission in the other side is considered in this analysis at the same time. In this work demand between O-D pairs is random and also the network flow pattern is subject to stochastic user equilibrium, specifically logit route choice model. Considering both demand and route choice, random is more applicable to design urban network programs. Epsilon-constraint is one of the methods to solve both linear and nonlinear multi-objective problems. In this work epsilon-constraint method is used to solve the problem. The problem was solved by keeping first objective (cost function) as the objective function of the problem and second objective as a constraint that should be less than an epsilon, where epsilon is an upper bound of the emission function. The value of epsilon should change from the worst to the best value of the emission function to generate the family of solutions representing Pareto set. A numerical example with 2 origin zones and 2 destination zones and 7 links is solved by GAMS and the set of Pareto points is obtained. There are 15 efficient solutions. According to these solutions as cost function value increases, emission function value decreases and vice versa.

Keywords: epsilon-constraint, multi-objective, network design, stochastic

Procedia PDF Downloads 620
10382 The Implementation of Secton Method for Finding the Root of Interpolation Function

Authors: Nur Rokhman

Abstract:

A mathematical function gives relationship between the variables composing the function. Interpolation can be viewed as a process of finding mathematical function which goes through some specified points. There are many interpolation methods, namely: Lagrange method, Newton method, Spline method etc. For some specific condition, such as, big amount of interpolation points, the interpolation function can not be written explicitly. This such function consist of computational steps. The solution of equations involving the interpolation function is a problem of solution of non linear equation. Newton method will not work on the interpolation function, for the derivative of the interpolation function cannot be written explicitly. This paper shows the use of Secton method to determine the numerical solution of the function involving the interpolation function. The experiment shows the fact that Secton method works better than Newton method in finding the root of Lagrange interpolation function.

Keywords: Secton method, interpolation, non linear function, numerical solution

Procedia PDF Downloads 357
10381 Flood Modeling in Urban Area Using a Well-Balanced Discontinuous Galerkin Scheme on Unstructured Triangular Grids

Authors: Rabih Ghostine, Craig Kapfer, Viswanathan Kannan, Ibrahim Hoteit

Abstract:

Urban flooding resulting from a sudden release of water due to dam-break or excessive rainfall is a serious threatening environment hazard, which causes loss of human life and large economic losses. Anticipating floods before they occur could minimize human and economic losses through the implementation of appropriate protection, provision, and rescue plans. This work reports on the numerical modelling of flash flood propagation in urban areas after an excessive rainfall event or dam-break. A two-dimensional (2D) depth-averaged shallow water model is used with a refined unstructured grid of triangles for representing the urban area topography. The 2D shallow water equations are solved using a second-order well-balanced discontinuous Galerkin scheme. Theoretical test case and three flood events are described to demonstrate the potential benefits of the scheme: (i) wetting and drying in a parabolic basin (ii) flash flood over a physical model of the urbanized Toce River valley in Italy; (iii) wave propagation on the Reyran river valley in consequence of the Malpasset dam-break in 1959 (France); and (iv) dam-break flood in October 1982 at the town of Sumacarcel (Spain). The capability of the scheme is also verified against alternative models. Computational results compare well with recorded data and show that the scheme is at least as efficient as comparable second-order finite volume schemes, with notable efficiency speedup due to parallelization.

Keywords: dam-break, discontinuous Galerkin scheme, flood modeling, shallow water equations

Procedia PDF Downloads 156
10380 Discontinuous Spacetime with Vacuum Holes as Explanation for Gravitation, Quantum Mechanics and Teleportation

Authors: Constantin Z. Leshan

Abstract:

Hole Vacuum theory is based on discontinuous spacetime that contains vacuum holes. Vacuum holes can explain gravitation, some laws of quantum mechanics and allow teleportation of matter. All massive bodies emit a flux of holes which curve the spacetime; if we increase the concentration of holes, it leads to length contraction and time dilation because the holes do not have the properties of extension and duration. In the limited case when space consists of holes only, the distance between every two points is equal to zero and time stops - outside of the Universe, the extension and duration properties do not exist. For this reason, the vacuum hole is the only particle in physics capable of describing gravitation using its own properties only. All microscopic particles must 'jump' continually and 'vibrate' due to the appearance of holes (impassable microscopic 'walls' in space), and it is the cause of the quantum behavior. Vacuum holes can explain the entanglement, non-locality, wave properties of matter, tunneling, uncertainty principle and so on. Particles do not have trajectories because spacetime is discontinuous and has impassable microscopic 'walls' due to the simple mechanical motion is impossible at small scale distances; it is impossible to 'trace' a straight line in the discontinuous spacetime because it contains the impassable holes. Spacetime 'boils' continually due to the appearance of the vacuum holes. For teleportation to be possible, we must send a body outside of the Universe by enveloping it with a closed surface consisting of vacuum holes. Since a material body cannot exist outside of the Universe, it reappears instantaneously in a random point of the Universe. Since a body disappears in one volume and reappears in another random volume without traversing the physical space between them, such a transportation method can be called teleportation (or Hole Teleportation). It is shown that Hole Teleportation does not violate causality and special relativity due to its random nature and other properties. Although Hole Teleportation has a random nature, it can be used for colonization of extrasolar planets by the help of the method called 'random jumps': after a large number of random teleportation jumps, there is a probability that the spaceship may appear near a habitable planet. We can create vacuum holes experimentally using the method proposed by Descartes: we must remove a body from the vessel without permitting another body to occupy this volume.

Keywords: border of the Universe, causality violation, perfect isolation, quantum jumps

Procedia PDF Downloads 399
10379 The Control of Wall Thickness Tolerance during Pipe Purchase Stage Based on Reliability Approach

Authors: Weichao Yu, Kai Wen, Weihe Huang, Yang Yang, Jing Gong

Abstract:

Metal-loss corrosion is a major threat to the safety and integrity of gas pipelines as it may result in the burst failures which can cause severe consequences that may include enormous economic losses as well as the personnel casualties. Therefore, it is important to ensure the corroding pipeline integrity and efficiency, considering the value of wall thickness, which plays an important role in the failure probability of corroding pipeline. Actually, the wall thickness is controlled during pipe purchase stage. For example, the API_SPEC_5L standard regulates the allowable tolerance of the wall thickness from the specified value during the pipe purchase. The allowable wall thickness tolerance will be used to determine the wall thickness distribution characteristic such as the mean value, standard deviation and distribution. Taking the uncertainties of the input variables in the burst limit-state function into account, the reliability approach rather than the deterministic approach will be used to evaluate the failure probability. Moreover, the cost of pipe purchase will be influenced by the allowable wall thickness tolerance. More strict control of the wall thickness usually corresponds to a higher pipe purchase cost. Therefore changing the wall thickness tolerance will vary both the probability of a burst failure and the cost of the pipe. This paper describes an approach to optimize the wall thickness tolerance considering both the safety and economy of corroding pipelines. In this paper, the corrosion burst limit-state function in Annex O of CSAZ662-7 is employed to evaluate the failure probability using the Monte Carlo simulation technique. By changing the allowable wall thickness tolerance, the parameters of the wall thickness distribution in the limit-state function will be changed. Using the reliability approach, the corresponding variations in the burst failure probability will be shown. On the other hand, changing the wall thickness tolerance will lead to a change in cost in pipe purchase. Using the variation of the failure probability and pipe cost caused by changing wall thickness tolerance specification, the optimal allowable tolerance can be obtained, and used to define pipe purchase specifications.

Keywords: allowable tolerance, corroding pipeline segment, operation cost, production cost, reliability approach

Procedia PDF Downloads 372
10378 Challenges and Opportunities in Computing Logistics Cost in E-Commerce Supply Chain

Authors: Pramod Ghadge, Swadesh Srivastava

Abstract:

Revenue generation of a logistics company depends on how the logistics cost of a shipment is calculated. Logistics cost of a shipment is a function of distance & speed of the shipment travel in a particular network, its volumetric size and dead weight. Logistics billing is based mainly on the consumption of the scarce resource (space or weight carrying capacity of a carrier). Shipment’s size or deadweight is a function of product and packaging weight, dimensions and flexibility. Hence, to arrive at a standard methodology to compute accurate cost to bill the customer, the interplay among above mentioned physical attributes along with their measurement plays a key role. This becomes even more complex for an ecommerce company, like Flipkart, which caters to shipments from both warehouse and marketplace in an unorganized non-standard market like India. In this paper, we will explore various methodologies to define a standard way of billing the non-standard shipments across a wide range of size, shape and deadweight. Those will be, usage of historical volumetric/dead weight data to arrive at a factor which can be used to compute the logistics cost of a shipment, also calculating the real/contour volume of a shipment to address the problem of irregular shipment shapes which cannot be solved by conventional bounding box volume measurements. We will also discuss certain key business practices and operational quality considerations needed to bring standardization and drive appropriate ownership in the ecosystem.

Keywords: contour volume, logistics, real volume, volumetric weight

Procedia PDF Downloads 244
10377 The Optimal Order Policy for the Newsvendor Model under Worker Learning

Authors: Sunantha Teyarachakul

Abstract:

We consider the worker-learning Newsvendor Model, under the case of lost-sales for unmet demand, with the research objective of proposing the cost-minimization order policy and lot size, scheduled to arrive at the beginning of the selling-period. In general, the New Vendor Model is used to find the optimal order quantity for the perishable items such as fashionable products or those with seasonal demand or short-life cycles. Technically, it is used when the product demand is stochastic and available for the single selling-season, and when there is only a one time opportunity for the vendor to purchase, with possibly of long ordering lead-times. Our work differs from the classical Newsvendor Model in that we incorporate the human factor (specifically worker learning) and its influence over the costs of processing units into the model. We describe this by using the well-known Wright’s Learning Curve. Most of the assumptions of the classical New Vendor Model are still maintained in our work, such as the constant per-unit cost of leftover and shortage, the zero initial inventory, as well as the continuous time. Our problem is challenging in the way that the best order quantity in the classical model, which is balancing the over-stocking and under-stocking costs, is no longer optimal. Specifically, when adding the cost-saving from worker learning to such expected total cost, the convexity of the cost function will likely not be maintained. This has called for a new way in determining the optimal order policy. In response to such challenges, we found a number of characteristics related to the expected cost function and its derivatives, which we then used in formulating the optimal ordering policy. Examples of such characteristics are; the optimal order quantity exists and is unique if the demand follows a Uniform Distribution; if the demand follows the Beta Distribution with some specific properties of its parameters, the second derivative of the expected cost function has at most two roots; and there exists the specific level of lot size that satisfies the first order condition. Our research results could be helpful for analysis of supply chain coordination and of the periodic review system for similar problems.

Keywords: inventory management, Newsvendor model, order policy, worker learning

Procedia PDF Downloads 393
10376 An Analytical Wall Function for 2-D Shock Wave/Turbulent Boundary Layer Interactions

Authors: X. Wang, T. J. Craft, H. Iacovides

Abstract:

When handling the near-wall regions of turbulent flows, it is necessary to account for the viscous effects which are important over the thin near-wall layers. Low-Reynolds- number turbulence models do this by including explicit viscous and also damping terms which become active in the near-wall regions, and using very fine near-wall grids to properly resolve the steep gradients present. In order to overcome the cost associated with the low-Re turbulence models, a more advanced wall function approach has been implemented within OpenFoam and tested together with a standard log-law based wall function in the prediction of flows which involve 2-D shock wave/turbulent boundary layer interactions (SWTBLIs). On the whole, from the calculation of the impinging shock interaction, the three turbulence modelling strategies, the Lauder-Sharma k-ε model with Yap correction (LS), the high-Re k-ε model with standard wall function (SWF) and analytical wall function (AWF), display good predictions of wall-pressure. However, the SWF approach tends to underestimate the tendency of the flow to separate as a result of the SWTBLI. The analytical wall function, on the other hand, is able to reproduce the shock-induced flow separation and returns predictions similar to those of the low-Re model, using a much coarser mesh.

Keywords: SWTBLIs, skin-friction, turbulence modeling, wall function

Procedia PDF Downloads 324
10375 Effect of Cost Control and Cost Reduction Techniques in Organizational Performance

Authors: Babatunde Akeem Lawal

Abstract:

In any organization, the primary aim is to maximize profit, but the major challenges facing them is the increase in cost of operation because of this there is increase in cost of production that could lead to inevitable cost control and cost reduction scheme which make it difficult for most organizations to operate at the cost efficient frontier. The study aims to critically examine and evaluate the application of cost control and cost reduction in organization performance and also to review budget as an effective tool of cost control and cost reduction. A descriptive survey research was adopted. A total number of 40 respondent retrieved were used for the study. The analysis of data collected was undertaken by applying appropriate statistical tools. Regression analysis was used to test the hypothesis with the use of SPSS. Based on the findings; it was evident that cost control has a positive impact on organizational performance and also the style of management has a positive impact on organizational performance.

Keywords: organization, cost reduction, cost control, performance, budget, profit

Procedia PDF Downloads 567
10374 Globally Convergent Sequential Linear Programming for Multi-Material Topology Optimization Using Ordered Solid Isotropic Material with Penalization Interpolation

Authors: Darwin Castillo Huamaní, Francisco A. M. Gomes

Abstract:

The aim of the multi-material topology optimization (MTO) is to obtain the optimal topology of structures composed by many materials, according to a given set of constraints and cost criteria. In this work, we seek the optimal distribution of materials in a domain, such that the flexibility of the structure is minimized, under certain boundary conditions and the intervention of external forces. In the case we have only one material, each point of the discretized domain is represented by two values from a function, where the value of the function is 1 if the element belongs to the structure or 0 if the element is empty. A common way to avoid the high computational cost of solving integer variable optimization problems is to adopt the Solid Isotropic Material with Penalization (SIMP) method. This method relies on the continuous interpolation function, power function, where the base variable represents a pseudo density at each point of domain. For proper exponent values, the SIMP method reduces intermediate densities, since values other than 0 or 1 usually does not have a physical meaning for the problem. Several extension of the SIMP method were proposed for the multi-material case. The one that we explore here is the ordered SIMP method, that has the advantage of not being based on the addition of variables to represent material selection, so the computational cost is independent of the number of materials considered. Although the number of variables is not increased by this algorithm, the optimization subproblems that are generated at each iteration cannot be solved by methods that rely on second derivatives, due to the cost of calculating the second derivatives. To overcome this, we apply a globally convergent version of the sequential linear programming method, which solves a linear approximation sequence of optimization problems.

Keywords: globally convergence, multi-material design ordered simp, sequential linear programming, topology optimization

Procedia PDF Downloads 288