Search results for: supply chain optimization
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6672

Search results for: supply chain optimization

4902 Multi-Objective Optimization for the Green Vehicle Routing Problem: Approach to Case Study of the Newspaper Distribution Problem

Authors: Julio C. Ferreira, Maria T. A. Steiner

Abstract:

The aim of this work is to present a solution procedure referred to here as the Multi-objective Optimization for Green Vehicle Routing Problem (MOOGVRP) to provide solutions for a case study. The proposed methodology consists of three stages to resolve Scenario A. Stage 1 consists of the “treatment” of data; Stage 2 consists of applying mathematical models of the p-Median Capacitated Problem (with the objectives of minimization of distances and homogenization of demands between groups) and the Asymmetric Traveling Salesman Problem (with the objectives of minimizing distances and minimizing time). The weighted method was used as the multi-objective procedure. In Stage 3, an analysis of the results is conducted, taking into consideration the environmental aspects related to the case study, more specifically with regard to fuel consumption and air pollutant emission. This methodology was applied to a (partial) database that addresses newspaper distribution in the municipality of Curitiba, Paraná State, Brazil. The preliminary findings for Scenario A showed that it was possible to improve the distribution of the load, reduce the mileage and the greenhouse gas by 17.32% and the journey time by 22.58% in comparison with the current scenario. The intention for future works is to use other multi-objective techniques and an expanded version of the database and explore the triple bottom line of sustainability.

Keywords: Asymmetric Traveling Salesman Problem, Green Vehicle Routing Problem, Multi-objective Optimization, p-Median Capacitated Problem

Procedia PDF Downloads 113
4901 Automatic and High Precise Modeling for System Optimization

Authors: Stephanie Chen, Mitja Echim, Christof Büskens

Abstract:

To describe and propagate the behavior of a system mathematical models are formulated. Parameter identification is used to adapt the coefficients of the underlying laws of science. For complex systems this approach can be incomplete and hence imprecise and moreover too slow to be computed efficiently. Therefore, these models might be not applicable for the numerical optimization of real systems, since these techniques require numerous evaluations of the models. Moreover not all quantities necessary for the identification might be available and hence the system must be adapted manually. Therefore, an approach is described that generates models that overcome the before mentioned limitations by not focusing on physical laws, but on measured (sensor) data of real systems. The approach is more general since it generates models for every system detached from the scientific background. Additionally, this approach can be used in a more general sense, since it is able to automatically identify correlations in the data. The method can be classified as a multivariate data regression analysis. In contrast to many other data regression methods this variant is also able to identify correlations of products of variables and not only of single variables. This enables a far more precise and better representation of causal correlations. The basis and the explanation of this method come from an analytical background: the series expansion. Another advantage of this technique is the possibility of real-time adaptation of the generated models during operation. Herewith system changes due to aging, wear or perturbations from the environment can be taken into account, which is indispensable for realistic scenarios. Since these data driven models can be evaluated very efficiently and with high precision, they can be used in mathematical optimization algorithms that minimize a cost function, e.g. time, energy consumption, operational costs or a mixture of them, subject to additional constraints. The proposed method has successfully been tested in several complex applications and with strong industrial requirements. The generated models were able to simulate the given systems with an error in precision less than one percent. Moreover the automatic identification of the correlations was able to discover so far unknown relationships. To summarize the above mentioned approach is able to efficiently compute high precise and real-time-adaptive data-based models in different fields of industry. Combined with an effective mathematical optimization algorithm like WORHP (We Optimize Really Huge Problems) several complex systems can now be represented by a high precision model to be optimized within the user wishes. The proposed methods will be illustrated with different examples.

Keywords: adaptive modeling, automatic identification of correlations, data based modeling, optimization

Procedia PDF Downloads 409
4900 X-Ray Fluorescence Molecular Imaging with Improved Sensitivity for Biomedical Applications

Authors: Guohua Cao, Xu Dong

Abstract:

X-ray Fluorescence Molecular Imaging (XFMI) holds great promise as a low-cost molecular imaging modality for biomedical applications with high chemical sensitivity. However, for in vivo biomedical applications, a key technical bottleneck is the relatively low chemical sensitivity of XFMI, especially at a reasonably low radiation dose. In laboratory x-ray source based XFMI, one of the main factors that limits the chemical sensitivity of XFMI is the scattered x-rays. We will present our latest findings on improving the chemical sensitivity of XFMI using excitation beam spectrum optimization. XFMI imaging experiments on two mouse-sized phantoms were conducted at three different excitation beam spectra. Our results show that the minimum detectable concentration (MDC) of iodine can be readily increased by five times via excitation spectrum optimization. Findings from this investigation could find use for in vivo pre-clinical small-animal XFMI in the future.

Keywords: molecular imaging, X-ray fluorescence, chemical sensitivity, X-ray scattering

Procedia PDF Downloads 188
4899 Optimization of the Transfer Molding Process by Implementation of Online Monitoring Techniques for Electronic Packages

Authors: Burcu Kaya, Jan-Martin Kaiser, Karl-Friedrich Becker, Tanja Braun, Klaus-Dieter Lang

Abstract:

Quality of the molded packages is strongly influenced by the process parameters of the transfer molding. To achieve a better package quality and a stable transfer molding process, it is necessary to understand the influence of the process parameters on the package quality. This work aims to comprehend the relationship between the process parameters, and to identify the optimum process parameters for the transfer molding process in order to achieve less voids and wire sweep. To achieve this, a DoE is executed for process optimization and a regression analysis is carried out. A systematic approach is represented to generate models which enable an estimation of the number of voids and wire sweep. Validation experiments are conducted to verify the model and the results are presented.

Keywords: dielectric analysis, electronic packages, epoxy molding compounds, transfer molding process

Procedia PDF Downloads 383
4898 Opportunities for Reducing Post-Harvest Losses of Cactus Pear (Opuntia Ficus-Indica) to Improve Small-Holder Farmers Income in Eastern Tigray, Northern Ethiopia: Value Chain Approach

Authors: Meron Zenaselase Rata, Euridice Leyequien Abarca

Abstract:

The production of major crops in Northern Ethiopia, especially the Tigray Region, is at subsistence level due to drought, erratic rainfall, and poor soil fertility. Since cactus pear is a drought-resistant plant, it is considered as a lifesaver fruit and a strategy for poverty reduction in a drought-affected area of the region. Despite its contribution to household income and food security in the area, the cactus pear sub-sector is experiencing many constraints with limited attention given to its post-harvest loss management. Therefore, this research was carried out to identify opportunities for reducing post-harvest losses and recommend possible strategies to reduce post-harvest losses, thereby improving production and smallholder’s income. Both probability and non-probability sampling techniques were employed to collect the data. Ganta Afeshum district was selected from Eastern Tigray, and two peasant associations (Buket and Golea) were also selected from the district purposively for being potential in cactus pear production. Simple random sampling techniques were employed to survey 30 households from each of the two peasant associations, and a semi-structured questionnaire was used as a tool for data collection. Moreover, in this research 2 collectors, 2 wholesalers, 1 processor, 3 retailers, 2 consumers were interviewed; and two focus group discussion was also done with 14 key farmers using semi-structured checklist; and key informant interview with governmental and non-governmental organizations were interviewed to gather more information about the cactus pear production, post-harvest losses, the strategies used to reduce the post-harvest losses and suggestions to improve the post-harvest management. To enter and analyze the quantitative data, SPSS version 20 was used, whereas MS-word were used to transcribe the qualitative data. The data were presented using frequency and descriptive tables and graphs. The data analysis was also done using a chain map, correlations, stakeholder matrix, and gross margin. Mean comparisons like ANOVA and t-test between variables were used. The analysis result shows that the present cactus pear value chain involves main actors and supporters. However, there is inadequate information flow and informal market linkages among actors in the cactus pear value chain. The farmer's gross margin is higher when they sell to the processor than sell to collectors. The significant postharvest loss in the cactus pear value chain is at the producer level, followed by wholesalers and retailers. The maximum and minimum volume of post-harvest losses at the producer level is 4212 and 240 kgs per season. The post-harvest loss was caused by limited farmers skill on-farm management and harvesting, low market price, limited market information, absence of producer organization, poor post-harvest handling, absence of cold storage, absence of collection centers, poor infrastructure, inadequate credit access, using traditional transportation system, absence of quality control, illegal traders, inadequate research and extension services and using inappropriate packaging material. Therefore, some of the recommendations were providing adequate practical training, forming producer organizations, and constructing collection centers.

Keywords: cactus pear, post-harvest losses, profit margin, value-chain

Procedia PDF Downloads 134
4897 Development of Automated Quality Management System for the Management of Heat Networks

Authors: Nigina Toktasynova, Sholpan Sagyndykova, Zhanat Kenzhebayeva, Maksat Kalimoldayev, Mariya Ishimova, Irbulat Utepbergenov

Abstract:

Any business needs a stable operation and continuous improvement, therefore it is necessary to constantly interact with the environment, to analyze the work of the enterprise in terms of employees, executives and consumers, as well as to correct any inconsistencies of certain types of processes and their aggregate. In the case of heat supply organizations, in addition to suppliers, local legislation must be considered which often is the main regulator of pricing of services. In this case, the process approach used to build a functional organizational structure in these types of businesses in Kazakhstan is a challenge not only in the implementation, but also in ways of analyzing the employee's salary. To solve these problems, we investigated the management system of heating enterprise, including strategic planning based on the balanced scorecard (BSC), quality management in accordance with the standards of the Quality Management System (QMS) ISO 9001 and analysis of the system based on expert judgment using fuzzy inference. To carry out our work we used the theory of fuzzy sets, the QMS in accordance with ISO 9001, BSC according to the method of Kaplan and Norton, method of construction of business processes according to the notation IDEF0, theory of modeling using Matlab software simulation tools and graphical programming LabVIEW. The results of the work are as follows: We determined possibilities of improving the management of heat-supply plant-based on QMS; after the justification and adaptation of software tool it has been used to automate a series of functions for the management and reduction of resources and for the maintenance of the system up to date; an application for the analysis of the QMS based on fuzzy inference has been created with novel organization of communication software with the application enabling the analysis of relevant data of enterprise management system.

Keywords: balanced scorecard, heat supply, quality management system, the theory of fuzzy sets

Procedia PDF Downloads 368
4896 Parameters Tuning of a PID Controller on a DC Motor Using Honey Bee and Genetic Algorithms

Authors: Saeid Jalilzadeh

Abstract:

PID controllers are widely used to control the industrial plants because of their robustness and simple structures. Tuning of the controller's parameters to get a desired response is difficult and time consuming. With the development of computer technology and artificial intelligence in automatic control field, all kinds of parameters tuning methods of PID controller have emerged in endlessly, which bring much energy for the study of PID controller, but many advanced tuning methods behave not so perfect as to be expected. Honey Bee algorithm (HBA) and genetic algorithm (GA) are extensively used for real parameter optimization in diverse fields of study. This paper describes an application of HBA and GA to the problem of designing a PID controller whose parameters comprise proportionality constant, integral constant and derivative constant. Presence of three parameters to optimize makes the task of designing a PID controller more challenging than conventional P, PI, and PD controllers design. The suitability of the proposed approach has been demonstrated through computer simulation using MATLAB/SIMULINK.

Keywords: controller, GA, optimization, PID, PSO

Procedia PDF Downloads 545
4895 Life Cycle Analysis of the Antibacterial Gel Product Using Iso 14040 and Recipe 2016 Method

Authors: Pablo Andres Flores Siguenza, Noe Rodrigo Guaman Guachichullca

Abstract:

Sustainable practices have received increasing attention from academics and companies in recent decades due to, among many factors, the market advantages they generate, global commitments, and policies aimed at reducing greenhouse gas emissions, addressing resource scarcity, and rethinking waste management. The search for ways to promote sustainability leads industries to abandon classical methods and resort to the use of innovative strategies, which in turn are based on quantitative analysis methods and tools such as life cycle analysis (LCA), which is the basis for sustainable production and consumption, since it is a method that analyzes objectively, methodically, systematically, and scientifically the environmental impact caused by a process/product during its entire life cycle. The objective of this study is to develop an LCA of the antibacterial gel product throughout its entire supply chain (SC) under the methodology of ISO 14044 with the help of Gabi software and the Recipe 2016 method. The selection of the case study product was made based on its relevance in the current context of the COVID-19 pandemic and its exponential increase in production. For the development of the LCA, data from a Mexican company are used, and 3 scenarios are defined to obtain the midpoint and endpoint environmental impacts both by phases and globally. As part of the results, the most outstanding environmental impact categories are climate change, fossil fuel depletion, and terrestrial ecotoxicity, and the stage that generates the most pollution in the entire SC is the extraction of raw materials. The study serves as a basis for the development of different sustainability strategies, demonstrates the usefulness of an LCA, and agrees with different authors on the role and importance of this methodology in sustainable development.

Keywords: sustainability, sustainable development, life cycle analysis, environmental impact, antibacterial gel

Procedia PDF Downloads 58
4894 A Versatile Data Processing Package for Ground-Based Synthetic Aperture Radar Deformation Monitoring

Authors: Zheng Wang, Zhenhong Li, Jon Mills

Abstract:

Ground-based synthetic aperture radar (GBSAR) represents a powerful remote sensing tool for deformation monitoring towards various geohazards, e.g. landslides, mudflows, avalanches, infrastructure failures, and the subsidence of residential areas. Unlike spaceborne SAR with a fixed revisit period, GBSAR data can be acquired with an adjustable temporal resolution through either continuous or discontinuous operation. However, challenges arise from processing high temporal-resolution continuous GBSAR data, including the extreme cost of computational random-access-memory (RAM), the delay of displacement maps, and the loss of temporal evolution. Moreover, repositioning errors between discontinuous campaigns impede the accurate measurement of surface displacements. Therefore, a versatile package with two complete chains is developed in this study in order to process both continuous and discontinuous GBSAR data and address the aforementioned issues. The first chain is based on a small-baseline subset concept and it processes continuous GBSAR images unit by unit. Images within a window form a basic unit. By taking this strategy, the RAM requirement is reduced to only one unit of images and the chain can theoretically process an infinite number of images. The evolution of surface displacements can be detected as it keeps temporarily-coherent pixels which are present only in some certain units but not in the whole observation period. The chain supports real-time processing of the continuous data and the delay of creating displacement maps can be shortened without waiting for the entire dataset. The other chain aims to measure deformation between discontinuous campaigns. Temporal averaging is carried out on a stack of images in a single campaign in order to improve the signal-to-noise ratio of discontinuous data and minimise the loss of coherence. The temporal-averaged images are then processed by a particular interferometry procedure integrated with advanced interferometric SAR algorithms such as robust coherence estimation, non-local filtering, and selection of partially-coherent pixels. Experiments are conducted using both synthetic and real-world GBSAR data. Displacement time series at the level of a few sub-millimetres are achieved in several applications (e.g. a coastal cliff, a sand dune, a bridge, and a residential area), indicating the feasibility of the developed GBSAR data processing package for deformation monitoring of a wide range of scientific and practical applications.

Keywords: ground-based synthetic aperture radar, interferometry, small baseline subset algorithm, deformation monitoring

Procedia PDF Downloads 163
4893 Multimodal Optimization of Density-Based Clustering Using Collective Animal Behavior Algorithm

Authors: Kristian Bautista, Ruben A. Idoy

Abstract:

A bio-inspired metaheuristic algorithm inspired by the theory of collective animal behavior (CAB) was integrated to density-based clustering modeled as multimodal optimization problem. The algorithm was tested on synthetic, Iris, Glass, Pima and Thyroid data sets in order to measure its effectiveness relative to CDE-based Clustering algorithm. Upon preliminary testing, it was found out that one of the parameter settings used was ineffective in performing clustering when applied to the algorithm prompting the researcher to do an investigation. It was revealed that fine tuning distance δ3 that determines the extent to which a given data point will be clustered helped improve the quality of cluster output. Even though the modification of distance δ3 significantly improved the solution quality and cluster output of the algorithm, results suggest that there is no difference between the population mean of the solutions obtained using the original and modified parameter setting for all data sets. This implies that using either the original or modified parameter setting will not have any effect towards obtaining the best global and local animal positions. Results also suggest that CDE-based clustering algorithm is better than CAB-density clustering algorithm for all data sets. Nevertheless, CAB-density clustering algorithm is still a good clustering algorithm because it has correctly identified the number of classes of some data sets more frequently in a thirty trial run with a much smaller standard deviation, a potential in clustering high dimensional data sets. Thus, the researcher recommends further investigation in the post-processing stage of the algorithm.

Keywords: clustering, metaheuristics, collective animal behavior algorithm, density-based clustering, multimodal optimization

Procedia PDF Downloads 234
4892 Water Resources and Sanitation in Public Schools of Datu Odin Sinsuat, Maguindanao

Authors: Lahaina U. Dilangalen

Abstract:

Using descriptive-experimental research methods, this study aimed to identify the main resources of water, assessed the water quality, sanitation and hygiene practices, and extent of implementation. Complete enumeration was done in 28 elementary public schools of Datu Odin Sinsuat Municipality. Questionnaires were given to the school advisers. Water samples were obtained from the same schools and were submitted to the Department of Science and Technology (DOST) Region XII for microbial analysis, specifically the presence of fecal coliform bacteria. Four water resources such as hand pump, faucet, deep well and spring were found being used in the 28 schools. Of water resources, the only treated was from the faucet. Most of the schools used the water for drinking and washing. Two schools strongly agreed, nine schools agreed and seventeen schools disagreed that they implemented DepEd Order no.56 s. 2009. In addition, two schools strongly agreed and twenty six agreed that they implemented DepEd Order No. 65 s. 2009. Five schools had water supply that were safe to drink while sixteen schools had water supply that were not safe to drink due to high fecal coliform count and did not undergo chemical treatment. The only safe for drinking were water resources that came from faucet because they were chemically treated. Seven out of 28 schools did not have water supply due to their location in mountainous areas. More than half of the schools did not comply with the DepEd Order No. 56 s. 2009 due to the lack of funds and support from the PTA and LGU. It is recommended that the Department of Education must have an urgent assessment of implementing both DepEd Orders No.56 and 65, to assure that the schoolchildren be protected from water and sanitation related ailments. Also, all water resources that are not treated must be used for washing only. Ideally, all the water resources must be treated to assure the safety of all school constituents. Moreover, the school administrators and teachers in the municipality must be provided copies of the results of this study for reference in implementing the said programs.

Keywords: assessment, drinking water, fecal coliform, groundwater

Procedia PDF Downloads 253
4891 Discrete Swarm with Passive Congregation for Cost Minimization of the Multiple Vehicle Routing Problem

Authors: Tarek Aboueldahab, Hanan Farag

Abstract:

Cost minimization of Multiple Vehicle Routing Problem becomes a critical issue in the field of transportation because it is NP-hard optimization problem and the search space is complex. Many researches use the hybridization of artificial intelligence (AI) models to solve this problem; however, it can not guarantee to reach the best solution due to the difficulty of searching the whole search space. To overcome this problem, we introduce the hybrid model of Discrete Particle Swarm Optimization (DPSO) with a passive congregation which enable searching the whole search space to compromise between both local and global search. The practical experiment shows that our model obviously outperforms other hybrid models in cost minimization.

Keywords: cost minimization, multi-vehicle routing problem, passive congregation, discrete swarm, passive congregation

Procedia PDF Downloads 99
4890 Development of an Optimization Method for Myoelectric Signal Processing by Active Matrix Sensing in Robot Rehabilitation

Authors: Noriyoshi Yamauchi, Etsuo Horikawa, Takunori Tsuji

Abstract:

Training by exoskeleton robot is drawing attention as a rehabilitation method for body paralysis seen in many cases, and there are many forms that assist with the myoelectric signal generated by exercise commands from the brain. Rehabilitation requires more frequent training, but it is one of the reasons that the technology is required for the identification of the myoelectric potential derivation site and attachment of the device is preventing the spread of paralysis. In this research, we focus on improving the efficiency of gait training by exoskeleton type robots, improvement of myoelectric acquisition and analysis method using active matrix sensing method, and improvement of walking rehabilitation and walking by optimization of robot control.

Keywords: active matrix sensing, brain machine interface (BMI), the central pattern generator (CPG), myoelectric signal processing, robot rehabilitation

Procedia PDF Downloads 386
4889 Faulty Sensors Detection in Planar Array Antenna Using Pelican Optimization Algorithm

Authors: Shafqat Ullah Khan, Ammar Nasir

Abstract:

Using planar antenna array (PAA) in radars, Broadcasting, satellite antennas, and sonar for the detection of targets, Helps provide instant beam pattern control. High flexibility and Adaptability are achieved by multiple beam steering by using a Planar array and are particularly needed in real-life Sanrio’s where the need arises for several high-directivity beams. Faulty sensors in planar arrays generate asymmetry, which leads to service degradation, radiation pattern distortion, and increased levels of sidelobe. The POA, a nature-inspired optimization algorithm, accurately determines faulty sensors within an array, enhancing the reliability and performance of planar array antennas through extensive simulations and experiments. The analysis was done for different types of faults in 7 x 7 and 8 x 8 planar arrays in MATLAB.

Keywords: Planar antenna array, , Pelican optimisation Algorithm, , Faculty sensor, Antenna arrays

Procedia PDF Downloads 84
4888 Optimization and Operation of Charging and Discharging Stations for Hybrid Cars and their Effects on the Electricity Distribution Network

Authors: Ali Heydarimoghim

Abstract:

In this paper, the optimal placement of charging and discharging stations is done to determine the location and capacity of the stations, reducing the cost of electric vehicle owners' losses, reducing the cost of distribution system losses, and reducing the costs associated with the stations. Also, observing the permissible limits of the bus voltage and the capacity of the stations and their distance are considered as constraints of the problem. Given the traffic situation in different areas of a city, we estimate the amount of energy required to charge and the amount of energy provided to discharge electric vehicles in each area. We then introduce the electricity distribution system of the city in question. Following are the scenarios for introducing the problem and introducing the objective and constraint functions. Finally, the simulation results for different scenarios are compared.

Keywords: charging & discharging stations, hybrid vehicles, optimization, replacement

Procedia PDF Downloads 139
4887 Enhancing Rupture Pressure Prediction for Corroded Pipes Through Finite Element Optimization

Authors: Benkouiten Imene, Chabli Ouerdia, Boutoutaou Hamid, Kadri Nesrine, Bouledroua Omar

Abstract:

Algeria is actively enhancing gas productivity by augmenting the supply flow. However, this effort has led to increased internal pressure, posing a potential risk to the pipeline's integrity, particularly in the presence of corrosion defects. Sonatrach relies on a vast network of pipelines spanning 24,000 kilometers for the transportation of gas and oil. The aging of these pipelines raises the likelihood of corrosion both internally and externally, heightening the risk of ruptures. To address this issue, a comprehensive inspection is imperative, utilizing specialized scraping tools. These advanced tools furnish a detailed assessment of all pipeline defects. It is essential to recalculate the pressure parameters to safeguard the corroded pipeline's integrity while ensuring the continuity of production. In this context, Sonatrach employs symbolic pressure limit calculations, such as ASME B31G (2009) and the modified ASME B31G (2012). The aim of this study is to perform a comparative analysis of various limit pressure calculation methods documented in the literature, namely DNV RP F-101, SHELL, P-CORRC, NETTO, and CSA Z662. This comparative assessment will be based on a dataset comprising 329 burst tests published in the literature. Ultimately, we intend to introduce a novel approach grounded in the finite element method, employing ANSYS software.

Keywords: pipeline burst pressure, burst test, corrosion defect, corroded pipeline, finite element method

Procedia PDF Downloads 58
4886 Regional Rates of Sand Supply to the New South Wales Coast: Southeastern Australia

Authors: Marta Ribo, Ian D. Goodwin, Thomas Mortlock, Phil O’Brien

Abstract:

Coastal behavior is best investigated using a sediment budget approach, based on the identification of sediment sources and sinks. Grain size distribution over the New South Wales (NSW) continental shelf has been widely characterized since the 1970’s. Coarser sediment has generally accumulated on the outer shelf, and/or nearshore zones, with the latter related to the presence of nearshore reef and bedrocks. The central part of the NSW shelf is characterized by the presence of fine sediments distributed parallel to the coastline. This study presents new grain size distribution maps along the NSW continental shelf, built using all available NSW and Commonwealth Government holdings. All available seabed bathymetric data form prior projects, single and multibeam sonar, and aerial LiDAR surveys were integrated into a single bathymetric surface for the NSW continental shelf. Grain size information was extracted from the sediment sample data collected in more than 30 studies. The information extracted from the sediment collections varied between reports. Thus, given the inconsistency of the grain size data, a common grain size classification was her defined using the phi scale. The new sediment distribution maps produced, together with new detailed seabed bathymetric data enabled us to revise the delineation of sediment compartments to more accurately reflect the true nature of sediment movement on the inner shelf and nearshore. Accordingly, nine primary mega coastal compartments were delineated along the NSW coast and shelf. The sediment compartments are bounded by prominent nearshore headlands and reefs, and major river and estuarine inlets that act as sediment sources and/or sinks. The new sediment grain size distribution was used as an input in the morphological modelling to quantify the sediment transport patterns (and indicative rates of transport), used to investigate sand supply rates and processes from the lower shoreface to the NSW coast. The rate of sand supply to the NSW coast from deep water is a major uncertainty in projecting future coastal response to sea-level rise. Offshore transport of sand is generally expected as beaches respond to rising sea levels but an onshore supply from the lower shoreface has the potential to offset some of the impacts of sea-level rise, such as coastline recession. Sediment exchange between the lower shoreface and sub-aerial beach has been modelled across the south, central, mid-north and far-north coast of NSW. Our model approach is that high-energy storm events are the primary agents of sand transport in deep water, while non-storm conditions are responsible for re-distributing sand within the beach and surf zone.

Keywords: New South Wales coast, off-shore transport, sand supply, sediment distribution maps

Procedia PDF Downloads 228
4885 On the Dwindling Supply of the Observable Cosmic Microwave Background Radiation

Authors: Jia-Chao Wang

Abstract:

The cosmic microwave background radiation (CMB) freed during the recombination era can be considered as a photon source of small duration; a one-time event happened everywhere in the universe simultaneously. If space is divided into concentric shells centered at an observer’s location, one can imagine that the CMB photons originated from the nearby shells would reach and pass the observer first, and those in shells farther away would follow as time goes forward. In the Big Bang model, space expands rapidly in a time-dependent manner as described by the scale factor. This expansion results in an event horizon coincident with one of the shells, and its radius can be calculated using cosmological calculators available online. Using Planck 2015 results, its value during the recombination era at cosmological time t = 0.379 million years (My) is calculated to be Revent = 56.95 million light-years (Mly). The event horizon sets a boundary beyond which the freed CMB photons will never reach the observer. The photons within the event horizon also exhibit a peculiar behavior. Calculated results show that the CMB observed today was freed in a shell located at 41.8 Mly away (inside the boundary set by Revent) at t = 0.379 My. These photons traveled 13.8 billion years (Gy) to reach here. Similarly, the CMB reaching the observer at t = 1, 5, 10, 20, 40, 60, 80, 100 and 120 Gy are calculated to be originated at shells of R = 16.98, 29.96, 37.79, 46.47, 53.66, 55.91, 56.62, 56.85 and 56.92 Mly, respectively. The results show that as time goes by, the R value approaches Revent = 56.95 Mly but never exceeds it, consistent with the earlier statement that beyond Revent the freed CMB photons will never reach the observer. The difference Revert - R can be used as a measure of the remaining observable CMB photons. Its value becomes smaller and smaller as R approaching Revent, indicating a dwindling supply of the observable CMB radiation. In this paper, detailed dwindling effects near the event horizon are analyzed with the help of online cosmological calculators based on the lambda cold dark matter (ΛCDM) model. It is demonstrated in the literature that assuming the CMB to be a blackbody at recombination (about 3000 K), then it will remain so over time under cosmological redshift and homogeneous expansion of space, but with the temperature lowered (2.725 K now). The present result suggests that the observable CMB photon density, besides changing with space expansion, can also be affected by the dwindling supply associated with the event horizon. This raises the question of whether the blackbody of CMB at recombination can remain so over time. Being able to explain the blackbody nature of the observed CMB is an import part of the success of the Big Bang model. The present results cast some doubts on that and suggest that the model may have an additional challenge to deal with.

Keywords: blackbody of CMB, CMB radiation, dwindling supply of CMB, event horizon

Procedia PDF Downloads 120
4884 A Dirty Page Migration Method in Process of Memory Migration Based on Pre-copy Technology

Authors: Kang Zijian, Zhang Tingyu, Burra Venkata Durga Kumar

Abstract:

This article investigates the challenges in memory migration during the live migration of virtual machines. We found three challenges probably existing in pre-copy technology. One of the main challenges is the challenge of downtime migration. Decrease the downtime could promise the normal work for a virtual machine. Although pre-copy technology is greatly decreasing the downtime, we still need to shut down the machine in order to finish the last round of data transfer. This paper provides an optimization scheme for the problems existing in pro-copy technology, mainly the optimization of the dirty page migration mechanism. The typical pre-copy technology copy n-1th’s dirty pages in nth turn. However, our idea is to create a double iteration method to solve this problem.

Keywords: virtual machine, pre-copy technology, memory migration process, downtime, dirty pages migration method

Procedia PDF Downloads 152
4883 Synthesis and Optimization of Bio Metal-Organic Framework with Permanent Porosity

Authors: Tia Kristian Tajnšek, Matjaž Mazaj, Nataša Zabukovec Logar

Abstract:

Metal-organic frameworks (MOFs) with their specific properties and the possibility of tuning the structure represent excellent candidates for use in the biomedical field. Their advantage lies in large pore surfaces and volumes, as well as the possibility of using bio-friendly or bioactive constituents. So-called bioMOFs are representatives of MOFs, which are constructed from at least one biomolecule (metal, a small bioactive molecule in metal clusters and/or linker) and are intended for bio-application (usually in the field of medicine; most commonly drug delivery). When designing a bioMOF for biomedical applications, we should adhere to some guidelines for an improved toxicological profile of the material. Such as (i) choosing an endogenous/nontoxic metal, (ii) GRAS (generally recognized as safe) linker, and (iii) nontoxic solvents. Design and synthesis of bioNICS-1 (bioMOF of National Institute of Chemistry Slovenia – 1) consider all these guidelines. Zinc (Zn) was chosen as an endogenous metal with an agreeable recommended daily intake (RDI) and LD50 value, and ascorbic acid (Vitamin C) was chosen as a GRAS and active linker. With these building blocks, we have synthesized a bioNICS-1 material. The synthesis was done in ethanol using a solvothermal method. The synthesis protocol was further optimized in three separate ways. Optimization of (i) synthesis parameters to improve the yield of the synthesis, (ii) input reactant ratio and addition of specific modulators for production of larger crystals, and (iii) differing of the heating source (conventional, microwave and ultrasound) to produce nano-crystals. With optimization strategies, the synthesis yield was increased. Larger crystals were prepared for structural analysis with the use of a proper species and amount of modulator. Synthesis protocol was adjusted to different heating sources, resulting in the production of nano-crystals of bioNICS-1 material. BioNICS-1 was further activated in ethanol and structurally characterized, resolving the crystal structure of new material.

Keywords: ascorbic acid, bioMOF, MOF, optimization, synthesis, zinc ascorbate

Procedia PDF Downloads 144
4882 CRYPTO COPYCAT: A Fashion Centric Blockchain Framework for Eliminating Fashion Infringement

Authors: Magdi Elmessiry, Adel Elmessiry

Abstract:

The fashion industry represents a significant portion of the global gross domestic product, however, it is plagued by cheap imitators that infringe on the trademarks which destroys the fashion industry's hard work and investment. While eventually the copycats would be found and stopped, the damage has already been done, sales are missed and direct and indirect jobs are lost. The infringer thrives on two main facts: the time it takes to discover them and the lack of tracking technologies that can help the consumer distinguish them. Blockchain technology is a new emerging technology that provides a distributed encrypted immutable and fault resistant ledger. Blockchain presents a ripe technology to resolve the infringement epidemic facing the fashion industry. The significance of the study is that a new approach leveraging the state of the art blockchain technology coupled with artificial intelligence is used to create a framework addressing the fashion infringement problem. It transforms the current focus on legal enforcement, which is difficult at best, to consumer awareness that is far more effective. The framework, Crypto CopyCat, creates an immutable digital asset representing the actual product to empower the customer with a near real time query system. This combination emphasizes the consumer's awareness and appreciation of the product's authenticity, while provides real time feedback to the producer regarding the fake replicas. The main findings of this study are that implementing this approach can delay the fake product penetration of the original product market, thus allowing the original product the time to take advantage of the market. The shift in the fake adoption results in reduced returns, which impedes the copycat market and moves the emphasis to the original product innovation.

Keywords: fashion, infringement, blockchain, artificial intelligence, textiles supply chain

Procedia PDF Downloads 261
4881 Design Optimization of Chevron Nozzles for Jet Noise Reduction

Authors: E. Manikandan, C. Chilambarasan, M. Sulthan Ariff Rahman, S. Kanagaraj, V. R. Sanal Kumar

Abstract:

The noise regulations around the major airports and rocket launching stations due to the environmental concern have made jet noise a crucial problem in the present day aero-acoustics research. The three main acoustic sources in jet nozzles are aerodynamics noise, noise from craft systems and engine and mechanical noise. Note that the majority of engine noise is due to the jet noise coming out from the exhaust nozzle. The previous studies reveal that the potential of chevron nozzles for aircraft engines noise reduction is promising owing to the fact that the jet noise continues to be the dominant noise component, especially during take-off. In this paper parametric analytical studies have been carried out for optimizing the number of chevron lobes, the lobe length and tip shape, and the level of penetration of the chevrons into the flow over a variety of flow conditions for various aerospace applications. The numerical studies have been carried out using a validated steady 3D density based, SST k-ω turbulence model with enhanced wall functions. In the numerical study, a fully implicit finite volume scheme of the compressible, Navier–Stokes equations is employed. We inferred that the geometry optimization of an environmental friendly chevron nozzle with a suitable number of chevron lobes with aerodynamically efficient tip contours for facilitating silent exit flow will enable a commendable sound reduction without much thrust penalty while comparing with the conventional supersonic nozzles with same area ratio.

Keywords: chevron nozzle, jet acoustic level, jet noise suppression, shape optimization of chevron nozzles

Procedia PDF Downloads 311
4880 Towards Sustainable Consumption: A Framework for Assessing Supplier's Commitment

Authors: O. O. Oguntoye

Abstract:

Product consumption constitutes an important consideration for sustainable development. Seeing how product consumption could be highly unsustainable, coupled with how existing policies on corporate responsibility do not particularly address the consumption aspect of product lifecycle, conducting this research became necessary. The research makes an attempt to provide a framework by which to gauge corporate responsibility of product suppliers in terms of their commitment towards the sustainable consumption of their products. Through an exploration of relevant literature, independently established ideas with which to assess a given product supplier were galvanised into a four-criterion framework. The criteria are: (1) Embeddedness of consumption as a factor in corporate sustainability policy, (2) Level of understanding of consumption behaviour, (3) Breadth of behaviour-influencing strategies adopted, and (4) Inclusiveness for all main dimensions of sustainability. This resulting framework was then applied in a case study involving a UK-based furniture supplier where interviews and content analysis of corporate documents were used as the mode for primary data collection. From the case study, it was found that the supplier had performed to different levels across the four themes of the assessment. Two major areas for improvement were however identified – one is for the furniture supplier to focus more proactively on understanding consumption behaviour and, two is for it to widen the scope of its current strategies for enhancing sustainable consumption of supplied furniture. As a generalisation, the framework presented here makes it possible for companies to reflect with a sense of guidance, how they have demonstrated commitment towards sustainable consumption through their values, culture, and operations. It also provides a foundation for developing standardized assessment which the current widely used frameworks such as the GRI, the Global Compact, and others do not cover. While these popularly used frameworks mainly focus on sustainability of companies within the production and supply chain management contexts (i.e. mostly ‘upstream’), the framework here provides an extension by bringing the ‘downstream’ or consumer bit into light.

Keywords: corporate sustainability, design for sustainable consumption, extended producer responsibility, sustainable consumer behaviour

Procedia PDF Downloads 419
4879 Solving Directional Overcurrent Relay Coordination Problem Using Artificial Bees Colony

Authors: M. H. Hussain, I. Musirin, A. F. Abidin, S. R. A. Rahim

Abstract:

This paper presents the implementation of Artificial Bees Colony (ABC) algorithm in solving Directional OverCurrent Relays (DOCRs) coordination problem for near-end faults occurring in fixed network topology. The coordination optimization of DOCRs is formulated as linear programming (LP) problem. The objective function is introduced to minimize the operating time of the associated relay which depends on the time multiplier setting. The proposed technique is to taken as a technique for comparison purpose in order to highlight its superiority. The proposed algorithms have been tested successfully on 8 bus test system. The simulation results demonstrated that the ABC algorithm which has been proved to have good search ability is capable in dealing with constraint optimization problems.

Keywords: artificial bees colony, directional overcurrent relay coordination problem, relay settings, time multiplier setting

Procedia PDF Downloads 331
4878 Organization of the Purchasing Function for Innovation

Authors: Jasna Prester, Ivana Rašić Bakarić, Božidar Matijević

Abstract:

Various prominent scholars and substantial practitioner-oriented literature on innovation orientation have shown positive effects on firm performance. There is a myriad of factors that influence and enhance innovation but it has been found in the literature that new product innovations accounted for an average of 14 percent of sales revenues for all firms. If there is one thing that has changed in innovation management during the last decade, it is the growing reliance on external partners. As a consequence, a new task for purchasing arises, as firms need to understand which suppliers actually do have high potential contributing to the innovativeness of the firm and which do not. Purchasing function in an organization is extremely important as it deals on an average of 50% or more of a firm's expenditures. In the nineties the purchasing department was largely seen as a transaction-oriented, clerical function but today purchasing integration provides a formal interface mechanism between purchasing and other firm functions that services other functions within the company. Purchasing function has to be organized differently to enable firm innovation potential. However, innovations are inherently risky. There are behavioral risk (that some partner will take advantage of the other party), technological risk in terms of complexity of products and processes of manufacturing and incoming materials and finally market risks, which in fact judge the value of the innovation. These risks are investigated in this work since it has been found in the literature that the higher the technological risk, higher will be the centralization of the purchasing function as an interface with other supply chain members. Most researches on organization of purchasing function were done by case study analysis of innovative firms. This work actually tends to prove or discard results found in the literature based on case study method. A large data set of 1493 companies, from 25 countries collected in the GMRG 4 survey served as a basis for analysis.

Keywords: purchasing function organization, innovation, technological risk, GMRG 4 survey

Procedia PDF Downloads 483
4877 Clients’ Priorities in Delivery of Green Projects: South African Perspective

Authors: C. Mothobiso, D. Root

Abstract:

Purpose: This study attempts to identify the clients’ main priorities when delivering green projects. The aim is to compare if the clients have the same interest that are similar in delivery of convectional buildings as compared to green buildings. The main purpose is to find why other clients are investing in green buildings while others are reluctant and adopting green building at a slow pace. Design/methodology/approach: A sample of construction professional accredited by the Green Building Council of South Africa (GBCSA) was sent a questionnaire to participate in the research. Since GBSCSA accredited professionals have knowledge and experience about the green buildings, they are chosen as the sample. The research is qualitative because it evaluates the perceptions and knowledge around the subject matter. Research limitations: The research focuses only on the South African construction clients. Findings: Findings reveal that private clients invest more on green buildings as compared to government and parastatal entities. Private clients prioritise on maximising returns on investments and they mainly invest on buildings that save energies and have low life cycle costs. Private clients are perceived to be more knowledgeable about the benefits of green building project as compared to government and Parastatals clients. Shortage of expertise and managerial skill leads to low adaptation of green buildings in the government and parastatal projects. Other factors, which seem to disintegrate the adoption of green buildings, are the readiness of supply chain within the industry and inappropriate procurements strategies adopted by clients. The evaluation of the clients’ priorities will enable the design team to come up with innovative ways to approach the design process so that clients’ priorities and needs are identified and met. Practical implications: The findings are indicating that clients’ needs and priorities have a huge impact on the delivery of the project in terms of time, quality and cost of the project.

Keywords: construction clients, design team, green construction and project deliver

Procedia PDF Downloads 273
4876 Performance Analysis and Multi-Objective Optimization of a Kalina Cycle for Low-Temperature Applications

Authors: Sadegh Sadeghi, Negar Shabani

Abstract:

From a thermal point of view, zeotropic mixtures are likely to be more efficient than azeotropic fluids in low-temperature thermodynamic cycles due to their suitable boiling characteristics. In this study, performance of a low-temperature Kalina cycle with R717/water working fluid used in different existing power plants is mathematically investigated. To analyze the behavior of the cycle, mass conservation, energy conservation, and exergy balance equations are presented. With regard to the similarity in molar mass of R717 (17.03 gr/mol) and water (18.01 gr/mol), there is no need to alter the size of Kalina system components such as turbine and pump. To optimize the cycle energy and exergy efficiencies simultaneously, a constrained multi-objective optimization is carried out applying an Artificial Bee Colony algorithm. The main motivation behind using this algorithm lies on its robustness, reliability, remarkable precision and high–speed convergence rate in dealing with complicated constrained multi-objective problems. Convergence rates of the algorithm for calculating the optimal energy and exergy efficiencies are presented. Subsequently, due to the importance of exergy concept in Kalina cycles, exergy destructions occurring in the components are computed. Finally, the impacts of pressure, temperature, mass fraction and mass flow rate on the energy and exergy efficiencies are elaborately studied.

Keywords: artificial bee colony algorithm, binary zeotropic mixture, constrained multi-objective optimization, energy efficiency, exergy efficiency, Kalina cycle

Procedia PDF Downloads 155
4875 Non-Linear Static Pushover Analysis of 15 Storied Reinforced Concrete Building Structure with Shear Wall

Authors: Hamid Nikzad, Shinta Yoshitomi

Abstract:

In this paper, nonlinear static pushover analysis is performed on 15 storied RC building structure with a shear wall to evaluate the seismic performance of the building. Section sizes of the members are obtained based on structural optimization method utilizing MATLAB frame optimizer, then the structure is simulated and designed in ETABS program conforming ACI 318-14 design code. The pushover curve has been generated by pushing the top node of the structure to the limited target displacement. Members failure due to the formation of plastic hinges, considering shear wall-frame structure was observed and the result of this study is presented based on current regulation of FEMA356, ASCE7-10, and ACI 318-14 design criteria

Keywords: structural optimization, linear static analysis, ETABS, MATLAB, RC moment frame, RC shear wall structures

Procedia PDF Downloads 162
4874 Bi-Criteria Vehicle Routing Problem for Possibility Environment

Authors: Bezhan Ghvaberidze

Abstract:

A multiple criteria optimization approach for the solution of the Fuzzy Vehicle Routing Problem (FVRP) is proposed. For the possibility environment the levels of movements between customers are calculated by the constructed simulation interactive algorithm. The first criterion of the bi-criteria optimization problem - minimization of the expectation of total fuzzy travel time on closed routes is constructed for the FVRP. A new, second criterion – maximization of feasibility of movement on the closed routes is constructed by the Choquet finite averaging operator. The FVRP is reduced to the bi-criteria partitioning problem for the so called “promising” routes which were selected from the all admissible closed routes. The convenient selection of the “promising” routes allows us to solve the reduced problem in the real-time computing. For the numerical solution of the bi-criteria partitioning problem the -constraint approach is used. An exact algorithm is implemented based on D. Knuth’s Dancing Links technique and the algorithm DLX. The Main objective was to present the new approach for FVRP, when there are some difficulties while moving on the roads. This approach is called FVRP for extreme conditions (FVRP-EC) on the roads. Also, the aim of this paper was to construct the solving model of the constructed FVRP. Results are illustrated on the numerical example where all Pareto-optimal solutions are found. Also, an approach for more complex model FVRP with time windows was developed. A numerical example is presented in which optimal routes are constructed for extreme conditions on the roads.

Keywords: combinatorial optimization, Fuzzy Vehicle routing problem, multiple objective programming, possibility theory

Procedia PDF Downloads 489
4873 Mechanical Study Printed Circuit Boards Bonding for Jefferson Laboratory Detector

Authors: F. Noto, F. De Persio, V. Bellini, G. Costa. F. Mammoliti, F. Meddi, C. Sutera, G. M. Urcioli

Abstract:

One plane X and one plane Y of silicon microstrip detectors will constitute the front part of the Super Bigbite Spectrometer that is under construction and that will be installed in the experimental Hall A of the Thomas Jefferson National Accelerator Facility (Jefferson Laboratory), located in Newport News, Virgina, USA. Each plane will be made up by two nearly identical, 300 μm thick, 10 cm x 10.3 cm wide silicon microstrip detectors with 50 um pitch, whose electronic signals will be transferred to the front-end electronic based on APV25 chips through C-shaped FR4 Printed Circuit Boards (PCB). A total of about 10000 strips are read-out. This paper treats the optimization of the detector support structure, the materials used through a finite element simulation. A very important aspect of the study will also cover the optimization of the bonding parameters between detector and electronics.

Keywords: FEM analysis, bonding, SBS tracker, mechanical structure

Procedia PDF Downloads 339