Search results for: generalized linear model
3294 Optimization by Means of Genetic Algorithm of the Equivalent Electrical Circuit Model of Different Order for Li-ion Battery Pack
Authors: V. Pizarro-Carmona, S. Castano-Solis, M. Cortés-Carmona, J. Fraile-Ardanuy, D. Jimenez-Bermejo
Abstract:
The purpose of this article is to optimize the Equivalent Electric Circuit Model (EECM) of different orders to obtain greater precision in the modeling of Li-ion battery packs. Optimization includes considering circuits based on 1RC, 2RC and 3RC networks, with a dependent voltage source and a series resistor. The parameters are obtained experimentally using tests in the time domain and in the frequency domain. Due to the high non-linearity of the behavior of the battery pack, Genetic Algorithm (GA) was used to solve and optimize the parameters of each EECM considered (1RC, 2RC and 3RC). The objective of the estimation is to minimize the mean square error between the measured impedance in the real battery pack and those generated by the simulation of different proposed circuit models. The results have been verified by comparing the Nyquist graphs of the estimation of the complex impedance of the pack. As a result of the optimization, the 2RC and 3RC circuit alternatives are considered as viable to represent the battery behavior. These battery pack models are experimentally validated using a hardware-in-the-loop (HIL) simulation platform that reproduces the well-known New York City cycle (NYCC) and Federal Test Procedure (FTP) driving cycles for electric vehicles. The results show that using GA optimization allows obtaining EECs with 2RC or 3RC networks, with high precision to represent the dynamic behavior of a battery pack in vehicular applications.
Keywords: Li-ion battery packs modeling optimized, EECM, GA, electric vehicle applications.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5443293 Convective Hot Air Drying of Different Varieties of Blanched Sweet Potato Slices
Authors: M. O. Oke, T. S. Workneh
Abstract:
Drying behavior of blanched sweet potato in a cabinet dryer using different five air temperatures (40-80°C) and ten sweet potato varieties sliced to 5mm thickness were investigated. The drying data were fitted to eight models. The Modified Henderson and Pabis model gave the best fit to the experimental moisture ratio data obtained during the drying of all the varieties while Newton (Lewis) and Wang and Singh models gave the least fit. The values of Deff obtained for Bophelo variety (1.27 x 10-9 to 1.77 x 10-9 m2/s) was the least while that of S191 (1.93 x 10-9 to 2.47 x 10-9 m2/s) was the highest which indicates that moisture diffusivity in sweet potato is affected by the genetic factor. Activation energy values ranged from 0.27-6.54 kJ/mol. The lower activation energy indicates that drying of sweet potato slices requires less energy and is hence a cost and energy saving method. The drying behavior of blanched sweet potato was investigated in a cabinet dryer. Drying time decreased considerably with increase in hot air temperature. Out of the eight models fitted, the Modified Henderson and Pabis model gave the best fit to the experimental moisture ratio data on all the varieties while Newton, Wang and Singh models gave the least. The lower activation energy (0.27 - 6.54 kJ/mol) obtained indicates that drying of sweet potato slices requires less energy and is hence a cost and energy saving method.
Keywords: Sweet Potato Slice, Drying Models, Moisture Ratio, Moisture Diffusivity, Activation Energy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30043292 Optimization and GIS-Based Intelligent Decision Support System for Urban Transportation Systems Analysis
Authors: Mohamad K. Hasan, Hameed Al-Qaheri
Abstract:
Optimization plays an important role in most real world applications that support decision makers to take the right decision regarding the strategic directions and operations of the system they manage. Solutions for traffic management and traffic congestion problems are considered major problems that most decision making authorities for cities around the world are looking for. This review paper gives a full description of the traffic problem as part of the transportation planning process and present a view as a framework of urban transportation system analysis where the core of the system is a transportation network equilibrium model that is based on optimization techniques and that can also be used for evaluating an alternative solution or a combination of alternative solutions for the traffic congestion. Different transportation network equilibrium models are reviewed from the sequential approach to the multiclass combining trip generation, trip distribution, modal split, trip assignment and departure time model. A GIS-Based intelligent decision support system framework for urban transportation system analysis is suggested for implementation where the selection of optimized alternative solutions, single or packages, will be based on an intelligent agent rather than human being which would lead to reduction in time, cost and the elimination of the difficulty, by human being, for finding the best solution to the traffic congestion problem.Keywords: Multiclass simultaneous transportation equilibrium models, transportation planning, urban transportation systems analysis, intelligent decision support system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23013291 The Research of Taiwan Green Building Materials (GBM) system and GBM Eco-Efficiency Model on Climate Change
Authors: Ting-Ting Hsieh, Che-Ming Chiang, Ming-Chin Ho, Kwang-Pang Lai
Abstract:
The globe Sustainability has become the subject of international attention, the key reason is that global climate change. Climate and disasters around the abnormal frequency multiplier, the global temperature of the catastrophe and disaster continue to occur throughout the world, as well as countries around the world. Currently there are many important international conferences and policy, it is a "global environmental sustainability " and "living human health " as the goal of development, including the APEC 2007 meeting to "climate Clean Energy" as the theme Sydney Declaration, 2008 World Economic Forum's "Carbon - promote Cool Earth energy efficiency improvement project", the EU proposed "Green Idea" program, the Japanese annual policy, "low-carbon society, sustainable eco-city environment (Eco City) "And from 2009 to 2010 to promote the "Eco-Point" to promote green energy and carbon reduction products .And the 2010 World Climate Change Conference (COP16 United Nations Climate Change Conference Copenhagen), the world has been the subject of Negative conservative "Environmental Protection ", "save energy consumption, " into a positive response to the "Sustainable " and" LOHAS", while Taiwan has actively put forward eco-cities, green building, green building materials and other related environmental response Measures, especially green building construction environment that is the basis of factors, the most widely used application level, and direct contact with human health and the key to sustainable planet. "Sustainable development "is a necessary condition for continuation of the Earth, "healthy and comfortable" is a necessary condition for the continuation of life, and improve the "quality" is a necessary condition for economic development, balance between the three is "to enhance the efficiency of ", According to the World Business Council for Sustainable Development (WBCSD) for the "environmental efficiency "(Eco-Efficiency) proposed: " the achievement of environmental efficiency, the price to be competitive in the provision of goods or services to meet people's needs, improve living Quality at the same time, the goods or services throughout the life cycle. Its impact on the environment and natural resource utilization and gradually reduced to the extent the Earth can load. "whichever is the economy "Economic" and " Ecologic". The research into the methodology to obtain the Taiwan Green Building Material Labeling product as the scope of the study, by investigating and weight analysis to explore green building environmental load (Ln) factor and the Green Building Quality (Qn) factor to Establish green building environmental efficiency assessment model (GBM Eco-Efficiency). And building materials for healthy green label products for priority assessment object, the object is set in the material evidence for the direct response to the environmental load from the floor class-based, explicit feedback correction to the Green Building environmental efficiency assessment model, "efficiency " as a starting point to achieve balance between human "health "and Earth "sustainable development of win-win strategy. The study is expected to reach 1.To establish green building materials and the quality of environmental impact assessment system, 2. To establish value of GBM Eco-Efficiency model, 3. To establish the GBM Eco-Efficiency model for application of green building material feedback mechanisms.
Keywords: Climate Change, Green Building Material (GBM), Eco-Efficiency, Life Cycle Assessment, Performance Evaluation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23313290 Estimating Spatial Disaggregation of Urban Thermal Responsiveness on Summer Diurnal Range with a Numerical Modeling Approach in Bangkok, Thailand
Authors: Manat Srivanit, Hokao Kazunori
Abstract:
Facing the concern of the population to its environment and to climatic change, city planners are now considering the urban climate in their choices of planning. The urban climate, representing different urban morphologies across central Bangkok metropolitan area (BMA), are used to investigates the effects of both the composition and configuration of variables of urban morphology indicators on the summer diurnal range of urban climate, using correlation analyses and multiple linear regressions. Results show first indicate that approximately 92.6% of the variation in the average maximum daytime near-surface air temperature (Ta) was explained jointly by the two composition variables of urban morphology indicators including open space ratio (OSR) and floor area ratio (FAR). It has been possible to determine the membership of sample areas to the local climate zones (LCZs) using these urban morphology descriptors automatically computed with GIS and remote sensed data. Finally result found the temperature differences among zones of large separation, such as the city center could be respectively from 35.48±1.04ºC (Mean±S.D.) warmer than the outskirt of Bangkok on average for maximum daytime near surface temperature to 28.27±0.21ºC for extreme event and, can exceed as 8ºC. A spatially disaggregation of urban thermal responsiveness map would be helpful for several reasons. First, it would localize urban areas concerned by different climate behavior over summer daytime and be a good indicator of urban climate variability. Second, when overlaid with a land cover map, this map may contribute to identify possible urban management strategies to reduce heat wave effects in BMA.
Keywords: Urban climate, Urban morphology, Local climate zone, Urban planning, GIS and remote sensing
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24643289 Conceptualizing Thoughtful Intelligence for Sustainable Decision Making
Authors: Musarrat Jabeen
Abstract:
Thoughtful intelligence offers a sustainable position to enhance the influence of decision-makers. Thoughtful Intelligence implies the understanding to realize the impact of one’s thoughts, words and actions on the survival, dignity and development of the individuals, groups and nations. Thoughtful intelligence has received minimal consideration in the area of Decision Support Systems, with an end goal to evaluate the quantity of knowledge and its viability. This pattern degraded the imbibed contribution of thoughtful intelligence required for sustainable decision making. Given the concern, this paper concentrates on the question: How to present a model of Thoughtful Decision Support System (TDSS)? The aim of this paper is to appreciate the concepts of thoughtful intelligence and insinuate a Decision Support System based on thoughtful intelligence. Thoughtful intelligence includes three dynamic competencies: i) Realization about long term impacts of decisions that are made in a specific time and space, ii) A great sense of taking actions, iii) Intense interconnectivity with people and nature and; seven associate competencies, of Righteousness, Purposefulness, Understanding, Contemplation, Sincerity, Mindfulness, and Nurturing. The study utilizes two methods: Focused group discussion to count prevailing Decision Support Systems; 70% results of focus group discussions found six decision support systems and the positive inexistence of thoughtful intelligence among decision support systems regarding sustainable decision making. Delphi focused on defining thoughtful intelligence to model (TDSS). 65% results helped to conceptualize (definition and description) of thoughtful intelligence. TDSS is offered here as an addition in the decision making literature. The clients are top leaders.
Keywords: Thoughtful intelligence, Sustainable decision making, Thoughtful decision support system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5933288 Improving Similarity Search Using Clustered Data
Authors: Deokho Kim, Wonwoo Lee, Jaewoong Lee, Teresa Ng, Gun-Ill Lee, Jiwon Jeong
Abstract:
This paper presents a method for improving object search accuracy using a deep learning model. A major limitation to provide accurate similarity with deep learning is the requirement of huge amount of data for training pairwise similarity scores (metrics), which is impractical to collect. Thus, similarity scores are usually trained with a relatively small dataset, which comes from a different domain, causing limited accuracy on measuring similarity. For this reason, this paper proposes a deep learning model that can be trained with a significantly small amount of data, a clustered data which of each cluster contains a set of visually similar images. In order to measure similarity distance with the proposed method, visual features of two images are extracted from intermediate layers of a convolutional neural network with various pooling methods, and the network is trained with pairwise similarity scores which is defined zero for images in identical cluster. The proposed method outperforms the state-of-the-art object similarity scoring techniques on evaluation for finding exact items. The proposed method achieves 86.5% of accuracy compared to the accuracy of the state-of-the-art technique, which is 59.9%. That is, an exact item can be found among four retrieved images with an accuracy of 86.5%, and the rest can possibly be similar products more than the accuracy. Therefore, the proposed method can greatly reduce the amount of training data with an order of magnitude as well as providing a reliable similarity metric.
Keywords: Visual search, deep learning, convolutional neural network, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8273287 Suicide Wrongful Death: Standard of Care Problems Involving the Inaccurate Discernment of Lethal Risk When Focusing on the Elicitation of Suicide Ideation
Authors: Bill D. Geis, Frederick Newman
Abstract:
Suicide and wrongful death forensic cases are the fastest rising tort in mental health law. Most suicide-related personal injury claims fall into the legal category of “wrongful death.” Though mental health experts may be called on to address a range of forensic questions in wrongful death cases, the central consultation that most experts provide is about the negligence element—specifically, the issue of whether the clinician met the clinical standard of care in assessing, treating, and managing the deceased person’s mental health care. Standards of care, varying from US state to state, are broad and address what a reasonable clinician might do in a similar circumstance. This fact leaves the issue of the suicide standard of care, in each case, up to forensic experts to put forth a reasoned estimate of what the standard of care should have been in the specific case under litigation. Because the general state guidelines for standard of care are broad, forensic experts are readily retained to provide scientific and clinical opinions about whether or not a clinician met the standard of care in their suicide assessment, treatment, and management of the case. In the past and in much of current practice, the assessment of suicide has centered on the elicitation of verbalized suicide ideation. But suicide ideation, in the matter of suicide risk determination, may be a necessary but insufficient target of lethal suicide risk assessment. Assessment of near-term suicide risk—assessment that goes beyond verbalized suicide ideation and relates to acute crisis variables—is likely needed. Specifically, such other or additional suicide risk variable assessment may be required in the context of lethal suicide risk situations, as opposed to the discernment of general, nonlethal suicide behavior as a standard of practice (whether a patient is having suicidal thoughts or exhibiting an ambivalent suicide attempt potential). In the current study, verbalized suicide ideation information was unhelpful in the assessment of lethal risk. The Lethal Suicide Risk Assessment, Acute Model, and other dynamic, near-term risk models (such as the Acute Suicide Affective Disorder Model and the Suicide Crisis Syndrome Model)—going beyond elicited suicide ideation—need to be incorporated into current clinical suicide assessment training and become the legal standard of care for expected clinical behavior. Without this expanded clinical assessment perspective, the standard of care for suicide assessment is out of sync with current knowledge—an emerging dilemma for the forensic evaluation of suicide wrongful death cases.
Keywords: Forensic evaluation, standard of care, suicide, suicide assessment, wrongful death.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2563286 Evaluating Efficiency of Nina Distribution Company Using Window Data Envelopment Analysis and Malmquist Index
Authors: Hossein Taherian Far, Ali Bazaee
Abstract:
Achieving continuous sustained economic growth and following economic development can be the target for all countries which are looking for it. In this regard, distribution industry plays an important role in growth and development of any nation. So, estimating the efficiency and productivity of the so called industry and identifying factors influencing it, is very necessary. The objective of the present study is to measure the efficiency and productivity of seven branches of Nina Distribution Company using window data envelopment analysis and Malmquist productivity index from spring 2013 to summer 2015. In this study, using criteria of fixed assets, payroll personnel, operating costs and duration of collection of receivables were selected as inputs and people and net sales, gross profit and percentage of coverage to customers were selected as outputs. Then, the process of performance window data envelopment analysis was driven and process efficiency has been measured using Malmquist index. The results indicate that the average technical efficiency of window Data Envelopment Analysis (DEA) model and fluctuating trend is sustainable. But the average management efficiency in window DEA model is related with negative growth (decline) of about 13%. The mean scale efficiency in all windows, except in the second one which is faced with 8%, shows growth of 18% compared to the first window. On the other hand, the mean change in total factor productivity in all branches of the industry shows average negative growth (decrease) of 12% which are the result of a negative change in technology.
Keywords: Nina Distribution Company branches, window data envelopment analysis, Malmquist productivity index.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11623285 Application of Single Tuned Passive Filters in Distribution Networks at the Point of Common Coupling
Authors: M. Almutairi, S. Hadjiloucas
Abstract:
The harmonic distortion of voltage is important in relation to power quality due to the interaction between the large diffusion of non-linear and time-varying single-phase and three-phase loads with power supply systems. However, harmonic distortion levels can be reduced by improving the design of polluting loads or by applying arrangements and adding filters. The application of passive filters is an effective solution that can be used to achieve harmonic mitigation mainly because filters offer high efficiency, simplicity, and are economical. Additionally, possible different frequency response characteristics can work to achieve certain required harmonic filtering targets. With these ideas in mind, the objective of this paper is to determine what size single tuned passive filters work in distribution networks best, in order to economically limit violations caused at a given point of common coupling (PCC). This article suggests that a single tuned passive filter could be employed in typical industrial power systems. Furthermore, constrained optimization can be used to find the optimal sizing of the passive filter in order to reduce both harmonic voltage and harmonic currents in the power system to an acceptable level, and, thus, improve the load power factor. The optimization technique works to minimize voltage total harmonic distortions (VTHD) and current total harmonic distortions (ITHD), where maintaining a given power factor at a specified range is desired. According to the IEEE Standard 519, both indices are viewed as constraints for the optimal passive filter design problem. The performance of this technique will be discussed using numerical examples taken from previous publications.
Keywords: Harmonics, passive filter, power factor, power quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21913284 A Theoretical Analysis of Air Cooling System Using Thermal Ejector under Variable Generator Pressure
Authors: Mohamed Ouzzane, Mahmoud Bady
Abstract:
Due to energy and environment context, research is looking for the use of clean and energy efficient system in cooling industry. In this regard, the ejector represents one of the promising solutions. The thermal ejector is a passive component used for thermal compression in refrigeration and cooling systems, usually activated by heat either waste or solar. The present study introduces a theoretical analysis of the cooling system which uses a gas ejector thermal compression. A theoretical model is developed and applied for the design and simulation of the ejector, as well as the whole cooling system. Besides the conservation equations of mass, energy and momentum, the gas dynamic equations, state equations, isentropic relations as well as some appropriate assumptions are applied to simulate the flow and mixing in the ejector. This model coupled with the equations of the other components (condenser, evaporator, pump, and generator) is used to analyze profiles of pressure and velocity (Mach number), as well as evaluation of the cycle cooling capacity. A FORTRAN program is developed to carry out the investigation. Properties of refrigerant R134a are calculated using real gas equations. Among many parameters, it is thought that the generator pressure is the cornerstone in the cycle, and hence considered as the key parameter in this investigation. Results show that the generator pressure has a great effect on the ejector and on the whole cooling system. At high generator pressures, strong shock waves inside the ejector are created, which lead to significant condenser pressure at the ejector exit. Additionally, at higher generator pressures, the designed system can deliver cooling capacity for high condensing pressure (hot season).
Keywords: Air cooling system, refrigeration, thermal ejector, thermal compression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6013283 Efficient Real-time Remote Data Propagation Mechanism for a Component-Based Approach to Distributed Manufacturing
Authors: V. Barot, S. McLeod, R. Harrison, A. A. West
Abstract:
Manufacturing Industries face a crucial change as products and processes are required to, easily and efficiently, be reconfigurable and reusable. In order to stay competitive and flexible, situations also demand distribution of enterprises globally, which requires implementation of efficient communication strategies. A prototype system called the “Broadcaster" has been developed with an assumption that the control environment description has been engineered using the Component-based system paradigm. This prototype distributes information to a number of globally distributed partners via an adoption of the circular-based data processing mechanism. The work highlighted in this paper includes the implementation of this mechanism in the domain of the manufacturing industry. The proposed solution enables real-time remote propagation of machine information to a number of distributed supply chain client resources such as a HMI, VRML-based 3D views and remote client instances regardless of their distribution nature and/ or their mechanisms. This approach is presented together with a set of evaluation results. Authors- main concentration surrounds the reliability and the performance metric of the adopted approach. Performance evaluation is carried out in terms of the response times taken to process the data in this domain and compared with an alternative data processing implementation such as the linear queue mechanism. Based on the evaluation results obtained, authors justify the benefits achieved from this proposed implementation and highlight any further research work that is to be carried out.
Keywords: Broadcaster, circular buffer, Component-based, distributed manufacturing, remote data propagation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13733282 Clean Sky 2 – Project PALACE: Aeration’s Experimental Sound Velocity Investigations for High-Speed Gerotor Simulations
Authors: Benoît Mary, Thibaut Gras, Gaëtan Fagot, Yvon Goth, Ilyes Mnassri-Cetim
Abstract:
A Gerotor pump is composed of an external and internal gear with conjugate cycloidal profiles. From suction to delivery ports, the fluid is transported inside cavities formed by teeth and driven by the shaft. From a geometric and conceptional side it is worth to note that the internal gear has one tooth less than the external one. Simcenter Amesim v.16 includes a new submodel for modelling the hydraulic Gerotor pumps behavior (THCDGP0). This submodel considers leakages between teeth tips using Poiseuille and Couette flows contributions. From the 3D CAD model of the studied pump, the “CAD import” tool takes out the main geometrical characteristics and the submodel THCDGP0 computes the evolution of each cavity volume and their relative position according to the suction or delivery areas. This module, based on international publications, presents robust results up to 6 000 rpm for pressure greater than atmospheric level. For higher rotational speeds or lower pressures, oil aeration and cavitation effects are significant and highly drop the pump’s performance. The liquid used in hydraulic systems always contains some gas, which is dissolved in the liquid at high pressure and tends to be released in a free form (i.e. undissolved as bubbles) when pressure drops. In addition to gas release and dissolution, the liquid itself may vaporize due to cavitation. To model the relative density of the equivalent fluid, modified Henry’s law is applied in Simcenter Amesim v.16 to predict the fraction of undissolved gas or vapor. Three parietal pressure sensors have been set up upstream from the pump to estimate the sound speed in the oil. Analytical models have been compared with the experimental sound speed to estimate the occluded gas content. Simcenter Amesim v.16 model was supplied by these previous analyses marks which have successfully improved the simulations results up to 14 000 rpm. This work provides a sound foundation for designing the next Gerotor pump generation reaching high rotation range more than 25 000 rpm. This improved module results will be compared to tests on this new pump demonstrator.
Keywords: Gerotor pump, high speed, simulations, aeronautic, aeration, cavitation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5683281 Working Children and Adolescents and the Vicious Circle of Poverty from the Perspective of Gunnar Myrdal’s Theory of Circular Cumulative Causation: Analysis and Implementation of a Probit Model to Brazil
Authors: J. Leige Lopes, L. Aparecida Bastos, R. Monteiro da Silva
Abstract:
The objective of this paper is to study the work of children and adolescents and the vicious circle of poverty from the perspective of Guinar Myrdal’s Theory of Circular Cumulative Causation. The objective is to show that if a person starts working in the juvenile phase of life they will be classified as poor or extremely poor when they are adult, which can to be observed in the case of Brazil, more specifically in the north and northeast. To do this, the methodology used was statistical and econometric analysis by applying a probit model. The main results show that: if people reside in the northeastern region of Brazil, and if they have a low educational level and if they start their professional life before the age 18, they will increase the likelihood that they will be poor or extremely poor. There is a consensus in the literature that one of the causes of the intergenerational transmission of poverty is related to child labor, this because when one starts their professional life while still in the toddler or adolescence stages of life, they end up sacrificing their studies. Because of their low level of education, children or adolescents are forced to perform low-paid functions and abandon school, becoming in the future, people who will be classified as poor or extremely poor. As a result of poverty, parents may be forced to send their children out to work when they are young, so that in the future they will also become poor adults, a process that is characterized as the "vicious circle of poverty."Keywords: Children, adolescents, Gunnar Myrdal, poverty, vicious circle.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16893280 Lessons to Management from the Control Loop Phenomenon
Authors: Raied Salman, Nazar Younis
Abstract:
In a none-super-competitive environment the concepts of closed system, management control remains to be the dominant guiding concept to management. The merits of closed loop have been the sources of most of the management literature and culture for many decades. It is a useful exercise to investigate and poke into the dynamics of the control loop phenomenon and draws some lessons to use for refining the practice of management. This paper examines the multitude of lessons abstracted from the behavior of the Input /output /feedback control loop model, which is the core of control theory. There are numerous lessons that can be learned from the insights this model would provide and how it parallels the management dynamics of the organization. It is assumed that an organization is basically a living system that interacts with the internal and external variables. A viable control loop is the one that reacts to the variation in the environment and provide or exert a corrective action. In managing organizations this is reflected in organizational structure and management control practices. This paper will report findings that were a result of examining several abstract scenarios that are exhibited in the design, operation, and dynamics of the control loop and how they are projected on the functioning of the organization. Valuable lessons are drawn in trying to find parallels and new paradigms, and how the control theory science is reflected in the design of the organizational structure and management practices. The paper is structured in a logical and perceptive format. Further research is needed to extend these findings.Keywords: Management theory, control theory, feed back, input/output, strategy, change, information technology, informationsystems, IS, organizational environment, organizations, opensystems, closed systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14333279 Methane versus Carbon Dioxide: Mitigation Prospects
Authors: Alexander J. Severinsky, Allen L. Sessoms
Abstract:
Atmospheric carbon dioxide (CO2) has dominated the discussion around the causes of climate change. This is a reflection of a 100-year time horizon for all greenhouse gases that became a norm. The 100-year time horizon is much too long – and yet, almost all mitigation efforts, including those set in the near-term frame of within 30 years, are still geared toward it. In this paper, we show that for a 30-year time horizon, methane (CH4) is the greenhouse gas whose radiative forcing exceeds that of CO2. In our analysis, we use the radiative forcing of greenhouse gases in the atmosphere, because they directly affect the rise in temperature on Earth. We found that in 2019, the radiative forcing (RF) of methane was ~2.5 W/m2 and that of carbon dioxide was ~2.1 W/m2. Under a business-as-usual (BAU) scenario until 2050, such forcing would be ~2.8 W/m2 and ~3.1 W/m2 respectively. There is a substantial spread in the data for anthropogenic and natural methane (CH4) emissions, along with natural gas, (which is primarily CH4), leakages from industrial production to consumption. For this reason, we estimate the minimum and maximum effects of a reduction of these leakages, and assume an effective immediate reduction by 80%. Such action may serve to reduce the annual radiative forcing of all CH4 emissions by ~15% to ~30%. This translates into a reduction of RF by 2050 from ~2.8 W/m2 to ~2.5 W/m2 in the case of the minimum effect that can be expected, and to ~2.15 W/m2 in the case of the maximum effort to reduce methane leakages. Under the BAU, we find that the RF of CO2 will increase from ~2.1 W/m2 now to ~3.1 W/m2 by 2050. We assume a linear reduction of 50% in anthropogenic emission over the course of the next 30 years, which would reduce the radiative forcing of CO2 from ~3.1 W/m2 to ~2.9 W/m2. In the case of "net zero," the other 50% of only anthropogenic CO2 emissions reduction would be limited to being either from sources of emissions or directly from the atmosphere. In this instance, the total reduction would be from ~3.1 W/m2 to ~2.7 W/m2, or ~0.4 W/m2. To achieve the same radiative forcing as in the scenario of maximum reduction of methane leakages of ~2.15 W/m2, an additional reduction of radiative forcing of CO2 would be approximately 2.7 -2.15 = 0.55 W/m2. In total, one would need to remove ~660 GT of CO2 from the atmosphere in order to match the maximum reduction of current methane leakages, and ~270 GT of CO2 from emitting sources, to reach "negative emissions". This amounts to over 900 GT of CO2.
Keywords: Methane Leakages, Methane Radiative Forcing, Methane Mitigation, Methane Net Zero.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6463278 Industry Symbiosis and Waste Glass Upgrading: A Feasibility Study in Liverpool towards Circular Economy
Authors: Han-Mei Chen, Rongxin Zhou, Taige Wang
Abstract:
Glass is widely used in everyday life, from glass bottles for beverages, to architectural glass for various forms of glazing. Although the mainstream of used glass is recycled in the UK, the single-use and then recycling procedure results in a lot of waste as it incorporates intact glass with smashing, re-melting and remanufacturing. These processes bring massive energy consumption with a huge loss of high embodied energy and economic value, compared to re-use which’s towards a ‘zero carbon’ target. As a tourism city, Liverpool has more glass bottle consumption than most less leisure focused cities. It is therefore vital for Liverpool to find an upgrading approach for the single-use glass bottles with a low carbon output. This project aims to assess the feasibility of an industrial symbiosis and upgrading framework of glass and to investigate the ways of achieving them. It is significant to Liverpool’s future industry strategy since it provides an opportunity to target on economy recovery for post-COVID by industry symbiosis and an up-grading waste management in Liverpool to respond to the climate emergency. In addition, it will influence the local government policy for glass bottle reuse and recycling in North West England, and as a good practice to be further recommended to other areas of the UK. First, critical literature review of glass waste strategies has been conducted in the UK, and world-wide industrial symbiosis practices. Second, mapping, data collection and analysis have shown the current life cycle chain and the strong links of glass reuse and upgrading potentials via site visits to 16 local waste recycling centres. The results of this research have demonstrated the understanding the influence of key factors on the development of a circular industrial symbiosis business model for beverage glass bottles. The current waste management procedures of glass bottle industry, its business model, supply chain and the material flow have been reviewed. The various potential opportunities for glass bottle up-valuing have been investigated towards an industrial symbiosis in Liverpool. Finally, an up-valuing business model has been developed for an industrial symbiosis framework of glass in Liverpool. For glass bottles, there are two possibilities: 1) focus on upgrading processes towards re-use rather than single-use and recycling, 2) focus on ‘smart’ re-use and recycling leading to optimised values in other sectors to create a wider industry symbiosis for a multi-level and circular economy.
Keywords: Glass bottles, industry symbiosis, smart reuse, waste upgrading.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2413277 Effective Stacking of Deep Neural Models for Automated Object Recognition in Retail Stores
Authors: Ankit Sinha, Soham Banerjee, Pratik Chattopadhyay
Abstract:
Automated product recognition in retail stores is an important real-world application in the domain of Computer Vision and Pattern Recognition. In this paper, we consider the problem of automatically identifying the classes of the products placed on racks in retail stores from an image of the rack and information about the query/product images. We improve upon the existing approaches in terms of effectiveness and memory requirement by developing a two-stage object detection and recognition pipeline comprising of a Faster-RCNN-based object localizer that detects the object regions in the rack image and a ResNet-18-based image encoder that classifies the detected regions into the appropriate classes. Each of the models is fine-tuned using appropriate data sets for better prediction and data augmentation is performed on each query image to prepare an extensive gallery set for fine-tuning the ResNet-18-based product recognition model. This encoder is trained using a triplet loss function following the strategy of online-hard-negative-mining for improved prediction. The proposed models are lightweight and can be connected in an end-to-end manner during deployment to automatically identify each product object placed in a rack image. Extensive experiments using Grozi-32k and GP-180 data sets verify the effectiveness of the proposed model.
Keywords: Retail stores, Faster-RCNN, object localization, ResNet-18, triplet loss, data augmentation, product recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5843276 Assessing the Suitability of South African Waste Foundry Sand as an Additive in Clay Masonry Products
Authors: Nthabiseng Portia Mahumapelo, Andre van Niekerk, Ndabenhle Sosibo, Nirdesh Singh
Abstract:
The foundry industry generates large quantities of solid waste in the form of waste foundry sand. The ever-increasing quantities of this type of industrial waste put pressure on land-filling space and its proper management has become a global concern. The South African foundry industry is not different when it comes to this solid waste generation. Utilizing the foundry waste sand in other applications has become an attractive avenue to deal with this waste stream. In the present paper, an evaluation was done on the suitability of foundry waste sand as an additive in clay masonry products. Purchased clay was added to the foundry waste sand sample in a 50/50 ratio. The mixture was named FC sample. The FC sample was mixed with water in a pan mixer until the mixture was consistent and suitable for extrusion. The FC sample was extruded and cut into briquettes. Water absorption, shrinkage and modulus of rupture tests were conducted on the resultant briquettes. Foundry waste sand and FC samples were respectively characterized mineralogically using X-Ray Diffraction, and the major and trace elements were determined using Inductively Coupled Plasma Optical Emission Spectroscopy. Adding purchased clay to the foundry waste sand positively influenced the workability of the test sample. Another positive characteristic was the low linear shrinkage, which indicated that products manufactured from the FC sample would not be susceptible to cracking. The water absorption values were acceptable and the unfired and fired strength values of the briquette’s samples were acceptable. In conclusion, tests showed that foundry waste sand can be used as an additive in masonry clay bricks, provided it is blended with good quality clay.
Keywords: Foundry waste sand, masonry clay bricks, modulus of rupture, shrinkage.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6603275 A Study of Indentation Energy in Three Points Bending of Sandwich beams with Composite Laminated Faces and Foam Core
Authors: M. Sadighi, H. Pouriayevali, M. Saadati
Abstract:
This paper deals with analysis of flexural stiffness, indentation and their energies in three point loading of sandwich beams with composite faces from Eglass/epoxy and cores from Polyurethane or PVC. Energy is consumed in three stages of indentation in laminated beam, indentation of sandwich beam and bending of sandwich beam. Theory of elasticity is chosen to present equations for indentation of laminated beam, then these equations have been corrected to offer better results. An analytical model has been used assuming an elastic-perfectly plastic compressive behavior of the foam core. Classical theory of beam is used to describe three point bending. Finite element (FE) analysis of static indentation sandwich beams is performed using the FE code ABAQUS. The foam core is modeled using the crushable foam material model and response of the foam core is experimentally characterized in uniaxial compression. Three point bending and indentation have been done experimentally in two cases of low velocity and higher velocity (quasi-impact) of loading. Results can describe response of beam in terms of core and faces thicknesses, core material, indentor diameter, energy absorbed, and length of plastic area in the testing. The experimental results are in good agreement with the analytical and FE analyses. These results can be used as an introduction for impact loading and energy absorbing of sandwich structures.Keywords: Three point Bending, Indentation, Foams, Composite laminated beam, Sandwich beams, Finite element
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25853274 Comparison between Experimental and Numerical Studies of Fully Encased Composite Columns
Authors: Md. Soebur Rahman, Mahbuba Begum, Raquib Ahsan
Abstract:
Composite column is a structural member that uses a combination of structural steel shapes, pipes or tubes with or without reinforcing steel bars and reinforced concrete to provide adequate load carrying capacity to sustain either axial compressive loads alone or a combination of axial loads and bending moments. Composite construction takes the advantages of the speed of construction, light weight and strength of steel, and the higher mass, stiffness, damping properties and economy of reinforced concrete. The most usual types of composite columns are the concrete filled steel tubes and the partially or fully encased steel profiles. Fully encased composite column (FEC) provides compressive strength, stability, stiffness, improved fire proofing and better corrosion protection. This paper reports experimental and numerical investigations of the behaviour of concrete encased steel composite columns subjected to short-term axial load. In this study, eleven short FEC columns with square shaped cross section were constructed and tested to examine the load-deflection behavior. The main variables in the test were considered as concrete compressive strength, cross sectional size and percentage of structural steel. A nonlinear 3-D finite element (FE) model has been developed to analyse the inelastic behaviour of steel, concrete, and longitudinal reinforcement as well as the effect of concrete confinement of the FEC columns. FE models have been validated against the current experimental study conduct in the laboratory and published experimental results under concentric load. It has been observed that FE model is able to predict the experimental behaviour of FEC columns under concentric gravity loads with good accuracy. Good agreement has been achieved between the complete experimental and the numerical load-deflection behaviour in this study. The capacities of each constituent of FEC columns such as structural steel, concrete and rebar's were also determined from the numerical study. Concrete is observed to provide around 57% of the total axial capacity of the column whereas the steel I-sections contributes to the rest of the capacity as well as ductility of the overall system. The nonlinear FE model developed in this study is also used to explore the effect of concrete strength and percentage of structural steel on the behaviour of FEC columns under concentric loads. The axial capacity of FEC columns has been found to increase significantly by increasing the strength of concrete.
Keywords: Composite, columns, experimental, finite element, fully encased, strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28593273 Fluidised Bed Gasification of Multiple Agricultural Biomass Derived Briquettes
Authors: Rukayya Ibrahim Muazu, Aiduan Li Borrion, Julia A. Stegemann
Abstract:
Biomass briquette gasification is regarded as a promising route for efficient briquette use in energy generation, fuels and other useful chemicals. However, previous research has been focused on briquette gasification in fixed bed gasifiers such as updraft and downdraft gasifiers. Fluidised bed gasifier has the potential to be effectively sized to medium or large scale. This study investigated the use of fuel briquettes produced from blends of rice husks and corn cobs biomass, in a bubbling fluidised bed gasifier. The study adopted a combination of numerical equations and Aspen Plus simulation software, to predict the product gas (syngas) composition base on briquette density and biomass composition (blend ratio of rice husks to corn cobs). The Aspen Plus model was based on an experimentally validated model from the literature. The results based on a briquette size 32 mm diameter and relaxed density range of 500 to 650kg/m3, indicated that fluidisation air required in the gasifier increased with increase in briquette density, and the fluidisation air showed to be the controlling factor compared with the actual air required for gasification of the biomass briquettes. The mass flowrate of CO2 in the predicted syngas composition increased with an increase in air flow, in the gasifier, while CO decreased and H2 was almost constant. The ratio of H2 to CO for various blends of rice husks and corn cobs did not significantly change at the designed process air, but a significant difference of 1.0 was observed between 10/90 and 90/10 % blend of rice husks and corn cobs.Keywords: Briquettes, fluidised bed, gasification, Aspen Plus, syngas.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25523272 A Study on Algorithm Fusion for Recognition and Tracking of Moving Robot
Authors: Jungho Choi, Youngwan Cho
Abstract:
This paper presents an algorithm for the recognition and tracking of moving objects, 1/10 scale model car is used to verify performance of the algorithm. Presented algorithm for the recognition and tracking of moving objects in the paper is as follows. SURF algorithm is merged with Lucas-Kanade algorithm. SURF algorithm has strong performance on contrast, size, rotation changes and it recognizes objects but it is slow due to many computational complexities. Processing speed of Lucas-Kanade algorithm is fast but the recognition of objects is impossible. Its optical flow compares the previous and current frames so that can track the movement of a pixel. The fusion algorithm is created in order to solve problems which occurred using the Kalman Filter to estimate the position and the accumulated error compensation algorithm was implemented. Kalman filter is used to create presented algorithm to complement problems that is occurred when fusion two algorithms. Kalman filter is used to estimate next location, compensate for the accumulated error. The resolution of the camera (Vision Sensor) is fixed to be 640x480. To verify the performance of the fusion algorithm, test is compared to SURF algorithm under three situations, driving straight, curve, and recognizing cars behind the obstacles. Situation similar to the actual is possible using a model vehicle. Proposed fusion algorithm showed superior performance and accuracy than the existing object recognition and tracking algorithms. We will improve the performance of the algorithm, so that you can experiment with the images of the actual road environment.Keywords: SURF, Optical Flow Lucas-Kanade, Kalman Filter, object recognition, object tracking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22923271 Sustainability Assessment of a Deconstructed Residential House
Authors: Atiq U. Zaman, Juliet Arnott
Abstract:
This paper analyses the various benefits and barriers of residential deconstruction in the context of environmental performance and circular economy based on a case study project in Christchurch, New Zealand. The case study project “Whole House Deconstruction” which aimed, firstly, to harvest materials from a residential house, secondly, to produce new products using the recovered materials, and thirdly, to organize an exhibition for the local public to promote awareness on resource conservation and sustainable deconstruction practices. Through a systematic deconstruction process, the project recovered around 12 tonnes of various construction materials, most of which would otherwise be disposed of to landfill in the traditional demolition approach. It is estimated that the deconstruction of a similar residential house could potentially prevent around 27,029 kg of carbon emission to the atmosphere by recovering and reusing the building materials. In addition, the project involved local designers to produce 400 artefacts using the recovered materials and to exhibit them to accelerate public awareness. The findings from this study suggest that the deconstruction project has significant environmental benefits, as well as social benefits by involving the local community and unemployed youth as a part of their professional skills development opportunities. However, the project faced a number of economic and institutional challenges. The study concludes that with proper economic models and appropriate institutional support a significant amount of construction and demolition waste can be reduced through a systematic deconstruction process. Traditionally, the greatest benefits from such projects are often ignored and remain unreported to wider audiences as most of the external and environmental costs have not been considered in the traditional linear economy.
Keywords: Circular economy, construction and demolition waste, resource recovery, systematic deconstruction, sustainable waste management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11133270 Applying Case-Based Reasoning in Supporting Strategy Decisions
Authors: S. M. Seyedhosseini, A. Makui, M. Ghadami
Abstract:
Globalization and therefore increasing tight competition among companies, have resulted to increase the importance of making well-timed decision. Devising and employing effective strategies, that are flexible and adaptive to changing market, stand a greater chance of being effective in the long-term. In other side, a clear focus on managing the entire product lifecycle has emerged as critical areas for investment. Therefore, applying wellorganized tools to employ past experience in new case, helps to make proper and managerial decisions. Case based reasoning (CBR) is based on a means of solving a new problem by using or adapting solutions to old problems. In this paper, an adapted CBR model with k-nearest neighbor (K-NN) is employed to provide suggestions for better decision making which are adopted for a given product in the middle of life phase. The set of solutions are weighted by CBR in the principle of group decision making. Wrapper approach of genetic algorithm is employed to generate optimal feature subsets. The dataset of the department store, including various products which are collected among two years, have been used. K-fold approach is used to evaluate the classification accuracy rate. Empirical results are compared with classical case based reasoning algorithm which has no special process for feature selection, CBR-PCA algorithm based on filter approach feature selection, and Artificial Neural Network. The results indicate that the predictive performance of the model, compare with two CBR algorithms, in specific case is more effective.
Keywords: Case based reasoning, Genetic algorithm, Groupdecision making, Product management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21743269 A Hybridized Competency-Based Teacher Candidate Selection System
Authors: R. Ramli, M. I. Ghazali, H. Ibrahim, M. M. Kasim, F. M. Kamal, S.Vikneswari
Abstract:
Teachers form the backbone of any educational system, hence selecting qualified candidates is very crucial. In Malaysia, the decision making in the selection process involves a few stages: Initial filtering through academic achievement, taking entry examination and going through an interview session. The last stage is the most challenging since it highly depends on human judgment. Therefore, this study sought to identify the selection criteria for teacher candidates that form the basis for an efficient multi-criteria teacher-candidate selection model for that last stage. The relevant criteria were determined from the literature and also based on expert input that is those who were involved in interviewing teacher candidates from a public university offering the formal training program. There are three main competency criteria that were identified which are content of knowledge, communication skills and personality. Further, each main criterion was divided into a few subcriteria. The Analytical Hierarchy Process (AHP) technique was employed to allocate weights for the criteria and later, integrated a Simple Weighted Average (SWA) scoring approach to develop the selection model. Subsequently, a web-based Decision Support System was developed to assist in the process of selecting the qualified teacher candidates. The Teacher-Candidate Selection (TeCaS) system is able to assist the panel of interviewers during the selection process which involves a large amount of complex qualitative judgments.
Keywords: Analytic Hierarchy Process, Simple Weighted Average, Decision Support System, Multi-criteria decision making problem.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21873268 The Impact of Online Advertising on Generation Y’s Purchase Decision in Malaysia
Authors: Mui Joo Tang, Eang Teng Chan
Abstract:
Advertising is commonly used to foster sales and reputation of an institution. It is at first the growth of print advertising that has increased the population and number of periodicals of newspaper and its circulation. The rise of Internet and online media has somehow blurred the role of media and advertising though the intention is still to reach out to audience and to increase sales. The relationship between advertising and audience on a product purchase through persuasion has been developing from print media to online media. From the changing media environment and audience, it is the concern of this research to study the impact of online advertising to such a relationship cycle. The content of online advertisements is much of text, multimedia, photo, audio and video. The messages of such content format may indeed bring impacts to its audience and its credibility. This study is therefore reflecting the effectiveness of online advertisement and its influences on generation Y in their purchasing behavior. This study uses Media Dependency Theory to analyze the relationship between the impact of online advertisement and media usage pattern of generation Y. Hierarchy of Effectiveness Model is used as a marketing communication model to study the effectiveness of advertising and further to determine the impact of online advertisement on generation Y in their purchasing decision making. This research uses online survey to reach out the sample of generation Y. The results have shown that online advertisements do not affect much on purchase decision making even though generation Y relies much on the media content including online advertisement for its information and believing in its credibility. There are few other external factors that may interrupt the effectiveness of online advertising. The very obvious influence of purchasing behavior is actually derived from the peers.
Keywords: Generation Y, online advertising, online media, persuasion, print media, purchase decision.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 59633267 Synthesis of Highly Sensitive Molecular Imprinted Sensor for Selective Determination of Doxycycline in Honey Samples
Authors: Nadia El Alami El Hassani, Soukaina Motia, Benachir Bouchikhi, Nezha El Bari
Abstract:
Doxycycline (DXy) is a cycline antibiotic, most frequently prescribed to treat bacterial infections in veterinary medicine. However, its broad antimicrobial activity and low cost, lead to an intensive use, which can seriously affect human health. Therefore, its spread in the food products has to be monitored. The scope of this work was to synthetize a sensitive and very selective molecularly imprinted polymer (MIP) for DXy detection in honey samples. Firstly, the synthesis of this biosensor was performed by casting a layer of carboxylate polyvinyl chloride (PVC-COOH) on the working surface of a gold screen-printed electrode (Au-SPE) in order to bind covalently the analyte under mild conditions. Secondly, DXy as a template molecule was bounded to the activated carboxylic groups, and the formation of MIP was performed by a biocompatible polymer by the mean of polyacrylamide matrix. Then, DXy was detected by measurements of differential pulse voltammetry (DPV). A non-imprinted polymer (NIP) prepared in the same conditions and without the use of template molecule was also performed. We have noticed that the elaborated biosensor exhibits a high sensitivity and a linear behavior between the regenerated current and the logarithmic concentrations of DXy from 0.1 pg.mL−1 to 1000 pg.mL−1. This technic was successfully applied to determine DXy residues in honey samples with a limit of detection (LOD) of 0.1 pg.mL−1 and an excellent selectivity when compared to the results of oxytetracycline (OXy) as analogous interfering compound. The proposed method is cheap, sensitive, selective, simple, and is applied successfully to detect DXy in honey with the recoveries of 87% and 95%. Considering these advantages, this system provides a further perspective for food quality control in industrial fields.Keywords: Electrochemical sensor, molecular imprinted polymer, doxycycline, food control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11743266 Optimal and Critical Path Analysis of State Transportation Network Using Neo4J
Authors: Pallavi Bhogaram, Xiaolong Wu, Min He, Onyedikachi Okenwa
Abstract:
A transportation network is a realization of a spatial network, describing a structure which permits either vehicular movement or flow of some commodity. Examples include road networks, railways, air routes, pipelines, and many more. The transportation network plays a vital role in maintaining the vigor of the nation’s economy. Hence, ensuring the network stays resilient all the time, especially in the face of challenges such as heavy traffic loads and large scale natural disasters, is of utmost importance. In this paper, we used the Neo4j application to develop the graph. Neo4j is the world's leading open-source, NoSQL, a native graph database that implements an ACID-compliant transactional backend to applications. The Southern California network model is developed using the Neo4j application and obtained the most critical and optimal nodes and paths in the network using centrality algorithms. The edge betweenness centrality algorithm calculates the critical or optimal paths using Yen's k-shortest paths algorithm, and the node betweenness centrality algorithm calculates the amount of influence a node has over the network. The preliminary study results confirm that the Neo4j application can be a suitable tool to study the important nodes and the critical paths for the major congested metropolitan area.
Keywords: Transportation network, critical path, connectivity reliability, network model, Neo4J application, optimal path, critical path, edge betweenness centrality index, node betweenness centrality index, Yen’s k-shortest paths.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8533265 Forming Limit Analysis of DP600-800 Steels
Authors: M. C. Cardoso, L. P. Moreira
Abstract:
In this work, the plastic behaviour of cold-rolled zinc coated dual-phase steel sheets DP600 and DP800 grades is firstly investigated with the help of uniaxial, hydraulic bulge and Forming Limit Curve (FLC) tests. The uniaxial tensile tests were performed in three angular orientations with respect to the rolling direction to evaluate the strain-hardening and plastic anisotropy. True stressstrain curves at large strains were determined from hydraulic bulge testing and fitted to a work-hardening equation. The limit strains are defined at both localized necking and fracture conditions according to Nakajima’s hemispherical punch procedure. Also, an elasto-plastic localization model is proposed in order to predict strain and stress based forming limit curves. The investigated dual-phase sheets showed a good formability in the biaxial stretching and drawing FLC regions. For both DP600 and DP800 sheets, the corresponding numerical predictions overestimated and underestimated the experimental limit strains in the biaxial stretching and drawing FLC regions, respectively. This can be attributed to the restricted failure necking condition adopted in the numerical model, which is not suitable to describe the tensile and shear fracture mechanisms in advanced high strength steels under equibiaxial and biaxial stretching conditions.Keywords: Advanced high strength steels, forming limit curve, numerical modeling, sheet metal forming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3456