Search results for: player metric optimization and analytics
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3978

Search results for: player metric optimization and analytics

1908 Towards the Need of Resilient Design and Its Assessment in South China

Authors: Alan Lai, Wilson Yik

Abstract:

With rapid urbanization, there has been a dramatic increase in global urban population in Asia and over half of population in Asia will live in urban regions in the near future. Facing with increasing exposure to climate-related stresses and shocks, most of the Asian cities will very likely to experience more frequent heat waves and flooding with rising sea levels, particularly the coastal cities will grapple for intense typhoons and storm surges. These climate changes have severe impacts in urban areas at the costs of infrastructure and population, for example, human health, wellbeing and high risks of dengue fever, malaria and diarrheal disease. With the increasing prominence of adaptation to climate changes, there have been changes in corresponding policies. Smaller cities have greater potentials for integrating the concept of resilience into their infrastructure as well as keeping pace with their rapid growths in population. It is therefore important to explore the potentials of Asian cities adapting to climate change and the opportunities of building climate resilience in urban planning and building design. Furthermore, previous studies have mainly attempted at exploiting the potential of resilience on a macro-level within urban planning rather than that on micro-level within the context of individual building. The resilience of individual building as a research field has not yet been much explored. Nonetheless, recent studies define that the resilience of an individual building is the one which is able to respond to physical damage and recover from such damage in a quickly and cost-effectively manner, while maintain its primary functions. There is also a need to develop an assessment tool to evaluate the resilience on building scale which is still largely uninvestigated as it should be regarded as a basic function of a building. Due to the lack of literature reporting metric for assessing building resilience with sustainability, the research will be designed as a case study to provide insight into the issue. The aim of this research project is to encourage and assist in developing neighborhood climate resilience design strategies for Hong Kong so as to bridge the gap between difference scales and that between theory and practice.

Keywords: resilience cities, building resilience, resilient buildings and infrastructure, climate resilience, hot and humid southeast area, high-density cities

Procedia PDF Downloads 159
1907 Optimizing Residential Housing Renovation Strategies at Territorial Scale: A Data Driven Approach and Insights from the French Context

Authors: Rit M., Girard R., Villot J., Thorel M.

Abstract:

In a scenario of extensive residential housing renovation, stakeholders need models that support decision-making through a deep understanding of the existing building stock and accurate energy demand simulations. To address this need, we have modified an optimization model using open data that enables the study of renovation strategies at both territorial and national scales. This approach provides (1) a definition of a strategy to simplify decision trees from theoretical combinations, (2) input to decision makers on real-world renovation constraints, (3) more reliable identification of energy-saving measures (changes in technology or behaviour), and (4) discrepancies between currently planned and actually achieved strategies. The main contribution of the studies described in this document is the geographic scale: all residential buildings in the areas of interest were modeled and simulated using national data (geometries and attributes). These buildings were then renovated, when necessary, in accordance with the environmental objectives, taking into account the constraints applicable to each territory (number of renovations per year) or at the national level (renovation of thermal deficiencies (Energy Performance Certificates F&G)). This differs from traditional approaches that focus only on a few buildings or archetypes. This model can also be used to analyze the evolution of a building stock as a whole, as it can take into account both the construction of new buildings and their demolition or sale. Using specific case studies of French territories, this paper highlights a significant discrepancy between the strategies currently advocated by decision-makers and those proposed by our optimization model. This discrepancy is particularly evident in critical metrics such as the relationship between the number of renovations per year and achievable climate targets or the financial support currently available to households and the remaining costs. In addition, users are free to seek optimizations for their building stock across a range of different metrics (e.g., financial, energy, environmental, or life cycle analysis). These results are a clear call to re-evaluate existing renovation strategies and take a more nuanced and customized approach. As the climate crisis moves inexorably forward, harnessing the potential of advanced technologies and data-driven methodologies is imperative.

Keywords: residential housing renovation, MILP, energy demand simulations, data-driven methodology

Procedia PDF Downloads 63
1906 Optimization of Gold Mining Parameters by Cyanidation

Authors: Della Saddam Housseyn

Abstract:

Gold, the quintessential noble metal, is one of the most popular metals today, given its ever-increasing cost in the international market. The Amesmessa gold deposit is one of the gold-producing deposits. The first step in our job is to analyze the ore (considered rich ore). Mineralogical and chemical analysis has shown that the general constitution of the ore is quartz in addition to other phases such as Al2O3, Fe2O3, CaO, dolomite. The second step consists of all the leaching tests carried out in rolling bottles. These tests were carried out on 14 samples to determine the maximum recovery rate and the optimum consumption of reagent (NaCN and CaO). Tests carried out on a pulp density at 50% solid, 500 ppm cyanide concentration and particle size less than 0.6 mm at alkaline pH gave a recovery rate of 94.37%.

Keywords: cyanide, DRX, FX, gold, leaching, rate of recovery, SAA

Procedia PDF Downloads 175
1905 Maintenance Optimization for a Multi-Component System Using Factored Partially Observable Markov Decision Processes

Authors: Ipek Kivanc, Demet Ozgur-Unluakin

Abstract:

Over the past years, technological innovations and advancements have played an important role in the industrial world. Due to technological improvements, the degree of complexity of the systems has increased. Hence, all systems are getting more uncertain that emerges from increased complexity, resulting in more cost. It is challenging to cope with this situation. So, implementing efficient planning of maintenance activities in such systems are getting more essential. Partially Observable Markov Decision Processes (POMDPs) are powerful tools for stochastic sequential decision problems under uncertainty. Although maintenance optimization in a dynamic environment can be modeled as such a sequential decision problem, POMDPs are not widely used for tackling maintenance problems. However, they can be well-suited frameworks for obtaining optimal maintenance policies. In the classical representation of the POMDP framework, the system is denoted by a single node which has multiple states. The main drawback of this classical approach is that the state space grows exponentially with the number of state variables. On the other side, factored representation of POMDPs enables to simplify the complexity of the states by taking advantage of the factored structure already available in the nature of the problem. The main idea of factored POMDPs is that they can be compactly modeled through dynamic Bayesian networks (DBNs), which are graphical representations for stochastic processes, by exploiting the structure of this representation. This study aims to demonstrate how maintenance planning of dynamic systems can be modeled with factored POMDPs. An empirical maintenance planning problem of a dynamic system consisting of four partially observable components deteriorating in time is designed. To solve the empirical model, we resort to Symbolic Perseus solver which is one of the state-of-the-art factored POMDP solvers enabling approximate solutions. We generate some more predefined policies based on corrective or proactive maintenance strategies. We execute the policies on the empirical problem for many replications and compare their performances under various scenarios. The results show that the computed policies from the POMDP model are superior to the others. Acknowledgment: This work is supported by the Scientific and Technological Research Council of Turkey (TÜBİTAK) under grant no: 117M587.

Keywords: factored representation, maintenance, multi-component system, partially observable Markov decision processes

Procedia PDF Downloads 130
1904 The Composting Process from a Waste Management Method to a Remediation Procedure

Authors: G. Petruzzelli, F. Pedron, M. Grifoni, F. Gorini, I. Rosellini, B. Pezzarossa

Abstract:

Composting is a controlled technology to enhance the natural aerobic process of organic wastes degradation. The resulting product is a humified material that is principally recyclable for agricultural purpose. The composting process is one of the most important tools for waste management, by the European Community legislation. In recent years composting has been increasingly used as a remediation technology to remove biodegradable contaminants from soil, and to modulate heavy metals bioavailability in phytoremediation strategies. An optimization in the recovery of resources from wastes through composting could enhance soil fertility and promote its use in the remediation biotechnologies of contaminated soils.

Keywords: agriculture, biopile, compost, soil clean-up, waste recycling

Procedia PDF Downloads 307
1903 Associations among Fetuin A, Cortisol and Thyroid Hormones in Children with Morbid Obesity and Metabolic Syndrome

Authors: Mustafa Metin Donma, Orkide Donma

Abstract:

Obesity is a disease with an ever-increasing prevalence throughout the world. The metabolic network associated with obesity is very complicated. In metabolic syndrome (MetS), it becomes even more difficult to understand. Within this context, hormones, cytokines, and many others participate in this complex matrix. The collaboration among all of these parameters is a matter of great wonder. Cortisol, as a stress hormone, is closely associated with obesity. Thyroid hormones are involved in the regulation of energy as well as glucose metabolism with all of its associates. Fetuin A is known for years; however, the involvement of this parameter in obesity discussions is rather new. Recently, it has been defined as one of the new generation markers of obesity. In this study, the aim was to introduce complex interactions among all to be able to make clear comparisons, at least for a part of this complicated matter. Morbid obese (MO) children participated in the study. Two groups with 46 MO children and 43 with MetS were constituted. All children included in the study were above 99th age- and sex-adjusted body mass index (BMI) percentiles according to World Health Organization criteria. Forty-three morbid obese children in the second group had also MetS components. Informed consent forms were filled by the parents of the participants. The institutional ethics committee has given approval for the study protocol. Data as well as the findings of the study were evaluated from a statistical point of view. Two groups were matched for their age and gender compositions. Significantly higher body mass index (BMI), waist circumference, thyrotropin, and insulin values were observed in the MetS group. Triiodothyronine concentrations did not differ between the groups. Elevated levels for thyroxin, cortisol, and fetuin-A were detected in the MetS group compared to the first group (p > 0.05). In MO MetS- group, cortisol was correlated with thyroxin and fetuin-A (p < 0.05). In the MO MetS+ group, none of these correlations were present. Instead, a correlation between cortisol and thyrotropin was found (p < 0.05). In conclusion, findings have shown that cortisol was the key player in severely obese children. The association of this hormone with the participants of thyroid hormone metabolism was quite important. The lack of association with fetuin A in the morbid obese MetS+ group has suggested the possible interference of MetS components in the behavior of this new generation obesity marker. The most remarkable finding of the study was the unique correlation between cortisol and thyrotropin in the morbid obese MetS+ group, suggesting that thyrotropin may serve as a target along with cortisol in the morbid obese MetS+ group. This association may deserve specific attention during the development of remedies against MetS in the pediatric population.

Keywords: children, cortisol, fetuin A, morbid obesity, thyrotropin

Procedia PDF Downloads 172
1902 Logistical Optimization of Nuclear Waste Flows during Decommissioning

Authors: G. Dottavio, M. F. Andrade, F. Renard, V. Cheutet, A.-L. Ladier, S. Vercraene, P. Hoang, S. Briet, R. Dachicourt, Y. Baizet

Abstract:

An important number of technological equipment and high-skilled workers over long periods of time have to be mobilized during nuclear decommissioning processes. The related operations generate complex flows of waste and high inventory levels, associated to information flows of heterogeneous types. Taking into account that more than 10 decommissioning operations are on-going in France and about 50 are expected toward 2025: A big challenge is addressed today. The management of decommissioning and dismantling of nuclear installations represents an important part of the nuclear-based energy lifecycle, since it has an environmental impact as well as an important influence on the electricity cost and therefore the price for end-users. Bringing new technologies and new solutions into decommissioning methodologies is thus mandatory to improve the quality, cost and delay efficiency of these operations. The purpose of our project is to improve decommissioning management efficiency by developing a decision-support framework dedicated to plan nuclear facility decommissioning operations and to optimize waste evacuation by means of a logistic approach. The target is to create an easy-to-handle tool capable of i) predicting waste flows and proposing the best decommissioning logistics scenario and ii) managing information during all the steps of the process and following the progress: planning, resources, delays, authorizations, saturation zones, waste volume, etc. In this article we present our results from waste nuclear flows simulation during decommissioning process, including discrete-event simulation supported by FLEXSIM 3-D software. This approach was successfully tested and our works confirms its ability to improve this type of industrial process by identifying the critical points of the chain and optimizing it by identifying improvement actions. This type of simulation, executed before the start of the process operations on the basis of a first conception, allow ‘what-if’ process evaluation and help to ensure quality of the process in an uncertain context. The simulation of nuclear waste flows before evacuation from the site will help reducing the cost and duration of the decommissioning process by optimizing the planning and the use of resources, transitional storage and expensive radioactive waste containers. Additional benefits are expected for the governance system of the waste evacuation since it will enable a shared responsibility of the waste flows.

Keywords: nuclear decommissioning, logistical optimization, decision-support framework, waste management

Procedia PDF Downloads 315
1901 Design and Implementation of an Image Based System to Enhance the Security of ATM

Authors: Seyed Nima Tayarani Bathaie

Abstract:

In this paper, an image-receiving system was designed and implemented through optimization of object detection algorithms using Haar features. This optimized algorithm served as face and eye detection separately. Then, cascading them led to a clear image of the user. Utilization of this feature brought about higher security by preventing fraud. This attribute results from the fact that services will be given to the user on condition that a clear image of his face has already been captured which would exclude the inappropriate person. In order to expedite processing and eliminating unnecessary ones, the input image was compressed, a motion detection function was included in the program, and detection window size was confined.

Keywords: face detection algorithm, Haar features, security of ATM

Procedia PDF Downloads 410
1900 Low Power, Highly Linear, Wideband LNA in Wireless SOC

Authors: Amir Mahdavi

Abstract:

In this paper a highly linear CMOS low noise amplifier (LNA) for ultra-wideband (UWB) applications is proposed. The proposed LNA uses a linearization technique to improve second and third-order intercept points (IIP3). The linearity is cured by repealing the common-mode section of all intermodulation components from the cascade topology current with optimization of biasing current use symmetrical and asymmetrical circuits for biasing. Simulation results show that maximum gain and noise figure are 6.9dB and 3.03-4.1dB over a 3.1–10.6 GHz, respectively. Power consumption of the LNA core and IIP3 are 2.64 mW and +4.9dBm respectively. The wideband input impedance matching of LNA is obtained by employing a degenerating inductor (|S11|<-9.1 dB). The circuit proposed UWB LNA is implemented using 0.18 μm based CMOS technology.

Keywords: highly linear LNA, low-power LNA, optimal bias techniques

Procedia PDF Downloads 275
1899 Summer STEM Institute in Environmental Science and Data Sciencefor Middle and High School Students at Pace University

Authors: Lauren B. Birney

Abstract:

Summer STEM Institute for Middle and High School Students at Pace University The STEM Collaboratory NYC® Summer Fellows Institute takes place on Pace University’s New York City campus during July and provides the following key features for all participants: (i) individual meetings with Pace faculty to discuss and refine future educational goals; (ii) mentorship, guidance, and new friendships with program leaders; and (iii) guest lectures from professionals in STEM disciplines and businesses. The Summer STEM Institute allows middle school and high school students to work in teams to conceptualize, develop, and build native mobile applications that teach and reinforce skills in the sciences and mathematics. These workshops enhance students’STEM problem solving techniques and teach advanced methods of computer science and engineering. Topics include: big data and analytics at the Big Data lab at Seidenberg, Data Science focused on social and environmental advancement and betterment; Natural Disasters and their Societal Influences; Algal Blooms and Environmental Impacts; Green CitiesNYC; STEM jobs and growth opportunities for the future; renew able energy and sustainable infrastructure; and climate and the economy. In order to better align the existing Summer STEM, Institute with the CCERS model and expand the overall network, Pace is actively recruiting new content area specialists from STEM industries and private sector enterprises to participate in an enhanced summer institute in order to1) nurture student progress and connect summer learning to school year curriculum, 2) increase peer-to-peer collaboration amongst STEM professionals and private sector technologists, and 3) develop long term funding and sponsorship opportunities for corporate sector partners to support CCERS schools and programs directly.

Keywords: environmental restoration science, citizen science, data science, STEM

Procedia PDF Downloads 82
1898 Well Inventory Data Entry: Utilization of Developed Technologies to Progress the Integrated Asset Plan

Authors: Danah Al-Selahi, Sulaiman Al-Ghunaim, Bashayer Sadiq, Fatma Al-Otaibi, Ali Ameen

Abstract:

In light of recent changes affecting the Oil & Gas Industry, optimization measures have become imperative for all companies globally, including Kuwait Oil Company (KOC). To keep abreast of the dynamic market, a detailed Integrated Asset Plan (IAP) was developed to drive optimization across the organization, which was facilitated through the in-house developed software “Well Inventory Data Entry” (WIDE). This comprehensive and integrated approach enabled centralization of all planned asset components for better well planning, enhancement of performance, and to facilitate continuous improvement through performance tracking and midterm forecasting. Traditionally, this was hard to achieve as, in the past, various legacy methods were used. This paper briefly describes the methods successfully adopted to meet the company’s objective. IAPs were initially designed using computerized spreadsheets. However, as data captured became more complex and the number of stakeholders requiring and updating this information grew, the need to automate the conventional spreadsheets became apparent. WIDE, existing in other aspects of the company (namely, the Workover Optimization project), was utilized to meet the dynamic requirements of the IAP cycle. With the growth of extensive features to enhance the planning process, the tool evolved into a centralized data-hub for all asset-groups and technical support functions to analyze and infer from, leading WIDE to become the reference two-year operational plan for the entire company. To achieve WIDE’s goal of operational efficiency, asset-groups continuously add their parameters in a series of predefined workflows that enable the creation of a structured process which allows risk factors to be flagged and helps mitigation of the same. This tool dictates assigned responsibilities for all stakeholders in a method that enables continuous updates for daily performance measures and operational use. The reliable availability of WIDE, combined with its user-friendliness and easy accessibility, created a platform of cross-functionality amongst all asset-groups and technical support groups to update contents of their respective planning parameters. The home-grown entity was implemented across the entire company and tailored to feed in internal processes of several stakeholders across the company. Furthermore, the implementation of change management and root cause analysis techniques captured the dysfunctionality of previous plans, which in turn resulted in the improvement of already existing mechanisms of planning within the IAP. The detailed elucidation of the 2 year plan flagged any upcoming risks and shortfalls foreseen in the plan. All results were translated into a series of developments that propelled the tool’s capabilities beyond planning and into operations (such as Asset Production Forecasts, setting KPIs, and estimating operational needs). This process exemplifies the ability and reach of applying advanced development techniques to seamlessly integrated the planning parameters of various assets and technical support groups. These techniques enables the enhancement of integrating planning data workflows that ultimately lay the founding plans towards an epoch of accuracy and reliability. As such, benchmarks of establishing a set of standard goals are created to ensure the constant improvement of the efficiency of the entire planning and operational structure.

Keywords: automation, integration, value, communication

Procedia PDF Downloads 139
1897 Optimization of Scheduling through Altering Layout Using Pro-Model

Authors: Zouhair Issa Ahmed, Ahmed Abdulrasool Ahmed, Falah Hassan Abdulsada

Abstract:

This paper presents a layout of a factory using Pro-Model simulation by choosing the best layout that gives the highest productivity and least work in process. The general problem is to find the best sequence in which jobs pass between the machines which are compatible with the technological constraints and optimal with respect to some performance criteria. The best simulation with Pro-Model program increased productivity and reduced work in process by balancing lines of production compared with the current layout of factory when productivity increased from 45 products to 180 products through 720 hours.

Keywords: scheduling, Pro-Model, simulation, balancing lines of production, layout planning, WIP

Procedia PDF Downloads 626
1896 A Compressor Map Optimizing Tool for Prediction of Compressor Off-Design Performance

Authors: Zhongzhi Hu, Jie Shen, Jiqiang Wang

Abstract:

A high precision aeroengine model is needed when developing the engine control system. Compared with other main components, the axial compressor is the most challenging component to simulate. In this paper, a compressor map optimizing tool based on the introduction of a modifiable β function is developed for FWorks (FADEC Works). Three parameters (d density, f fitting coefficient, k₀ slope of the line β=0) are introduced to the β function to make it modifiable. The comparison of the traditional β function and the modifiable β function is carried out for a certain type of compressor. The interpolation errors show that both methods meet the modeling requirements, while the modifiable β function can predict compressor performance more accurately for some areas of the compressor map where the users are interested in.

Keywords: beta function, compressor map, interpolation error, map optimization tool

Procedia PDF Downloads 260
1895 The Risks of 'Techtopia': Reviewing the Negative Lessons of Smart City Development

Authors: Amanda Grace Ahl, Matthew Brummer

Abstract:

‘Smart cities’ are not always as ‘smart’ as the term suggests, which is not often covered in the associated academic and public policy literatures. In what has become known as the smart city approach to urban planning, governments around the world are seeking to harness the power of information and communications technology with increasingly advanced data analytics to address major social, economic, and environmental issues reshaping the ways people live. The definitional and theoretical boundaries of the smart city framework are broad and at times ambiguous, as is empirical treatment of the topic. However, and for all the disparity, in investigating any number of institutional and policy prescriptions to the challenges faced by current and emerging metropoles, scholarly thought has hinged overwhelmingly on value-positive conceptions of informatics-centered design. From enhanced quality of services, to increased efficiency of resources, to improved communication between societal stakeholders, the smart city design is championed as a technological wellspring capable of providing answers to the systemic issues stymying a utopian image of the city. However, it is argued that this ‘techtopia’, has resulted in myopia within the discipline as to value-negative implications of such planning, such as weaknesses in practicality, scalability, social equity and affordability of solutions. In order to more carefully examine this observation - that ‘stupid’ represents an omitted variable bias in the study of ‘smart’ - this paper reviews critical cases of unsuccessful smart city developments. It is argued that also understanding the negative factors affiliated with the development processes is imperative for the advancement of theoretical foundations, policies, and strategies to further the smart city as an equitable, holistic urban innovation. What emerges from the process-tracing carried out in this study are distinctly negative lessons of smart city projects, the significance of which are vital for understanding how best to conceive smart urban planning in the 21st century.

Keywords: case study, city management, innovation system, negative lessons, smart city development

Procedia PDF Downloads 407
1894 The Finance of Happiness: Thinking Finance from the Science of Happiness Perspective

Authors: Renaud Gaucher

Abstract:

Research on happiness has developed significantly in the past fifty years and economics and the political science are starting to be influenced by advances in the field. Until recently, finance has stayed outside this movement. The goal of our research is to integrate finance into this movement conceptually. We explain the why, the what and the how of the finance of happiness. We then study the relationship between corporate finance and happiness. We discuss the optimization of the relationship between the financial performance of a firm and the happiness at work of its employees, and the reduction of financial risk by developing goods that foster the happiness of their users. Finally we look at the development of happiness investment funds, that is investment funds founded on happiness research, and the best ways to share risks and earnings to build a happier society.

Keywords: finance, happiness, investment fund, risk

Procedia PDF Downloads 185
1893 Preliminary Experience in Multiple Green Health Hospital Construction

Authors: Ming-Jyh Chen, Wen-Ming Huang, Yi-Chu Liu, Li-Hui Yang

Abstract:

Introduction: Social responsibility is the key to sustainable organizational development. Under the ground Green Health Hospital Declaration signed by our superintendent, we have launched comprehensive energy conservation management in medical services, the community, and the staff’s life. To execute environment-friendly promotion with robust strategies, we build up a low-carbon medical system and community with smart green public construction promotion as well as intensifying energy conservation education and communication. Purpose/Methods: With the support of the board and the superintendent, we construct an energy management team, commencing with an environment-friendly system, management, education, and ISO 50001 energy management system; we have ameliorated energy performance and energy efficiency and continuing. Results: In the year 2021, we have achieved multiple goals. The energy management system efficiently controls diesel, natural gas, and electricity usage. About 5% of the consumption is saved when compared to the numbers from 2018 and 2021. Our company develops intelligent services and promotes various paperless electronic operations to provide people with a vibrant and environmentally friendly lifestyle. The goal is to save 68.6% on printing and photocopying by reducing 35.15 million sheets of paper yearly. We strengthen the concept of environmental protection classification among colleagues. In the past two years, the amount of resource recycling has reached more than 650 tons, and the resource recycling rate has reached 70%. The annual growth rate of waste recycling is about 28 metric tons. Conclusions: To build a green medical system with “high efficacy, high value, low carbon, low reliance,” energy stewardship, economic prosperity, and social responsibility are our principles when it comes to formulation of energy conservation management strategies, converting limited sources to efficient usage, developing clean energy, and continuing with sustainable energy.

Keywords: energy efficiency, environmental education, green hospital, sustainable development

Procedia PDF Downloads 75
1892 Designing a Low Power Consumption Mote in Wireless Sensor Network

Authors: Saidi Nabiha, Khaled Zaatouri, Walid Fajraoui, Tahar Ezzeddine

Abstract:

The market of Wireless Sensor Network WSN has a great potential and development opportunities. Researchers are focusing on optimization in many fields like efficient deployment and routing protocols. In this article, we will concentrate on energy efficiency for WSN because WSN nodes are habitually deployed in severe No Man’s Land with batteries are not rechargeable, so reducing energy consumption represents an important challenge to extend the life of the network. We will present the design of new WSN mote based on ultra low power STM32L microcontrollers and the ZIGBEE transceiver CC2520. We will compare it to existent motes and we will conclude that our mote is promising in energy consumption.

Keywords: component, WSN mote, power consumption, STM32L, sensors, CC2520

Procedia PDF Downloads 571
1891 Design Optimization of the Primary Containment Building of a Pressurized Water Reactor

Authors: M. Hossain, A. H. Khan, M. A. R. Sarkar

Abstract:

Primary containment structure is one of the five safety layers of a nuclear facility which is needed to be designed in such a manner that it can withstand the pressure and excessive radioactivity during accidental situations. It is also necessary to ensure minimization of cost with maximum possible safety in order to make the design economically feasible and attractive. This paper attempts to identify the optimum design conditions for primary containment structure considering both mechanical and radiation safety keeping the economic aspects in mind. This work takes advantage of commercial simulation software to identify the suitable conditions without the requirement of costly experiments. Generated data may be helpful for further studies.

Keywords: PWR, concrete containment, finite element approach, neutron attenuation, Von Mises stress

Procedia PDF Downloads 179
1890 About the Case Portfolio Management Algorithms and Their Applications

Authors: M. Chumburidze, N. Salia, T. Namchevadze

Abstract:

This work deal with case processing problems in business. The task of strategic credit requirements management of cases portfolio is discussed. The information model of credit requirements in a binary tree diagram is considered. The algorithms to solve issues of prioritizing clusters of cases in business have been investigated. An implementation of priority queues to support case management operations has been presented. The corresponding pseudo codes for the programming application have been constructed. The tools applied in this development are based on binary tree ordering algorithms, optimization theory, and business management methods.

Keywords: credit network, case portfolio, binary tree, priority queue, stack

Procedia PDF Downloads 137
1889 Development of Technologies for Biotransformation of Aquatic Biological Resources for the Production of Functional, Specialized, Therapeutic, Preventive, and Microbiological Products

Authors: Kira Rysakova, Vitaly Novikov

Abstract:

An improved method of obtaining enzymatic collagen hydrolysate from the tissues of marine hydrobionts is proposed, which allows to obtain hydrolysate without pre-isolation of pure collagen. The method can be used to isolate enzymatic collagen hydrolysate from the waste of industrial processing of Red King crab and non-traditional objects - marine holothurias. Comparative analysis of collagen hydrolysates has shown the possibility of their use in a number of nutrient media, but this requires additional optimization of their composition and biological tests on wide sets of test strains of microorganisms.

Keywords: collagen hydrolysate, marine hydrobionts, red king crab, marine holothurias, enzymes, exclusive HPLC

Procedia PDF Downloads 161
1888 Optimization of Transmission Loss on a Series-Coupled Muffler by Taguchi Method

Authors: Jing-Fung Lin, Jer-Jia Sheu

Abstract:

In this study, an approach has been developed for the noise reduction of a muffler. The transmission loss (TL) in the muffler is maximized by the use of a double-chamber muffler, and a baffle with a hole is inserted between chambers. Taguchi method is used to optimize the design for the acoustical performance of the muffler. The TL performance is evaluated by COMSOL software. The excellent parameter combination for the maximum TL is attained as high as 35.30 dB in a wide frequency range from 10 Hz to 1400 Hz. The influence sequence of four parameters on TL is determined by the range analysis. The effects of length and expansion ratio of the first chamber on TL performance for the excellent program were discussed. Comparisons of the TL results from different designs are made.

Keywords: acoustics, baffle, chamber, muffler, Taguchi method, transmission loss

Procedia PDF Downloads 110
1887 Flexible Coupling between Gearbox and Pump (High Speed Machine)

Authors: Naif Mohsen Alharbi

Abstract:

This paper present failure occurred on flexible coupling installed at oil anf gas operation. Also it presents maintenance ideas implemented on the flexible coupling installed to transmit high torque from gearbox to pump. Basically, the machine train is including steam turbine which drives the pump and there is gearbox located in between for speed reduction. investigation are identifying the root causes, solving and developing the technology designs or bad actor. This report provides the study intentionally for continues operation optimization, utilize the advanced opportunity and implement a improvement. Objective: The main objectives of the investigation are identifying the root causes, solving and developing the technology designs or bad actor. Ultimately, fulfilling the operation productivity, also ensuring better technology, quality and design by solutions. This report provides the study intentionally for continues operation optimization, utilize the advanced opportunity and implemet improvement. Method: The method used in this project was a very focused root cause analysis procedure that incorporated engineering analysis and measurements. The analysis method extensively covers the measuring of the complete coupling dimensions. Including the membranes thickness, hubs, bore diameter and total length, dismantle flexible coupling to diagnose how deep the coupling has been affected. Also, defining failure modes, so that the causes could be identified and verified. Moreover, Vibration analysis and metallurgy test. Lastly applying several solutions by advanced tools (will be mentioned in detail). Results and observation: Design capacity: Coupling capacity is an inadequate to fulfil 100% of operating conditions. Therefore, design modification of service factor to be at least 2.07 is crucial to address this issue and prevent recurrence of similar scenario, especially for the new upgrading project. Discharge fluctuation: High torque flexible coupling encountered during the operation. Therefore, discharge valve behaviour, tuning, set point and general conditions revaluated and modified subsequently, it can be used as baseline for upcoming Coupling design project. Metallurgy test: Material of flexible coupling membrane (discs) tested at the lab, for a detailed metallurgical investigation, better material grade has been selected for our operating conditions,

Keywords: high speed machine, reliabilty, flexible coupling, rotating equipment

Procedia PDF Downloads 60
1886 Enhanced Tensor Tomographic Reconstruction: Integrating Absorption, Refraction and Temporal Effects

Authors: Lukas Vierus, Thomas Schuster

Abstract:

A general framework is examined for dynamic tensor field tomography within an inhomogeneous medium characterized by refraction and absorption, treated as an inverse source problem concerning the associated transport equation. Guided by Fermat’s principle, the Riemannian metric within the specified domain is determined by the medium's refractive index. While considerable literature exists on the inverse problem of reconstructing a tensor field from its longitudinal ray transform within a static Euclidean environment, limited inversion formulas and algorithms are available for general Riemannian metrics and time-varying tensor fields. It is established that tensor field tomography, akin to an inverse source problem for a transport equation, persists in dynamic scenarios. Framing dynamic tensor tomography as an inverse source problem embodies a comprehensive perspective within this domain. Ensuring well-defined forward mappings necessitates establishing existence and uniqueness for the underlying transport equations. However, the bilinear forms of the associated weak formulations fail to meet the coercivity condition. Consequently, recourse to viscosity solutions is taken, demonstrating their unique existence within suitable Sobolev spaces (in the static case) and Sobolev-Bochner spaces (in the dynamic case), under a specific assumption restricting variations in the refractive index. Notably, the adjoint problem can also be reformulated as a transport equation, with analogous results regarding uniqueness. Analytical solutions are expressed as integrals over geodesics, facilitating more efficient evaluation of forward and adjoint operators compared to solving partial differential equations. Certainly, here's the revised sentence in English: Numerical experiments are conducted using a Nesterov-accelerated Landweber method, encompassing various fields, absorption coefficients, and refractive indices, thereby illustrating the enhanced reconstruction achieved through this holistic modeling approach.

Keywords: attenuated refractive dynamic ray transform of tensor fields, geodesics, transport equation, viscosity solutions

Procedia PDF Downloads 40
1885 A Topology-Based Dynamic Repair Strategy for Enhancing Urban Road Network Resilience under Flooding

Authors: Xuhui Lin, Qiuchen Lu, Yi An, Tao Yang

Abstract:

As global climate change intensifies, extreme weather events such as floods increasingly threaten urban infrastructure, making the vulnerability of urban road networks a pressing issue. Existing static repair strategies fail to adapt to the rapid changes in road network conditions during flood events, leading to inefficient resource allocation and suboptimal recovery. The main research gap lies in the lack of repair strategies that consider both the dynamic characteristics of networks and the progression of flood propagation. This paper proposes a topology-based dynamic repair strategy that adjusts repair priorities based on real-time changes in flood propagation and traffic demand. Specifically, a novel method is developed to assess and enhance the resilience of urban road networks during flood events. The method combines road network topological analysis, flood propagation modelling, and traffic flow simulation, introducing a local importance metric to dynamically evaluate the significance of road segments across different spatial and temporal scales. Using London's road network and rainfall data as a case study, the effectiveness of this dynamic strategy is compared to traditional and Transport for London (TFL) strategies. The most significant highlight of the research is that the dynamic strategy substantially reduced the number of stranded vehicles across different traffic demand periods, improving efficiency by up to 35.2%. The advantage of this method lies in its ability to adapt in real-time to changes in network conditions, enabling more precise resource allocation and more efficient repair processes. This dynamic strategy offers significant value to urban planners, traffic management departments, and emergency response teams, helping them better respond to extreme weather events like floods, enhance overall urban resilience, and reduce economic losses and social impacts.

Keywords: Urban resilience, road networks, flood response, dynamic repair strategy, topological analysis

Procedia PDF Downloads 25
1884 Economic Evaluation of Degradation by Corrosion of an On-Grid Battery Energy Storage System: A Case Study in Algeria Territory

Authors: Fouzia Brihmat

Abstract:

Economic planning models, which are used to build microgrids and distributed energy resources, are the current norm for expressing such confidence (DER). These models often decide both short-term DER dispatch and long-term DER investments. This research investigates the most cost-effective hybrid (photovoltaic-diesel) renewable energy system (HRES) based on Total Net Present Cost (TNPC) in an Algerian Saharan area, which has a high potential for solar irradiation and has a production capacity of 1GW/h. Lead-acid batteries have been around much longer and are easier to understand, but have limited storage capacity. Lithium-ion batteries last longer, are lighter, but generally more expensive. By combining the advantages of each chemistry, we produce cost-effective high-capacity battery banks that operate solely on AC coupling. The financial implications of this research describe the corrosion process that occurs at the interface between the active material and grid material of the positive plate of a lead-acid battery. The best cost study for the HRES is completed with the assistance of the HOMER Pro MATLAB Link. Additionally, during the course of the project's 20 years, the system is simulated for each time step. In this model, which takes into consideration decline in solar efficiency, changes in battery storage levels over time, and rises in fuel prices above the rate of inflation. The trade-off is that the model is more accurate, but it took longer to compute. As a consequence, the model is more precise, but the computation takes longer. We initially utilized the Optimizer to run the model without MultiYear in order to discover the best system architecture. The optimal system for the single-year scenario is the Danvest generator, which has 760 kW, 200 kWh of the necessary quantity of lead-acid storage, and a somewhat lower COE of $0.309/kWh. Different scenarios that account for fluctuations in the gasified biomass generator's production of electricity have been simulated, and various strategies to guarantee the balance between generation and consumption have been investigated. The technological optimization of the same system has been finished and is being reviewed in a recent paper study.

Keywords: battery, corrosion, diesel, economic planning optimization, hybrid energy system, lead-acid battery, multi-year planning, microgrid, price forecast, PV, total net present cost

Procedia PDF Downloads 81
1883 A Reduced Ablation Model for Laser Cutting and Laser Drilling

Authors: Torsten Hermanns, Thoufik Al Khawli, Wolfgang Schulz

Abstract:

In laser cutting as well as in long pulsed laser drilling of metals, it can be demonstrated that the ablation shape (the shape of cut faces respectively the hole shape) that is formed approaches a so-called asymptotic shape such that it changes only slightly or not at all with further irradiation. These findings are already known from the ultrashort pulse (USP) ablation of dielectric and semiconducting materials. The explanation for the occurrence of an asymptotic shape in laser cutting and long pulse drilling of metals is identified, its underlying mechanism numerically implemented, tested and clearly confirmed by comparison with experimental data. In detail, there now is a model that allows the simulation of the temporal (pulse-resolved) evolution of the hole shape in laser drilling as well as the final (asymptotic) shape of the cut faces in laser cutting. This simulation especially requires much less in the way of resources, such that it can even run on common desktop PCs or laptops. Individual parameters can be adjusted using sliders – the simulation result appears in an adjacent window and changes in real time. This is made possible by an application-specific reduction of the underlying ablation model. Because this reduction dramatically decreases the complexity of calculation, it produces a result much more quickly. This means that the simulation can be carried out directly at the laser machine. Time-intensive experiments can be reduced and set-up processes can be completed much faster. The high speed of simulation also opens up a range of entirely different options, such as metamodeling. Suitable for complex applications with many parameters, metamodeling involves generating high-dimensional data sets with the parameters and several evaluation criteria for process and product quality. These sets can then be used to create individual process maps that show the dependency of individual parameter pairs. This advanced simulation makes it possible to find global and local extreme values through mathematical manipulation. Such simultaneous optimization of multiple parameters is scarcely possible by experimental means. This means that new methods in manufacturing such as self-optimization can be executed much faster. However, the software’s potential does not stop there; time-intensive calculations exist in many areas of industry. In laser welding or laser additive manufacturing, for example, the simulation of thermal induced residual stresses still uses up considerable computing capacity or is even not possible. Transferring the principle of reduced models promises substantial savings there, too.

Keywords: asymptotic ablation shape, interactive process simulation, laser drilling, laser cutting, metamodeling, reduced modeling

Procedia PDF Downloads 209
1882 Mixed Integer Programming-Based One-Class Classification Method for Process Monitoring

Authors: Younghoon Kim, Seoung Bum Kim

Abstract:

One-class classification plays an important role in detecting outlier and abnormality from normal observations. In the previous research, several attempts were made to extend the scope of application of the one-class classification techniques to statistical process control problems. For most previous approaches, such as support vector data description (SVDD) control chart, the design of the control limits is commonly based on the assumption that the proportion of abnormal observations is approximately equal to an expected Type I error rate in Phase I process. Because of the limitation of the one-class classification techniques based on convex optimization, we cannot make the proportion of abnormal observations exactly equal to expected Type I error rate: controlling Type I error rate requires to optimize constraints with integer decision variables, but convex optimization cannot satisfy the requirement. This limitation would be undesirable in theoretical and practical perspective to construct effective control charts. In this work, to address the limitation of previous approaches, we propose the one-class classification algorithm based on the mixed integer programming technique, which can solve problems formulated with continuous and integer decision variables. The proposed method minimizes the radius of a spherically shaped boundary subject to the number of normal data to be equal to a constant value specified by users. By modifying this constant value, users can exactly control the proportion of normal data described by the spherically shaped boundary. Thus, the proportion of abnormal observations can be made theoretically equal to an expected Type I error rate in Phase I process. Moreover, analogous to SVDD, the boundary can be made to describe complex structures by using some kernel functions. New multivariate control chart applying the effectiveness of the algorithm is proposed. This chart uses a monitoring statistic to characterize the degree of being an abnormal point as obtained through the proposed one-class classification. The control limit of the proposed chart is established by the radius of the boundary. The usefulness of the proposed method was demonstrated through experiments with simulated and real process data from a thin film transistor-liquid crystal display.

Keywords: control chart, mixed integer programming, one-class classification, support vector data description

Procedia PDF Downloads 170
1881 On the Optimization of a Decentralized Photovoltaic System

Authors: Zaouche Khelil, Talha Abdelaziz, Berkouk El Madjid

Abstract:

In this paper, we present a grid-tied photovoltaic system. The studied topology is structured around a seven-level inverter, supplying a non-linear load. A three-stage step-up DC/DC converter ensures DC-link balancing. The presented system allows the extraction of all the available photovoltaic power. This extracted energy feeds the local load; the surplus energy is injected into the electrical network. During poor weather conditions, where the photovoltaic panels cannot meet the energy needs of the load, the missing power is supplied by the electrical network. At the common connexion point, the network current shows excellent spectral performances.

Keywords: seven-level inverter, multi-level DC/DC converter, photovoltaic, non-linear load

Procedia PDF Downloads 180
1880 Overview of Adaptive Spline interpolation

Authors: Rongli Gai, Zhiyuan Chang

Abstract:

At this stage, in view of various situations in the interpolation process, most researchers use self-adaptation to adjust the interpolation process, which is also one of the current and future research hotspots in the field of CNC machining. In the interpolation process, according to the overview of the spline curve interpolation algorithm, the adaptive analysis is carried out from the factors affecting the interpolation process. The adaptive operation is reflected in various aspects, such as speed, parameters, errors, nodes, feed rates, random Period, sensitive point, step size, curvature, adaptive segmentation, adaptive optimization, etc. This paper will analyze and summarize the research of adaptive imputation in the direction of the above factors affecting imputation.

Keywords: adaptive algorithm, CNC machining, interpolation constraints, spline curve interpolation

Procedia PDF Downloads 189
1879 Water Quality Trading with Equitable Total Maximum Daily Loads

Authors: S. Jamshidi, E. Feizi Ashtiani, M. Ardestani, A. Feizi Ashtiani

Abstract:

Waste load allocation (WLA) strategies usually intend to find economical policies for water resource management. Water quality trading (WQT) is an approach that uses discharge permit market to reduce total environmental protection costs. This primarily requires assigning discharge limits known as total maximum daily loads (TMDLs). These are determined by monitoring organizations with respect to the receiving water quality and remediation capabilities. The purpose of this study is to compare two approaches of TMDL assignment for WQT policy in small catchment area of Haraz River, in north of Iran. At first, TMDLs are assigned uniformly for the whole point sources to keep the concentrations of BOD and dissolved oxygen (DO) at the standard level at checkpoint (terminus point). This was simply simulated and controlled by Qual2kw software. In the second scenario, TMDLs are assigned using multi objective particle swarm optimization (MOPSO) method in which the environmental violation at river basin and total treatment costs are minimized simultaneously. In both scenarios, the equity index and the WLA based on trading discharge permits (TDP) are calculated. The comparative results showed that using economically optimized TMDLs (2nd scenario) has slightly more cost savings rather than uniform TMDL approach (1st scenario). The former annually costs about 1 M$ while the latter is 1.15 M$. WQT can decrease these annual costs to 0.9 and 1.1 M$, respectively. In other word, these approaches may save 35 and 45% economically in comparison with command and control policy. It means that using multi objective decision support systems (DSS) may find more economical WLA, however its outcome is not necessarily significant in comparison with uniform TMDLs. This may be due to the similar impact factors of dischargers in small catchments. Conversely, using uniform TMDLs for WQT brings more equity that makes stakeholders not feel that much envious of difference between TMDL and WQT allocation. In addition, for this case, determination of TMDLs uniformly would be much easier for monitoring. Consequently, uniform TMDL for TDP market is recommended as a sustainable approach. However, economical TMDLs can be used for larger watersheds.

Keywords: waste load allocation (WLA), water quality trading (WQT), total maximum daily loads (TMDLs), Haraz River, multi objective particle swarm optimization (MOPSO), equity

Procedia PDF Downloads 388