Search results for: hybrid optimization
1291 Short-Term Operation Planning for Energy Management of Exhibition Hall
Authors: Yooncheol Lee, Jeongmin Kim, Kwang Ryel Ryu
Abstract:
This paper deals with the establishment of a short-term operational plan for an air conditioner for efficient energy management of exhibition hall. The short-term operational plan is composed of a time series of operational schedules, which we have searched using genetic algorithms. Establishing operational schedule should be considered the future trends of the variables affecting the exhibition hall environment. To reflect continuously changing factors such as external temperature and occupant, short-term operational plans should be updated in real time. But it takes too much time to evaluate a short-term operational plan using EnergyPlus, a building emulation tool. For that reason, it is difficult to update the operational plan in real time. To evaluate the short-term operational plan, we designed prediction models based on machine learning with fast evaluation speed. This model, which was created by learning the past operational data, is accurate and fast. The collection of operational data and the verification of operational plans were made using EnergyPlus. Experimental results show that the proposed method can save energy compared to the reactive control method.Keywords: exhibition hall, energy management, predictive model, simulation-based optimization
Procedia PDF Downloads 3401290 Comparative Analysis of the Computer Methods' Usage for Calculation of Hydrocarbon Reserves in the Baltic Sea
Authors: Pavel Shcherban, Vlad Golovanov
Abstract:
Nowadays, the depletion of hydrocarbon deposits on the land of the Kaliningrad region leads to active geological exploration and development of oil and natural gas reserves in the southeastern part of the Baltic Sea. LLC 'Lukoil-Kaliningradmorneft' implements a comprehensive program for the development of the region's shelf in 2014-2023. Due to heterogeneity of reservoir rocks in various open fields, as well as with ambiguous conclusions on the contours of deposits, additional geological prospecting and refinement of the recoverable oil reserves are carried out. The key element is use of an effective technique of computer stock modeling at the first stage of processing of the received data. The following step uses information for the cluster analysis, which makes it possible to optimize the field development approaches. The article analyzes the effectiveness of various methods for reserves' calculation and computer modelling methods of the offshore hydrocarbon fields. Cluster analysis allows to measure influence of the obtained data on the development of a technical and economic model for mining deposits. The relationship between the accuracy of the calculation of recoverable reserves and the need of modernization of existing mining infrastructure, as well as the optimization of the scheme of opening and development of oil deposits, is observed.Keywords: cluster analysis, computer modelling of deposits, correction of the feasibility study, offshore hydrocarbon fields
Procedia PDF Downloads 1671289 Depolymerization of Lignin in Sugarcane Bagasse by Hydrothermal Liquefaction to Optimize Catechol Formation
Authors: Nirmala Deenadayalu, Kwanele B. Mazibuko, Lethiwe D. Mthembu
Abstract:
Sugarcane bagasse is the residue obtained after the extraction of sugar from the sugarcane. The main aim of this work was to produce catechol from sugarcane bagasse. The optimization of catechol production was investigated using a Box-Behnken design of experiments. The sugarcane bagasse was heated in a Parr reactor at a set temperature. The reactions were carried out at different temperatures (100-250) °C, catalyst loading (1% -10% KOH (m/v)) and reaction times (60 – 240 min) at 17 bar pressure. The solid and liquid fractions were then separated by vacuum filtration. The liquid fraction was analyzed for catechol using high-pressure liquid chromatography (HPLC) and characterized for the functional groups using Fourier transform infrared spectroscopy (FTIR). The optimized condition for catechol production was 175 oC, 240 min, and 10 % KOH with a catechol yield of 79.11 ppm. Since the maximum time was 240 min and 10 % KOH, a further series of experiments were conducted at 175 oC, 260 min, and 20 % KOH and yielded 2.46 ppm catechol, which was a large reduction in catechol produced. The HPLC peak for catechol was obtained at 2.5 min for the standards and the samples. The FTIR peak at 1750 cm⁻¹ was due to the C=C vibration band of the aromatic ring in the catechol present for both the standard and the samples. The peak at 3325 cm⁻¹ was due to the hydrogen-bonded phenolic OH vibration bands for the catechol. The ANOVA analysis was also performed on the set of experimental data to obtain the factors that most affected the amount of catechol produced.Keywords: catechol, sugarcane bagasse, lignin, hydrothermal liquefaction
Procedia PDF Downloads 1031288 Multiaxial Stress Based High Cycle Fatigue Model for Adhesive Joint Interfaces
Authors: Martin Alexander Eder, Sergei Semenov
Abstract:
Many glass-epoxy composite structures, such as large utility wind turbine rotor blades (WTBs), comprise of adhesive joints with typically thick bond lines used to connect the different components during assembly. Performance optimization of rotor blades to increase power output by simultaneously maintaining high stiffness-to-low-mass ratios entails intricate geometries in conjunction with complex anisotropic material behavior. Consequently, adhesive joints in WTBs are subject to multiaxial stress states with significant stress gradients depending on the local joint geometry. Moreover, the dynamic aero-elastic interaction of the WTB with the airflow generates non-proportional, variable amplitude stress histories in the material. Empiricism shows that a prominent failure type in WTBs is high cycle fatigue failure of adhesive bond line interfaces, which in fact over time developed into a design driver as WTB sizes increase rapidly. Structural optimization employed at an early design stage, therefore, sets high demands on computationally efficient interface fatigue models capable of predicting the critical locations prone for interface failure. The numerical stress-based interface fatigue model presented in this work uses the Drucker-Prager criterion to compute three different damage indices corresponding to the two interface shear tractions and the outward normal traction. The two-parameter Drucker-Prager model was chosen because of its ability to consider shear strength enhancement under compression and shear strength reduction under tension. The governing interface damage index is taken as the maximum of the triple. The damage indices are computed through the well-known linear Palmgren-Miner rule after separate rain flow-counting of the equivalent shear stress history and the equivalent pure normal stress history. The equivalent stress signals are obtained by self-similar scaling of the Drucker-Prager surface whose shape is defined by the uniaxial tensile strength and the shear strength such that it intersects with the stress point at every time step. This approach implicitly assumes that the damage caused by the prevailing multiaxial stress state is the same as the damage caused by an amplified equivalent uniaxial stress state in the three interface directions. The model was implemented as Python plug-in for the commercially available finite element code Abaqus for its use with solid elements. The model was used to predict the interface damage of an adhesively bonded, tapered glass-epoxy composite cantilever I-beam tested by LM Wind Power under constant amplitude compression-compression tip load in the high cycle fatigue regime. Results show that the model was able to predict the location of debonding in the adhesive interface between the webfoot and the cap. Moreover, with a set of two different constant life diagrams namely in shear and tension, it was possible to predict both the fatigue lifetime and the failure mode of the sub-component with reasonable accuracy. It can be concluded that the fidelity, robustness and computational efficiency of the proposed model make it especially suitable for rapid fatigue damage screening of large 3D finite element models subject to complex dynamic load histories.Keywords: adhesive, fatigue, interface, multiaxial stress
Procedia PDF Downloads 1711287 Separation Performance of CO₂ by Mixed Matrix Membrane Comprising Carbide-Derived Carbon
Authors: Musa Najimu, Isam Aljundi
Abstract:
In this study, the development of mixed matrix membrane (MMM) containing carbide-derived carbon (CDC) for the separation of CO₂ was investigated. MMM with four different loadings (0.1 to 2 wt%) were prepared by the dry/wet phase inversion technique. Prior to this, the formula of the control polysulfone (PSF) membrane was optimized in terms of the PSF concentration in a mixture of NMP/THF solvents and ethanol. Prepared samples were characterized and tested for CO₂ and CH₄ gas permeation. The optimization of the control PSF membrane revealed that 30 wt% PSF is the critical polymer concentration in the formulation. Characterization results unveiled reinforcement of thermal stability and improved polarity imparted by CDC in the MMM, in addition to uniform dispersion of filler up to 1 wt% loading. Furthermore, the incorporation of CDC in PSF membrane formulation enhanced both the CO₂ permeance and ideal selectivity over the control membrane. A CDC loading of 0.5 wt% resulted in the highest CO₂ permeance of 5.5 GPU corresponding to 120% increase in permeance while a CDC loading of 1 wt% resulted in the highest selectivity (CO₂ /CH₄) of 27 corresponding to 29% increase in selectivity. Studies of operating temperature effect showed that an optimum operating temperature for M1.0 membrane is 20 ⁰C. In addition, the feed pressure studies showed that high pressure feeds will favor high performance of the membrane and a good CO₂ /CH₄ separation.Keywords: carbide derived carbon, mixed matrix membrane, CO₂ separation, polysulfone
Procedia PDF Downloads 2081286 Two Efficient Heuristic Algorithms for the Integrated Production Planning and Warehouse Layout Problem
Authors: Mohammad Pourmohammadi Fallah, Maziar Salahi
Abstract:
In the literature, a mixed-integer linear programming model for the integrated production planning and warehouse layout problem is proposed. To solve the model, the authors proposed a Lagrangian relax-and-fix heuristic that takes a significant amount of time to stop with gaps above 5$\%$ for large-scale instances. Here, we present two heuristic algorithms to solve the problem. In the first one, we use a greedy approach by allocating warehouse locations with less reservation costs and also less transportation costs from the production area to locations and from locations to the output point to items with higher demands. Then a smaller model is solved. In the second heuristic, first, we sort items in descending order according to the fraction of the sum of the demands for that item in the time horizon plus the maximum demand for that item in the time horizon and the sum of all its demands in the time horizon. Then we categorize the sorted items into groups of 3, 4, or 5 and solve a small-scale optimization problem for each group, hoping to improve the solution of the first heuristic. Our preliminary numerical results show the effectiveness of the proposed heuristics.Keywords: capacitated lot-sizing, warehouse layout, mixed-integer linear programming, heuristics algorithm
Procedia PDF Downloads 1981285 Left to Right-Right Most Parsing Algorithm with Lookahead
Authors: Jamil Ahmed
Abstract:
Left to Right-Right Most (LR) parsing algorithm is a widely used algorithm of syntax analysis. It is contingent on a parsing table, whereas the parsing tables are extracted from the grammar. The parsing table specifies the actions to be taken during parsing. It requires that the parsing table should have no action conflicts for the same input symbol. This requirement imposes a condition on the class of grammars over which the LR algorithms work. However, there are grammars for which the parsing tables hold action conflicts. In such cases, the algorithm needs a capability of scanning (looking-ahead) next input symbols ahead of the current input symbol. In this paper, a ‘Left to Right’-‘Right Most’ parsing algorithm with lookahead capability is introduced. The 'look-ahead' capability in the LR parsing algorithm is the major contribution of this paper. The practicality of the proposed algorithm is substantiated by the parser implementation of the Context Free Grammar (CFG) of an already proposed programming language 'State Controlled Object Oriented Programming' (SCOOP). SCOOP’s Context Free Grammar has 125 productions and 192 item sets. This algorithm parses SCOOP while the grammar requires to ‘look ahead’ the input symbols due to action conflicts in its parsing table. Proposed LR parsing algorithm with lookahead capability can be viewed as an optimization of ‘Simple Left to Right’-‘Right Most’ (SLR) parsing algorithm.Keywords: left to right-right most parsing, syntax analysis, bottom-up parsing algorithm
Procedia PDF Downloads 1271284 Surface Functionalized Biodegradable Polymersome for Targeted Drug Delivery
Authors: Susmita Roy, Madhavan Nallani
Abstract:
In recent years' polymersomes, self-assembled polymeric vesicles emerge from block copolymers, have been widely investigated due to their enhance stability and unique advantageous properties compared to their phospholipid counterpart, liposomes, dendrimers, and micelles. It provides a distinctive platform for advanced therapeutics and the creation of complex (bio) catalytically active systems for research in Nanomedicine and synthetic biology. Inspired by nature, where compartmentalization of biological components is all ubiquitous, we are interested in developing a platform technology of self-assembled multifunctional compartments with applications in areas from targeted drug/gene delivery, biosensing, pharmaceutical to cosmetics. Polymersome surfaces can be a proper choice of derivatization with a controlled amount of functional groups. To achieve site-specific targeting of polymersomes, biological recognition motives can be attached to the polymersomes surface by standard bioconjugation techniques, (like esterification, amidation, thiol-maleimide coupling, click-chemistry routes or other coupling methods). Herein, we are developing easy going, one-step bioconjugation strategies for site-specific surface functionalized biodegradable polymeric and/or polymer-lipid hybrid vesicles for targeted drug delivery. Biodegradable polymer, polycaprolactone-b-polyethylene glycol (PCL-PEG), polylactic acid-b-polyethylene glycol (PLA-PEG) and phospholipid, 1-palmitoyl-2- oleoyl-sn-glycero-3-phosphocholine (POPC) has been widely used for numerous vesicle formulations. Some of these drug-loaded formulations are being tested on mice for controlled release. These surface functionalized polymersomes are also appropriate for membrane protein reconstitution/insertion, antibodies conjugation and various bioconjugation with diverse targeted molecules for controlled drug delivery.Keywords: drug delivery, membrane protein, polymersome, surface modification
Procedia PDF Downloads 1551283 Molecular Engineering of Intrinsically Microporous Polybenzimidazole for Energy-efficient Gas Separation
Authors: Mahmoud Abdulhamid, Rifan Hardian, Prashant Bhatt, Shuvo Datta, Adrian Ramirez, Jorge Gascon, Mohamed Eddaoudi, Gyorgy Szekely
Abstract:
Polybenzimidazole (PBI) is a high-performance polymer that exhibits high thermal and chemical stability. However, it suffers from low porosity and low fractional free volume, which hinder its application as separation material. Herein, we demonstrate the molecular engineering of gas separation materials by manipulating a PBI backbone possessing kinked moieties. PBI was selected as it contains NH groups which increase the affinity towards CO₂, increase sorption capacity, and favors CO₂ over other gasses. We have designed and synthesized an intrinsically microporous polybenzimidazole (iPBI) featuring a spirobisindane structure. Introducing a kinked moiety in conjunction with crosslinking enhanced the polymer properties, markedly increasing the gas separation performance. In particular, the BET surface area of PBI increased 30-fold by replacing a flat benzene ring with a kinked structure. iPBI displayed a good CO₂ uptake of 1.4 mmol g⁻¹ at 1 bar and 3.6 mmol g⁻¹ at 10 bar. Gas sorption uptake and breakthrough experiments were conducted using mixtures of CO₂/CH₄ (50%/50%) and CO₂/N₂ (50%/50%), which revealed the high selectivity of CO₂ over both CH₄ and N₂. The obtained CO₂/N₂ selectivity is attractive for power plant flue gas application requiring CO₂ capturing materials. Energy and process simulations of biogas CO₂ removal demonstrated that up to 70% of the capture energy could be saved when iPBI was used rather than the current amine technology (methyl diethanolamine [MDEA]). Similarly, the combination of iPBI and MDEA in a hybrid system exhibited the highest CO₂ capture yield (99%), resulting in nearly 50% energy saving. The concept of enhancing the porosity of PBI using kinked moieties provides new scope for designing highly porous polybenzimidazoles for various separation processes.Keywords: polybenzimidazole (PBI), intrinsically microporous polybenzimidazole (iPBI), gas separation, pnergy and process simulations
Procedia PDF Downloads 921282 Numerical Analysis of Engine Performance and Emission of a 2-Stroke Opposed Piston Hydrogen Engine
Authors: Bahamin Bazooyar, Xinyan Wang, Hua Zhao
Abstract:
As a zero-carbon fuel, hydrogen can be used in combustion engines to avoid carbon emissions. This paper numerically investigates the engine performance of a two-stroke opposed piston hydrogen engine by using three-dimensional (3D) Computational Fluid Dynamics (CFD) simulations. The engine displacement is 12.2 cm, and the compression ratio of 39. RANS simulations with the k-ε turbulence model and coupled chemistry combustion models are performed at an engine speed of 4500 rpm and hydrogen flow rate of up to 100 gr/s. In order to model the hydrogen injection process, the hydrogen nozzle was meshed with refined mesh, and injection pressure varied between 100 and 200 bars. In order to optimize the hydrogen combustion process, the injection timing was optimized between 15 before the top dead center and 10. The results showed that the combustion efficiency was mostly influenced by the injection pressures due to its impact on the fuel/air mixing and charge inhomogeneity. Nitrogen oxide (NOₓ) emissions are well correlated with engine peak temperatures, demonstrating that the thermal NO mechanism is dominant under engine conditions. Through the optimization of hydrogen injection timing and pressure, the peak thermal efficiency of 45 and NOx emission of 15 ppm/kWh can be achieved at an injection timing of 350 CA and pressure of 160 bars.Keywords: engine, hydrogen, diesel, two-stroke, opposed-piston, decarbonisation
Procedia PDF Downloads 151281 Integrated Wastewater Reuse Project of the Faculty of Sciences AinChock, Morocco
Authors: Nihad Chakri, Btissam El Amrani, Faouzi Berrada, Fouad Amraoui
Abstract:
In Morocco, water scarcity requires the exploitation of non-conventional resources. Rural areas are under-equipped with sanitation infrastructure, unlike urban areas. Decentralized and low-cost solutions could improve the quality of life of the population and the environment. In this context, the Faculty of Sciences Ain Chock "FSAC" has undertaken an integrated project to treat part of its wastewater using a decentralized compact system. The project will propose alternative solutions that are inexpensive and adapted to the context of peri-urban and rural areas in order to treat the wastewater generated and use it for irrigation, watering, and cleaning. For this purpose, several tests were carried out in the laboratory in order to develop a liquid waste treatment system optimized for local conditions. Based on the results obtained at the laboratory scale of the different proposed scenarios, we designed and implemented a prototype of a mini wastewater treatment plant for the Faculty. In this article, we will outline the steps of dimensioning, construction, and monitoring of the mini-station in our Faculty.Keywords: wastewater, purification, optimization, vertical filter, MBBR process, sizing, decentralized pilot, reuse, irrigation, sustainable development
Procedia PDF Downloads 1161280 Optimization of Element Type for FE Model and Verification of Analyses with Physical Tests
Authors: Mustafa Tufekci, Caner Guven
Abstract:
In Automotive Industry, sliding door systems that are also used as body closures, are safety members. Extreme product tests are realized to prevent failures in a design process, but these tests realized experimentally result in high costs. Finite element analysis is an effective tool used for the design process. These analyses are used before production of a prototype for validation of design according to customer requirement. In result of this, the substantial amount of time and cost is saved. Finite element model is created for geometries that are designed in 3D CAD programs. Different element types as bar, shell and solid, can be used for creating mesh model. The cheaper model can be created by the selection of element type, but combination of element type that was used in model, number and geometry of element and degrees of freedom affects the analysis result. Sliding door system is a good example which used these methods for this study. Structural analysis was realized for sliding door mechanism by using FE models. As well, physical tests that have same boundary conditions with FE models were realized. Comparison study for these element types, were done regarding test and analyses results then the optimum combination was achieved.Keywords: finite element analysis, sliding door mechanism, element type, structural analysis
Procedia PDF Downloads 3301279 Discrete-Event Modeling and Simulation Methodologies: Past, Present and Future
Authors: Gabriel Wainer
Abstract:
Modeling and Simulation methods have been used to better analyze the behavior of complex physical systems, and it is now common to use simulation as a part of the scientific and technological discovery process. M&S advanced thanks to the improvements in computer technology, which, in many cases, resulted in the development of simulation software using ad-hoc techniques. Formal M&S appeared in order to try to improve the development task of very complex simulation systems. Some of these techniques proved to be successful in providing a sound base for the development of discrete-event simulation models, improving the ease of model definition and enhancing the application development tasks; reducing costs and favoring reuse. The DEVS formalism is one of these techniques, which proved to be successful in providing means for modeling while reducing development complexity and costs. DEVS model development is based on a sound theoretical framework. The independence of M&S tasks made possible to run DEVS models on different environments (personal computers, parallel computers, real-time equipment, and distributed simulators) and middleware. We will present a historical perspective of discrete-event M&S methodologies, showing different modeling techniques. We will introduce DEVS origins and general ideas, and compare it with some of these techniques. We will then show the current status of DEVS M&S, and we will discuss a technological perspective to solve current M&S problems (including real-time simulation, interoperability, and model-centered development techniques). We will show some examples of the current use of DEVS, including applications in different fields. We will finally show current open topics in the area, which include advanced methods for centralized, parallel or distributed simulation, the need for real-time modeling techniques, and our view in these fields.Keywords: modeling and simulation, discrete-event simulation, hybrid systems modeling, parallel and distributed simulation
Procedia PDF Downloads 3231278 An Optimal Algorithm for Finding (R, Q) Policy in a Price-Dependent Order Quantity Inventory System with Soft Budget Constraint
Authors: S. Hamid Mirmohammadi, Shahrazad Tamjidzad
Abstract:
This paper is concerned with the single-item continuous review inventory system in which demand is stochastic and discrete. The budget consumed for purchasing the ordered items is not restricted but it incurs extra cost when exceeding specific value. The unit purchasing price depends on the quantity ordered under the all-units discounts cost structure. In many actual systems, the budget as a resource which is occupied by the purchased items is limited and the system is able to confront the resource shortage by charging more costs. Thus, considering the resource shortage costs as a part of system costs, especially when the amount of resource occupied by the purchased item is influenced by quantity discounts, is well motivated by practical concerns. In this paper, an optimization problem is formulated for finding the optimal (R, Q) policy, when the system is influenced by the budget limitation and a discount pricing simultaneously. Properties of the cost function are investigated and then an algorithm based on a one-dimensional search procedure is proposed for finding an optimal (R, Q) policy which minimizes the expected system costs .Keywords: (R, Q) policy, stochastic demand, backorders, limited resource, quantity discounts
Procedia PDF Downloads 6421277 Optimal Pressure Control and Burst Detection for Sustainable Water Management
Authors: G. K. Viswanadh, B. Rajasekhar, G. Venkata Ramana
Abstract:
Water distribution networks play a vital role in ensuring a reliable supply of clean water to urban areas. However, they face several challenges, including pressure control, pump speed optimization, and burst event detection. This paper combines insights from two studies to address these critical issues in Water distribution networks, focusing on the specific context of Kapra Municipality, India. The first part of this research concentrates on optimizing pressure control and pump speed in complex Water distribution networks. It utilizes the EPANET- MATLAB Toolkit to integrate EPANET functionalities into the MATLAB environment, offering a comprehensive approach to network analysis. By optimizing Pressure Reduce Valves (PRVs) and variable speed pumps (VSPs), this study achieves remarkable results. In the Benchmark Water Distribution System (WDS), the proposed PRV optimization algorithm reduces average leakage by 20.64%, surpassing the previous achievement of 16.07%. When applied to the South-Central and East zone WDS of Kapra Municipality, it identifies PRV locations that were previously missed by existing algorithms, resulting in average leakage reductions of 22.04% and 10.47%. These reductions translate to significant daily Water savings, enhancing Water supply reliability and reducing energy consumption. The second part of this research addresses the pressing issue of burst event detection and localization within the Water Distribution System. Burst events are a major contributor to Water losses and repair expenses. The study employs wireless sensor technology to monitor pressure and flow rate in real time, enabling the detection of pipeline abnormalities, particularly burst events. The methodology relies on transient analysis of pressure signals, utilizing Cumulative Sum and Wavelet analysis techniques to robustly identify burst occurrences. To enhance precision, burst event localization is achieved through meticulous analysis of time differentials in the arrival of negative pressure waveforms across distinct pressure sensing points, aided by nodal matrix analysis. To evaluate the effectiveness of this methodology, a PVC Water pipeline test bed is employed, demonstrating the algorithm's success in detecting pipeline burst events at flow rates of 2-3 l/s. Remarkably, the algorithm achieves a localization error of merely 3 meters, outperforming previously established algorithms. This research presents a significant advancement in efficient burst event detection and localization within Water pipelines, holding the potential to markedly curtail Water losses and the concomitant financial implications. In conclusion, this combined research addresses critical challenges in Water distribution networks, offering solutions for optimizing pressure control, pump speed, burst event detection, and localization. These findings contribute to the enhancement of Water Distribution System, resulting in improved Water supply reliability, reduced Water losses, and substantial cost savings. The integrated approach presented in this paper holds promise for municipalities and utilities seeking to improve the efficiency and sustainability of their Water distribution networks.Keywords: pressure reduce valve, complex networks, variable speed pump, wavelet transform, burst detection, CUSUM (Cumulative Sum), water pipeline monitoring
Procedia PDF Downloads 881276 Hybridization of Manually Extracted and Convolutional Features for Classification of Chest X-Ray of COVID-19
Authors: M. Bilal Ishfaq, Adnan N. Qureshi
Abstract:
COVID-19 is the most infectious disease these days, it was first reported in Wuhan, the capital city of Hubei in China then it spread rapidly throughout the whole world. Later on 11 March 2020, the World Health Organisation (WHO) declared it a pandemic. Since COVID-19 is highly contagious, it has affected approximately 219M people worldwide and caused 4.55M deaths. It has brought the importance of accurate diagnosis of respiratory diseases such as pneumonia and COVID-19 to the forefront. In this paper, we propose a hybrid approach for the automated detection of COVID-19 using medical imaging. We have presented the hybridization of manually extracted and convolutional features. Our approach combines Haralick texture features and convolutional features extracted from chest X-rays and CT scans. We also employ a minimum redundancy maximum relevance (MRMR) feature selection algorithm to reduce computational complexity and enhance classification performance. The proposed model is evaluated on four publicly available datasets, including Chest X-ray Pneumonia, COVID-19 Pneumonia, COVID-19 CTMaster, and VinBig data. The results demonstrate high accuracy and effectiveness, with 0.9925 on the Chest X-ray pneumonia dataset, 0.9895 on the COVID-19, Pneumonia and Normal Chest X-ray dataset, 0.9806 on the Covid CTMaster dataset, and 0.9398 on the VinBig dataset. We further evaluate the effectiveness of the proposed model using ROC curves, where the AUC for the best-performing model reaches 0.96. Our proposed model provides a promising tool for the early detection and accurate diagnosis of COVID-19, which can assist healthcare professionals in making informed treatment decisions and improving patient outcomes. The results of the proposed model are quite plausible and the system can be deployed in a clinical or research setting to assist in the diagnosis of COVID-19.Keywords: COVID-19, feature engineering, artificial neural networks, radiology images
Procedia PDF Downloads 761275 Leveraging Business to Business Collaborations to Optimize Reverse Haul Logistics
Authors: Pallav Singh, Rajesh Yabaji, Rajesh Dhir, Chanakya Hridaya
Abstract:
Supply Chain Costs for the Indian Industries have been on an exponential trend due to steep inflation on fundamental cost factors – Fuel, Labour, Rents. In this changing context organizations have been focusing on adopting multiple approaches to keep logistics costs under control to protect the profit margins. The lever of ‘Business to Business (B2B) collaboration’ can be used by organizations to garner higher value. Given the context of Indian Logistics Industry the penetration of B2B Collaboration initiatives have been limited. This paper outlines a structured framework for adoption of B2B collaboration through discussion of a successful initiative between ITC’s Leaf Tobacco Business and a leading Indian Media House. Multiple barriers to such a collaborative process exist which need to be addressed through comprehensive structured approaches. This paper outlines a generic framework approach to B2B collaboration for the Indian Logistics Space, outlining the guidelines for arriving at potential opportunities, identification of collaborators, effective tie-up process, design of operations and sustenance factors. The generic methods outlined can be used in any other industry and also builds a foundation for further research on many topics.Keywords: business to business collaboration, reverse haul logistics, transportation cost optimization, exports logistics
Procedia PDF Downloads 3291274 Reduction of Aerodynamic Drag Using Vortex Generators
Authors: Siddharth Ojha, Varun Dua
Abstract:
Classified as one of the most important reasons of aerodynamic drag in the sedan automobiles is the fluid flow separation near the vehicle’s rear end. To retard the separation of flow, bump-shaped vortex generators are being tested for its implementation to the roof end of a sedan vehicle. Frequently used in the aircrafts to prevent the separation of fluid flow, vortex generators themselves produce drag, but they also substantially reduce drag by preventing flow separation at the downstream. The net effects of vortex generators can be calculated by summing the positive and negative impacts and effects. Since this effect depends on dimensions and geometry of vortex generators, those present on the vehicle roof are optimized for maximum efficiency and performance. The model was tested through ANSYS CFD analysis and modeling. The model was tested in the wind tunnel for observing it’s properties such as aerodynamic drag and flow separation and a major time lag was gained by employing vortex generators in the scaled model. Major conclusions which were recorded during the analysis were a substantial 24% reduction in the aerodynamic drag and 14% increase in the efficiency of the sedan automobile as the flow separation from the surface is delayed. This paper presents the results of optimization, the effect of vortex generators in the flow field and the mechanism by which these effects occur and are regulated.Keywords: aerodynamics, aerodynamic devices, body, computational fluid dynamics (CFD), flow visualization
Procedia PDF Downloads 2231273 Hot Air Flow Annealing of MAPbI₃ Perovskite: Structural and Optical Properties
Authors: Mouad Ouafi, Lahoucine Atourki, Larbi Laanab, Erika Vega, Miguel Mollar, Bernabe Marib, Boujemaa Jaber
Abstract:
Despite the astonishing emergence of the methylammonium lead triiodide perovskite as a promising light harvester for solar cells, their physical properties in solution-processed MAPbI₃ are still crucial and need to be improved. The objective of this work is to investigate the hot airflow effect during the growth of MAPbI₃ films using the spin-coating process on their structural, optical and morphological proprieties. The experimental results show that many physical proprieties of the perovskite strongly depend on the air flow temperature and the optimization which has a beneficial effect on the perovskite quality. In fact, a clear improvement of the crystallinity and the crystallite size of MAPbI₃ perovskite is demonstrated by the XRD analyses, when the airflow temperature is increased up to 100°C. Alternatively, as far as the surface morphology is concerned, SEM micrographs show that significant homogenous nucleation, uniform surface distribution and pin holes free with highest surface coverture of 98% are achieved when the airflow temperature reaches 100°C. At this temperature, the improvement is also observed when considering the optical properties of the films. By contrast, a remarkable degradation of the MAPbI₃ perovskites associated to the PbI₂ phase formation is noticed, when the hot airflow temperature is higher than 100°C, especially 300°C.Keywords: hot air flow, crystallinity, surface coverage, perovskite morphology
Procedia PDF Downloads 1651272 Optimization in Locating Firefighting Stations Using GIS Data and AHP Model; A Case Study on Arak City
Authors: Hasan Heydari
Abstract:
In recent decades, locating urban services is one of the significant discussions in urban planning. Among these considerations, cities require more accurate planning in order to supply citizen needs, especially part of urban safety. In order to gain this goal, one of the main tasks of urban planners and managers is specifying suitable sites to locate firefighting stations. This study has been done to reach this purpose. Therefore effective criteria consist of coverage radius, population density, proximity to pathway network, land use (compatible and incompatible neighborhood) have been specified. After that, descriptive and local information of the criteria was provided and their layers were created in ArcGIS 9.3. Using Analytic Hierarchy Process (AHP) these criteria and their sub-criteria got the weights. These layers were classified regarding their weights and finally were overlaid by Index Overlay Model and provided the final site selection map for firefighting stations of Arak city. The results gained by analyzing in GIS environment indicate the existing fire station don’t cover the whole city sufficiently and some of the stations have established on the unsuitable sites. The output map indicates the best sites to locate firefighting stations of Arak.Keywords: site-selection, firefighting stations, analytic hierarchy process (AHP), GIS, index overlay model
Procedia PDF Downloads 3491271 Wall Heat Flux Mapping in Liquid Rocket Combustion Chamber with Different Jet Impingement Angles
Authors: O. S. Pradeep, S. Vigneshwaran, K. Praveen Kumar, K. Jeyendran, V. R. Sanal Kumar
Abstract:
The influence of injector attitude on wall heat flux plays an important role in predicting the start-up transient and also determining the combustion chamber wall durability of liquid rockets. In this paper comprehensive numerical studies have been carried out on an idealized liquid rocket combustion chamber to examine the transient wall heat flux during its start-up transient at different injector attitude. Numerical simulations have been carried out with the help of a validated 2d axisymmetric, double precision, pressure-based, transient, species transport, SST k-omega model with laminar finite rate model for governing turbulent-chemistry interaction for four cases with different jet intersection angles, viz., 0o, 30o, 45o, and 60o. We concluded that the jets intersection angle is having a bearing on the time and location of the maximum wall-heat flux zone of the liquid rocket combustion chamber during the start-up transient. We also concluded that the wall heat flux mapping in liquid rocket combustion chamber during the start-up transient is a meaningful objective for the chamber wall material selection and the lucrative design optimization of the combustion chamber for improving the payload capability of the rocket.Keywords: combustion chamber, injector, liquid rocket, rocket engine wall heat flux
Procedia PDF Downloads 4891270 A Hybrid Watermarking Scheme Using Discrete and Discrete Stationary Wavelet Transformation For Color Images
Authors: Bülent Kantar, Numan Ünaldı
Abstract:
This paper presents a new method which includes robust and invisible digital watermarking on images that is colored. Colored images are used as watermark. Frequency region is used for digital watermarking. Discrete wavelet transform and discrete stationary wavelet transform are used for frequency region transformation. Low, medium and high frequency coefficients are obtained by applying the two-level discrete wavelet transform to the original image. Low frequency coefficients are obtained by applying one level discrete stationary wavelet transform separately to all frequency coefficient of the two-level discrete wavelet transformation of the original image. For every low frequency coefficient obtained from one level discrete stationary wavelet transformation, watermarks are added. Watermarks are added to all frequency coefficients of two-level discrete wavelet transform. Totally, four watermarks are added to original image. In order to get back the watermark, the original and watermarked images are applied with two-level discrete wavelet transform and one level discrete stationary wavelet transform. The watermark is obtained from difference of the discrete stationary wavelet transform of the low frequency coefficients. A total of four watermarks are obtained from all frequency of two-level discrete wavelet transform. Obtained watermark results are compared with real watermark results, and a similarity result is obtained. A watermark is obtained from the highest similarity values. Proposed methods of watermarking are tested against attacks of the geometric and image processing. The results show that proposed watermarking method is robust and invisible. All features of frequencies of two level discrete wavelet transform watermarking are combined to get back the watermark from the watermarked image. Watermarks have been added to the image by converting the binary image. These operations provide us with better results in getting back the watermark from watermarked image by attacking of the geometric and image processing.Keywords: watermarking, DWT, DSWT, copy right protection, RGB
Procedia PDF Downloads 5381269 Seismic Fragility Assessment of Strongback Steel Braced Frames Subjected to Near-Field Earthquakes
Authors: Mohammadreza Salek Faramarzi, Touraj Taghikhany
Abstract:
In this paper, seismic fragility assessment of a recently developed hybrid structural system, known as the strongback system (SBS) is investigated. In this system, to mitigate the occurrence of the soft-story mechanism and improve the distribution of story drifts over the height of the structure, an elastic vertical truss is formed. The strengthened members of the braced span are designed to remain substantially elastic during levels of excitation where soft-story mechanisms are likely to occur and impose a nearly uniform story drift distribution. Due to the distinctive characteristics of near-field ground motions, it seems to be necessary to study the effect of these records on seismic performance of the SBS. To this end, a set of 56 near-field ground motion records suggested by FEMA P695 methodology is used. For fragility assessment, nonlinear dynamic analyses are carried out in OpenSEES based on the recommended procedure in HAZUS technical manual. Four damage states including slight, moderate, extensive, and complete damage (collapse) are considered. To evaluate each damage state, inter-story drift ratio and floor acceleration are implemented as engineering demand parameters. Further, to extend the evaluation of the collapse state of the system, a different collapse criterion suggested in FEMA P695 is applied. It is concluded that SBS can significantly increase the collapse capacity and consequently decrease the collapse risk of the structure during its life time. Comparing the observing mean annual frequency (MAF) of exceedance of each damage state against the allowable values presented in performance-based design methods, it is found that using the elastic vertical truss, improves the structural response effectively.Keywords: IDA, near-fault, probabilistic performance assessment, seismic fragility, strongback system, uncertainty
Procedia PDF Downloads 1181268 Tool for Maxillary Sinus Quantification in Computed Tomography Exams
Authors: Guilherme Giacomini, Ana Luiza Menegatti Pavan, Allan Felipe Fattori Alves, Marcela de Oliveira, Fernando Antonio Bacchim Neto, José Ricardo de Arruda Miranda, Seizo Yamashita, Diana Rodrigues de Pina
Abstract:
The maxillary sinus (MS), part of the paranasal sinus complex, is one of the most enigmatic structures in modern humans. The literature has suggested that MSs function as olfaction accessories, to heat or humidify inspired air, for thermoregulation, to impart resonance to the voice and others. Thus, the real function of the MS is still uncertain. Furthermore, the MS anatomy is complex and varies from person to person. Many diseases may affect the development process of sinuses. The incidence of rhinosinusitis and other pathoses in the MS is comparatively high, so, volume analysis has clinical value. Providing volume values for MS could be helpful in evaluating the presence of any abnormality and could be used for treatment planning and evaluation of the outcome. The computed tomography (CT) has allowed a more exact assessment of this structure, which enables a quantitative analysis. However, this is not always possible in the clinical routine, and if possible, it involves much effort and/or time. Therefore, it is necessary to have a convenient, robust, and practical tool correlated with the MS volume, allowing clinical applicability. Nowadays, the available methods for MS segmentation are manual or semi-automatic. Additionally, manual methods present inter and intraindividual variability. Thus, the aim of this study was to develop an automatic tool to quantity the MS volume in CT scans of paranasal sinuses. This study was developed with ethical approval from the authors’ institutions and national review panels. The research involved 30 retrospective exams of University Hospital, Botucatu Medical School, São Paulo State University, Brazil. The tool for automatic MS quantification, developed in Matlab®, uses a hybrid method, combining different image processing techniques. For MS detection, the algorithm uses a Support Vector Machine (SVM), by features such as pixel value, spatial distribution, shape and others. The detected pixels are used as seed point for a region growing (RG) segmentation. Then, morphological operators are applied to reduce false-positive pixels, improving the segmentation accuracy. These steps are applied in all slices of CT exam, obtaining the MS volume. To evaluate the accuracy of the developed tool, the automatic method was compared with manual segmentation realized by an experienced radiologist. For comparison, we used Bland-Altman statistics, linear regression, and Jaccard similarity coefficient. From the statistical analyses for the comparison between both methods, the linear regression showed a strong association and low dispersion between variables. The Bland–Altman analyses showed no significant differences between the analyzed methods. The Jaccard similarity coefficient was > 0.90 in all exams. In conclusion, the developed tool to quantify MS volume proved to be robust, fast, and efficient, when compared with manual segmentation. Furthermore, it avoids the intra and inter-observer variations caused by manual and semi-automatic methods. As future work, the tool will be applied in clinical practice. Thus, it may be useful in the diagnosis and treatment determination of MS diseases. Providing volume values for MS could be helpful in evaluating the presence of any abnormality and could be used for treatment planning and evaluation of the outcome. The computed tomography (CT) has allowed a more exact assessment of this structure which enables a quantitative analysis. However, this is not always possible in the clinical routine, and if possible, it involves much effort and/or time. Therefore, it is necessary to have a convenient, robust and practical tool correlated with the MS volume, allowing clinical applicability. Nowadays, the available methods for MS segmentation are manual or semi-automatic. Additionally, manual methods present inter and intraindividual variability. Thus, the aim of this study was to develop an automatic tool to quantity the MS volume in CT scans of paranasal sinuses. This study was developed with ethical approval from the authors’ institutions and national review panels. The research involved 30 retrospective exams of University Hospital, Botucatu Medical School, São Paulo State University, Brazil. The tool for automatic MS quantification, developed in Matlab®, uses a hybrid method, combining different image processing techniques. For MS detection, the algorithm uses a Support Vector Machine (SVM), by features such as pixel value, spatial distribution, shape and others. The detected pixels are used as seed point for a region growing (RG) segmentation. Then, morphological operators are applied to reduce false-positive pixels, improving the segmentation accuracy. These steps are applied in all slices of CT exam, obtaining the MS volume. To evaluate the accuracy of the developed tool, the automatic method was compared with manual segmentation realized by an experienced radiologist. For comparison, we used Bland-Altman statistics, linear regression and Jaccard similarity coefficient. From the statistical analyses for the comparison between both methods, the linear regression showed a strong association and low dispersion between variables. The Bland–Altman analyses showed no significant differences between the analyzed methods. The Jaccard similarity coefficient was > 0.90 in all exams. In conclusion, the developed tool to automatically quantify MS volume proved to be robust, fast and efficient, when compared with manual segmentation. Furthermore, it avoids the intra and inter-observer variations caused by manual and semi-automatic methods. As future work, the tool will be applied in clinical practice. Thus, it may be useful in the diagnosis and treatment determination of MS diseases.Keywords: maxillary sinus, support vector machine, region growing, volume quantification
Procedia PDF Downloads 5041267 IoT and Advanced Analytics Integration in Biogas Modelling
Authors: Rakesh Choudhary, Ajay Kumar, Deepak Sharma
Abstract:
The main goal of this paper is to investigate the challenges and benefits of IoT integration in biogas production. This overview explains how the inclusion of IoT can enhance biogas production efficiency. Therefore, such collected data can be explored by advanced analytics, including Artificial intelligence (AI) and Machine Learning (ML) algorithms, consequently improving bio-energy processes. To boost biogas generation efficiency, this report examines the use of IoT devices for real-time data collection on key parameters, e.g., pH, temperature, gas composition, and microbial growth. Real-time monitoring through big data has made it possible to detect diverse, complex trends in the process of producing biogas. The Informed by advanced analytics can also help in improving bio-energy production as well as optimizing operational conditions. Moreover, IoT allows remote observation, control and management, which decreases manual intervention needed whilst increasing process effectiveness. Such a paradigm shift in the incorporation of IoT technologies into biogas production systems helps to achieve higher productivity levels as well as more practical biomass quality biomethane through real-time monitoring-based proactive decision-making, thus driving continuous performance improvement.Keywords: internet of things, biogas, renewable energy, sustainability, anaerobic digestion, real-time monitoring, optimization
Procedia PDF Downloads 221266 Improving Fingerprinting-Based Localization System Using Generative AI
Authors: Getaneh Berie Tarekegn
Abstract:
A precise localization system is crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarming, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. The most common method for providing continuous positioning services in outdoor environments is by using a global navigation satellite system (GNSS). Due to nonline-of-sight, multipath, and weather conditions, GNSS systems do not perform well in dense urban, urban, and suburban areas.This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. It also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 0.39 m, and more than 90% of the errors are less than 0.82 m. According to numerical results, SRCLoc improves positioning performance and reduces radio map construction costs significantly compared to traditional methods.Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine
Procedia PDF Downloads 611265 Simulation and Characterization of Compact Magnetic Proton Recoil Spectrometer for Fast Neutron Spectra Measurements
Authors: Xingyu Peng, Qingyuan Hu, Xuebin Zhu, Xi Yuan
Abstract:
Neutron spectrometry has contributed much to the development of nuclear physics since 1932 and has also become an importance tool in several other fields, notably nuclear technology, fusion plasma diagnostics and radiation protection. Compared with neutron fluxes, neutron spectra can provide more detailed information on the internal physical process of neutron sources, such as fast neutron reactors, fusion plasma, fission-fusion hybrid reactors, and so on. However, high performance neutron spectrometer is not so commonly available as it requires the use of large and complex instrumentation. This work describes the development and characterization of a compact magnetic proton recoil (MPR) spectrometer for high-resolution measurements of fast neutron spectra. The compact MPR spectrometer is featured by its large recoil angle, small size permanent analysis magnet, short beam transport line and dual-purpose detector array for both steady state and pulsed neutron spectra measurement. A 3-dimensional electromagnetic particle transport code is developed to simulate the response function of the spectrometer. Simulation results illustrate that the performance of the spectrometer is mainly determined by n-p recoil foil and proton apertures, and an overall energy resolution of 3% is achieved for 14 MeV neutrons. Dedicated experiments using alpha source and mono-energetic neutron beam are employed to verify the simulated response function of the compact MPR spectrometer. These experimental results show a good agreement with the simulated ones, which indicates that the simulation code possesses good accuracy and reliability. The compact MPR spectrometer described in this work is a valuable tool for fast neutron spectra measurements for the fission or fusion devices.Keywords: neutron spectrometry, magnetic proton recoil spectrometer, neutron spectra, fast neutron
Procedia PDF Downloads 2041264 Elaboration and Characterization of Self-Compacting Mortar Based Biopolymer
Authors: I. Djefour, M. Saidi, I. Tlemsani, S. Toubal
Abstract:
Lignin is a molecule derived from wood and also generated as waste from the paper industry. With a view to its valorization and protection of the environment, we are interested in its use as a superplasticizer-type adjuvant in mortars and concretes to improve their mechanical strengths. The additives of the concrete have a very strong influence on the properties of the fresh and / or hardened concrete. This study examines the development and use of industrial waste and lignin extracted from a renewable natural source (wood) in cementitious materials. The use of these resources is known at present as a definite resurgence of interest in the development of building materials. Physicomechanical characteristics of mortars are determined by optimization quantity of the natural superplasticizer. The results show that the mechanical strengths of mortars based on natural adjuvant have improved by 20% (64 MPa) for a W/C ratio = 0.4, and the amount of natural adjuvant of dry extract needed is 40 times smaller than commercial adjuvant. This study has a scientific impact (improving the performance of the mortar with an increase in compactness and reduction of the quantity of water), ecological use of the lignin waste generated by the paper industry) and economic reduction of the cost price necessary to elaboration of self-compacting mortars and concretes).Keywords: biopolymer (lignin), industrial waste, mechanical resistances, self compacting mortars (SCM)
Procedia PDF Downloads 1681263 The Impact of Artificial Intelligence on Spare Parts Technology
Authors: Amir Andria Gad Shehata
Abstract:
Minimizing the inventory cost, optimizing the inventory quantities, and increasing system operational availability are the main motivations to enhance forecasting demand of spare parts in a major power utility company in Medina. This paper reports in an effort made to optimize the orders quantities of spare parts by improving the method of forecasting the demand. The study focuses on equipment that has frequent spare parts purchase orders with uncertain demand. The pattern of the demand considers a lumpy pattern which makes conventional forecasting methods less effective. A comparison was made by benchmarking various methods of forecasting based on experts’ criteria to select the most suitable method for the case study. Three actual data sets were used to make the forecast in this case study. Two neural networks (NN) approaches were utilized and compared, namely long short-term memory (LSTM) and multilayer perceptron (MLP). The results as expected, showed that the NN models gave better results than traditional forecasting method (judgmental method). In addition, the LSTM model had a higher predictive accuracy than the MLP model.Keywords: spare part, spare part inventory, inventory model, optimization, maintenanceneural network, LSTM, MLP, forecasting demand, inventory management
Procedia PDF Downloads 661262 Predictive Analytics in Traffic Flow Management: Integrating Temporal Dynamics and Traffic Characteristics to Estimate Travel Time
Authors: Maria Ezziani, Rabie Zine, Amine Amar, Ilhame Kissani
Abstract:
This paper introduces a predictive model for urban transportation engineering, which is vital for efficient traffic management. Utilizing comprehensive datasets and advanced statistical techniques, the model accurately forecasts travel times by considering temporal variations and traffic dynamics. Machine learning algorithms, including regression trees and neural networks, are employed to capture sequential dependencies. Results indicate significant improvements in predictive accuracy, particularly during peak hours and holidays, with the incorporation of traffic flow and speed variables. Future enhancements may integrate weather conditions and traffic incidents. The model's applications range from adaptive traffic management systems to route optimization algorithms, facilitating congestion reduction and enhancing journey reliability. Overall, this research extends beyond travel time estimation, offering insights into broader transportation planning and policy-making realms, empowering stakeholders to optimize infrastructure utilization and improve network efficiency.Keywords: predictive analytics, traffic flow, travel time estimation, urban transportation, machine learning, traffic management
Procedia PDF Downloads 86