Search results for: Analytic Network Process (ANP)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18936

Search results for: Analytic Network Process (ANP)

15876 Preliminary Study of Desiccant Cooling System under Algerian Climates

Authors: N. Hatraf, N. Moummi

Abstract:

The interest in air conditioning using renewable energies is increasing. The thermal energy produced from the solar energy can be converted to useful cooling and heating through the thermochemical or thermophysical processes by using thermally activated energy conversion systems. The ambient air contains so much water that very high dehumidification rates are required. For a continuous dehumidification of the process air, the water adsorbed on the desiccant material has to be removed, which is done by allowing hot air to flow through the desiccant material (regeneration). A solid desiccant cooling system transfers moisture from the inlet air to the silica gel by using two processes: Absorption process and the regeneration process. The main aim of this paper is to study how the dehumidification rate, the generation temperature and many other factors influence the efficiency of a solid desiccant system by using TRNSYS software. The results show that the desiccant system could be used to decrease the humidity rate of the entering air.

Keywords: dehumidification, efficiency, humidity, Trnsys

Procedia PDF Downloads 423
15875 Comparison of E-learning and Face-to-Face Learning Models Through the Early Design Stage in Architectural Design Education

Authors: Gülay Dalgıç, Gildis Tachir

Abstract:

Architectural design studios are ambiencein where architecture design is realized as a palpable product in architectural education. In the design studios that the architect candidate will use in the design processthe information, the methods of approaching the design problem, the solution proposals, etc., are set uptogetherwith the studio coordinators. The architectural design process, on the other hand, is complex and uncertain.Candidate architects work in a process that starts with abstre and ill-defined problems. This process starts with the generation of alternative solutions with the help of representation tools, continues with the selection of the appropriate/satisfactory solution from these alternatives, and then ends with the creation of an acceptable design/result product. In the studio ambience, many designs and thought relationships are evaluated, the most important step is the early design phase. In the early design phase, the first steps of converting the information are taken, and converted information is used in the constitution of the first design decisions. This phase, which positively affects the progress of the design process and constitution of the final product, is complex and fuzzy than the other phases of the design process. In this context, the aim of the study is to investigate the effects of face-to-face learning model and e-learning model on the early design phase. In the study, the early design phase was defined by literature research. The data of the defined early design phase criteria were obtained with the feedback graphics created for the architect candidates who performed e-learning in the first year of architectural education and continued their education with the face-to-face learning model. The findings of the data were analyzed with the common graphics program. It is thought that this research will contribute to the establishment of a contemporary architectural design education model by reflecting the evaluation of the data and results on architectural education.

Keywords: education modeling, architecture education, design education, design process

Procedia PDF Downloads 126
15874 Modeling of Combustion Process in the Piston Aircraft Engine Using a MCFM-3Z Model

Authors: Marcin Szlachetka, Konrad Pietrykowski

Abstract:

Modeling of a combustion process in a 9-cylinder aircraft engine is presented. The simulations of the combustion process in the IC engine have provided the information on the spatial and time distributions of selected quantities within the combustion chamber of the engine. The numerical analysis results have been compared with the results of indication process of the engine on the test stand. Modeling of combustion process an auto-ignited IC engine in the AVL Fire was carried out within the study. For the calculations, a ECFM-3Z model was used. Verification of simulation results was carried out by comparison of the pressure in the cylinder. The courses of indicated pressure, obtained from the simulations and during the engine tests mounted on a test stand were compared. The engine was braked by the propeller, which results in an adequate external power characteristics. The test object is a modified ASz-62IR engine with the injection system. The engine was running at take-off power. To check the optimum ignition timing regarding power, calculations, tests were performed for 7 different moments of ignition. The analyses of temperature distribution in the cylinder depending on the moments of ignition were carried out. Additional the course of pressure in the cylinder at different angles of ignition delays of the second spark plug were examined. The swirling of the mixture in the combustion chamber was also analysed. It has been shown that the largest vortexes occur in the middle of the chamber, and gets smaller, closer to the combustion chamber walls. This work has been financed by the Polish National Centre for Research and Development, INNOLOT, under Grant Agreement No. INNOLOT/I/1/NCBR/2013.

Keywords: CFD, combustion, internal combustion engine, aircraft engine

Procedia PDF Downloads 361
15873 Advanced Hybrid Particle Swarm Optimization for Congestion and Power Loss Reduction in Distribution Networks with High Distributed Generation Penetration through Network Reconfiguration

Authors: C. Iraklis, G. Evmiridis, A. Iraklis

Abstract:

Renewable energy sources and distributed power generation units already have an important role in electrical power generation. A mixture of different technologies penetrating the electrical grid, adds complexity in the management of distribution networks. High penetration of distributed power generation units creates node over-voltages, huge power losses, unreliable power management, reverse power flow and congestion. This paper presents an optimization algorithm capable of reducing congestion and power losses, both described as a function of weighted sum. Two factors that describe congestion are being proposed. An upgraded selective particle swarm optimization algorithm (SPSO) is used as a solution tool focusing on the technique of network reconfiguration. The upgraded SPSO algorithm is achieved with the addition of a heuristic algorithm specializing in reduction of power losses, with several scenarios being tested. Results show significant improvement in minimization of losses and congestion while achieving very small calculation times.

Keywords: congestion, distribution networks, loss reduction, particle swarm optimization, smart grid

Procedia PDF Downloads 431
15872 Finite Element Simulation of Deep Drawing Process to Minimize Earing

Authors: Pawan S. Nagda, Purnank S. Bhatt, Mit K. Shah

Abstract:

Earing defect in drawing process is highly undesirable not only because it adds on an additional trimming operation but also because the uneven material flow demands extra care. The objective of this work is to study the earing problem in the Deep Drawing of circular cup and to optimize the blank shape to reduce the earing. A finite element model is developed for 3-D numerical simulation of cup forming process in ABAQUS. Extra-deep-drawing (EDD) steel sheet has been used for simulation. Properties and tool design parameters were used as input for simulation. Earing was observed in the simulated cup and it was measured at various angles with respect to rolling direction. To reduce the earing defect initial blank shape was modified with the help of anisotropy coefficient. Modified blanks showed notable reduction in earing.

Keywords: anisotropy, deep drawing, earing, finite element simulation

Procedia PDF Downloads 364
15871 A Mixed Integer Linear Programming Model for Container Collection

Authors: J. Van Engeland, C. Lavigne, S. De Jaeger

Abstract:

In the light of the transition towards a more circular economy, recovery of products, parts or materials will gain in importance. Additionally, the EU proximity principle related to waste management and emissions generated by transporting large amounts of end-of-life products, shift attention to local recovery networks. The Flemish inter-communal cooperation for municipal solid waste management Meetjesland (IVM) is currently investigating the set-up of such a network. More specifically, the network encompasses the recycling of polyvinyl chloride (PVC), which is collected in separate containers. When these containers are full, a truck should transport them to the processor which can recycle the PVC into new products. This paper proposes a model to optimize the container collection. The containers are located at different Civic Amenity sites (CA sites) in a certain region. Since people can drop off their waste at these CA sites, the containers will gradually fill up during a planning horizon. If a certain container is full, it has to be collected and replaced by an empty container. The collected waste is then transported to a single processor. To perform this collection and transportation of containers, the responsible firm has a set of vehicles stationed at a single depot and different personnel crews. A vehicle can load exactly one container. If a trailer is attached to the vehicle, it can load an additional container. Each day of the planning horizon, the different crews and vehicles leave the depot to collect containers at the different sites. After loading one or two containers, the crew has to drive to the processor for unloading the waste and to pick up empty containers. Afterwards, the crew can again visit sites or it can return to the depot to end its collection work for that day. All along the collection process, the crew has to respect the opening hours of the sites. In order to allow for some flexibility, a crew is allowed to wait a certain amount of time at the gate of a site until it opens. The problem described can be modelled as a variant to the PVRP-TW (Periodic Vehicle Routing Problem with Time Windows). However, a vehicle can at maximum load two containers, hence only two subsequent site visits are possible. For that reason, we will refer to the model as a model for building tactical waste collection schemes. The goal is to a find a schedule describing which crew should visit which CA site on which day to minimize the number of trucks and the routing costs. The model was coded in IBM CPLEX Optimization studio and applied to a number of test instances. Good results were obtained, and specific suggestions concerning route and truck costs could be made. For a large range of input parameters, collection schemes using two trucks are obtained.

Keywords: container collection, crew scheduling, mixed integer linear programming, waste management

Procedia PDF Downloads 115
15870 Public Health Emergency Management (PHEM) to COVID-19 Pandemic in North-Eastern Part of Thailand

Authors: Orathai Srithongtham, Ploypailin Mekathepakorn, Tossaphong Buraman, Pontida Moonpradap, Rungrueng Kitpati, Chulapon Kratet, Worayuth Nak-ai, Suwaree Charoenmukkayanan, Peeranuch Keawkanya

Abstract:

The COVID-19 pandemic was effect to the health security of the Thai people. The PHEM principle was essential to the surveillance, prevention, and control of COVID-19. This study aimed to present the process of prevention and control of COVID-19 from February 29, 2021- April 30, 2022, and the factors and conditions influent the successful outcome. The study areas were three provinces. The target group was 37 people, composed of public health personnel. The data was collected in-depth, and group interviews followed the non-structure interview guide and were analyzed by content analysis. The components of COVID-19 prevention and control were found in the process of PHEM as follows; 1) Emergency Operation Center (EOC) with an incidence command system (ICS) from the district to provincial level and to propose the provincial measure, 2) Provincial Communicable Disease Committee (PCDC) to decide the provincial measure 3) The measure for surveillance, prevention, control, and treatment of COVID-19, and 4) outcomes and best practices for surveillance and control of COVID-19. The success factors of 4S and EC were as follows; Space: prepare the quarantine (HQ, LQ), Cohort Ward (CW), field hospital, and community isolation and home isolation to face with the patient and risky group, Staff network from various organization and group cover the community leader and Health Volunteer (HV), Stuff the management and sharing of the medical and non-medical equipment, System of Covid-19 respond were EOC, ICS, Joint Investigation Team (JIT) and Communicable Disease Control Unit (CDCU) for monitoring the real-time of surveillance and control of COVID-19 output, Environment management in hospital and the community with Infections Control (IC) principle, and Culture in term of social capital on “the relationship of Isan people” supported the patient provide the good care and support. The structure of PHEM, Isan’s Culture, and good preparation was a significant factor in the three provinces.

Keywords: public health, emergency management, covid-19, pandemic

Procedia PDF Downloads 73
15869 Machine Learning Techniques for Estimating Ground Motion Parameters

Authors: Farid Khosravikia, Patricia Clayton

Abstract:

The main objective of this study is to evaluate the advantages and disadvantages of various machine learning techniques in forecasting ground-motion intensity measures given source characteristics, source-to-site distance, and local site condition. Intensity measures such as peak ground acceleration and velocity (PGA and PGV, respectively) as well as 5% damped elastic pseudospectral accelerations at different periods (PSA), are indicators of the strength of shaking at the ground surface. Estimating these variables for future earthquake events is a key step in seismic hazard assessment and potentially subsequent risk assessment of different types of structures. Typically, linear regression-based models, with pre-defined equations and coefficients, are used in ground motion prediction. However, due to the restrictions of the linear regression methods, such models may not capture more complex nonlinear behaviors that exist in the data. Thus, this study comparatively investigates potential benefits from employing other machine learning techniques as a statistical method in ground motion prediction such as Artificial Neural Network, Random Forest, and Support Vector Machine. The algorithms are adjusted to quantify event-to-event and site-to-site variability of the ground motions by implementing them as random effects in the proposed models to reduce the aleatory uncertainty. All the algorithms are trained using a selected database of 4,528 ground-motions, including 376 seismic events with magnitude 3 to 5.8, recorded over the hypocentral distance range of 4 to 500 km in Oklahoma, Kansas, and Texas since 2005. The main reason of the considered database stems from the recent increase in the seismicity rate of these states attributed to petroleum production and wastewater disposal activities, which necessities further investigation in the ground motion models developed for these states. Accuracy of the models in predicting intensity measures, generalization capability of the models for future data, as well as usability of the models are discussed in the evaluation process. The results indicate the algorithms satisfy some physically sound characteristics such as magnitude scaling distance dependency without requiring pre-defined equations or coefficients. Moreover, it is shown that, when sufficient data is available, all the alternative algorithms tend to provide more accurate estimates compared to the conventional linear regression-based method, and particularly, Random Forest outperforms the other algorithms. However, the conventional method is a better tool when limited data is available.

Keywords: artificial neural network, ground-motion models, machine learning, random forest, support vector machine

Procedia PDF Downloads 112
15868 Exploring the Activity Fabric of an Intelligent Environment with Hierarchical Hidden Markov Theory

Authors: Chiung-Hui Chen

Abstract:

The Internet of Things (IoT) was designed for widespread convenience. With the smart tag and the sensing network, a large quantity of dynamic information is immediately presented in the IoT. Through the internal communication and interaction, meaningful objects provide real-time services for users. Therefore, the service with appropriate decision-making has become an essential issue. Based on the science of human behavior, this study employed the environment model to record the time sequences and locations of different behaviors and adopted the probability module of the hierarchical Hidden Markov Model for the inference. The statistical analysis was conducted to achieve the following objectives: First, define user behaviors and predict the user behavior routes with the environment model to analyze user purposes. Second, construct the hierarchical Hidden Markov Model according to the logic framework, and establish the sequential intensity among behaviors to get acquainted with the use and activity fabric of the intelligent environment. Third, establish the intensity of the relation between the probability of objects’ being used and the objects. The indicator can describe the possible limitations of the mechanism. As the process is recorded in the information of the system created in this study, these data can be reused to adjust the procedure of intelligent design services.

Keywords: behavior, big data, hierarchical hidden Markov model, intelligent object

Procedia PDF Downloads 221
15867 Online Monitoring and Control of Continuous Mechanosynthesis by UV-Vis Spectrophotometry

Authors: Darren A. Whitaker, Dan Palmer, Jens Wesholowski, James Flaherty, John Mack, Ahmad B. Albadarin, Gavin Walker

Abstract:

Traditional mechanosynthesis has been performed by either ball milling or manual grinding. However, neither of these techniques allow the easy application of process control. The temperature may change unpredictably due to friction in the process. Hence the amount of energy transferred to the reactants is intrinsically non-uniform. Recently, it has been shown that the use of Twin-Screw extrusion (TSE) can overcome these limitations. Additionally, TSE enables a platform for continuous synthesis or manufacturing as it is an open-ended process, with feedstocks at one end and product at the other. Several materials including metal-organic frameworks (MOFs), co-crystals and small organic molecules have been produced mechanochemically using TSE. The described advantages of TSE are offset by drawbacks such as increased process complexity (a large number of process parameters) and variation in feedstock flow impacting on product quality. To handle the above-mentioned drawbacks, this study utilizes UV-Vis spectrophotometry (InSpectroX, ColVisTec) as an online tool to gain real-time information about the quality of the product. Additionally, this is combined with real-time process information in an Advanced Process Control system (PharmaMV, Perceptive Engineering) allowing full supervision and control of the TSE process. Further, by characterizing the dynamic behavior of the TSE, a model predictive controller (MPC) can be employed to ensure the process remains under control when perturbed by external disturbances. Two reactions were studied; a Knoevenagel condensation reaction of barbituric acid and vanillin and, the direct amidation of hydroquinone by ammonium acetate to form N-Acetyl-para-aminophenol (APAP) commonly known as paracetamol. Both reactions could be carried out continuously using TSE, nuclear magnetic resonance (NMR) spectroscopy was used to confirm the percentage conversion of starting materials to product. This information was used to construct partial least squares (PLS) calibration models within the PharmaMV development system, which relates the percent conversion to product to the acquired UV-Vis spectrum. Once this was complete, the model was deployed within the PharmaMV Real-Time System to carry out automated optimization experiments to maximize the percentage conversion based on a set of process parameters in a design of experiments (DoE) style methodology. With the optimum set of process parameters established, a series of PRBS process response tests (i.e. Pseudo-Random Binary Sequences) around the optimum were conducted. The resultant dataset was used to build a statistical model and associated MPC. The controller maximizes product quality whilst ensuring the process remains at the optimum even as disturbances such as raw material variability are introduced into the system. To summarize, a combination of online spectral monitoring and advanced process control was used to develop a robust system for optimization and control of two TSE based mechanosynthetic processes.

Keywords: continuous synthesis, pharmaceutical, spectroscopy, advanced process control

Procedia PDF Downloads 157
15866 The Effect of the Addition of Additives on the Properties of Bisamide Organogels

Authors: Elmira Ghanbari, Jan Van Esch, Stephen J. Picken, Sahil Aggarwal

Abstract:

Organogels are formed by the assembly of low molecular weight gelators (LMWG) into fibrous structures. The assembly of these molecules into crystalline fibrous structures occurs as a result of reversible interactions such as π-stacking, hydrogen-bonding, and van der Waals interactions. Bisamide organogelators with two amide groups have been used as one of LMWGs which show efficient assembly behavior via hydrogen bonding for network formation, the formation of a crystalline network for solvent entrapment. In this study, different bisamide gelators with different lengths of alkyl chains have been added to the bisamide parent gels. The effect of the addition of bisamide additives on the gelation of bisamide gels is described. Investigation of the thermal properties of the gels by differential scanning calorimetry and dropping ball techniques indicated that the bisamide gels can be formed by the addition of a high concentration of the second bisamide components. The microstructure of the gels with different gelator components has been visualized with scanning electron microscopy (SEM) which has shown systematic woven, platelet-like, and a combination of those morphologies for different gels. Examining the addition of a range of bisamide additives with different structural characteristics than the parent bisamide gels has confirmed the effect of the molecular structure on the morphology of the bisamide gels and their final properties.

Keywords: bisamide organogelator additives, gel morphology, gel properties, self-assembly

Procedia PDF Downloads 183
15865 Optimization of Surface Roughness by Taguchi’s Method for Turning Process

Authors: Ashish Ankus Yerunkar, Ravi Terkar

Abstract:

Study aimed at evaluating the best process environment which could simultaneously satisfy requirements of both quality as well as productivity with special emphasis on reduction of cutting tool flank wear, because reduction in flank wear ensures increase in tool life. The predicted optimal setting ensured minimization of surface roughness. Purpose of this paper is focused on the analysis of optimum cutting conditions to get lowest surface roughness in turning SCM 440 alloy steel by Taguchi method. Design for the experiment was done using Taguchi method and 18 experiments were designed by this process and experiments conducted. The results are analyzed using ANOVA method. Taguchi method has depicted that the depth of cut has significant role to play in producing lower surface roughness followed by feed. The Cutting speed has lesser role on surface roughness from the tests. The vibrations of the machine tool, tool chattering are the other factors which may contribute poor surface roughness to the results and such factors ignored for analyses. The inferences by this method will be useful to other researches for similar type of study and may be vital for further research on tool vibrations, cutting forces etc.

Keywords: surface roughness (ra), machining, dry turning, taguchi method, turning process, anova method, mahr perthometer

Procedia PDF Downloads 359
15864 AI Software Algorithms for Drivers Monitoring within Vehicles Traffic - SiaMOTO

Authors: Ioan Corneliu Salisteanu, Valentin Dogaru Ulieru, Mihaita Nicolae Ardeleanu, Alin Pohoata, Bogdan Salisteanu, Stefan Broscareanu

Abstract:

Creating a personalized statistic for an individual within the population using IT systems, based on the searches and intercepted spheres of interest they manifest, is just one 'atom' of the artificial intelligence analysis network. However, having the ability to generate statistics based on individual data intercepted from large demographic areas leads to reasoning like that issued by a human mind with global strategic ambitions. The DiaMOTO device is a technical sensory system that allows the interception of car events caused by a driver, positioning them in time and space. The device's connection to the vehicle allows the creation of a source of data whose analysis can create psychological, behavioural profiles of the drivers involved. The SiaMOTO system collects data from many vehicles equipped with DiaMOTO, driven by many different drivers with a unique fingerprint in their approach to driving. In this paper, we aimed to explain the software infrastructure of the SiaMOTO system, a system designed to monitor and improve driver driving behaviour, as well as the criteria and algorithms underlying the intelligent analysis process.

Keywords: artificial intelligence, data processing, driver behaviour, driver monitoring, SiaMOTO

Procedia PDF Downloads 65
15863 System for the Detecting of Fake Profiles on Online Social Networks Using Machine Learning and the Bio-Inspired Algorithms

Authors: Sekkal Nawel, Mahammed Nadir

Abstract:

The proliferation of online activities on Online Social Networks (OSNs) has captured significant user attention. However, this growth has been hindered by the emergence of fraudulent accounts that do not represent real individuals and violate privacy regulations within social network communities. Consequently, it is imperative to identify and remove these profiles to enhance the security of OSN users. In recent years, researchers have turned to machine learning (ML) to develop strategies and methods to tackle this issue. Numerous studies have been conducted in this field to compare various ML-based techniques. However, the existing literature still lacks a comprehensive examination, especially considering different OSN platforms. Additionally, the utilization of bio-inspired algorithms has been largely overlooked. Our study conducts an extensive comparison analysis of various fake profile detection techniques in online social networks. The results of our study indicate that supervised models, along with other machine learning techniques, as well as unsupervised models, are effective for detecting false profiles in social media. To achieve optimal results, we have incorporated six bio-inspired algorithms to enhance the performance of fake profile identification results.

Keywords: machine learning, bio-inspired algorithm, detection, fake profile, system, social network

Procedia PDF Downloads 50
15862 Study on Inverse Solution from Remote Displacements to Reservoir Process during Flow Injection

Authors: Sumei Cai, Hong Li

Abstract:

Either during water or gas injection into reservoir, in order to understand the areal flow pressure distribution underground, associated bounding deformation is prevalently monitored by ground or downhole tiltmeters. In this paper, an inverse solution to elastic response of far field displacements induced by reservoir pressure change due to flow injection was studied. Furthermore, the fundamental theory on inverse solution to elastic problem as well as its spatial smoothing approach is presented. Taking advantage of source code development based on Boundary Element Method, numerical analysis on the monitoring data of ground surface displacements to further understand the behavior of reservoir process was developed. Numerical examples were also conducted to verify the effectiveness.

Keywords: remote displacement, inverse problem, boundary element method, BEM, reservoir process

Procedia PDF Downloads 104
15861 Cutting Tool-Life Test of Ceramic Insert for Engine Sleeve

Authors: Adam Janásek, Marek Pagáč

Abstract:

The article is looking for an experimental determination of tool life tests for ceramic cutting inserts. Mentioned experimental determination should provide an added information about cutting process. The mechanism of tool wear, cutting temperature in machining, quality machined surface and machining process itself is the information, which are important for whole manufacturing process. Mainly, the roughness plays very important role in determining how a real object will interact with its environment. The main aim was to determine the number of machined inserts, tool life and micro-geometry, as well. On the basis of previous tests the tool-wear was measured at constant cutting parameter which is more typical for high volume manufacturing processes.

Keywords: ceramic, insert, machining, surface roughness, tool-life, tool-wear

Procedia PDF Downloads 476
15860 Electrocardiogram-Based Heartbeat Classification Using Convolutional Neural Networks

Authors: Jacqueline Rose T. Alipo-on, Francesca Isabelle F. Escobar, Myles Joshua T. Tan, Hezerul Abdul Karim, Nouar Al Dahoul

Abstract:

Electrocardiogram (ECG) signal analysis and processing are crucial in the diagnosis of cardiovascular diseases, which are considered one of the leading causes of mortality worldwide. However, the traditional rule-based analysis of large volumes of ECG data is time-consuming, labor-intensive, and prone to human errors. With the advancement of the programming paradigm, algorithms such as machine learning have been increasingly used to perform an analysis of ECG signals. In this paper, various deep learning algorithms were adapted to classify five classes of heartbeat types. The dataset used in this work is the synthetic MIT-BIH Arrhythmia dataset produced from generative adversarial networks (GANs). Various deep learning models such as ResNet-50 convolutional neural network (CNN), 1-D CNN, and long short-term memory (LSTM) were evaluated and compared. ResNet-50 was found to outperform other models in terms of recall and F1 score using a five-fold average score of 98.88% and 98.87%, respectively. 1-D CNN, on the other hand, was found to have the highest average precision of 98.93%.

Keywords: heartbeat classification, convolutional neural network, electrocardiogram signals, generative adversarial networks, long short-term memory, ResNet-50

Procedia PDF Downloads 107
15859 Application of Italian Guidelines for Existing Bridge Management

Authors: Giovanni Menichini, Salvatore Giacomo Morano, Gloria Terenzi, Luca Salvatori, Maurizio Orlando

Abstract:

The “Guidelines for Risk Classification, Safety Assessment, and Structural Health Monitoring of Existing Bridges” were recently approved by the Italian Government to define technical standards for managing the national network of existing bridges. These guidelines provide a framework for risk mitigation and safety assessment of bridges, which are essential elements of the built environment and form the basis for the operation of transport systems. Within the guideline framework, a workflow based on three main points was proposed: (1) risk-based, i.e., based on typical parameters of hazard, vulnerability, and exposure; (2) multi-level, i.e., including six assessment levels of increasing complexity; and (3) multirisk, i.e., assessing structural/foundational, seismic, hydrological, and landslide risks. The paper focuses on applying the Italian Guidelines to specific case studies, aiming to identify the parameters that predominantly influence the determination of the “class of attention”. The significance of each parameter is determined via sensitivity analysis. Additionally, recommendations for enhancing the process of assigning the class of attention are proposed.

Keywords: bridge safety assessment, Italian guidelines implementation, risk classification, structural health monitoring

Procedia PDF Downloads 39
15858 A Hybrid Feature Selection and Deep Learning Algorithm for Cancer Disease Classification

Authors: Niousha Bagheri Khulenjani, Mohammad Saniee Abadeh

Abstract:

Learning from very big datasets is a significant problem for most present data mining and machine learning algorithms. MicroRNA (miRNA) is one of the important big genomic and non-coding datasets presenting the genome sequences. In this paper, a hybrid method for the classification of the miRNA data is proposed. Due to the variety of cancers and high number of genes, analyzing the miRNA dataset has been a challenging problem for researchers. The number of features corresponding to the number of samples is high and the data suffer from being imbalanced. The feature selection method has been used to select features having more ability to distinguish classes and eliminating obscures features. Afterward, a Convolutional Neural Network (CNN) classifier for classification of cancer types is utilized, which employs a Genetic Algorithm to highlight optimized hyper-parameters of CNN. In order to make the process of classification by CNN faster, Graphics Processing Unit (GPU) is recommended for calculating the mathematic equation in a parallel way. The proposed method is tested on a real-world dataset with 8,129 patients, 29 different types of tumors, and 1,046 miRNA biomarkers, taken from The Cancer Genome Atlas (TCGA) database.

Keywords: cancer classification, feature selection, deep learning, genetic algorithm

Procedia PDF Downloads 100
15857 Public Bus Transport Passenger Safety Evaluations in Ghana: A Phenomenological Constructivist Exploration

Authors: Enoch F. Sam, Kris Brijs, Stijn Daniels, Tom Brijs, Geert Wets

Abstract:

Notwithstanding the growing body of literature that recognises the importance of personal safety to public transport (PT) users, it remains unclear what PT users consider regarding their safety. In this study, we explore the criteria PT users in Ghana use to assess bus safety. This knowledge will afford a better understanding of PT users’ risk perceptions and assessments which may contribute to theoretical models of PT risk perceptions. We utilised phenomenological research methodology, with data drawn from 61 purposively sampled participants. Data collection (through focus group discussions and in-depth interviews) and analyses were done concurrently to the point of saturation. Our inductive data coding and analyses through the constant comparison and content analytic techniques resulted in 4 code categories (conceptual dimensions), 27 codes (safety items/criteria), and 100 quotations (data segments). Of the number of safety criteria participants use to assess bus safety, vehicle condition, driver’s marital status, and transport operator’s safety records were the most considered. With each criterion, participants rightly demonstrated its respective relevance to bus safety. These findings imply that investment in and maintenance of safer vehicles, and responsible and safety-conscious drivers, and prioritization of passengers’ safety are key-targets for public bus/minibus operators in Ghana.

Keywords: safety evaluations, public bus/minibus, passengers, phenomenology, Ghana

Procedia PDF Downloads 314
15856 A Wideband CMOS Power Amplifier with 23.3 dB S21, 10.6 dBm Psat and 12.3% PAE for 60 GHz WPAN and 77 GHz Automobile Radar Systems

Authors: Yo-Sheng Lin, Chien-Chin Wang, Yun-Wen Lin, Chien-Yo Lee

Abstract:

A wide band power amplifier (PA) for 60 GHz and 77 GHz direct-conversion transceiver using standard 90 nm CMOS technology is reported. The PA comprises a cascode input stage with a wide band T-type input-matching network and inductive interconnection and load, followed by a common-source (CS) gain stage and a CS output stage. To increase the saturated output power (PSAT) and power-added efficiency (PAE), the output stage adopts a two-way power dividing and combining architecture. Instead of the area-consumed Wilkinson power divider and combiner, miniature low-loss transmission-line inductors are used at the input and output terminals of each of the output stages for wide band input and output impedance matching to 100 ohm. This in turn results in further PSAT and PAE enhancement. The PA consumes 92.2 mW and achieves maximum power gain (S21) of 23.3 dB at 56 GHz, and S21 of 21.7 dB and 14 dB, respectively, at 60 GHz and 77 GHz. In addition, the PA achieves excellent saturated output power (PSAT) of 10.6 dB and maximum power added efficiency (PAE) of 12.3% at 60 GHz. At 77 GHz, the PA achieves excellent PSAT of 10.4 dB and maximum PAE of 6%. These results demonstrate the proposed wide band PA architecture is very promising for 60 GHz wireless personal local network (WPAN) and 77 GHz automobile radar systems.

Keywords: 60 GHz, 77 GHz, PA, WPAN, automotive radar

Procedia PDF Downloads 565
15855 Prediction of Embankment Fires at Railway Infrastructure Using Machine Learning, Geospatial Data and VIIRS Remote Sensing Imagery

Authors: Jan-Peter Mund, Christian Kind

Abstract:

In view of the ongoing climate change and global warming, fires along railways in Germany are occurring more frequently, with sometimes massive consequences for railway operations and affected railroad infrastructure. In the absence of systematic studies within the infrastructure network of German Rail, little is known about the causes of such embankment fires. Since a further increase in these hazards is to be expected in the near future, there is a need for a sound knowledge of triggers and drivers for embankment fires as well as methodical knowledge of prediction tools. Two predictable future trends speak for the increasing relevance of the topic: through the intensification of the use of rail for passenger and freight transport (e.g..: doubling of annual passenger numbers by 2030, compared to 2019), there will be more rail traffic and also more maintenance and construction work on the railways. This research project approach uses satellite data to identify historical embankment fires along rail network infrastructure. The team links data from these fires with infrastructure and weather data and trains a machine-learning model with the aim of predicting fire hazards on sections of the track. Companies reflect on the results and use them on a pilot basis in precautionary measures.

Keywords: embankment fires, railway maintenance, machine learning, remote sensing, VIIRS data

Procedia PDF Downloads 75
15854 Evaluating Reliability Indices in 3 Critical Feeders at Lorestan Electric Power Distribution Company

Authors: Atefeh Pourshafie, Homayoun Bakhtiari

Abstract:

The main task of power distribution companies is to supply the power required by customers in an acceptable level of quality and reliability. Some key performance indicators for electric power distribution companies are those evaluating the continuity of supply within the network. More than other problems, power outages (due to lightning, flood, fire, earthquake, etc.) challenge economy and business. In addition, end users expect a reliable power supply. Reliability indices are evaluated on an annual basis by the specialized holding company of Tavanir (Power Produce, Transmission& distribution company of Iran) . Evaluation of reliability indices is essential for distribution companies, and with regard to the privatization of distribution companies, it will be of particular importance to evaluate these indices and to plan for their improvement in a not too distant future. According to IEEE-1366 standard, there are too many indices; however, the most common reliability indices include SAIFI, SAIDI and CAIDI. These indices describe the period and frequency of blackouts in the reporting period (annual or any desired timeframe). This paper calculates reliability indices for three sample feeders in Lorestan Electric Power Distribution Company and defines the threshold values in a ten-month period. At the end, strategies are introduced to reach the threshold values in order to increase customers' satisfaction.

Keywords: power, distribution network, reliability, outage

Procedia PDF Downloads 456
15853 Synthesis and Electromagnetic Wave Absorbing Property of Amorphous Carbon Nanotube Networks on a 3D Graphene Aerogel/BaFe₁₂O₁₉ Nanorod Composite

Authors: Tingkai Zhao, Jingtian Hu, Xiarong Peng, Wenbo Yang, Tiehu Li

Abstract:

Homogeneous amorphous carbon nanotube (ACNT) networks have been synthesized using floating catalyst chemical vapor deposition method on a three-dimensional (3D) graphene aerogel (GA)/BaFe₁₂O₁₉ nanorod (BNR) composite which prepared by a self-propagating combustion process. The as-synthesized ACNT/GA/BNR composite which has 3D network structures could be directly used as a good absorber in the electromagnetic wave absorbent materials. The experimental results indicated that the maximum absorbing peak of ACNT/GA/BNR composite with a thickness of 2 mm was -18.35 dB at 10.64 GHz in the frequency range of 2-18 GHz. The bandwidth of the reflectivity below -10 dB is 3.32 GHz. The 3D graphene aerogel structures which composed of dense interlined tubes and amorphous structure of ACNTs bearing quantities of dihedral angles could consume the incident waves through multiple reflection and scattering inside the 3D web structures. The interlinked ACNTs have both the virtues of amorphous CNTs (multiple reflections inside the wall) and crystalline CNTs (high conductivity), consuming the electromagnetic wave as resistance heat. ACNT/GA/BNR composite has a good electromagnetic wave absorbing performance.

Keywords: amorphous carbon nanotubes, graphene aerogel, barium ferrite nanorod, electromagnetic wave absorption

Procedia PDF Downloads 267
15852 Disaster Resilience Analysis of Atlanta Interstate Highway System within the Perimeter

Authors: Mengmeng Liu, J. David Frost

Abstract:

Interstate highway system within the Atlanta Perimeter plays an important role in residents’ daily life. The serious influence of Atlanta I-85 Collapses implies that transportation system in the region lacks a cohesive and comprehensive transportation plan. Therefore, disaster resilience analysis of the transportation system is necessary. Resilience is the system’s capability to persist or to maintain transportation services when exposed to changes or shocks. This paper analyzed the resilience of the whole transportation system within the Perimeter and see how removing interstates within the Perimeter will affect the resilience of the transportation system. The data used in the paper are Atlanta transportation networks and LEHD Origin-Destination Employment Statistics data. First, we calculate the traffic flow on each road section based on LEHD data assuming each trip travel along the shortest travel time paths. Second, we calculate the measure of resilience, which is flow-based connectivity and centrality of the transportation network, and see how they will change if we remove each section of interstates from the current transportation system. Finally, we get the resilience function curve of the interstates and identify the most resilient interstates section. The resilience analysis results show that the framework of calculation resilience is effective and can provide some useful information for the transportation planning and sustainability analysis of the transportation infrastructures.

Keywords: connectivity, interstate highway system, network analysis, resilience analysis

Procedia PDF Downloads 240
15851 Anaerobic Digestion Batch Study of Taxonomic Variations in Microbial Communities during Adaptation of Consortium to Different Lignocellulosic Substrates Using Targeted Sequencing

Authors: Priyanka Dargode, Suhas Gore, Manju Sharma, Arvind Lali

Abstract:

Anaerobic digestion has been widely used for production of methane from different biowastes. However, the complexity of microbial communities involved in the process is poorly understood. The performance of biogas production process concerning the process productivity is closely coupled to its microbial community structure and syntrophic interactions amongst the community members. The present study aims at understanding taxonomic variations occurring in any starter inoculum when acclimatised to different lignocellulosic biomass (LBM) feedstocks relating to time of digestion. The work underlines use of high throughput Next Generation Sequencing (NGS) for validating the changes in taxonomic patterns of microbial communities. Biomethane Potential (BMP) batches were set up with different pretreated and non-pretreated LBM residues using the same microbial consortium and samples were withdrawn for studying the changes in microbial community in terms of its structure and predominance with respect to changes in metabolic profile of the process. DNA of samples withdrawn at different time intervals with reference to performance changes of the digestion process, was extracted followed by its 16S rRNA amplicon sequencing analysis using Illumina Platform. Biomethane potential and substrate consumption was monitored using Gas Chromatography(GC) and reduction in COD (Chemical Oxygen Demand) respectively. Taxonomic analysis by QIIME server data revealed that microbial community structure changes with different substrates as well as at different time intervals. It was observed that biomethane potential of each substrate was relatively similar but, the time required for substrate utilization and its conversion to biomethane was different for different substrates. This could be attributed to the nature of substrate and consequently the discrepancy between the dominance of microbial communities with regards to different substrate and at different phases of anaerobic digestion process. Knowledge of microbial communities involved would allow a rational substrate specific consortium design which will help to reduce consortium adaptation period and enhance the substrate utilisation resulting in improved efficacy of biogas process.

Keywords: amplicon sequencing, biomethane potential, community predominance, taxonomic analysis

Procedia PDF Downloads 510
15850 Mixed Sub-Fractional Brownian Motion

Authors: Mounir Zili

Abstract:

We will introduce a new extension of the Brownian motion, that could serve to get a good model of many natural phenomena. It is a linear combination of a finite number of sub-fractional Brownian motions; that is why we will call it the mixed sub-fractional Brownian motion. We will present some basic properties of this process. Among others, we will check that our process is non-Markovian and that it has non-stationary increments. We will also give the conditions under which it is a semimartingale. Finally, the main features of its sample paths will be specified.

Keywords: mixed Gaussian processes, Sub-fractional Brownian motion, sample paths

Procedia PDF Downloads 472
15849 Large Scale Production of Polyhydroxyalkanoates (PHAs) from Waste Water: A Study of Techno-Economics, Energy Use, and Greenhouse Gas Emissions

Authors: Cora Fernandez Dacosta, John A. Posada, Andrea Ramirez

Abstract:

The biodegradable family of polymers polyhydroxyalkanoates are interesting substitutes for convectional fossil-based plastics. However, the manufacturing and environmental impacts associated with their production via intracellular bacterial fermentation are strongly dependent on the raw material used and on energy consumption during the extraction process, limiting their potential for commercialization. Industrial wastewater is studied in this paper as a promising alternative feedstock for waste valorization. Based on results from laboratory and pilot-scale experiments, a conceptual process design, techno-economic analysis and life cycle assessment are developed for the large-scale production of the most common type of polyhydroxyalkanoate, polyhydroxbutyrate. Intracellular polyhydroxybutyrate is obtained via fermentation of microbial community present in industrial wastewater and the downstream processing is based on chemical digestion with surfactant and hypochlorite. The economic potential and environmental performance results help identifying bottlenecks and best opportunities to scale-up the process prior to industrial implementation. The outcome of this research indicates that the fermentation of wastewater towards PHB presents advantages compared to traditional PHAs production from sugars because the null environmental burdens and financial costs of the raw material in the bioplastic production process. Nevertheless, process optimization is still required to compete with the petrochemicals counterparts.

Keywords: circular economy, life cycle assessment, polyhydroxyalkanoates, waste valorization

Procedia PDF Downloads 438
15848 Planning a Haemodialysis Process by Minimum Time Control of Hybrid Systems with Sliding Motion

Authors: Radoslaw Pytlak, Damian Suski

Abstract:

The aim of the paper is to provide a computational tool for planning a haemodialysis process. It is shown that optimization methods can be used to obtain the most effective treatment focused on removing both urea and phosphorus during the process. In order to achieve that, the IV–compartment model of phosphorus kinetics is applied. This kinetics model takes into account a rebound phenomenon that can occur during haemodialysis and results in a hybrid model of the process. Furthermore, vector fields associated with the model equations are such that it is very likely that using the most intuitive objective functions in the planning problem could lead to solutions which include sliding motions. Therefore, building computational tools for solving the problem of planning a haemodialysis process has required constructing numerical algorithms for solving optimal control problems with hybrid systems. The paper concentrates on minimum time control of hybrid systems since this control objective is the most suitable for the haemodialysis process considered in the paper. The presented approach to optimal control problems with hybrid systems is different from the others in several aspects. First of all, it is assumed that a hybrid system can exhibit sliding modes. Secondly, the system’s motion on the switching surface is described by index 2 differential–algebraic equations, and that guarantees accurate tracking of the sliding motion surface. Thirdly, the gradients of the problem’s functionals are evaluated with the help of adjoint equations. The adjoint equations presented in the paper take into account sliding motion and exhibit jump conditions at transition times. The optimality conditions in the form of the weak maximum principle for optimal control problems with hybrid systems exhibiting sliding modes and with piecewise constant controls are stated. The presented sensitivity analysis can be used to construct globally convergent algorithms for solving considered problems. The paper presents numerical results of solving the haemodialysis planning problem.

Keywords: haemodialysis planning process, hybrid systems, optimal control, sliding motion

Procedia PDF Downloads 183
15847 Crow Search Algorithm-Based Task Offloading Strategies for Fog Computing Architectures

Authors: Aniket Ganvir, Ritarani Sahu, Suchismita Chinara

Abstract:

The rapid digitization of various aspects of life is leading to the creation of smart IoT ecosystems, where interconnected devices generate significant amounts of valuable data. However, these IoT devices face constraints such as limited computational resources and bandwidth. Cloud computing emerges as a solution by offering ample resources for offloading tasks efficiently despite introducing latency issues, especially for time-sensitive applications like fog computing. Fog computing (FC) addresses latency concerns by bringing computation and storage closer to the network edge, minimizing data travel distance, and enhancing efficiency. Offloading tasks to fog nodes or the cloud can conserve energy and extend IoT device lifespan. The offloading process is intricate, with tasks categorized as full or partial, and its optimization presents an NP-hard problem. Traditional greedy search methods struggle to address the complexity of task offloading efficiently. To overcome this, the efficient crow search algorithm (ECSA) has been proposed as a meta-heuristic optimization algorithm. ECSA aims to effectively optimize computation offloading, providing solutions to this challenging problem.

Keywords: IoT, fog computing, task offloading, efficient crow search algorithm

Procedia PDF Downloads 33