Search results for: neural simulated annealing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3620

Search results for: neural simulated annealing

470 Poly(L-Lactic Acid) Scaffolds for Bone Tissue Engineering

Authors: Aleksandra BužArovska, Gordana Bogoeva Gaceva

Abstract:

Biodegradable polymers have received significant scientific attention in tissue engineering (TE) application, in particular their composites consisting of inorganic nanoparticles. In the last 15 years, they are subject of intensive research by many groups, aiming to develop polymer scaffolds with defined biodegradability, porosity and adequate mechanical stability. The most important characteristic making these materials attractive for TE is their biodegradability, a process that could be time controlled and long enough to enable generation of a new tissue as a replacement for the degraded polymer scaffold. In this work poly(L-lactic acid) scaffolds, filled with TiO2 nanoparticles functionalized with oleic acid, have been prepared by thermally induced phase separation method (TIPS). The functionalization of TiO2 nanoparticles with oleic acid was performed in order to improve the nanoparticles dispersibility within the polymer matrix and at the same time to inhibit the cytotoxicity of the nanofiller. The oleic acid was chosen as amphiphilic molecule belonging to the fatty acid family because of its non-toxicity and possibility for mediation between the hydrophilic TiO2 nanoparticles and hydrophobic PLA matrix. The produced scaffolds were characterized with thermogravimetric analysis (TGA), differential scanning calorimetry (DSC), scanning electron microscopy (SEM) and mechanical compression measurements. The bioactivity for bone tissue engineering application was tested in supersaturated simulated body fluid. The degradation process was followed by Fourier transform infrared spectroscopy (FTIR). The results showed anisotropic morphology with elongated open pores (100 µm), high porosity (around 92%) and perfectly dispersed nanofiller. The compression moduli up to 10 MPa were identified independent on the nanofiller content. Functionalized TiO2 nanoparticles induced formation of hydroxyapatite clusters as much as unfunctionalized TiO2. The prepared scaffolds showed properties ideal for scaffold vascularization, cell attachment, growth and proliferation.

Keywords: biodegradation, bone tissue engineering, mineralization, PLA scaffolds

Procedia PDF Downloads 267
469 Dynamic Model for Forecasting Rainfall Induced Landslides

Authors: R. Premasiri, W. A. H. A. Abeygunasekara, S. M. Hewavidana, T. Jananthan, R. M. S. Madawala, K. Vaheeshan

Abstract:

Forecasting the potential for disastrous events such as landslides has become one of the major necessities in the current world. Most of all, the landslides occurred in Sri Lanka are found to be triggered mostly by intense rainfall events. The study area is the landslide near Gerandiella waterfall which is located by the 41st kilometer post on Nuwara Eliya-Gampala main road in Kotmale Division in Sri Lanka. The landslide endangers the entire Kotmale town beneath the slope. Geographic Information System (GIS) platform is very much useful when it comes to the need of emulating the real-world processes. The models are used in a wide array of applications ranging from simple evaluations to the levels of forecast future events. This project investigates the possibility of developing a dynamic model to map the spatial distribution of the slope stability. The model incorporates several theoretical models including the infinite slope model, Green Ampt infiltration model and Perched ground water flow model. A series of rainfall values can be fed to the model as the main input to simulate the dynamics of slope stability. Hydrological model developed using GIS is used to quantify the perched water table height, which is one of the most critical parameters affecting the slope stability. Infinite slope stability model is used to quantify the degree of slope stability in terms of factor of safety. DEM was built with the use of digitized contour data. Stratigraphy was modeled in Surfer using borehole data and resistivity images. Data available from rainfall gauges and piezometers were used in calibrating the model. During the calibration, the parameters were adjusted until a good fit between the simulated ground water levels and the piezometer readings was obtained. This model equipped with the predicted rainfall values can be used to forecast of the slope dynamics of the area of interest. Therefore it can be investigated the slope stability of rainfall induced landslides by adjusting temporal dimensions.

Keywords: factor of safety, geographic information system, hydrological model, slope stability

Procedia PDF Downloads 420
468 AI Peer Review Challenge: Standard Model of Physics vs 4D GEM EOS

Authors: David A. Harness

Abstract:

Natural evolution of ATP cognitive systems is to meet AI peer review standards. ATP process of axiom selection from Mizar to prove a conjecture would be further refined, as in all human and machine learning, by solving the real world problem of the proposed AI peer review challenge: Determine which conjecture forms the higher confidence level constructive proof between Standard Model of Physics SU(n) lattice gauge group operation vs. present non-standard 4D GEM EOS SU(n) lattice gauge group spatially extended operation in which the photon and electron are the first two trace angular momentum invariants of a gravitoelectromagnetic (GEM) energy momentum density tensor wavetrain integration spin-stress pressure-volume equation of state (EOS), initiated via 32 lines of Mathematica code. Resulting gravitoelectromagnetic spectrum ranges from compressive through rarefactive of the central cosmological constant vacuum energy density in units of pascals. Said self-adjoint group operation exclusively operates on the stress energy momentum tensor of the Einstein field equations, introducing quantization directly on the 4D spacetime level, essentially reformulating the Yang-Mills virtual superpositioned particle compounded lattice gauge groups quantization of the vacuum—into a single hyper-complex multi-valued GEM U(1) × SU(1,3) lattice gauge group Planck spacetime mesh quantization of the vacuum. Thus the Mizar corpus already contains all of the axioms required for relevant DeepMath premise selection and unambiguous formal natural language parsing in context deep learning.

Keywords: automated theorem proving, constructive quantum field theory, information theory, neural networks

Procedia PDF Downloads 177
467 Characterization and Correlation of Neurodegeneration and Biological Markers of Model Mice with Traumatic Brain Injury and Alzheimer's Disease

Authors: J. DeBoard, R. Dietrich, J. Hughes, K. Yurko, G. Harms

Abstract:

Alzheimer’s disease (AD) is a predominant type of dementia and is likely a major cause of neural network impairment. The pathogenesis of this neurodegenerative disorder has yet to be fully elucidated. There are currently no known cures for the disease, and the best hope is to be able to detect it early enough to impede its progress. Beyond age and genetics, another prevalent risk factor for AD might be traumatic brain injury (TBI), which has similar neurodegenerative hallmarks. Our research focuses on obtaining information and methods to be able to predict when neurodegenerative effects might occur at a clinical level by observation of events at a cellular and molecular level in model mice. First, we wish to introduce our evidence that brain damage can be observed via brain imaging prior to the noticeable loss of neuromuscular control in model mice of AD. We then show our evidence that some blood biomarkers might be able to be early predictors of AD in the same model mice. Thus, we were interested to see if we might be able to predict which mice might show long-term neurodegenerative effects due to differing degrees of TBI and what level of TBI causes further damage and earlier death to the AD model mice. Upon application of TBIs via an apparatus to effectively induce extremely mild to mild TBIs, wild-type (WT) mice and AD mouse models were tested for cognition, neuromuscular control, olfactory ability, blood biomarkers, and brain imaging. Experiments are currently still in process, and more results are therefore forthcoming. Preliminary data suggest that neuromotor control diminishes as well as olfactory function for both AD and WT mice after the administration of five consecutive mild TBIs. Also, seizure activity increases significantly for both AD and WT after the administration of the five TBI treatment. If future data supports these findings, important implications about the effect of TBI on those at risk for AD might be possible.

Keywords: Alzheimer's disease, blood biomarker, neurodegeneration, neuromuscular control, olfaction, traumatic brain injury

Procedia PDF Downloads 140
466 Investigating the Steam Generation Potential of Lithium Bromide Based CuO Nanofluid under Simulated Solar Flux

Authors: Tamseela Habib, Muhammad Amjad, Muhammad Edokali, Masome Moeni, Olivia Pickup, Ali Hassanpour

Abstract:

Nanofluid-assisted steam generation is rapidly attracting attention amongst the scientific community since it can be applied in a wide range of industrial processes. Because of its high absorption rate of solar energy, nanoparticle-based solar steam generation could be a major contributor to many applications, including water desalination, sterilization and power generation. Lithium bromide-based iron oxide nanofluids have been previously studied in steam generation, which showed promising results. However, the efficiency of the system could be improved if a more heat-conductive nanofluid system could be utilised. In the current paper, we report on an experimental investigation of the photothermal conversion properties of functionalised Copper oxide (CuO) nanoparticles used in Lithium Bromide salt solutions. CuO binary nanofluid was prepared by chemical functionalization with polyethyleneimine (PEI). Long-term stability evaluation of prepared binary nanofluid was done by a high-speed centrifuge analyser which showed a 0.06 Instability index suggesting low agglomeration and sedimentation tendencies. This stability is also supported by the measurements from dynamic light scattering (DLS), transmission electron microscope (TEM), and ultraviolet-visible (UV-Vis) spectrophotometer. The fluid rheology is also characterised, which suggests the system exhibits a Newtonian fluid behavior. The photothermal conversion efficiency of different concentrations of CuO was experimentally investigated under a solar simulator. Experimental results reveal that the binary nanofluid in this study can remarkably increase the solar energy trapping efficiency and evaporation rate as compared to conventional fluids due to localized solar energy harvesting by the surface of the nanofluid. It was found that 0.1wt% CuO NP is the optimum nanofluid concentration for enhanced sensible and latent heat efficiencies.

Keywords: nanofluids, vapor absorption refrigeration system, steam generation, high salinity

Procedia PDF Downloads 82
465 Hybridization of Manually Extracted and Convolutional Features for Classification of Chest X-Ray of COVID-19

Authors: M. Bilal Ishfaq, Adnan N. Qureshi

Abstract:

COVID-19 is the most infectious disease these days, it was first reported in Wuhan, the capital city of Hubei in China then it spread rapidly throughout the whole world. Later on 11 March 2020, the World Health Organisation (WHO) declared it a pandemic. Since COVID-19 is highly contagious, it has affected approximately 219M people worldwide and caused 4.55M deaths. It has brought the importance of accurate diagnosis of respiratory diseases such as pneumonia and COVID-19 to the forefront. In this paper, we propose a hybrid approach for the automated detection of COVID-19 using medical imaging. We have presented the hybridization of manually extracted and convolutional features. Our approach combines Haralick texture features and convolutional features extracted from chest X-rays and CT scans. We also employ a minimum redundancy maximum relevance (MRMR) feature selection algorithm to reduce computational complexity and enhance classification performance. The proposed model is evaluated on four publicly available datasets, including Chest X-ray Pneumonia, COVID-19 Pneumonia, COVID-19 CTMaster, and VinBig data. The results demonstrate high accuracy and effectiveness, with 0.9925 on the Chest X-ray pneumonia dataset, 0.9895 on the COVID-19, Pneumonia and Normal Chest X-ray dataset, 0.9806 on the Covid CTMaster dataset, and 0.9398 on the VinBig dataset. We further evaluate the effectiveness of the proposed model using ROC curves, where the AUC for the best-performing model reaches 0.96. Our proposed model provides a promising tool for the early detection and accurate diagnosis of COVID-19, which can assist healthcare professionals in making informed treatment decisions and improving patient outcomes. The results of the proposed model are quite plausible and the system can be deployed in a clinical or research setting to assist in the diagnosis of COVID-19.

Keywords: COVID-19, feature engineering, artificial neural networks, radiology images

Procedia PDF Downloads 74
464 Applying Multiplicative Weight Update to Skin Cancer Classifiers

Authors: Animish Jain

Abstract:

This study deals with using Multiplicative Weight Update within artificial intelligence and machine learning to create models that can diagnose skin cancer using microscopic images of cancer samples. In this study, the multiplicative weight update method is used to take the predictions of multiple models to try and acquire more accurate results. Logistic Regression, Convolutional Neural Network (CNN), and Support Vector Machine Classifier (SVMC) models are employed within the Multiplicative Weight Update system. These models are trained on pictures of skin cancer from the ISIC-Archive, to look for patterns to label unseen scans as either benign or malignant. These models are utilized in a multiplicative weight update algorithm which takes into account the precision and accuracy of each model through each successive guess to apply weights to their guess. These guesses and weights are then analyzed together to try and obtain the correct predictions. The research hypothesis for this study stated that there would be a significant difference in the accuracy of the three models and the Multiplicative Weight Update system. The SVMC model had an accuracy of 77.88%. The CNN model had an accuracy of 85.30%. The Logistic Regression model had an accuracy of 79.09%. Using Multiplicative Weight Update, the algorithm received an accuracy of 72.27%. The final conclusion that was drawn was that there was a significant difference in the accuracy of the three models and the Multiplicative Weight Update system. The conclusion was made that using a CNN model would be the best option for this problem rather than a Multiplicative Weight Update system. This is due to the possibility that Multiplicative Weight Update is not effective in a binary setting where there are only two possible classifications. In a categorical setting with multiple classes and groupings, a Multiplicative Weight Update system might become more proficient as it takes into account the strengths of multiple different models to classify images into multiple categories rather than only two categories, as shown in this study. This experimentation and computer science project can help to create better algorithms and models for the future of artificial intelligence in the medical imaging field.

Keywords: artificial intelligence, machine learning, multiplicative weight update, skin cancer

Procedia PDF Downloads 75
463 Calibration and Validation of ArcSWAT Model for Estimation of Surface Runoff and Sediment Yield from Dhangaon Watershed

Authors: M. P. Tripathi, Priti Tiwari

Abstract:

Soil and Water Assessment Tool (SWAT) is a distributed parameter continuous time model and was tested on daily and fortnightly basis for a small agricultural watershed (Dhangaon) of Chhattisgarh state in India. The SWAT model recently interfaced with ArcGIS and called as ArcSWAT. The watershed and sub-watershed boundaries, drainage networks, slope and texture maps were generated in the environment of ArcGIS of ArcSWAT. Supervised classification method was used for land use/cover classification from satellite imageries of the years 2009 and 2012. Manning's roughness coefficient 'n' for overland flow and channel flow and Fraction of Field Capacity (FFC) were calibrated for monsoon season of the years 2009 and 2010. The model was validated on a daily basis for the years 2011 and 2012 by using the observed daily rainfall and temperature data. Calibration and validation results revealed that the model was predicting the daily surface runoff and sediment yield satisfactorily. Sensitivity analysis showed that the annual sediment yield was inversely proportional to the overland and channel 'n' values whereas; annual runoff and sediment yields were directly proportional to the FFC. The model was also tested (calibrated and validated) for the fortnightly runoff and sediment yield for the year 2009-10 and 2011-12, respectively. Simulated values of fortnightly runoff and sediment yield for the calibration and validation years compared well with their observed counterparts. The calibration and validation results revealed that the ArcSWAT model could be used for identification of critical sub-watershed and for developing management scenarios for the Dhangaon watershed. Further, the model should be tested for simulating the surface runoff and sediment yield using generated rainfall and temperature before applying it for developing the management scenario for the critical or priority sub-watersheds.

Keywords: watershed, hydrologic and water quality, ArcSWAT model, remote sensing, GIS, runoff and sediment yield

Procedia PDF Downloads 376
462 Bimetallic MOFs Based Membrane for the Removal of Heavy Metal Ions from the Industrial Wastewater

Authors: Muhammad Umar Mushtaq, Muhammad Bilal Khan Niazi, Nouman Ahmad, Dooa Arif

Abstract:

Apart from organic dyes, heavy metals such as Pb, Ni, Cr, and Cu are present in textile effluent and pose a threat to humans and the environment. Many studies on removing heavy metallic ions from textile wastewater have been conducted in recent decades using metal-organic frameworks (MOFs). In this study new polyether sulfone ultrafiltration membrane, modified with Cu/Co and Cu/Zn-based bimetal-organic frameworks (MOFs), was produced. Phase inversion was used to produce the membrane, and atomic force microscopy (AFM), scanning electron microscopy (SEM) were used to characterize it. The bimetallic MOFs-based membrane structure is complex and can be comprehended using characterization techniques. The bimetallic MOF-based filtration membranes are designed to selectively adsorb specific contaminants while allowing the passage of water molecules, improving the ultrafiltration efficiency. MOFs' adsorption capacity and selectivity are enhanced by functionalizing them with particular chemical groups or incorporating them into composite membranes with other materials, such as polymers. The morphology and performance of the bimetallic MOF-based membrane were investigated regarding pure water flux and metal ion rejection. The advantages of developed bimetallic MOFs based membranes for wastewater treatment include enhanced adsorption capacity because of the presence of two metals in their structure, which provides additional binding sites for contaminants, leading to a higher adsorption capacity and more efficient removal of pollutants from wastewater. Based on the experimental findings, bimetallic MOF-based membranes are more capable of rejecting metal ions from industrial wastewater than conventional membranes that have already been developed. Furthermore, the difficulties associated with operational parameters, including pressure gradients and velocity profiles, are simulated using Ansys Fluent software. The simulation results obtained for the operating parameters are in complete agreement with the experimental results.

Keywords: bimetallic MOFs, heavy metal ions, industrial wastewater treatment, ultrafiltration.

Procedia PDF Downloads 88
461 Policy Recommendations for Reducing CO2 Emissions in Kenya's Electricity Generation, 2015-2030

Authors: Paul Kipchumba

Abstract:

Kenya is an East African Country lying at the Equator. It had a population of 46 million in 2015 with an annual growth rate of 2.7%, making a population of at least 65 million in 2030. Kenya’s GDP in 2015 was about 63 billion USD with per capita GDP of about 1400 USD. The rural population is 74%, whereas urban population is 26%. Kenya grapples with not only access to energy but also with energy security. There is direct correlation between economic growth, population growth, and energy consumption. Kenya’s energy composition is at least 74.5% from renewable energy with hydro power and geothermal forming the bulk of it; 68% from wood fuel; 22% from petroleum; 9% from electricity; and 1% from coal and other sources. Wood fuel is used by majority of rural and poor urban population. Electricity is mostly used for lighting. As of March 2015 Kenya had installed electricity capacity of 2295 MW, making a per capital electricity consumption of 0.0499 KW. The overall retail cost of electricity in 2015 was 0.009915 USD/ KWh (KES 19.85/ KWh), for installed capacity over 10MW. The actual demand for electricity in 2015 was 3400 MW and the projected demand in 2030 is 18000 MW. Kenya is working on vision 2030 that aims at making it a prosperous middle income economy and targets 23 GW of generated electricity. However, cost and non-cost factors affect generation and consumption of electricity in Kenya. Kenya does not care more about CO2 emissions than on economic growth. Carbon emissions are most likely to be paid by future costs of carbon emissions and penalties imposed on local generating companies by sheer disregard of international law on C02 emissions and climate change. The study methodology was a simulated application of carbon tax on all carbon emitting sources of electricity generation. It should cost only USD 30/tCO2 tax on all emitting sources of electricity generation to have solar as the only source of electricity generation in Kenya. The country has the best evenly distributed global horizontal irradiation. Solar potential after accounting for technology efficiencies such as 14-16% for solar PV and 15-22% for solar thermal is 143.94 GW. Therefore, the paper recommends adoption of solar power for generating all electricity in Kenya in order to attain zero carbon electricity generation in the country.

Keywords: co2 emissions, cost factors, electricity generation, non-cost factors

Procedia PDF Downloads 361
460 Modeling of Void Formation in 3D Woven Fabric During Resin Transfer Moulding

Authors: Debabrata Adhikari, Mikhail Matveev, Louise Brown, Jan Kočí, Andy Long

Abstract:

Resin transfer molding (RTM) is increasingly used for manufacturing high-quality composite structures due to its additional advantages over prepregs of low-cost out-of-autoclave processing. However, to retain the advantages, it is critical to reduce the void content during the injection. Reinforcements commonly used in RTM, such as woven fabrics, have dual-scale porosity with mesoscale pores between the yarns and the micro-scale pores within the yarns. Due to the fabric geometry and the nature of the dual-scale flow, the flow front during injection creates a complicated fingering formation which leads to void formation. Analytical modeling of void formation for woven fabrics has been widely studied elsewhere. However, there is scope for improvement to the reduction in void formation in 3D fabrics wherein the in-plane yarn layers are confined by additional through-thickness binder yarns. In the present study, the structural morphology of the tortuous pore spaces in the 3D fabric has been studied and implemented using open-source software TexGen. An analytical model for the void and the fingering formation has been implemented based on an idealized unit cell model of the 3D fabric. Since the pore spaces between the yarns are free domains, the region is treated as flow-through connected channels, whereas intra-yarn flow has been modeled using Darcy’s law with an additional term to account for capillary pressure. Later the void fraction has been characterised using the criterion of void formation by comparing the fill time for inter and intra yarn flow. Moreover, the dual-scale two-phase flow of resin with air has been simulated in the commercial CFD solver OpenFOAM/ANSYS to predict the probable location of voids and validate the analytical model. The use of an idealised unit cell model will give the insight to optimise the mesoscale geometry of the reinforcement and injection parameters to minimise the void content during the LCM process.

Keywords: 3D fiber, void formation, RTM, process modelling

Procedia PDF Downloads 93
459 A Method for Solid-Liquid Separation of Cs+ from Radioactive Waste by Using Ionic Liquids and Extractants

Authors: J. W. Choi, S. Y. Cho, H. J. Lee, W. Z. Oh, S. J. Choi

Abstract:

Ionic liquids (ILs), which is alternative to conventional organic solvent, were used for extraction of Cs ions. ILs, as useful environment friendly green solvents, have been recently applied as replacement for traditional volatile organic compounds (VOCs) in liquid/liquid extraction of heavy metal ions as well as organic and inorganic species and pollutants. Thus, Ionic liquids were used for extraction of Cs ions from the liquid radioactive waste. In most cases, Cs ions present in radioactive wastes in very low concentration, approximately less than 1ppm. Therefore, unlike established extraction system the required amount of ILs as extractant is comparatively very small. This extraction method involves cation exchange mechanism in which Cs ion transfers to the organic phase and binds to one crown ether by chelation in exchange of single ILs cation, IL_cation+, transfer to the aqueous phase. In this extraction system showed solid-liquid separation in which the Ionic liquid 1-ethyl-3-methylimidazolium bis(trifluoromethylsulfonly)imide (C2mimTf2N) and the crown ether Dicyclohexano-18-crown-6 (DCH18C6) both were used here in very little amount as solvent and as extractant, respectively. 30 mM of CsNO3 was used as simulated waste solution cesium ions. Generally, in liquid-liquid extraction, the molar ratio of CE:Cs+:ILs was 1:5~10:>100, while our applied molar ratio of CE:Cs+:ILs was 1:2:1~10. The quantity of CE and Cs ions were fixed to 0.6 and 1.2 mmol, respectively. The phenomenon of precipitation showed two kinds of separation: solid-liquid separation in the ratio of 1:2:1 and 1:2:2; solid-liquid-liquid separation (3 phase) in the ratio of 1:2:5 and 1:2:10. In the last system, 3 phases were precipitate-ionic liquids-aqueous. The precipitate was verified to consist of Cs+, DCH18C6, Tf2N- based on the cation exchange mechanism. We analyzed precipitate using scanning electron microscopy with X-ray microanalysis (SEM-EDS), an elemental analyser, Fourier transform infrared spectroscopy (FT-IR) and differential scanning calorimetry (DSC). The experimental results showed an easy extraction method and confirmed the composition of solid precipitate. We also obtained information that complex formation ratio of Cs+ to DCH18C6 is 0.88:1 regardless of C2mimTf2N quantities.

Keywords: extraction, precipitation, solid-liquid seperation, ionic liquid, precipitate

Procedia PDF Downloads 414
458 Use Cloud-Based Watson Deep Learning Platform to Train Models Faster and More Accurate

Authors: Susan Diamond

Abstract:

Machine Learning workloads have traditionally been run in high-performance computing (HPC) environments, where users log in to dedicated machines and utilize the attached GPUs to run training jobs on huge datasets. Training of large neural network models is very resource intensive, and even after exploiting parallelism and accelerators such as GPUs, a single training job can still take days. Consequently, the cost of hardware is a barrier to entry. Even when upfront cost is not a concern, the lead time to set up such an HPC environment takes months from acquiring hardware to set up the hardware with the right set of firmware, software installed and configured. Furthermore, scalability is hard to achieve in a rigid traditional lab environment. Therefore, it is slow to react to the dynamic change in the artificial intelligent industry. Watson Deep Learning as a service, a cloud-based deep learning platform that mitigates the long lead time and high upfront investment in hardware. It enables robust and scalable sharing of resources among the teams in an organization. It is designed for on-demand cloud environments. Providing a similar user experience in a multi-tenant cloud environment comes with its own unique challenges regarding fault tolerance, performance, and security. Watson Deep Learning as a service tackles these challenges and present a deep learning stack for the cloud environments in a secure, scalable and fault-tolerant manner. It supports a wide range of deep-learning frameworks such as Tensorflow, PyTorch, Caffe, Torch, Theano, and MXNet etc. These frameworks reduce the effort and skillset required to design, train, and use deep learning models. Deep Learning as a service is used at IBM by AI researchers in areas including machine translation, computer vision, and healthcare. 

Keywords: deep learning, machine learning, cognitive computing, model training

Procedia PDF Downloads 205
457 Modelling and Simulation of Aero-Elastic Vibrations Using System Dynamic Approach

Authors: Cosmas Pandit Pagwiwoko, Ammar Khaled Abdelaziz Abdelsamia

Abstract:

Flutter as a phenomenon of flow-induced and self-excited vibration has to be recognized considering its harmful effect on the structure especially in a stage of aircraft design. This phenomenon is also important for a wind energy harvester based on the fluttering surface due to its effective operational velocity range. This multi-physics occurrence can be presented by two governing equations in both fluid and structure simultaneously in respecting certain boundary conditions on the surface of the body. In this work, the equations are resolved separately by two distinct solvers, one-time step of each domain. The modelling and simulation of this flow-structure interaction in ANSYS show the effectiveness of this loosely coupled method in representing flutter phenomenon however the process is time-consuming for design purposes. Therefore, another technique using the same weak coupled aero-structure is proposed by using system dynamics approach. In this technique, the aerodynamic forces were calculated using singularity function for a range of frequencies and certain natural mode shapes are transformed into time domain by employing an approximation model of fraction rational function in Laplace variable. The representation of structure in a multi-degree-of-freedom coupled with a transfer function of aerodynamic forces can then be simulated in time domain on a block-diagram platform such as Simulink MATLAB. The dynamic response of flutter at certain velocity can be evaluated with another established flutter calculation in frequency domain k-method. In this method, a parameter of artificial structural damping is inserted in the equation of motion to assure the energy balance of flow and vibrating structure. The simulation in time domain is particularly interested as it enables to apply the structural non-linear factors accurately. Experimental tests on a fluttering airfoil in the wind tunnel are also conducted to validate the method.

Keywords: flutter, flow-induced vibration, flow-structure interaction, non-linear structure

Procedia PDF Downloads 313
456 Cadmium Telluride Quantum Dots (CdTe QDs)-Thymine Conjugate Based Fluorescence Biosensor for Sensitive Determination of Nucleobases/Nucleosides

Authors: Lucja Rodzik, Joanna Lewandowska-Lancucka, Michal Szuwarzynski, Krzysztof Szczubialka, Maria Nowakowska

Abstract:

The analysis of nucleobases is of great importance for bioscience since their abnormal concentration in body fluids suggests the deficiency and mutation of the immune system, and it is considered to be an important parameter for diagnosis of various diseases. The presented conjugate meets the need for development of the effective, selective and highly sensitive sensor for nucleobase/nucleoside detection. The novel, highly fluorescent cadmium telluride quantum dots (CdTe QDs) functionalized with thymine and stabilized with thioglycolic acid (TGA) conjugates has been developed and thoroughly characterized. Successful formation of the material was confirmed by elemental analysis, and UV–Vis fluorescence and FTIR spectroscopies. The crystalline structure of the obtained product was characterized with X-ray diffraction (XRD) method. The composition of CdTe QDs and their thymine conjugate was also examined using X-ray photoelectron spectroscopy (XPS). The size of the CdTe-thymine was 3-6 nm as demonstrated using atomic force microscopy (AFM) and high resolution transmission electron microscopy (HRTEM) imaging. The plasmon resonance fluorescence band at 540 nm on excitation at 351 nm was observed for these nanoparticles. The intensity of this band increased with the increase in the amount of conjugated thymine with no shift in its position. Based on the fluorescence measurements, it was found that the CdTe-thymine conjugate interacted efficiently and selectively not only with adenine, a nucleobase complementary to thymine, but also with nucleosides and adenine-containing modified nucleosides, i.e., 5′-deoxy-5′-(methylthio)adenosine (MTA) and 2’-O-methyladenosine, the urinary tumor markers which allow monitoring of the disease progression. The applicability of the CdTe-thymine sensor for the real sample analysis was also investigated in simulated urine conditions. High sensitivity and selectivity of CdTe-thymine fluorescence towards adenine, adenosine and modified adenosine suggest that obtained conjugate can be potentially useful for development of the biosensor for complementary nucleobase/nucleoside detection.

Keywords: CdTe quantum dots, conjugate, sensor, thymine

Procedia PDF Downloads 409
455 Using Hyperspectral Sensor and Machine Learning to Predict Water Potentials of Wild Blueberries during Drought Treatment

Authors: Yongjiang Zhang, Kallol Barai, Umesh R. Hodeghatta, Trang Tran, Vikas Dhiman

Abstract:

Detecting water stress on crops early and accurately is crucial to minimize its impact. This study aims to measure water stress in wild blueberry crops non-destructively by analyzing proximal hyperspectral data. The data collection took place in the summer growing season of 2022. A drought experiment was conducted on wild blueberries in the randomized block design in the greenhouse, incorporating various genotypes and irrigation treatments. Hyperspectral data ( spectral range: 400-1000 nm) using a handheld spectroradiometer and leaf water potential data using a pressure chamber were collected from wild blueberry plants. Machine learning techniques, including multiple regression analysis and random forest models, were employed to predict leaf water potential (MPa). We explored the optimal wavelength bands for simple differences (RY1-R Y2), simple ratios (RY1/RY2), and normalized differences (|RY1-R Y2|/ (RY1-R Y2)). NDWI ((R857 - R1241)/(R857 + R1241)), SD (R2188 – R2245), and SR (R1752 / R1756) emerged as top predictors for predicting leaf water potential, significantly contributing to the highest model performance. The base learner models achieved an R-squared value of approximately 0.81, indicating their capacity to explain 81% of the variance. Research is underway to develop a neural vegetation index (NVI) that automates the process of index development by searching for specific wavelengths in the space ratio of linear functions of reflectance. The NVI framework could work across species and predict different physiological parameters.

Keywords: hyperspectral reflectance, water potential, spectral indices, machine learning, wild blueberries, optimal bands

Procedia PDF Downloads 64
454 Polymeric Micelles Based on Block Copolymer α-Tocopherol Succinate-g-Carboxymethyl Chitosan for Tamoxifen Delivery

Authors: Sunil K. Jena, Sanjaya K. Samal, Mahesh Chand, Abhay T. Sangamwar

Abstract:

Tamoxifen (TMX) and its analogues are approved as a first line therapy for the treatment of estrogen receptor-positive tumors. However, clinical development of TMX has been hampered by its low bioavailability and severe hepatotoxicity. Herein, we attempt to design a new drug delivery vehicle that could enhance the pharmacokinetic performance of TMX. Initially, high-molecular weight carboxymethyl chitosan was hydrolyzed to low-molecular weight carboxymethyl chitosan (LMW CMC) with hydrogen peroxide under the catalysis of phosphotungstic acid. Amphiphilic block copolymers of LMW CMC were synthesized via amidation reaction between the carboxyl group of α-tocopherol succinate (TS) and an amine group of LMW CMC. These amphiphilic block copolymers were self-assembled to nanosize core-shell-structural micelles in the aqueous medium. The critical micelle concentration (CMC) decreased with the increasing substitution of TS on LMW CMC, which ranged from 1.58 × 10-6 to 7.94 × 10-8 g/mL. Maximum TMX loading up to 8.08 ± 0.98% was achieved with Cmc-TS4.5 (TMX/Cmc-TS4.5 with 1:8 weight ratio). Both blank and TMX-loaded polymeric micelles (TMX-PM) of Cmc-TS4.5 exhibits spherical shape with the particle size below 200 nm. TMX-PM has been found to be stable in the gastrointestinal conditions and released only 44.5% of the total drug content by the first 72 h in simulated gastric fluid (SGF), pH 1.2. However, the presence of pepsin does not significantly increased the TMX release in SGF, pH 1.2, released only about 46.2% by the first 72 h suggesting its inability to cleave the peptide bond. In contrast, the release of TMX from TMX-PM4.5 in SIF, pH 6.8 (without pancreatin) was slow and sustained, released only about 10.43% of the total drug content within the first 30 min and nearly about 12.41% by the first 72 h. The presence of pancreatin in SIF, pH 6.8 led to an improvement in drug release. About 28.09% of incorporated TMX was released in the presence of pancreatin in 72 h. A cytotoxicity study demonstrated that TMX-PM exhibited time-delayed cytotoxicity in human MCF-7 breast cancer cells. Pharmacokinetic studies on Sprague-Dawley rats revealed a remarkable increase in oral bioavailability (1.87-fold) with significant (p < 0.0001) enhancement in AUC0-72 h, t1/2 and MRT of TMX-PM4.5 than that of TMX-suspension. Thus, the results suggested that CMC-TS micelles are a promising carrier for TMX delivery.

Keywords: carboxymethyl chitosan, d-α-tocopherol succinate, pharmacokinetic, polymeric micelles, tamoxifen

Procedia PDF Downloads 325
453 Selection of Optimal Reduced Feature Sets of Brain Signal Analysis Using Heuristically Optimized Deep Autoencoder

Authors: Souvik Phadikar, Nidul Sinha, Rajdeep Ghosh

Abstract:

In brainwaves research using electroencephalogram (EEG) signals, finding the most relevant and effective feature set for identification of activities in the human brain is a big challenge till today because of the random nature of the signals. The feature extraction method is a key issue to solve this problem. Finding those features that prove to give distinctive pictures for different activities and similar for the same activities is very difficult, especially for the number of activities. The performance of a classifier accuracy depends on this quality of feature set. Further, more number of features result in high computational complexity and less number of features compromise with the lower performance. In this paper, a novel idea of the selection of optimal feature set using a heuristically optimized deep autoencoder is presented. Using various feature extraction methods, a vast number of features are extracted from the EEG signals and fed to the autoencoder deep neural network. The autoencoder encodes the input features into a small set of codes. To avoid the gradient vanish problem and normalization of the dataset, a meta-heuristic search algorithm is used to minimize the mean square error (MSE) between encoder input and decoder output. To reduce the feature set into a smaller one, 4 hidden layers are considered in the autoencoder network; hence it is called Heuristically Optimized Deep Autoencoder (HO-DAE). In this method, no features are rejected; all the features are combined into the response of responses of the hidden layer. The results reveal that higher accuracy can be achieved using optimal reduced features. The proposed HO-DAE is also compared with the regular autoencoder to test the performance of both. The performance of the proposed method is validated and compared with the other two methods recently reported in the literature, which reveals that the proposed method is far better than the other two methods in terms of classification accuracy.

Keywords: autoencoder, brainwave signal analysis, electroencephalogram, feature extraction, feature selection, optimization

Procedia PDF Downloads 111
452 Simulation and Design of an Aerospace Mission Powered by “Candy” Type Fuel Engines

Authors: N. Hernández Huertas, F. Rojas Mora

Abstract:

Sounding rockets are aerospace vehicles that were developed in the mid-20th century, and since then numerous investigations have been executed with the aim of innovate in this type of technology. However, the costs associated to the production of this type of technology are usually quite high, and therefore the challenge that exists today is to be able to reduce them. In this way, the main objective of this document is to present the design process of a Colombian aerospace mission capable to reach the thermosphere using low-cost “Candy” type solid fuel engines. This mission is the latest development of the Uniandes Aerospace Project (PUA for its Spanish acronym), which is an undergraduate and postgraduate research group at Universidad de los Andes (Bogotá, Colombia), dedicated to incurring in this type of technology. In this way, the investigations that have been carried out on Candy-type solid fuel, which is a compound of potassium nitrate and sorbitol, have allowed the production of engines powerful enough to reach space, and which represents a unique technological advance in Latin America and an important development in experimental rocketry. In this way, following the engineering iterative design methodology was possible to design a 2-stage sounding rocket with 1 solid fuel engine in each one, which was then simulated in RockSim V9.0 software and reached an apogee of approximately 150 km above sea level. Similarly, a speed equal to 5 Mach was obtained, which after performing a finite element analysis, it was shown that the rocket is strong enough to be able to withstand such speeds. Under these premises, it was demonstrated that it is possible to build a high-power aerospace mission at low cost, using Candy-type solid fuel engines. For this reason, the feasibility of carrying out similar missions clearly depends on the ability to replicate the engines in the best way, since as mentioned above, the design of the rocket is adequate to reach supersonic speeds and reach space. Consequently, with a team of at least 3 members, the mission can be obtained in less than 3 months. Therefore, when publishing this project, it is intended to be a reference for future research in this field and benefit the industry.

Keywords: aerospace missions, Candy type solid propellant engines, design of solid rockets, experimental rocketry, low costs missions

Procedia PDF Downloads 108
451 A Preliminary Kinematic Comparison of Vive and Vicon Systems for the Accurate Tracking of Lumbar Motion

Authors: Yaghoubi N., Moore Z., Van Der Veen S. M., Pidcoe P. E., Thomas J. S., Dexheimer B.

Abstract:

Optoelectronic 3D motion capture systems, such as the Vicon kinematic system, are widely utilized in biomedical research to track joint motion. These systems are considered powerful and accurate measurement tools with <2 mm average error. However, these systems are costly and may be difficult to implement and utilize in a clinical setting. 3D virtual reality (VR) is gaining popularity as an affordable and accessible tool to investigate motor control and perception in a controlled, immersive environment. The HTC Vive VR system includes puck-style trackers that seamlessly integrate into its VR environments. These affordable, wireless, lightweight trackers may be more feasible for clinical kinematic data collection. However, the accuracy of HTC Vive Trackers (3.0), when compared to optoelectronic 3D motion capture systems, remains unclear. In this preliminary study, we compared the HTC Vive Tracker system to a Vicon kinematic system in a simulated lumbar flexion task. A 6-DOF robot arm (SCORBOT ER VII, Eshed Robotec/RoboGroup, Rosh Ha’Ayin, Israel) completed various reaching movements to mimic increasing levels of hip flexion (15°, 30°, 45°). Light reflective markers, along with one HTC Vive Tracker (3.0), were placed on the rigid segment separating the elbow and shoulder of the robot. We compared position measures simultaneously collected from both systems. Our preliminary analysis shows no significant differences between the Vicon motion capture system and the HTC Vive tracker in the Z axis, regardless of hip flexion. In the X axis, we found no significant differences between the two systems at 15 degrees of hip flexion but minimal differences at 30 and 45 degrees, ranging from .047 cm ± .02 SE (p = .03) at 30 degrees hip flexion to .194 cm ± .024 SE (p < .0001) at 45 degrees of hip flexion. In the Y axis, we found a minimal difference for 15 degrees of hip flexion only (.743 cm ± .275 SE; p = .007). This preliminary analysis shows that the HTC Vive Tracker may be an appropriate, affordable option for gross motor motion capture when the Vicon system is not available, such as in clinical settings. Further research is needed to compare these two motion capture systems in different body poses and for different body segments.

Keywords: lumbar, vivetracker, viconsystem, 3dmotion, ROM

Procedia PDF Downloads 98
450 Mechanical and Tribological Performances of (Nb: H-D: a-C) Thin Films for Biomedical Applications

Authors: Sara Khamseh, Kambiz Javanruee, Hamid Khorsand

Abstract:

Plenty of metallic materials are used for biomedical applications like hip joints and screws. Besides, it is reported that metal platforms such as stainless steel show significant deterioration because of wear and friction. The surface of metal substrates has been coated with a variety of multicomponent coatings to prevail these problems. The carbon-based multicomponent coatings such as metal-added amorphous carbon and diamond coatings are crucially important because of their remarkable tribological performance and chemical stability. In the current study, H-D contained Nb: (a-C) multicomponent coatings (H-D: hexagonal diamond, a-C: amorphous carbon) coated on A 304 steel substrates using an unbalanced magnetron (UBM) sputtering system. The effects of Nb and H-D content and ID/IG ratio on microstructure, mechanical and tribological characteristics of (Nb: H-D: a-C) composite coatings were investigated. The results of Raman spectroscopy represented that a-C phase with a Graphite-like structure (GLC with high value of sp2 carbon bonding) is formed, and its domain size increased with increasing Nb content of the coatings. Moreover, the Nb played a catalyst for the formation of the H-D phase. The nanoindentation hardness value of the coatings ranged between ~17 to ~35 GPa and (Nb: H-D: a-C) composite coatings with more H-D content represented higher hardness and plasticity index. It seems that the existence of extra-hard H-D particles straightly increased hardness. The tribological performance of the coatings was evaluated using the pin-on-disc method under the wet environment of SBF (Simulated Body Fluid). The COF value of the (Nb: H-D: a-C) coatings decreased with an increasing ID/IG ratio. The lower coefficient of friction is a result of the lamelliform array of graphitic domains. Also, the wear rate of the coatings decreased with increasing H-D content of the coatings. Based on the literature, a-C coatings with high hardness and H3/E2 ratio represent lower wear rates and better tribological performance. According to the nanoindentation analysis, hardness and H3/E2 ratio of (Nb: H-D: a-C) multicomponent coatings increased with increasing H-D content, which in turn decreased the wear rate of the coatings. The mechanical and tribological potency of (Nb: H-D: a-C) composite coatings on A 304 steel substrates paved the way for the development of innovative advanced coatings to ameliorate the performance of A 304 steel for biomedical applications.

Keywords: COF, mechanical properties, (Nb: H-D: a-C) coatings, wear rate

Procedia PDF Downloads 97
449 Energy Loss Reduction in Oil Refineries through Flare Gas Recovery Approaches

Authors: Majid Amidpour, Parisa Karimi, Marzieh Joda

Abstract:

For the last few years, release of burned undesirable by-products has become a challenging issue in oil industries. Flaring, as one of the main sources of air contamination, involves detrimental and long-lasting effects on human health and is considered a substantial reason for energy losses worldwide. This research involves studying the implications of two main flare gas recovery methods at three oil refineries, all in Iran as the case I, case II, and case III in which the production capacities are increasing respectively. In the proposed methods, flare gases are converted into more valuable products, before combustion by the flare networks. The first approach involves collecting, compressing and converting the flare gas to smokeless fuel which can be used in the fuel gas system of the refineries. The other scenario includes utilizing the flare gas as a feed into liquefied petroleum gas (LPG) production unit already established in the refineries. The processes of these scenarios are simulated, and the capital investment is calculated for each procedure. The cumulative profits of the scenarios are evaluated using Net Present Value method. Furthermore, the sensitivity analysis based on total propane and butane mole fraction is carried out to make a rational comparison for LPG production approach, and the results are illustrated for different mole fractions of propane and butane. As the mole fraction of propane and butane contained in LPG differs in summer and winter seasons, the results corresponding to LPG scenario are demonstrated for each season. The results of the simulations show that cumulative profit in fuel gas production scenario and LPG production rate increase with the capacity of the refineries. Moreover, the investment return time in LPG production method experiences a decline, followed by a rising trend with an increase in C3 and C4 content. The minimum value of time return occurs at propane and butane sum concentration values of 0.7, 0.6, and 0.7 in case I, II, and III, respectively. Based on comparison of the time of investment return and cumulative profit, fuel gas production is the superior scenario for three case studies.

Keywords: flare gas reduction, liquefied petroleum gas, fuel gas, net present value method, sensitivity analysis

Procedia PDF Downloads 156
448 Improving Chest X-Ray Disease Detection with Enhanced Data Augmentation Using Novel Approach of Diverse Conditional Wasserstein Generative Adversarial Networks

Authors: Malik Muhammad Arslan, Muneeb Ullah, Dai Shihan, Daniyal Haider, Xiaodong Yang

Abstract:

Chest X-rays are instrumental in the detection and monitoring of a wide array of diseases, including viral infections such as COVID-19, tuberculosis, pneumonia, lung cancer, and various cardiac and pulmonary conditions. To enhance the accuracy of diagnosis, artificial intelligence (AI) algorithms, particularly deep learning models like Convolutional Neural Networks (CNNs), are employed. However, these deep learning models demand a substantial and varied dataset to attain optimal precision. Generative Adversarial Networks (GANs) can be employed to create new data, thereby supplementing the existing dataset and enhancing the accuracy of deep learning models. Nevertheless, GANs have their limitations, such as issues related to stability, convergence, and the ability to distinguish between authentic and fabricated data. In order to overcome these challenges and advance the detection and classification of CXR normal and abnormal images, this study introduces a distinctive technique known as DCWGAN (Diverse Conditional Wasserstein GAN) for generating synthetic chest X-ray (CXR) images. The study evaluates the effectiveness of this Idiosyncratic DCWGAN technique using the ResNet50 model and compares its results with those obtained using the traditional GAN approach. The findings reveal that the ResNet50 model trained on the DCWGAN-generated dataset outperformed the model trained on the classic GAN-generated dataset. Specifically, the ResNet50 model utilizing DCWGAN synthetic images achieved impressive performance metrics with an accuracy of 0.961, precision of 0.955, recall of 0.970, and F1-Measure of 0.963. These results indicate the promising potential for the early detection of diseases in CXR images using this Inimitable approach.

Keywords: CNN, classification, deep learning, GAN, Resnet50

Procedia PDF Downloads 83
447 Estimation of Snow and Ice Melt Contributions to Discharge from the Glacierized Hunza River Basin, Karakoram, Pakistan

Authors: Syed Hammad Ali, Rijan Bhakta Kayastha, Danial Hashmi, Richard Armstrong, Ahuti Shrestha, Iram Bano, Javed Hassan

Abstract:

This paper presents the results of a semi-distributed modified positive degree-day model (MPDDM) for estimating snow and ice melt contributions to discharge from the glacierized Hunza River basin, Pakistan. The model uses daily temperature data, daily precipitation data, and positive degree day factors for snow and ice melt. The model is calibrated for the period 1995-2001 and validated for 2002-2013, and demonstrates close agreements between observed and simulated discharge with Nash–Sutcliffe Efficiencies of 0.90 and 0.88, respectively. Furthermore, the Weather Research and Forecasting model projected temperature, and precipitation data from 2016-2050 are used for representative concentration pathways RCP4.5 and RCP8.5, and bias correction was done using a statistical approach for future discharge estimation. No drastic changes in future discharge are predicted for the emissions scenarios. The aggregate snow-ice melt contribution is 39% of total discharge in the period 1993-2013. Snow-ice melt contribution ranges from 35% to 63% during the high flow period (May to October), which constitutes 89% of annual discharge; in the low flow period (November to April) it ranges from 0.02% to 17%, which constitutes 11 % of the annual discharge. The snow-ice melt contribution to total discharge will increase gradually in the future and reach up to 45% in 2041-2050. From a sensitivity analysis, it is found that the combination of a 2°C temperature rise and 20% increase in precipitation shows a 10% increase in discharge. The study allows us to evaluate the impact of climate change in such basins and is also useful for the future prediction of discharge to define hydropower potential, inform other water resource management in the area, to understand future changes in snow-ice melt contribution to discharge, and offer a possible evaluation of future water quantity and availability.

Keywords: climate variability, future discharge projection, positive degree day, regional climate model, water resource management

Procedia PDF Downloads 288
446 Comparative Evaluation of Root Uptake Models for Developing Moisture Uptake Based Irrigation Schedules for Crops

Authors: Vijay Shankar

Abstract:

In the era of water scarcity, effective use of water via irrigation requires good methods for determining crop water needs. Implementation of irrigation scheduling programs requires an accurate estimate of water use by the crop. Moisture depletion from the root zone represents the consequent crop evapotranspiration (ET). A numerical model for simulating soil water depletion in the root zone has been developed by taking into consideration soil physical properties, crop and climatic parameters. The governing differential equation for unsaturated flow of water in the soil is solved numerically using the fully implicit finite difference technique. The water uptake by plants is simulated by using three different sink functions. The non-linear model predictions are in good agreement with field data and thus it is possible to schedule irrigations more effectively. The present paper describes irrigation scheduling based on moisture depletion from the different layers of the root zone, obtained using different sink functions for three cash, oil and forage crops: cotton, safflower and barley, respectively. The soil is considered at a moisture level equal to field capacity prior to planting. Two soil moisture regimes are then imposed for irrigated treatment, one wherein irrigation is applied whenever soil moisture content is reduced to 50% of available soil water; and other wherein irrigation is applied whenever soil moisture content is reduced to 75% of available soil water. For both the soil moisture regimes it has been found that the model incorporating a non-linear sink function which provides best agreement of computed root zone moisture depletion with field data, is most effective in scheduling irrigations. Simulation runs with this moisture uptake function result in saving 27.3 to 45.5% & 18.7 to 37.5%, 12.5 to 25% % &16.7 to 33.3% and 16.7 to 33.3% & 20 to 40% irrigation water for cotton, safflower and barley respectively, under 50 & 75% moisture depletion regimes over other moisture uptake functions considered in the study. Simulation developed can be used for an optimized irrigation planning for different crops, choosing a suitable soil moisture regime depending upon the irrigation water availability and crop requirements.

Keywords: irrigation water, evapotranspiration, root uptake models, water scarcity

Procedia PDF Downloads 329
445 A Deep Learning Approach to Real Time and Robust Vehicular Traffic Prediction

Authors: Bikis Muhammed, Sehra Sedigh Sarvestani, Ali R. Hurson, Lasanthi Gamage

Abstract:

Vehicular traffic events have overly complex spatial correlations and temporal interdependencies and are also influenced by environmental events such as weather conditions. To capture these spatial and temporal interdependencies and make more realistic vehicular traffic predictions, graph neural networks (GNN) based traffic prediction models have been extensively utilized due to their capability of capturing non-Euclidean spatial correlation very effectively. However, most of the already existing GNN-based traffic prediction models have some limitations during learning complex and dynamic spatial and temporal patterns due to the following missing factors. First, most GNN-based traffic prediction models have used static distance or sometimes haversine distance mechanisms between spatially separated traffic observations to estimate spatial correlation. Secondly, most GNN-based traffic prediction models have not incorporated environmental events that have a major impact on the normal traffic states. Finally, most of the GNN-based models did not use an attention mechanism to focus on only important traffic observations. The objective of this paper is to study and make real-time vehicular traffic predictions while incorporating the effect of weather conditions. To fill the previously mentioned gaps, our prediction model uses a real-time driving distance between sensors to build a distance matrix or spatial adjacency matrix and capture spatial correlation. In addition, our prediction model considers the effect of six types of weather conditions and has an attention mechanism in both spatial and temporal data aggregation. Our prediction model efficiently captures the spatial and temporal correlation between traffic events, and it relies on the graph attention network (GAT) and Bidirectional bidirectional long short-term memory (Bi-LSTM) plus attention layers and is called GAT-BILSTMA.

Keywords: deep learning, real time prediction, GAT, Bi-LSTM, attention

Procedia PDF Downloads 68
444 150 KVA Multifunction Laboratory Test Unit Based on Power-Frequency Converter

Authors: Bartosz Kedra, Robert Malkowski

Abstract:

This paper provides description and presentation of laboratory test unit built basing on 150 kVA power frequency converter and Simulink RealTime platform. Assumptions, based on criteria which load and generator types may be simulated using discussed device, are presented, as well as control algorithm structure. As laboratory setup contains transformer with thyristor controlled tap changer, a wider scope of setup capabilities is presented. Information about used communication interface, data maintenance, and storage solution as well as used Simulink real-time features is presented. List and description of all measurements are provided. Potential of laboratory setup modifications is evaluated. For purposes of Rapid Control Prototyping, a dedicated environment was used Simulink RealTime. Therefore, load model Functional Unit Controller is based on a PC computer with I/O cards and Simulink RealTime software. Simulink RealTime was used to create real-time applications directly from Simulink models. In the next step, applications were loaded on a target computer connected to physical devices that provided opportunity to perform Hardware in the Loop (HIL) tests, as well as the mentioned Rapid Control Prototyping process. With Simulink RealTime, Simulink models were extended with I/O cards driver blocks that made automatic generation of real-time applications and performing interactive or automated runs on a dedicated target computer equipped with a real-time kernel, multicore CPU, and I/O cards possible. Results of performed laboratory tests are presented. Different load configurations are described and experimental results are presented. This includes simulation of under frequency load shedding, frequency and voltage dependent characteristics of groups of load units, time characteristics of group of different load units in a chosen area and arbitrary active and reactive power regulation basing on defined schedule.

Keywords: MATLAB, power converter, Simulink Real-Time, thyristor-controlled tap changer

Procedia PDF Downloads 320
443 A Parallel Computation Based on GPU Programming for a 3D Compressible Fluid Flow Simulation

Authors: Sugeng Rianto, P.W. Arinto Yudi, Soemarno Muhammad Nurhuda

Abstract:

A computation of a 3D compressible fluid flow for virtual environment with haptic interaction can be a non-trivial issue. This is especially how to reach good performances and balancing between visualization, tactile feedback interaction, and computations. In this paper, we describe our approach of computation methods based on parallel programming on a GPU. The 3D fluid flow solvers have been developed for smoke dispersion simulation by using combinations of the cubic interpolated propagation (CIP) based fluid flow solvers and the advantages of the parallelism and programmability of the GPU. The fluid flow solver is generated in the GPU-CPU message passing scheme to get rapid development of haptic feedback modes for fluid dynamic data. A rapid solution in fluid flow solvers is developed by applying cubic interpolated propagation (CIP) fluid flow solvers. From this scheme, multiphase fluid flow equations can be solved simultaneously. To get more acceleration in the computation, the Navier-Stoke Equations (NSEs) is packed into channels of texel, where computation models are performed on pixels that can be considered to be a grid of cells. Therefore, despite of the complexity of the obstacle geometry, processing on multiple vertices and pixels can be done simultaneously in parallel. The data are also shared in global memory for CPU to control the haptic in providing kinaesthetic interaction and felling. The results show that GPU based parallel computation approaches provide effective simulation of compressible fluid flow model for real-time interaction in 3D computer graphic for PC platform. This report has shown the feasibility of a new approach of solving the compressible fluid flow equations on the GPU. The experimental tests proved that the compressible fluid flowing on various obstacles with haptic interactions on the few model obstacles can be effectively and efficiently simulated on the reasonable frame rate with a realistic visualization. These results confirm that good performances and balancing between visualization, tactile feedback interaction, and computations can be applied successfully.

Keywords: CIP, compressible fluid, GPU programming, parallel computation, real-time visualisation

Procedia PDF Downloads 430
442 Assessment, Diagnosis and Treatment, Simulation for the Nurse Practitioner Student

Authors: Helen Coronel, Will Brewer, Peggy Bergeron, Clarissa Hall, Victoria Casson

Abstract:

Simulation-based training provides the nurse practitioner (NP) student with a safe and controlled environment in which they can practice a real-life scenario. This type of learning fosters critical thinking skills essential to practice. The expectation of this study was that students would have an increase in their competency and confidence after performing the simulation. Approximately 8.4% of Americans suffer from depression. The state of Alabama is ranked 47th out of 50 for access to mental health care. As a result of this significant shortage of mental health providers, primary care providers are frequently put in the position of screening for and treating mental health conditions, such as depression. Family nurse practitioners are often utilized as primary care providers, making their ability to assess, diagnose and treat these disorders a necessary skill. The expected outcome of this simulation is an increase in confidence, competency and the empowerment of the nurse practitioner student’s ability to assess, diagnose and treat a common mood disorder they may encounter in practice. The Kirkpatrick Module was applied for this study. A non-experimental design using descriptive statistical analysis was utilized. The simulation was based on a common psychiatric mood disorder frequently observed in primary care and mental health clinics. Students were asked to watch a voiceover power point presentation prior to their on-campus simulation. The presentation included training on the assessment, diagnosis, and treatment of a patient with depression. Prior to the simulation, the students completed a pre-test, then participated in the simulation, and completed a post-test when done. Apple iPads were utilized to access a simulated health record. Major findings of the study support an increase in students’ competency and confidence when assessing, diagnosing, and treating an adult patient with depression.

Keywords: advanced practice, nurse practitioner, simulation, primary care, depression

Procedia PDF Downloads 92
441 DeepLig: A de-novo Computational Drug Design Approach to Generate Multi-Targeted Drugs

Authors: Anika Chebrolu

Abstract:

Mono-targeted drugs can be of limited efficacy against complex diseases. Recently, multi-target drug design has been approached as a promising tool to fight against these challenging diseases. However, the scope of current computational approaches for multi-target drug design is limited. DeepLig presents a de-novo drug discovery platform that uses reinforcement learning to generate and optimize novel, potent, and multitargeted drug candidates against protein targets. DeepLig’s model consists of two networks in interplay: a generative network and a predictive network. The generative network, a Stack- Augmented Recurrent Neural Network, utilizes a stack memory unit to remember and recognize molecular patterns when generating novel ligands from scratch. The generative network passes each newly created ligand to the predictive network, which then uses multiple Graph Attention Networks simultaneously to forecast the average binding affinity of the generated ligand towards multiple target proteins. With each iteration, given feedback from the predictive network, the generative network learns to optimize itself to create molecules with a higher average binding affinity towards multiple proteins. DeepLig was evaluated based on its ability to generate multi-target ligands against two distinct proteins, multi-target ligands against three distinct proteins, and multi-target ligands against two distinct binding pockets on the same protein. With each test case, DeepLig was able to create a library of valid, synthetically accessible, and novel molecules with optimal and equipotent binding energies. We propose that DeepLig provides an effective approach to design multi-targeted drug therapies that can potentially show higher success rates during in-vitro trials.

Keywords: drug design, multitargeticity, de-novo, reinforcement learning

Procedia PDF Downloads 92