Search results for: fast motion estimation; low-complexity motion estimation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4920

Search results for: fast motion estimation; low-complexity motion estimation

390 Determination of Circulating Tumor Cells in Breast Cancer Patients by Electrochemical Biosensor

Authors: Gökçe Erdemir, İlhan Yaylım, Serap Erdem-Kuruca, Musa Mutlu Can

Abstract:

It has been determined that the main reason for the death of cancer disease is caused by metastases rather than the primary tumor. The cells that leave the primary tumor and enter the circulation and cause metastasis in the secondary organs are called "circulating tumor cells" (CTCs). The presence and number of circulating tumor cells has been associated with poor prognosis in many major types of cancer, including breast, prostate, and colorectal cancer. It is thought that knowledge of circulating tumor cells, which are seen as the main cause of cancer-related deaths due to metastasis, plays a key role in the diagnosis and treatment of cancer. The fact that tissue biopsies used in cancer diagnosis and follow-up are an invasive method and are insufficient in understanding the risk of metastasis and the progression of the disease have led to new searches. Liquid biopsy tests performed with a small amount of blood sample taken from the patient for the detection of CTCs are easy and reliable, as well as allowing more than one sample to be taken over time to follow the prognosis. However, since these cells are found in very small amounts in the blood, it is very difficult to capture them and specially designed analytical techniques and devices are required. Methods based on the biological and physical properties of the cells are used to capture these cells in the blood. Early diagnosis is very important in following the prognosis of tumors of epithelial origin such as breast, lung, colon and prostate. Molecules such as EpCAM, vimentin, and cytokeratins are expressed on the surface of cells that pass into the circulation from very few primary tumors and reach secondary organs from the circulation, and are used in the diagnosis of cancer in the early stage. For example, increased EpCAM expression in breast and prostate cancer has been associated with prognosis. These molecules can be determined in some blood or body fluids to be taken from patients. However, more sensitive methods are required to be able to determine when they are at a low level according to the course of the disease. The aim is to detect these molecules found in very few cancer cells with the help of sensitive, fast-sensing biosensors, first in breast cancer cells reproduced in vitro and then in blood samples taken from breast cancer patients. In this way, cancer cells can be diagnosed early and easily and effectively treated.

Keywords: electrochemical biosensors, breast cancer, circulating tumor cells, EpCAM, Vimentin, Cytokeratins

Procedia PDF Downloads 259
389 Psychological Aspects of Quality of Life in Patients with Primary and Metastatic Bone Tumors

Authors: O. Yu Shchelkova, E. B. Usmanova

Abstract:

Introduction: Last decades scientific research of quality of life (QoL) is developing fast worldwide. QoL concept pays attention to emotional experience of disease in patients, particularly to personal sense of possibility to satisfy actual needs and possibility of full social functioning in spite of disease limitations. QoL in oncological patients is studied intensively. Nevertheless, the issue of QoL in patients with bone tumors focused on psychological factors of QoL and relation to disease impact on QoL is not discussed. The aim of the study was to reveal the basic aspects and personality factors of QoL in patients with bone tumor. Results: Study participants were 139 patients with bone tumors. The diagnoses were osteosarcoma (n=42), giant cell tumor (n=32), chondrosarcoma (n=32), Ewing sarcoma (n=10) and bone metastases (n=23). The study revealed that patients with bone metastases assess their health significantly worse than other patients. Besides patients with osteosarcoma evaluate their general health higher than patients with giant cell tumors. Social functioning in patients with chondrosarcoma is higher than in patients with bone metastases and patients with giant cell tumor. Patients with chondrosarcoma have higher physical functioning and less restricted in daily activities than patients with bone metastases. Patients with bone metastases characterize their pain as more widespread than patients with primary bone tumors and have more functional restrictions due to bone incision. Moreover, the study revealed personality significant influence on QoL related to bone tumors. Such characteristics in structure of personality as high degree of self-consciousness, personal resources, cooperation and disposition to positive reappraisal in difficult situation correspond to higher QoL. Otherwise low personal resources and slight problem solving behaviour, low degree of self-consciousness and high social dependence correspond to decrease of QoL in patients with bone tumors. Conclusion: Patients with bone metastasis have lower QoL compared to patients with primary bone tumors. Patients with giant cell tumor have the worth quality of life among patients with primary bone tumors. Furthermore, the results revealed differences in QoL parameters associated with personality characteristics in patients with bone tumors. Such psychological factors as future goals, interest in life and emotional saturation, besides high degree of personal resources and cooperation influence on increasing QoL in patients with bone tumors.

Keywords: quality of life, psychological factors, bone tumor, personality

Procedia PDF Downloads 139
388 Exploring the Potential of Modular Housing Designs for the Emergency Housing Need in Türkiye after the February Earthquake in 2023

Authors: Hailemikael Negussie, Sebla Arın Ensarioğlu

Abstract:

In February 2023 Southeastern Türkiye and Northwestern Syria were hit by two consecutive earthquakes with high magnitude leaving thousands dead and thousands more homeless. The housing crisis in the affected areas has resulted in the need for a fast and qualified solution. There are a number of solutions, one of which is the use of modular designs to rebuild the cities that have been affected. Modular designs are prefabricated building components that can be quickly and efficiently assembled on-site, making them ideal to build structures with faster speed and higher quality. These structures are flexible, adaptable, and can be customized to meet the specific needs of the inhabitants, in addition to being more energy-efficient and sustainable. The prefabricated nature also assures that the quality of the products can be easily controlled. The reason for the collapse of most of the buildings during the earthquakes was found out to be the lack of quality during the construction stage. Using modular designs allows a higher control over the quality of the construction materials being used. The use of modular designs for a project of this scale presents some challenges, including the high upfront cost to design and manufacture components. However, if implemented correctly, modular designs can offer an effective and efficient solution to the urgent housing needs. The aim of this paper is to explore the potential of modular housing for mid- and long-term earthquake-resistant housing needs in the affected disaster zones after the earthquakes of February 2023. In the scope of this paper the adaptability of modular, prefabricated housing designs for the post-disaster environment, the advantages and disadvantages of this system will be examined. Elements such as; the current conditions of the region where the destruction happened, climatic data, topographic factors will be examined. Additionally, the paper will examine; examples of similar local and international modular post-earthquake housing projects. The region is projected to enter a rapid reconstruction phase in the following periods. Therefore, this paper will present a proposal for a system that can be used to produce safe and healthy urbanization policies without causing new aggrievements while meeting the housing needs of the people in the affected regions.

Keywords: post-disaster housing, earthquake-resistant design, modular design, housing, Türkiye

Procedia PDF Downloads 86
387 Understanding the Processwise Entropy Framework in a Heat-powered Cooling Cycle

Authors: P. R. Chauhan, S. K. Tyagi

Abstract:

Adsorption refrigeration technology offers a sustainable and energy-efficient cooling alternative over traditional refrigeration technologies for meeting the fast-growing cooling demands. With its ability to utilize natural refrigerants, low-grade heat sources, and modular configurations, it has the potential to revolutionize the cooling industry. Despite these benefits, the commercial viability of this technology is hampered by several fundamental limiting constraints, including its large size, low uptake capacity, and poor performance as a result of deficient heat and mass transfer characteristics. The primary cause of adequate heat and mass transfer characteristics and magnitude of exergy loss in various real processes of adsorption cooling system can be assessed by the entropy generation rate analysis, i. e. Second law of Thermodynamics. Therefore, this article presents the second law of thermodynamic-based investigation in terms of entropy generation rate (EGR) to identify the energy losses in various processes of the HPCC-based adsorption system using MATLAB R2021b software. The adsorption technology-based cooling system consists of two beds made up of silica gel and arranged in a single stage, while the water is employed as a refrigerant, coolant, and hot fluid. The variation in process-wise EGR is examined corresponding to cycle time, and a comparative analysis is also presented. Moreover, the EGR is also evaluated in the external units, such as the heat source and heat sink unit used for regeneration and heat dump, respectively. The research findings revealed that the combination of adsorber and desorber, which operates across heat reservoirs with a higher temperature gradient, shares more than half of the total amount of EGR. Moreover, the EGR caused by the heat transfer process is determined to be the highest, followed by a heat sink, heat source, and mass transfer, respectively. in case of heat transfer process, the operation of the valve is determined to be responsible for more than half (54.9%) of the overall EGR during the heat transfer. However, the combined contribution of the external units, such as the source (18.03%) and sink (21.55%), to the total EGR, is 35.59%. The analysis and findings of the present research are expected to pinpoint the source of the energy waste in HPCC based adsorption cooling systems.

Keywords: adsorption cooling cycle, heat transfer, mass transfer, entropy generation, silica gel-water

Procedia PDF Downloads 106
386 An Object-Oriented Modelica Model of the Water Level Swell during Depressurization of the Reactor Pressure Vessel of the Boiling Water Reactor

Authors: Rafal Bryk, Holger Schmidt, Thomas Mull, Ingo Ganzmann, Oliver Herbst

Abstract:

Prediction of the two-phase water mixture level during fast depressurization of the Reactor Pressure Vessel (RPV) resulting from an accident scenario is an important issue from the view point of the reactor safety. Since the level swell may influence the behavior of some passive safety systems, it has been recognized that an assumption which at the beginning may be considered as a conservative one, not necessary leads to a conservative result. This paper discusses outcomes obtained during simulations of the water dynamics and heat transfer during sudden depressurization of a vessel filled up to a certain level with liquid water under saturation conditions and with the rest of the vessel occupied by saturated steam. In case of the pressure decrease e.g. due to the main steam line break, the liquid water evaporates abruptly, being a reason thereby, of strong transients in the vessel. These transients and the sudden emergence of void in the region occupied at the beginning by liquid, cause elevation of the two-phase mixture. In this work, several models calculating the water collapse and swell levels are presented and validated against experimental data. Each of the models uses different approach to calculate void fraction. The object-oriented models were developed with the Modelica modelling language and the OpenModelica environment. The models represent the RPV of the Integral Test Facility Karlstein (INKA) – a dedicated test rig for simulation of KERENA – a new Boiling Water Reactor design of Framatome. The models are based on dynamic mass and energy equations. They are divided into several dynamic volumes in each of which, the fluid may be single-phase liquid, steam or a two-phase mixture. The heat transfer between the wall of the vessel and the fluid is taken into account. Additional heat flow rate may be applied to the first volume of the vessel in order to simulate the decay heat of the reactor core in a similar manner as it is simulated at INKA. The comparison of the simulations results against the reference data shows a good agreement.

Keywords: boiling water reactor, level swell, Modelica, RPV depressurization, thermal-hydraulics

Procedia PDF Downloads 209
385 Modelling of Phase Transformation Kinetics in Post Heat-Treated Resistance Spot Weld of AISI 1010 Mild Steel

Authors: B. V. Feujofack Kemda, N. Barka, M. Jahazi, D. Osmani

Abstract:

Automobile manufacturers are constantly seeking means to reduce the weight of car bodies. The usage of several steel grades in auto body assembling has been found to be a good technique to enlighten vehicles weight. This few years, the usage of dual phase (DP) steels, transformation induced plasticity (TRIP) steels and boron steels in some parts of the auto body have become a necessity because of their lightweight. However, these steels are martensitic, when they undergo a fast heat treatment, the resultant microstructure is essential, made of martensite. Resistance spot welding (RSW), one of the most used techniques in assembling auto bodies, becomes problematic in the case of these steels. RSW being indeed a process were steel is heated and cooled in a very short period of time, the resulting weld nugget is mostly fully martensitic, especially in the case of DP, TRIP and boron steels but that also holds for plain carbon steels as AISI 1010 grade which is extensively used in auto body inner parts. Martensite in its turn must be avoided as most as possible when welding steel because it is the principal source of brittleness and it weakens weld nugget. Thus, this work aims to find a mean to reduce martensite fraction in weld nugget when using RSW for assembling. The prediction of phase transformation kinetics during RSW has been done. That phase transformation kinetics prediction has been made possible through the modelling of the whole welding process, and a technique called post weld heat treatment (PWHT) have been applied in order to reduce martensite fraction in the weld nugget. Simulation has been performed for AISI 1010 grade, and results show that the application of PWHT leads to the formation of not only martensite but also ferrite, bainite and pearlite during the cooling of weld nugget. Welding experiments have been done in parallel and micrographic analyses show the presence of several phases in the weld nugget. Experimental weld geometry and phase proportions are in good agreement with simulation results, showing here the validity of the model.

Keywords: resistance spot welding, AISI 1010, modeling, post weld heat treatment, phase transformation, kinetics

Procedia PDF Downloads 116
384 Numerical Modelling of Wind Dispersal Seeds of Bromeliad Tillandsia recurvata L. (L.) Attached to Electric Power Lines

Authors: Bruna P. De Souza, Ricardo C. De Almeida

Abstract:

In some cities in the State of Parana – Brazil and in other countries atmospheric bromeliads (Tillandsia spp - Bromeliaceae) are considered weeds in trees, electric power lines, satellite dishes and other artificial supports. In this study, a numerical model was developed to simulate the seed dispersal of the Tillandsia recurvata species by wind with the objective of evaluating seeds displacement in the city of Ponta Grossa – PR, Brazil, since it is considered that the region is already infested. The model simulates the dispersal of each individual seed integrating parameters from the atmospheric boundary layer (ABL) and the local wind, simulated by the Weather Research Forecasting (WRF) mesoscale atmospheric model for the 2012 to 2015 period. The dispersal model also incorporates the approximate number of bromeliads and source height data collected from most infested electric power lines. The seeds terminal velocity, which is an important input data but was not available in the literature, was measured by an experiment with fifty-one seeds of Tillandsia recurvata. Wind is the main dispersal agent acting on plumed seeds whereas atmospheric turbulence is a determinant factor to transport the seeds to distances beyond 200 meters as well as to introduce random variability in the seed dispersal process. Such variability was added to the model through the application of an Inverse Fast Fourier Transform to wind velocity components energy spectra based on boundary-layer meteorology theory and estimated from micrometeorological parameters produced by the WRF model. Seasonal and annual wind means were obtained from the surface wind data simulated by WRF for Ponta Grossa. The mean wind direction is assumed to be the most probable direction of bromeliad seed trajectory. Moreover, the atmospheric turbulence effect and dispersal distances were analyzed in order to identify likely regions of infestation around Ponta Grossa urban area. It is important to mention that this model could be applied to any species and local as long as seed’s biological data and meteorological data for the region of interest are available.

Keywords: atmospheric turbulence, bromeliad, numerical model, seed dispersal, terminal velocity, wind

Procedia PDF Downloads 140
383 Evaluation of the Phenolic Composition of Curcumin from Different Turmeric (Curcuma longa L.) Extracts: A Comprehensive Study Based on Chemical Turmeric Extract, Turmeric Tea and Fresh Turmeric Juice

Authors: Beyza Sukran Isik, Gokce Altin, Ipek Yalcinkaya, Evren Demircan, Asli Can Karaca, Beraat Ozcelik

Abstract:

Turmeric (Curcuma longa L.), is used as a food additive (spice), preservative and coloring agent in Asian countries, including China and South East Asia. It is also considered as a medicinal plant. Traditional Indian medicine evaluates turmeric powder for the treatment of biliary disorders, rheumatism, and sinusitis. It has rich polyphenol content. Turmeric has yellow color mainly because of the presence of three major pigments; curcumin 1,7-bis(4-hydroxy-3-methoxyphenyl)-1, 6-heptadiene-3,5-dione), demethoxy-curcumin and bis demothoxy-curcumin. These curcuminoids are recognized to have high antioxidant activities. Curcumin is the major constituent of Curcuma species. Method: To prepare turmeric tea, 0.5 gram of turmeric powder was brewed with 250 ml of water at 90°C, 10 minutes. 500 grams of fresh turmeric washed and shelled prior to squeezing. Both turmeric tea and turmeric juice pass through 45 lm filters and stored at -20°C in the dark for further analyses. Curcumin was extracted from 20 grams of turmeric powder by 70 ml ethanol solution (95:5 ethanol/water v/v) in a water bath at 80°C, 6 hours. Extraction was contributed for 2 hours at the end of 6 hours by addition of 30 ml ethanol. Ethanol was removed by rotary evaporator. Remained extract stored at -20°C in the dark. Total phenolic content and phenolic profile were determined by spectrophotometric analysis and ultra-fast liquid chromatography (UFLC), respectively. Results: The total phenolic content of ethanolic extract of turmeric, turmeric juice, and turmeric tea were determined 50.72, 31.76 and 29.68 ppt, respectively. The ethanolic extract of turmeric, turmeric juice, and turmeric tea have been injected into UFLC and analyzed for curcumin contents. The curcumin content in ethanolic extract of turmeric, turmeric juice, and turmeric tea were 4067.4, 156.7 ppm and 1.1 ppm, respectively. Significance: Turmeric is known as a good source of curcumin. According to the results, it can be stated that its tea is not sufficient way for curcumin consumption. Turmeric juice can be preferred to turmeric tea for higher curcumin content. Ethanolic extract of turmeric showed the highest content of turmeric in both spectrophotometric and chromatographic analyses. Nonpolar solvents and carriers which have polar binding sites have to be considered for curcumin consumption due to its nonpolar nature.

Keywords: phenolic compounds, spectrophotometry, turmeric, UFLC

Procedia PDF Downloads 198
382 An Adaptive Conversational AI Approach for Self-Learning

Authors: Airy Huang, Fuji Foo, Aries Prasetya Wibowo

Abstract:

In recent years, the focus of Natural Language Processing (NLP) development has been gradually shifting from the semantics-based approach to deep learning one, which performs faster with fewer resources. Although it performs well in many applications, the deep learning approach, due to the lack of semantics understanding, has difficulties in noticing and expressing a novel business case with a pre-defined scope. In order to meet the requirements of specific robotic services, deep learning approach is very labor-intensive and time consuming. It is very difficult to improve the capabilities of conversational AI in a short time, and it is even more difficult to self-learn from experiences to deliver the same service in a better way. In this paper, we present an adaptive conversational AI algorithm that combines both semantic knowledge and deep learning to address this issue by learning new business cases through conversations. After self-learning from experience, the robot adapts to the business cases originally out of scope. The idea is to build new or extended robotic services in a systematic and fast-training manner with self-configured programs and constructed dialog flows. For every cycle in which a chat bot (conversational AI) delivers a given set of business cases, it is trapped to self-measure its performance and rethink every unknown dialog flows to improve the service by retraining with those new business cases. If the training process reaches a bottleneck and incurs some difficulties, human personnel will be informed of further instructions. He or she may retrain the chat bot with newly configured programs, or new dialog flows for new services. One approach employs semantics analysis to learn the dialogues for new business cases and then establish the necessary ontology for the new service. With the newly learned programs, it completes the understanding of the reaction behavior and finally uses dialog flows to connect all the understanding results and programs, achieving the goal of self-learning process. We have developed a chat bot service mounted on a kiosk, with a camera for facial recognition and a directional microphone array for voice capture. The chat bot serves as a concierge with polite conversation for visitors. As a proof of concept. We have demonstrated to complete 90% of reception services with limited self-learning capability.

Keywords: conversational AI, chatbot, dialog management, semantic analysis

Procedia PDF Downloads 135
381 Study of Mixing Conditions for Different Endothelial Dysfunction in Arteriosclerosis

Authors: Sara Segura, Diego Nuñez, Miryam Villamil

Abstract:

In this work, we studied the microscale interaction of foreign substances with blood inside an artificial transparent artery system that represents medium and small muscular arteries. This artery system had channels ranging from 75 μm to 930 μm and was fabricated using glass and transparent polymer blends like Phenylbis(2,4,6-trimethylbenzoyl) phosphine oxide, Poly(ethylene glycol) and PDMS in order to be monitored in real time. The setup was performed using a computer controlled precision micropump and a high resolution optical microscope capable of tracking fluids at fast capture. Observation and analysis were performed using a real time software that reconstructs the fluid dynamics determining the flux velocity, injection dependency, turbulence and rheology. All experiments were carried out with fully computer controlled equipment. Interactions between substances like water, serum (0.9% sodium chloride and electrolyte with a ratio of 4 ppm) and blood cells were studied at microscale as high as 400nm of resolution and the analysis was performed using a frame-by-frame observation and HD-video capture. These observations lead us to understand the fluid and mixing behavior of the interest substance in the blood stream and to shed a light on the use of implantable devices for drug delivery at arteries with different Endothelial dysfunction. Several substances were tested using the artificial artery system. Initially, Milli-Q water was used as a control substance for the study of the basic fluid dynamics of the artificial artery system. However, serum and other low viscous substances were pumped into the system with the presence of other liquids to study the mixing profiles and behaviors. Finally, mammal blood was used for the final test while serum was injected. Different flow conditions, pumping rates, and time rates were evaluated for the determination of the optimal mixing conditions. Our results suggested the use of a very fine controlled microinjection for better mixing profiles with and approximately rate of 135.000 μm3/s for the administration of drugs inside arteries.

Keywords: artificial artery, drug delivery, microfluidics dynamics, arteriosclerosis

Procedia PDF Downloads 294
380 Production of Pig Iron by Smelting of Blended Pre-Reduced Titaniferous Magnetite Ore and Hematite Ore Using Lean Grade Coal

Authors: Bitan Kumar Sarkar, Akashdeep Agarwal, Rajib Dey, Gopes Chandra Das

Abstract:

The rapid depletion of high-grade iron ore (Fe2O3) has gained attention on the use of other sources of iron ore. Titaniferous magnetite ore (TMO) is a special type of magnetite ore having high titania content (23.23% TiO2 present in this case). Due to high TiO2 content and high density, TMO cannot be treated by the conventional smelting reduction. In this present work, the TMO has been collected from high-grade metamorphic terrain of the Precambrian Chotanagpur gneissic complex situated in the eastern part of India (Shaltora area, Bankura district, West Bengal) and the hematite ore has been collected from Visakhapatnam Steel Plant (VSP), Visakhapatnam. At VSP, iron ore is received from Bailadila mines, Chattisgarh of M/s. National Mineral Development Corporation. The preliminary characterization of TMO and hematite ore (HMO) has been investigated by WDXRF, XRD and FESEM analyses. Similarly, good quality of coal (mainly coking coal) is also getting depleted fast. The basic purpose of this work is to find how lean grade coal can be utilised along with TMO for smelting to produce pig iron. Lean grade coal has been characterised by using TG/DTA, proximate and ultimate analyses. The boiler grade coal has been found to contain 28.08% of fixed carbon and 28.31% of volatile matter. TMO fines (below 75 μm) and HMO fines (below 75 μm) have been separately agglomerated with lean grade coal fines (below 75 μm) in the form of briquettes using binders like bentonite and molasses. These green briquettes are dried first in oven at 423 K for 30 min and then reduced isothermally in tube furnace over the temperature range of 1323 K, 1373 K and 1423 K for 30 min & 60 min. After reduction, the reduced briquettes are characterized by XRD and FESEM analyses. The best reduced TMO and HMO samples are taken and blended in three different weight percentage ratios of 1:4, 1:8 and 1:12 of TMO:HMO. The chemical analysis of three blended samples is carried out and degree of metallisation of iron is found to contain 89.38%, 92.12% and 93.12%, respectively. These three blended samples are briquetted using binder like bentonite and lime. Thereafter these blended briquettes are separately smelted in raising hearth furnace at 1773 K for 30 min. The pig iron formed is characterized using XRD, microscopic analysis. It can be concluded that 90% yield of pig iron can be achieved when the blend ratio of TMO:HMO is 1:4.5. This means for 90% yield, the maximum TMO that could be used in the blend is about 18%.

Keywords: briquetting reduction, lean grade coal, smelting reduction, TMO

Procedia PDF Downloads 319
379 Case-Based Reasoning Application to Predict Geological Features at Site C Dam Construction Project

Authors: Shahnam Behnam Malekzadeh, Ian Kerr, Tyson Kaempffer, Teague Harper, Andrew Watson

Abstract:

The Site C Hydroelectric dam is currently being constructed in north-eastern British Columbia on sub-horizontal sedimentary strata that dip approximately 15 meters from one bank of the Peace River to the other. More than 615 pressure sensors (Vibrating Wire Piezometers) have been installed on bedding planes (BPs) since construction began, with over 80 more planned before project completion. These pressure measurements are essential to monitor the stability of the rock foundation during and after construction and for dam safety purposes. BPs are identified by their clay gouge infilling, which varies in thickness from less than 1 to 20 mm and can be challenging to identify as the core drilling process often disturbs or washes away the gouge material. Without the use of depth predictions from nearby boreholes, stratigraphic markers, and downhole geophysical data, it is difficult to confidently identify BP targets for the sensors. In this paper, a Case-Based Reasoning (CBR) method was used to develop an empirical model called the Bedding Plane Elevation Prediction (BPEP) to help geologists and geotechnical engineers to predict geological features and bedding planes at new locations in a fast and accurate manner. To develop CBR, a database was developed based on 64 pressure sensors already installed on key bedding planes BP25, BP28, and BP31 on the Right Bank, including bedding plane elevations and coordinates. Thirteen (20%) of the most recent cases were selected to validate and evaluate the accuracy of the developed model, while the similarity was defined as the distance between previous cases and recent cases to predict the depth of significant BPs. The average difference between actual BP elevations and predicted elevations for above BPs was ±55cm, while the actual results showed that 69% of predicted elevations were within ±79 cm of actual BP elevations while 100% of predicted elevations for new cases were within ±99cm range. Eventually, the actual results will be used to develop the database and improve BPEP to perform as a learning machine to predict more accurate BP elevations for future sensor installations.

Keywords: case-based reasoning, geological feature, geology, piezometer, pressure sensor, core logging, dam construction

Procedia PDF Downloads 77
378 A Fourier Method for Risk Quantification and Allocation of Credit Portfolios

Authors: Xiaoyu Shen, Fang Fang, Chujun Qiu

Abstract:

Herewith we present a Fourier method for credit risk quantification and allocation in the factor-copula model framework. The key insight is that, compared to directly computing the cumulative distribution function of the portfolio loss via Monte Carlo simulation, it is, in fact, more efficient to calculate the transformation of the distribution function in the Fourier domain instead and inverting back to the real domain can be done in just one step and semi-analytically, thanks to the popular COS method (with some adjustments). We also show that the Euler risk allocation problem can be solved in the same way since it can be transformed into the problem of evaluating a conditional cumulative distribution function. Once the conditional or unconditional cumulative distribution function is known, one can easily calculate various risk metrics. The proposed method not only fills the niche in literature, to the best of our knowledge, of accurate numerical methods for risk allocation but may also serve as a much faster alternative to the Monte Carlo simulation method for risk quantification in general. It can cope with various factor-copula model choices, which we demonstrate via examples of a two-factor Gaussian copula and a two-factor Gaussian-t hybrid copula. The fast error convergence is proved mathematically and then verified by numerical experiments, in which Value-at-Risk, Expected Shortfall, and conditional Expected Shortfall are taken as examples of commonly used risk metrics. The calculation speed and accuracy are tested to be significantly superior to the MC simulation for real-sized portfolios. The computational complexity is, by design, primarily driven by the number of factors instead of the number of obligors, as in the case of Monte Carlo simulation. The limitation of this method lies in the "curse of dimension" that is intrinsic to multi-dimensional numerical integration, which, however, can be relaxed with the help of dimension reduction techniques and/or parallel computing, as we will demonstrate in a separate paper. The potential application of this method has a wide range: from credit derivatives pricing to economic capital calculation of the banking book, default risk charge and incremental risk charge computation of the trading book, and even to other risk types than credit risk.

Keywords: credit portfolio, risk allocation, factor copula model, the COS method, Fourier method

Procedia PDF Downloads 165
377 3D-Printing Compressible Macroporous Polymer Using Poly-Pickering-High Internal Phase Emulsions as Micromixer

Authors: Hande Barkan-Ozturk, Angelika Menner, Alexander Bismarck

Abstract:

Microfluidic mixing technology grew rapidly in the past few years due to its many advantages over the macro-scale mixing, especially the ability to use small amounts of internal volume and also very high surface-to-volume ratio. The Reynold number identify whether the mixing is operated by the laminar or turbulence flow. Therefore, mixing with very fast kinetic can be achieved by diminishing the channel dimensions to decrease Reynold number and the laminar flow can be accomplished. Moreover, by using obstacles in the micromixer, the mixing length and the contact area between the species have been increased. Therefore, the channel geometry and its surface property have great importance to reach satisfactory mixing results. Since poly(-merised) High Internal Phase Emulsions (polyHIPEs) have more than 74% porosity and their pores are connected each other with pore throats, which cause high permeability, they are ideal candidate to build a micromixer. The HIPE precursor is commonly produced by using an overhead stirrer to obtain relatively large amount of emulsion in batch process. However, we will demonstrate that a desired amount of emulsion can be prepared continuously with micromixer build from polyHIPE, and such HIPE can subsequently be employed as ink in 3D printing process. In order to produce the micromixer a poly-Pickering(St-co-DVB)HIPE with 80% porosity was prepared with modified silica particles as stabilizer and surfactant Hypermer 2296 to obtain open porous structure and after coating of the surface, the three 1/16' ' PTFE tubes to transfer continuous (CP) and internal phases (IP) and the other is to collect the emulsion were placed. Afterwards, the two phases were injected in the ratio 1:3 CP:IP with syringe dispensers, respectively, and highly viscoelastic H(M)IPE, which can be used as an ink in 3D printing process, was gathered continuously. After the polymerisation of the resultant emulsion, polyH(M)IPE has interconnected porous structure identical to the monolithic polyH(M)IPE indicating that the emulsion can be prepared constantly with poly-Pickering-HIPE as micromixer and it can be used to prepare desired pattern with a 3D printer. Moreover, the morphological properties of the emulsion can be adjustable by changing flow ratio, flow speed and structure of the micromixer.

Keywords: 3D-Printing, emulsification, macroporous polymer, micromixer, polyHIPE

Procedia PDF Downloads 159
376 Multi-Walled Carbon Nanotubes Doped Poly (3,4 Ethylenedioxythiophene) Composites Based Electrochemical Nano-Biosensor for Organophosphate Detection

Authors: Navpreet Kaur, Himkusha Thakur, Nirmal Prabhakar

Abstract:

One of the most publicized and controversial issue in crop production is the use of agrichemicals- also known as pesticides. This is evident in many reports that Organophosphate (OP) insecticides, among the broad range of pesticides are mainly involved in acute and chronic poisoning cases. Therefore, detection of OPs is very necessary for health protection, food and environmental safety. In our study, a nanocomposite of poly (3,4 ethylenedioxythiophene) (PEDOT) and multi-walled carbon nanotubes (MWCNTs) has been deposited electrochemically onto the surface of fluorine doped tin oxide sheets (FTO) for the analysis of malathion OP. The -COOH functionalization of MWCNTs has been done for the covalent binding with amino groups of AChE enzyme. The use of PEDOT-MWCNT films exhibited an excellent conductivity, enables fast transfer kinetics and provided a favourable biocompatible microenvironment for AChE, for the significant malathion OP detection. The prepared PEDOT-MWCNT/FTO and AChE/PEDOT-MWCNT/FTO nano-biosensors were characterized by Fourier transform infrared spectrometry (FTIR), Field emission-scanning electron microscopy (FE-SEM) and electrochemical studies. Electrochemical studies were done using Cyclic Voltammetry (CV) or Differential Pulse Voltammetry (DPV) and Electrochemical Impedance Spectroscopy (EIS). Various optimization studies were done for different parameters including pH (7.5), AChE concentration (50 mU), substrate concentration (0.3 mM) and inhibition time (10 min). The detection limit for malathion OP was calculated to be 1 fM within the linear range 1 fM to 1 µM. The activity of inhibited AChE enzyme was restored to 98% of its original value by 2-pyridine aldoxime methiodide (2-PAM) (5 mM) treatment for 11 min. The oxime 2-PAM is able to remove malathion from the active site of AChE by means of trans-esterification reaction. The storage stability and reusability of the prepared nano-biosensor is observed to be 30 days and seven times, respectively. The application of the developed nano-biosensor has also been evaluated for spiked lettuce sample. Recoveries of malathion from the spiked lettuce sample ranged between 96-98%. The low detection limit obtained by the developed nano-biosensor made them reliable, sensitive and a low cost process.

Keywords: PEDOT-MWCNT, malathion, organophosphates, acetylcholinesterase, nano-biosensor, oxime (2-PAM)

Procedia PDF Downloads 432
375 Applications of Artificial Intelligence (AI) in Cardiac imaging

Authors: Angelis P. Barlampas

Abstract:

The purpose of this study is to inform the reader, about the various applications of artificial intelligence (AI), in cardiac imaging. AI grows fast and its role is crucial in medical specialties, which use large amounts of digital data, that are very difficult or even impossible to be managed by human beings and especially doctors.Artificial intelligence (AI) refers to the ability of computers to mimic human cognitive function, performing tasks such as learning, problem-solving, and autonomous decision making based on digital data. Whereas AI describes the concept of using computers to mimic human cognitive tasks, machine learning (ML) describes the category of algorithms that enable most current applications described as AI. Some of the current applications of AI in cardiac imaging are the follows: Ultrasound: Automated segmentation of cardiac chambers across five common views and consequently quantify chamber volumes/mass, ascertain ejection fraction and determine longitudinal strain through speckle tracking. Determine the severity of mitral regurgitation (accuracy > 99% for every degree of severity). Identify myocardial infarction. Distinguish between Athlete’s heart and hypertrophic cardiomyopathy, as well as restrictive cardiomyopathy and constrictive pericarditis. Predict all-cause mortality. CT Reduce radiation doses. Calculate the calcium score. Diagnose coronary artery disease (CAD). Predict all-cause 5-year mortality. Predict major cardiovascular events in patients with suspected CAD. MRI Segment of cardiac structures and infarct tissue. Calculate cardiac mass and function parameters. Distinguish between patients with myocardial infarction and control subjects. It could potentially reduce costs since it would preclude the need for gadolinium-enhanced CMR. Predict 4-year survival in patients with pulmonary hypertension. Nuclear Imaging Classify normal and abnormal myocardium in CAD. Detect locations with abnormal myocardium. Predict cardiac death. ML was comparable to or better than two experienced readers in predicting the need for revascularization. AI emerge as a helpful tool in cardiac imaging and for the doctors who can not manage the overall increasing demand, in examinations such as ultrasound, computed tomography, MRI, or nuclear imaging studies.

Keywords: artificial intelligence, cardiac imaging, ultrasound, MRI, CT, nuclear medicine

Procedia PDF Downloads 77
374 Genetic Improvement Potential for Wood Production in Melaleuca cajuputi

Authors: Hong Nguyen Thi Hai, Ryota Konda, Dat Kieu Tuan, Cao Tran Thanh, Khang Phung Van, Hau Tran Tin, Harry Wu

Abstract:

Melaleuca cajuputi is a moderately fast-growing species and considered as a multi-purpose tree as it provides fuelwood, piles and frame poles in construction, leaf essential oil and honey. It occurs in Australia, Papua New Guinea, and South-East Asia. M. cajuputi plantation can be harvested on 6-7 year rotations for wood products. Its timber can also be used for pulp and paper, fiber and particle board, producing quality charcoal and potentially sawn timber. However, most reported M. cajuputi breeding programs have been focused on oil production rather than wood production. In this study, breeding program of M. cajuputi aimed to improve wood production was examined by estimating genetic parameters for growth (tree height, diameter at breast height (DBH), and volume), stem form, stiffness (modulus of elasticity (MOE)), bark thickness and bark ratio in a half-sib family progeny trial including 80 families in the Mekong Delta of Vietnam. MOE is one of the key wood properties of interest to the wood industry. Non-destructive wood stiffness was measured indirectly by acoustic velocity using FAKOPP Microsecond Timer and especially unaffected by bark mass. Narrow-sense heritability for the seven traits ranged from 0.13 to 0.27 at age 7 years. MOE and stem form had positive genetic correlations with growth while the negative correlation between bark ratio and growth was also favorable. Breeding for simultaneous improvement of multiple traits, faster growth with higher MOE and reduction of bark ratio should be possible in M. cajuputi. Index selection based on volume and MOE showed genetic gains of 31 % in volume, 6 % in MOE and 13 % in stem form. In addition, heritability and age-age genetic correlations for growth traits increased with time and optimal early selection age for growth of M. cajuputi based on DBH alone was 4 years. Selected thinning resulted in an increase of heritability due to considerable reduction of phenotypic variation but little effect on genetic variation.

Keywords: acoustic velocity, age-age correlation, bark thickness, heritability, Melaleuca cajuputi, stiffness, thinning effect

Procedia PDF Downloads 180
373 Factors Associated with Increase of Diabetic Foot Ulcers in Diabetic Patients in Nyahururu County Hospital

Authors: Daniel Wachira

Abstract:

The study aims to determine factors contributing to increasing rates of DFU among DM patients attending clinics in Nyahururu County referral hospital, Lakipia County. The study objectives include;- To determine the demographic factors contributing to increased rates of DFU among DM patients, determining the sociocultural factors that contribute to increased rates of DFU among DM patients and determining the health facility factors contributing to increased rates of DFU among DM patients attending DM clinic at Nyahururu county referral hospital, Laikipia County. This study will adopt a descriptive cross-sectional study design. It involves the collection of data at a one-time point without follow-up. This method is fast and inexpensive, there is no loss to follow up as the data is collected at one time point and associations between variables can be determined. The study population includes all DM patients with or without DFU. The sampling technique that will be used is the probability sampling method, a simple random method of sampling will be used. The study will employ the use of questionnaires to collect the required information. Questionnaires will be a research administered questionnaires. The questionnaire developed was done in consultation with other research experts (supervisor) to ensure reliability. The questionnaire designed will be pre-tested by hand delivering them to a sample 10% of the sample size at J.M Kariuki Memorial hospital, Nyandarua county and thereafter collecting them dully filled followed by refining of errors to ensure it is valid for collection of data relevant for this study. Refining of errors on the questionnaires to ensure it was valid for collection of data relevant for this study. Data collection will begin after the approval of the project. Questionnaires will be administered only to the participants who met the selection criteria by the researcher and those who agreed to participate in the study to collect key information with regard to the objectives of the study. The study's authority will be obtained from the National Commission of Science and Technology and Innovation. Permission will also be obtained from the Nyahururu County referral hospital administration staff. The purpose of the study will be explained to the respondents in order to secure informed consent, and no names will be written on the questionnaires. All the information will be treated with maximum confidentiality by not disclosing who the respondent was and the information.

Keywords: diabetes, foot ulcer, social factors, hospital factors

Procedia PDF Downloads 15
372 Research on the Spatial Organization and Collaborative Innovation of Innovation Corridors from the Perspective of Ecological Niche: A Case Study of Seven Municipal Districts in Jiangsu Province, China

Authors: Weikang Peng

Abstract:

The innovation corridor is an important spatial carrier to promote regional collaborative innovation, and its development process is the spatial re-organization process of regional innovation resources. This paper takes the Nanjing-Zhenjiang G312 Industrial Innovation Corridor, which involves seven municipal districts in Jiangsu Province, as empirical evidence. Based on multi-source spatial big data in 2010, 2016, and 2022, this paper applies triangulated irregular network (TIN), head/tail breaks, regional innovation ecosystem (RIE) niche fitness evaluation model, and social network analysis to carry out empirical research on the spatial organization and functional structural evolution characteristics of innovation corridors and their correlation with the structural evolution of collaborative innovation network. The results show, first, the development of innovation patches in the corridor has fractal characteristics in time and space and tends to be multi-center and cluster layout along the Nanjing Bypass Highway and National Highway G312. Second, there are large differences in the spatial distribution pattern of niche fitness in the corridor in various dimensions, and the niche fitness of innovation patches along the highway has increased significantly. Third, the scale of the collaborative innovation network in the corridor is expanding fast. The core of the network is shifting from the main urban area to the periphery of the city along the highway, with small-world and hierarchical levels, and the core-edge network structure is highlighted. With the development of the Innovation Corridor, the main collaborative mode in the corridor is changing from collaboration within innovation patches to collaboration between innovation patches, and innovation patches with high ecological suitability tend to be the active areas of collaborative innovation. Overall, polycentric spatial layout, graded functional structure, diversified innovation clusters, and differentiated environmental support play an important role in effectively constructing collaborative innovation linkages and the stable expansion of the scale of collaborative innovation within the innovation corridor.

Keywords: innovation corridor development, spatial structure, niche fitness evaluation model, head/tail breaks, innovation network

Procedia PDF Downloads 18
371 Product Separation of Green Processes and Catalyst Recycling of a Homogeneous Polyoxometalate Catalyst Using Nanofiltration Membranes

Authors: Dorothea Voß, Tobias Esser, Michael Huber, Jakob Albert

Abstract:

The growing world population and the associated increase in demand for energy and consumer goods, as well as increasing waste production, requires the development of sustainable processes. In addition, the increasing environmental awareness of our society is a driving force for the requirement that processes must be as resource and energy efficient as possible. In this context, the use of polyoxometalate catalysts (POMs) has emerged as a promising approach for the development of green processes. POMs are bifunctional polynuclear metal-oxo-anion cluster characterized by a strong Brønsted acidity, a high proton mobility combined with fast multi-electron transfer and tunable redox potential. In addition, POMs are soluble in many commonly known solvents and exhibit resistance to hydrolytic and oxidative degradation. Due to their structure and excellent physicochemical properties, POMs are efficient acid and oxidation catalysts that have attracted much attention in recent years. Oxidation processes with molecular oxygen are worth mentioning here. However, the fact that the POM catalysts are homogeneous poses a challenge for downstream processing of product solutions and recycling of the catalysts. In this regard, nanofiltration membranes have gained increasing interest in recent years, particularly due to their relative sustainability advantage over other technologies and their unique properties such as increased selectivity towards multivalent ions. In order to establish an efficient downstream process for the highly selective separation of homogeneous POM catalysts from aqueous solutions using nanofiltration membranes, a laboratory-scale membrane system was designed and constructed. By varying various process parameters, a sensitivity analysis was performed on a model system to develop an optimized method for the recovery of POM catalysts. From this, process-relevant key figures such as the rejection of various system components were derived. These results form the basis for further experiments on other systems to test the transferability to serval separation tasks with different POMs and products, as well as for recycling experiments of the catalysts in processes on laboratory scale.

Keywords: downstream processing, nanofiltration, polyoxometalates, homogeneous catalysis, green chemistry

Procedia PDF Downloads 87
370 Thermal Analysis of Adsorption Refrigeration System Using Silicagel–Methanol Pair

Authors: Palash Soni, Vivek Kumar Gaba, Shubhankar Bhowmick, Bidyut Mazumdar

Abstract:

Refrigeration technology is a fast developing field at the present era since it has very wide application in both domestic and industrial areas. It started from the usage of simple ice coolers to store food stuffs to the present sophisticated cold storages along with other air conditioning system. A variety of techniques are used to bring down the temperature below the ambient. Adsorption refrigeration technology is a novel, advanced and promising technique developed in the past few decades. It gained attention due to its attractive property of exploiting unlimited natural sources like solar energy, geothermal energy or even waste heat recovery from plants or from the exhaust of locomotives to fulfill its energy need. This will reduce the exploitation of non-renewable resources and hence reduce pollution too. This work is aimed to develop a model for a solar adsorption refrigeration system and to simulate the same for different operating conditions. In this system, the mechanical compressor is replaced by a thermal compressor. The thermal compressor uses renewable energy such as solar energy and geothermal energy which makes it useful for those areas where electricity is not available. Refrigerants normally in use like chlorofluorocarbon/perfluorocarbon have harmful effects like ozone depletion and greenhouse warming. It is another advantage of adsorption systems that it can replace these refrigerants with less harmful natural refrigerants like water, methanol, ammonia, etc. Thus the double benefit of reduction in energy consumption and pollution can be achieved. A thermodynamic model was developed for the proposed adsorber, and a universal MATLAB code was used to simulate the model. Simulations were carried out for a different operating condition for the silicagel-methanol working pair. Various graphs are plotted between regeneration temperature, adsorption capacities, the coefficient of performance, desorption rate, specific cooling power, adsorption/desorption times and mass. The results proved that adsorption system could be installed successfully for refrigeration purpose as it has saving in terms of power and reduction in carbon emission even though the efficiency is comparatively less as compared to conventional systems. The model was tested for its compliance in a cold storage refrigeration with a cooling load of 12 TR.

Keywords: adsorption, refrigeration, renewable energy, silicagel-methanol

Procedia PDF Downloads 205
369 Evaluation of Commercial Back-analysis Package in Condition Assessment of Railways

Authors: Shadi Fathi, Moura Mehravar, Mujib Rahman

Abstract:

Over the years,increased demands on railways, the emergence of high-speed trains and heavy axle loads, ageing, and deterioration of the existing tracks, is imposing costly maintenance actions on the railway sector. The need for developing a fast andcost-efficient non-destructive assessment method for the structural evaluation of railway tracksis therefore critically important. The layer modulus is the main parameter used in the structural design and evaluation of the railway track substructure (foundation). Among many recently developed NDTs, Falling Weight Deflectometer (FWD) test, widely used in pavement evaluation, has shown promising results for railway track substructure monitoring. The surface deflection data collected by FWD are used to estimate the modulus of substructure layers through the back-analysis technique. Although there are different commerciallyavailableback-analysis programs are used for pavement applications, there are onlya limited number of research-based techniques have been so far developed for railway track evaluation. In this paper, the suitability, accuracy, and reliability of the BAKFAAsoftware are investigated. The main rationale for selecting BAKFAA as it has a relatively straightforward user interfacethat is freely available and widely used in highway and airport pavement evaluation. As part of the study, a finite element (FE) model of a railway track section near Leominsterstation, Herefordshire, UK subjected to the FWD test, was developed and validated against available field data. Then, a virtual experimental database (including 218 sets of FWD testing data) was generated using theFE model and employed as the measured database for the BAKFAA software. This database was generated considering various layers’ moduli for each layer of track substructure over a predefined range. The BAKFAA predictions were compared against the cone penetration test (CPT) data (available from literature; conducted near to Leominster station same section as the FWD was performed). The results reveal that BAKFAA overestimatesthe layers’ moduli of each substructure layer. To adjust the BAKFA with the CPT data, this study introduces a correlation model to make the BAKFAA applicable in railway applications.

Keywords: back-analysis, bakfaa, railway track substructure, falling weight deflectometer (FWD), cone penetration test (CPT)

Procedia PDF Downloads 128
368 An Evaluation and Guidance for mHealth Apps

Authors: Tareq Aljaber

Abstract:

The number of mobile health apps is growing at a fast frequency as it's nearly doubled in a year between 2015 and 2016. Though, there is a lack of an effective evaluation framework to verify the usability and reliability of mobile phone health education applications which would help saving time and effort for the numerous user groups. This abstract describing a framework for evaluating mobile applications in specifically mobile health education applications, along with a guidance select tool to assist different users to select the most suitable mobile health education apps. The effective framework outcome is intended to meet the requirements and needs of the different stakeholder groups additionally to enhancing the development of mobile health education applications with software engineering approaches, by producing new and more effective techniques to evaluate such software. This abstract highlights the significance and consequences of mobile health education apps, before focusing the light on the required to create an effective evaluation framework for these apps. An explanation of the effective evaluation framework is going to be delivered in the abstract, beside with some specific evaluation metrics: an efficient hybrid of selected heuristic evaluation (HE) and usability evaluation (UE) metrics to enable the determination of the usefulness and usability of health education mobile apps. Moreover, an explanation of the qualitative and quantitative outcomes for the effective evaluation framework was accomplished using Epocrates mobile phone app in addition to some other mobile phone apps. This proposed framework-An Evaluation Framework for Mobile Health Education Apps-consists of a hybrid of 5 metrics designated from a larger set in usability evaluation and heuristic evaluation, illuminated grounded on 15 unstructured interviews from software developers (SD), health professionals (HP) and patients (P). These five metrics corresponding to explicit facets of usability recognised through a requirements analysis of typical stakeholders of mobile health apps. These five hybrid selected metrics were scattered across 24 specific questionnaire questions, which are available on request from first author. This questionnaire has been sent to 81 participants distributed in three sets of stakeholders from software developers (SD), health professionals (HP) and patients/general users (P/GU) on the purpose of ranking three sets of mobile health education applications. Finally, the outcomes from the questionnaire data helped us to approach our aims which are finding the profile for different stakeholders, finding the profile for different mobile health educations application packages, ranking different mobile health education application and guide us to build the select guidance too which is apart from the Evaluation Framework for Mobile Health Education Apps.

Keywords: evaluation framework, heuristic evaluation, usability evaluation, metrics

Procedia PDF Downloads 400
367 The Feminism of Data Privacy and Protection in Africa

Authors: Olayinka Adeniyi, Melissa Omino

Abstract:

The field of data privacy and data protection in Africa is still an evolving area, with many African countries yet to enact legislation on the subject. While African Governments are bringing their legislation to speed in this field, how patriarchy pervades every sector of African thought and manifests in society needs to be considered. Moreover, the laws enacted ought to be inclusive, especially towards women. This, in a nutshell, is the essence of data feminism. Data feminism is a new way of thinking about data science and data ethics that is informed by the ideas of intersectional feminism. Feminising data privacy and protection will involve thinking women, considering women in the issues of data privacy and protection, particularly in legislation, as is the case in this paper. The line of thought of women inclusion is not uncommon when even international and regional human rights specific for women only came long after the general human rights. The consideration is that these should have been inserted or rather included in the original general instruments in the first instance. Since legislation on data privacy is coming in this century, having seen the rights and shortcomings of earlier instruments, then the cue should be taken to ensure inclusive wholistic legislation for data privacy and protection in the first instance. Data feminism is arguably an area that has been scantily researched, albeit a needful one. With the spate of increase in the violence against women spiraling in the cyber world, compounding the issue of COVID-19 and the needful response of governments, and the effect of these on women and their rights, fast forward, the research on the feminism of data privacy and protection in Africa becomes inevitable. This paper seeks to answer the questions, what is data feminism in the African context, why is it important in the issue of data privacy and protection legislation; what are the laws, if any, existing on data privacy and protection in Africa, are they women inclusive, if not, why; what are the measures put in place for the privacy and protection of women in Africa, and how can this be made possible. The paper aims to investigate the issue of data privacy and protection in Africa, the legal framework, and the protection or provision that it has for women if any. It further aims to research the importance and necessity of feminizing data privacy and protection, the effect of lack of it, the challenges or bottlenecks in attaining this feat and the possibilities of accessing data privacy and protection for African women. The paper also researches the emerging practices of data privacy and protection of women in other jurisprudences. It approaches the research through the methodology of review of papers, analysis of laws, and reports. It seeks to contribute to the existing literature in the field and is explorative in its suggestion. It suggests a draft of some clauses to make any data privacy and protection legislation women inclusive. It would be useful for policymaking, academic, and public enlightenment.

Keywords: feminism, women, law, data, Africa

Procedia PDF Downloads 205
366 Design of Identification Based Adaptive Control for Fermentation Process in Bioreactor

Authors: J. Ritonja

Abstract:

The biochemical technology has been developing extremely fast since the middle of the last century. The main reason for such development represents a requirement for large production of high-quality biologically manufactured products such as pharmaceuticals, foods, and beverages. The impact of the biochemical industry on the world economy is enormous. The great importance of this industry also results in intensive development in scientific disciplines relevant to the development of biochemical technology. In addition to developments in the fields of biology and chemistry, which enable to understand complex biochemical processes, development in the field of control theory and applications is also very important. In the paper, the control for the biochemical reactor for the milk fermentation was studied. During the fermentation process, the biophysical quantities must be precisely controlled to obtain the high-quality product. To control these quantities, the bioreactor’s stirring drive and/or heating system can be used. Available commercial biochemical reactors are equipped with open loop or conventional linear closed loop control system. Due to the outstanding parameters variations and the partial nonlinearity of the biochemical process, the results obtained with these control systems are not satisfactory. To improve the fermentation process, the self-tuning adaptive control system was proposed. The use of the self-tuning adaptive control is suggested because the parameters’ variations of the studied biochemical process are very slow in most cases. To determine the linearized mathematical model of the fermentation process, the recursive least square identification method was used. Based on the obtained mathematical model the linear quadratic regulator was tuned. The parameters’ identification and the controller’s synthesis are executed on-line and adapt the controller’s parameters to the fermentation process’ dynamics during the operation. The use of the proposed combination represents the original solution for the control of the milk fermentation process. The purpose of the paper is to contribute to the progress of the control systems for the biochemical reactors. The proposed adaptive control system was tested thoroughly. From the obtained results it is obvious that the proposed adaptive control system assures much better following of the reference signal as a conventional linear control system with fixed control parameters.

Keywords: adaptive control, biochemical reactor, linear quadratic regulator, recursive least square identification

Procedia PDF Downloads 124
365 Experimental Study of Nucleate Pool Boiling Heat Transfer Characteristics on Laser-Processed Copper Surfaces of Different Patterns

Authors: Luvindran Sugumaran, Mohd Nashrul Mohd Zubir, Kazi Md Salim Newaz, Tuan Zaharinie Tuan Zahari, Suazlan Mt Aznam, Aiman Mohd Halil

Abstract:

With the fast growth of integrated circuits and the trend towards making electronic devices smaller, the heat dissipation load of electronic devices has continued to go over the limit. The high heat flux element would not only harm the operation and lifetime of the equipment but would also impede the performance upgrade brought about by the iteration of technological updates, which would have a direct negative impact on the economic and production cost benefits of rising industries. Hence, in high-tech industries like radar, information and communication, electromagnetic power, and aerospace, the development and implementation of effective heat dissipation technologies were urgently required. Pool boiling is favored over other cooling methods because of its capacity to dissipate a high heat flux at a low wall superheat without the usage of mechanical components. Enhancing the pool boiling performance by increasing the heat transfer coefficient via surface modification techniques has received a lot of attention. There are several surface modification methods feasible today, but the stability and durability of surface modification are the greatest priority. Thus, laser machining is an interesting choice for surface modification due to its low production cost, high scalability, and repeatability. In this study, different patterns of laser-processed copper surfaces are fabricated to investigate the nucleate pool boiling heat transfer performance of distilled water. The investigation showed that there is a significant enhancement in the pool boiling heat transfer performance of the laser-processed surface compared to the reference surface due to the notable increase in nucleation frequency and nucleation site density. It was discovered that the heat transfer coefficients increased when both the surface area ratio and the ratio of peak-to-valley height of the microstructure were raised. It is believed that the development of microstructures on the surface as a result of laser processing is the primary factor in the enhancement of heat transfer performance.

Keywords: heat transfer coefficient, laser processing, micro structured surface, pool boiling

Procedia PDF Downloads 86
364 Triploid Rainbow Trout (Oncorhynchus mykiss) for Better Aquaculture and Ecological Risk Management

Authors: N. N. Pandey, Raghvendra Singh, Biju S. Kamlam, Bipin K. Vishwakarma, Preetam Kala

Abstract:

The rainbow trout (Oncorhynchus mykiss) is an exotic salmonid fish, well known for its fast growth, tremendous ability to thrive in diverse conditions, delicious flesh and hard fighting nature in Europe and other countries. Rainbow trout farming has a great potential for its contribution to the mainstream economy of Himalayan states in India and other temperate countries. These characteristics establish them as one of the most widely introduced and cultured fish across the globe, and its farming is also prominent in the cold water regions of India. Nevertheless, genetic fatigue, slow growth, early maturity, and low productivity are limiting the expansion of trout production. Moreover, farms adjacent to natural streams or other water sources are subject to escape of domesticated rainbow trout into the wild, which is a serious environmental concern as the escaped fish is subject to contaminate and disrupt the receiving ecosystem. A decline in production traits due to early maturity prolongs the culture duration and affects the profit margin of rainbow trout farms in India. A viable strategy that could overcome these farming constraints in large scale operation is the production of triploid fish that are sterile and more heterozygous. For better triploidy induction rate (TR), heat shock at 28°C for 10 minutes and pressure shock 9500 psi pressure for 5 minutes is applied to green eggs with 90-100% of triploidy success and 72-80% survival upto swim-up fry stage. There is 20% better growth in aquaculture with triploids rainbow trout over diploids. As compared to wild diploid fish, larger sized and fitter triploid rainbow trout in natural waters attract to trout anglers, and support the development of recreational fisheries by state fisheries departments without the risk of contaminating existing gene pools and disrupting local fish diversity. Overall, enhancement of productivity in rainbow trout farms and trout production in coldwater regions, development of lucrative trout angling and better ecological management is feasible with triploid rainbow trout.

Keywords: rainbow trout, triploids fish, heat shock, pressure shock, trout angling

Procedia PDF Downloads 123
363 Effects of Lower and Upper Body Plyometric Training on Electrocardiogram Parameters of University Athletes

Authors: T. N. Uzor, C. O. Akosile, G. O. Emeahara

Abstract:

Plyometric training is a form of specialised strength training that uses fast muscular contractions to improve power and speed in sports conditioning by coaches and athletes. Despite its useful role in sports conditioning programme, the information about plyometric training on the athletes cardiovascular health especially Electrocardiogram (ECG) has not been established in the literature. The purpose of the study was to determine the effects of lower and upper body plyometric training on ECG of athletes. The study was guided by three null hypotheses. Quasi–experimental research design was adopted for the study. Seventy-two university male athletes constituted the population of the study. Thirty male athletes aged 18 to 24 years volunteered to participate in the study, but only twenty-three completed the study. The volunteered athletes were apparently healthy, physically active and free of any lower and upper extremity bone injuries for past one year and they had no medical or orthopedic injuries that may affect their participation in the study. Ten subjects were purposively assigned to one of the three groups: lower body plyometric training (LBPT), upper body plyometric training (UBPT), and control (C). Training consisted of six plyometric exercises: lower (ankle hops, squat jumps, tuck jumps) and upper body plyometric training (push-ups, medicine ball-chest throws and side throws) with moderate intensity. The general data were collated and analysed using Statistical Package for Social Science (SPSS version 22.0). The research questions were answered using mean and standard deviation, while paired samples t-test was also used to test for the hypotheses. The results revealed that athletes who were trained using LBPT had reduced ECG parameters better than those in the control group. The results also revealed that athletes who were trained using both LBPT and UBPT indicated lack of significant differences following ten weeks plyometric training than those in the control group in the ECG parameters except in Q wave, R wave and S wave (QRS) complex. Based on the findings of the study, it was recommended among others that coaches should include both LBPT and UBPT as part of athletes’ overall training programme from primary to tertiary institution to optimise performance as well as reduce the risk of cardiovascular diseases and promotes good healthy lifestyle.

Keywords: concentric, eccentric, electrocardiogram, plyometric

Procedia PDF Downloads 142
362 Adversarial Attacks and Defenses on Deep Neural Networks

Authors: Jonathan Sohn

Abstract:

Deep neural networks (DNNs) have shown state-of-the-art performance for many applications, including computer vision, natural language processing, and speech recognition. Recently, adversarial attacks have been studied in the context of deep neural networks, which aim to alter the results of deep neural networks by modifying the inputs slightly. For example, an adversarial attack on a DNN used for object detection can cause the DNN to miss certain objects. As a result, the reliability of DNNs is undermined by their lack of robustness against adversarial attacks, raising concerns about their use in safety-critical applications such as autonomous driving. In this paper, we focus on studying the adversarial attacks and defenses on DNNs for image classification. There are two types of adversarial attacks studied which are fast gradient sign method (FGSM) attack and projected gradient descent (PGD) attack. A DNN forms decision boundaries that separate the input images into different categories. The adversarial attack slightly alters the image to move over the decision boundary, causing the DNN to misclassify the image. FGSM attack obtains the gradient with respect to the image and updates the image once based on the gradients to cross the decision boundary. PGD attack, instead of taking one big step, repeatedly modifies the input image with multiple small steps. There is also another type of attack called the target attack. This adversarial attack is designed to make the machine classify an image to a class chosen by the attacker. We can defend against adversarial attacks by incorporating adversarial examples in training. Specifically, instead of training the neural network with clean examples, we can explicitly let the neural network learn from the adversarial examples. In our experiments, the digit recognition accuracy on the MNIST dataset drops from 97.81% to 39.50% and 34.01% when the DNN is attacked by FGSM and PGD attacks, respectively. If we utilize FGSM training as a defense method, the classification accuracy greatly improves from 39.50% to 92.31% for FGSM attacks and from 34.01% to 75.63% for PGD attacks. To further improve the classification accuracy under adversarial attacks, we can also use a stronger PGD training method. PGD training improves the accuracy by 2.7% under FGSM attacks and 18.4% under PGD attacks over FGSM training. It is worth mentioning that both FGSM and PGD training do not affect the accuracy of clean images. In summary, we find that PGD attacks can greatly degrade the performance of DNNs, and PGD training is a very effective way to defend against such attacks. PGD attacks and defence are overall significantly more effective than FGSM methods.

Keywords: deep neural network, adversarial attack, adversarial defense, adversarial machine learning

Procedia PDF Downloads 193
361 An Evaluation of Medical Waste in Health Facilities through Data Envelopment Analysis (DEA) Method: Turkey-Amasya Public Hospitals Union Model

Authors: Murat Iskender Aktaş, Sadi Ergin, Rasime Acar Aktaş

Abstract:

In the light of fast-paced changes and developments in the health sector, the Ministry of Health started a new structuring with decree law numbered 663 within the scope of the Project of Transformation in Health. Accordingly, hospitals should ensure patient satisfaction through more efficient, more effective use of resources and sustainable finance by placing patients in the centre and should operate to increase efficiency to its maximum level while doing these. Within this study, in order to find out how efficient the hospitals were in terms of medical waste management between the years 2011-2014, the data from six hospitals of Amasya Public Hospitals Union were evaluated separately through Data Envelopment Analysis (DEA) method. First of all, input variables were determined. Input variables were the number of patients admitted to polyclinics, the number of inpatients in clinics, the number of patients who were operated and the number of patients who applied to the laboratory. Output variable was the cost of medical wastes in Turkish liras. Each hospital’s total medical waste level before and after public hospitals union; the amounts of average medical waste per patient admitted to polyclinics, per inpatient in clinics, per patient admitted to laboratory and per operated patient were compared within each group. In addition, average medical waste levels and costs were compared for Turkey in general and Europe in general. Paired samples t-test was used to find out whether the changes (increase-decrease) after public hospitals union were statistically significant. The health facilities that were unsuccessful in terms of medical waste management before and after public hospital union and the factors that caused this failure were determined. Based on the results, for each health facility that was ineffective in terms of medical waste management, the level of improvement required for each input was determined. The results of the study showed that there was an improvement in medical waste management applications after the health facilities became a member of public hospitals union; their medical waste levels were lower than the average of Turkey and Europe while the averages of cost of disposal were the highest.

Keywords: medical waste management, cost of medical waste, public hospitals, data envelopment analysis

Procedia PDF Downloads 415