Search results for: gas distribution network
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9379

Search results for: gas distribution network

7459 Investigation of Processing Conditions on Rheological Features of Emulsion Gels and Oleogels Stabilized by Biopolymers

Authors: M. Sarraf, J. E. Moros, M. C. Sánchez

Abstract:

Oleogels are self-standing systems that are able to trap edible liquid oil into a tridimensional network and also help to use less fat by forming crystallization oleogelators. There are different ways to generate oleogelation and oil structuring, including direct dispersion, structured biphasic systems, oil sorption, and indirect method (emulsion-template). The selection of processing conditions as well as the composition of the oleogels is essential to obtain a stable oleogel with characteristics suitable for its purpose. In this sense, one of the ingredients widely used in food products to produce oleogels and emulsions is polysaccharides. Basil seed gum (BSG), with the scientific name Ocimum basilicum, is a new native polysaccharide with high viscosity and pseudoplastic behavior because of its high molecular weight in the food industry. Also, proteins can stabilize oil in water due to the presence of amino and carboxyl moieties that result in surface activity. Whey proteins are widely used in the food industry due to available, cheap ingredients, nutritional and functional characteristics such as emulsifier and a gelling agent, thickening, and water-binding capacity. In general, the interaction of protein and polysaccharides has a significant effect on the food structures and their stability, like the texture of dairy products, by controlling the interactions in macromolecular systems. Using edible oleogels as oil structuring helps for targeted delivery of a component trapped in a structural network. Therefore, the development of efficient oleogel is essential in the food industry. A complete understanding of the important points, such as the ratio oil phase, processing conditions, and concentrations of biopolymers that affect the formation and stability of the emulsion, can result in crucial information in the production of a suitable oleogel. In this research, the effects of oil concentration and pressure used in the manufacture of the emulsion prior to obtaining the oleogel have been evaluated through the analysis of droplet size and rheological properties of obtained emulsions and oleogels. The results show that the emulsion prepared in the high-pressure homogenizer (HPH) at higher pressure values has smaller droplet sizes and a higher uniformity in the size distribution curve. On the other hand, in relation to the rheological characteristics of the emulsions and oleogels obtained, the predominantly elastic character of the systems must be noted, as they present values of the storage modulus higher than those of losses, also showing an important plateau zone, typical of structured systems. In the same way, if steady-state viscous flow tests have been analyzed on both emulsions and oleogels, the result is that, once again, the pressure used in the homogenizer is an important factor for obtaining emulsions with adequate droplet size and the subsequent oleogel. Thus, various routes for trapping oil inside a biopolymer matrix with adjustable mechanical properties could be applied for the creation of the three-dimensional network in order to the oil absorption and creating oleogel.

Keywords: basil seed gum, particle size, viscoelastic properties, whey protein

Procedia PDF Downloads 70
7458 The Market Structure Simulation of Heterogenous Firms

Authors: Arunas Burinskas, Manuela Tvaronavičienė

Abstract:

Although the new trade theories, unlike the theories of an industrial organisation, see the structure of the market and competition between enterprises through their heterogeneity according to various parameters, they do not pay any particular attention to the analysis of the market structure and its development. In this article, although we relied mainly on models developed by the scholars of new trade theory, we proposed a different approach. In our simulation model, we model market demand according to normal distribution function, while on the supply side (as it is in the new trade theory models), productivity is modeled with the Pareto distribution function. The results of the simulation show that companies with higher productivity (lower marginal costs) do not pass on all the benefits of such economies to buyers. However, even with higher marginal costs, firms can choose to offer higher value-added goods to stay in the market. In general, the structure of the market is formed quickly enough and depends on the skills available to firms.

Keywords: market, structure, simulation, heterogenous firms

Procedia PDF Downloads 152
7457 The Effects of Street Network Layout on Walking to School

Authors: Ayse Ozbil, Gorsev Argin, Demet Yesiltepe

Abstract:

Data for this cross-sectional study were drawn from questionnaires conducted in 10 elementary schools (1000 students, ages 12-14) located in Istanbul, Turkey. School environments (1600 meter buffers around the school) were evaluated through GIS-based land-use data (parcel level land use density) and street-level topography. Street networks within the same buffers were evaluated by using angular segment analysis (Integration and Choice) implemented in Depthmap as well as two segment-based connectivity measures, namely Metric and Directional Reach implemented in GIS. Segment Angular Integration measures how accessible each space from all the others within the radius using the least angle measure of distance. Segment Angular Choice which measures how many times a space is selected on journeys between all pairs of origins and destinations. Metric Reach captures the density of streets and street connections accessible from each individual road segment. Directional Reach measures the extent to which the entire street network is accessible with few direction changes. In addition, socio-economic characteristics (annual income, car ownership, education-level) of parents, obtained from parental questionnaires, were also included in the analysis. It is shown that surrounding street network configuration is strongly associated with both walk-mode shares and average walking distances to/from schools when controlling for parental socio-demographic attributes as well as land-use compositions and topographic features in school environments. More specifically, findings suggest that the scale at which urban form has an impact on pedestrian travel is considerably larger than a few blocks around the school.

Keywords: Istanbul, street network layout, urban form, walking to/from school

Procedia PDF Downloads 413
7456 Quantum Inspired Security on a Mobile Phone

Authors: Yu Qin, Wanjiaman Li

Abstract:

The widespread use of mobile electronic devices increases the complexities of mobile security. This thesis aims to provide a secure communication environment for smartphone users. Some research proves that the one-time pad is one of the securest encryption methods, and that the key distribution problem can be solved by using the QKD (quantum key distribution). The objective of this project is to design an Android APP (application) to exchange several random keys between mobile phones. Inspired by QKD, the developed APP uses the quick response (QR) code as a carrier to dispatch large amounts of one-time keys. After evaluating the performance of APP, it allows the mobile phone to capture and decode 1800 bytes of random data in 600ms. The continuous scanning mode of APP is designed to improve the overall transmission performance and user experience, and the maximum transmission rate of this mode is around 2200 bytes/s. The omnidirectional readability and error correction capability of QR code gives it a better real-life application, and the features of adequate storage capacity and quick response optimize overall transmission efficiency. The security of this APP is guaranteed since QR code is exchanged face-to-face, eliminating the risk of being eavesdropped. Also, the id of QR code is the only message that would be transmitted through the whole communication. The experimental results show this project can achieve superior transmission performance, and the correlation between the transmission rate of the system and several parameters, such as the QR code size, has been analyzed. In addition, some existing technologies and the main findings in the context of the project are summarized and critically compared in detail.

Keywords: one-time pad, QKD (quantum key distribution), QR code, application

Procedia PDF Downloads 148
7455 Enhancing Scalability in Ethereum Network Analysis: Methods and Techniques

Authors: Stefan K. Behfar

Abstract:

The rapid growth of the Ethereum network has brought forth the urgent need for scalable analysis methods to handle the increasing volume of blockchain data. In this research, we propose efficient methodologies for making Ethereum network analysis scalable. Our approach leverages a combination of graph-based data representation, probabilistic sampling, and parallel processing techniques to achieve unprecedented scalability while preserving critical network insights. Data Representation: We develop a graph-based data representation that captures the underlying structure of the Ethereum network. Each block transaction is represented as a node in the graph, while the edges signify temporal relationships. This representation ensures efficient querying and traversal of the blockchain data. Probabilistic Sampling: To cope with the vastness of the Ethereum blockchain, we introduce a probabilistic sampling technique. This method strategically selects a representative subset of transactions and blocks, allowing for concise yet statistically significant analysis. The sampling approach maintains the integrity of the network properties while significantly reducing the computational burden. Graph Convolutional Networks (GCNs): We incorporate GCNs to process the graph-based data representation efficiently. The GCN architecture enables the extraction of complex spatial and temporal patterns from the sampled data. This combination of graph representation and GCNs facilitates parallel processing and scalable analysis. Distributed Computing: To further enhance scalability, we adopt distributed computing frameworks such as Apache Hadoop and Apache Spark. By distributing computation across multiple nodes, we achieve a significant reduction in processing time and enhanced memory utilization. Our methodology harnesses the power of parallelism, making it well-suited for large-scale Ethereum network analysis. Evaluation and Results: We extensively evaluate our methodology on real-world Ethereum datasets covering diverse time periods and transaction volumes. The results demonstrate its superior scalability, outperforming traditional analysis methods. Our approach successfully handles the ever-growing Ethereum data, empowering researchers and developers with actionable insights from the blockchain. Case Studies: We apply our methodology to real-world Ethereum use cases, including detecting transaction patterns, analyzing smart contract interactions, and predicting network congestion. The results showcase the accuracy and efficiency of our approach, emphasizing its practical applicability in real-world scenarios. Security and Robustness: To ensure the reliability of our methodology, we conduct thorough security and robustness evaluations. Our approach demonstrates high resilience against adversarial attacks and perturbations, reaffirming its suitability for security-critical blockchain applications. Conclusion: By integrating graph-based data representation, GCNs, probabilistic sampling, and distributed computing, we achieve network scalability without compromising analytical precision. This approach addresses the pressing challenges posed by the expanding Ethereum network, opening new avenues for research and enabling real-time insights into decentralized ecosystems. Our work contributes to the development of scalable blockchain analytics, laying the foundation for sustainable growth and advancement in the domain of blockchain research and application.

Keywords: Ethereum, scalable network, GCN, probabilistic sampling, distributed computing

Procedia PDF Downloads 81
7454 Comparative Evaluation of Pharmacologically Guided Approaches (PGA) to Determine Maximum Recommended Starting Dose (MRSD) of Monoclonal Antibodies for First Clinical Trial

Authors: Ibraheem Husain, Abul Kalam Najmi, Karishma Chester

Abstract:

First-in-human (FIH) studies are a critical step in clinical development of any molecule that has shown therapeutic promise in preclinical evaluations, since preclinical research and safety studies into clinical development is a crucial step for successful development of monoclonal antibodies for guidance in pharmaceutical industry for the treatment of human diseases. Therefore, comparison between USFDA and nine pharmacologically guided approaches (PGA) (simple allometry, maximum life span potential, brain weight, rule of exponent (ROE), two species methods and one species methods) were made to determine maximum recommended starting dose (MRSD) for first in human clinical trials using four drugs namely Denosumab, Bevacizumab, Anakinra and Omalizumab. In our study, the predicted pharmacokinetic (pk) parameters and the estimated first-in-human dose of antibodies were compared with the observed human values. The study indicated that the clearance and volume of distribution of antibodies can be predicted with reasonable accuracy in human and a good estimate of first human dose can be obtained from the predicted human clearance and volume of distribution. A pictorial method evaluation chart was also developed based on fold errors for simultaneous evaluation of various methods.

Keywords: clinical pharmacology (CPH), clinical research (CRE), clinical trials (CTR), maximum recommended starting dose (MRSD), clearance and volume of distribution

Procedia PDF Downloads 378
7453 Dynamic Economic Load Dispatch Using Quadratic Programming: Application to Algerian Electrical Network

Authors: A. Graa, I. Ziane, F. Benhamida, S. Souag

Abstract:

This paper presents a comparative analysis study of an efficient and reliable quadratic programming (QP) to solve economic load dispatch (ELD) problem with considering transmission losses in a power system. The proposed QP method takes care of different unit and system constraints to find optimal solution. To validate the effectiveness of the proposed QP solution, simulations have been performed using Algerian test system. Results obtained with the QP method have been compared with other existing relevant approaches available in literatures. Experimental results show a proficiency of the QP method over other existing techniques in terms of robustness and its optimal search.

Keywords: economic dispatch, quadratic programming, Algerian network, dynamic load

Procedia PDF Downloads 569
7452 Evaluation of the Internal Quality for Pineapple Based on the Spectroscopy Approach and Neural Network

Authors: Nonlapun Meenil, Pisitpong Intarapong, Thitima Wongsheree, Pranchalee Samanpiboon

Abstract:

In Thailand, once pineapples are harvested, they must be classified into two classes based on their sweetness: sweet and unsweet. This paper has studied and developed the assessment of internal quality of pineapples using a low-cost compact spectroscopy sensor according to the Spectroscopy approach and Neural Network (NN). During the experiments, Batavia pineapples were utilized, generating 100 samples. The extracted pineapple juice of each sample was used to determine the Soluble Solid Content (SSC) labeling into sweet and unsweet classes. In terms of experimental equipment, the sensor cover was specifically designed to install the sensor and light source to read the reflectance at a five mm depth from pineapple flesh. By using a spectroscopy sensor, data on visible and near-infrared reflectance (Vis-NIR) were collected. The NN was used to classify the pineapple classes. Before the classification step, the preprocessing methods, which are Class balancing, Data shuffling, and Standardization were applied. The 510 nm and 900 nm reflectance values of the middle parts of pineapples were used as features of the NN. With the Sequential model and Relu activation function, 100% accuracy of the training set and 76.67% accuracy of the test set were achieved. According to the abovementioned information, using a low-cost compact spectroscopy sensor has achieved favorable results in classifying the sweetness of the two classes of pineapples.

Keywords: neural network, pineapple, soluble solid content, spectroscopy

Procedia PDF Downloads 83
7451 Calibration of Hybrid Model and Arbitrage-Free Implied Volatility Surface

Authors: Kun Huang

Abstract:

This paper investigates whether the combination of local and stochastic volatility models can be calibrated exactly to any arbitrage-free implied volatility surface of European option. The risk neutral Brownian Bridge density is applied for calibration of the leverage function of our Hybrid model. Furthermore, the tails of marginal risk neutral density are generated by Generalized Extreme Value distribution in order to capture the properties of asset returns. The local volatility is generated from the arbitrage-free implied volatility surface using stochastic volatility inspired parameterization.

Keywords: arbitrage free implied volatility, calibration, extreme value distribution, hybrid model, local volatility, risk-neutral density, stochastic volatility

Procedia PDF Downloads 272
7450 Conventional Four Steps Travel Demand Modeling for Kabul New City

Authors: Ahmad Mansoor Stanikzai, Yoshitaka Kajita

Abstract:

This research is a very essential towards transportation planning of Kabul New City. In this research, the travel demand of Kabul metropolitan area (Existing and Kabul New City) are evaluated for three different target years (2015, current, 2025, mid-term, 2040, long-term). The outcome of this study indicates that, though currently the vehicle volume is less the capacity of existing road networks, Kabul city is suffering from daily traffic congestions. This is mainly due to lack of transportation management, the absence of proper policies, improper public transportation system and violation of traffic rules and regulations by inhabitants. On the other hand, the observed result indicates that the current vehicle to capacity ratio (VCR) which is the most used index to judge traffic status in the city is around 0.79. This indicates the inappropriate traffic condition of the city. Moreover, by the growth of population in mid-term (2025) and long-term (2040) and in the case of no development in the road network and transportation system, the VCR value will dramatically increase to 1.40 (2025) and 2.5 (2040). This can be a critical situation for an urban area from an urban transportation perspective. Thus, by introducing high-capacity public transportation system and the development of road network in Kabul New City and integrating these links with the existing city road network, significant improvements were observed in the value of VCR.

Keywords: Afghanistan, Kabul new city, planning, policy, urban transportation

Procedia PDF Downloads 334
7449 Voltage Profile Enhancement in the Unbalanced Distribution Systems during Fault Conditions

Authors: K. Jithendra Gowd, Ch. Sai Babu, S. Sivanagaraju

Abstract:

Electric power systems are daily exposed to service interruption mainly due to faults and human accidental interference. Short circuit currents are responsible for several types of disturbances in power systems. The fault currents are high and the voltages are reduced at the time of fault. This paper presents two suitable methods, consideration of fault resistance and Distributed Generator are implemented and analyzed for the enhancement of voltage profile during fault conditions. Fault resistance is a critical parameter of electric power systems operation due to its stochastic nature. If not considered, this parameter may interfere in fault analysis studies and protection scheme efficiency. The effect of Distributed Generator is also considered. The proposed methods are tested on the IEEE 37 bus test systems and the results are compared.

Keywords: distributed generation, electrical distribution systems, fault resistance

Procedia PDF Downloads 519
7448 Non-intrusive Hand Control of Drone Using an Inexpensive and Streamlined Convolutional Neural Network Approach

Authors: Evan Lowhorn, Rocio Alba-Flores

Abstract:

The purpose of this work is to develop a method for classifying hand signals and using the output in a drone control algorithm. To achieve this, methods based on Convolutional Neural Networks (CNN) were applied. CNN's are a subset of deep learning, which allows grid-like inputs to be processed and passed through a neural network to be trained for classification. This type of neural network allows for classification via imaging, which is less intrusive than previous methods using biosensors, such as EMG sensors. Classification CNN's operate purely from the pixel values in an image; therefore they can be used without additional exteroceptive sensors. A development bench was constructed using a desktop computer connected to a high-definition webcam mounted on a scissor arm. This allowed the camera to be pointed downwards at the desk to provide a constant solid background for the dataset and a clear detection area for the user. A MATLAB script was created to automate dataset image capture at the development bench and save the images to the desktop. This allowed the user to create their own dataset of 12,000 images within three hours. These images were evenly distributed among seven classes. The defined classes include forward, backward, left, right, idle, and land. The drone has a popular flip function which was also included as an additional class. To simplify control, the corresponding hand signals chosen were the numerical hand signs for one through five for movements, a fist for land, and the universal “ok” sign for the flip command. Transfer learning with PyTorch (Python) was performed using a pre-trained 18-layer residual learning network (ResNet-18) to retrain the network for custom classification. An algorithm was created to interpret the classification and send encoded messages to a Ryze Tello drone over its 2.4 GHz Wi-Fi connection. The drone’s movements were performed in half-meter distance increments at a constant speed. When combined with the drone control algorithm, the classification performed as desired with negligible latency when compared to the delay in the drone’s movement commands.

Keywords: classification, computer vision, convolutional neural networks, drone control

Procedia PDF Downloads 217
7447 Hybrid Approach for Controlling Inductive Load Fed by a Multicellular Converter by Using the Petri Nets

Authors: I. Bentchikou, A. Tlemcani, F. Boudjema, D. Boukhetala, N. Ould Cherchali

Abstract:

In this paper, hybrid approach is proposed to regulate the voltages of the floating capacitor multicell inverter and the current in the load. This structure makes it possible to ensure the distribution of the voltage stresses on the various low-voltage semiconductor components connected in series. And as the problem and to keep a constant voltage across the capacitors. Thus, it is necessary to ensure a distribution balanced voltages at the terminals of floating capacitors thanks to Algorithm develop for this, using the Petri nets. So we consider a three-cell converter represented as a hybrid system with eight modes of operation. The operating modes of the system are governed by the control reference voltage and a reference current. Finally, we present the results of the simulation with MATLAB/SIMULINK to illustrate the performances of this approach.

Keywords: hybrid control, floating condensers, multicellular converter, petri nets

Procedia PDF Downloads 131
7446 Secured Cancer Care and Cloud Services in Internet of Things /Wireless Sensor Network Based Medical Systems

Authors: Adeniyi Onasanya, Maher Elshakankiri

Abstract:

In recent years, the Internet of Things (IoT) has constituted a driving force of modern technological advancement, and it has become increasingly common as its impacts are seen in a variety of application domains, including healthcare. IoT is characterized by the interconnectivity of smart sensors, objects, devices, data, and applications. With the unprecedented use of IoT in industrial, commercial and domestic, it becomes very imperative to harness the benefits and functionalities associated with the IoT technology in (re)assessing the provision and positioning of healthcare to ensure efficient and improved healthcare delivery. In this research, we are focusing on two important services in healthcare systems, which are cancer care services and business analytics/cloud services. These services incorporate the implementation of an IoT that provides solution and framework for analyzing health data gathered from IoT through various sensor networks and other smart devices in order to improve healthcare delivery and to help health care providers in their decision-making process for enhanced and efficient cancer treatment. In addition, we discuss the wireless sensor network (WSN), WSN routing and data transmission in the healthcare environment. Finally, some operational challenges and security issues with IoT-based healthcare system are discussed.

Keywords: IoT, smart health care system, business analytics, (wireless) sensor network, cancer care services, cloud services

Procedia PDF Downloads 180
7445 Form of Distribution of Traffic Accident and Environment Factors of Road Affecting of Traffic Accident in Dusit District, Only Area Responsible of Samsen Police Station

Authors: Musthaya Patchanee

Abstract:

This research aimed to study form of traffic distribution and environmental factors of road that affect traffic accidents in Dusit District, only areas responsible of Samsen Police Station. Data used in this analysis is the secondary data of traffic accident case from year 2011. Observed area units are 15 traffic lines that are under responsible of Samsen Police Station. Technique and method used are the Cartographic Method, the Correlation Analysis, and the Multiple Regression Analysis. The results of form of traffic accidents show that, the Samsen Road area had most traffic accidents (24.29%), second was Rachvithi Road (18.10%), third was Sukhothai Road (15.71%), fourth was Rachasrima Road (12.38%), and fifth was Amnuaysongkram Road (7.62%). The result from Dusit District, only areas responsible of Samsen police station, has suggested that the scale of accidents have high positive correlation with statistic significant at level 0.05 and the frequency of travel (r=0.857). Traffic intersection point (r=0.763)and traffic control equipments (r=0.713) are relevant factors respectively. By using the Multiple Regression Analysis, travel frequency is the only one that has considerable influences on traffic accidents in Dusit district only Samsen Police Station area. Also, a factor in frequency of travel can explain the change in traffic accidents scale to 73.40 (R2 = 0.734). By using the Multiple regression summation from analysis was Y ̂=-7.977+0.044X6.

Keywords: form of traffic distribution, environmental factors of road, traffic accidents, Dusit district

Procedia PDF Downloads 394
7444 Comparison of Receiver Operating Characteristic Curve Smoothing Methods

Authors: D. Sigirli

Abstract:

The Receiver Operating Characteristic (ROC) curve is a commonly used statistical tool for evaluating the diagnostic performance of screening and diagnostic test with continuous or ordinal scale results which aims to predict the presence or absence probability of a condition, usually a disease. When the test results were measured as numeric values, sensitivity and specificity can be computed across all possible threshold values which discriminate the subjects as diseased and non-diseased. There are infinite numbers of possible decision thresholds along the continuum of the test results. The ROC curve presents the trade-off between sensitivity and the 1-specificity as the threshold changes. The empirical ROC curve which is a non-parametric estimator of the ROC curve is robust and it represents data accurately. However, especially for small sample sizes, it has a problem of variability and as it is a step function there can be different false positive rates for a true positive rate value and vice versa. Besides, the estimated ROC curve being in a jagged form, since the true ROC curve is a smooth curve, it underestimates the true ROC curve. Since the true ROC curve is assumed to be smooth, several smoothing methods have been explored to smooth a ROC curve. These include using kernel estimates, using log-concave densities, to fit parameters for the specified density function to the data with the maximum-likelihood fitting of univariate distributions or to create a probability distribution by fitting the specified distribution to the data nd using smooth versions of the empirical distribution functions. In the present paper, we aimed to propose a smooth ROC curve estimation based on the boundary corrected kernel function and to compare the performances of ROC curve smoothing methods for the diagnostic test results coming from different distributions in different sample sizes. We performed simulation study to compare the performances of different methods for different scenarios with 1000 repetitions. It is seen that the performance of the proposed method was typically better than that of the empirical ROC curve and only slightly worse compared to the binormal model when in fact the underlying samples were generated from the normal distribution.

Keywords: empirical estimator, kernel function, smoothing, receiver operating characteristic curve

Procedia PDF Downloads 155
7443 Association of Transforming Growth Factor-β1 Gene 1800469 C > T and 1982073 C > T Polymorphism with Type 2 Diabetic Foot Ulcer Patient in Cipto Mangunkusumo National Hospital Jakarta

Authors: Dedy Pratama, Akhmadu Muradi, Hilman Ibrahim, Patrianef Darwis, Alexander Jayadi Utama, Raden Suhartono, D. Suryandari, Luluk Yunaini, Tom Ch Adriani

Abstract:

Objective: Diabetic Foot Ulcer (DFU) is one of the complications of Type 2 Diabetes Mellitus (T2DM) that can lead to disability and death. Inadequate vascularization condition will affect healing process of DFU. Therefore, we investigated the expression of polymorphism TGF- β1 in the relation of the occurrence of DFU in T2DM. Methods: We designed a case-control study to investigate the polymorphism TGF- β1 gene 1800469 C > T and 1982073 C > T in T2DM in Cipto Mangunkusumo National Hospital (RSCM) Jakarta from June to December 2016. We used PCR techniques and compared the results in a group of T2DM patients with DFU as the case study and without DFU as the control group. Results: There were 203 patients, 102 patients with DFU and 101 patients control without DFU. 49,8% is male and 50,2% female with mean age about 56 years. Distribution of wild-type genotype TGF-B1 1800469 C > T wild type CC was found in 44,8%, the number of mutant heterozygote CT was 10,8% and mutant homozygote is 11,3%. Distribution of TGF-B1 1982073 C>T wild type CC was 32,5%, mutant heterozygote is 38,9% and mutant homozygote 25,1%. Conclusion: Distribution of alleles from TGF-B1 1800469 C > T is C 75% and T 25% and from TGF-B1 1982073 C > T is C53,8% and T 46,2%. In the other word polymorphism TGF- β1 plays a role in the occurrence and healing process of the DFU in T2DM patients.

Keywords: diabetic foot ulcers, diabetes mellitus, polymorphism, TGF-β1

Procedia PDF Downloads 291
7442 Nelder-Mead Parametric Optimization of Elastic Metamaterials with Artificial Neural Network Surrogate Model

Authors: Jiaqi Dong, Qing-Hua Qin, Yi Xiao

Abstract:

Some of the most fundamental challenges of elastic metamaterials (EMMs) optimization can be attributed to the high consumption of computational power resulted from finite element analysis (FEA) simulations that render the optimization process inefficient. Furthermore, due to the inherent mesh dependence of FEA, minuscule geometry features, which often emerge during the later stages of optimization, induce very fine elements, resulting in enormously high time consumption, particularly when repetitive solutions are needed for computing the objective function. In this study, a surrogate modelling algorithm is developed to reduce computational time in structural optimization of EMMs. The surrogate model is constructed based on a multilayer feedforward artificial neural network (ANN) architecture, trained with prepopulated eigenfrequency data prepopulated from FEA simulation and optimized through regime selection with genetic algorithm (GA) to improve its accuracy in predicting the location and width of the primary elastic band gap. With the optimized ANN surrogate at the core, a Nelder-Mead (NM) algorithm is established and its performance inspected in comparison to the FEA solution. The ANNNM model shows remarkable accuracy in predicting the band gap width and a reduction of time consumption by 47%.

Keywords: artificial neural network, machine learning, mechanical metamaterials, Nelder-Mead optimization

Procedia PDF Downloads 131
7441 Classification of Generative Adversarial Network Generated Multivariate Time Series Data Featuring Transformer-Based Deep Learning Architecture

Authors: Thrivikraman Aswathi, S. Advaith

Abstract:

As there can be cases where the use of real data is somehow limited, such as when it is hard to get access to a large volume of real data, we need to go for synthetic data generation. This produces high-quality synthetic data while maintaining the statistical properties of a specific dataset. In the present work, a generative adversarial network (GAN) is trained to produce multivariate time series (MTS) data since the MTS is now being gathered more often in various real-world systems. Furthermore, the GAN-generated MTS data is fed into a transformer-based deep learning architecture that carries out the data categorization into predefined classes. Further, the model is evaluated across various distinct domains by generating corresponding MTS data.

Keywords: GAN, transformer, classification, multivariate time series

Procedia PDF Downloads 135
7440 Using Dynamic Bayesian Networks to Characterize and Predict Job Placement

Authors: Xupin Zhang, Maria Caterina Bramati, Enrest Fokoue

Abstract:

Understanding the career placement of graduates from the university is crucial for both the qualities of education and ultimate satisfaction of students. In this research, we adapt the capabilities of dynamic Bayesian networks to characterize and predict students’ job placement using data from various universities. We also provide elements of the estimation of the indicator (score) of the strength of the network. The research focuses on overall findings as well as specific student groups including international and STEM students and their insight on the career path and what changes need to be made. The derived Bayesian network has the potential to be used as a tool for simulating the career path for students and ultimately helps universities in both academic advising and career counseling.

Keywords: dynamic bayesian networks, indicator estimation, job placement, social networks

Procedia PDF Downloads 386
7439 Task Based Functional Connectivity within Reward Network in Food Image Viewing Paradigm Using Functional MRI

Authors: Preetham Shankapal, Jill King, Kori Murray, Corby Martin, Paula Giselman, Jason Hicks, Owen Carmicheal

Abstract:

Activation of reward and satiety networks in the brain while processing palatable food cues, as well as functional connectivity during rest has been studied using functional Magnetic Resonance Imaging of the brain in various obesity phenotypes. However, functional connectivity within the reward and satiety network during food cue processing is understudied. 14 obese individuals underwent two fMRI scans during viewing of Macronutrient Picture System images. Each scan included two blocks of images of High Sugar/High Fat (HSHF), High Carbohydrate/High Fat (HCHF), Low Sugar/Low Fat (LSLF) and also non-food images. Seed voxels within seven food reward relevant ROIs: Insula, putamen and cingulate, precentral, parahippocampal, medial frontal and superior temporal gyri were isolated based on a prior meta-analysis. Beta series correlation for task-related functional connectivity between these seed voxels and the rest of the brain was computed. Voxel-level differences in functional connectivity were calculated between: first and the second scan; individuals who saw novel (N=7) vs. Repeated (N=7) images in the second scan; and between the HC/HF, HSHF blocks vs LSLF and non-food blocks. Computations and analysis showed that during food image viewing, reward network ROIs showed significant functional connectivity with each other and with other regions responsible for attentional and motor control, including inferior parietal lobe and precentral gyrus. These functional connectivity values were heightened among individuals who viewed novel HS/HF images in the second scan. In the second scan session, functional connectivity was reduced within the reward network but increased within attention, memory and recognition regions, suggesting habituation to reward properties and increased recollection of previously viewed images. In conclusion it can be inferred that Functional Connectivity within reward network and between reward and other brain regions, varies by important experimental conditions during food photography viewing, including habituation to shown foods.

Keywords: fMRI, functional connectivity, task-based, beta series correlation

Procedia PDF Downloads 275
7438 Assessing the Influence of Station Density on Geostatistical Prediction of Groundwater Levels in a Semi-arid Watershed of Karnataka

Authors: Sakshi Dhumale, Madhushree C., Amba Shetty

Abstract:

The effect of station density on the geostatistical prediction of groundwater levels is of critical importance to ensure accurate and reliable predictions. Monitoring station density directly impacts the accuracy and reliability of geostatistical predictions by influencing the model's ability to capture localized variations and small-scale features in groundwater levels. This is particularly crucial in regions with complex hydrogeological conditions and significant spatial heterogeneity. Insufficient station density can result in larger prediction uncertainties, as the model may struggle to adequately represent the spatial variability and correlation patterns of the data. On the other hand, an optimal distribution of monitoring stations enables effective coverage of the study area and captures the spatial variability of groundwater levels more comprehensively. In this study, we investigate the effect of station density on the predictive performance of groundwater levels using the geostatistical technique of Ordinary Kriging. The research utilizes groundwater level data collected from 121 observation wells within the semi-arid Berambadi watershed, gathered over a six-year period (2010-2015) from the Indian Institute of Science (IISc), Bengaluru. The dataset is partitioned into seven subsets representing varying sampling densities, ranging from 15% (12 wells) to 100% (121 wells) of the total well network. The results obtained from different monitoring networks are compared against the existing groundwater monitoring network established by the Central Ground Water Board (CGWB). The findings of this study demonstrate that higher station densities significantly enhance the accuracy of geostatistical predictions for groundwater levels. The increased number of monitoring stations enables improved interpolation accuracy and captures finer-scale variations in groundwater levels. These results shed light on the relationship between station density and the geostatistical prediction of groundwater levels, emphasizing the importance of appropriate station densities to ensure accurate and reliable predictions. The insights gained from this study have practical implications for designing and optimizing monitoring networks, facilitating effective groundwater level assessments, and enabling sustainable management of groundwater resources.

Keywords: station density, geostatistical prediction, groundwater levels, monitoring networks, interpolation accuracy, spatial variability

Procedia PDF Downloads 69
7437 Community Health Commodities Distribution of integrated HIV and Non-Communicable Disease Services during COVID-19 Pandemic – Eswatini Case Study

Authors: N. Dlamini, Mpumelelo G. Ndlela, Philisiwe Dlamini, Nicholus Kisyeri, Bhekizitha Sithole

Abstract:

Accessing health services during the COVID-19 pandemic have exacerbated scarcity to routine medication. To ensure continuous accessibility to services, Eswatini launched Community Health Commodities Distribution (CHCD). Eligible Antiretroviral Therapy(ART) stable clients (VL<1,000) and patients on Non-Communicable Disease (NCD) medications were attended at community pick up points (PUP) based on distance between clients’ residence and the public health facility. Services provided includes ART and Pre-Exposure prophylaxis (PrEP) refills and NCD drug refills). The number of community PUP was 14% higher than health facility visits. Among all medications and commodities distributed between April and October 2020 at the PUP, 64% were HIV-related (HIV rapid test, HIVST, VL test, PrEP meds), and 36% were NCD related. The rapid roll out of CHCD during COVID-19 pandemic reduced the risk of COVID-19 transmission to clients as travel to health facilities was eliminated. It Additionally increased access to commodities during COVID-19-driven lockdown, decongested health facilities, integrated model of care, and increase service coverage. It was also noted that CHCD added different curative and HIV related services based on client specific needs and availability of the commodities.

Keywords: community health commodities distribution, pick up points, antiretroviral therapy, pre-exposure prophylaxis

Procedia PDF Downloads 143
7436 Microfluidic Fluid Shear Mechanotransduction Device Using Linear Optimization of Hydraulic Channels

Authors: Sanat K. Dash, Rama S. Verma, Sarit K. Das

Abstract:

A logarithmic microfluidic shear device was designed and fabricated for cellular mechanotransduction studies. The device contains four cell culture chambers in which flow was modulated to achieve a logarithmic increment. Resistance values were optimized to make the device compact. The network of resistances was developed according to a unique combination of series and parallel resistances as found via optimization. Simulation results done in Ansys 16.1 matched the analytical calculations and showed the shear stress distribution at different inlet flow rates. Fabrication of the device was carried out using conventional photolithography and PDMS soft lithography. Flow profile was validated taking DI water as working fluid and measuring the volume collected at all four outlets. Volumes collected at the outlets were in accordance with the simulation results at inlet flow rates ranging from 1 ml/min to 0.1 ml/min. The device can exert fluid shear stresses ranging four orders of magnitude on the culture chamber walls which will cover shear stress values from interstitial flow to blood flow. This will allow studying cell behavior in the long physiological range of shear stress in a single run reducing number of experiments.

Keywords: microfluidics, mechanotransduction, fluid shear stress, physiological shear

Procedia PDF Downloads 133
7435 Chairussyuhur Arman, Totti Tjiptosumirat, Muhammad Gunawan, Mastur, Joko Priyono, Baiq Tri Ratna Erawati

Authors: Maria M. Giannakou, Athanasios K. Ziliaskopoulos

Abstract:

Transmission pipelines carrying natural gas are often routed through populated cities, industrial and environmentally sensitive areas. While the need for these networks is unquestionable, there are serious concerns about the risk these lifeline networks pose to the people, to their habitat and to the critical infrastructures, especially in view of natural disasters such as earthquakes. This work presents an Integrated Pipeline Risk Management methodology (IPRM) for assessing the hazard associated with a natural gas pipeline failure due to natural or manmade disasters. IPRM aims to optimize the allocation of the available resources to countermeasures in order to minimize the impacts of pipeline failure to humans, the environment, the infrastructure and the economic activity. A proposed knapsack mathematical programming formulation is introduced that optimally selects the proper mitigation policies based on the estimated cost – benefit ratios. The proposed model is demonstrated with a small numerical example. The vulnerability analysis of these pipelines and the quantification of consequences from such failures can be useful for natural gas industries on deciding which mitigation measures to implement on the existing pipeline networks with the minimum cost in an acceptable level of hazard.

Keywords: cost benefit analysis, knapsack problem, natural gas distribution network, risk management, risk mitigation

Procedia PDF Downloads 299
7434 Self-Action of Pyroelectric Spatial Soliton in Undoped Lithium Niobate Samples with Pyroelectric Mechanism of Nonlinear Response

Authors: Anton S. Perin, Vladimir M. Shandarov

Abstract:

Compensation for the nonlinear diffraction of narrow laser beams with wavelength of 532 and the formation of photonic waveguides and waveguide circuits due to the contribution of pyroelectric effect to the nonlinear response of lithium niobate crystal have been experimentally demonstrated. Complete compensation for the linear and nonlinear diffraction broadening of light beams is obtained upon uniform heating of an undoped sample from room temperature to 55 degrees Celsius. An analysis of the light-field distribution patterns and the corresponding intensity distribution profiles allowed us to estimate the spacing for the channel waveguides. The observed behavior of bright soliton beams may be caused by their coherent interaction, which manifests itself in repulsion for anti-phase light fields and in attraction for in-phase light fields. The experimental results of this study showed a fundamental possibility of forming optically complex waveguide structures in lithium niobate crystals with pyroelectric mechanism of nonlinear response. The topology of these structures is determined by the light field distribution on the input face of crystalline sample. The optical induction of channel waveguide elements by interacting spatial solitons makes it possible to design optical systems with a more complex topology and a possibility of their dynamic reconfiguration.

Keywords: self-action, soliton, lithium niobate, piroliton, photorefractive effect, pyroelectric effect

Procedia PDF Downloads 171
7433 A Multi-Output Network with U-Net Enhanced Class Activation Map and Robust Classification Performance for Medical Imaging Analysis

Authors: Jaiden Xuan Schraut, Leon Liu, Yiqiao Yin

Abstract:

Computer vision in medical diagnosis has achieved a high level of success in diagnosing diseases with high accuracy. However, conventional classifiers that produce an image to-label result provides insufficient information for medical professionals to judge and raise concerns over the trust and reliability of a model with results that cannot be explained. In order to gain local insight into cancerous regions, separate tasks such as imaging segmentation need to be implemented to aid the doctors in treating patients, which doubles the training time and costs which renders the diagnosis system inefficient and difficult to be accepted by the public. To tackle this issue and drive AI-first medical solutions further, this paper proposes a multi-output network that follows a U-Net architecture for image segmentation output and features an additional convolutional neural networks (CNN) module for auxiliary classification output. Class activation maps are a method of providing insight into a convolutional neural network’s feature maps that leads to its classification but in the case of lung diseases, the region of interest is enhanced by U-net-assisted Class Activation Map (CAM) visualization. Therefore, our proposed model combines image segmentation models and classifiers to crop out only the lung region of a chest X-ray’s class activation map to provide a visualization that improves the explainability and is able to generate classification results simultaneously which builds trust for AI-led diagnosis systems. The proposed U-Net model achieves 97.61% accuracy and a dice coefficient of 0.97 on testing data from the COVID-QU-Ex Dataset which includes both diseased and healthy lungs.

Keywords: multi-output network model, U-net, class activation map, image classification, medical imaging analysis

Procedia PDF Downloads 207
7432 Unsupervised Neural Architecture for Saliency Detection

Authors: Natalia Efremova, Sergey Tarasenko

Abstract:

We propose a novel neural network architecture for visual saliency detections, which utilizes neuro physiologically plausible mechanisms for extraction of salient regions. The model has been significantly inspired by recent findings from neuro physiology and aimed to simulate the bottom-up processes of human selective attention. Two types of features were analyzed: color and direction of maximum variance. The mechanism we employ for processing those features is PCA, implemented by means of normalized Hebbian learning and the waves of spikes. To evaluate performance of our model we have conducted psychological experiment. Comparison of simulation results with those of experiment indicates good performance of our model.

Keywords: neural network models, visual saliency detection, normalized Hebbian learning, Oja's rule, psychological experiment

Procedia PDF Downloads 354
7431 Optrix: Energy Aware Cross Layer Routing Using Convex Optimization in Wireless Sensor Networks

Authors: Ali Shareef, Aliha Shareef, Yifeng Zhu

Abstract:

Energy minimization is of great importance in wireless sensor networks in extending the battery lifetime. One of the key activities of nodes in a WSN is communication and the routing of their data to a centralized base-station or sink. Routing using the shortest path to the sink is not the best solution since it will cause nodes along this path to fail prematurely. We propose a cross-layer energy efficient routing protocol Optrix that utilizes a convex formulation to maximize the lifetime of the network as a whole. We further propose, Optrix-BW, a novel convex formulation with bandwidth constraint that allows the channel conditions to be accounted for in routing. By considering this key channel parameter we demonstrate that Optrix-BW is capable of congestion control. Optrix is implemented in TinyOS, and we demonstrate that a relatively large topology of 40 nodes can converge to within 91% of the optimal routing solution. We describe the pitfalls and issues related with utilizing a continuous form technique such as convex optimization with discrete packet based communication systems as found in WSNs. We propose a routing controller mechanism that allows for this transformation. We compare Optrix against the Collection Tree Protocol (CTP) and we found that Optrix performs better in terms of convergence to an optimal routing solution, for load balancing and network lifetime maximization than CTP.

Keywords: wireless sensor network, Energy Efficient Routing

Procedia PDF Downloads 395
7430 Network and Sentiment Analysis of U.S. Congressional Tweets

Authors: Chaitanya Kanakamedala, Hansa Pradhan, Carter Gilbert

Abstract:

Social media platforms, such as Twitter, are excellent datasets for understanding human interactions and sentiments. This report explores social dynamics among US Congressional members through a network analysis applied to a dataset of tweets spanning 2008 to 2017 from the ’US Congressional Tweets Dataset’. In this report, we preform network analysis where connections between users (edges) are established based on a similarity threshold: two tweets are connected if the tweets they post are similar. By utilizing the Natural Language Toolkit (NLTK) and NetworkX, we quantified tweet similarity and constructed a graph comprising various interconnected components. Each component represents a cluster of users with closely aligned content. We then preform sentiment analysis on each cluster to explore the prevalent emotions and opinions within these groups. Our findings reveal that despite the initial expectation of distinct ideological divisions typically aligning with party lines, the analysis exposed a high degree of topical convergence across tweets from different political affiliations. The analysis preformed in this report not only highlights the potential of social media as a tool for political communication but also suggests a complex layer of interaction that transcends traditional partisan boundaries, reflecting a complicated landscape of politics in the digital age.

Keywords: natural language processing, sentiment analysis, centrality analysis, topic modeling

Procedia PDF Downloads 41