Search results for: normalized architectures
420 Language Processing of Seniors with Alzheimer’s Disease: From the Perspective of Temporal Parameters
Authors: Lai Yi-Hsiu
Abstract:
The present paper aims to examine the language processing of Chinese-speaking seniors with Alzheimer’s disease (AD) from the perspective of temporal cues. Twenty healthy adults, 17 healthy seniors, and 13 seniors with AD in Taiwan participated in this study to tell stories based on two sets of pictures. Nine temporal cues were fetched and analyzed. Oral productions in Mandarin Chinese were compared and discussed to examine to what extent and in what way these three groups of participants performed with significant differences. Results indicated that the age effects were significant in filled pauses. The dementia effects were significant in mean duration of pauses, empty pauses, filled pauses, lexical pauses, normalized mean duration of filled pauses and lexical pauses. The findings reported in the current paper help characterize the nature of language processing in seniors with or without AD, and contribute to the interactions between the AD neural mechanism and their temporal parameters.Keywords: language processing, Alzheimer’s disease, Mandarin Chinese, temporal cues
Procedia PDF Downloads 446419 Self-Supervised Pretraining on Sequences of Functional Magnetic Resonance Imaging Data for Transfer Learning to Brain Decoding Tasks
Authors: Sean Paulsen, Michael Casey
Abstract:
In this work we present a self-supervised pretraining framework for transformers on functional Magnetic Resonance Imaging (fMRI) data. First, we pretrain our architecture on two self-supervised tasks simultaneously to teach the model a general understanding of the temporal and spatial dynamics of human auditory cortex during music listening. Our pretraining results are the first to suggest a synergistic effect of multitask training on fMRI data. Second, we finetune the pretrained models and train additional fresh models on a supervised fMRI classification task. We observe significantly improved accuracy on held-out runs with the finetuned models, which demonstrates the ability of our pretraining tasks to facilitate transfer learning. This work contributes to the growing body of literature on transformer architectures for pretraining and transfer learning with fMRI data, and serves as a proof of concept for our pretraining tasks and multitask pretraining on fMRI data.Keywords: transfer learning, fMRI, self-supervised, brain decoding, transformer, multitask training
Procedia PDF Downloads 90418 Finding Bicluster on Gene Expression Data of Lymphoma Based on Singular Value Decomposition and Hierarchical Clustering
Authors: Alhadi Bustaman, Soeganda Formalidin, Titin Siswantining
Abstract:
DNA microarray technology is used to analyze thousand gene expression data simultaneously and a very important task for drug development and test, function annotation, and cancer diagnosis. Various clustering methods have been used for analyzing gene expression data. However, when analyzing very large and heterogeneous collections of gene expression data, conventional clustering methods often cannot produce a satisfactory solution. Biclustering algorithm has been used as an alternative approach to identifying structures from gene expression data. In this paper, we introduce a transform technique based on singular value decomposition to identify normalized matrix of gene expression data followed by Mixed-Clustering algorithm and the Lift algorithm, inspired in the node-deletion and node-addition phases proposed by Cheng and Church based on Agglomerative Hierarchical Clustering (AHC). Experimental study on standard datasets demonstrated the effectiveness of the algorithm in gene expression data.Keywords: agglomerative hierarchical clustering (AHC), biclustering, gene expression data, lymphoma, singular value decomposition (SVD)
Procedia PDF Downloads 278417 Transformers in Gene Expression-Based Classification
Authors: Babak Forouraghi
Abstract:
A genetic circuit is a collection of interacting genes and proteins that enable individual cells to implement and perform vital biological functions such as cell division, growth, death, and signaling. In cell engineering, synthetic gene circuits are engineered networks of genes specifically designed to implement functionalities that are not evolved by nature. These engineered networks enable scientists to tackle complex problems such as engineering cells to produce therapeutics within the patient's body, altering T cells to target cancer-related antigens for treatment, improving antibody production using engineered cells, tissue engineering, and production of genetically modified plants and livestock. Construction of computational models to realize genetic circuits is an especially challenging task since it requires the discovery of flow of genetic information in complex biological systems. Building synthetic biological models is also a time-consuming process with relatively low prediction accuracy for highly complex genetic circuits. The primary goal of this study was to investigate the utility of a pre-trained bidirectional encoder transformer that can accurately predict gene expressions in genetic circuit designs. The main reason behind using transformers is their innate ability (attention mechanism) to take account of the semantic context present in long DNA chains that are heavily dependent on spatial representation of their constituent genes. Previous approaches to gene circuit design, such as CNN and RNN architectures, are unable to capture semantic dependencies in long contexts as required in most real-world applications of synthetic biology. For instance, RNN models (LSTM, GRU), although able to learn long-term dependencies, greatly suffer from vanishing gradient and low-efficiency problem when they sequentially process past states and compresses contextual information into a bottleneck with long input sequences. In other words, these architectures are not equipped with the necessary attention mechanisms to follow a long chain of genes with thousands of tokens. To address the above-mentioned limitations of previous approaches, a transformer model was built in this work as a variation to the existing DNA Bidirectional Encoder Representations from Transformers (DNABERT) model. It is shown that the proposed transformer is capable of capturing contextual information from long input sequences with attention mechanism. In a previous work on genetic circuit design, the traditional approaches to classification and regression, such as Random Forrest, Support Vector Machine, and Artificial Neural Networks, were able to achieve reasonably high R2 accuracy levels of 0.95 to 0.97. However, the transformer model utilized in this work with its attention-based mechanism, was able to achieve a perfect accuracy level of 100%. Further, it is demonstrated that the efficiency of the transformer-based gene expression classifier is not dependent on presence of large amounts of training examples, which may be difficult to compile in many real-world gene circuit designs.Keywords: transformers, generative ai, gene expression design, classification
Procedia PDF Downloads 59416 Flexural Response of Glass Fiber Reinforced Polymer Sandwich Panels with 3D Woven Honeycomb Core
Authors: Elif Kalkanli, Constantinos Soutis
Abstract:
The use of textile preform in the advanced fields including aerospace, automotive and marine has exponentially grown in recent years. These preforms offer excellent advantages such as being lightweight and low-cost, and also, their suitability for creating different fiber architectures with different materials whilst improved mechanical properties in certain aspects. In this study, a novel honeycomb core is developed by a 3Dweaving process. The assembly of the layers is achieved thanks to innovative weaving design. Polyester yarn is selected for the 3D woven honeycomb core (3DWHC). The core is used to manufacture a sandwich panel with 2x2 twill glass fiber composite face sheets. These 3DWHC sandwich panels will be tested in three-point bending. The in-plane and out-of-plane (through-the-thickness) mechanical response of the core will be examined as a function of cell size in addition to the flexural response of the sandwich panel. The failure mechanisms of the core and the sandwich skins will be reported in addition to flexural strength and stiffness. Possible engineering applications will be identified.Keywords: 3D woven, assembly, failure modes, honeycomb sandwich panel
Procedia PDF Downloads 206415 Rare Earth Element (REE) Geochemistry of Tepeköy Sandstones (Central Anatolia, Turkey)
Authors: Mehmet Yavuz Hüseyinca, Şuayip Küpeli
Abstract:
Sandstones from Upper Eocene - Oligocene Tepeköy formation (Member of Mezgit Group) that exposed on the eastern edge of Tuz Gölü (Salt Lake) were analyzed for their rare earth element (REE) contents. Average concentrations of ΣREE, ΣLREE (Total light rare earth elements) and ΣHREE (Total heavy rare earth elements) were determined as 31.37, 26.47 and 4.55 ppm respectively. These values are lower than UCC (Upper continental crust) which indicates grain size and/or CaO dilution effect. The chondrite-normalized REE pattern is characterized by the average ratios of (La/Yb)cn = 6.20, (La/Sm)cn = 4.06, (Gd/Lu)cn = 1.10, Eu/Eu* = 0.99 and Ce/Ce* = 0.94. Lower values of ΣLREE/ΣHREE (Average 5.97) and (La/Yb)cn suggest lower fractionation of overall REE. Moreover (La/Sm)cn and (Gd/Lu)cn ratios define less inclined LREE and almost flat HREE pattern when compared with UCC. Almost no Ce anomaly (Ce/Ce*) emphasizes that REE were originated from terrigenous material. Also depleted LREE and no Eu anomaly (Eu/Eu*) suggest an undifferentiated mafic provenance for the sandstones.Keywords: central Anatolia, provenance, rare earth elements, REE, Tepeköy sandstone
Procedia PDF Downloads 475414 Functions and Effects of Green Facades in the Developing Countries: Case Study of Tehran
Authors: S. Jahani, V. Choopankareh
Abstract:
Many people lost their life caused by environmental pollution every year. The negative effects of environmental crises appear to be much higher in Asian countries. The most important environmental issue in the developing countries and especially in Tehran, to our best knowledge, is air pollution that has affected many aspects of life in society. Environmental topics related to technology’s development have been salient issues among the main concerns of designers. Green facades are the most considerable solutions which designers and architectures are focused on, all over the world. But there are lots of behavioral and psychological problems about this point. In this line, this excavation has tried to reveal the cultural and psychological influences of green façade in developing countries like Tehran. Green façades in developing countries are so useless, although they are so expensive. As a matter of fact, users consider green facade as a decorative item. This research is an attempt to recognize the reasons which show green façades as worthless element. Also, some solutions are presented to promote green façades in the developing countries as an intrinsic solution. There are so many environmental threats, especially about air pollution, for a city as Tehran, which might be solved by green facades.Keywords: air pollution, developing countries, effects, green facades
Procedia PDF Downloads 276413 Challenging Convections: Rethinking Literature Review Beyond Citations
Authors: Hassan Younis
Abstract:
Purpose: The objective of this study is to review influential papers in the sustainability and supply chain studies domain, leveraging insights from this review to develop a structured framework for academics and researchers. This framework aims to assist scholars in identifying the most impactful publications for their scholarly pursuits. Subsequently, the study will apply and trial the developed framework on selected scholarly articles within the sustainability and supply chain studies domain to evaluate its efficacy, practicality, and reliability. Design/Methodology/Approach: Utilizing the "Publish or Perish" tool, a search was conducted to locate papers incorporating "sustainability" and "supply chain" in their titles. After rigorous filtering steps, a panel of university professors identified five crucial criteria for evaluating research robustness: average yearly citation counts (25%), scholarly contribution (25%), alignment of findings with objectives (15%), methodological rigor (20%), and journal impact factor (15%). These five evaluation criteria are abbreviated as “ACMAJ" framework. Each paper then received a tiered score (1-3) for each criterion, normalized within its category, and summed using weighted averages to calculate a Final Normalized Score (FNS). This systematic approach allows for objective comparison and ranking of the research based on its impact, novelty, rigor, and publication venue. Findings: The study's findings highlight the lack of structured frameworks for assessing influential sustainability research in supply chain management, which often results in a dependence on citation counts. A complete model that incorporates five essential criteria has been suggested as a response. By conducting a methodical trial on specific academic articles in the field of sustainability and supply chain studies, the model demonstrated its effectiveness as a tool for identifying and selecting influential research papers that warrant additional attention. This work aims to fill a significant deficiency in existing techniques by providing a more comprehensive approach to identifying and ranking influential papers in the field. Practical Implications: The developed framework helps scholars identify the most influential sustainability and supply chain publications. Its validation serves the academic community by offering a credible tool and helping researchers, students, and practitioners find and choose influential papers. This approach aids field literature reviews and study suggestions. Analysis of major trends and topics deepens our grasp of this critical study area's changing terrain. Originality/Value: The framework stands as a unique contribution to academia, offering scholars an important and new tool to identify and validate influential publications. Its distinctive capacity to efficiently guide scholars, learners, and professionals in selecting noteworthy publications, coupled with the examination of key patterns and themes, adds depth to our understanding of the evolving landscape in this critical field of study.Keywords: supply chain management, sustainability, framework, model
Procedia PDF Downloads 52412 Survey on Fiber Optic Deployment for Telecommunications Operators in Ghana: Coverage Gap, Recommendations and Research Directions
Authors: Francis Padi, Solomon Nunoo, John Kojo Annan
Abstract:
The paper "Survey on Fiber Optic Deployment for Telecommunications Operators in Ghana: Coverage Gap, Recommendations and Research Directions" presents a comprehensive survey on the deployment of fiber optic networks for telecommunications operators in Ghana. It addresses the challenges encountered by operators using microwave transmission systems for backhauling traffic and emphasizes the advantages of deploying fiber optic networks. The study delves into the coverage gap, provides recommendations, and outlines research directions to enhance the telecommunications infrastructure in Ghana. Additionally, it evaluates next-generation optical access technologies and architectures tailored to operators' needs. The paper also investigates current technological solutions and regulatory, technical, and economical dimensions related to sharing mobile telecommunication networks in emerging countries. Overall, this paper offers valuable insights into fiber optic network deployment for telecommunications operators in Ghana and suggests strategies to meet the increasing demand for data and mobile applications.Keywords: survey on fiber optic deployment, coverage gap, recommendations, research directions
Procedia PDF Downloads 21411 Development and Investigation of Sustainable Wireless Sensor Networks for forest Ecosystems
Authors: Shathya Duobiene, Gediminas Račiukaitis
Abstract:
Solar-powered wireless sensor nodes work best when they operate continuously with minimal energy consumption. Wireless Sensor Networks (WSNs) are a new technology opens up wide studies, and advancements are expanding the prevalence of numerous monitoring applications and real-time aid for environments. The Selective Surface Activation Induced by Laser (SSAIL) technology is an exciting development that gives the design of WSNs more flexibility in terms of their shape, dimensions, and materials. This research work proposes a methodology for using SSAIL technology for forest ecosystem monitoring by wireless sensor networks. WSN monitoring the temperature and humidity were deployed, and their architectures are discussed. The paper presents the experimental outcomes of deploying newly built sensor nodes in forested areas. Finally, a practical method is offered to extend the WSN's lifespan and ensure its continued operation. When operational, the node is independent of the base station's power supply and uses only as much energy as necessary to sense and transmit data.Keywords: internet of things (IoT), wireless sensor network, sensor nodes, SSAIL technology, forest ecosystem
Procedia PDF Downloads 72410 Myoelectric Analysis for the Assessment of Muscle Functions and Fatigue Monitoring of Upper Extremity for Stroke Patients Performing Robot-Assisted Bilateral Training
Authors: Hsiao-Lung Chan, Ching-Yi Wu, Yan-Zou Lin, Yo Chiao, Ya-Ju Chang
Abstract:
Robot-assisted bilateral arm training has demonstrated useful to improve motor control in stroke patients and save human resources. In clinics, the efficiency of this treatment is mostly performed by comparing functional scales before and after rehabilitation. However, most of these assessments are based on behavior evaluation. The underlying improvement of muscle activation and coordination is unknown. Moreover, stroke patients are easier to have muscle fatigue under robot-assisted rehabilitation due to the weakness of muscles. This safety issue is still less studied. In this study, EMG analysis was applied during training. Our preliminary results showed the co-contraction index and co-contraction area index can delineate the improved muscle coordination of biceps brachii vs. flexor carpiradialis. Moreover, the smoothed, normalized cycle-by-cycle median frequency of left and right extensor carpiradialis decreased as the training progress, implying the occurrence of muscle fatigue.Keywords: robot-assisted rehabilitation, strokes, muscle coordination, muscle fatigue
Procedia PDF Downloads 475409 Solution of the Nonrelativistic Radial Wave Equation of Hydrogen Atom Using the Green's Function Approach
Authors: F. U. Rahman, R. Q. Zhang
Abstract:
This work aims to develop a systematic numerical technique which can be easily extended to many-body problem. The Lippmann Schwinger equation (integral form of the Schrodinger wave equation) is solved for the nonrelativistic radial wave of hydrogen atom using iterative integration scheme. As the unknown wave function appears on both sides of the Lippmann Schwinger equation, therefore an approximate wave function is used in order to solve the equation. The Green’s function is obtained by the method of Laplace transform for the radial wave equation with excluded potential term. Using the Lippmann Schwinger equation, the product of approximate wave function, the Green’s function and the potential term is integrated iteratively. Finally, the wave function is normalized and plotted against the standard radial wave for comparison. The outcome wave function converges to the standard wave function with the increasing number of iteration. Results are verified for the first fifteen states of hydrogen atom. The method is efficient and consistent and can be applied to complex systems in future.Keywords: Green’s function, hydrogen atom, Lippmann Schwinger equation, radial wave
Procedia PDF Downloads 394408 Wave Propagation In Functionally Graded Lattice Structures Under Impact Loads
Authors: Mahmood Heshmati, Farhang Daneshmand
Abstract:
Material scientists and engineers have introduced novel materials with complex geometries due to the recent technological advances and promotion of manufacturing methods. Among them, lattice structures with graded architectures denoted by functionally graded porous materials (FGPMs) have been developed to optimize the structural response. FGPMs are achieved by tailoring the size and density of the internal pores in one or more directions that lead to the desired mechanical properties and structural responses. Also, FGPMs provide more flexible transition and the possibility of designing and fabricating structural elements with complex and variable properties. In this paper, wave propagation in lattice structures with functionally graded (FG) porosity is investigated in order to examine the ability of shock absorbing effect. The behavior of FG porous beams with different porosity distributions under impact load and the effects of porosity distribution and porosity content on the wave speed are studied. Important conclusions are made, along with a discussion of the future scope of studies on FGPMs structures.Keywords: functionally graded, porous materials, wave propagation, impact load, finite element
Procedia PDF Downloads 91407 On the Utility of Bidirectional Transformers in Gene Expression-Based Classification
Authors: Babak Forouraghi
Abstract:
A genetic circuit is a collection of interacting genes and proteins that enable individual cells to implement and perform vital biological functions such as cell division, growth, death, and signaling. In cell engineering, synthetic gene circuits are engineered networks of genes specifically designed to implement functionalities that are not evolved by nature. These engineered networks enable scientists to tackle complex problems such as engineering cells to produce therapeutics within the patient's body, altering T cells to target cancer-related antigens for treatment, improving antibody production using engineered cells, tissue engineering, and production of genetically modified plants and livestock. Construction of computational models to realize genetic circuits is an especially challenging task since it requires the discovery of the flow of genetic information in complex biological systems. Building synthetic biological models is also a time-consuming process with relatively low prediction accuracy for highly complex genetic circuits. The primary goal of this study was to investigate the utility of a pre-trained bidirectional encoder transformer that can accurately predict gene expressions in genetic circuit designs. The main reason behind using transformers is their innate ability (attention mechanism) to take account of the semantic context present in long DNA chains that are heavily dependent on the spatial representation of their constituent genes. Previous approaches to gene circuit design, such as CNN and RNN architectures, are unable to capture semantic dependencies in long contexts, as required in most real-world applications of synthetic biology. For instance, RNN models (LSTM, GRU), although able to learn long-term dependencies, greatly suffer from vanishing gradient and low-efficiency problem when they sequentially process past states and compresses contextual information into a bottleneck with long input sequences. In other words, these architectures are not equipped with the necessary attention mechanisms to follow a long chain of genes with thousands of tokens. To address the above-mentioned limitations, a transformer model was built in this work as a variation to the existing DNA Bidirectional Encoder Representations from Transformers (DNABERT) model. It is shown that the proposed transformer is capable of capturing contextual information from long input sequences with an attention mechanism. In previous works on genetic circuit design, the traditional approaches to classification and regression, such as Random Forrest, Support Vector Machine, and Artificial Neural Networks, were able to achieve reasonably high R2 accuracy levels of 0.95 to 0.97. However, the transformer model utilized in this work, with its attention-based mechanism, was able to achieve a perfect accuracy level of 100%. Further, it is demonstrated that the efficiency of the transformer-based gene expression classifier is not dependent on the presence of large amounts of training examples, which may be difficult to compile in many real-world gene circuit designs.Keywords: machine learning, classification and regression, gene circuit design, bidirectional transformers
Procedia PDF Downloads 61406 Energy-Efficient Contact Selection Method for CARD in Wireless Ad-Hoc Networks
Authors: Mehdi Assefi, Keihan Hataminezhad
Abstract:
One of the efficient architectures for exploring the resources in wireless ad-hoc networks is contact-based architecture. In this architecture, each node assigns a unique zone for itself and each node keeps all information from inside the zone, as well as some from outside the zone, which is called contact. Reducing the overlap between different zones of a node and its contacts increases its performance, therefore Edge Method (EM) is designed for this purpose. Contacts selected by EM do not have any overlap with their sources, but for choosing the contact a vast amount of information must be transmitted. In this article, we will offer a new protocol for contact selection, which is called PEM. The objective would be reducing the volume of transmitted information, using Non-Uniform Dissemination Probabilistic Protocols. Consumed energy for contact selection is a function of the size of transmitted information between nodes. Therefore, by reducing the content of contact selection message using the PEM will decrease the consumed energy. For evaluation of the PEM we applied the simulation method. Results indicated that PEM consumes less energy compared to EM, and by increasing the number of nodes (level of nodes), performance of PEM will improve in comparison with EM.Keywords: wireless ad-hoc networks, contact selection, method for CARD, energy-efficient
Procedia PDF Downloads 290405 Post-Exercise Recovery Tracking Based on Electrocardiography-Derived Features
Authors: Pavel Bulai, Taras Pitlik, Tatsiana Kulahava, Timofei Lipski
Abstract:
The method of Electrocardiography (ECG) interpretation for post-exercise recovery tracking was developed. Metabolic indices (aerobic and anaerobic) were designed using ECG-derived features. This study reports the associations between aerobic and anaerobic indices and classical parameters of the person’s physiological state, including blood biochemistry, glycogen concentration and VO2max changes. During the study 9 participants, healthy, physically active medium trained men and women, which trained 2-4 times per week for at least 9 weeks, fulfilled (i) ECG monitoring using Apple Watch Series 4 (AWS4); (ii) blood biochemical analysis; (iii) maximal oxygen consumption (VO2max) test, (iv) bioimpedance analysis (BIA). ECG signals from a single-lead wrist-wearable device were processed with detection of QRS-complex. Aerobic index (AI) was derived as the normalized slope of QR segment. Anaerobic index (ANI) was derived as the normalized slope of SJ segment. Biochemical parameters, glycogen content and VO2max were evaluated eight times within 3-60 hours after training. ECGs were recorded 5 times per day, plus before and after training, cycloergometry and BIA. The negative correlation between AI and blood markers of the muscles functional status including creatine phosphokinase (r=-0.238, p < 0.008), aspartate aminotransferase (r=-0.249, p < 0.004) and uric acid (r = -0.293, p<0.004) were observed. ANI was also correlated with creatine phosphokinase (r= -0.265, p < 0.003), aspartate aminotransferase (r = -0.292, p < 0.001), lactate dehydrogenase (LDH) (r = -0.190, p < 0.050). So, when the level of muscular enzymes increases during post-exercise fatigue, AI and ANI decrease. During recovery, the level of metabolites is restored, and metabolic indices rising is registered. It can be concluded that AI and ANI adequately reflect the physiology of the muscles during recovery. One of the markers of an athlete’s physiological state is the ratio between testosterone and cortisol (TCR). TCR provides a relative indication of anabolic-catabolic balance and is considered to be more sensitive to training stress than measuring testosterone and cortisol separately. AI shows a strong negative correlation with TCR (r=-0.437, p < 0.001) and correctly represents post-exercise physiology. In order to reveal the relation between the ECG-derived metabolic indices and the state of the cardiorespiratory system, direct measurements of VO2max were carried out at various time points after training sessions. The negative correlation between AI and VO2max (r = -0.342, p < 0.001) was obtained. These data testifying VO2max rising during fatigue are controversial. However, some studies have revealed increased stroke volume after training, that agrees with findings. It is important to note that post-exercise increase in VO2max does not mean an athlete’s readiness for the next training session, because the recovery of the cardiovascular system occurs over a substantially longer period. Negative correlations registered for ANI with glycogen (r = -0.303, p < 0.001), albumin (r = -0.205, p < 0.021) and creatinine (r = -0.268, p < 0.002) reflect the dehydration status of participants after training. Correlations between designed metabolic indices and physiological parameters revealed in this study can be considered as the sufficient evidence to use these indices for assessing the state of person’s aerobic and anaerobic metabolic systems after training during fatigue, recovery and supercompensation.Keywords: aerobic index, anaerobic index, electrocardiography, supercompensation
Procedia PDF Downloads 115404 A DOE Study of Ultrasound Intensified Removal of Phenol
Authors: P. R. Rahul, A. Kannan
Abstract:
Ultrasound-aided adsorption of phenol by Granular Activated Carbon (GAC) was investigated at different frequencies ranging from 35 kHz, 58 kHz, and 192 kHz. Other factors influencing adsorption such as Adsorbent dosage (g/L), the initial concentration of the phenol solution (ppm) and RPM was also considered along with the frequency variable. However, this study involved calorimetric measurements which helped is determining the effect of frequency on the % removal of phenol from the power dissipated to the system was normalized. It was found that low frequency (35 kHz) cavitation effects had a profound influence on the % removal of phenol per unit power. This study also had cavitation mapping of the ultrasonic baths, and it showed that the effect of cavitation on the adsorption system is irrespective of the position of the vessel. Hence, the vessel was placed at the center of the bath. In this study, novel temperature control and monitoring system to make sure that the system is under proper condition while operations. From the BET studies, it was found that there was only 5% increase in the surface area and hence it was concluded that ultrasound doesn’t profoundly alter the equilibrium value of the adsorption system. DOE studies indicated that adsorbent dosage has a higher influence on the % removal in comparison with other factors.Keywords: ultrasound, adsorption, granulated activated carbon, phenol
Procedia PDF Downloads 283403 Exact Energy Spectrum and Expectation Values of the Inverse Square Root Potential Model
Authors: Benedict Ita, Peter Okoi
Abstract:
In this work, the concept of the extended Nikiforov-Uvarov technique is discussed and employed to obtain the exact bound state energy eigenvalues and the corresponding normalized eigenfunctions of the inverse square root potential. With expressions for the exact energy eigenvalues and corresponding eigenfunctions, the expressions for the expectation values of the inverse separation-squared, kinetic energy, and the momentum-squared of the potential are presented using the Hellmann Feynman theorem. For visualization, algorithms written and implemented in Python language are used to generate tables and plots for l-states of the energy eigenvalues and some expectation values. The results obtained here may find suitable applications in areas like atomic and molecular physics, chemical physics, nuclear physics, and solid-state physics.Keywords: Schrodinger equation, Nikoforov-Uvarov method, inverse square root potential, diatomic molecules, Python programming, Hellmann-Feynman theorem, second order differential equation, matrix algebra
Procedia PDF Downloads 19402 Efficient Deep Neural Networks for Real-Time Strawberry Freshness Monitoring: A Transfer Learning Approach
Authors: Mst. Tuhin Akter, Sharun Akter Khushbu, S. M. Shaqib
Abstract:
A real-time system architecture is highly effective for monitoring and detecting various damaged products or fruits that may deteriorate over time or become infected with diseases. Deep learning models have proven to be effective in building such architectures. However, building a deep learning model from scratch is a time-consuming and costly process. A more efficient solution is to utilize deep neural network (DNN) based transfer learning models in the real-time monitoring architecture. This study focuses on using a novel strawberry dataset to develop effective transfer learning models for the proposed real-time monitoring system architecture, specifically for evaluating and detecting strawberry freshness. Several state-of-the-art transfer learning models were employed, and the best performing model was found to be Xception, demonstrating higher performance across evaluation metrics such as accuracy, recall, precision, and F1-score.Keywords: strawberry freshness evaluation, deep neural network, transfer learning, image augmentation
Procedia PDF Downloads 90401 Optimized Preprocessing for Accurate and Efficient Bioassay Prediction with Machine Learning Algorithms
Authors: Jeff Clarine, Chang-Shyh Peng, Daisy Sang
Abstract:
Bioassay is the measurement of the potency of a chemical substance by its effect on a living animal or plant tissue. Bioassay data and chemical structures from pharmacokinetic and drug metabolism screening are mined from and housed in multiple databases. Bioassay prediction is calculated accordingly to determine further advancement. This paper proposes a four-step preprocessing of datasets for improving the bioassay predictions. The first step is instance selection in which dataset is categorized into training, testing, and validation sets. The second step is discretization that partitions the data in consideration of accuracy vs. precision. The third step is normalization where data are normalized between 0 and 1 for subsequent machine learning processing. The fourth step is feature selection where key chemical properties and attributes are generated. The streamlined results are then analyzed for the prediction of effectiveness by various machine learning algorithms including Pipeline Pilot, R, Weka, and Excel. Experiments and evaluations reveal the effectiveness of various combination of preprocessing steps and machine learning algorithms in more consistent and accurate prediction.Keywords: bioassay, machine learning, preprocessing, virtual screen
Procedia PDF Downloads 274400 Bone Fracture Detection with X-Ray Images Using Mobilenet V3 Architecture
Authors: Ashlesha Khanapure, Harsh Kashyap, Abhinav Anand, Sanjana Habib, Anupama Bidargaddi
Abstract:
Technologies that are developing quickly are being developed daily in a variety of disciplines, particularly the medical field. For the purpose of detecting bone fractures in X-ray pictures of different body segments, our work compares the ResNet-50 and MobileNetV3 architectures. It evaluates accuracy and computing efficiency with X-rays of the elbow, hand, and shoulder from the MURA dataset. Through training and validation, the models are evaluated on normal and fractured images. While ResNet-50 showcases superior accuracy in fracture identification, MobileNetV3 showcases superior speed and resource optimization. Despite ResNet-50’s accuracy, MobileNetV3’s swifter inference makes it a viable choice for real-time clinical applications, emphasizing the importance of balancing computational efficiency and accuracy in medical imaging. We created a graphical user interface (GUI) for MobileNet V3 model bone fracture detection. This research underscores MobileNetV3’s potential to streamline bone fracture diagnoses, potentially revolutionizing orthopedic medical procedures and enhancing patient care.Keywords: CNN, MobileNet V3, ResNet-50, healthcare, MURA, X-ray, fracture detection
Procedia PDF Downloads 63399 Mobile WiMAX Network based Wireless Communication on Rail: An Analysis
Authors: Vinod Kumar Jatav, Dr. Vrijendra Singh
Abstract:
WiMAX is an emerging wireless technology designed by WiMAX forum. WiMAX technology delivers broadband internet access with QoS, mobility and robust security. WiMAX is among the prominent mobile broadband wireless technology which laid the foundation for the next generation networks (NGN). The next-generation communication system for railway should facilitate high level network availability, fast mobility for high speed trains with reliability, high handover rate, the firmness of train operations, and high QoS. The system should also be capable to provide various railway services by transmitting big data efficiently. One of the most promising technologies for the next generation railway wireless communication is Mobile WiMAX. This paper analyses some of the network architectures for railway wireless communication and considers the elementary concepts to facilitate the users with broadband internet access on trains. The paper aims to recognize the suitability of Mobile WiMAX technology for the special requirements of broadband internet facilities and wireless telecommunication services of Railways.Keywords: Broadband internet, IEEE 802.16e, mobile WiMAX, Railway wireless communication
Procedia PDF Downloads 524398 SEM Image Classification Using CNN Architectures
Authors: Güzi̇n Ti̇rkeş, Özge Teki̇n, Kerem Kurtuluş, Y. Yekta Yurtseven, Murat Baran
Abstract:
A scanning electron microscope (SEM) is a type of electron microscope mainly used in nanoscience and nanotechnology areas. Automatic image recognition and classification are among the general areas of application concerning SEM. In line with these usages, the present paper proposes a deep learning algorithm that classifies SEM images into nine categories by means of an online application to simplify the process. The NFFA-EUROPE - 100% SEM data set, containing approximately 21,000 images, was used to train and test the algorithm at 80% and 20%, respectively. Validation was carried out using a separate data set obtained from the Middle East Technical University (METU) in Turkey. To increase the accuracy in the results, the Inception ResNet-V2 model was used in view of the Fine-Tuning approach. By using a confusion matrix, it was observed that the coated-surface category has a negative effect on the accuracy of the results since it contains other categories in the data set, thereby confusing the model when detecting category-specific patterns. For this reason, the coated-surface category was removed from the train data set, hence increasing accuracy by up to 96.5%.Keywords: convolutional neural networks, deep learning, image classification, scanning electron microscope
Procedia PDF Downloads 125397 A Multi-agent System Framework for Stakeholder Analysis of Local Energy Systems
Authors: Mengqiu Deng, Xiao Peng, Yang Zhao
Abstract:
The development of local energy systems requires the collective involvement of different actors from various levels of society. However, the stakeholder analysis of local energy systems still has been under-developed. This paper proposes an multi-agent system (MAS) framework to facilitate the development of stakeholder analysis of local energy systems. The framework takes into account the most influencing stakeholders, including prosumers/consumers, system operators, energy companies and government bodies. Different stakeholders are modeled based on agent architectures for example the belief-desire-intention (BDI) to better reflect their motivations and interests in participating in local energy systems. The agent models of different stakeholders are then integrated in one model of the whole energy system. An illustrative case study is provided to elaborate how to develop a quantitative agent model for different stakeholders, as well as to demonstrate the practicability of the proposed framework. The findings from the case study indicate that the suggested framework and agent model can serve as analytical instruments for enhancing the government’s policy-making process by offering a systematic view of stakeholder interconnections in local energy systems.Keywords: multi-agent system, BDI agent, local energy systems, stakeholders
Procedia PDF Downloads 87396 Single Cell Analysis of Circulating Monocytes in Prostate Cancer Patients
Authors: Leander Van Neste, Kirk Wojno
Abstract:
The innate immune system reacts to foreign insult in several unique ways, one of which is phagocytosis of perceived threats such as cancer, bacteria, and viruses. The goal of this study was to look for evidence of phagocytosed RNA from tumor cells in circulating monocytes. While all monocytes possess phagocytic capabilities, the non-classical CD14+/FCGR3A+ monocytes and the intermediate CD14++/FCGR3A+ monocytes most actively remove threatening ‘external’ cellular materials. Purified CD14-positive monocyte samples from fourteen patients recently diagnosed with clinically localized prostate cancer (PCa) were investigated by single-cell RNA sequencing using the 10X Genomics protocol followed by paired-end sequencing on Illumina’s NovaSeq. Similarly, samples were processed and used as controls, i.e., one patient underwent biopsy but was found not to harbor prostate cancer (benign), three young, healthy men, and three men previously diagnosed with prostate cancer that recently underwent (curative) radical prostatectomy (post-RP). Sequencing data were mapped using 10X Genomics’ CellRanger software and viable cells were subsequently identified using CellBender, removing technical artifacts such as doublets and non-cellular RNA. Next, data analysis was performed in R, using the Seurat package. Because the main goal was to identify differences between PCa patients and ‘control’ patients, rather than exploring differences between individual subjects, the individual Seurat objects of all 21 patients were merged into one Seurat object per Seurat’s recommendation. Finally, the single-cell dataset was normalized as a whole prior to further analysis. Cell identity was assessed using the SingleR and cell dex packages. The Monaco Immune Data was selected as the reference dataset, consisting of bulk RNA-seq data of sorted human immune cells. The Monaco classification was supplemented with normalized PCa data obtained from The Cancer Genome Atlas (TCGA), which consists of bulk RNA sequencing data from 499 prostate tumor tissues (including 1 metastatic) and 52 (adjacent) normal prostate tissues. SingleR was subsequently run on the combined immune cell and PCa datasets. As expected, the vast majority of cells were labeled as having a monocytic origin (~90%), with the most noticeable difference being the larger number of intermediate monocytes in the PCa patients (13.6% versus 7.1%; p<.001). In men harboring PCa, 0.60% of all purified monocytes were classified as harboring PCa signals when the TCGA data were included. This was 3-fold, 7.5-fold, and 4-fold higher compared to post-RP, benign, and young men, respectively (all p<.001). In addition, with 7.91%, the number of unclassified cells, i.e., cells with pruned labels due to high uncertainty of the assigned label, was also highest in men with PCa, compared to 3.51%, 2.67%, and 5.51% of cells in post-RP, benign, and young men, respectively (all p<.001). It can be postulated that actively phagocytosing cells are hardest to classify due to their dual immune cell and foreign cell nature. Hence, the higher number of unclassified cells and intermediate monocytes in PCa patients might reflect higher phagocytic activity due to tumor burden. This also illustrates that small numbers (~1%) of circulating peripheral blood monocytes that have interacted with tumor cells might still possess detectable phagocytosed tumor RNA.Keywords: circulating monocytes, phagocytic cells, prostate cancer, tumor immune response
Procedia PDF Downloads 162395 Backward-Facing Step Measurements at Different Reynolds Numbers Using Acoustic Doppler Velocimetry
Authors: Maria Amelia V. C. Araujo, Billy J. Araujo, Brian Greenwood
Abstract:
The flow over a backward-facing step is characterized by the presence of flow separation, recirculation and reattachment, for a simple geometry. This type of fluid behaviour takes place in many practical engineering applications, hence the reason for being investigated. Historically, fluid flows over a backward-facing step have been examined in many experiments using a variety of measuring techniques such as laser Doppler velocimetry (LDV), hot-wire anemometry, particle image velocimetry or hot-film sensors. However, some of these techniques cannot conveniently be used in separated flows or are too complicated and expensive. In this work, the applicability of the acoustic Doppler velocimetry (ADV) technique is investigated to such type of flows, at various Reynolds numbers corresponding to different flow regimes. The use of this measuring technique in separated flows is very difficult to find in literature. Besides, most of the situations where the Reynolds number effect is evaluated in separated flows are in numerical modelling. The ADV technique has the advantage in providing nearly non-invasive measurements, which is important in resolving turbulence. The ADV Nortek Vectrino+ was used to characterize the flow, in a recirculating laboratory flume, at various Reynolds Numbers (Reh = 3738, 5452, 7908 and 17388) based on the step height (h), in order to capture different flow regimes, and the results compared to those obtained using other measuring techniques. To compare results with other researchers, the step height, expansion ratio and the positions upstream and downstream the step were reproduced. The post-processing of the AVD records was performed using a customized numerical code, which implements several filtering techniques. Subsequently, the Vectrino noise level was evaluated by computing the power spectral density for the stream-wise horizontal velocity component. The normalized mean stream-wise velocity profiles, skin-friction coefficients and reattachment lengths were obtained for each Reh. Turbulent kinetic energy, Reynolds shear stresses and normal Reynolds stresses were determined for Reh = 7908. An uncertainty analysis was carried out, for the measured variables, using the moving block bootstrap technique. Low noise levels were obtained after implementing the post-processing techniques, showing their effectiveness. Besides, the errors obtained in the uncertainty analysis were relatively low, in general. For Reh = 7908, the normalized mean stream-wise velocity and turbulence profiles were compared directly with those acquired by other researchers using the LDV technique and a good agreement was found. The ADV technique proved to be able to characterize the flow properly over a backward-facing step, although additional caution should be taken for measurements very close to the bottom. The ADV measurements showed reliable results regarding: a) the stream-wise velocity profiles; b) the turbulent shear stress; c) the reattachment length; d) the identification of the transition from transitional to turbulent flows. Despite being a relatively inexpensive technique, acoustic Doppler velocimetry can be used with confidence in separated flows and thus very useful for numerical model validation. However, it is very important to perform adequate post-processing of the acquired data, to obtain low noise levels, thus decreasing the uncertainty.Keywords: ADV, experimental data, multiple Reynolds number, post-processing
Procedia PDF Downloads 148394 Determination of Verapamil Hydrochloride in the Tablet and Injection Solution by the Verapamil-Sensitive Electrode and Possibilities of Application in Pharmaceutical Analysis
Authors: Faisal A. Salih, V. V. Egorov
Abstract:
Verapamil is a drug used in medicine for arrhythmia, angina, and hypertension as a calcium channel blocker. In this study, a Verapamil-selective electrode was prepared, and the concentrations of the components in the membrane were as follows: PVC (32.8 wt %), O-NPhOE (66.6 wt %), and KTPClPB (0.6 wt % or approximately 0.01 M). The inner solution containing verapamil hydrochloride 1 x 10⁻³ M was introduced, and the electrodes were conditioned overnight in 1 x 10⁻³ M verapamil hydrochloride solution in 1 x 10⁻³ M orthophosphoric acid. These studies have demonstrated that O-NPhOE and KTPClPB are the best plasticizers and ion exchangers, while both direct potentiometry and potentiometric titration methods can be used for the determination of verapamil hydrochloride in tablets and injection solutions. Normalized weights of verapamil per tablet (80.4±0.2, 80.7±0.2, 81.0±0.4 mg) were determined by direct potentiometry and potentiometric titration, respectively. Weights of verapamil per average tablet weight determined by the methods of direct potentiometry and potentiometric titration were" 80.4±0.2, 80.7±0.2 mg determined for the same set of tablets, respectively. The masses of verapamil in solutions for injection, determined by direct potentiometry for two ampoules from one set, were (5.00±0.015, 5.004±0.006) mg. In all cases, good reproducibility and excellent correspondence with the declared quantities were observed.Keywords: verapamil, potentiometry, ion-selective electrode, lipophilic physiologically active amines
Procedia PDF Downloads 86393 Hate Speech Detection in Tunisian Dialect
Authors: Helmi Baazaoui, Mounir Zrigui
Abstract:
This study addresses the challenge of hate speech detection in Tunisian Arabic text, a critical issue for online safety and moderation. Leveraging the strengths of the AraBERT model, we fine-tuned and evaluated its performance against the Bi-LSTM model across four distinct datasets: T-HSAB, TNHS, TUNIZI-Dataset, and a newly compiled dataset with diverse labels such as Offensive Language, Racism, and Religious Intolerance. Our experimental results demonstrate that AraBERT significantly outperforms Bi-LSTM in terms of Recall, Precision, F1-Score, and Accuracy across all datasets. The findings underline the robustness of AraBERT in capturing the nuanced features of Tunisian Arabic and its superior capability in classification tasks. This research not only advances the technology for hate speech detection but also provides practical implications for social media moderation and policy-making in Tunisia. Future work will focus on expanding the datasets and exploring more sophisticated architectures to further enhance detection accuracy, thus promoting safer online interactions.Keywords: hate speech detection, Tunisian Arabic, AraBERT, Bi-LSTM, Gemini annotation tool, social media moderation
Procedia PDF Downloads 11392 The Predictive Value of Serum Bilirubin in the Post-Transplant De Novo Malignancy: A Data Mining Approach
Authors: Nasim Nosoudi, Amir Zadeh, Hunter White, Joshua Conrad, Joon W. Shim
Abstract:
De novo Malignancy has become one of the major causes of death after transplantation, so early cancer diagnosis and detection can drastically improve survival rates post-transplantation. Most previous work focuses on using artificial intelligence (AI) to predict transplant success or failure outcomes. In this work, we focused on predicting de novo malignancy after liver transplantation using AI. We chose the patients that had malignancy after liver transplantation with no history of malignancy pre-transplant. Their donors were cancer-free as well. We analyzed 254,200 patient profiles with post-transplant malignancy from the US Organ Procurement and Transplantation Network (OPTN). Several popular data mining methods were applied to the resultant dataset to build predictive models to characterize de novo malignancy after liver transplantation. Recipient's bilirubin, creatinine, weight, gender, number of days recipient was on the transplant waiting list, Epstein Barr Virus (EBV), International normalized ratio (INR), and ascites are among the most important factors affecting de novo malignancy after liver transplantationKeywords: De novo malignancy, bilirubin, data mining, transplantation
Procedia PDF Downloads 105391 The Conflict between Empowerment and Exploitation: The Hypersexualization of Women in the Media
Authors: Seung Won Park
Abstract:
Pornographic images are becoming increasingly normalized as innovations in media technology arise, the porn industry explosively grows, and transnational capitalism spreads due to government deregulation and privatization of media. As the media evolves, pornography has become more and more violent and non-consensual; this growth of ‘raunch culture’ reifies the traditional power balance between men and women in which men are dominant, and women are submissive. This male domination objectifies and commodifies women, reducing them to merely sexual objects for the gratification of men. Women are exposed to pornographic images at younger and younger ages, providing unhealthy sexual role models and teaching them lessons on sexual behavior before the onset of puberty. The increasingly sexualized depiction of women in particular positions them as appropriately desirable and available to men. As a result, women are not only viewed as sexual prey but also end up treating themselves primarily as sexual objects, basing their worth off of their sexuality alone. Although many scholars are aware of and have written on the great lack of agency exercised by women in these representations, the general public tends to view some of these women as being empowered, rather than exploited. Scholarly discourse is constrained by the popular misconception that the construction of women’s sexuality in the media is controlled by women themselves.Keywords: construction of gender, hypersexualization, media, objectification
Procedia PDF Downloads 296