Search results for: deep brain stimulation (DBS)
3068 Neural Synchronization - The Brain’s Transfer of Sensory Data
Authors: David Edgar
Abstract:
To understand how the brain’s subconscious and conscious functions, we must conquer the physics of Unity, which leads to duality’s algorithm. Where the subconscious (bottom-up) and conscious (top-down) processes function together to produce and consume intelligence, we use terms like ‘time is relative,’ but we really do understand the meaning. In the brain, there are different processes and, therefore, different observers. These different processes experience time at different rates. A sensory system such as the eyes cycles measurement around 33 milliseconds, the conscious process of the frontal lobe cycles at 300 milliseconds, and the subconscious process of the thalamus cycle at 5 milliseconds. Three different observers experience time differently. To bridge observers, the thalamus, which is the fastest of the processes, maintains a synchronous state and entangles the different components of the brain’s physical process. The entanglements form a synchronous cohesion between the brain components allowing them to share the same state and execute in the same measurement cycle. The thalamus uses the shared state to control the firing sequence of the brain’s linear subconscious process. Sharing state also allows the brain to cheat on the amount of sensory data that must be exchanged between components. Only unpredictable motion is transferred through the synchronous state because predictable motion already exists in the shared framework. The brain’s synchronous subconscious process is entirely based on energy conservation, where prediction regulates energy usage. So, the eyes every 33 milliseconds dump their sensory data into the thalamus every day. The thalamus is going to perform a motion measurement to identify the unpredictable motion in the sensory data. Here is the trick. The thalamus conducts its measurement based on the original observation time of the sensory system (33 ms), not its own process time (5 ms). This creates a data payload of synchronous motion that preserves the original sensory observation. Basically, a frozen moment in time (Flat 4D). The single moment in time can then be processed through the single state maintained by the synchronous process. Other processes, such as consciousness (300 ms), can interface with the synchronous state to generate awareness of that moment. Now, synchronous data traveling through a separate faster synchronous process creates a theoretical time tunnel where observation time is tunneled through the synchronous process and is reproduced on the other side in the original time-relativity. The synchronous process eliminates time dilation by simply removing itself from the equation so that its own process time does not alter the experience. To the original observer, the measurement appears to be instantaneous, but in the thalamus, a linear subconscious process generating sensory perception and thought production is being executed. It is all just occurring in the time available because other observation times are slower than thalamic measurement time. For life to exist in the physical universe requires a linear measurement process, it just hides by operating at a faster time relativity. What’s interesting is time dilation is not the problem; it’s the solution. Einstein said there was no universal time.Keywords: neural synchronization, natural intelligence, 99.95% IoT data transmission savings, artificial subconscious intelligence (ASI)
Procedia PDF Downloads 1273067 Neuropsychology of Social Awareness: A Research Study Applied to University Students in Greece
Authors: Argyris Karapetsas, Maria Bampou, Andriani Mitropoulou
Abstract:
The aim of the present work is to study the role of brain function in social awareness processing. Mind controls all the psychosomatic functions. Mind’s functioning enables individual not only to recognize one's own self and propositional attitudes, but also to assign such attitudes to other individuals, and to consider such observed mental states in the elucidation of behavior. Participants and Methods: Twenty (n=20) undergraduate students (mean age 18 years old) were involved in this study. Students participated in a clinical assessment, being conducted in Laboratory of Neuropsychology, at University of Thessaly, in Volos, Greece. Assessment included both electrophysiological (i.e.Event Related Potentials (ERPs) esp.P300 waveform) and neuropsychological tests (Raven's Progressive Matrices (RPM) and Sally-Anne test). Results: Initial assessment’s results confirmed statistically significant differences between the males and females, as well as in score performance to the tests applied. Strong correlations emerged between prefrontal lobe functioning, RPM, Sally-Anne test and P300 latencies. Also, significant dysfunction of mind has been found, regarding its three dimensions (straight, circular and helical). At the end of the assessment, students received consultation and appropriate guidelines in order to improve their intrapersonal and interpersonal skills. Conclusions: Mind and social awareness phenomena play a vital role in human development and may act as determinants of the quality of one’s own life. Meanwhile, brain function is highly correlated with social awareness and it seems that different set of brain structures are involved in social behavior.Keywords: brain activity, emotions, ERP's, social awareness
Procedia PDF Downloads 1933066 Dosimetric Analysis of Intensity Modulated Radiotherapy versus 3D Conformal Radiotherapy in Adult Primary Brain Tumors: Regional Cancer Centre, India
Authors: Ravi Kiran Pothamsetty, Radha Rani Ghosh, Baby Paul Thaliath
Abstract:
Radiation therapy has undergone many advancements and evloved from 2D to 3D. Recently, with rapid pace of drug discoveries, cutting edge technology, and clinical trials has made innovative advancements in computer technology and treatment planning and upgraded to intensity modulated radiotherapy (IMRT) which delivers in homogenous dose to tumor and normal tissues. The present study was a hospital-based experience comparing two different conformal radiotherapy techniques for brain tumors. This analytical study design has been conducted at Regional Cancer Centre, India from January 2014 to January 2015. Ten patients have been selected after inclusion and exclusion criteria. All the patients were treated on Artiste Siemens Linac Accelerator. The tolerance level for maximum dose was 6.0 Gyfor lenses and 54.0 Gy for brain stem, optic chiasm and optical nerves as per RTOG criteria. Mean and standard deviation values of PTV98%, PTV 95% and PTV 2% in IMRT were 93.16±2.9, 95.01±3.4 and 103.1±1.1 respectively; for 3DCRT were 91.4±4.7, 94.17±2.6 and 102.7±0.39 respectively. PTV max dose (%) in IMRT and 3D-CRT were 104.7±0.96 and 103.9±1.0 respectively. Maximum dose to the tumor can be delivered with IMRT with acceptable toxicity limits. Variables such as expertise, location of tumor, patient condition, and TPS influence the outcome of the treatment.Keywords: brain tumors, intensity modulated radiotherapy (IMRT), three dimensional conformal radiotherapy (3D-CRT), radiation therapy oncology group (RTOG)
Procedia PDF Downloads 2393065 A Conv-Long Short-term Memory Deep Learning Model for Traffic Flow Prediction
Authors: Ali Reza Sattarzadeh, Ronny J. Kutadinata, Pubudu N. Pathirana, Van Thanh Huynh
Abstract:
Traffic congestion has become a severe worldwide problem, affecting everyday life, fuel consumption, time, and air pollution. The primary causes of these issues are inadequate transportation infrastructure, poor traffic signal management, and rising population. Traffic flow forecasting is one of the essential and effective methods in urban congestion and traffic management, which has attracted the attention of researchers. With the development of technology, undeniable progress has been achieved in existing methods. However, there is a possibility of improvement in the extraction of temporal and spatial features to determine the importance of traffic flow sequences and extraction features. In the proposed model, we implement the convolutional neural network (CNN) and long short-term memory (LSTM) deep learning models for mining nonlinear correlations and their effectiveness in increasing the accuracy of traffic flow prediction in the real dataset. According to the experiments, the results indicate that implementing Conv-LSTM networks increases the productivity and accuracy of deep learning models for traffic flow prediction.Keywords: deep learning algorithms, intelligent transportation systems, spatiotemporal features, traffic flow prediction
Procedia PDF Downloads 1713064 Radar Fault Diagnosis Strategy Based on Deep Learning
Authors: Bin Feng, Zhulin Zong
Abstract:
Radar systems are critical in the modern military, aviation, and maritime operations, and their proper functioning is essential for the success of these operations. However, due to the complexity and sensitivity of radar systems, they are susceptible to various faults that can significantly affect their performance. Traditional radar fault diagnosis strategies rely on expert knowledge and rule-based approaches, which are often limited in effectiveness and require a lot of time and resources. Deep learning has recently emerged as a promising approach for fault diagnosis due to its ability to learn features and patterns from large amounts of data automatically. In this paper, we propose a radar fault diagnosis strategy based on deep learning that can accurately identify and classify faults in radar systems. Our approach uses convolutional neural networks (CNN) to extract features from radar signals and fault classify the features. The proposed strategy is trained and validated on a dataset of measured radar signals with various types of faults. The results show that it achieves high accuracy in fault diagnosis. To further evaluate the effectiveness of the proposed strategy, we compare it with traditional rule-based approaches and other machine learning-based methods, including decision trees, support vector machines (SVMs), and random forests. The results demonstrate that our deep learning-based approach outperforms the traditional approaches in terms of accuracy and efficiency. Finally, we discuss the potential applications and limitations of the proposed strategy, as well as future research directions. Our study highlights the importance and potential of deep learning for radar fault diagnosis. It suggests that it can be a valuable tool for improving the performance and reliability of radar systems. In summary, this paper presents a radar fault diagnosis strategy based on deep learning that achieves high accuracy and efficiency in identifying and classifying faults in radar systems. The proposed strategy has significant potential for practical applications and can pave the way for further research.Keywords: radar system, fault diagnosis, deep learning, radar fault
Procedia PDF Downloads 903063 Novel Synthesis of Metal Oxide Nanoparticles from Type IV Deep Eutectic Solvents
Authors: Lorenzo Gontrani, Marilena Carbone, Domenica Tommasa Donia, Elvira Maria Bauer, Pietro Tagliatesta
Abstract:
One of the fields where DES shows remarkable added values is the synthesis Of inorganic materials, in particular nanoparticles. In this field, the higher- ent and highly-tunable nano-homogeneities of DES structure give origin to a marked templating effect, a precious role that has led to the recent bloom of a vast number of studies exploiting these new synthesis media to prepare Nanomaterials and composite structures of various kinds. In this contribution, the most recent developments in the field will be reviewed, and some ex-citing examples of novel metal oxide nanoparticles syntheses using non-toxic type-IV Deep Eutectic Solvents will be described. The prepared materials possess nanometric dimensions and show flower-like shapes. The use of the pre- pared nanoparticles as fluorescent materials for the detection of various contaminants is under development.Keywords: metal deep eutectic solvents, nanoparticles, inorganic synthesis, type IV DES, lamellar
Procedia PDF Downloads 1353062 The Effect Study of Meditation Music in the Elderly
Authors: Metee Pigultong
Abstract:
The research aims at 1) composition of meditation music, 2) study of the meditation time reliability. The population is the older adults who meditated practitioners in the Thepnimitra Temple, Don Mueang District, Bangkok. The sample group was the older persons who meditated practitioners from the age of 60 with five volunteers. The research methodology was time-series to conduct the research progression. The research instruments included: 1) meditation music, 2) brain wave recording form. The research results found that 1) the music combines the binaural beats suitable for the meditation of the older persons, consisting of the following features: a) The tempo rate of the meditation music is no more than 60 beats per minute. b) The musical instruments for the meditation music arrangement include only 4-5 pieces. c) The meditation music arrangement needs to consider the nature of the right instrument. d) Digital music instruments are suitable for composition. e) The pure-tone sound combined in music must generate a brain frequency at the level of 10 Hz. 2) After the researcher conducted a 3-weeks brain training procedure, the researcher performed three tests for the reliability level using Cronbach's Alpha method. The result showed that the meditation reliability had the level = .475 as a moderate concentration.Keywords: binaural beats, music therapy, meditation, older person, the Buddhist meditated practitioners
Procedia PDF Downloads 1913061 IoT Based Approach to Healthcare System for a Quadriplegic Patient Using EEG
Authors: R. Gautam, P. Sastha Kanagasabai, G. N. Rathna
Abstract:
The proposed healthcare system enables quadriplegic patients, people with severe motor disabilities to send commands to electronic devices and monitor their vitals. The growth of Brain-Computer-Interface (BCI) has led to rapid development in 'assistive systems' for the disabled called 'assistive domotics'. Brain-Computer-Interface is capable of reading the brainwaves of an individual and analyse it to obtain some meaningful data. This processed data can be used to assist people having speech disorders and sometimes people with limited locomotion to communicate. In this Project, Emotiv EPOC Headset is used to obtain the electroencephalogram (EEG). The obtained data is processed to communicate pre-defined commands over the internet to the desired mobile phone user. Other Vital Information like the heartbeat, blood pressure, ECG and body temperature are monitored and uploaded to the server. Data analytics enables physicians to scan databases for a specific illness. The Data is processed in Intel Edison, system on chip (SoC). Patient metrics are displayed via Intel IoT Analytics cloud service.Keywords: brain computer interface, Intel Edison, Emotiv EPOC, IoT analytics, electroencephalogram
Procedia PDF Downloads 1863060 Facial Emotion Recognition with Convolutional Neural Network Based Architecture
Authors: Koray U. Erbas
Abstract:
Neural networks are appealing for many applications since they are able to learn complex non-linear relationships between input and output data. As the number of neurons and layers in a neural network increase, it is possible to represent more complex relationships with automatically extracted features. Nowadays Deep Neural Networks (DNNs) are widely used in Computer Vision problems such as; classification, object detection, segmentation image editing etc. In this work, Facial Emotion Recognition task is performed by proposed Convolutional Neural Network (CNN)-based DNN architecture using FER2013 Dataset. Moreover, the effects of different hyperparameters (activation function, kernel size, initializer, batch size and network size) are investigated and ablation study results for Pooling Layer, Dropout and Batch Normalization are presented.Keywords: convolutional neural network, deep learning, deep learning based FER, facial emotion recognition
Procedia PDF Downloads 2743059 Whole Body Cooling Hypothermia Treatment Modelling Using a Finite Element Thermoregulation Model
Authors: Ana Beatriz C. G. Silva, Luiz Carlos Wrobel, Fernando Luiz B. Ribeiro
Abstract:
This paper presents a thermoregulation model using the finite element method to perform numerical analyses of brain cooling procedures as a contribution to the investigation on the use of therapeutic hypothermia after ischemia in adults. The use of computational methods can aid clinicians to observe body temperature using different cooling methods without the need of invasive techniques, and can thus be a valuable tool to assist clinical trials simulating different cooling options that can be used for treatment. In this work, we developed a FEM package applied to the solution of the continuum bioheat Pennes equation. Blood temperature changes were considered using a blood pool approach and a lumped analysis for intravascular catheter method of blood cooling. Some analyses are performed using a three-dimensional mesh based on a complex geometry obtained from computed tomography medical images, considering a cooling blanket and a intravascular catheter. A comparison is made between the results obtained and the effects of each case in brain temperature reduction in a required time, maintenance of body temperature at moderate hypothermia levels and gradual rewarming.Keywords: brain cooling, finite element method, hypothermia treatment, thermoregulation
Procedia PDF Downloads 3113058 Distributed System Computing Resource Scheduling Algorithm Based on Deep Reinforcement Learning
Authors: Yitao Lei, Xingxiang Zhai, Burra Venkata Durga Kumar
Abstract:
As the quantity and complexity of computing in large-scale software systems increase, distributed system computing becomes increasingly important. The distributed system realizes high-performance computing by collaboration between different computing resources. If there are no efficient resource scheduling resources, the abuse of distributed computing may cause resource waste and high costs. However, resource scheduling is usually an NP-hard problem, so we cannot find a general solution. However, some optimization algorithms exist like genetic algorithm, ant colony optimization, etc. The large scale of distributed systems makes this traditional optimization algorithm challenging to work with. Heuristic and machine learning algorithms are usually applied in this situation to ease the computing load. As a result, we do a review of traditional resource scheduling optimization algorithms and try to introduce a deep reinforcement learning method that utilizes the perceptual ability of neural networks and the decision-making ability of reinforcement learning. Using the machine learning method, we try to find important factors that influence the performance of distributed system computing and help the distributed system do an efficient computing resource scheduling. This paper surveys the application of deep reinforcement learning on distributed system computing resource scheduling proposes a deep reinforcement learning method that uses a recurrent neural network to optimize the resource scheduling, and proposes the challenges and improvement directions for DRL-based resource scheduling algorithms.Keywords: resource scheduling, deep reinforcement learning, distributed system, artificial intelligence
Procedia PDF Downloads 1113057 Effect of Depth on the Distribution of Zooplankton in Wushishi Lake Minna, Niger State, Nigeria
Authors: Adamu Zubairu Mohammed, Fransis Oforum Arimoro, Salihu Maikudi Ibrahim, Y. I. Auta, T. I. Arowosegbe, Y. Abdullahi
Abstract:
The present study was conducted to evaluate the effect of depth on the distribution of zooplankton and some physicochemical parameters in Tungan Kawo Lake (Wushishi dam). Water and zooplankton samples were collected from the surface, 3.0 meters deep and 6.0 meters deep, for a period of 24 hours for six months. Standard procedures were adopted for the determination of physicochemical parameters. Results have shown significant differences in the pH, DO, BOD Hardness, Na, and Mg. A total of 1764 zooplankton were recorded, comprising 35 species, with cladocera having 18 species (58%), 14 species of copepoda (41%), 3 species of diptera (1.0%). Results show that more of the zooplankton were recorded in the 3.0 meters-deep region compared to the two other depts and a significant difference was observed in the distribution of Ceriodaphnia dubia, Daphnia laevis, and Leptodiaptomus coloradensis. Though the most abundant zooplankton was recorded in the 3.0 meters deep, Leptodiaptomus coloradesnsis, which was observed in the 6.0 meters deep as the most individual observed, this was followed by Daphnia laevis. Canonical correspondence analysis between physicochemical parameters and the zooplankton indicated a good relationship in the Lake. Ceriodaphnia dubia was found to have a good association with oxygen, sodium, and potassium, while Daphnia laevis and Leptodiaptomus coloradensis are in good relationship with magnesium and phosphorus. It was generally observed that this depth does not have much influence on the distribution of zooplankton in Wushishi Lake.Keywords: zooplankton, standard procedures, canonical correspondence analysis, Wushishi, canonical, physicochemical parameter
Procedia PDF Downloads 903056 Development of Web-Based Iceberg Detection Using Deep Learning
Authors: A. Kavya Sri, K. Sai Vineela, R. Vanitha, S. Rohith
Abstract:
Large pieces of ice that break from the glaciers are known as icebergs. The threat that icebergs pose to navigation, production of offshore oil and gas services, and underwater pipelines makes their detection crucial. In this project, an automated iceberg tracking method using deep learning techniques and satellite images of icebergs is to be developed. With a temporal resolution of 12 days and a spatial resolution of 20 m, Sentinel-1 (SAR) images can be used to track iceberg drift over the Southern Ocean. In contrast to multispectral images, SAR images are used for analysis in meteorological conditions. This project develops a web-based graphical user interface to detect and track icebergs using sentinel-1 images. To track the movement of the icebergs by using temporal images based on their latitude and longitude values and by comparing the center and area of all detected icebergs. Testing the accuracy is done by precision and recall measures.Keywords: synthetic aperture radar (SAR), icebergs, deep learning, spatial resolution, temporal resolution
Procedia PDF Downloads 913055 Deep Supervision Based-Unet to Detect Buildings Changes from VHR Aerial Imagery
Authors: Shimaa Holail, Tamer Saleh, Xiongwu Xiao
Abstract:
Building change detection (BCD) from satellite imagery is an essential topic in urbanization monitoring, agricultural land management, and updating geospatial databases. Recently, methods for detecting changes based on deep learning have made significant progress and impressive results. However, it has the problem of being insensitive to changes in buildings with complex spectral differences, and the features being extracted are not discriminatory enough, resulting in incomplete buildings and irregular boundaries. To overcome these problems, we propose a dual Siamese network based on the Unet model with the addition of a deep supervision strategy (DS) in this paper. This network consists of a backbone (encoder) based on ImageNet pre-training, a fusion block, and feature pyramid networks (FPN) to enhance the step-by-step information of the changing regions and obtain a more accurate BCD map. To train the proposed method, we created a new dataset (EGY-BCD) of high-resolution and multi-temporal aerial images captured over New Cairo in Egypt to detect building changes for this purpose. The experimental results showed that the proposed method is effective and performs well with the EGY-BCD dataset regarding the overall accuracy, F1-score, and mIoU, which were 91.6 %, 80.1 %, and 73.5 %, respectively.Keywords: building change detection, deep supervision, semantic segmentation, EGY-BCD dataset
Procedia PDF Downloads 1203054 Computer Aided Analysis of Breast Based Diagnostic Problems from Mammograms Using Image Processing and Deep Learning Methods
Authors: Ali Berkan Ural
Abstract:
This paper presents the analysis, evaluation, and pre-diagnosis of early stage breast based diagnostic problems (breast cancer, nodulesorlumps) by Computer Aided Diagnosing (CAD) system from mammogram radiological images. According to the statistics, the time factor is crucial to discover the disease in the patient (especially in women) as possible as early and fast. In the study, a new algorithm is developed using advanced image processing and deep learning method to detect and classify the problem at earlystagewithmoreaccuracy. This system first works with image processing methods (Image acquisition, Noiseremoval, Region Growing Segmentation, Morphological Operations, Breast BorderExtraction, Advanced Segmentation, ObtainingRegion Of Interests (ROIs), etc.) and segments the area of interest of the breast and then analyzes these partly obtained area for cancer detection/lumps in order to diagnosis the disease. After segmentation, with using the Spectrogramimages, 5 different deep learning based methods (specified Convolutional Neural Network (CNN) basedAlexNet, ResNet50, VGG16, DenseNet, Xception) are applied to classify the breast based problems.Keywords: computer aided diagnosis, breast cancer, region growing, segmentation, deep learning
Procedia PDF Downloads 963053 Development of Electroencephalograph Collection System in Language-Learning Self-Study System That Can Detect Learning State of the Learner
Authors: Katsuyuki Umezawa, Makoto Nakazawa, Manabu Kobayashi, Yutaka Ishii, Michiko Nakano, Shigeichi Hirasawa
Abstract:
This research aims to develop a self-study system equipped with an artificial teacher who gives advice to students by detecting the learners and to evaluate language learning in a unified framework. 'Detecting the learners' means that the system understands the learners' learning conditions, such as each learner’s degree of understanding, the difference in each learner’s thinking process, the degree of concentration or boredom in learning, and problem solving for each learner, which can be interpreted from learning behavior. In this paper, we propose a system to efficiently collect brain waves from learners by focusing on only the brain waves among the biological information for 'detecting the learners'. The conventional Electroencephalograph (EEG) measurement method during learning using a simple EEG has the following disadvantages. (1) The start and end of EEG measurement must be done manually by the experiment participant or staff. (2) Even when the EEG signal is weak, it may not be noticed, and the data may not be obtained. (3) Since the acquired EEG data is stored in each PC, there is a possibility that the time of data acquisition will be different in each PC. This time, we developed a system to collect brain wave data on the server side. This system overcame the above disadvantages.Keywords: artificial teacher, e-learning, self-study system, simple EEG
Procedia PDF Downloads 1453052 Nonparametric Sieve Estimation with Dependent Data: Application to Deep Neural Networks
Authors: Chad Brown
Abstract:
This paper establishes general conditions for the convergence rates of nonparametric sieve estimators with dependent data. We present two key results: one for nonstationary data and another for stationary mixing data. Previous theoretical results often lack practical applicability to deep neural networks (DNNs). Using these conditions, we derive convergence rates for DNN sieve estimators in nonparametric regression settings with both nonstationary and stationary mixing data. The DNN architectures considered adhere to current industry standards, featuring fully connected feedforward networks with rectified linear unit activation functions, unbounded weights, and a width and depth that grows with sample size.Keywords: sieve extremum estimates, nonparametric estimation, deep learning, neural networks, rectified linear unit, nonstationary processes
Procedia PDF Downloads 423051 The Effects of Electrical Muscle Stimulation (EMS) towards Male Skeletal Muscle Mass
Authors: Mohd Faridz Ahmad, Amirul Hakim Hasbullah
Abstract:
Electrical Muscle Stimulation (EMS) has been introduced to the world in the 19th and 20th centuries and has globally gained increasing attention on its usefulness. EMS is known as the application of electrical current transcutaneous to muscles through electrodes to induce involuntary contractions that can lead to the increment of muscle mass and strength. This study can be used as an alternative to help people especially those living a sedentary lifestyle to improve their muscle activity without having to go through a heavy workout session. Therefore, this study intended to investigate the effectiveness of EMS training in 5 weeks interventions towards male body composition. It was a quasi-experimental design, held at the Impulse Studio Bangsar, which examined the effects of EMS training towards skeletal muscle mass among the subjects. Fifteen subjects (n = 15) were selected to assist in this study. The demographic data showed that, the average age of the subjects was 43.07 years old ± 9.90, height (173.4 cm ± 9.09) and weight was (85.79 kg ± 18.07). Results showed that there was a significant difference on the skeletal muscle mass (p = 0.01 < 0.05), upper body (p = 0.01 < 0.05) and lower body (p = 0.00 < 0.05). Therefore, the null hypothesis has been rejected in this study. As a conclusion, the application of EMS towards body composition can increase the muscle size and strength. This method has been proven to be able to improve athlete strength and thus, may be implemented in the sports science area of knowledge.Keywords: body composition, EMS, skeletal muscle mass, strength
Procedia PDF Downloads 4903050 Brain-Computer Interfaces That Use Electroencephalography
Authors: Arda Ozkurt, Ozlem Bozkurt
Abstract:
Brain-computer interfaces (BCIs) are devices that output commands by interpreting the data collected from the brain. Electroencephalography (EEG) is a non-invasive method to measure the brain's electrical activity. Since it was invented by Hans Berger in 1929, it has led to many neurological discoveries and has become one of the essential components of non-invasive measuring methods. Despite the fact that it has a low spatial resolution -meaning it is able to detect when a group of neurons fires at the same time-, it is a non-invasive method, making it easy to use without possessing any risks. In EEG, electrodes are placed on the scalp, and the voltage difference between a minimum of two electrodes is recorded, which is then used to accomplish the intended task. The recordings of EEGs include, but are not limited to, the currents along dendrites from synapses to the soma, the action potentials along the axons connecting neurons, and the currents through the synaptic clefts connecting axons with dendrites. However, there are some sources of noise that may affect the reliability of the EEG signals as it is a non-invasive method. For instance, the noise from the EEG equipment, the leads, and the signals coming from the subject -such as the activity of the heart or muscle movements- affect the signals detected by the electrodes of the EEG. However, new techniques have been developed to differentiate between those signals and the intended ones. Furthermore, an EEG device is not enough to analyze the data from the brain to be used by the BCI implication. Because the EEG signal is very complex, to analyze it, artificial intelligence algorithms are required. These algorithms convert complex data into meaningful and useful information for neuroscientists to use the data to design BCI devices. Even though for neurological diseases which require highly precise data, invasive BCIs are needed; non-invasive BCIs - such as EEGs - are used in many cases to help disabled people's lives or even to ease people's lives by helping them with basic tasks. For example, EEG is used to detect before a seizure occurs in epilepsy patients, which can then prevent the seizure with the help of a BCI device. Overall, EEG is a commonly used non-invasive BCI technique that has helped develop BCIs and will continue to be used to detect data to ease people's lives as more BCI techniques will be developed in the future.Keywords: BCI, EEG, non-invasive, spatial resolution
Procedia PDF Downloads 713049 Field Prognostic Factors on Discharge Prediction of Traumatic Brain Injuries
Authors: Mohammad Javad Behzadnia, Amir Bahador Boroumand
Abstract:
Introduction: Limited facility situations require allocating the most available resources for most casualties. Accordingly, Traumatic Brain Injury (TBI) is the one that may need to transport the patient as soon as possible. In a mass casualty event, deciding when the facilities are restricted is hard. The Extended Glasgow Outcome Score (GOSE) has been introduced to assess the global outcome after brain injuries. Therefore, we aimed to evaluate the prognostic factors associated with GOSE. Materials and Methods: In a multicenter cross-sectional study conducted on 144 patients with TBI admitted to trauma emergency centers. All the patients with isolated TBI who were mentally and physically healthy before the trauma entered the study. The patient’s information was evaluated, including demographic characteristics, duration of hospital stays, mechanical ventilation on admission laboratory measurements, and on-admission vital signs. We recorded the patients’ TBI-related symptoms and brain computed tomography (CT) scan findings. Results: GOSE assessments showed an increasing trend by the comparison of on-discharge (7.47 ± 1.30), within a month (7.51 ± 1.30), and within three months (7.58 ± 1.21) evaluations (P < 0.001). On discharge, GOSE was positively correlated with Glasgow Coma Scale (GCS) (r = 0.729, P < 0.001) and motor GCS (r = 0.812, P < 0.001), and inversely with age (r = −0.261, P = 0.002), hospitalization period (r = −0.678, P < 0.001), pulse rate (r = −0.256, P = 0.002) and white blood cell (WBC). Among imaging signs and trauma-related symptoms in univariate analysis, intracranial hemorrhage (ICH), interventricular hemorrhage (IVH) (P = 0.006), subarachnoid hemorrhage (SAH) (P = 0.06; marginally at P < 0.1), subdural hemorrhage (SDH) (P = 0.032), and epidural hemorrhage (EDH) (P = 0.037) were significantly associated with GOSE at discharge in multivariable analysis. Conclusion: Our study showed some predictive factors that could help to decide which casualty should transport earlier to a trauma center. According to the current study findings, GCS, pulse rate, WBC, and among imaging signs and trauma-related symptoms, ICH, IVH, SAH, SDH, and EDH are significant independent predictors of GOSE at discharge in TBI patients.Keywords: field, Glasgow outcome score, prediction, traumatic brain injury.
Procedia PDF Downloads 763048 Comparison Study between Deep Mixed Columns and Encased Sand Column for Soft Clay Soil in Egypt
Authors: Walid El Kamash
Abstract:
Sand columns (or granular piles) can be employed as soil strengthening for flexible constructions such as road embankments, oil storage tanks in addition to multistory structures. The challenge of embedding the sand columns in soft soil is that the surrounding soft soil cannot avail the enough confinement stress in order to keep the form of the sand column. Therefore, the sand columns which were installed in such soil will lose their ability to perform needed load-bearing capacity. The encasement, besides increasing the strength and stiffness of the sand column, prevents the lateral squeezing of sands when the column is installed even in extremely soft soils, thus enabling quicker and more economical installation. This paper investigates the improvement in load capacity of the sand column by encasement through a comprehensive parametric study using the 3-D finite difference analysis for the soft clay of soil in Egypt. Moreover, the study was extended to include a comparison study between encased sand column and Deep Mixed columns (DM). The study showed that confining the sand by geosynthetic resulted in an increment of shear strength. That result paid the attention to use encased sand stone rather than deep mixed columns due to relative high permeability of the first material.Keywords: encased sand column, Deep mixed column, numerical analysis, improving soft soil
Procedia PDF Downloads 3783047 Comparing the Detection of Autism Spectrum Disorder within Males and Females Using Machine Learning Techniques
Authors: Joseph Wolff, Jeffrey Eilbott
Abstract:
Autism Spectrum Disorders (ASD) are a spectrum of social disorders characterized by deficits in social communication, verbal ability, and interaction that can vary in severity. In recent years, researchers have used magnetic resonance imaging (MRI) to help detect how neural patterns in individuals with ASD differ from those of neurotypical (NT) controls for classification purposes. This study analyzed the classification of ASD within males and females using functional MRI data. Functional connectivity (FC) correlations among brain regions were used as feature inputs for machine learning algorithms. Analysis was performed on 558 cases from the Autism Brain Imaging Data Exchange (ABIDE) I dataset. When trained specifically on females, the algorithm underperformed in classifying the ASD subset of our testing population. Although the subject size was relatively smaller in the female group, the manual matching of both male and female training groups helps explain the algorithm’s bias, indicating the altered sex abnormalities in functional brain networks compared to typically developing peers. These results highlight the importance of taking sex into account when considering how generalizations of findings on males with ASD apply to females.Keywords: autism spectrum disorder, machine learning, neuroimaging, sex differences
Procedia PDF Downloads 2093046 Utilizing Federated Learning for Accurate Prediction of COVID-19 from CT Scan Images
Authors: Jinil Patel, Sarthak Patel, Sarthak Thakkar, Deepti Saraswat
Abstract:
Recently, the COVID-19 outbreak has spread across the world, leading the World Health Organization to classify it as a global pandemic. To save the patient’s life, the COVID-19 symptoms have to be identified. But using an AI (Artificial Intelligence) model to identify COVID-19 symptoms within the allotted time was challenging. The RT-PCR test was found to be inadequate in determining the COVID status of a patient. To determine if the patient has COVID-19 or not, a Computed Tomography Scan (CT scan) of patient is a better alternative. It will be challenging to compile and store all the data from various hospitals on the server, though. Federated learning, therefore, aids in resolving this problem. Certain deep learning models help to classify Covid-19. This paper will have detailed work of certain deep learning models like VGG19, ResNet50, MobileNEtv2, and Deep Learning Aggregation (DLA) along with maintaining privacy with encryption.Keywords: federated learning, COVID-19, CT-scan, homomorphic encryption, ResNet50, VGG-19, MobileNetv2, DLA
Procedia PDF Downloads 733045 Trans-Activator of Transcription-Tagged Active AKT1 Variants for Delivery to Mammalian Cells
Authors: Tarana Siddika, Ilka U. Heinemann, Patrick O’Donoghue
Abstract:
Protein kinase B (AKT1) is a serine/threonine kinase and central transducer of cell survival pathways. Typical approaches to study AKT1 biology in cells rely on growth factor or insulin stimulation that activates AKT1 via phosphorylation at two key regulatory sites (Threonine308, Serine473), yet cell stimulation also activates many other kinases and fails to differentiate the effect of the two main activating sites of AKT1 on downstream substrate phosphorylation and cell growth. While both AKT1 activating sites are associated with disease and used as clinical markers, in some cancers, high levels of Threonine308 phosphorylation are associated with poor prognosis while in others poor survival correlates with high Serine473 levels. To produce cells with specific AKT1 activity, a system was developed to deliver active AKT1 to human cells. AKT1 phospho-variants were produced from Escherichia coli with programmed phosphorylation by genetic code expansion. Tagging of AKT1 with an N-terminal cell penetrating peptide tag derived from the human immunodeficiency virus trans-activator of transcription (TAT) helped to enter AKT1 proteins in mammalian cells. The TAT-tag did not alter AKT1 kinase activity and was necessary and sufficient to rapidly deliver AKT1 protein variants that persisted in human cells for 24 h without the need to use transfection reagents. TAT-pAKT1T308, TAT-pAKT1S473 and TAT-pAKT1T308S473 proteins induced selective phosphorylation of the known AKT1 substrate GSK-3αβ, and downstream stimulation of the AKT1 pathway as evidenced by phosphorylation of ribosomal protein S6 at Serine240/244 in transfected cells. Increase in cell growth and proliferation was observed due to the transfection of different phosphorylated AKT1 protein variants compared to cells with TAT-AKT1 protein. The data demonstrate efficient delivery of AKT1 with programmed phosphorylation to human cells, thus establishing a cell-based model system to investigate signaling that is dependent on specific AKT1 activity and phosphorylation.Keywords: cell penetrating peptide, cell signaling, protein kinase b (AKT1), phosphorylation
Procedia PDF Downloads 1183044 Impact of Interventions on Brain Functional Connectivity in Young Male Basketball Players: A Comparative Study
Authors: Mohammad Khazaei, Reza Rostami, Hassan Gharayagh Zandi, Ruhollah Basatnia, Mahboubeh Ghayour Najafabadi
Abstract:
Introduction: This study delves into the influence of diverse interventions on brain functional connectivity among young male basketball players. Given the significance of understanding how interventions affect cognitive functions in athletes, particularly in the context of basketball, this research contributes to the growing body of knowledge in sports neuroscience. Methods: Three distinct groups were selected for comprehensive investigation: the Motivational Interview Group, Placebo Consumption Group, and Ritalin Consumption Group. The study involved assessing brain functional connectivity using various frequency bands (Delta, Theta, Alpha, Beta1, Beta2, Gamma, and Total Band) before and after the interventions. The participants were subjected to specific interventions corresponding to their assigned groups. Results: The findings revealed substantial differences in brain functional connectivity across the studied groups. The Motivational Interview Group exhibited optimal outcomes in PLI (Total Band) connectivity. The Placebo Consumption Group demonstrated a marked impact on PLV (Alpha) connectivity, and the Ritalin Consumption Group experienced a considerable enhancement in imCoh (Total Band) connectivity. Discussion: The observed variations in brain functional connectivity underscore the nuanced effects of different interventions on young male basketball players. The enhanced connectivity in specific frequency bands suggests potential cognitive and performance improvements. Notably, the Motivational Interview and Placebo Consumption groups displayed unique patterns, emphasizing the multifaceted nature of interventions. These findings contribute to the understanding of tailored interventions for optimizing cognitive functions in young male basketball players. Conclusion: This study provides valuable insights into the intricate relationship between interventions and brain functional connectivity in young male basketball players. Further research with expanded sample sizes and more sophisticated statistical analyses is recommended to corroborate and expand upon these initial findings. The implications of this study extend to the broader field of sports neuroscience, aiding in the development of targeted interventions for athletes in various disciplines.Keywords: electroencephalography, Ritalin, Placebo effect, motivational interview
Procedia PDF Downloads 643043 Physics Informed Deep Residual Networks Based Type-A Aortic Dissection Prediction
Abstract:
Purpose: Acute Type A aortic dissection is a well-known cause of extremely high mortality rate. A highly accurate and cost-effective non-invasive predictor is critically needed so that the patient can be treated at earlier stage. Although various CFD approaches have been tried to establish some prediction frameworks, they are sensitive to uncertainty in both image segmentation and boundary conditions. Tedious pre-processing and demanding calibration procedures requirement further compound the issue, thus hampering their clinical applicability. Using the latest physics informed deep learning methods to establish an accurate and cost-effective predictor framework are amongst the main goals for a better Type A aortic dissection treatment. Methods: Via training a novel physics-informed deep residual network, with non-invasive 4D MRI displacement vectors as inputs, the trained model can cost-effectively calculate all these biomarkers: aortic blood pressure, WSS, and OSI, which are used to predict potential type A aortic dissection to avoid the high mortality events down the road. Results: The proposed deep learning method has been successfully trained and tested with both synthetic 3D aneurysm dataset and a clinical dataset in the aortic dissection context using Google colab environment. In both cases, the model has generated aortic blood pressure, WSS, and OSI results matching the expected patient’s health status. Conclusion: The proposed novel physics-informed deep residual network shows great potential to create a cost-effective, non-invasive predictor framework. Additional physics-based de-noising algorithm will be added to make the model more robust to clinical data noises. Further studies will be conducted in collaboration with big institutions such as Cleveland Clinic with more clinical samples to further improve the model’s clinical applicability.Keywords: type-a aortic dissection, deep residual networks, blood flow modeling, data-driven modeling, non-invasive diagnostics, deep learning, artificial intelligence.
Procedia PDF Downloads 893042 A Deep Learning Based Integrated Model For Spatial Flood Prediction
Authors: Vinayaka Gude Divya Sampath
Abstract:
The research introduces an integrated prediction model to assess the susceptibility of roads in a future flooding event. The model consists of deep learning algorithm for forecasting gauge height data and Flood Inundation Mapper (FIM) for spatial flooding. An optimal architecture for Long short-term memory network (LSTM) was identified for the gauge located on Tangipahoa River at Robert, LA. Dropout was applied to the model to evaluate the uncertainty associated with the predictions. The estimates are then used along with FIM to identify the spatial flooding. Further geoprocessing in ArcGIS provides the susceptibility values for different roads. The model was validated based on the devastating flood of August 2016. The paper discusses the challenges for generalization the methodology for other locations and also for various types of flooding. The developed model can be used by the transportation department and other emergency response organizations for effective disaster management.Keywords: deep learning, disaster management, flood prediction, urban flooding
Procedia PDF Downloads 1473041 Single-Camera Basketball Tracker through Pose and Semantic Feature Fusion
Authors: Adrià Arbués-Sangüesa, Coloma Ballester, Gloria Haro
Abstract:
Tracking sports players is a widely challenging scenario, specially in single-feed videos recorded in tight courts, where cluttering and occlusions cannot be avoided. This paper presents an analysis of several geometric and semantic visual features to detect and track basketball players. An ablation study is carried out and then used to remark that a robust tracker can be built with Deep Learning features, without the need of extracting contextual ones, such as proximity or color similarity, nor applying camera stabilization techniques. The presented tracker consists of: (1) a detection step, which uses a pretrained deep learning model to estimate the players pose, followed by (2) a tracking step, which leverages pose and semantic information from the output of a convolutional layer in a VGG network. Its performance is analyzed in terms of MOTA over a basketball dataset with more than 10k instances.Keywords: basketball, deep learning, feature extraction, single-camera, tracking
Procedia PDF Downloads 1383040 DLtrace: Toward Understanding and Testing Deep Learning Information Flow in Deep Learning-Based Android Apps
Authors: Jie Zhang, Qianyu Guo, Tieyi Zhang, Zhiyong Feng, Xiaohong Li
Abstract:
With the widespread popularity of mobile devices and the development of artificial intelligence (AI), deep learning (DL) has been extensively applied in Android apps. Compared with traditional Android apps (traditional apps), deep learning based Android apps (DL-based apps) need to use more third-party application programming interfaces (APIs) to complete complex DL inference tasks. However, existing methods (e.g., FlowDroid) for detecting sensitive information leakage in Android apps cannot be directly used to detect DL-based apps as they are difficult to detect third-party APIs. To solve this problem, we design DLtrace; a new static information flow analysis tool that can effectively recognize third-party APIs. With our proposed trace and detection algorithms, DLtrace can also efficiently detect privacy leaks caused by sensitive APIs in DL-based apps. Moreover, using DLtrace, we summarize the non-sequential characteristics of DL inference tasks in DL-based apps and the specific functionalities provided by DL models for such apps. We propose two formal definitions to deal with the common polymorphism and anonymous inner-class problems in the Android static analyzer. We conducted an empirical assessment with DLtrace on 208 popular DL-based apps in the wild and found that 26.0% of the apps suffered from sensitive information leakage. Furthermore, DLtrace has a more robust performance than FlowDroid in detecting and identifying third-party APIs. The experimental results demonstrate that DLtrace expands FlowDroid in understanding DL-based apps and detecting security issues therein.Keywords: mobile computing, deep learning apps, sensitive information, static analysis
Procedia PDF Downloads 1793039 Postmortem Analysis of Lidocaine in Women Died of Criminal Abortion
Authors: Mohammed A. Arishy, Sultan M. Alharbi, Mohammed A. Hakami, Farid M. Abualsail, Mohammad A. Attafi, Riyadh M. Tobaiqi, Hussain M. Alsalem, Ibraheem M. Attafi
Abstract:
Lidocaine is the most common local anesthetics used for para cervical block to reduce pain associated with surgical abortion. A 25-year-old pregnant woman who. She died before reaching hospital, and she was undergoing criminal abortion during the first trimester. In post-mortem investigations and autopsy shows no clear finding; therefore, toxic substances must be suspected and searched for routinely toxicology analysis. In this case report, the postmortem concentration of lidocaine was detected blood, brain, liver, kidney, and stomach. For lidocaine identification and quantification, sample was extracted using solid phase extraction and analyzed by GC-MS (Shimadzu, Japan). Initial screening and confirmatory analysis results showed that only lidocaine was detected in all collected samples, and no other toxic substances or alcohol were detected. The concentrations of lidocaine in samples were 19, 17, 14, 7, and 3 ug/m in the brain, blood, kidney, liver, and stomach, respectively. Lidocaine blood concentration (17 ug/ml) was toxic level and may result in death. Among the tissues, brain showed the highest level of lidocaine, followed by the kidney, liver, and stomach.Keywords: forensic toxicology, GC-MS, lidocaine, postmortem
Procedia PDF Downloads 210