Search results for: edge AI device
428 A Comprehensive Methodology for Voice Segmentation of Large Sets of Speech Files Recorded in Naturalistic Environments
Authors: Ana Londral, Burcu Demiray, Marcus Cheetham
Abstract:
Speech recording is a methodology used in many different studies related to cognitive and behaviour research. Modern advances in digital equipment brought the possibility of continuously recording hours of speech in naturalistic environments and building rich sets of sound files. Speech analysis can then extract from these files multiple features for different scopes of research in Language and Communication. However, tools for analysing a large set of sound files and automatically extract relevant features from these files are often inaccessible to researchers that are not familiar with programming languages. Manual analysis is a common alternative, with a high time and efficiency cost. In the analysis of long sound files, the first step is the voice segmentation, i.e. to detect and label segments containing speech. We present a comprehensive methodology aiming to support researchers on voice segmentation, as the first step for data analysis of a big set of sound files. Praat, an open source software, is suggested as a tool to run a voice detection algorithm, label segments and files and extract other quantitative features on a structure of folders containing a large number of sound files. We present the validation of our methodology with a set of 5000 sound files that were collected in the daily life of a group of voluntary participants with age over 65. A smartphone device was used to collect sound using the Electronically Activated Recorder (EAR): an app programmed to record 30-second sound samples that were randomly distributed throughout the day. Results demonstrated that automatic segmentation and labelling of files containing speech segments was 74% faster when compared to a manual analysis performed with two independent coders. Furthermore, the methodology presented allows manual adjustments of voiced segments with visualisation of the sound signal and the automatic extraction of quantitative information on speech. In conclusion, we propose a comprehensive methodology for voice segmentation, to be used by researchers that have to work with large sets of sound files and are not familiar with programming tools.Keywords: automatic speech analysis, behavior analysis, naturalistic environments, voice segmentation
Procedia PDF Downloads 281427 Charge Transport of Individual Thermoelectric Bi₂Te₃ Core-Poly(3,4-Ethylenedioxythiophene):Polystyrenesulfonate Shell Nanowires Determined Using Conductive Atomic Force Microscopy and Spectroscopy
Authors: W. Thongkham, K. Sinthiptharakoon, K. Tantisantisom, A. Klamchuen, P. Khanchaitit, K. Jiramitmongkon, C. Lertsatitthanakorn, M. Liangruksa
Abstract:
Due to demands of sustainable energy, thermoelectricity converting waste heat into electrical energy has become one of the intensive fields of worldwide research. However, such harvesting technology has shown low device performance in the temperature range below 150℃. In this work, a hybrid nanowire of inorganic bismuth telluride (Bi₂Te₃) and organic poly(3,4-ethylenedioxythiophene):polystyrenesulfonate (PEDOT:PSS) synthesized using a simple in-situ one-pot synthesis, enhancing efficiency of the nanowire-incorporated PEDOT:PSS-based thermoelectric converter is highlighted. Since the improvement is ascribed to the increased electrical conductivity of the thermoelectric host material, the individual hybrid nanowires are investigated using voltage-dependent conductive atomic force microscopy (CAFM) and spectroscopy (CAFS) considering that the electrical transport measurement can be performed either on insulating or conducting areas of the sample. Correlated with detailed chemical information on the crystalline structure and compositional profile of the nanowire core-shell structure, an electrical transporting pathway through the nanowire and the corresponding electronic-band structure have been determined, in which the native oxide layer on the Bi₂Te₃ surface is not considered, and charge conduction on the topological surface states of Bi₂Te₃ is suggested. Analyzing the core-shell nanowire synthesized using the conventional mixing of as-prepared Bi₂Te₃ nanowire with PEDOT:PSS for comparison, the oxide-removal effect of the in-situ encapsulating polymeric layer is further supported. The finding not only provides a structural information for mechanistic determination of the thermoelectricity, but it also encourages new approach toward more appropriate encapsulation and consequently higher efficiency of the nanowire-based thermoelectric generation.Keywords: electrical transport measurement, hybrid Bi₂Te₃-PEDOT:PSS nanowire, nanoencapsulation, thermoelectricity, topological insulator
Procedia PDF Downloads 205426 Monitoring of Water Quality Using Wireless Sensor Network: Case Study of Benue State of Nigeria
Authors: Desmond Okorie, Emmanuel Prince
Abstract:
Availability of portable water has been a global challenge especially to the developing continents/nations such as Africa/Nigeria. The World Health Organization WHO has produced the guideline for drinking water quality GDWQ which aims at ensuring water safety from source to consumer. Portable water parameters test include physical (colour, odour, temperature, turbidity), chemical (PH, dissolved solids) biological (algae, plytoplankton). This paper discusses the use of wireless sensor networks to monitor water quality using efficient and effective sensors that have the ability to sense, process and transmit sensed data. The integration of wireless sensor network to a portable sensing device offers the feasibility of sensing distribution capability, on site data measurements and remote sensing abilities. The current water quality tests that are performed in government water quality institutions in Benue State Nigeria are carried out in problematic locations that require taking manual water samples to the institution laboratory for examination, to automate the entire process based on wireless sensor network, a system was designed. The system consists of sensor node containing one PH sensor, one temperature sensor, a microcontroller, a zigbee radio and a base station composed by a zigbee radio and a PC. Due to the advancement of wireless sensor network technology, unexpected contamination events in water environments can be observed continuously. local area network (LAN) wireless local area network (WLAN) and internet web-based also commonly used as a gateway unit for data communication via local base computer using standard global system for mobile communication (GSM). The improvement made on this development show a water quality monitoring system and prospect for more robust and reliable system in the future.Keywords: local area network, Ph measurement, wireless sensor network, zigbee
Procedia PDF Downloads 172425 Skull Extraction for Quantification of Brain Volume in Magnetic Resonance Imaging of Multiple Sclerosis Patients
Authors: Marcela De Oliveira, Marina P. Da Silva, Fernando C. G. Da Rocha, Jorge M. Santos, Jaime S. Cardoso, Paulo N. Lisboa-Filho
Abstract:
Multiple Sclerosis (MS) is an immune-mediated disease of the central nervous system characterized by neurodegeneration, inflammation, demyelination, and axonal loss. Magnetic resonance imaging (MRI), due to the richness in the information details provided, is the gold standard exam for diagnosis and follow-up of neurodegenerative diseases, such as MS. Brain atrophy, the gradual loss of brain volume, is quite extensive in multiple sclerosis, nearly 0.5-1.35% per year, far off the limits of normal aging. Thus, the brain volume quantification becomes an essential task for future analysis of the occurrence atrophy. The analysis of MRI has become a tedious and complex task for clinicians, who have to manually extract important information. This manual analysis is prone to errors and is time consuming due to various intra- and inter-operator variability. Nowadays, computerized methods for MRI segmentation have been extensively used to assist doctors in quantitative analyzes for disease diagnosis and monitoring. Thus, the purpose of this work was to evaluate the brain volume in MRI of MS patients. We used MRI scans with 30 slices of the five patients diagnosed with multiple sclerosis according to the McDonald criteria. The computational methods for the analysis of images were carried out in two steps: segmentation of the brain and brain volume quantification. The first image processing step was to perform brain extraction by skull stripping from the original image. In the skull stripper for MRI images of the brain, the algorithm registers a grayscale atlas image to the grayscale patient image. The associated brain mask is propagated using the registration transformation. Then this mask is eroded and used for a refined brain extraction based on level-sets (edge of the brain-skull border with dedicated expansion, curvature, and advection terms). In the second step, the brain volume quantification was performed by counting the voxels belonging to the segmentation mask and converted in cc. We observed an average brain volume of 1469.5 cc. We concluded that the automatic method applied in this work can be used for the brain extraction process and brain volume quantification in MRI. The development and use of computer programs can contribute to assist health professionals in the diagnosis and monitoring of patients with neurodegenerative diseases. In future works, we expect to implement more automated methods for the assessment of cerebral atrophy and brain lesions quantification, including machine-learning approaches. Acknowledgements: This work was supported by a grant from Brazilian agency Fundação de Amparo à Pesquisa do Estado de São Paulo (number 2019/16362-5).Keywords: brain volume, magnetic resonance imaging, multiple sclerosis, skull stripper
Procedia PDF Downloads 146424 Quality Assurance Comparison of Map Check 2, Epid, and Gafchromic® EBT3 Film for IMRT Treatment Planning
Authors: Khalid Iqbal, Saima Altaf, M. Akram, Muhammad Abdur Rafaye, Saeed Ahmad Buzdar
Abstract:
Objective: Verification of patient-specific intensity modulated radiation therapy (IMRT) plans using different 2-D detectors has become increasingly popular due to their ease of use and immediate readout of the results. The purpose of this study was to test and compare various 2-D detectors for dosimetric quality assurance (QA) of intensity-modulated radiotherapy (IMRT) with the vision to find alternative QA methods. Material and Methods: Twenty IMRT patients (12 of brain and 8 of the prostate) were planned on Eclipse treatment planning system using Varian Clinac DHX on both energies 6MV and 15MV. Verification plans of all such patients were also made and delivered to Map check2, EPID (Electronic portal imaging device) and Gafchromic EBT3. Gamma index analyses were performed using different criteria to evaluate and compare the dosimetric results. Results: Statistical analysis shows the passing rate of 99.55%, 97.23% and 92.9% for 6MV and 99.53%, 98.3% and 94.85% for 15 MV energy using a criteria of ±5% of 3mm, ±3% of 3mm and ±3% of 2mm respectively for brain, whereas using ±5% of 3mm and ±3% of 3mm gamma evaluation criteria, the passing rate is 94.55% and 90.45% for 6MV and 95.25%9 and 95% for 15 MV energy for the case of prostate using EBT3 film. Map check 2 results shows the passing rates of 98.17%, 97.68% and 86.78% for 6MV energy and 94.87%,97.46% and 88.31% for 15 MV energy respectively for brain using a criteria of ±5% of 3mm, ±3% of 3mm and ±3% of 2mm, whereas using ±5% of 3mm and ±3% of 3mm gamma evaluation criteria gives the passing rate of 97.7% and 96.4% for 6MV and 98.75%9 and 98.05% for 15 MV energy for the case of prostate. EPID 6 MV and gamma analysis shows the passing rate of 99.56%, 98.63% and 98.4% for the brain, 100% and 99.9% for prostate using the same criteria as for map check 2 and EBT 3 film. Conclusion: The results demonstrate excellent passing rates were obtained for all dosimeter when compared with the planar dose distributions for 6 MV IMRT fields as well as for 15 MV. EPID results are better than EBT3 films and map check 2 because it is likely that part of this difference is real, and part is due to manhandling and different treatment set up verification which contributes dose distribution difference. Overall all three dosimeter exhibits results within limits according to AAPM report.120.Keywords: gafchromic EBT3, radiochromic film dosimetry, IMRT verification, EPID
Procedia PDF Downloads 421423 Numerical Methodology to Support the Development of a Double Chamber Syringe
Authors: Lourenço Bastos, Filipa Carneiro, Bruno Vale, Rita Marques Joana Silva, Ricardo Freitas, Ângelo Marques, Sara Cortez, Alberta Coelho, Pedro Parreira, Liliana Sousa, Anabela Salgueiro, Bruno Silva
Abstract:
The process of flushing is considered to be an adequate technique to reduce the risk of infection during the clinical practice of venous catheterization. Nonetheless, there is still a lack of adhesion to this method, in part due to the complexity of this procedure. The project SeringaDuo aimed to develop an innovative double-chamber syringe for intravenous sequential administration of drugs and serums. This device served the purpose of improving the adherence to the practice, through the reduction of manipulations needed, which also improves patient safety, and though the promotion of flushing practice by health professionals, by simplifying this task. To assist on the development of this innovative syringe, a numerical methodology was developed and validated in order to predict the syringe’s mechanical and flow behavior during the fluids’ loading and administration phases, as well as to allow the material behavior evaluation during its production. For this, three commercial numerical simulation software was used, namely ABAQUS, ANSYS/FLUENT, and MOLDFLOW. This methodology aimed to evaluate the concepts feasibility and to optimize the geometries of the syringe’s components, creating this way an iterative process for product development based on numerical simulations, validated by the production of prototypes. Through this methodology, it was possible to achieve a final design that fulfils all the characteristics and specifications defined. This iterative process based on numerical simulations is a powerful tool for product development that allows obtaining fast and accurate results without the strict need for prototypes. An iterative process can be implemented, consisting of consecutive constructions and evaluations of new concepts, to obtain an optimized solution, which fulfils all the predefined specifications and requirements.Keywords: Venous catheterization, flushing, syringe, numerical simulation
Procedia PDF Downloads 167422 Formulation of Hybrid Nanopowder-Molecular Ink for Fabricating Critical Material-Free Cu₂ZnSnS₄ Thin Film Solar Absorber
Authors: Anies Mutiari, Neha Bansal, Martin Artner, Veronika Mayer, Juergen Roth, Mathias Weil, Rachmat Adhi Wibowo
Abstract:
Cu₂ZnSnS₄ (CZTS) compound (mineral name kesterite) has attracted considerable interests for photovoltaic application owing to its optoelectrical properties. Moreover, its elemental abundance in Earth’s crust offers a comparative advantage for envisaged large-scale photovoltaic deployment without any material shortage issues. In this contribution, we present an innovative route to prepare CZTS solar absorber layer for photovoltaic application from low-cost and up-scalable process. CZTS layers were spin coated on the Molybdenum-coated glass from two inks composed of different solvents; dimethylsulfoxide (DMSO) and ultrapure water. Into each solvent; 0.57M CuCl₂, 0.39M ZnCl₂, 0.53M SnCl₂, and 1.85M Thiourea or Na₂S₂O₃, as well as pre-synthesized CZTS nanopowder, were added as sources of Cu, Zn, Sn and S in the ink. The crystallisation of ink into CZTS dense layers was carried out by firstly annealing the as-deposited CZTS layer in open air at 300°C for 1 minute, followed by sulfurisation at 560–620°C under atmospheric pressure for 120 minutes. Complementary electron microscopy, grazing incidence X-ray diffraction and Raman spectroscopy investigations suggest that both solvents can be used for preparing high quality and device relevant CZTS solar absorber layers. The sulphurisation crystallizes the as-deposited CZTS into highly polycrystalline CZTS layer with tetragonal structure demonstrated by the presence of tetrahedrally-shaped grains with the size of 1 µm. An advancement of the CZTS layer preparation was made by gradual substitution of volatile organic compound solvent of DMSO with ultrapure water. It is revealed that by using similar air annealing and sulphurisation process, dense and compact CZTS layers can also be fabricated from an ink with reduced volatile organic compound content.Keywords: kesterite, solar ink, spin coating, photovoltaics
Procedia PDF Downloads 171421 Health Literacy: Collaboration between Clinician and Patient
Authors: Cathy Basterfield
Abstract:
Issue: To engage in one’s own health care, health professionals need to be aware of an individual’s specific skills and abilities for best communication. One of the most discussed is health literacy. One of the assumed skills and abilities for adults is an individuals’ health literacy. Background: A review of publicly available health content appears to assume all adult readers will have a broad and full capacity to read at a high level of literacy, often at a post-school education level. Health information writers and clinicians need to recognise one critical area for why there may be little or no change in a person’s behaviour, or no-shows to appointments. Perhaps unintentionally, they are miscommunicating with the majority of the adult population. Health information contains many literacy domains. It usually includes technical medical terms or jargon. Many fact sheets and other information require scientific literacy with or without specific numerical literacy. It may include graphs, percentages, timing, distance, or weights. Each additional word or concept in these domains decreases the readers' ability to meaningfully read, understand and know what to do with the information. An attempt to begin to read the heading where long or unfamiliar words are used will reduce the readers' motivation to attempt to read. Critically people who have low literacy are overwhelmed when pages are covered with lots of words. People attending a health environment may be unwell or anxious about a diagnosis. These make it harder to read, understand and know what to do with the information. But access to health information must consider an even wider range of adults, including those with poor school attainment, migrants, and refugees. It is also homeless people, people with mental health illnesses, or people who are ageing. People with low literacy also may include people with lifelong disabilities, people with acquired disabilities, people who read English as a second (or third) language, people who are Deaf, or people who are vision impaired. Outcome: This paper will discuss Easy English, which is developed for adults. It uses the audiences’ everyday words, short sentences, short words, and no jargon. It uses concrete language and concrete, specific images to support the text. It has been developed in Australia since the mid-2000s. This paper will showcase various projects in the health domain which use Easy English to improve the understanding and functional use of written information for the large numbers of adults in our communities who do not have the health literacy to manage a range of day to day reading tasks. See examples from consent forms, fact sheets and choice options, instructions, and other functional documents, where Easy English has been developed. This paper will ask individuals to reflect on their own work practice and consider what written information must be available in Easy English. It does not matter how cutting-edge a new treatment is; when adults can not read or understand what it is about and the positive and negative outcomes, they are less likely to be engaged in their own health journey.Keywords: health literacy, inclusion, Easy English, communication
Procedia PDF Downloads 126420 BFDD-S: Big Data Framework to Detect and Mitigate DDoS Attack in SDN Network
Authors: Amirreza Fazely Hamedani, Muzzamil Aziz, Philipp Wieder, Ramin Yahyapour
Abstract:
Software-defined networking in recent years came into the sight of so many network designers as a successor to the traditional networking. Unlike traditional networks where control and data planes engage together within a single device in the network infrastructure such as switches and routers, the two planes are kept separated in software-defined networks (SDNs). All critical decisions about packet routing are made on the network controller, and the data level devices forward the packets based on these decisions. This type of network is vulnerable to DDoS attacks, degrading the overall functioning and performance of the network by continuously injecting the fake flows into it. This increases substantial burden on the controller side, and the result ultimately leads to the inaccessibility of the controller and the lack of network service to the legitimate users. Thus, the protection of this novel network architecture against denial of service attacks is essential. In the world of cybersecurity, attacks and new threats emerge every day. It is essential to have tools capable of managing and analyzing all this new information to detect possible attacks in real-time. These tools should provide a comprehensive solution to automatically detect, predict and prevent abnormalities in the network. Big data encompasses a wide range of studies, but it mainly refers to the massive amounts of structured and unstructured data that organizations deal with on a regular basis. On the other hand, it regards not only the volume of the data; but also that how data-driven information can be used to enhance decision-making processes, security, and the overall efficiency of a business. This paper presents an intelligent big data framework as a solution to handle illegitimate traffic burden on the SDN network created by the numerous DDoS attacks. The framework entails an efficient defence and monitoring mechanism against DDoS attacks by employing the state of the art machine learning techniques.Keywords: apache spark, apache kafka, big data, DDoS attack, machine learning, SDN network
Procedia PDF Downloads 169419 Optimizing the Design Parameters of Acoustic Power Transfer Model to Achieve High Power Intensity and Compact System
Authors: Ariba Siddiqui, Amber Khan
Abstract:
The need for bio-implantable devices in the field of medical sciences has been increasing day by day; however, the charging of these devices is a major issue. Batteries, a very common method of powering the implants, have a limited lifetime and bulky nature. Therefore, as a replacement of batteries, acoustic power transfer (APT) technology is being accepted as the most suitable technique to wirelessly power the medical implants in the present scenario. The basic model of APT consists of piezoelectric transducers that work on the principle of converse piezoelectric effect at the transmitting end and direct piezoelectric effect at the receiving end. This paper provides mechanistic insight into the parameters affecting the design and efficient working of acoustic power transfer systems. The optimum design considerations have been presented that will help to compress the size of the device and augment the intensity of the pressure wave. A COMSOL model of the PZT (Lead Zirconate Titanate) transducer was developed. The model was simulated and analyzed on a frequency spectrum. The simulation results displayed that the efficiency of these devices is strongly dependent on the frequency of operation, and a wrong choice of the operating frequency leads to the high absorption of acoustic field inside the tissue (medium), poor power strength, and heavy transducers, which in effect influence the overall configuration of the acoustic systems. Considering all the tradeoffs, the simulations were performed again by determining an optimum frequency (900 kHz) that resulted in the reduction of the transducer's thickness to 1.96 mm and augmented the power strength with an intensity of 432 W/m². Thus, the results obtained after the second simulation contribute to lesser attenuation, lightweight systems, high power intensity, and also comply with safety limits provided by the U.S Food and Drug Administration (FDA). It was also found that the chosen operating frequency enhances the directivity of the acoustic wave at the receiver side.Keywords: acoustic power, bio-implantable, COMSOL, Lead Zirconate Titanate, piezoelectric, transducer
Procedia PDF Downloads 174418 Reactors with Effective Mixing as a Solutions for Micro-Biogas Plant
Authors: M. Zielinski, M. Debowski, P. Rusanowska, A. Glowacka-Gil, M. Zielinska, A. Cydzik-Kwiatkowska, J. Kazimierowicz
Abstract:
Technologies for the micro-biogas plant with heating and mixing systems are presented as a part of the Research Coordination for a Low-Cost Biomethane Production at Small and Medium Scale Applications (Record Biomap). The main objective of the Record Biomap project is to build a network of operators and scientific institutions interested in cooperation and the development of promising technologies in the sector of small and medium-sized biogas plants. The activities carried out in the project will bridge the gap between research and market and reduce the time of implementation of new, efficient technological and technical solutions. Reactor with simultaneously mixing and heating system is a concrete tank with a rectangular cross-section. In the reactor, heating is integrated with the mixing of substrate and anaerobic sludge. This reactor is solution dedicated for substrates with high solids content, which cannot be introduced to the reactor with pumps, even with positive displacement pumps. Substrates are poured to the reactor and then with a screw pump, they are mixed with anaerobic sludge. The pumped sludge, flowing through the screw pump, is simultaneously heated by a heat exchanger. The level of the fermentation sludge inside the reactor chamber is above the bottom edge of the cover. Cover of the reactor is equipped with the screw pump driver. Inside the reactor, an electric motor is installed that is driving a screw pump. The heated sludge circulates in the digester. The post-fermented sludge is collected using a drain well. The inlet to the drain well is below the level of the sludge in the digester. The biogas is discharged from the reactor by the biogas intake valve located on the cover. The technology is very useful for fermentation of lignocellulosic biomass and substrates with high content of dry mass (organic wastes). The other technology is a reactor for micro-biogas plant with a pressure mixing system. The reactor has a form of plastic or concrete tank with a circular cross-section. The effective mixing of sludge is ensured by profiled at 90° bottom of the tank. Substrates for fermentation are supplied by an inlet well. The inlet well is equipped with a cover that eliminates odour release. The introduction of a new portion of substrates is preceded by pumping of digestate to the disposal well. Optionally, digestate can gravitationally flow to digestate storage tank. The obtained biogas is discharged into the separator. The valve supplies biogas to the blower. The blower presses the biogas from the fermentation chamber in such a way as to facilitate the introduction of a new portion of substrates. Biogas is discharged from the reactor by valve that enables biogas removal but prevents suction from outside the reactor.Keywords: biogas, digestion, heating system, mixing system
Procedia PDF Downloads 154417 A Review on Development of Pedicle Screws and Characterization of Biomaterials for Fixation in Lumbar Spine
Authors: Shri Dubey, Jamal Ghorieshi
Abstract:
Instability of the lumbar spine is caused by various factors that include degenerative disc, herniated disc, traumatic injuries, and other disorders. Pedicle screws are widely used as a main fixation device to construct rigid linkages of vertebrae to provide a fully functional and stable spine. Various technologies and methods have been used to restore the stabilization. However, loosening of pedicle screws is the main cause of concerns for neurosurgeons. This could happen due to poor bone quality with osteoporosis as well as types of pedicle screw used. Compatibilities and stabilities of pedicle screws with bone depend on design (thread design, length, and diameter) and material. Grip length and pullout strength affect the motion and stability of the spine when it goes through different phases such as extension, flexion, and rotation. Pullout strength of augmented pedicle screws is increased in both primary and salvage procedures by 119% (p = 0.001) and 162% (p = 0.01), respectively. Self-centering pedicle screws at different trajectories (0°, 10°, 20°, and 30°) show the same pullout strength as insertion in a straight-ahead trajectory. The outer cylindrical and inner conical shape of pedicle screws show the highest pullout strength in Grades 5 and 15 foams (synthetic bone). An outer cylindrical and inner conical shape with a V-shape thread exhibit the highest pullout strength in all foam grades. The maximum observed pullout strength is at axial pullout configuration at 0°. For Grade 15 (240 kg/m³) foam, there is a decline in pull out strength. The largest decrease in pullout strength is reported for Grade 10 (160 kg/m³) foam. The maximum pullout strength of 2176 N (0.32-g/cm³ Sawbones) on all densities. Type 1 Pedicle screw shows the best fixation due to smaller conical core diameter and smaller thread pitch (Screw 2 with 2 mm; Screws 1 and 3 with 3 mm).Keywords: polymethylmethacrylate, PMMA, classical pedicle screws, CPS, expandable poly-ether-ether-ketone shell, EPEEKS, includes translaminar facet screw, TLFS, poly-ether-ether-ketone, PEEK, transfacetopedicular screw, TFPS
Procedia PDF Downloads 155416 Real Time Detection of Application Layer DDos Attack Using Log Based Collaborative Intrusion Detection System
Authors: Farheen Tabassum, Shoab Ahmed Khan
Abstract:
The brutality of attacks on networks and decisive infrastructures are on the climb over recent years and appears to continue to do so. Distributed Denial of service attack is the most prevalent and easy attack on the availability of a service due to the easy availability of large botnet computers at cheap price and the general lack of protection against these attacks. Application layer DDoS attack is DDoS attack that is targeted on wed server, application server or database server. These types of attacks are much more sophisticated and challenging as they get around most conventional network security devices because attack traffic often impersonate normal traffic and cannot be recognized by network layer anomalies. Conventional techniques of single-hosted security systems are becoming gradually less effective in the face of such complicated and synchronized multi-front attacks. In order to protect from such attacks and intrusion, corporation among all network devices is essential. To overcome this issue, a collaborative intrusion detection system (CIDS) is proposed in which multiple network devices share valuable information to identify attacks, as a single device might not be capable to sense any malevolent action on its own. So it helps us to take decision after analyzing the information collected from different sources. This novel attack detection technique helps to detect seemingly benign packets that target the availability of the critical infrastructure, and the proposed solution methodology shall enable the incident response teams to detect and react to DDoS attacks at the earliest stage to ensure that the uptime of the service remain unaffected. Experimental evaluation shows that the proposed collaborative detection approach is much more effective and efficient than the previous approaches.Keywords: Distributed Denial-of-Service (DDoS), Collaborative Intrusion Detection System (CIDS), Slowloris, OSSIM (Open Source Security Information Management tool), OSSEC HIDS
Procedia PDF Downloads 354415 Nickel Removal from Industrial Wastewater by Eucalyptus Leaves and Poplar Ashes
Authors: Negin Bayat, Nahid HasanZadeh
Abstract:
Effluents of different industries such as metalworking, battery industry, mining, including heavy metal are considered problematic issues for both humans and the environment. These heavy metals include cadmium, copper, zinc, nickel, chromium, cyanide, lead, etc. Different physicochemical and biological methods are used to remove heavy metals, such as sedimentation, coagulation, flotation, chemical precipitation, filtration, membrane processes (reverse osmosis and nanofiltration), ion exchange, biological methods, adsorption with activated carbon, etc. These methods are generally either expensive or ineffective. In recent years, considerable attention has been given to the removal of heavy metal ions from solution by absorption using discarded and low-cost materials. In this study, nickel removal using an adsorption process by eucalyptus powdered leaves and poplar ash was investigated. This is an applied study. The effect of various parameters on metal removal, such as pH, amount of adsorbent, contact time, and stirring speed, was studied using a discontinuous method. This research was conducted in aqueous solutions on the laboratory scale. Then, optimum absorption conditions were obtained. Then, the study was conducted on real wastewater samples. In addition, the nickel concentration in the wastewater before and after the absorption process was measured. In all experiments, the remaining nickel was measured using an atomic absorption spectrometry device at 382 nm wavelength after an appropriate time and filtration. The results showed that increasing both adsorbent and pH parameters increase the metal removal rate. Nickel removal increased at the first 60 minutes. Then, the absorption rate remained constant and reached equilibrium. A desired removal rate with 40 mg in 100 ml adsorbent solution at pH = 9.5 was observed. According to the obtained results, the best absorption rate was observed at 40 mg dose using a combination of eucalyptus leaves and poplar ash in this study, which was equal to 99.76%. Thus, this combined method can be used as an inexpensive and effective absorbent for the removal of nickel from aqueous solutions.Keywords: absorption, wastewater, nickel, poplar ash, eucalyptus leaf, treatment
Procedia PDF Downloads 19414 Determination of Rare Earth Element Patterns in Uranium Matrix for Nuclear Forensics Application: Method Development for Inductively Coupled Plasma Mass Spectrometry (ICP-MS) Measurements
Authors: Bernadett Henn, Katalin Tálos, Éva Kováss Széles
Abstract:
During the last 50 years, the worldwide permeation of the nuclear techniques induces several new problems in the environmental and in the human life. Nowadays, due to the increasing of the risk of terrorism worldwide, the potential occurrence of terrorist attacks using also weapon of mass destruction containing radioactive or nuclear materials as e.g. dirty bombs, is a real threat. For instance, the uranium pellets are one of the potential nuclear materials which are suitable for making special weapons. The nuclear forensics mainly focuses on the determination of the origin of the confiscated or found nuclear and other radioactive materials, which could be used for making any radioactive dispersive device. One of the most important signatures in nuclear forensics to find the origin of the material is the determination of the rare earth element patterns (REE) in the seized or found radioactive or nuclear samples. The concentration and the normalized pattern of the REE can be used as an evidence of uranium origin. The REE are the fourteen Lanthanides in addition scandium and yttrium what are mostly found together and really low concentration in uranium pellets. The problems of the REE determination using ICP-MS technique are the uranium matrix (high concentration of uranium) and the interferences among Lanthanides. In this work, our aim was to develop an effective chemical sample preparation process using extraction chromatography for separation the uranium matrix and the rare earth elements from each other following some publications can be found in the literature and modified them. Secondly, our purpose was the optimization of the ICP-MS measuring process for REE concentration. During method development, in the first step, a REE model solution was used in two different types of extraction chromatographic resins (LN® and TRU®) and different acidic media for environmental testing the Lanthanides separation. Uranium matrix was added to the model solution and was proved in the same conditions. Methods were tested and validated using REE UOC (uranium ore concentrate) reference materials. Samples were analyzed by sector field mass spectrometer (ICP-SFMS).Keywords: extraction chromatography, nuclear forensics, rare earth elements, uranium
Procedia PDF Downloads 308413 Fructose-Aided Cross-Linked Enzyme Aggregates of Laccase: An Insight on Its Chemical and Physical Properties
Authors: Bipasa Dey, Varsha Panwar, Tanmay Dutta
Abstract:
Laccase, a multicopper oxidase (EC 1.10.3.2) have been at the forefront as a superior industrial biocatalyst. They are versatile in terms of bestowing sustainable and ecological catalytic reactions such as polymerisation, xenobiotic degradation and bioremediation of phenolic and non-phenolic compounds. Regardless of the wide biotechnological applications, the critical limiting factors viz. reusability, retrieval, and storage stability still prevail. This can cause an impediment in their applicability. Crosslinked enzyme aggregates (CLEAs) have emerged as a promising technique that rehabilitates these essential facets, albeit at the expense of their enzymatic activity. The carrier free crosslinking method prevails over the carrier-bound immobilisation in conferring high productivity, low production cost owing to the absence of additional carrier and circumvent any non-catalytic ballast which could dilute the volumetric activity. To the best of our knowledge, the ε-amino group of lysyl residue is speculated as the best choice for forming Schiff’s base with glutaraldehyde. Despite being most preferrable, excess glutaraldehyde can bring about disproportionate and undesirable crosslinking within the catalytic site and hence could deliver undesirable catalytic losses. Moreover, the surface distribution of lysine residues in Trametes versicolor laccase is significantly less. Thus, to mitigate the adverse effect of glutaraldehyde in conjunction with scaling down the degradation or catalytic loss of the enzyme, crosslinking with inert substances like gelatine, collagen, Bovine serum albumin (BSA) or excess lysine is practiced. Analogous to these molecules, sugars have been well known as a protein stabiliser. It helps to retain the structural integrity, specifically secondary structure of the protein during aggregation by changing the solvent properties. They are comprehended to avert protein denaturation or enzyme deactivation during precipitation. We prepared crosslinked enzyme aggregates (CLEAs) of laccase from T. versicolor with the aid of sugars. The sugar CLEAs were compared with the classic BSA and glutaraldehyde laccase CLEAs concerning physico-chemical properties. The activity recovery for the fructose CLEAs were found to be ~20% higher than the non-sugar CLEA. Moreover, the 𝐾𝑐𝑎𝑡𝐾𝑚⁄ values of the CLEAs were two and three-fold higher than BSA-CLEA and GACLEA, respectively. The half-life (t1/2) deciphered by sugar-CLEA was higher than the t1/2 of GA-CLEAs and free enzyme, portraying more thermal stability. Besides, it demonstrated extraordinarily high pH stability, which was analogous to BSA-CLEA. The promising attributes of increased storage stability and recyclability (>80%) gives more edge to the sugar-CLEAs over conventional CLEAs of their corresponding free enzyme. Thus, sugar-CLEA prevails in furnishing the rudimentary properties required for a biocatalyst and holds many prospects.Keywords: cross-linked enzyme aggregates, laccase immobilization, enzyme reusability, enzyme stability
Procedia PDF Downloads 102412 Methodology for the Integration of Object Identification Processes in Handling and Logistic Systems
Authors: L. Kiefer, C. Richter, G. Reinhart
Abstract:
The uprising complexity in production systems due to an increasing amount of variants up to customer innovated products leads to requirements that hierarchical control systems are not able to fulfil. Therefore, factory planners can install autonomous manufacturing systems. The fundamental requirement for an autonomous control is the identification of objects within production systems. In this approach an attribute-based identification is focused for avoiding dose-dependent identification costs. Instead of using an identification mark (ID) like a radio frequency identification (RFID)-Tag, an object type is directly identified by its attributes. To facilitate that it’s recommended to include the identification and the corresponding sensors within handling processes, which connect all manufacturing processes and therefore ensure a high identification rate and reduce blind spots. The presented methodology reduces the individual effort to integrate identification processes in handling systems. First, suitable object attributes and sensor systems for object identification in a production environment are defined. By categorising these sensor systems as well as handling systems, it is possible to match them universal within a compatibility matrix. Based on that compatibility further requirements like identification time are analysed, which decide whether the combination of handling and sensor system is well suited for parallel handling and identification within an autonomous control. By analysing a list of more than thousand possible attributes, first investigations have shown, that five main characteristics (weight, form, colour, amount, and position of subattributes as drillings) are sufficient for an integrable identification. This knowledge limits the variety of identification systems and leads to a manageable complexity within the selection process. Besides the procedure, several tools, as an example a sensor pool are presented. These tools include the generated specific expert knowledge and simplify the selection. The primary tool is a pool of preconfigured identification processes depending on the chosen combination of sensor and handling device. By following the defined procedure and using the created tools, even laypeople out of other scientific fields can choose an appropriate combination of handling devices and sensors which enable parallel handling and identification.Keywords: agent systems, autonomous control, handling systems, identification
Procedia PDF Downloads 177411 Practical Skill Education for Doctors in Training: Economical and Efficient Methods for Students to Receive Hands-on Experience
Authors: Nathaniel Deboever, Malcolm Breeze, Adrian Sheen
Abstract:
Basic surgical and suturing techniques are a fundamental requirement for all doctors. In order to gain confidence and competence, doctors in training need to obtain sufficient teaching and just as importantly: practice. Young doctors with an apt level of expertise on these simple surgical skills, which are often used in the Emergency Department, can help alleviate some pressure during a busy evening. Unfortunately, learning these skills can be quite difficult during medical school or even during junior doctor years. The aim of this project was to adequately train medical students attending University of Sydney’s Nepean Clinical School through a series of workshops highlighting practical skills, with hopes to further extend this program to junior doctors in the hospital. The sessions instructed basic skills via tutorials, demonstrations, and lastly, the sessions cemented these proficiencies with practical sessions. During such an endeavor, it is fundamental to employ models that appropriately resemble what students will encounter in the clinical setting. The sustainability of workshops is similarly important to the continuity of such a program. To address both these challenges, the authors have developed models including suturing platforms, knot tying, and vessel ligation stations, as well as a shave and punch biopsy models and ophthalmologic foreign body device. The unique aspect of this work is that we utilized hands-on teaching sessions, to address a gap in doctors-in-training and junior doctor curriculum. Presented to you through this poster are our approaches to creating models that do not employ animal products and therefore do not necessitate particular facilities or discarding requirements. Covering numerous skills that would be beneficial to all young doctors, these models are easily replicable and affordable. This exciting work allows for countless sessions at low cost, providing enough practice for students to perform these skills confidently as it has been shown through attendee questionnaires.Keywords: medical education, surgical models, surgical simulation, surgical skills education
Procedia PDF Downloads 157410 Dust Particle Removal from Air in a Self-Priming Submerged Venturi Scrubber
Authors: Manisha Bal, Remya Chinnamma Jose, B.C. Meikap
Abstract:
Dust particles suspended in air are a major source of air pollution. A self-priming submerged venturi scrubber proven very effective in cases of handling nuclear power plant accidents is an efficient device to remove dust particles from the air and thus aids in pollution control. Venturi scrubbers are compact, have a simple mode of operation, no moving parts, easy to install and maintain when compared to other pollution control devices and can handle high temperatures and corrosive and flammable gases and dust particles. In the present paper, fly ash particles recognized as a high air pollutant substance emitted mostly from thermal power plants is considered as the dust particle. Its exposure through skin contact, inhalation and indigestion can lead to health risks and in severe cases can even root to lung cancer. The main focus of this study is on the removal of fly ash particles from polluted air using a self-priming venturi scrubber in submerged conditions using water as the scrubbing liquid. The venturi scrubber comprising of three sections: converging section, throat and diverging section is submerged inside a water tank. The liquid enters the throat due to the pressure difference composed of the hydrostatic pressure of the liquid and static pressure of the gas. The high velocity dust particles atomize the liquid droplets at the throat and this interaction leads to its absorption into water and thus removal of fly ash from the air. Detailed investigation on the scrubbing of fly ash has been done in this literature. Experiments were conducted at different throat gas velocities, water levels and fly ash inlet concentrations to study the fly ash removal efficiency. From the experimental results, the highest fly ash removal efficiency of 99.78% is achieved at the throat gas velocity of 58 m/s, water level of height 0.77m with fly ash inlet concentration of 0.3 x10⁻³ kg/Nm³ in the submerged condition. The effect of throat gas velocity, water level and fly ash inlet concentration on the removal efficiency has also been evaluated. Furthermore, experimental results of removal efficiency are validated with the developed empirical model.Keywords: dust particles, fly ash, pollution control, self-priming venturi scrubber
Procedia PDF Downloads 164409 Micro-Droplet Formation in a Microchannel under the Effect of an Electric Field: Experiment
Authors: Sercan Altundemir, Pinar Eribol, A. Kerem Uguz
Abstract:
Microfluidics systems allow many-large scale laboratory applications to be miniaturized on a single device in order to reduce cost and advance fluid control. Moreover, such systems enable to generate and control droplets which have a significant role on improved analysis for many chemical and biological applications. For example, they can be employed as the model for cells in microfluidic systems. In this work, the interfacial instability of two immiscible Newtonian liquids flowing in a microchannel is investigated. When two immiscible liquids are in laminar regime, a flat interface is formed between them. If a direct current electric field is applied, the interface may deform, i.e. may become unstable and it may be ruptured and form micro-droplets. First, the effect of thickness ratio, total flow rate, viscosity ratio of the silicone oil and ethylene glycol liquid couple on the critical voltage at which the interface starts to destabilize is investigated. Then the droplet sizes are measured under the effect of these parameters at various voltages. Moreover, the effect of total flow rate on the time elapsed for the interface to be ruptured to form droplets by hitting the wall of the channel is analyzed. It is observed that an increase in the viscosity or the thickness ratio of the silicone oil to the ethylene glycol has a stabilizing effect, i.e. a higher voltage is needed while the total flow rate has no effect on it. However, it is observed that an increase in the total flow rate results in shortening of the elapsed time for the interface to hit the wall. Moreover, the droplet size decreases down to 0.1 μL with an increase in the applied voltage, the viscosity ratio or the total flow rate or a decrease in the thickness ratio. In addition to these observations, two empirical models for determining the critical electric number, i.e., the dimensionless voltage and the droplet size and another model which is a combination of both models, for determining the droplet size at the critical voltage are established.Keywords: droplet formation, electrohydrodynamics, microfluidics, two-phase flow
Procedia PDF Downloads 176408 Assessing Overall Thermal Conductance Value of Low-Rise Residential Home Exterior Above-Grade Walls Using Infrared Thermography Methods
Authors: Matthew D. Baffa
Abstract:
Infrared thermography is a non-destructive test method used to estimate surface temperatures based on the amount of electromagnetic energy radiated by building envelope components. These surface temperatures are indicators of various qualitative building envelope deficiencies such as locations and extent of heat loss, thermal bridging, damaged or missing thermal insulation, air leakage, and moisture presence in roof, floor, and wall assemblies. Although infrared thermography is commonly used for qualitative deficiency detection in buildings, this study assesses its use as a quantitative method to estimate the overall thermal conductance value (U-value) of the exterior above-grade walls of a study home. The overall U-value of exterior above-grade walls in a home provides useful insight into the energy consumption and thermal comfort of a home. Three methodologies from the literature were employed to estimate the overall U-value by equating conductive heat loss through the exterior above-grade walls to the sum of convective and radiant heat losses of the walls. Outdoor infrared thermography field measurements of the exterior above-grade wall surface and reflective temperatures and emissivity values for various components of the exterior above-grade wall assemblies were carried out during winter months at the study home using a basic thermal imager device. The overall U-values estimated from each methodology from the literature using the recorded field measurements were compared to the nominal exterior above-grade wall overall U-value calculated from materials and dimensions detailed in architectural drawings of the study home. The nominal overall U-value was validated through calendarization and weather normalization of utility bills for the study home as well as various estimated heat loss quantities from a HOT2000 computer model of the study home and other methods. Under ideal environmental conditions, the estimated overall U-values deviated from the nominal overall U-value between ±2% to ±33%. This study suggests infrared thermography can estimate the overall U-value of exterior above-grade walls in low-rise residential homes with a fair amount of accuracy.Keywords: emissivity, heat loss, infrared thermography, thermal conductance
Procedia PDF Downloads 313407 Magnetic Resonance Imaging in Cochlear Implant Patients without Magnet Removal: A Safe and Effective Workflow Management Program
Authors: Yunhe Chen, Xinyun Liu, Qian Wang, Jianan Li
Abstract:
Background Cochlear implants (CIs) are currently the primary effective treatment for severe or profound sensorineural hearing loss. As China's population ages and the number of young children rises, the demand for MRI for CI patients is expected to increase. Methods Reviewed MRI cases of 25 CI patients between 2015 and 2024, assessed imaging auditory outcomes and adverse reactions. Use the adverse event record sheet and accompanying medication sheet to record follow-up measures. Results Most CI patients undergoing MRI may face risks such as artifacts, pain, redness, swelling, tissue damage, bleeding, and magnet displacement or demagnetization. Twenty-five CI patients in our hospital were reviewed. Seven patient underwent 3.0 T MR, the others underwent 1.5 T MR. The manufacturers are 18 cases in Austria, 5 cases in Australia and 2 cases in Nurotron. Among them, one patient with bilateral CI underwent 1.5 T MR examination after head pressure bandaging, and the left magnet was displaced (CI24RE Series, Australia). This patient underwent surgical replacement of the magnet under general anesthesia. Six days after the operation, the patient's feedback indicated that the performance of the cochlear implant was consistent with the previous results following the reactivation of the external device. Based on the experience of our hospital, we proposed the feasible management scheme of MRI examination procedure for CI patients. This plan should include a module for confirming MRI imaging parameters, informed consent, educational materials for patients, and other safety measures to ensure that patients receive imaging results safely and effectively, implify clinical. Conclusion As indications for both MRI and cochlear implantation expand,the number of MRI studies recommended for patients with cochlear implants will also increase. The process and management scheme proposed in this study can help to obtain imaging results safely and effectively, and reduce clinical stress.Keywords: cochlear implantation, MRI, magnet, displacement
Procedia PDF Downloads 14406 Oral Versus Iontophoresis Nonsteroidal Anti-Inflammatory Drugs in Tennis Elbow
Authors: Moustafa Ali Elwan, Ibrahim Salem Abdelrafa, Ashraf Moharm
Abstract:
Nonsteroidal anti-inflammatory drugs (NSAIDs) are among the most commonly prescribed oral and topical drugs worldwide. Moreover, NSAIDs are responsible for most of all adverse drug reactions. For several decades, there are numerous attempts to use the cutaneous layers as a gate into the body for the local delivery of the therapeutic agent. Transdermal drug delivery is a validated technology contributing significantly to global pharmaceutical care. Transdermal Drug Delivery systems can be improved by using therapeutic agents. Moreover, Transdermal Drug Delivery systems can be improved by using chemical enhancers like ultrasound or iontophoresis. Iontophoresis provides a mechanism to enhance the penetration of hydrophilic and charged molecules across the skin. Objective: to compare the drug administration by ‘iontophoresis’ versus the oral rule. Methods: This study was conducted at the Faculty of Physical Therapy, Modern University for technology and information, Cairo, Egypt, on 20 participants (8 female & 12 male) who complained of tennis elbow. Their mean age was (25.45 ± 3.98) years, and all participants were assessed in many aspects: Pain threshold was assessed by algometer. Range of motion was assessed by electro goniometer, and isometric strength was assessed by a portable hand-held dynamometer. Then Participants were randomly assigned into two groups: group A was treated with oral NSAID (diclofenac) while group B was treated via administration of NSAIDs (diclofenac) via an iontophoresis device. All the participants were subjected to blood samples analysis in both pre-administration of the drug and post-administration of the drug for 24 hours (sample/every 6 hours). Results: The results demonstrated that there was a significant improvement in group b, “iontophoresis NSAIDs group,” more than in group B,” oral NSAIDs group,” in all measurements ‘ pain threshold, strength, and range of motion. Also, the iontophoresis method shows higher maximum plasma concentrations (Cmax) and concentration-time curves than the oral method.Keywords: diclofenac, iontophoresis, NSAIDs, oral, tennis elbow
Procedia PDF Downloads 115405 Cerebral Pulsatility Mediates the Link Between Physical Activity and Executive Functions in Older Adults with Cardiovascular Risk Factors: A Longitudinal NIRS Study
Authors: Hanieh Mohammadi, Sarah Fraser, Anil Nigam, Frederic Lesage, Louis Bherer
Abstract:
A chronically higher cerebral pulsatility is thought to damage cerebral microcirculation, leading to cognitive decline in older adults. Although it is widely known that regular physical activity is linked to improvement in some cognitive domains, including executive functions, the mediating role of cerebral pulsatility on this link remains to be elucidated. This study assessed the impact of 6 months of regular physical activity upon changes in an optical index of cerebral pulsatility and the role of physical activity for the improvement of executive functions. 27 older adults (aged 57-79, 66.7% women) with cardiovascular risk factors (CVRF) were enrolled in the study. The participants completed the behavioral Stroop test, which was extracted from the Delis-Kaplan executive functions system battery at baseline (T0) and after 6 months (T6) of physical activity. Near-infrared spectroscopy (NIRS) was applied for an innovative approach to indexing cerebral pulsatility in the brain microcirculation at T0 and T6. The participants were at standing rest while a NIRS device recorded hemodynamics data from frontal and motor cortex subregions at T0 and T6. The cerebral pulsatility index of interest was cerebral pulse amplitude, which was extracted from the pulsatile component of NIRS data. Our data indicated that 6 months of physical activity was associated with a reduction in the response time for the executive functions, including inhibition (T0: 56.33± 18.2 to T6: 53.33± 15.7,p= 0.038)and Switching(T0: 63.05± 5.68 to T6: 57.96 ±7.19,p< 0.001) conditions of the Stroop test. Also, physical activity was associated with a reduction in cerebral pulse amplitude (T0: 0.62± 0.05 to T6: 0.55± 0.08, p < 0.001). Notably, cerebral pulse amplitude was a significant mediator of the link between physical activity and response to the Stroop test for both inhibition (β=0.33 (0.61,0.23),p< 0.05)and switching (β=0.42 (0.69,0.11),p <0.01) conditions. This study suggests that regular physical activity may support cognitive functions through the improvement of cerebral pulsatility in older adults with CVRF.Keywords: near-infrared spectroscopy, cerebral pulsatility, physical activity, cardiovascular risk factors, executive functions
Procedia PDF Downloads 195404 From Text to Data: Sentiment Analysis of Presidential Election Political Forums
Authors: Sergio V Davalos, Alison L. Watkins
Abstract:
User generated content (UGC) such as website post has data associated with it: time of the post, gender, location, type of device, and number of words. The text entered in user generated content (UGC) can provide a valuable dimension for analysis. In this research, each user post is treated as a collection of terms (words). In addition to the number of words per post, the frequency of each term is determined by post and by the sum of occurrences in all posts. This research focuses on one specific aspect of UGC: sentiment. Sentiment analysis (SA) was applied to the content (user posts) of two sets of political forums related to the US presidential elections for 2012 and 2016. Sentiment analysis results in deriving data from the text. This enables the subsequent application of data analytic methods. The SASA (SAIL/SAI Sentiment Analyzer) model was used for sentiment analysis. The application of SASA resulted with a sentiment score for each post. Based on the sentiment scores for the posts there are significant differences between the content and sentiment of the two sets for the 2012 and 2016 presidential election forums. In the 2012 forums, 38% of the forums started with positive sentiment and 16% with negative sentiment. In the 2016 forums, 29% started with positive sentiment and 15% with negative sentiment. There also were changes in sentiment over time. For both elections as the election got closer, the cumulative sentiment score became negative. The candidate who won each election was in the more posts than the losing candidates. In the case of Trump, there were more negative posts than Clinton’s highest number of posts which were positive. KNIME topic modeling was used to derive topics from the posts. There were also changes in topics and keyword emphasis over time. Initially, the political parties were the most referenced and as the election got closer the emphasis changed to the candidates. The performance of the SASA method proved to predict sentiment better than four other methods in Sentibench. The research resulted in deriving sentiment data from text. In combination with other data, the sentiment data provided insight and discovery about user sentiment in the US presidential elections for 2012 and 2016.Keywords: sentiment analysis, text mining, user generated content, US presidential elections
Procedia PDF Downloads 192403 Wearable System for Prolonged Cooling and Dehumidifying of PPE in Hot Environments
Abstract:
While personal protective equipment (PPE) prevents the healthcare personnel from exposing to harmful surroundings, it creates a barrier to the dissipation of body heat and perspiration, leading to severe heat stress during prolonged exposure, especially in hot environments. It has been found that most of the existed personal cooling strategies have limitations in achieving effective cooling performance with long duration and lightweight. This work aimed to develop a lightweight (<1.0 kg) and less expensive wearable air cooling and dehumidifying system (WCDS) that can be applied underneath the protective clothing and provide 50W mean cooling power for more than 5 hours at 35°C environmental temperature without compromising the protection of PPE. For the WCDS, blowers will be used to activate an internal air circulation inside the clothing microclimate, which doesn't interfere with the protection of PPE. An air cooling and dehumidifying chamber (ACMR) with a specific design will be developed to reduce the air temperature and humidity inside the protective clothing. Then the cooled and dried air will be supplied to upper chest and back areas through a branching tubing system for personal cooling. A detachable ice cooling unit will be applied from the outside of the PPE to extract heat from the clothing microclimate. This combination allows for convenient replacement of the cooling unit to refresh the cooling effect, which can realize a continuous cooling function without taking off the PPE or adding too much weight. A preliminary thermal manikin test showed that the WCDS was able to reduce the microclimate temperature inside the PPE averagely by about 8°C for 60 minutes when the environmental temperature was 28.0 °C and 33.5 °C, respectively. Replacing the ice cooling unit every hour can maintain this cooling effect, while the longest operation duration is determined by the battery of the blowers, which can last for about 6 hours. This unique design is especially helpful for the PPE users, such as health care workers in infectious and hot environments when continuous cooling and dehumidifying are needed, but the change of protective clothing may increase the risk of infection. The new WCDS will not only improve the thermal comfort of PPE users but can also extend their safe working duration.Keywords: personal thermal management, heat stress, ppe, health care workers, wearable device
Procedia PDF Downloads 79402 On Cloud Computing: A Review of the Features
Authors: Assem Abdel Hamed Mousa
Abstract:
The Internet of Things probably already influences your life. And if it doesn’t, it soon will, say computer scientists; Ubiquitous computing names the third wave in computing, just now beginning. First were mainframes, each shared by lots of people. Now we are in the personal computing era, person and machine staring uneasily at each other across the desktop. Next comes ubiquitous computing, or the age of calm technology, when technology recedes into the background of our lives. Alan Kay of Apple calls this "Third Paradigm" computing. Ubiquitous computing is essentially the term for human interaction with computers in virtually everything. Ubiquitous computing is roughly the opposite of virtual reality. Where virtual reality puts people inside a computer-generated world, ubiquitous computing forces the computer to live out here in the world with people. Virtual reality is primarily a horse power problem; ubiquitous computing is a very difficult integration of human factors, computer science, engineering, and social sciences. The approach: Activate the world. Provide hundreds of wireless computing devices per person per office, of all scales (from 1" displays to wall sized). This has required new work in operating systems, user interfaces, networks, wireless, displays, and many other areas. We call our work "ubiquitous computing". This is different from PDA's, dynabooks, or information at your fingertips. It is invisible; everywhere computing that does not live on a personal device of any sort, but is in the woodwork everywhere. The initial incarnation of ubiquitous computing was in the form of "tabs", "pads", and "boards" built at Xerox PARC, 1988-1994. Several papers describe this work, and there are web pages for the Tabs and for the Boards (which are a commercial product now): Ubiquitous computing will drastically reduce the cost of digital devices and tasks for the average consumer. With labor intensive components such as processors and hard drives stored in the remote data centers powering the cloud , and with pooled resources giving individual consumers the benefits of economies of scale, monthly fees similar to a cable bill for services that feed into a consumer’s phone.Keywords: internet, cloud computing, ubiquitous computing, big data
Procedia PDF Downloads 382401 Postmortem Magnetic Resonance Imaging as an Objective Method for the Differential Diagnosis of a Stillborn and a Neonatal Death
Authors: Uliana N. Tumanova, Sergey M. Voevodin, Veronica A. Sinitsyna, Alexandr I. Shchegolev
Abstract:
An important part of forensic and autopsy research in perinatology is the answer to the question of life and stillbirth. Postmortem magnetic resonance imaging (MRI) is an objective non-invasive research method that allows to store data for a long time and not to exhume the body to clarify the diagnosis. The purpose of the research is to study the possibilities of a postmortem MRI to determine the stillbirth and death of a newborn who had spontaneous breathing and died on the first day after birth. MRI and morphological data of a study of 23 stillborn bodies, prenatally dead at a gestational age of 22-39 weeks (Group I) and the bodies of 16 newborns who died from 2 to 24 hours after birth (Group II) were compared. Before the autopsy, postmortem MRI was performed on the Siemens Magnetom Verio 3T device in the supine position of the body. The control group for MRI studies consisted of 7 live newborns without lung disease (Group III). On T2WI in the sagittal projection was measured MR-signal intensity (SI) in the lung tissue (L) and shoulder muscle (M). During the autopsy, a pulmonary swimming test was evaluated, and macro- and microscopic studies were performed. According to the postmortem MRI, the highest values of mean SI of the lung (430 ± 27.99) and of the muscle (405.5 ± 38.62) on T2WI were detected in group I and exceeded the corresponding value of group II by 2.7 times. The lowest values were found in the control group - 77.9 ± 12.34 and 119.7 ± 6.3, respectively. In the group II, the lung SI was 1.6 times higher than the muscle SI, whereas in the group I and in the control group, the muscle SI was 2.1 times and 1.8 times larger than the lung. On the basis of clinical and morphological data, we calculated the formula for determining the breathing index (BI) during postmortem MRI: BI = SIL x SIM / 100. The mean value of BI in the group I (1801.14 ± 241.6) (values ranged from 756 to 3744) significantly higher than the corresponding average value of BI in the group II (455.89 ± 137.32, p < 0.05) (305-638.4). In the control group, the mean BI value was 91.75 ± 13.3 (values ranged from 53 to 154). The BI with the results of pulmonary swimming tests and microscopic examination of the lungs were compared. The boundary value of BI for the differential diagnosis of stillborn and newborn death was 700. Using the postmortem MRI allows to differentiate the stillborn with the death of the breathing newborn.Keywords: lung, newborn, postmortem MRI, stillborn
Procedia PDF Downloads 128400 Integration of a Protective Film to Enhance the Longevity and Performance of Miniaturized Ion Sensors
Authors: Antonio Ruiz Gonzalez, Kwang-Leong Choy
Abstract:
The measurement of electrolytes has a high value in the clinical routine. Ions are present in all body fluids with variable concentrations and are involved in multiple pathologies such as heart failures and chronic kidney disease. In the case of dissolved potassium, although a high concentration in the blood (hyperkalemia) is relatively uncommon in the general population, it is one of the most frequent acute electrolyte abnormalities. In recent years, the integration of thin films technologies in this field has allowed the development of highly sensitive biosensors with ultra-low limits of detection for the assessment of metals in liquid samples. However, despite the current efforts in the miniaturization of sensitive devices and their integration into portable systems, only a limited number of successful examples used commercially can be found. This fact can be attributed to a high cost involved in their production and the sustained degradation of the electrodes over time, which causes a signal drift in the measurements. Thus, there is an unmet necessity for the development of low-cost and robust sensors for the real-time monitoring of analyte concentrations in patients to allow the early detection and diagnosis of diseases. This paper reports a thin film ion-selective sensor for the evaluation of potassium ions in aqueous samples. As an alternative for this fabrication method, aerosol assisted chemical vapor deposition (AACVD), was applied due to cost-effectivity and fine control over the film deposition. Such a technique does not require vacuum and is suitable for the coating of large surface areas and structures with complex geometries. This approach allowed the fabrication of highly homogeneous surfaces with well-defined microstructures onto 50 nm thin gold layers. The degradative processes of the ubiquitously employed poly (vinyl chloride) membranes in contact with an electrolyte solution were studied, including the polymer leaching process, mechanical desorption of nanoparticles and chemical degradation over time. Rational design of a protective coating based on an organosilicon material in combination with cellulose to improve the long-term stability of the sensors was then carried out, showing an improvement in the performance after 5 weeks. The antifouling properties of such coating were assessed using a cutting-edge quartz microbalance sensor, allowing the quantification of the adsorbed proteins in the nanogram range. A correlation between the microstructural properties of the films with the surface energy and biomolecules adhesion was then found and used to optimize the protective film.Keywords: hyperkalemia, drift, AACVD, organosilicon
Procedia PDF Downloads 123399 Electrochemical Biosensor Based on Chitosan-Gold Nanoparticles, Carbon Nanotubes for Detection of Ovarian Cancer Biomarker
Authors: Parvin Samadi Pakchin, Reza Saber, Hossein Ghanbari, Yadollah Omidi
Abstract:
Ovarian cancer is one of the leading cause of mortality among the gynecological malignancies, and it remains the one of the most prevalent cancer in females worldwide. Tumor markers are biochemical molecules in blood or tissues which can indicates cancers occurrence in the human body. So, the sensitive and specific detection of cancer markers typically recruited for diagnosing and evaluating cancers. Recently extensive research efforts are underway to achieve a simple, inexpensive and accurate device for detection of cancer biomarkers. Compared with conventional immunoassay techniques, electrochemical immunosensors are of great interest, because they are specific, simple, inexpensive, easy to handling and miniaturization. Moreover, in the past decade nanotechnology has played a crucial role in the development of biosensors. In this study, a signal-off electrochemical immunosensor for the detection of CA125 antigen has been developed using chitosan-gold nanoparticles (CS-AuNP) and multi-wall carbon nanotubes (MWCNT) composites. Toluidine blue (TB) is used as redox probe which is immobilized on the electrode surface. CS-AuNP is synthesized by a simple one step method that HAuCl4 is reduced by NH2 groups of chitosan. The CS-AuNP-MWCNT modified electrode has shown excellent electrochemical performance compared with bare Au electrode. MWCNTs and AuNPs increased electrochemical conductivity and accelerate electrons transfer between solution and electrode surface while excessive amine groups on chitosan lead to the effective loading of the biological material (CA125 antibody) and TB on the electrode surface. The electrochemical, immobilization and sensing properties CS-AuNP-MWCNT-TB modified electrodes are characterized by cyclic voltammetry, electrochemical impedance spectroscopy, differential pulse voltammetry and square wave voltammetry with Fe(CN)63−/4−as an electrochemical redox indicator.Keywords: signal-off electrochemical biosensor, CA125, ovarian cancer, chitosan-gold nanoparticles
Procedia PDF Downloads 290