Search results for: computer processing of large databases
12085 Rapid Processing Techniques Applied to Sintered Nickel Battery Technologies for Utility Scale Applications
Authors: J. D. Marinaccio, I. Mabbett, C. Glover, D. Worsley
Abstract:
Through use of novel modern/rapid processing techniques such as screen printing and Near-Infrared (NIR) radiative curing, process time for the sintering of sintered nickel plaques, applicable to alkaline nickel battery chemistries, has been drastically reduced from in excess of 200 minutes with conventional convection methods to below 2 minutes using NIR curing methods. Steps have also been taken to remove the need for forming gas as a reducing agent by implementing carbon as an in-situ reducing agent, within the ink formulation.Keywords: batteries, energy, iron, nickel, storage
Procedia PDF Downloads 44012084 Automatic Classification of Lung Diseases from CT Images
Authors: Abobaker Mohammed Qasem Farhan, Shangming Yang, Mohammed Al-Nehari
Abstract:
Pneumonia is a kind of lung disease that creates congestion in the chest. Such pneumonic conditions lead to loss of life of the severity of high congestion. Pneumonic lung disease is caused by viral pneumonia, bacterial pneumonia, or Covidi-19 induced pneumonia. The early prediction and classification of such lung diseases help to reduce the mortality rate. We propose the automatic Computer-Aided Diagnosis (CAD) system in this paper using the deep learning approach. The proposed CAD system takes input from raw computerized tomography (CT) scans of the patient's chest and automatically predicts disease classification. We designed the Hybrid Deep Learning Algorithm (HDLA) to improve accuracy and reduce processing requirements. The raw CT scans have pre-processed first to enhance their quality for further analysis. We then applied a hybrid model that consists of automatic feature extraction and classification. We propose the robust 2D Convolutional Neural Network (CNN) model to extract the automatic features from the pre-processed CT image. This CNN model assures feature learning with extremely effective 1D feature extraction for each input CT image. The outcome of the 2D CNN model is then normalized using the Min-Max technique. The second step of the proposed hybrid model is related to training and classification using different classifiers. The simulation outcomes using the publically available dataset prove the robustness and efficiency of the proposed model compared to state-of-art algorithms.Keywords: CT scan, Covid-19, deep learning, image processing, lung disease classification
Procedia PDF Downloads 15512083 Numerical Implementation and Testing of Fractioning Estimator Method for the Box-Counting Dimension of Fractal Objects
Authors: Abraham Terán Salcedo, Didier Samayoa Ochoa
Abstract:
This work presents a numerical implementation of a method for estimating the box-counting dimension of self-avoiding curves on a planar space, fractal objects captured on digital images; this method is named fractioning estimator. Classical methods of digital image processing, such as noise filtering, contrast manipulation, and thresholding, among others, are used in order to obtain binary images that are suitable for performing the necessary computations of the fractioning estimator. A user interface is developed for performing the image processing operations and testing the fractioning estimator on different captured images of real-life fractal objects. To analyze the results, the estimations obtained through the fractioning estimator are compared to the results obtained through other methods that are already implemented on different available software for computing and estimating the box-counting dimension.Keywords: box-counting, digital image processing, fractal dimension, numerical method
Procedia PDF Downloads 8312082 Clinch Process Simulation Using Diffuse Elements
Authors: Benzegaou Ali, Brani Benabderrahmane
Abstract:
This work describes a numerical study of the TOX–clinching process using diffuse elements. A computer code baptized SEMA "Static Explicit Method Analysis" is developed to simulate the clinch joining process. The FE code is based on an Updated Lagrangian scheme. The used resolution method is based on an explicit static approach. The integration of the elasto-plastic behavior law is realized with an algorithm of Simo and Taylor. The tools are represented by plane facets.Keywords: diffuse elements, numerical simulation, clinching, contact, large deformation
Procedia PDF Downloads 36312081 Tracking Performance Evaluation of Robust Back-Stepping Control Design for a Nonlinear Electro-Hydraulic Servo System
Authors: Maria Ahmadnezhad, Mohammad Reza Soltanpour
Abstract:
Electrohydraulic servo systems have been used in industry in a wide number of applications. Its dynamics are highly nonlinear and also have large extent of model uncertainties and external disturbances. In this thesis, a robust back-stepping control (RBSC) scheme is proposed to overcome the problem of disturbances and system uncertainties effectively and to improve the tracking performance of EHS systems. In order to implement the proposed control scheme, the system uncertainties in EHS systems are considered as total leakage coefficient and effective oil volume. In addition, in order to obtain the virtual controls for stabilizing system, the update rule for the system uncertainty term is induced by the Lyapunov control function (LCF). To verify the performance and robustness of the proposed control system, computer simulation of the proposed control system using Matlab/Simulink Software is executed. From the computer simulation, it was found that the RBSC system produces the desired tracking performance and has robustness to the disturbances and system uncertainties of EHS systems.Keywords: electro hydraulic servo system, back-stepping control, robust back-stepping control, Lyapunov redesign
Procedia PDF Downloads 29612080 Association between Hypertensive Disorders of Pregnancy and the Development of Offspring Mental and Behavioural Problems: Systematic Review and Meta-Analysis
Authors: Berihun Dachew, Abdullah Mamun, Joemer Maravilla, Rosa Alati
Abstract:
Background: Hypertensive disorders of pregnancy are a major cause of maternal and childhood morbidity and mortality worldwide. However, its effect on offspring mental and behavioural disorders is unclear. Aims:The aim of this study was to provide the best scientific evidence regarding the association between hypertensive disorders of pregnancy and offspring mental and behavioural problems. Methods: We systematically searched Scopus, PubMed, Cochrane, EMBASE, CINAH and PsycINFO databases. A total of 23 studies (11 included in meta-analysis) were identified. A qualitative analysis was conducted by summarizing, comparing, and contrasting the abstracted data for all included studies. For quantitative analysis, relative risk (RR) with 95% confidence interval (95% CI) was used as pooled effect size. Heterogeneity was assessed by measuring Cochran’s Q and I2 test statistics. Results: Of the 23 studies included in this review, 15 studies found that hypertensive disorders of pregnancy had a negative impact for at least one mental or behavioural problem. The pooled effect of 11 studies included in the meta-analysis showed that preeclampsia was associated with increased risk of offspring schizophrenia (RR=1.37; 95% CI, 1.08-1.72). Conclusions: Intrauterine exposure to pre-eclampsia increased the risk of schizophrenia among offspring. However, we found inconclusive finding on the effect of hypertensive disorders of pregnancy and other mental and behavioural problems. Further high quality, large sample, mother child cohort studies are needed to further progress this area of research.Keywords: behavioural disorders, hypertensive disorders of pregnancy, mental disorders, offspring
Procedia PDF Downloads 23812079 Influence of Displacement Amplitude and Vertical Load on the Horizontal Dynamic and Static Behavior of Helical Wire Rope Isolators
Authors: Nicolò Vaiana, Mariacristina Spizzuoco, Giorgio Serino
Abstract:
In this paper, the results of experimental tests performed on a Helical Wire Rope Isolator (HWRI) are presented in order to describe the dynamic and static behavior of the selected metal device in three different displacements ranges, namely small, relatively large, and large displacements ranges, without and under the effect of a vertical load. A testing machine, allowing to apply horizontal displacement or load histories to the tested bearing with a constant vertical load, has been adopted to perform the dynamic and static tests. According to the experimental results, the dynamic behavior of the tested device depends on the applied displacement amplitude. Indeed, the HWRI displays a softening and a hardening stiffness at small and relatively large displacements, respectively, and a stronger nonlinear stiffening behavior at large displacements. Furthermore, the experimental tests reveal that the application of a vertical load allows to have a more flexible device with higher damping properties and that the applied vertical load affects much less the dynamic response of the metal device at large displacements. Finally, a decrease in the static to dynamic effective stiffness ratio with increasing displacement amplitude has been observed.Keywords: base isolation, earthquake engineering, experimental hysteresis loops, wire rope isolators
Procedia PDF Downloads 43312078 Mechanical Analysis and Characterization of Friction Stir Processed Aluminium Alloy
Authors: Jaswinder Kumar, Kulbir Singh Sandhu
Abstract:
Friction stir processing (FSP) is a solid-state surface processing technique. A single-pass FSP was performed on Aluminum alloy at combinations of different tool rotational speeds with cylindrical threaded pin profiled tool. The effect of these parameters on tribological properties was studied. The wear resistance is found to be increased from base metal to a single pass FSP sample. The results revealed that with an increase in tool rotational speed, the wear rate increases. The high heat generation causes matrix softening, which results in an increased wear rate; on the other hand, high heat generation leads to coarse grains, which also affected tribological properties. Furthermore, Microstructure results showed that FSPed alloy has a more refined grain structure as compare to the base material, which may be resulted in enhancement of hardness and resistance to wear in FSP.Keywords: friction stir processing, aluminium alloy, microhardness, microstructure
Procedia PDF Downloads 10912077 A 3D Bioprinting System for Engineering Cell-Embedded Hydrogels by Digital Light Processing
Authors: Jimmy Jiun-Ming Su, Yuan-Min Lin
Abstract:
Bioprinting has been applied to produce 3D cellular constructs for tissue engineering. Microextrusion printing is the most common used method. However, printing low viscosity bioink is a challenge for this method. Herein, we developed a new 3D printing system to fabricate cell-laden hydrogels via a DLP-based projector. The bioprinter is assembled from affordable equipment including a stepper motor, screw, LED-based DLP projector, open source computer hardware and software. The system can use low viscosity and photo-polymerized bioink to fabricate 3D tissue mimics in a layer-by-layer manner. In this study, we used gelatin methylacrylate (GelMA) as bioink for stem cell encapsulation. In order to reinforce the printed construct, surface modified hydroxyapatite has been added in the bioink. We demonstrated the silanization of hydroxyapatite could improve the crosslinking between the interface of hydroxyapatite and GelMA. The results showed that the incorporation of silanized hydroxyapatite into the bioink had an enhancing effect on the mechanical properties of printed hydrogel, in addition, the hydrogel had low cytotoxicity and promoted the differentiation of embedded human bone marrow stem cells (hBMSCs) and retinal pigment epithelium (RPE) cells. Moreover, this bioprinting system has the ability to generate microchannels inside the engineered tissues to facilitate diffusion of nutrients. We believe this 3D bioprinting system has potential to fabricate various tissues for clinical applications and regenerative medicine in the future.Keywords: bioprinting, cell encapsulation, digital light processing, GelMA hydrogel
Procedia PDF Downloads 18112076 Computer-Aided Classification of Liver Lesions Using Contrasting Features Difference
Authors: Hussein Alahmer, Amr Ahmed
Abstract:
Liver cancer is one of the common diseases that cause the death. Early detection is important to diagnose and reduce the incidence of death. Improvements in medical imaging and image processing techniques have significantly enhanced interpretation of medical images. Computer-Aided Diagnosis (CAD) systems based on these techniques play a vital role in the early detection of liver disease and hence reduce liver cancer death rate. This paper presents an automated CAD system consists of three stages; firstly, automatic liver segmentation and lesion’s detection. Secondly, extracting features. Finally, classifying liver lesions into benign and malignant by using the novel contrasting feature-difference approach. Several types of intensity, texture features are extracted from both; the lesion area and its surrounding normal liver tissue. The difference between the features of both areas is then used as the new lesion descriptors. Machine learning classifiers are then trained on the new descriptors to automatically classify liver lesions into benign or malignant. The experimental results show promising improvements. Moreover, the proposed approach can overcome the problems of varying ranges of intensity and textures between patients, demographics, and imaging devices and settings.Keywords: CAD system, difference of feature, fuzzy c means, lesion detection, liver segmentation
Procedia PDF Downloads 32512075 Content-Aware Image Augmentation for Medical Imaging Applications
Authors: Filip Rusak, Yulia Arzhaeva, Dadong Wang
Abstract:
Machine learning based Computer-Aided Diagnosis (CAD) is gaining much popularity in medical imaging and diagnostic radiology. However, it requires a large amount of high quality and labeled training image datasets. The training images may come from different sources and be acquired from different radiography machines produced by different manufacturers, digital or digitized copies of film radiographs, with various sizes as well as different pixel intensity distributions. In this paper, a content-aware image augmentation method is presented to deal with these variations. The results of the proposed method have been validated graphically by plotting the removed and added seams of pixels on original images. Two different chest X-ray (CXR) datasets are used in the experiments. The CXRs in the datasets defer in size, some are digital CXRs while the others are digitized from analog CXR films. With the proposed content-aware augmentation method, the Seam Carving algorithm is employed to resize CXRs and the corresponding labels in the form of image masks, followed by histogram matching used to normalize the pixel intensities of digital radiography, based on the pixel intensity values of digitized radiographs. We implemented the algorithms, resized the well-known Montgomery dataset, to the size of the most frequently used Japanese Society of Radiological Technology (JSRT) dataset and normalized our digital CXRs for testing. This work resulted in the unified off-the-shelf CXR dataset composed of radiographs included in both, Montgomery and JSRT datasets. The experimental results show that even though the amount of augmentation is large, our algorithm can preserve the important information in lung fields, local structures, and global visual effect adequately. The proposed method can be used to augment training and testing image data sets so that the trained machine learning model can be used to process CXRs from various sources, and it can be potentially used broadly in any medical imaging applications.Keywords: computer-aided diagnosis, image augmentation, lung segmentation, medical imaging, seam carving
Procedia PDF Downloads 22212074 Improved Processing Speed for Text Watermarking Algorithm in Color Images
Authors: Hamza A. Al-Sewadi, Akram N. A. Aldakari
Abstract:
Copyright protection and ownership proof of digital multimedia are achieved nowadays by digital watermarking techniques. A text watermarking algorithm for protecting the property rights and ownership judgment of color images is proposed in this paper. Embedding is achieved by inserting texts elements randomly into the color image as noise. The YIQ image processing model is found to be faster than other image processing methods, and hence, it is adopted for the embedding process. An optional choice of encrypting the text watermark before embedding is also suggested (in case required by some applications), where, the text can is encrypted using any enciphering technique adding more difficulty to hackers. Experiments resulted in embedding speed improvement of more than double the speed of other considered systems (such as least significant bit method, and separate color code methods), and a fairly acceptable level of peak signal to noise ratio (PSNR) with low mean square error values for watermarking purposes.Keywords: steganography, watermarking, time complexity measurements, private keys
Procedia PDF Downloads 14312073 A Study on Bilingual Semantic Processing: Category Effects and Age Effects
Authors: Lai Yi-Hsiu
Abstract:
The present study addressed the nature of bilingual semantic processing in Mandarin Chinese and Southern Min and examined category effects and age effects. Nineteen bilingual adults of Mandarin Chinese and Southern Min, nine monolingual seniors of Mandarin Chinese, and ten monolingual seniors of Southern Min in Taiwan individually completed two semantic tasks: Picture naming and category fluency tasks. The instruments for the naming task were sixty black-and-white pictures, including thirty-five object pictures and twenty-five action pictures. The category fluency task also consisted of two semantic categories – objects (or nouns) and actions (or verbs). The reaction time for each picture/question was additionally calculated and analyzed. Oral productions in Mandarin Chinese and in Southern Min were compared and discussed to examine the category effects and age effects. The results of the category fluency task indicated that the content of information of these seniors was comparatively deteriorated, and thus they produced a smaller number of semantic-lexical items. Significant group differences were also found in the reaction time results. Category effects were significant for both adults and seniors in the semantic fluency task. The findings of the present study will help characterize the nature of the bilingual semantic processing of adults and seniors, and contribute to the fields of contrastive and corpus linguistics.Keywords: bilingual semantic processing, aging, Mandarin Chinese, Southern Min
Procedia PDF Downloads 57112072 Bacteriophage Lysis Of Physiologically Stressed Listeria Monocytogenes In A Simulated Seafood Processing Environment
Authors: Geevika J. Ganegama Arachchi, Steve H. Flint, Lynn McIntyre, Cristina D. Cruz, Beatrice M. Dias-Wanigasekera, Craig Billington, J. Andrew Hudson, Anthony N. Mutukumira
Abstract:
In seafood processing plants, Listeriamonocytogenes(L. monocytogenes)likely exists in a metabolically stressed state due to the nutrient-deficient environment, processing treatments such as heating, curing, drying, and freezing, and exposure to detergents and disinfectants. Stressed L. monocytogenes cells have been shown to be as pathogenic as unstressed cells. This study investigated lytic efficacy of (LiMN4L, LiMN4p, and LiMN17) which were previouslycharacterized as virulent against physiologically stressed cells of three seafood borne L. monocytogenesstrains (19CO9, 19DO3, and 19EO3).Physiologically compromised cells ofL. monocytogenesstrains were prepared by aging cultures in TrypticaseSoy Broth at 15±1°C for 72 h; heat injuringcultures at 54±1 - 55±1°C for 40 - 60 min;salt-stressing cultures in Milli-Q water were incubated at 25±1°C in darkness for three weeks; and incubating cultures in 9% (w/v) NaCl at 15±1°C for 72 h. Low concentrations of physiologically compromised cells of three L. monocytogenesstrainswere challenged in vitrowith high titre of three phages in separate experiments using Fish Broth medium (aqueous fish extract) at 15 °C in order to mimic the environment of seafood processing plant. Each phage, when present at ≈9 log10 PFU/ml, reduced late exponential phase cells of L. monocytogenes suspended in fish protein broth at ≈2-3 log10 CFU/ml to a non-detectable level (< 10 CFU/ml). Each phage, when present at ≈8.5 log10 PFU/ml, reduced both heat-injured cells present at 2.5-3.6 log10 CFU/ml and starved cells that were showed coccoid shape, present at ≈2-3 log10 CFU/ml to < 10 CFU/ml after 30 min. Phages also reduced salt-stressed cellspresent at ≈3 log10 CFU/ml by > 2 log10. L. monocytogenes (≈8 log10 CFU/ml) were reduced to below the detection limit (1 CFU/ml) by the three successive phage infections over 16 h, indicating that emergence of spontaneous phage resistance was infrequent. The three virulent phages showed high decontamination potential for physiologically stressed L. monocytogenes strains from seafood processing environments.Keywords: physiologically stressed L. monocytogenes, heat injured, seafood processing environment, virulent phage
Procedia PDF Downloads 13512071 An Interpolation Tool for Data Transfer in Two-Dimensional Ice Accretion Problems
Authors: Marta Cordero-Gracia, Mariola Gomez, Olivier Blesbois, Marina Carrion
Abstract:
One of the difficulties in icing simulations is for extended periods of exposure, when very large ice shapes are created. As well as being large, they can have complex shapes, such as a double horn. For icing simulations, these configurations are currently computed in several steps. The icing step is stopped when the ice shapes become too large, at which point a new mesh has to be created to allow for further CFD and ice growth simulations to be performed. This can be very costly, and is a limiting factor in the simulations that can be performed. A way to avoid the costly human intervention in the re-meshing step of multistep icing computation is to use mesh deformation instead of re-meshing. The aim of the present work is to apply an interpolation method based on Radial Basis Functions (RBF) to transfer deformations from surface mesh to volume mesh. This deformation tool has been developed specifically for icing problems. It is able to deal with localized, sharp and large deformations, unlike the tools traditionally used for more smooth wing deformations. This tool will be presented along with validation on typical two-dimensional icing shapes.Keywords: ice accretion, interpolation, mesh deformation, radial basis functions
Procedia PDF Downloads 31312070 The Effect of Computer-Mediated vs. Face-to-Face Instruction on L2 Pragmatics: A Meta-Analysis
Authors: Marziyeh Yousefi, Hossein Nassaji
Abstract:
This paper reports the results of a meta-analysis of studies on the effects of instruction mode on learning second language pragmatics during the last decade (from 2006 to 2016). After establishing related inclusion/ exclusion criteria, 39 published studies were retrieved and included in the present meta-analysis. Studies were later coded for face-to-face and computer-assisted mode of instruction. Statistical procedures were applied to obtain effect sizes. It was found that Computer-Assisted-Language-Learning studies generated larger effects than Face-to-Face instruction.Keywords: meta-analysis, effect size, L2 pragmatics, comprehensive meta-analysis, face-to-face, computer-assisted language learning
Procedia PDF Downloads 22312069 A Supervised Learning Data Mining Approach for Object Recognition and Classification in High Resolution Satellite Data
Authors: Mais Nijim, Rama Devi Chennuboyina, Waseem Al Aqqad
Abstract:
Advances in spatial and spectral resolution of satellite images have led to tremendous growth in large image databases. The data we acquire through satellites, radars and sensors consists of important geographical information that can be used for remote sensing applications such as region planning, disaster management. Spatial data classification and object recognition are important tasks for many applications. However, classifying objects and identifying them manually from images is a difficult task. Object recognition is often considered as a classification problem, this task can be performed using machine-learning techniques. Despite of many machine-learning algorithms, the classification is done using supervised classifiers such as Support Vector Machines (SVM) as the area of interest is known. We proposed a classification method, which considers neighboring pixels in a region for feature extraction and it evaluates classifications precisely according to neighboring classes for semantic interpretation of region of interest (ROI). A dataset has been created for training and testing purpose; we generated the attributes by considering pixel intensity values and mean values of reflectance. We demonstrated the benefits of using knowledge discovery and data-mining techniques, which can be on image data for accurate information extraction and classification from high spatial resolution remote sensing imagery.Keywords: remote sensing, object recognition, classification, data mining, waterbody identification, feature extraction
Procedia PDF Downloads 34012068 Limbic Involvement in Visual Processing
Authors: Deborah Zelinsky
Abstract:
The retina filters millions of incoming signals into a smaller amount of exiting optic nerve fibers that travel to different portions of the brain. Most of the signals are for eyesight (called "image-forming" signals). However, there are other faster signals that travel "elsewhere" and are not directly involved with eyesight (called "non-image-forming" signals). This article centers on the neurons of the optic nerve connecting to parts of the limbic system. Eye care providers are currently looking at parvocellular and magnocellular processing pathways without realizing that those are part of an enormous "galaxy" of all the body systems. Lenses are modifying both non-image and image-forming pathways, taking A.M. Skeffington's seminal work one step further. Almost 100 years ago, he described the Where am I (orientation), Where is It (localization), and What is It (identification) pathways. Now, among others, there is a How am I (animation) and a Who am I (inclination, motivation, imagination) pathway. Classic eye testing considers pupils and often assesses posture and motion awareness, but classical prescriptions often overlook limbic involvement in visual processing. The limbic system is composed of the hippocampus, amygdala, hypothalamus, and anterior nuclei of the thalamus. The optic nerve's limbic connections arise from the intrinsically photosensitive retinal ganglion cells (ipRGC) through the "retinohypothalamic tract" (RHT). There are two main hypothalamic nuclei with direct photic inputs. These are the suprachiasmatic nucleus and the paraventricular nucleus. Other hypothalamic nuclei connected with retinal function, including mood regulation, appetite, and glucose regulation, are the supraoptic nucleus and the arcuate nucleus. The retino-hypothalamic tract is often overlooked when we prescribe eyeglasses. Each person is different, but the lenses we choose are influencing this fast processing, which affects each patient's aiming and focusing abilities. These signals arise from the ipRGC cells that were only discovered 20+ years ago and do not address the campana retinal interneurons that were only discovered 2 years ago. As eyecare providers, we are unknowingly altering such factors as lymph flow, glucose metabolism, appetite, and sleep cycles in our patients. It is important to know what we are prescribing as the visual processing evaluations expand past the 20/20 central eyesight.Keywords: neuromodulation, retinal processing, retinohypothalamic tract, limbic system, visual processing
Procedia PDF Downloads 8512067 A Design System for Complex Profiles of Machine Members Using a Synthetic Curve
Authors: N. Sateesh, C. S. P. Rao, K. Satyanarayana, C. Rajashekar
Abstract:
This paper proposes a development of a CAD/CAM system for complex profiles of various machine members using a synthetic curve i.e. B-spline. Conventional methods in designing and manufacturing of complex profiles are tedious and time consuming. Even programming those on a computer numerical control (CNC) machine can be a difficult job because of the complexity of the profiles. The system developed provides graphical and numerical representation B-spline profile for any given input. In this paper, the system is applicable to represent a cam profile with B-spline and attempt is made to improve the follower motion.Keywords: plate-cams, cam profile, b-spline, computer numerical control (CNC), computer aided design and computer aided manufacturing (CAD/CAM), R-D-R-D (rise-dwell-return-dwell)
Procedia PDF Downloads 61112066 Multivariate Data Analysis for Automatic Atrial Fibrillation Detection
Authors: Zouhair Haddi, Stephane Delliaux, Jean-Francois Pons, Ismail Kechaf, Jean-Claude De Haro, Mustapha Ouladsine
Abstract:
Atrial fibrillation (AF) has been considered as the most common cardiac arrhythmia, and a major public health burden associated with significant morbidity and mortality. Nowadays, telemedical approaches targeting cardiac outpatients situate AF among the most challenged medical issues. The automatic, early, and fast AF detection is still a major concern for the healthcare professional. Several algorithms based on univariate analysis have been developed to detect atrial fibrillation. However, the published results do not show satisfactory classification accuracy. This work was aimed at resolving this shortcoming by proposing multivariate data analysis methods for automatic AF detection. Four publicly-accessible sets of clinical data (AF Termination Challenge Database, MIT-BIH AF, Normal Sinus Rhythm RR Interval Database, and MIT-BIH Normal Sinus Rhythm Databases) were used for assessment. All time series were segmented in 1 min RR intervals window and then four specific features were calculated. Two pattern recognition methods, i.e., Principal Component Analysis (PCA) and Learning Vector Quantization (LVQ) neural network were used to develop classification models. PCA, as a feature reduction method, was employed to find important features to discriminate between AF and Normal Sinus Rhythm. Despite its very simple structure, the results show that the LVQ model performs better on the analyzed databases than do existing algorithms, with high sensitivity and specificity (99.19% and 99.39%, respectively). The proposed AF detection holds several interesting properties, and can be implemented with just a few arithmetical operations which make it a suitable choice for telecare applications.Keywords: atrial fibrillation, multivariate data analysis, automatic detection, telemedicine
Procedia PDF Downloads 26812065 Continuous Processing Approaches for Tunable Asymmetric Photochemical Synthesis
Authors: Amanda C. Evans
Abstract:
Enabling technologies such as continuous processing (CP) approaches can provide the tools needed to control and manipulate reactivities and transform chemical reactions into micro-controlled in-flow processes. Traditional synthetic approaches can be radically transformed by the application of CP, facilitating the pairing of chemical methodologies with technologies from other disciplines. CP supports sustainable processes that controllably generate reaction specificity utilizing supramolecular interactions. Continuous photochemical processing is an emerging field of investigation. The use of light to drive chemical reactivity is not novel, but the controlled use of specific and tunable wavelengths of light to selectively generate molecular structure under continuous processing conditions is an innovative approach towards chemical synthesis. This investigation focuses on the use of circularly polarized (cp) light as a sustainable catalyst for the CP generation of asymmetric molecules. Chiral photolysis has already been achieved under batch, solid-phase conditions: using synchrotron-sourced cp light, asymmetric photolytic selectivities of up to 4.2% enantiomeric excess (e.e.) have been reported. In order to determine the optimal wavelengths to use for irradiation with cp light for any given molecular building block, CD and anisotropy spectra for each building block of interest have been generated in two different solvents (water, hexafluoroisopropanol) across a range of wavelengths (130-400 nm). These spectra are being used to support a series of CP experiments using cp light to generate enantioselectivity.Keywords: anisotropy, asymmetry, flow chemistry, active pharmaceutical ingredients
Procedia PDF Downloads 15712064 A Fact-Finding Analysis on the Expulsions Made under Title 42 in Us
Authors: Avi Shrivastava
Abstract:
Title 42, an emergency health decree, has forced the federal authorities to turn away asylum seekers and all other border crossers since last year. When Title 42 was first deployed in immigration detention centers, where many migrants are held when they arrive at the U.S.-Mexico border, the Trump administration embraced it as a strategy. Expulsions Policy and New Border Challenges will be examined in regard to Title 42 concerns. Humanitarian measures for refugees arriving at the US-Mexico border are the focus of this article. To a large extent, this article addresses the implications of the United States' use of Title 42 in expelling refugees and the possible ramifications of doing away with it. A secondary data collecting strategy was used to gather the information for this study, allowing researchers to examine a large number of previously collected data sets. Information about Title 42 may be found in a variety of places, such as scholarly publications, newspapers, books, and the internet. The inquiry employed qualitative and explanatory research approaches. The claim that 1.7 million individuals were forced to leave the country as a result of it was withdrawn. Since CBP and ICE were limited in their ability to process deportees, it employed a very random patchwork technique in selecting the expelled individuals. As a consequence, repeat offenders, particularly those who were single, got a reduced punishment. The government will be compelled to focus on long-overdue but vital border enhancements if expulsions are halted. Title 42 provisions may help expedite the processing of asylum and other types of humanitarian relief. The government is prepared for an increase in arrivals, but ending the program would lead to a return to arrival levels seen during the Title 42 period.Keywords: migrants, refugees, title 42, medical, trump administration
Procedia PDF Downloads 8712063 Time Delay Estimation Using Signal Envelopes for Synchronisation of Recordings
Authors: Sergei Aleinik, Mikhail Stolbov
Abstract:
In this work, a method of time delay estimation for dual-channel acoustic signals (speech, music, etc.) recorded under reverberant conditions is investigated. Standard methods based on cross-correlation of the signals show poor results in cases involving strong reverberation, large distances between microphones and asynchronous recordings. Under similar conditions, a method based on cross-correlation of temporal envelopes of the signals delivers a delay estimation of acceptable quality. This method and its properties are described and investigated in detail, including its limits of applicability. The method’s optimal parameter estimation and a comparison with other known methods of time delay estimation are also provided.Keywords: cross-correlation, delay estimation, signal envelope, signal processing
Procedia PDF Downloads 48512062 Signal Processing Techniques for Adaptive Beamforming with Robustness
Authors: Ju-Hong Lee, Ching-Wei Liao
Abstract:
Adaptive beamforming using antenna array of sensors is useful in the process of adaptively detecting and preserving the presence of the desired signal while suppressing the interference and the background noise. For conventional adaptive array beamforming, we require a prior information of either the impinging direction or the waveform of the desired signal to adapt the weights. The adaptive weights of an antenna array beamformer under a steered-beam constraint are calculated by minimizing the output power of the beamformer subject to the constraint that forces the beamformer to make a constant response in the steering direction. Hence, the performance of the beamformer is very sensitive to the accuracy of the steering operation. In the literature, it is well known that the performance of an adaptive beamformer will be deteriorated by any steering angle error encountered in many practical applications, e.g., the wireless communication systems with massive antennas deployed at the base station and user equipment. Hence, developing effective signal processing techniques to deal with the problem due to steering angle error for array beamforming systems has become an important research work. In this paper, we present an effective signal processing technique for constructing an adaptive beamformer against the steering angle error. The proposed array beamformer adaptively estimates the actual direction of the desired signal by using the presumed steering vector and the received array data snapshots. Based on the presumed steering vector and a preset angle range for steering mismatch tolerance, we first create a matrix related to the direction vector of signal sources. Two projection matrices are generated from the matrix. The projection matrix associated with the desired signal information and the received array data are utilized to iteratively estimate the actual direction vector of the desired signal. The estimated direction vector of the desired signal is then used for appropriately finding the quiescent weight vector. The other projection matrix is set to be the signal blocking matrix required for performing adaptive beamforming. Accordingly, the proposed beamformer consists of adaptive quiescent weights and partially adaptive weights. Several computer simulation examples are provided for evaluating and comparing the proposed technique with the existing robust techniques.Keywords: adaptive beamforming, robustness, signal blocking, steering angle error
Procedia PDF Downloads 12412061 Wait-Optimized Scheduler Algorithm for Efficient Process Scheduling in Computer Systems
Authors: Md Habibur Rahman, Jaeho Kim
Abstract:
Efficient process scheduling is a crucial factor in ensuring optimal system performance and resource utilization in computer systems. While various algorithms have been proposed over the years, there are still limitations to their effectiveness. This paper introduces a new Wait-Optimized Scheduler (WOS) algorithm that aims to minimize process waiting time by dividing them into two layers and considering both process time and waiting time. The WOS algorithm is non-preemptive and prioritizes processes with the shortest WOS. In the first layer, each process runs for a predetermined duration, and any unfinished process is subsequently moved to the second layer, resulting in a decrease in response time. Whenever the first layer is free or the number of processes in the second layer is twice that of the first layer, the algorithm sorts all the processes in the second layer based on their remaining time minus waiting time and sends one process to the first layer to run. This ensures that all processes eventually run, optimizing waiting time. To evaluate the performance of the WOS algorithm, we conducted experiments comparing its performance with traditional scheduling algorithms such as First-Come-First-Serve (FCFS) and Shortest-Job-First (SJF). The results showed that the WOS algorithm outperformed the traditional algorithms in reducing the waiting time of processes, particularly in scenarios with a large number of short tasks with long wait times. Our study highlights the effectiveness of the WOS algorithm in improving process scheduling efficiency in computer systems. By reducing process waiting time, the WOS algorithm can improve system performance and resource utilization. The findings of this study provide valuable insights for researchers and practitioners in developing and implementing efficient process scheduling algorithms.Keywords: process scheduling, wait-optimized scheduler, response time, non-preemptive, waiting time, traditional scheduling algorithms, first-come-first-serve, shortest-job-first, system performance, resource utilization
Procedia PDF Downloads 9112060 Fractional Residue Number System
Authors: Parisa Khoshvaght, Mehdi Hosseinzadeh
Abstract:
During the past few years, the Residue Number System (RNS) has been receiving considerable interest due to its parallel and fault-tolerant properties. This system is a useful tool for Digital Signal Processing (DSP) since it can support parallel, carry-free, high-speed and low power arithmetic. One of the drawbacks of Residue Number System is the fractional numbers, that is, the corresponding circuit is very hard to realize in conventional CMOS technology. In this paper, we propose a method in which the numbers of transistors are significantly reduced. The related delay is extremely diminished, in the first glance we use this method to solve concerning problem of one decimal functional number some how this proposition can be extended to generalize the idea. Another advantage of this method is the independency on the kind of moduli.Keywords: computer arithmetic, residue number system, number system, one-Hot, VLSI
Procedia PDF Downloads 49512059 Algorithm for Path Recognition in-between Tree Rows for Agricultural Wheeled-Mobile Robots
Authors: Anderson Rocha, Pedro Miguel de Figueiredo Dinis Oliveira Gaspar
Abstract:
Machine vision has been widely used in recent years in agriculture, as a tool to promote the automation of processes and increase the levels of productivity. The aim of this work is the development of a path recognition algorithm based on image processing to guide a terrestrial robot in-between tree rows. The proposed algorithm was developed using the software MATLAB, and it uses several image processing operations, such as threshold detection, morphological erosion, histogram equalization and the Hough transform, to find edge lines along tree rows on an image and to create a path to be followed by a mobile robot. To develop the algorithm, a set of images of different types of orchards was used, which made possible the construction of a method capable of identifying paths between trees of different heights and aspects. The algorithm was evaluated using several images with different characteristics of quality and the results showed that the proposed method can successfully detect a path in different types of environments.Keywords: agricultural mobile robot, image processing, path recognition, hough transform
Procedia PDF Downloads 14612058 Using Automated Agents to Facilitate Instructions in a Large Online Course
Authors: David M Gilstrap
Abstract:
In an online course with a large enrollment, the potential exists for the instructor to become overburdened with having to respond to students’ emails, which consequently decreases the instructor’s efficiency in teaching the course. Repetition of instructions is an effective way of reducing confusion among students, which in turn increases their efficiencies, as well. World of Turf is the largest online course at Michigan State University, which employs Brightspace as its management system (LMS) software. Recently, the LMS upgraded its capabilities to utilize agents, which are auto generated email notifications to students based on certain criteria. Agents are additional tools that can enhance course design. They can be run on-demand or according to a schedule. Agents can be timed to effectively remind students of approaching deadlines. The content of these generated emails can also include reinforced instructions. With a large online course, even a small percentage of students that either do not read or do not comprehend the course syllabus or do not notice instructions on course pages can result in numerous emails to the instructor, often near the deadlines for assignments. Utilizing agents to decrease the number of emails from students has enabled the instructor to efficiently instruct more than one thousand students per semester without any graduate student teaching assistants.Keywords: agents, Brightspace, large enrollment, learning management system, repetition of instructions
Procedia PDF Downloads 20312057 Selection of Pichia kudriavzevii Strain for the Production of Single-Cell Protein from Cassava Processing Waste
Authors: Phakamas Rachamontree, Theerawut Phusantisampan, Natthakorn Woravutthikul, Peerapong Pornwongthong, Malinee Sriariyanun
Abstract:
A total of 115 yeast strains isolated from local cassava processing wastes were measured for crude protein content. Among these strains, the strain MSY-2 possessed the highest protein concentration (>3.5 mg protein/mL). By using molecular identification tools, it was identified to be a strain of Pichia kudriavzevii based on similarity of D1/D2 domain of 26S rDNA region. In this study, to optimize the protein production by MSY-2 strain, Response Surface Methodology (RSM) was applied. The tested parameters were the carbon content, nitrogen content, and incubation time. Here, the value of regression coefficient (R2) = 0.7194 could be explained by the model, which is high to support the significance of the model. Under the optimal condition, the protein content was produced up to 3.77 g per L of the culture and MSY-2 strain contain 66.8 g protein per 100 g of cell dry weight. These results revealed the plausibility of applying the novel strain of yeast in single-cell protein production.Keywords: single cell protein, response surface methodology, yeast, cassava processing waste
Procedia PDF Downloads 40312056 Clinical Outcomes of Critically Ill Patients with Sepsis Receiving Extended and Standard Meropenem Infusion in Malaysian Hospitals
Authors: Fahmi Hassan, Noorizan Abdul Aziz, Yahaya Hassan, Hazlinda Abu Hassan
Abstract:
Sepsis incidence in critical care settings is a major problem in health care. Extended antibiotic infusion is thought to be superior to traditional dosing especially when treating critically ill patients with sepsis. We compared clinical outcomes of critically ill patients with sepsis receiving 30-minute meropenem infusion and three-hour meropenem infusion. A retrospective case-control study was conducted among septic patients treated with meropenem infusion in ICUs of three hospitals. Patients included in the study received either extended or standard meropenem infusion as per the practice of individual settings. Outcomes and clinical data were retrospectively collected from the electronic databases and patients’ files. A total of 108 patients received extended meropenem infusion while another 117 patients received standard meropenem infusion. Patients receiving the extended meropenem infusion were found to have a significantly lower shorter length of hospital and ICU stay. It was also found that among those receiving extended meropenem infusion, 54.7% (64/117) had a reduction of SAPS II score, while only 44% (48/108) of patients receiving standard meropenem infusion had reduced scores. This study will strengthen the evidence in using extended meropenem infusion as a standard practice in critical care settings. As this is the first study of its kind done in Malaysia, it proves that prolonged meropenem infusion may be beneficial to critically ill patients with sepsis. However, randomized clinical trials with large sample size should be carried out in local settings in order to minimize other confounders that may influence with the result of the study.Keywords: antibiotics, beta lactams, critical care, extended infusion, meropenem
Procedia PDF Downloads 408