Search results for: normalization of the auto- correlation function
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8757

Search results for: normalization of the auto- correlation function

8697 Development of Membrane Reactor for Auto Thermal Reforming of Dimethyl Ether for Hydrogen Production

Authors: Tie-Qing Zhang, Seunghun Jung, Young-Bae Kim

Abstract:

This research is devoted to developing a membrane reactor to flexibly meet the hydrogen demand of onboard fuel cells, which is an important part of green energy development. Among many renewable chemical products, dimethyl ether (DME) has the advantages of low reaction temperature (400 °C in this study), high hydrogen atom content, low toxicity, and easy preparation. Autothermal reforming, on the other hand, has a high hydrogen recovery rate and exhibits thermal neutrality during the reaction process, so the additional heat source in the hydrogen production process can be omitted. Therefore, the DME auto thermal reforming process was adopted in this study. To control the temperature of the reaction catalyst bed and hydrogen production rate, a Model Predictive Control (MPC) scheme was designed. Taking the above two variables as the control objectives, stable operation of the reformer can be achieved by controlling the flow rates of DME, steam, and high-purity air in real-time. To prevent catalyst poisoning in the fuel cell, the hydrogen needs to be purified to reduce the carbon monoxide content to below 50 ppm. Therefore, a Pd-Ag hydrogen semi-permeable membrane with a thickness of 3-5 μm was inserted into the auto thermal reactor, and the permeation efficiency of hydrogen was improved by steam purging on the permeation side. Finally, hydrogen with a purity of 99.99 was obtained.

Keywords: hydrogen production, auto thermal reforming, membrane, fuel cell

Procedia PDF Downloads 63
8696 Author Name Disambiguation for Biomedical Literature

Authors: Parthiban Srinivasan

Abstract:

PubMed provides online access to the National Library of Medicine database (MEDLINE) and other publications, which contain close to 25 million scientific citations from 1865 to the present. There are close to 80 million author name instances in those close to 25 million citations. For any work of literature, a fundamental issue is to identify the individual(s) who wrote it, and conversely, to identify all of the works that belong to a given individual. Due to the lack of universal standards for name information, there are two aspects of name ambiguity: name synonymy (a single author with multiple name representations), and name homonymy (multiple authors sharing the same name representation). In this talk, we present some results from our extensive work in author name disambiguation for PubMed citations. Information will be presented on the effectiveness and shortcomings of different aspects of successful name disambiguation such as parsing, validation, standardization and normalization.

Keywords: disambiguation, normalization, parsing, PubMed

Procedia PDF Downloads 273
8695 Investigation of the Speckle Pattern Effect for Displacement Assessments by Digital Image Correlation

Authors: Salim Çalışkan, Hakan Akyüz

Abstract:

Digital image correlation has been accustomed as a versatile and efficient method for measuring displacements on the article surfaces by comparing reference subsets in undeformed images with the define target subset in the distorted image. The theoretical model points out that the accuracy of the digital image correlation displacement data can be exactly anticipated based on the divergence of the image noise and the sum of the squares of the subset intensity gradients. The digital image correlation procedure locates each subset of the original image in the distorted image. The software then determines the displacement values of the centers of the subassemblies, providing the complete displacement measures. In this paper, the effect of the speckle distribution and its effect on displacements measured out plane displacement data as a function of the size of the subset was investigated. Nine groups of speckle patterns were used in this study: samples are sprayed randomly by pre-manufactured patterns of three different hole diameters, each with three coverage ratios, on a computer numerical control punch press. The resulting displacement values, referenced at the center of the subset, are evaluated based on the average of the displacements of the pixel’s interior the subset.

Keywords: digital image correlation, speckle pattern, experimental mechanics, tensile test, aluminum alloy

Procedia PDF Downloads 41
8694 Correlation between Creatinine Level with Erectile Dysfunction among Diabetics in Temerloh Health Clinic

Authors: Mohammad Zainie Bin Hassan

Abstract:

Background: Erectile dysfunction (ED) is a complication commonly seen among men with diabetes which can be assessed based upon International Index of Erectile Function (IIEF-5) questionnaire. Creatinine level is a blood test that indicates kidney functionality. Object: To evaluate the association between ED, determined by the IIEF-5scores and Creatinine level in diabetic men attending Temerloh Health Clinic, Pahang, Malaysia.Hence, to identify raising Creatinine level related with ED or not. Methods: All married diabetic patients will be investigated face to face after consented for answering the IIEF-5 questionnaire. Creatinine level will be taken by using standard method.Patients with no sexual partner, refuse to answer the questionnaire, cancer, stroke, heart disease and language barrier will be excluded.Data obtained from IIEF-5 score and Creatinine level will be analyzed by using Pearson correlation. All statistical value determined by p=0.05. ED will be categorized accordingly to IIEF-5 scores: no ED (22-25), mild (17-21), moderate (12-16), severe (8-11) and very severe (1-7). Results: A total of 450 patients were investigated with 385 patients were included (85.6% respondant rate) and 65 patients were excluded in this study with age range from 29 to 85 years old. 7% had no ED, 28% mild ED, 34% moderate ED, 16% severe ED and 15% had very severe ED. There was a significant negative correlation between Creatinine level and IIEF-5 scores (r=-0.218, p <0.001). This result implicated that poor kidney function which indicated by high Creatinine level associated significantly with erectile dysfunction. 93% had ED with a different range of severity which triggers for appropriate aggressive ED management among diabetics. Conclusion: The high level of Creatinine is associated with erectile dysfunction among diabetics in Temerloh Health Clinic.

Keywords: correlation, creatinine level, erectile dysfunction, ED, diabetes

Procedia PDF Downloads 352
8693 Throughput of Point Coordination Function (PCF)

Authors: Faisel Eltuhami Alzaalik, Omar Imhemed Alramli, Ahmed Mohamed Elaieb

Abstract:

The IEEE 802.11 defines two modes of MAC, distributed coordination function (DCF) and point coordination function (PCF) mode. The first sub-layer of the MAC is the distributed coordination function (DCF). A contention algorithm is used via DCF to provide access to all traffic. The point coordination function (PCF) is the second sub-layer used to provide contention-free service. PCF is upper DCF and it uses features of DCF to establish guarantee access of its users. Some papers and researches that have been published in this technology were reviewed in this paper, as well as talking briefly about the distributed coordination function (DCF) technology. The simulation of the PCF function have been applied by using a simulation program called network simulator (NS2) and have been found out the throughput of a transmitter system by using this function.

Keywords: DCF, PCF, throughput, NS2

Procedia PDF Downloads 540
8692 Determinants of Economic Growth in Pakistan: A Structural Vector Auto Regression Approach

Authors: Muhammad Ajmair

Abstract:

This empirical study followed structural vector auto regression (SVAR) approach proposed by the so-called AB-model of Amisano and Giannini (1997) to check the impact of relevant macroeconomic determinants on economic growth in Pakistan. Before that auto regressive distributive lag (ARDL) bound testing technique and time varying parametric approach along with general to specific approach was employed to find out relevant significant determinants of economic growth. To our best knowledge, no author made such a study that employed auto regressive distributive lag (ARDL) bound testing and time varying parametric approach with general to specific approach in empirical literature, but current study will bridge this gap. Annual data was taken from World Development Indicators (2014) during period 1976-2014. The widely-used Schwarz information criterion and Akaike information criterion were considered for the lag length in each estimated equation. Main findings of the study are that remittances received, gross national expenditures and inflation are found to be the best relevant positive and significant determinants of economic growth. Based on these empirical findings, we conclude that government should focus on overall economic growth augmenting factors while formulating any policy relevant to the concerned sector.

Keywords: economic growth, gross national expenditures, inflation, remittances

Procedia PDF Downloads 172
8691 Assessment of Pre-Processing Influence on Near-Infrared Spectra for Predicting the Mechanical Properties of Wood

Authors: Aasheesh Raturi, Vimal Kothiyal, P. D. Semalty

Abstract:

We studied mechanical properties of Eucalyptus tereticornis using FT-NIR spectroscopy. Firstly, spectra were pre-processed to eliminate useless information. Then, prediction model was constructed by partial least squares regression. To study the influence of pre-processing on prediction of mechanical properties for NIR analysis of wood samples, we applied various pretreatment methods like straight line subtraction, constant offset elimination, vector-normalization, min-max normalization, multiple scattering. Correction, first derivative, second derivatives and their combination with other treatment such as First derivative + straight line subtraction, First derivative+ vector normalization and First derivative+ multiplicative scattering correction. The data processing methods in combination of preprocessing with different NIR regions, RMSECV, RMSEP and optimum factors/rank were obtained by optimization process of model development. More than 350 combinations were obtained during optimization process. More than one pre-processing method gave good calibration/cross-validation and prediction/test models, but only the best calibration/cross-validation and prediction/test models are reported here. The results show that one can safely use NIR region between 4000 to 7500 cm-1 with straight line subtraction, constant offset elimination, first derivative and second derivative preprocessing method which were found to be most appropriate for models development.

Keywords: FT-NIR, mechanical properties, pre-processing, PLS

Procedia PDF Downloads 312
8690 Hypocalcaemia Inducing Heart Failure: A Rare Presentation

Authors: A. Kherraf, M. Bouziane, L. Azzouzi, R. Habbal

Abstract:

Introduction: Hypocalcaemia is a rare cause of heart failure. We report the clinical case of a young patient with reversible dilated cardiomyopathy secondary to hypocalcaemia in the context of hyperparathyroidism. Clinical case: We report the clinical case of a 23-year-old patient with a history of thyroidectomy for papillary thyroid carcinoma 3 years previously, who presented to the emergency room with a progressive onset dyspnea and edema of the lower limbs. Clinical examination showed hypotension at 90/70 mmHg, tachycardia at 102 bpm, and edema of the lower limbs. The ECG showed a regular sinus rhythm with a prolonged corrected QT interval to 520ms. The chest x-ray showed cardiomegaly. Echocardiography revealed dilated cardiomyopathy with biventricular dysfunction and a left ventricular ejection fraction of 45%, as well as moderate mitral insufficiency by restriction of the posterior mitral leaflet, moderate tricuspid insufficiency, and a dilated inferior vena cava with a pulmonary arterial pressure estimated at 46 mmHg. Blood tests revealed severe hypocalcemia at 38 mg / l with normal albumin and thyroxine levels, as well as hyperphosphatemia and increased TSH. The patient received calcium intake and vitamin D supplementation and was treated with beta blockers, ACE inhibitors, and diuretics with good progress and progressive normalization of cardiac function. Discussion: The cardiovascular manifestations of hypocalcaemia usually appear with deeply low serum calcium levels. This can lead to hypotension, arrhythmias, ventricular fibrillation, prolonged QT interval, or even heart failure. Heart failure is a rare and serious complication of hypocalcemia but most often characterized by complete normalization of myocardial function after treatment. The etiology of the hypocalcaemia, in this case, was probably related to accidental parathyroid removal during thyroidectomy. This is why careful monitoring of calcium levels is recommended after surgery. Conclusion: Hypocalcemic heart failure is rare but reversible heart disease. Systematic monitoring of serum calcium should be performed in all patients after thyroid surgery to avoid any complications related to hypoparathyroidism.

Keywords: hypocalcemia, heart failure, thyroid surgery, hypoparathyroidism

Procedia PDF Downloads 115
8689 Enhancement of Underwater Haze Image with Edge Reveal Using Pixel Normalization

Authors: M. Dhana Lakshmi, S. Sakthivel Murugan

Abstract:

As light passes from source to observer in the water medium, it is scattered by the suspended particulate matter. This scattering effect will plague the captured images with non-uniform illumination, blurring details, halo artefacts, weak edges, etc. To overcome this, pixel normalization with an Amended Unsharp Mask (AUM) filter is proposed to enhance the degraded image. To validate the robustness of the proposed technique irrespective of atmospheric light, the considered datasets are collected on dual locations. For those images, the maxima and minima pixel intensity value is computed and normalized; then the AUM filter is applied to strengthen the blurred edges. Finally, the enhanced image is obtained with good illumination and contrast. Thus, the proposed technique removes the effect of scattering called de-hazing and restores the perceptual information with enhanced edge detail. Both qualitative and quantitative analyses are done on considering the standard non-reference metric called underwater image sharpness measure (UISM), and underwater image quality measure (UIQM) is used to measure color, sharpness, and contrast for both of the location images. It is observed that the proposed technique has shown overwhelming performance compared to other deep-based enhancement networks and traditional techniques in an adaptive manner.

Keywords: underwater drone imagery, pixel normalization, thresholding, masking, unsharp mask filter

Procedia PDF Downloads 161
8688 Evaluation of Spatial Correlation Length and Karhunen-Loeve Expansion Terms for Predicting Reliability Level of Long-Term Settlement in Soft Soils

Authors: Mehrnaz Alibeikloo, Hadi Khabbaz, Behzad Fatahi

Abstract:

The spectral random field method is one of the widely used methods to obtain more reliable and accurate results in geotechnical problems involving material variability. Karhunen-Loeve (K-L) expansion method was applied to perform random field discretization of cross-correlated creep parameters. Karhunen-Loeve expansion method is based on eigenfunctions and eigenvalues of covariance function adopting Kernel integral solution. In this paper, the accuracy of Karhunen-Loeve expansion was investigated to predict long-term settlement of soft soils adopting elastic visco-plastic creep model. For this purpose, a parametric study was carried to evaluate the effect of K-L expansion terms and spatial correlation length on the reliability of results. The results indicate that small values of spatial correlation length require more K-L expansion terms. Moreover, by increasing spatial correlation length, the coefficient of variation (COV) of creep settlement increases, confirming more conservative and safer prediction.

Keywords: Karhunen-Loeve expansion, long-term settlement, reliability analysis, spatial correlation length

Procedia PDF Downloads 131
8687 3D Printing Perceptual Models of Preference Using a Fuzzy Extreme Learning Machine Approach

Authors: Xinyi Le

Abstract:

In this paper, 3D printing orientations were determined through our perceptual model. Some FDM (Fused Deposition Modeling) 3D printers, which are widely used in universities and industries, often require support structures during the additive manufacturing. After removing the residual material, some surface artifacts remain at the contact points. These artifacts will damage the function and visual effect of the model. To prevent the impact of these artifacts, we present a fuzzy extreme learning machine approach to find printing directions that avoid placing supports in perceptually significant regions. The proposed approach is able to solve the evaluation problem by combing both the subjective knowledge and objective information. Our method combines the advantages of fuzzy theory, auto-encoders, and extreme learning machine. Fuzzy set theory is applied for dealing with subjective preference information, and auto-encoder step is used to extract good features without supervised labels before extreme learning machine. An extreme learning machine method is then developed successfully for training and learning perceptual models. The performance of this perceptual model will be demonstrated on both natural and man-made objects. It is a good human-computer interaction practice which draws from supporting knowledge on both the machine side and the human side.

Keywords: 3d printing, perceptual model, fuzzy evaluation, data-driven approach

Procedia PDF Downloads 403
8686 Implication of the Exchange-Correlation on Electromagnetic Wave Propagation in Single-Wall Carbon Nanotubes

Authors: A. Abdikian

Abstract:

Using the linearized quantum hydrodynamic model (QHD) and by considering the role of quantum parameter (Bohm’s potential) and electron exchange-correlation potential in conjunction with Maxwell’s equations, electromagnetic wave propagation in a single-walled carbon nanotubes was studied. The electronic excitations are described. By solving the mentioned equations with appropriate boundary conditions and by assuming the low-frequency electromagnetic waves, two general expressions of dispersion relations are derived for the transverse magnetic (TM) and transverse electric (TE) modes, respectively. The dispersion relations are analyzed numerically and it was found that the dependency of dispersion curves with the exchange-correlation effects (which have been ignored in previous works) in the low frequency would be limited. Moreover, it has been realized that asymptotic behaviors of the TE and TM modes are similar in single wall carbon nanotubes (SWCNTs). The results show that by adding the function of electron exchange-correlation potential lead to the phenomena and make to extend the validity range of QHD model. The results can be important in the study of collective phenomena in nanostructures.

Keywords: transverse magnetic, transverse electric, quantum hydrodynamic model, electron exchange-correlation potential, single-wall carbon nanotubes

Procedia PDF Downloads 425
8685 Attempt to Reuse Used-PCs as Distributed Storage

Authors: Toshiya Kawato, Shin-ichi Motomura, Masayuki Higashino, Takao Kawamura

Abstract:

Storage for storing data is indispensable. If a storage capacity becomes insufficient, we can increase its capacity by adding new disks. It is, however, difficult to add a new disk when a budget is not enough. On the other hand, there are many unused idle resources such as used personal computers despite those use value. In order to solve those problems, used personal computers can be reused as storage. In this paper, we attempt to reuse used-PCs as a distributed storage. First, we list up the characteristics of used-PCs and design a storage system that utilizes its characteristics. Next, we experimentally implement an auto-construction system that automatically constructs a distributed storage environment in used-PCs.

Keywords: distributed storage, used personal computer, idle resource, auto construction

Procedia PDF Downloads 221
8684 Glycosaminoglycan, a Cartilage Erosion Marker in Synovial Fluid of Osteoarthritis Patients Strongly Correlates with WOMAC Function Subscale

Authors: Priya Kulkarni, Soumya Koppikar, Narendrakumar Wagh, Dhanshri Ingle, Onkar Lande, Abhay Harsulkar

Abstract:

Cartilage is an extracellular matrix composed of aggrecan, which imparts it with a great tensile strength, stiffness and resilience. Disruption in cartilage metabolism leading to progressive degeneration is a characteristic feature of Osteoarthritis (OA). The process involves enzymatic depolymerisation of cartilage specific proteoglycan, releasing free glycosaminoglycan (GAG). This released GAG in synovial fluid (SF) of knee joint serves as a direct measure of cartilage loss, however, limited due to its invasive nature. Western Ontario and McMaster Universities Arthritis Index (WOMAC) is widely used for assessing pain, stiffness and physical-functions in OA patients. The scale is comprised of three subscales namely, pain, stiffness and physical-function, intends to measure patient’s perspective of disease severity as well as efficacy of prescribed treatment. Twenty SF samples obtained from OA patients were analysed for their GAG values in SF using DMMB based assay. LK 1.0 vernacular version was used to attain WOMAC scale. The results were evaluated using SAS University software (Edition 1.0) for statistical significance. All OA patients revealed higher GAG values compared to the control value of 78.4±30.1µg/ml (obtained from our non-OA patients). Average WOMAC calculated was 51.3 while pain, stiffness and function estimated were 9.7, 3.9 and 37.7, respectively. Interestingly, a strong statistical correlation was established between WOMAC function subscale and GAG (p = 0.0102). This subscale is based on day-to-day activities like stair-use, bending, walking, getting in/out of car, rising from bed. However, pain and stiffness subscale did not show correlation with any of the studied markers and endorsed the atypical inflammation in OA pathology. On one side, where knee pain showed poor correlation with GAG, it is often noted that radiography is insensitive to cartilage degenerative changes; thus OA remains undiagnosed for long. Moreover, active cartilage degradation phase remains elusive to both, patient and clinician. Through analysis of large number of OA patients we have established a close association of Kellgren-Lawrence grades and increased cartilage loss. A direct attempt to correlate WOMAC and radiographic progression of OA with various biomarkers has not been attempted so far. We found a good correlation in GAG levels in SF and the function subscale.

Keywords: cartilage, Glycosaminoglycan, synovial fluid, western ontario and McMaster Universities Arthritis Index

Procedia PDF Downloads 418
8683 Hydrographic Mapping Based on the Concept of Fluvial-Geomorphological Auto-Classification

Authors: Jesús Horacio, Alfredo Ollero, Víctor Bouzas-Blanco, Augusto Pérez-Alberti

Abstract:

Rivers have traditionally been classified, assessed and managed in terms of hydrological, chemical and / or biological criteria. Geomorphological classifications had in the past a secondary role, although proposals like River Styles Framework, Catchment Baseline Survey or Stroud Rural Sustainable Drainage Project did incorporate geomorphology for management decision-making. In recent years many studies have been attracted to the geomorphological component. The geomorphological processes and their associated forms determine the structure of a river system. Understanding these processes and forms is a critical component of the sustainable rehabilitation of aquatic ecosystems. The fluvial auto-classification approach suggests that a river is a self-built natural system, with processes and forms designed to effectively preserve their ecological function (hydrologic, sedimentological and biological regime). Fluvial systems are formed by a wide range of elements with multiple non-linear interactions on different spatial and temporal scales. Besides, the fluvial auto-classification concept is built using data from the river itself, so that each classification developed is peculiar to the river studied. The variables used in the classification are specific stream power and mean grain size. A discriminant analysis showed that these variables are the best characterized processes and forms. The statistical technique applied allows to get an individual discriminant equation for each geomorphological type. The geomorphological classification was developed using sites with high naturalness. Each site is a control point of high ecological and geomorphological quality. The changes in the conditions of the control points will be quickly recognizable, and easy to apply a right management measures to recover the geomorphological type. The study focused on Galicia (NW Spain) and the mapping was made analyzing 122 control points (sites) distributed over eight river basins. In sum, this study provides a method for fluvial geomorphological classification that works as an open and flexible tool underlying the fluvial auto-classification concept. The hydrographic mapping is the visual expression of the results, such that each river has a particular map according to its geomorphological characteristics. Each geomorphological type is represented by a particular type of hydraulic geometry (channel width, width-depth ratio, hydraulic radius, etc.). An alteration of this geometry is indicative of a geomorphological disturbance (whether natural or anthropogenic). Hydrographic mapping is also dynamic because its meaning changes if there is a modification in the specific stream power and/or the mean grain size, that is, in the value of their equations. The researcher has to check annually some of the control points. This procedure allows to monitor the geomorphology quality of the rivers and to see if there are any alterations. The maps are useful to researchers and managers, especially for conservation work and river restoration.

Keywords: fluvial auto-classification concept, mapping, geomorphology, river

Procedia PDF Downloads 346
8682 The Complex Relationship Between IQ and Attention Deficit Hyperactivity Disorder Symptoms: Insights From Behaviors, Cognition, and Brain in 5,138 Children With Attention Deficit Hyperactivity Disorder

Authors: Ningning Liu, Gaoding Jia, Yinshan Wang, Haimei Li, Xinian Zuo, Yufeng Wang, Lu Liu, Qiujin Qian

Abstract:

Background: There has been speculation that a high IQ may not necessarily provide protection against attention deficit hyperactivity disorder (ADHD), and there may be a U-shaped correlation between IQ and ADHD symptoms. However, this speculation has not been validated in the ADHD population in any study so far. Method: We conducted a study with 5,138 children who have been professionally diagnosed with ADHD and have a wide range of IQ levels. General Linear Models were used to determine the optimal model between IQ and ADHD core symptoms with sex and age as covariates. The ADHD symptoms we looked at included the total scores (TO), inattention (IA) and hyperactivity/impulsivity (HI). Wechsler Intelligence scale were used to assess IQ [Full-Scale IQ (FSIQ), Verbal IQ (VIQ), and Performance IQ (PIQ)]. Furthermore, we examined the correlation between IQ and the execution function [Behavior Rating Inventory of Executive Function (BRIEF)], as well as between IQ and brain surface area, to determine if the associations between IQ and ADHD symptoms are reflected in executive functions and brain structure. Results: Consistent with previous research, the results indicated that FSIQ and VIQ both showed a linear negative correlation with the TO and IA scores of ADHD. However, PIQ showed an inverted U-shaped relationship with the TO and HI scores of ADHD, with 103 as the peak point. These findings were also partially reflected in the relationship between IQ and executive functions, as well as IQ and brain surface area. Conclusion: To sum up, the relationship between IQ and ADHD symptoms is not straightforward. Our study confirms long-standing academic hypotheses and finds that PIQ exhibits an inverted U-shaped relationship with ADHD symptoms. This study enhances our understanding of symptoms and behaviors of ADHD with varying IQ characteristics and provides some evidence for targeted clinical intervention.

Keywords: ADHD, IQ, execution function, brain imaging

Procedia PDF Downloads 19
8681 The Influence of Applying Mechanical Chest Compression Systems on the Effectiveness of Cardiopulmonary Resuscitation in Out-of-Hospital Cardiac Arrest

Authors: Slawomir Pilip, Michal Wasilewski, Daniel Celinski, Leszek Szpakowski, Grzegorz Michalak

Abstract:

The aim of the study was to evaluate the effectiveness of cardiopulmonary resuscitation taken by Medical Emergency Teams (MET) at the place of an accident including the usage of mechanical chest compression systems. In the period of January-May 2017, there were 137 cases of a sudden cardiac arrest in a chosen region of Eastern Poland with 360.000 inhabitants. Medical records and questionnaires filled by METs were analysed to prove the effectiveness of cardiopulmonary resuscitations that were considered to be effective when an early indication of spontaneous circulation was provided and the patient was taken to hospital. A chest compression system used by METs was applied in 60 cases (Lucas3 - 34 patients; Auto Pulse - 24 patients). The effectiveness of cardiopulmonary resuscitation among patients who were employed a chest compression system was much higher (43,3%) than the manual cardiac massage (36,4%). Thus, the usage of Lucas3 chest compression system resulted in 47% while Auto Pulse was 33,3%. The average ambulance arrival time could have had a significant impact on the subsequent effectiveness of cardiopulmonary resuscitation in these cases. Ambulances equipped with Lucas3 reached the destination within 8 minutes, and those with Auto Pulse needed 12,1 minutes. Moreover, taking effective basic life support (BLS) by bystanders before the ambulance arrival was much more frequent for ambulances with Lucas3 than Auto Pulse. Therefore, the percentage of BLS among the group of patients who were employed Lucas3 by METs was 26,5%, and 20,8% for Auto Pulse. The total percentage of taking BLS by bystanders before the ambulance arrival resulted in 25% of patients who were later applied a chest compression system by METs. Not only was shockable cardiac rhythm obtained in 47% of these cases, but an early indication of spontaneous circulation was also provided in all these patients. Both Lucas3 and Auto Pulse were evaluated to be significantly useful in improving the effectiveness of cardiopulmonary resuscitation by 97% of Medical Emergency Teams. Therefore, implementation of chest compression systems essentially makes the cardiopulmonary resuscitation even more effective. The ambulance arrival time, taking successful BLS by bystanders before the ambulance arrival and the presence of shockable cardiac rhythm determine an early indication of spontaneous circulation among patients after a sudden cardiac arrest.

Keywords: cardiac arrest, effectiveness, mechanical chest compression systems, resuscitation

Procedia PDF Downloads 224
8680 Deep Learning Based on Image Decomposition for Restoration of Intrinsic Representation

Authors: Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Kensuke Nakamura, Dongeun Choi, Byung-Woo Hong

Abstract:

Artefacts are commonly encountered in the imaging process of clinical computed tomography (CT) where the artefact refers to any systematic discrepancy between the reconstructed observation and the true attenuation coefficient of the object. It is known that CT images are inherently more prone to artefacts due to its image formation process where a large number of independent detectors are involved, and they are assumed to yield consistent measurements. There are a number of different artefact types including noise, beam hardening, scatter, pseudo-enhancement, motion, helical, ring, and metal artefacts, which cause serious difficulties in reading images. Thus, it is desired to remove nuisance factors from the degraded image leaving the fundamental intrinsic information that can provide better interpretation of the anatomical and pathological characteristics. However, it is considered as a difficult task due to the high dimensionality and variability of data to be recovered, which naturally motivates the use of machine learning techniques. We propose an image restoration algorithm based on the deep neural network framework where the denoising auto-encoders are stacked building multiple layers. The denoising auto-encoder is a variant of a classical auto-encoder that takes an input data and maps it to a hidden representation through a deterministic mapping using a non-linear activation function. The latent representation is then mapped back into a reconstruction the size of which is the same as the size of the input data. The reconstruction error can be measured by the traditional squared error assuming the residual follows a normal distribution. In addition to the designed loss function, an effective regularization scheme using residual-driven dropout determined based on the gradient at each layer. The optimal weights are computed by the classical stochastic gradient descent algorithm combined with the back-propagation algorithm. In our algorithm, we initially decompose an input image into its intrinsic representation and the nuisance factors including artefacts based on the classical Total Variation problem that can be efficiently optimized by the convex optimization algorithm such as primal-dual method. The intrinsic forms of the input images are provided to the deep denosing auto-encoders with their original forms in the training phase. In the testing phase, a given image is first decomposed into the intrinsic form and then provided to the trained network to obtain its reconstruction. We apply our algorithm to the restoration of the corrupted CT images by the artefacts. It is shown that our algorithm improves the readability and enhances the anatomical and pathological properties of the object. The quantitative evaluation is performed in terms of the PSNR, and the qualitative evaluation provides significant improvement in reading images despite degrading artefacts. The experimental results indicate the potential of our algorithm as a prior solution to the image interpretation tasks in a variety of medical imaging applications. This work was supported by the MISP(Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by the IITP(Institute for Information and Communications Technology Promotion).

Keywords: auto-encoder neural network, CT image artefact, deep learning, intrinsic image representation, noise reduction, total variation

Procedia PDF Downloads 165
8679 Prediction of Formation Pressure Using Artificial Intelligence Techniques

Authors: Abdulmalek Ahmed

Abstract:

Formation pressure is the main function that affects drilling operation economically and efficiently. Knowing the pore pressure and the parameters that affect it will help to reduce the cost of drilling process. Many empirical models reported in the literature were used to calculate the formation pressure based on different parameters. Some of these models used only drilling parameters to estimate pore pressure. Other models predicted the formation pressure based on log data. All of these models required different trends such as normal or abnormal to predict the pore pressure. Few researchers applied artificial intelligence (AI) techniques to predict the formation pressure by only one method or a maximum of two methods of AI. The objective of this research is to predict the pore pressure based on both drilling parameters and log data namely; weight on bit, rotary speed, rate of penetration, mud weight, bulk density, porosity and delta sonic time. A real field data is used to predict the formation pressure using five different artificial intelligence (AI) methods such as; artificial neural networks (ANN), radial basis function (RBF), fuzzy logic (FL), support vector machine (SVM) and functional networks (FN). All AI tools were compared with different empirical models. AI methods estimated the formation pressure by a high accuracy (high correlation coefficient and low average absolute percentage error) and outperformed all previous. The advantage of the new technique is its simplicity, which represented from its estimation of pore pressure without the need of different trends as compared to other models which require a two different trend (normal or abnormal pressure). Moreover, by comparing the AI tools with each other, the results indicate that SVM has the advantage of pore pressure prediction by its fast processing speed and high performance (a high correlation coefficient of 0.997 and a low average absolute percentage error of 0.14%). In the end, a new empirical correlation for formation pressure was developed using ANN method that can estimate pore pressure with a high precision (correlation coefficient of 0.998 and average absolute percentage error of 0.17%).

Keywords: Artificial Intelligence (AI), Formation pressure, Artificial Neural Networks (ANN), Fuzzy Logic (FL), Support Vector Machine (SVM), Functional Networks (FN), Radial Basis Function (RBF)

Procedia PDF Downloads 126
8678 Correlation between Potential Intelligence Explanatory Study in the Perspective of Multiple Intelligence Theory by Using Dermatoglyphics and Culture Approaches

Authors: Efnie Indrianie

Abstract:

Potential Intelligence constitutes one essential factor in every individual. This intelligence can be a provision for the development of Performance Intelligence if it is supported by surrounding environment. Fingerprint analysis is a method in recognizing this Potential Intelligence. This method is grounded on pattern and number of finger print outlines that are assumed symmetrical with the number of nerves in our brain, in which these areas have their own function among another. These brain’s functions are later being transposed into intelligence components in accordance with the Multiple Intelligences theory. This research tested the correlation between Potential Intelligence and the components of its Performance Intelligence. Statistical test results that used Pearson correlation showed that five components of Potential Intelligence correlated with Performance Intelligence. Those five components are Logic-Math, Logic, Linguistic, Music, Kinesthetic, and Intrapersonal. Also, this research indicated that cultural factor had a big role in shaping intelligence.

Keywords: potential intelligence, performance intelligence, multiple intelligences, fingerprint, environment, brain

Procedia PDF Downloads 496
8677 A Normalized Non-Stationary Wavelet Based Analysis Approach for a Computer Assisted Classification of Laryngoscopic High-Speed Video Recordings

Authors: Mona K. Fehling, Jakob Unger, Dietmar J. Hecker, Bernhard Schick, Joerg Lohscheller

Abstract:

Voice disorders origin from disturbances of the vibration patterns of the two vocal folds located within the human larynx. Consequently, the visual examination of vocal fold vibrations is an integral part within the clinical diagnostic process. For an objective analysis of the vocal fold vibration patterns, the two-dimensional vocal fold dynamics are captured during sustained phonation using an endoscopic high-speed camera. In this work, we present an approach allowing a fully automatic analysis of the high-speed video data including a computerized classification of healthy and pathological voices. The approach bases on a wavelet-based analysis of so-called phonovibrograms (PVG), which are extracted from the high-speed videos and comprise the entire two-dimensional vibration pattern of each vocal fold individually. Using a principal component analysis (PCA) strategy a low-dimensional feature set is computed from each phonovibrogram. From the PCA-space clinically relevant measures can be derived that quantify objectively vibration abnormalities. In the first part of the work it will be shown that, using a machine learning approach, the derived measures are suitable to distinguish automatically between healthy and pathological voices. Within the approach the formation of the PCA-space and consequently the extracted quantitative measures depend on the clinical data, which were used to compute the principle components. Therefore, in the second part of the work we proposed a strategy to achieve a normalization of the PCA-space by registering the PCA-space to a coordinate system using a set of synthetically generated vibration patterns. The results show that owing to the normalization step potential ambiguousness of the parameter space can be eliminated. The normalization further allows a direct comparison of research results, which bases on PCA-spaces obtained from different clinical subjects.

Keywords: Wavelet-based analysis, Multiscale product, normalization, computer assisted classification, high-speed laryngoscopy, vocal fold analysis, phonovibrogram

Procedia PDF Downloads 236
8676 Study on Optimal Control Strategy of PM2.5 in Wuhan, China

Authors: Qiuling Xie, Shanliang Zhu, Zongdi Sun

Abstract:

In this paper, we analyzed the correlation relationship among PM2.5 from other five Air Quality Indices (AQIs) based on the grey relational degree, and built a multivariate nonlinear regression equation model of PM2.5 and the five monitoring indexes. For the optimal control problem of PM2.5, we took the partial large Cauchy distribution of membership equation as satisfaction function. We established a nonlinear programming model with the goal of maximum performance to price ratio. And the optimal control scheme is given.

Keywords: grey relational degree, multiple linear regression, membership function, nonlinear programming

Procedia PDF Downloads 261
8675 Spatial Point Process Analysis of Dengue Fever in Tainan, Taiwan

Authors: Ya-Mei Chang

Abstract:

This research is intended to apply spatio-temporal point process methods to the dengue fever data in Tainan. The spatio-temporal intensity function of the dataset is assumed to be separable. The kernel estimation is a widely used approach to estimate intensity functions. The intensity function is very helpful to study the relation of the spatio-temporal point process and some covariates. The covariate effects might be nonlinear. An nonparametric smoothing estimator is used to detect the nonlinearity of the covariate effects. A fitted parametric model could describe the influence of the covariates to the dengue fever. The correlation between the data points is detected by the K-function. The result of this research could provide useful information to help the government or the stakeholders making decisions.

Keywords: dengue fever, spatial point process, kernel estimation, covariate effect

Procedia PDF Downloads 324
8674 A Deterministic Approach for Solving the Hull and White Interest Rate Model with Jump Process

Authors: Hong-Ming Chen

Abstract:

This work considers the resolution of the Hull and White interest rate model with the jump process. A deterministic process is adopted to model the random behavior of interest rate variation as deterministic perturbations, which is depending on the time t. The Brownian motion and jumps uncertainty are denoted as the integral functions piecewise constant function w(t) and point function θ(t). It shows that the interest rate function and the yield function of the Hull and White interest rate model with jump process can be obtained by solving a nonlinear semi-infinite programming problem. A relaxed cutting plane algorithm is then proposed for solving the resulting optimization problem. The method is calibrated for the U.S. treasury securities at 3-month data and is used to analyze several effects on interest rate prices, including interest rate variability, and the negative correlation between stock returns and interest rates. The numerical results illustrate that our approach essentially generates the yield functions with minimal fitting errors and small oscillation.

Keywords: optimization, interest rate model, jump process, deterministic

Procedia PDF Downloads 131
8673 On a Univalent Function and the Integral Means of Its Derivative

Authors: Shatha S. Alhily

Abstract:

The purpose of this research paper is to show all the possible values of the pth power of the integrable function which make the integral means of the derivative of univalent function existing and finite.

Keywords: derivative, integral means, self conformal maps, univalent function

Procedia PDF Downloads 597
8672 Artificial Neural Networks Application on Nusselt Number and Pressure Drop Prediction in Triangular Corrugated Plate Heat Exchanger

Authors: Hany Elsaid Fawaz Abdallah

Abstract:

This study presents a new artificial neural network(ANN) model to predict the Nusselt Number and pressure drop for the turbulent flow in a triangular corrugated plate heat exchanger for forced air and turbulent water flow. An experimental investigation was performed to create a new dataset for the Nusselt Number and pressure drop values in the following range of dimensionless parameters: The plate corrugation angles (from 0° to 60°), the Reynolds number (from 10000 to 40000), pitch to height ratio (from 1 to 4), and Prandtl number (from 0.7 to 200). Based on the ANN performance graph, the three-layer structure with {12-8-6} hidden neurons has been chosen. The training procedure includes back-propagation with the biases and weight adjustment, the evaluation of the loss function for the training and validation dataset and feed-forward propagation of the input parameters. The linear function was used at the output layer as the activation function, while for the hidden layers, the rectified linear unit activation function was utilized. In order to accelerate the ANN training, the loss function minimization may be achieved by the adaptive moment estimation algorithm (ADAM). The ‘‘MinMax’’ normalization approach was utilized to avoid the increase in the training time due to drastic differences in the loss function gradients with respect to the values of weights. Since the test dataset is not being used for the ANN training, a cross-validation technique is applied to the ANN network using the new data. Such procedure was repeated until loss function convergence was achieved or for 4000 epochs with a batch size of 200 points. The program code was written in Python 3.0 using open-source ANN libraries such as Scikit learn, TensorFlow and Keras libraries. The mean average percent error values of 9.4% for the Nusselt number and 8.2% for pressure drop for the ANN model have been achieved. Therefore, higher accuracy compared to the generalized correlations was achieved. The performance validation of the obtained model was based on a comparison of predicted data with the experimental results yielding excellent accuracy.

Keywords: artificial neural networks, corrugated channel, heat transfer enhancement, Nusselt number, pressure drop, generalized correlations

Procedia PDF Downloads 52
8671 Multi-Objectives Genetic Algorithm for Optimizing Machining Process Parameters

Authors: Dylan Santos De Pinho, Nabil Ouerhani

Abstract:

Energy consumption of machine-tools is becoming critical for machine-tool builders and end-users because of economic, ecological and legislation-related reasons. Many machine-tool builders are seeking for solutions that allow the reduction of energy consumption of machine-tools while preserving the same productivity rate and the same quality of machined parts. In this paper, we present the first results of a project conducted jointly by academic and industrial partners to reduce the energy consumption of a Swiss-Type lathe. We employ genetic algorithms to find optimal machining parameters – the set of parameters that lead to the best trade-off between energy consumption, part quality and tool lifetime. Three main machining process parameters are considered in our optimization technique, namely depth of cut, spindle rotation speed and material feed rate. These machining process parameters have been identified as the most influential ones in the configuration of the Swiss-type machining process. A state-of-the-art multi-objective genetic algorithm has been used. The algorithm combines three fitness functions, which are objective functions that permit to evaluate a set of parameters against the three objectives: energy consumption, quality of the machined parts, and tool lifetime. In this paper, we focus on the investigation of the fitness function related to energy consumption. Four different energy consumption related fitness functions have been investigated and compared. The first fitness function refers to the Kienzle cutting force model. The second fitness function uses the Material Removal Rate (RMM) as an indicator of energy consumption. The two other fitness functions are non-deterministic, learning-based functions. One fitness function uses a simple Neural Network to learn the relation between the process parameters and the energy consumption from experimental data. Another fitness function uses Lasso regression to determine the same relation. The goal is, then, to find out which fitness functions predict best the energy consumption of a Swiss-Type machining process for the given set of machining process parameters. Once determined, these functions may be used for optimization purposes – determine the optimal machining process parameters leading to minimum energy consumption. The performance of the four fitness functions has been evaluated. The Tornos DT13 Swiss-Type Lathe has been used to carry out the experiments. A mechanical part including various Swiss-Type machining operations has been selected for the experiments. The evaluation process starts with generating a set of CNC (Computer Numerical Control) programs for machining the part at hand. Each CNC program considers a different set of machining process parameters. During the machining process, the power consumption of the spindle is measured. All collected data are assigned to the appropriate CNC program and thus to the set of machining process parameters. The evaluation approach consists in calculating the correlation between the normalized measured power consumption and the normalized power consumption prediction for each of the four fitness functions. The evaluation shows that the Lasso and Neural Network fitness functions have the highest correlation coefficient with 97%. The fitness function “Material Removal Rate” (MRR) has a correlation coefficient of 90%, whereas the Kienzle-based fitness function has a correlation coefficient of 80%.

Keywords: adaptive machining, genetic algorithms, smart manufacturing, parameters optimization

Procedia PDF Downloads 119
8670 Constructing a Bayesian Network for Solar Energy in Egypt Using Life Cycle Analysis and Machine Learning Algorithms

Authors: Rawaa H. El-Bidweihy, Hisham M. Abdelsalam, Ihab A. El-Khodary

Abstract:

In an era where machines run and shape our world, the need for a stable, non-ending source of energy emerges. In this study, the focus was on the solar energy in Egypt as a renewable source, the most important factors that could affect the solar energy’s market share throughout its life cycle production were analyzed and filtered, the relationships between them were derived before structuring a Bayesian network. Also, forecasted models were built for multiple factors to predict the states in Egypt by 2035, based on historical data and patterns, to be used as the nodes’ states in the network. 37 factors were found to might have an impact on the use of solar energy and then were deducted to 12 factors that were chosen to be the most effective to the solar energy’s life cycle in Egypt, based on surveying experts and data analysis, some of the factors were found to be recurring in multiple stages. The presented Bayesian network could be used later for scenario and decision analysis of using solar energy in Egypt, as a stable renewable source for generating any type of energy needed.

Keywords: ARIMA, auto correlation, Bayesian network, forecasting models, life cycle, partial correlation, renewable energy, SARIMA, solar energy

Procedia PDF Downloads 116
8669 A Neural Network Classifier for Identifying Duplicate Image Entries in Real-Estate Databases

Authors: Sergey Ermolin, Olga Ermolin

Abstract:

A Deep Convolution Neural Network with Triplet Loss is used to identify duplicate images in real-estate advertisements in the presence of image artifacts such as watermarking, cropping, hue/brightness adjustment, and others. The effects of batch normalization, spatial dropout, and various convergence methodologies on the resulting detection accuracy are discussed. For comparative Return-on-Investment study (per industry request), end-2-end performance is benchmarked on both Nvidia Titan GPUs and Intel’s Xeon CPUs. A new real-estate dataset from San Francisco Bay Area is used for this work. Sufficient duplicate detection accuracy is achieved to supplement other database-grounded methods of duplicate removal. The implemented method is used in a Proof-of-Concept project in the real-estate industry.

Keywords: visual recognition, convolutional neural networks, triplet loss, spatial batch normalization with dropout, duplicate removal, advertisement technologies, performance benchmarking

Procedia PDF Downloads 305
8668 Stress Perception, Social Supports and Family Function among Military Inpatients with Adjustment Disorders in Taiwan

Authors: Huey-Fang Sun, Wei-Kai Weng, Mei-Kuang Chao, Hui-Shan Hsu, Tsai-Yin Shih

Abstract:

Psycho-social stress is important for mental illness and the presence of emotional and behavioral symptoms to an identifiable event is the central feature of adjustment disorders. However, whether patients with adjustment disorders have been raised in family with poor family functions and social supports and have higher stress perception than their peer group when they both experienced a similar stressful environment remains unknown. The specific aims of the study are to investigate the correlation among the family function, social supports and the level of stress perception and to test the hypothesis that military patients with adjustment disorders would have lower family function, lower social supports and higher stress perception than their healthy colleagues recruited in the same cohort for military services given their common exposure to similar stressful environments. Methods: The study was conducted in four hospitals of northern part of Taiwan from July 1, 2015 to June 30, 2017 and a matched case-control study design was used. The inclusion criteria for potential patient participants were psychiatric inpatients that serviced in military during the study period and met the diagnosis of adjustment disorders. Patients who had been admitted to psychiatric ward before or had illiteracy problem were excluded. A healthy military control sample matched by the same military service unit, gender, and recruited cohort was invited to participate the study as well. Totally 74 participants (37 patients and 37 controls) completed the consent forms and filled out the research questionnaires. Questionnaires used in the study included Perceived Stress Scale (PSS) as a measure of stress perception; Family APGAR as a measure of family function, and Multidimensional Scale of Perceived Social Support (MSPSS) as a measure of social supports. Pearson correlation analysis and t-test were applied for statistical analysis. Results: The analysis results showed that PSS level significantly negatively correlated with three social support subscales (family subscale, r= -.37, P < .05; friend subscale, r= -.38, P < .05; significant other subscale, r= -.39, P < .05). A negative correlation between PSS level and Family APGAR only reached a borderline significant level (P= .06). The t-test results for PSS scores, Family APGAR levels, and three subscale scores of MSPSS between patient and control participants were all significantly different (P < .001, P < .05, P < .05, P < .05, P < .05, respectively) and the patient participants had higher stress perception scores, lower social supports and lower family function scores than the healthy control participants. Conclusions: Our study suggested that family function and social supports were negatively correlated with patients’ subjective stress perception. Military patients with adjustment disorders tended to have higher stress perception and lower family function and social supports than those military peers who remained healthy and still provided services in their military units.

Keywords: adjustment disorders, family function, social support, stress perception

Procedia PDF Downloads 169