Search results for: computational imaging
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3152

Search results for: computational imaging

1772 Numerical Study of Two Mechanical Stirring Systems for Yield Stress Fluid

Authors: Amine Benmoussa, Mebrouk Rebhi, Rahmani Lakhdar

Abstract:

Mechanically agitated vessels are commonly used for various operations within a wide range process in chemical, pharmaceutical, polymer, biochemical, mineral, petroleum industries. Depending on the purpose of the operation carried out in mixer, the best choice for geometry of the tank and agitator type can vary widely. In this paper, the laminar 2D agitation flow and power consumption of viscoplastic fluids with straight and circular gate impellers in a stirring tank is studied by using computational fluid dynamics (CFD), where the velocity profile, the velocity fields and power consumption was analyzed.

Keywords: CFD, mechanical stirring, power consumption, yield stress fluid

Procedia PDF Downloads 331
1771 Nonlinear Dynamic Analysis of Base-Isolated Structures Using a Partitioned Solution Approach and an Exponential Model

Authors: Nicolò Vaiana, Filip C. Filippou, Giorgio Serino

Abstract:

The solution of the nonlinear dynamic equilibrium equations of base-isolated structures adopting a conventional monolithic solution approach, i.e. an implicit single-step time integration method employed with an iteration procedure, and the use of existing nonlinear analytical models, such as differential equation models, to simulate the dynamic behavior of seismic isolators can require a significant computational effort. In order to reduce numerical computations, a partitioned solution method and a one dimensional nonlinear analytical model are presented in this paper. A partitioned solution approach can be easily applied to base-isolated structures in which the base isolation system is much more flexible than the superstructure. Thus, in this work, the explicit conditionally stable central difference method is used to evaluate the base isolation system nonlinear response and the implicit unconditionally stable Newmark’s constant average acceleration method is adopted to predict the superstructure linear response with the benefit in avoiding iterations in each time step of a nonlinear dynamic analysis. The proposed mathematical model is able to simulate the dynamic behavior of seismic isolators without requiring the solution of a nonlinear differential equation, as in the case of widely used differential equation model. The proposed mixed explicit-implicit time integration method and nonlinear exponential model are adopted to analyze a three dimensional seismically isolated structure with a lead rubber bearing system subjected to earthquake excitation. The numerical results show the good accuracy and the significant computational efficiency of the proposed solution approach and analytical model compared to the conventional solution method and mathematical model adopted in this work. Furthermore, the low stiffness value of the base isolation system with lead rubber bearings allows to have a critical time step considerably larger than the imposed ground acceleration time step, thus avoiding stability problems in the proposed mixed method.

Keywords: base-isolated structures, earthquake engineering, mixed time integration, nonlinear exponential model

Procedia PDF Downloads 266
1770 Efficient Filtering of Graph Based Data Using Graph Partitioning

Authors: Nileshkumar Vaishnav, Aditya Tatu

Abstract:

An algebraic framework for processing graph signals axiomatically designates the graph adjacency matrix as the shift operator. In this setup, we often encounter a problem wherein we know the filtered output and the filter coefficients, and need to find out the input graph signal. Solution to this problem using direct approach requires O(N3) operations, where N is the number of vertices in graph. In this paper, we adapt the spectral graph partitioning method for partitioning of graphs and use it to reduce the computational cost of the filtering problem. We use the example of denoising of the temperature data to illustrate the efficacy of the approach.

Keywords: graph signal processing, graph partitioning, inverse filtering on graphs, algebraic signal processing

Procedia PDF Downloads 298
1769 Downhole Logging and Dynamics Data Resolving Lithology-Related Drilling Behavior

Authors: Christopher Viens, Steve Krase

Abstract:

Terms such as “riding a hard streak”, “formation push”, and “fighting formation” are commonly used in the directional drilling world to explain BHA behavior that causes unwanted trajectory change. Theories about downhole directional tendencies are commonly speculated from various personal experiences with little merit due to the lack of hard data to reveal the actual mechanisms behind the phenomenon, leaving interpretation of the root cause up to personal perception. Understanding and identifying in real time the lithological factors that influence the BHA to change or hold direction adds tremendous value in terms reducing sliding time and targeting zones for optimal ROP. Utilizing surface drilling parameters and employing downhole measurements of azimuthal gamma, continuous inclination, and bending moment, a direct measure of the rock related directional phenomenon have been captured and quantified. Furthermore, identifying continuous zones of like lithology with consistent bit to rock interaction has value from a reservoir characterization and completions standpoint. The paper will show specific examples of lithology related directional tendencies from the Spraberry and Wolfcamp in the Delaware Basin.

Keywords: Azimuthal gamma imaging, bending moment, continuous inclination, downhole dynamics measurements, high frequency data

Procedia PDF Downloads 274
1768 Solving Extended Linear Complementarity Problems (XLCP) - Wood and Environment

Authors: Liberto Pombal, Christian Dieter Jaekel

Abstract:

The objective of this work is to establish theoretical and numerical conditions for Solving Extended Linear Complementarity Problems (XLCP), with emphasis on the Horizontal Linear Complementarity Problem (HLCP). Two new strategies for solving complementarity problems are presented, using differentiable and penalized functions, which resulted in a natural formalization for the Linear Horizontal case. The computational results of all suggested strategies are also discussed in depth in this paper. The implication in practice allows solving and optimizing, in an innovative way, the (forestry) problems of the value chain of the industrial wood sector in Angola.

Keywords: complementarity, box constrained, optimality conditions, wood and environment

Procedia PDF Downloads 40
1767 A Case Study on Utility of 18FDG-PET/CT Scan in Identifying Active Extra Lymph Nodes and Staging of Breast Cancer

Authors: Farid Risheq, M. Zaid Alrisheq, Shuaa Al-Sadoon, Karim Al-Faqih, Mays Abdulazeez

Abstract:

Breast cancer is the most frequently diagnosed cancer worldwide, and a common cause of death among women. Various conventional anatomical imaging tools are utilized for diagnosis, histological assessment and TNM (Tumor, Node, Metastases) staging of breast cancer. Biopsy of sentinel lymph node is becoming an alternative to the axillary lymph node dissection. Advances in 18-Fluoro-Deoxi-Glucose Positron Emission Tomography/Computed Tomography (18FDG-PET/CT) imaging have facilitated breast cancer diagnosis utilizing biological trapping of 18FDG inside lesion cells, expressed as Standardized Uptake Value (SUVmax). Objective: To present the utility of 18FDG uptake PET/CT scans in detecting active extra lymph nodes and distant occult metastases for breast cancer staging. Subjects and Methods: Four female patients were presented with initially classified TNM stages of breast cancer based on conventional anatomical diagnostic techniques. 18FDG-PET/CT scans were performed one hour post 18FDG intra-venous injection of (300-370) MBq, and (7-8) bed/130sec. Transverse, sagittal, and coronal views; fused PET/CT and MIP modality were reconstructed for each patient. Results: A total of twenty four lesions in breast, extended lesions to lung, liver, bone and active extra lymph nodes were detected among patients. The initial TNM stage was significantly changed post 18FDG-PET/CT scan for each patient, as follows: Patient-1: Initial TNM-stage: T1N1M0-(stage I). Finding: Two lesions in right breast (3.2cm2, SUVmax=10.2), (1.8cm2, SUVmax=6.7), associated with metastases to two right axillary lymph nodes. Final TNM-stage: T1N2M0-(stage II). Patient-2: Initial TNM-stage: T2N2M0-(stage III). Finding: Right breast lesion (6.1cm2, SUVmax=15.2), associated with metastases to right internal mammary lymph node, two right axillary lymph nodes, and sclerotic lesions in right scapula. Final TNM-stage: T2N3M1-(stage IV). Patient-3: Initial TNM-stage: T2N0M1-(stage III). Finding: Left breast lesion (11.1cm2, SUVmax=18.8), associated with metastases to two lymph nodes in left hilum, and three lesions in both lungs. Final TNM-stage: T2N2M1-(stage IV). Patient-4: Initial TNM-stage: T4N1M1-(stage III). Finding: Four lesions in upper outer quadrant area of right breast (largest: 12.7cm2, SUVmax=18.6), in addition to one lesion in left breast (4.8cm2, SUVmax=7.1), associated with metastases to multiple lesions in liver (largest: 11.4cm2, SUV=8.0), and two bony-lytic lesions in left scapula and cervicle-1. No evidence of regional or distant lymph node involvement. Final TNM-stage: T4N0M2-(stage IV). Conclusions: Our results demonstrated that 18FDG-PET/CT scans had significantly changed the TNM stages of breast cancer patients. While the T factor was unchanged, N and M factors showed significant variations. A single session of PET/CT scan was effective in detecting active extra lymph nodes and distant occult metastases, which were not identified by conventional diagnostic techniques, and might advantageously replace bone scan, and contrast enhanced CT of chest, abdomen and pelvis. Applying 18FDG-PET/CT scan early in the investigation, might shorten diagnosis time, helps deciding adequate treatment protocol, and could improve patients’ quality of life and survival. Trapping of 18FDG in malignant lesion cells, after a PET/CT scan, increases the retention index (RI%) for a considerable time, which might help localize sentinel lymph node for biopsy using a hand held gamma probe detector. Future work is required to demonstrate its utility.

Keywords: axillary lymph nodes, breast cancer staging, fluorodeoxyglucose positron emission tomography/computed tomography, lymph nodes

Procedia PDF Downloads 297
1766 Identifying the True Extend of Glioblastoma Based on Preoperative FLAIR Images

Authors: B. Shukir, L. Szivos, D. Kis, P. Barzo

Abstract:

Glioblastoma is the most malignant brain tumor. In general, the survival rate varies between (14-18) months. Glioblastoma consists a solid and infiltrative part. The standard therapeutic management of glioblastoma is maximum safe resection followed by chemo-radiotherapy. It’s hypothesized that the pretumoral hyperintense region in fluid attenuated inversion recovery (FLAIR) images includes both vasogenic edema and infiltrated tumor cells. In our study, we aimed to define the sensitivity and specificity of hyperintense FLAIR images preoperatively to examine how well it can define the true extent of glioblastoma. (16) glioblastoma patients included in this study. Hyperintense FLAIR region were delineated preoperatively as tumor mask. The infiltrative part of glioblastoma considered the regions where the tumor recurred on the follow up MRI. The recurrence on the CE-T1 images was marked as the recurrence masks. According to (AAL3) and (JHU white matter labels) atlas, the brain divided into cortical and subcortical regions respectively. For calculating specificity and sensitivity, the FLAIR and the recurrence masks overlapped counting how many regions affected by both . The average sensitivity and specificity was 83% and 85% respectively. Individually, the sensitivity and specificity varied between (31-100)%, and (100-58)% respectively. These results suggest that despite FLAIR being as an effective radiologic imaging tool its prognostic value remains controversial and probabilistic tractography remain more reliable available method for identifying the true extent of glioblastoma.

Keywords: brain tumors, glioblastoma, MRI, FLAIR

Procedia PDF Downloads 29
1765 A Survey of Attacks and Security Requirements in Wireless Sensor Networks

Authors: Vishnu Pratap Singh Kirar

Abstract:

Wireless sensor network (WSN) is a network of many interconnected networked systems, they equipped with energy resources and they are used to detect other physical characteristics. On WSN, there are many researches are performed in past decades. WSN applicable in many security systems govern by military and in many civilian related applications. Thus, the security of WSN gets attention of researchers and gives an opportunity for many future aspects. Still, there are many other issues are related to deployment and overall coverage, scalability, size, energy efficiency, quality of service (QoS), computational power and many more. In this paper we discus about various applications and security related issue and requirements of WSN.

Keywords: wireless sensor network (WSN), wireless network attacks, wireless network security, security requirements

Procedia PDF Downloads 468
1764 Computational Tool for Surface Electromyography Analysis; an Easy Way for Non-Engineers

Authors: Fabiano Araujo Soares, Sauro Emerick Salomoni, Joao Paulo Lima da Silva, Igor Luiz Moura, Adson Ferreira da Rocha

Abstract:

This paper presents a tool developed in the Matlab platform. It was developed to simplify the analysis of surface electromyography signals (S-EMG) in a way accessible to users that are not familiarized with signal processing procedures. The tool receives data by commands in window fields and generates results as graphics and excel tables. The underlying math of each S-EMG estimator is presented. Setup window and result graphics are presented. The tool was presented to four non-engineer users and all of them managed to appropriately use it after a 5 minutes instruction period.

Keywords: S-EMG estimators, electromyography, surface electromyography, ARV, RMS, MDF, MNF, CV

Procedia PDF Downloads 535
1763 Managing Subretinal Bleeds with Intravitreal Aflibercept

Authors: Prachi Abhishek Dave, Abhishek Dave

Abstract:

Purpose: The purpose of this study is to elucidate the role of intravitreal injection Aflibercept in managing complex cases of Wet Age Related Macular Degeneration (ARMD) and the gratifying visual recovery experienced with a minimally invasive procedure. Methods: A 73-year-old gentleman presented with a drop in vision in the left eye for 25 days. On examination, his best corrected visual acuity (BCVA) in the Right eye (OD) was 6/60, and finger counting close to face in the Left eye (OS). On multimodal imaging, he was diagnosed to have a scarred Wet ARMD in OD and an active Wet ARMD with a large subretinal bleed secondary to Wet ARMD in OS. Treatment management options included monotherapy with an Injection Aflibercept or an intravitreal gas injection with tPA followed by Injection Aflibercept. Considering his one-eyed status, the patient decided to go for Aflibercept monotherapy. Results: After 3 monthly injections of injection Aflibercept, the subretinal bleed reduced, the subretinal fluid resolved, and his vision in OS improved to 6/9. He is on a regular follow-up and has not needed any further injections in OS and he maintains 6/9 vision. Conclusions: Conventional treatment guidelines for a large subretinal bleed dictate the use of gas followed by intravitreal Injection Aflibercept. However, gas has its own limitations of causing a rise in intraocular pressure and a transient loss of vision, which is particularly troublesome in one-eyed patients. Injection Aflibercept offers a much safer, less invasive, and elegant treatment option for such patients with equally good or even better visual outcomes.

Keywords: wet ARMD, subretinal bleed, intravitreal injections, aflibercept, EYELEA, intravitreal gas

Procedia PDF Downloads 24
1762 A Comparative Study of Sampling-Based Uncertainty Propagation with First Order Error Analysis and Percentile-Based Optimization

Authors: M. Gulam Kibria, Shourav Ahmed, Kais Zaman

Abstract:

In system analysis, the information on the uncertain input variables cause uncertainty in the system responses. Different probabilistic approaches for uncertainty representation and propagation in such cases exist in the literature. Different uncertainty representation approaches result in different outputs. Some of the approaches might result in a better estimation of system response than the other approaches. The NASA Langley Multidisciplinary Uncertainty Quantification Challenge (MUQC) has posed challenges about uncertainty quantification. Subproblem A, the uncertainty characterization subproblem, of the challenge posed is addressed in this study. In this subproblem, the challenge is to gather knowledge about unknown model inputs which have inherent aleatory and epistemic uncertainties in them with responses (output) of the given computational model. We use two different methodologies to approach the problem. In the first methodology we use sampling-based uncertainty propagation with first order error analysis. In the other approach we place emphasis on the use of Percentile-Based Optimization (PBO). The NASA Langley MUQC’s subproblem A is developed in such a way that both aleatory and epistemic uncertainties need to be managed. The challenge problem classifies each uncertain parameter as belonging to one the following three types: (i) An aleatory uncertainty modeled as a random variable. It has a fixed functional form and known coefficients. This uncertainty cannot be reduced. (ii) An epistemic uncertainty modeled as a fixed but poorly known physical quantity that lies within a given interval. This uncertainty is reducible. (iii) A parameter might be aleatory but sufficient data might not be available to adequately model it as a single random variable. For example, the parameters of a normal variable, e.g., the mean and standard deviation, might not be precisely known but could be assumed to lie within some intervals. It results in a distributional p-box having the physical parameter with an aleatory uncertainty, but the parameters prescribing its mathematical model are subjected to epistemic uncertainties. Each of the parameters of the random variable is an unknown element of a known interval. This uncertainty is reducible. From the study, it is observed that due to practical limitations or computational expense, the sampling is not exhaustive in sampling-based methodology. That is why the sampling-based methodology has high probability of underestimating the output bounds. Therefore, an optimization-based strategy to convert uncertainty described by interval data into a probabilistic framework is necessary. This is achieved in this study by using PBO.

Keywords: aleatory uncertainty, epistemic uncertainty, first order error analysis, uncertainty quantification, percentile-based optimization

Procedia PDF Downloads 224
1761 Theoretical and Experimental Investigation of Fe and Ni-TCNQ on Graphene

Authors: A. Shahsavar, Z. Jakub

Abstract:

Due to the outstanding properties of the 2D metal-organic frameworks (MOF), intensive computational and experimental studies have been done. However, the lack of fundamental studies of MOFs on the graphene backbone is observed. This work studies Fe and Ni as metal and tetracyanoquinodimethane (TCNQ) with a high electron affinity as an organic linker functionalized on graphene. Here we present DFT calculations results to unveil the electronic and magnetic properties of iron and nickel-TCNQ physisorbed on graphene. Adsorption and Fermi energies, structural, and magnetic properties will be reported. Our experimental observations prove Fe- and NiTCNQ@Gr/Ir(111) are thermally highly stable up to 500 and 250°C, respectively, making them promising materials for single-atom catalysts or high-density storage media.

Keywords: DFT, graphene, MTCNQ, self-assembly

Procedia PDF Downloads 96
1760 Case of A Huge Retroperitoneal Abscess Spanning from the Diaphragm to the Pelvic Brim

Authors: Christopher Leung, Tony Kim, Rebecca Lendzion, Scott Mackenzie

Abstract:

Retroperitoneal abscesses are a rare but serious condition with often delayed diagnosis, non-specific symptoms, multiple causes and high morbidity/mortality. With the advent of more readily available cross-sectional imaging, retroperitoneal abscesses are treated earlier and better outcomes are achieved. Occasionally, a retroperitoneal abscess is present as a huge retroperitoneal abscess, as evident in this 53-year-old male. With a background of chronic renal disease and left partial nephrectomy, this gentleman presented with a one-month history of left flank pain without any other symptoms, including fevers or abdominal pain. CT abdomen and pelvis demonstrated a huge retroperitoneal abscess spanning from the diaphragm, abutting the spleen, down to the iliopsoas muscle and abutting the iliac vessels at the pelvic brim. This large retroperitoneal abscess required open drainage as well as drainage by interventional radiology. A long course of intravenous antibiotics and multiple drainages was required to drain the abscess. His blood culture and fluid culture grew Proteus species suggesting a urinary source, likely from his non-functioning kidney, which had a partial nephrectomy. Such a huge retroperitoneal abscess has rarely been described in the literature. The learning point here is that the basic principle of source control and antibiotics is paramount in treating retroperitoneal abscesses regardless of the size of the abscess.

Keywords: retroperitoneal abscess, retroperitoneal mass, sepsis, genitourinary infection

Procedia PDF Downloads 203
1759 Numerical Flow Simulation around HSP Propeller in Open Water and behind a Vessel Wake Using RANS CFD Code

Authors: Kadda Boumediene, Mohamed Bouzit

Abstract:

The prediction of the flow around marine propellers and vessel hulls propeller interaction is one of the challenges of Computational fluid dynamics (CFD). The CFD has emerged as a potential tool in recent years and has promising applications. The objective of the current study is to predict the hydrodynamic performances of HSP marine propeller in open water and behind a vessel. The unsteady 3-D flow was modeled numerically along with respectively the K-ω standard and K-ω SST turbulence models for steady and unsteady cases. The hydrodynamic performances such us a torque and thrust coefficients and efficiency show good agreement with the experiment results.

Keywords: seiun maru propeller, steady, unstead, CFD, HSP

Procedia PDF Downloads 285
1758 Stereoscopic Motion Design: Design Futures

Authors: Edgar Teixeira, Eurico Carrapatoso

Abstract:

As 3D displays become increasingly affordable, while production techniques and computational resources to create stereoscopic content being ever more accessible, a new dimension is literally introduced along with new expressive and immersive potentialities in support of designing for the screen. Prospective design visionaries have already at the reach of their hands an innovative and powerful visualization technology, which enables them to actively envision future trends and vanguardist directions. This paper explores the aesthetic and informational potentialities of stereoscopic motion graphics, providing insight on the application of 3D displays in design practice, proposing strategies to investigate stereoscopic communication, discussing potential repercussions to extant theory and impacts on audience.

Keywords: design, visual communication, technology, stereoscopy, 3D media

Procedia PDF Downloads 387
1757 Design Aspects for Developing a Microfluidics Diagnostics Device Used for Low-Cost Water Quality Monitoring

Authors: Wenyu Guo, Malachy O’Rourke, Mark Bowkett, Michael Gilchrist

Abstract:

Many devices for real-time monitoring of surface water have been developed in the past few years to provide early warning of pollutions and so to decrease the risk of environmental pollution efficiently. One of the most common methodologies used in the detection system is a colorimetric process, in which a container with fixed volume is filled with target ions and reagents to combine a colorimetric dye. The colorimetric ions can sensitively absorb a specific-wavelength radiation beam, and its absorbance rate is proportional to the concentration of the fully developed product, indicating the concentration of target nutrients in the pre-mixed water samples. In order to achieve precise and rapid detection effect, channels with dimensions in the order of micrometers, i.e., microfluidic systems have been developed and introduced into these diagnostics studies. Microfluidics technology largely reduces the surface to volume ratios and decrease the samples/reagents consumption significantly. However, species transport in such miniaturized channels is limited by the low Reynolds numbers in the regimes. Thus, the flow is extremely laminar state, and diffusion is the dominant mass transport process all over the regimes of the microfluidic channels. The objective of this present work has been to analyse the mixing effect and chemistry kinetics in a stop-flow microfluidic device measuring Nitride concentrations in fresh water samples. In order to improve the temporal resolution of the Nitride microfluidic sensor, we have used computational fluid dynamics to investigate the influence that the effectiveness of the mixing process between the sample and reagent within a microfluidic device exerts on the time to completion of the resulting chemical reaction. This computational approach has been complemented by physical experiments. The kinetics of the Griess reaction involving the conversion of sulphanilic acid to a diazonium salt by reaction with nitrite in acidic solution is set in the Laminar Finite-rate chemical reaction in the model. Initially, a methodology was developed to assess the degree of mixing of the sample and reagent within the device. This enabled different designs of the mixing channel to be compared, such as straight, square wave and serpentine geometries. Thereafter, the time to completion of the Griess reaction within a straight mixing channel device was modeled and the reaction time validated with experimental data. Further simulations have been done to compare the reaction time to effective mixing within straight, square wave and serpentine geometries. Results show that square wave channels can significantly improve the mixing effect and provides a low standard deviations of the concentrations of nitride and reagent, while for straight channel microfluidic patterns the corresponding values are 2-3 orders of magnitude greater, and consequently are less efficiently mixed. This has allowed us to design novel channel patterns of micro-mixers with more effective mixing that can be used to detect and monitor levels of nutrients present in water samples, in particular, Nitride. Future generations of water quality monitoring and diagnostic devices will easily exploit this technology.

Keywords: nitride detection, computational fluid dynamics, chemical kinetics, mixing effect

Procedia PDF Downloads 187
1756 Computational Simulations and Assessment of the Application of Non-Circular TAVI Devices

Authors: Jonathon Bailey, Neil Bressloff, Nick Curzen

Abstract:

Transcatheter Aortic Valve Implantation (TAVI) devices are stent-like frames with prosthetic leaflets on the inside, which are percutaneously implanted. The device in a crimped state is fed through the arteries to the aortic root, where the device frame is opened through either self-expansion or balloon expansion, which reveals the prosthetic valve within. The frequency at which TAVI is being used to treat aortic stenosis is rapidly increasing. In time, TAVI is likely to become the favoured treatment over Surgical Valve Replacement (SVR). Mortality after TAVI has been associated with severe Paravalvular Aortic Regurgitation (PAR). PAR occurs when the frame of the TAVI device does not make an effective seal against the internal surface of the aortic root, allowing blood to flow backwards about the valve. PAR is common in patients and has been reported to some degree in as much as 76% of cases. Severe PAR (grade 3 or 4) has been reported in approximately 17% of TAVI patients resulting in post-procedural mortality increases from 6.7% to 16.5%. TAVI devices, like SVR devices, are circular in cross-section as the aortic root is often considered to be approximately circular in shape. In reality, however, the aortic root is often non-circular. The ascending aorta, aortic sino tubular junction, aortic annulus and left ventricular outflow tract have an average ellipticity ratio of 1.07, 1.09, 1.29, and 1.49 respectively. An elliptical aortic root does not severely affect SVR, as the leaflets are completely removed during the surgical procedure. However, an elliptical aortic root can inhibit the ability of the circular Balloon-Expandable (BE) TAVI devices to conform to the interior of the aortic root wall, which increases the risk of PAR. Self-Expanding (SE) TAVI devices are considered better at conforming to elliptical aortic roots, however the valve leaflets were not designed for elliptical function, furthermore the incidence of PAR is greater in SE devices than BE devices (19.8% vs. 12.2% respectively). If a patient’s aortic root is too severely elliptical, they will not be suitable for TAVI, narrowing the treatment options to SVR. It therefore follows that in order to increase the population who can undergo TAVI, and reduce the risk associated with TAVI, non-circular devices should be developed. Computational simulations were employed to further advance our understanding of non-circular TAVI devices. Radial stiffness of the TAVI devices in multiple directions, frame bending stiffness and resistance to balloon induced expansion are all computationally simulated. Finally, a simulation has been developed that demonstrates the expansion of TAVI devices into a non-circular patient specific aortic root model in order to assess the alterations in deployment dynamics, PAR and the stresses induced in the aortic root.

Keywords: tavi, tavr, fea, par, fem

Procedia PDF Downloads 428
1755 Towards Overturning the Dismal Mathematics Performance in Schools by Capitalizing on the Overlooked Cognitive Prowess for Adolescents to Learn Mathematics

Authors: Dudu Ka Ruth Mkhize

Abstract:

Adolescents are at the front and centre of poor mathematics performance in schools. Literature has concluded in some countries that there is a permanent and perpetual mathematics crisis in schools of the persistent poor performance in mathematics by teens. There is no shortage of interventions and research to solve this problem. However, none has capitalised on the cognitive prowess of adolescents, which was revealed at the turn of the century by the introduction of neuroimaging technologies such as structural and functional magnetic resonance imaging (sMRI and fMRI). This research found that brain growth during adolescence results in enhanced cognitive abilities essential for mathematics learning. This paper is based on the four-year case study of rural high school adolescents who had a negative attitude towards mathematics and hence were failing mathematics. But through a ten-day intervention where teaching revolved around invoking their cognitive ability, their attitude and motivation for mathematics changed for the better. The paper concludes that despite educational psychology being part of teacher education as well as education systems, there are numerous overlooked gems of psychological theories which have the potential to enhance academic achievement for youth in schools. A recommendation is made to take cues from positive psychology, whose establishment was a rejection of the dominance of the disease model in psychology. Similarly, the general perspective of poor mathematics performance can take a u-turn towards the cognitive ability acquired by adolescents because of their developmental stage.

Keywords: adolescence, cognitive growth, mathematics performance

Procedia PDF Downloads 50
1754 The Role of Vitamin D Supplementation in Augmenting IFN-γ Production in Response to Mycobacterium Tuberculosis Infection: A Randomized Controlled Trial

Authors: Muhammad Imran Hussain, Ramisha Ibtisam, Tayyaba Fatima, Huba Khalid, Ayesha Aziz, Khansa, Adan Sitara, Anam Shahzad, Aymen Jabeen

Abstract:

Vitamin D supports the immune system fight TB by inhibiting Interferon-gamma (IFN-γ) and lowering host inflammation. The purpose of the research was to see if giving the vitamin D supplements to TB patients affected their prognosis. A randomized placebo control study of 200 TB patients was performed among which 106 received 400,000 IU of injectable vitamin D3 and 94 received placebo for 2 doses. Assessment was carried out at the end of every month for 3 months. IFN-γ responses to whole blood stimulation generated by the Mycobacterium tuberculosis sonicate (MTBs) antigen and early secreted and T cell activated 6 kDa (ESAT6) were assessed at 0 and 12 weeks. The statistical analysis used descriptive statistics (mean and standard deviation), Friedman's test and Fisher's test. The vitamin D group gained significantly more weight (+3.90 pounds) and had less persistent lung disease on imaging (1.33 zones vs. 1.84 zones). They also had a 50% decrease in cavity size. Additionally, patients with low baseline serum concentrations of 25-(OH)D had a significant increase in MTB-induced IFN-γ production after taking vitamin D supplements. Vitamin D administration in large amounts can hasten the recovery of TB patients. The findings point is a therapeutically useful activity of Vitamin D's in the management for tuberculosis.

Keywords: tuberculosis, vitamin D, interferon gamma, protein, infection

Procedia PDF Downloads 37
1753 Segmentation of Liver Using Random Forest Classifier

Authors: Gajendra Kumar Mourya, Dinesh Bhatia, Akash Handique, Sunita Warjri, Syed Achaab Amir

Abstract:

Nowadays, Medical imaging has become an integral part of modern healthcare. Abdominal CT images are an invaluable mean for abdominal organ investigation and have been widely studied in the recent years. Diagnosis of liver pathologies is one of the major areas of current interests in the field of medical image processing and is still an open problem. To deeply study and diagnose the liver, segmentation of liver is done to identify which part of the liver is mostly affected. Manual segmentation of the liver in CT images is time-consuming and suffers from inter- and intra-observer differences. However, automatic or semi-automatic computer aided segmentation of the Liver is a challenging task due to inter-patient Liver shape and size variability. In this paper, we present a technique for automatic segmenting the liver from CT images using Random Forest Classifier. Random forests or random decision forests are an ensemble learning method for classification that operate by constructing a multitude of decision trees at training time and outputting the class that is the mode of the classes of the individual trees. After comparing with various other techniques, it was found that Random Forest Classifier provide a better segmentation results with respect to accuracy and speed. We have done the validation of our results using various techniques and it shows above 89% accuracy in all the cases.

Keywords: CT images, image validation, random forest, segmentation

Procedia PDF Downloads 295
1752 Numerical Study of a 6080HP Open Drip Proof (ODP) Motor

Authors: Feng-Hisang Lai

Abstract:

CFD(Computational Fluid Dynamics) is conducted to numerically study the flow and heat transfer features of a two-pole, 6,080HP, 60Hz, 3,150V open drip-proof (ODP) motor. The stator and rotor cores in this high voltage induction motor are segmented with the use of spacers for cooling purposes, which leads to difficulties in meshing when the entire system is to be simulated. The system is divided into 4 parts, meshed separately and then combined using interfaces. The deviation between the CFD and experimental results in temperature and flow rate is less than 10%. The internal flow is further examined and a final design is proposed to reduce the winding temperature by 10 degrees.

Keywords: CFD, open drip proof, induction motor, cooling

Procedia PDF Downloads 181
1751 In silico Model of Transamination Reaction Mechanism

Authors: Sang-Woo Han, Jong-Shik Shin

Abstract:

w-Transaminase (w-TA) is broadly used for synthesizing chiral amines with a high enantiopurity. However, the reaction mechanism of w-TA has been not well studied, contrary to a-transaminase (a-TA) such as AspTA. Here, we propose in silico model on the reaction mechanism of w-TA. Based on the modeling results which showed large free energy gaps between external aldimine and quinonoid on deamination (or ketimine and quinonoid on amination), withdrawal of Ca-H seemed as a critical step which determines the reaction rate on both amination and deamination reactions, which is consistent with previous researches. Hyperconjugation was also observed in both external aldimine and ketimine which weakens Ca-H bond to elevate Ca-H abstraction.

Keywords: computational modeling, reaction intermediates, w-transaminase, in silico model

Procedia PDF Downloads 528
1750 Impact of Modifying the Surface Materials on the Radiative Heat Transfer Phenomenon

Authors: Arkadiusz Urzędowski, Dorota Wójcicka-Migasiuk, Andrzej Sachajdak, Magdalena Paśnikowska-Łukaszuk

Abstract:

Due to the impact of climate changes and inevitability to reduce greenhouse gases, the need to use low-carbon and sustainable construction has increased. In this work, it is investigated how texture of the surface building materials and radiative heat transfer phenomenon in flat multilayer can be correlated. Attempts to test the surface emissivity are taken however, the trustworthiness of measurement results remains a concern since sensor size and thickness are common problems. This paper presents an experimental method to studies surface emissivity with use self constructed thermal sensors and thermal imaging technique. The surface of building materials was modified by mechanical and chemical treatment affecting the reduction of the emissivity. For testing the shaping surface of materials and mapping its three-dimensional structure, scanning profilometry were used in a laboratory. By comparing the results of laboratory tests and performed analysis of 3D computer fluid dynamics software, it can be shown that a change in the surface coverage of materials affects the heat transport by radiation between layers. Motivated by recent advancements in variational inference, this publication evaluates the potential use a dedicated data processing approach, and properly constructed temperature sensors, the influence of the surface emissivity on the phenomenon of radiation and heat transport in the entire partition can be determined.

Keywords: heat transfer, surface roughness, surface emissivity, radiation

Procedia PDF Downloads 76
1749 Enhanced Face Recognition with Daisy Descriptors Using 1BT Based Registration

Authors: Sevil Igit, Merve Meric, Sarp Erturk

Abstract:

In this paper, it is proposed to improve Daisy descriptor based face recognition using a novel One-Bit Transform (1BT) based pre-registration approach. The 1BT based pre-registration procedure is fast and has low computational complexity. It is shown that the face recognition accuracy is improved with the proposed approach. The proposed approach can facilitate highly accurate face recognition using DAISY descriptor with simple matching and thereby facilitate a low-complexity approach.

Keywords: face recognition, Daisy descriptor, One-Bit Transform, image registration

Procedia PDF Downloads 349
1748 Sensitivity Analysis during the Optimization Process Using Genetic Algorithms

Authors: M. A. Rubio, A. Urquia

Abstract:

Genetic algorithms (GA) are applied to the solution of high-dimensional optimization problems. Additionally, sensitivity analysis (SA) is usually carried out to determine the effect on optimal solutions of changes in parameter values of the objective function. These two analyses (i.e., optimization and sensitivity analysis) are computationally intensive when applied to high-dimensional functions. The approach presented in this paper consists in performing the SA during the GA execution, by statistically analyzing the data obtained of running the GA. The advantage is that in this case SA does not involve making additional evaluations of the objective function and, consequently, this proposed approach requires less computational effort than conducting optimization and SA in two consecutive steps.

Keywords: optimization, sensitivity, genetic algorithms, model calibration

Procedia PDF Downloads 422
1747 A Metaheuristic for the Layout and Scheduling Problem in a Job Shop Environment

Authors: Hernández Eva Selene, Reyna Mary Carmen, Rivera Héctor, Barragán Irving

Abstract:

We propose an approach that jointly addresses the layout of a facility and the scheduling of a sequence of jobs. In real production, these two problems are interrelated. However, they are treated separately in the literature. Our approach is an extension of the job shop problem with transportation delay, where the location of the machines is selected among possible sites. The model minimizes the makespan, using the short processing times rule with two algorithms; the first one considers all the permutations for the location of machines, and the second only a heuristic to select some specific permutations that reduces computational time. Some instances are proved and compared with literature.

Keywords: layout problem, job shop scheduling problem, concurrent scheduling and layout problem, metaheuristic

Procedia PDF Downloads 592
1746 Modeling False Statements in Texts

Authors: Francielle A. Vargas, Thiago A. S. Pardo

Abstract:

According to the standard philosophical definition, lying is saying something that you believe to be false with the intent to deceive. For deception detection, the FBI trains its agents in a technique named statement analysis, which attempts to detect deception based on parts of speech (i.e., linguistics style). This method is employed in interrogations, where the suspects are first asked to make a written statement. In this poster, we model false statements using linguistics style. In order to achieve this, we methodically analyze linguistic features in a corpus of fake news in the Portuguese language. The results show that they present substantial lexical, syntactic and semantic variations, as well as punctuation and emotion distinctions.

Keywords: deception detection, linguistics style, computational linguistics, natural language processing

Procedia PDF Downloads 202
1745 Efficient Signcryption Scheme with Provable Security for Smart Card

Authors: Jayaprakash Kar, Daniyal M. Alghazzawi

Abstract:

The article proposes a novel construction of signcryption scheme with provable security which is most suited to implement on smart card. It is secure in random oracle model and the security relies on Decisional Bilinear Diffie-Hellmann Problem. The proposed scheme is secure against adaptive chosen ciphertext attack (indistiguishbility) and adaptive chosen message attack (unforgebility). Also, it is inspired by zero-knowledge proof. The two most important security goals for smart card are Confidentiality and authenticity. These functions are performed in one logical step in low computational cost.

Keywords: random oracle, provable security, unforgebility, smart card

Procedia PDF Downloads 584
1744 Fruit Identification System in Sweet Orange Citrus (L.) Osbeck Using Thermal Imaging and Fuzzy

Authors: Ingrid Argote, John Archila, Marcelo Becker

Abstract:

In agriculture, intelligent systems applications have generated great advances in automating some of the processes in the production chain. In order to improve the efficiency of those systems is proposed a vision system to estimate the amount of fruits in sweet orange trees. This work presents a system proposal using capture of thermal images and fuzzy logic. A bibliographical review has been done to analyze the state-of-the-art of the different systems used in fruit recognition, and also the different applications of thermography in agricultural systems. The algorithm developed for this project uses the metrics of the fuzzines parameter to the contrast improvement and segmentation of the image, for the counting algorith m was used the Hough transform. In order to validate the proposed algorithm was created a bank of images of sweet orange Citrus (L.) Osbeck acquired in the Maringá Farm. The tests with the algorithm Indicated that the variation of the tree branch temperature and the fruit is not very high, Which makes the process of image segmentation using this differentiates, This Increases the amount of false positives in the fruit counting algorithm. Recognition of fruits isolated with the proposed algorithm present an overall accuracy of 90.5 % and grouped fruits. The accuracy was 81.3 %. The experiments show the need for a more suitable hardware to have a better recognition of small temperature changes in the image.

Keywords: Agricultural systems, Citrus, Fuzzy logic, Thermal images.

Procedia PDF Downloads 218
1743 Virtual Screening and in Silico Toxicity Property Prediction of Compounds against Mycobacterium tuberculosis Lipoate Protein Ligase B (LipB)

Authors: Junie B. Billones, Maria Constancia O. Carrillo, Voltaire G. Organo, Stephani Joy Y. Macalino, Inno A. Emnacen, Jamie Bernadette A. Sy

Abstract:

The drug discovery and development process is generally known to be a very lengthy and labor-intensive process. Therefore, in order to be able to deliver prompt and effective responses to cure certain diseases, there is an urgent need to reduce the time and resources needed to design, develop, and optimize potential drugs. Computer-aided drug design (CADD) is able to alleviate this issue by applying computational power in order to streamline the whole drug discovery process, starting from target identification to lead optimization. This drug design approach can be predominantly applied to diseases that cause major public health concerns, such as tuberculosis. Hitherto, there has been no concrete cure for this disease, especially with the continuing emergence of drug resistant strains. In this study, CADD is employed for tuberculosis by first identifying a key enzyme in the mycobacterium’s metabolic pathway that would make a good drug target. One such potential target is the lipoate protein ligase B enzyme (LipB), which is a key enzyme in the M. tuberculosis metabolic pathway involved in the biosynthesis of the lipoic acid cofactor. Its expression is considerably up-regulated in patients with multi-drug resistant tuberculosis (MDR-TB) and it has no known back-up mechanism that can take over its function when inhibited, making it an extremely attractive target. Using cutting-edge computational methods, compounds from AnalytiCon Discovery Natural Derivatives database were screened and docked against the LipB enzyme in order to rank them based on their binding affinities. Compounds which have better binding affinities than LipB’s known inhibitor, decanoic acid, were subjected to in silico toxicity evaluation using the ADMET and TOPKAT protocols. Out of the 31,692 compounds in the database, 112 of these showed better binding energies than decanoic acid. Furthermore, 12 out of the 112 compounds showed highly promising ADMET and TOPKAT properties. Future studies involving in vitro or in vivo bioassays may be done to further confirm the therapeutic efficacy of these 12 compounds, which eventually may then lead to a novel class of anti-tuberculosis drugs.

Keywords: pharmacophore, molecular docking, lipoate protein ligase B (LipB), ADMET, TOPKAT

Procedia PDF Downloads 406