Search results for: computational vision
1489 Necrotising Anterior Scleritis and Scleroderma: A Rare Association
Authors: Angeliki Vassila, Dimitrios Kalogeropoulos, Rania Rawashdeh, Nigel Hall, Najiha Rahman, Mark Fabian, Suresh Thulasidharan, Hossain Parwez
Abstract:
Introduction: Necrotising scleritis is a severe form of scleritis and poses a significant threat to vision. It can manifest in various systemic autoimmune disorders, systemic vasculitis, or as a consequence of microbial infections. The objective of this study is to present a case of necrotizing scleritis associated with scleroderma, which was further complicated by a secondary Staphylococcus epidermidis infection. Methods: This is a retrospective analysis that examines the medical records of a patient who was hospitalised in the Eye Unit at University Hospital Southampton. Results: A 78-year-old woman presented at the eye casualty department of our unit with a two-week history of progressively worsening pain in her left eye. She received a diagnosis of necrotising scleritis and was admitted to the hospital for further treatment. It was decided to commence a three-day course of intravenous methylprednisolone followed by a tapering regimen of oral steroids. Additionally, a conjunctival swab was taken, and two days later, it revealed the presence of S. epidermidis, indicating a potential secondary infection. Given this finding, she was also prescribed topical (Ofloxacin 0.3% - four times daily) and oral (Ciprofloxacin 750mg – twice daily) antibiotics. The inflammation and symptoms gradually improved, leading to the patient being scheduled for a scleral graft and applying an amniotic membrane to cover the area of scleral thinning. Conclusions: Rheumatoid arthritis and granulomatosis with polyangiitis are the most commonly identifiable systemic diseases associated with necrotising scleritis. Although association with scleroderma is extremely rare, early identification and treatment are necessary to prevent scleritis-related complications.Keywords: scleritis, necrotizing scleritis, scleroderma, autoimmune disease
Procedia PDF Downloads 291488 3D Human Face Reconstruction in Unstable Conditions
Authors: Xiaoyuan Suo
Abstract:
3D object reconstruction is a broad research area within the computer vision field involving many stages and still open problems. One of the existing challenges in this field lies with micromotion, such as the facial expressions on the appearance of the human or animal face. Similar literatures in this field focuses on 3D reconstruction in stable conditions such as an existing image or photos taken in a rather static environment, while the purpose of this work is to discuss a flexible scan system using multiple cameras that can correctly reconstruct 3D stable and moving objects -- human face with expression in particular. Further, a mathematical model is proposed at the end of this literature to automate the 3D object reconstruction process. The reconstruction process takes several stages. Firstly, a set of simple 2D lines would be projected onto the object and hence a set of uneven curvy lines can be obtained, which represents the 3D numerical data of the surface. The lines and their shapes will help to identify object’s 3D construction in pixels. With the two-recorded angles and their distance from the camera, a simple mathematical calculation would give the resulting coordinate of each projected line in an absolute 3D space. This proposed research will benefit many practical areas, including but not limited to biometric identification, authentications, cybersecurity, preservation of cultural heritage, drama acting especially those with rapid and complex facial gestures, and many others. Specifically, this will (I) provide a brief survey of comparable techniques existing in this field. (II) discuss a set of specialized methodologies or algorithms for effective reconstruction of 3D objects. (III)implement, and testing the developed methodologies. (IV) verify findings with data collected from experiments. (V) conclude with lessons learned and final thoughts.Keywords: 3D photogrammetry, 3D object reconstruction, facial expression recognition, facial recognition
Procedia PDF Downloads 1501487 Radical Web Text Classification Using a Composite-Based Approach
Authors: Kolade Olawande Owoeye, George R. S. Weir
Abstract:
The widespread of terrorism and extremism activities on the internet has become a major threat to the government and national securities due to their potential dangers which have necessitated the need for intelligence gathering via web and real-time monitoring of potential websites for extremist activities. However, the manual classification for such contents is practically difficult or time-consuming. In response to this challenge, an automated classification system called composite technique was developed. This is a computational framework that explores the combination of both semantics and syntactic features of textual contents of a web. We implemented the framework on a set of extremist webpages dataset that has been subjected to the manual classification process. Therein, we developed a classification model on the data using J48 decision algorithm, this is to generate a measure of how well each page can be classified into their appropriate classes. The classification result obtained from our method when compared with other states of arts, indicated a 96% success rate in classifying overall webpages when matched against the manual classification.Keywords: extremist, web pages, classification, semantics, posit
Procedia PDF Downloads 1451486 The Use of Degradation Measures to Design Reliability Test Plans
Authors: Stephen V. Crowder, Jonathan W. Lane
Abstract:
With short production development times, there is an increased need to demonstrate product reliability relatively quickly with minimal testing. In such cases there may be few if any observed failures. Thus it may be difficult to assess reliability using the traditional reliability test plans that measure only time (or cycles) to failure. For many components, degradation measures will contain important information about performance and reliability. These measures can be used to design a minimal test plan, in terms of number of units placed on test and duration of the test, necessary to demonstrate a reliability goal. In this work we present a case study involving an electronic component subject to degradation. The data, consisting of 42 degradation paths of cycles to failure, are first used to estimate a reliability function. Bootstrapping techniques are then used to perform power studies and develop a minimal reliability test plan for future production of this component.Keywords: degradation measure, time to failure distribution, bootstrap, computational science
Procedia PDF Downloads 5311485 Evaluation of Progressive Collapse of Transmission Tower
Authors: Jeong-Hwan Choi, Hyo-Sang Park, Tae-Hyung Lee
Abstract:
The transmission tower is one of the crucial lifeline structures in a modern society, and it needs to be protected against extreme loading conditions. However, the transmission tower is a very complex structure and, therefore, it is very difficult to simulate the actual damage and the collapse behavior of the tower structure. In this study, the actual collapse behavior of the transmission tower due to lateral loading conditions such as wind load is evaluated through the computational simulation. For that, a progressive collapse procedure is applied to the simulation. In this procedure, after running the simulation, if a member of the tower structure fails, the failed member is removed and the simulation run again. The 154kV transmission tower is selected for this study. The simulation is performed by nonlinear static analysis procedure, namely pushover analysis, using OpenSEES, an earthquake simulation platform. Three-dimensional finite element models of those towers are developed.Keywords: transmission tower, OpenSEES, pushover, progressive collapse
Procedia PDF Downloads 3571484 Application of a Universal Distortion Correction Method in Stereo-Based Digital Image Correlation Measurement
Authors: Hu Zhenxing, Gao Jianxin
Abstract:
Stereo-based digital image correlation (also referred to as three-dimensional (3D) digital image correlation (DIC)) is a technique for both 3D shape and surface deformation measurement of a component, which has found increasing applications in academia and industries. The accuracy of the reconstructed coordinate depends on many factors such as configuration of the setup, stereo-matching, distortion, etc. Most of these factors have been investigated in literature. For instance, the configuration of a binocular vision system determines the systematic errors. The stereo-matching errors depend on the speckle quality and the matching algorithm, which can only be controlled in a limited range. And the distortion is non-linear particularly in a complex imaging acquisition system. Thus, the distortion correction should be carefully considered. Moreover, the distortion function is difficult to formulate in a complex imaging acquisition system using conventional models in such cases where microscopes and other complex lenses are involved. The errors of the distortion correction will propagate to the reconstructed 3D coordinates. To address the problem, an accurate mapping method based on 2D B-spline functions is proposed in this study. The mapping functions are used to convert the distorted coordinates into an ideal plane without distortions. This approach is suitable for any image acquisition distortion models. It is used as a prior process to convert the distorted coordinate to an ideal position, which enables the camera to conform to the pin-hole model. A procedure of this approach is presented for stereo-based DIC. Using 3D speckle image generation, numerical simulations were carried out to compare the accuracy of both the conventional method and the proposed approach.Keywords: distortion, stereo-based digital image correlation, b-spline, 3D, 2D
Procedia PDF Downloads 4981483 Numerical Solution Speedup of the Laplace Equation Using FPGA Hardware
Authors: Abbas Ebrahimi, Mohammad Zandsalimy
Abstract:
The main purpose of this study is to investigate the feasibility of using FPGA (Field Programmable Gate Arrays) chips as alternatives for the conventional CPUs to accelerate the numerical solution of the Laplace equation. FPGA is an integrated circuit that contains an array of logic blocks, and its architecture can be reprogrammed and reconfigured after manufacturing. Complex circuits for various applications can be designed and implemented using FPGA hardware. The reconfigurable hardware used in this paper is an SoC (System on a Chip) FPGA type that integrates both microprocessor and FPGA architectures into a single device. In the present study the Laplace equation is implemented and solved numerically on both reconfigurable hardware and CPU. The precision of results and speedups of the calculations are compared together. The computational process on FPGA, is up to 20 times faster than a conventional CPU, with the same data precision. An analytical solution is used to validate the results.Keywords: accelerating numerical solutions, CFD, FPGA, hardware definition language, numerical solutions, reconfigurable hardware
Procedia PDF Downloads 3821482 Numerical Investigation of Aerodynamic Analysis on Passenger Vehicle
Authors: Cafer Görkem Pınar, İlker Coşar, Serkan Uzun, Atahan Çelebi, Mehmet Ali Ersoy, Ali Pınarbaşı
Abstract:
In this study, it was numerically investigated that a 1:1 scale model of the Renault Clio MK4 SW brand vehicle aerodynamic analysis was performed in the commercial computational fluid dynamics (CFD) package program of ANSYS CFX 2021 R1 under steady, subsonic, and 3-D conditions. The model of vehicle used for the analysis was made independent of the number of mesh elements, and the k-epsilon turbulence model was applied during the analysis. Results were interpreted as streamlines, pressure gradient, and turbulent kinetic energy contours around the vehicle at 50 km/h and 100 km/h speeds. In addition, the validity of the analysis was decided by comparing the drag coefficient of the vehicle with the values in the literature. As a result, the pressure gradient contours of the taillight of the Renault Clio MK4 SW vehicle were examined, and the behavior of the total force at speeds of 50 km/h and 100 km/h was interpreted.Keywords: CFD, k-epsilon, aerodynamics, drag coefficient, taillight
Procedia PDF Downloads 1431481 A High-Level Co-Evolutionary Hybrid Algorithm for the Multi-Objective Job Shop Scheduling Problem
Authors: Aydin Teymourifar, Gurkan Ozturk
Abstract:
In this paper, a hybrid distributed algorithm has been suggested for the multi-objective job shop scheduling problem. Many new approaches are used at design steps of the distributed algorithm. Co-evolutionary structure of the algorithm and competition between different communicated hybrid algorithms, which are executed simultaneously, causes to efficient search. Using several machines for distributing the algorithms, at the iteration and solution levels, increases computational speed. The proposed algorithm is able to find the Pareto solutions of the big problems in shorter time than other algorithm in the literature. Apache Spark and Hadoop platforms have been used for the distribution of the algorithm. The suggested algorithm and implementations have been compared with results of the successful algorithms in the literature. Results prove the efficiency and high speed of the algorithm.Keywords: distributed algorithms, Apache Spark, Hadoop, job shop scheduling, multi-objective optimization
Procedia PDF Downloads 3631480 Generation of Quasi-Measurement Data for On-Line Process Data Analysis
Authors: Hyun-Woo Cho
Abstract:
For ensuring the safety of a manufacturing process one should quickly identify an assignable cause of a fault in an on-line basis. To this end, many statistical techniques including linear and nonlinear methods have been frequently utilized. However, such methods possessed a major problem of small sample size, which is mostly attributed to the characteristics of empirical models used for reference models. This work presents a new method to overcome the insufficiency of measurement data in the monitoring and diagnosis tasks. Some quasi-measurement data are generated from existing data based on the two indices of similarity and importance. The performance of the method is demonstrated using a real data set. The results turn out that the presented methods are able to handle the insufficiency problem successfully. In addition, it is shown to be quite efficient in terms of computational speed and memory usage, and thus on-line implementation of the method is straightforward for monitoring and diagnosis purposes.Keywords: data analysis, diagnosis, monitoring, process data, quality control
Procedia PDF Downloads 4811479 Roof and Road Network Detection through Object Oriented SVM Approach Using Low Density LiDAR and Optical Imagery in Misamis Oriental, Philippines
Authors: Jigg L. Pelayo, Ricardo G. Villar, Einstine M. Opiso
Abstract:
The advances of aerial laser scanning in the Philippines has open-up entire fields of research in remote sensing and machine vision aspire to provide accurate timely information for the government and the public. Rapid mapping of polygonal roads and roof boundaries is one of its utilization offering application to disaster risk reduction, mitigation and development. The study uses low density LiDAR data and high resolution aerial imagery through object-oriented approach considering the theoretical concept of data analysis subjected to machine learning algorithm in minimizing the constraints of feature extraction. Since separating one class from another in distinct regions of a multi-dimensional feature-space, non-trivial computing for fitting distribution were implemented to formulate the learned ideal hyperplane. Generating customized hybrid feature which were then used in improving the classifier findings. Supplemental algorithms for filtering and reshaping object features are develop in the rule set for enhancing the final product. Several advantages in terms of simplicity, applicability, and process transferability is noticeable in the methodology. The algorithm was tested in the different random locations of Misamis Oriental province in the Philippines demonstrating robust performance in the overall accuracy with greater than 89% and potential to semi-automation. The extracted results will become a vital requirement for decision makers, urban planners and even the commercial sector in various assessment processes.Keywords: feature extraction, machine learning, OBIA, remote sensing
Procedia PDF Downloads 3621478 Creating Systems Change: Implementing Cross-Sector Initiatives within the Justice System to Support Ontarians with Mental Health and Addictions Needs
Authors: Tania Breton, Dorina Simeonov, Shauna MacEachern
Abstract:
Ontario’s 10 Year Mental Health and Addictions Strategy has included the establishment of 18 Service Collaborative across the province; cross-sector tables in a specific region coming together to explore mental health and addiction system needs and adopting an intervention to address that need. The process is community led and supported by implementation teams from the Centre for Addiction and Mental Health (CAMH), using the framework of implementation science (IS) to enable evidence-based and sustained change. These justice initiatives are focused on the intersection of the justice system and the mental health and addiction systems. In this presentation, we will share the learnings, achievements and challenges of implementing innovative practices to the mental health and addictions needs of Ontarians within the justice system. Specifically, we will focus on the key points across the justice system - from early intervention and trauma-informed, culturally appropriate services to post-sentence support and community reintegration. Our approach to this work involves external implementation support from the CAMH team including coaching, knowledge exchange, evaluation, Aboriginal engagement and health equity expertise. Agencies supported the implementation of tools and processes which changed practice at the local level. These practices are being scaled up across Ontario and community agencies have come together in an unprecedented collaboration and there is a shared vision of the issues overlapping between the mental health, addictions and justice systems. Working with ministry partners has allowed space for innovation and created an environment where better approaches can be nurtured and spread.Keywords: implementation, innovation, early identification, mental health and addictions, prevention, systems
Procedia PDF Downloads 3621477 Using Scilab® as New Introductory Method in Numerical Calculations and Programming for Computational Fluid Dynamics (CFD)
Authors: Nicoly Coelho, Eduardo Vieira Vilas Boas, Paulo Orestes Formigoni
Abstract:
Faced with the remarkable developments in the various segments of modern engineering, provided by the increasing technological development, professionals of all educational areas need to overcome the difficulties generated due to the good understanding of those who are starting their academic journey. Aiming to overcome these difficulties, this article aims at an introduction to the basic study of numerical methods applied to fluid mechanics and thermodynamics, demonstrating the modeling and simulations with its substance, and a detailed explanation of the fundamental numerical solution for the use of finite difference method, using SCILAB, a free software easily accessible as it is free and can be used for any research center or university, anywhere, both in developed and developing countries. It is known that the Computational Fluid Dynamics (CFD) is a necessary tool for engineers and professionals who study fluid mechanics, however, the teaching of this area of knowledge in undergraduate programs faced some difficulties due to software costs and the degree of difficulty of mathematical problems involved in this way the matter is treated only in postgraduate courses. This work aims to bring the use of DFC low cost in teaching Transport Phenomena for graduation analyzing a small classic case of fundamental thermodynamics with Scilab® program. The study starts from the basic theory involving the equation the partial differential equation governing heat transfer problem, implies the need for mastery of students, discretization processes that include the basic principles of series expansion Taylor responsible for generating a system capable of convergence check equations using the concepts of Sassenfeld, finally coming to be solved by Gauss-Seidel method. In this work we demonstrated processes involving both simple problems solved manually, as well as the complex problems that required computer implementation, for which we use a small algorithm with less than 200 lines in Scilab® in heat transfer study of a heated plate in rectangular shape on four sides with different temperatures on either side, producing a two-dimensional transport with colored graphic simulation. With the spread of computer technology, numerous programs have emerged requiring great researcher programming skills. Thinking that this ability to program DFC is the main problem to be overcome, both by students and by researchers, we present in this article a hint of use of programs with less complex interface, thus enabling less difficulty in producing graphical modeling and simulation for DFC with an extension of the programming area of experience for undergraduates.Keywords: numerical methods, finite difference method, heat transfer, Scilab
Procedia PDF Downloads 3871476 Excited State Structural Dynamics of Retinal Isomerization Revealed by a Femtosecond X-Ray Laser
Authors: Przemyslaw Nogly, Tobias Weinert, Daniel James, Sergio Carbajo, Dmitry Ozerov, Antonia Furrer, Dardan Gashi, Veniamin Borin, Petr Skopintsev, Kathrin Jaeger, Karol Nass, Petra Bath, Robert Bosman, Jason Koglin, Matthew Seaberg, Thomas Lane, Demet Kekilli, Steffen Brünle, Tomoyuki Tanaka, Wenting Wu, Christopher Milne, Thomas A. White, Anton Barty, Uwe Weierstall, Valerie Panneels, Eriko Nango, So Iwata, Mark Hunter, Igor Schapiro, Gebhard Schertler, Richard Neutze, Jörg Standfuss
Abstract:
Ultrafast isomerization of retinal is the primary step in a range of photoresponsive biological functions including vision in humans and ion-transport across bacterial membranes. We studied the sub-picosecond structural dynamics of retinal isomerization in the light-driven proton pump bacteriorhodopsin using an X-ray laser. Twenty snapshots with near-atomic spatial and temporal resolution in the femtosecond regime show how the excited all-trans retinal samples conformational states within the protein binding pocket prior to passing through a highly-twisted geometry and emerging in the 13-cis conformation. The aspartic acid residues and functional water molecules in proximity of the retinal Schiff base respond collectively to formation and decay of the initial excited state and retinal isomerization. These observations reveal how the protein scaffold guides this remarkably efficient photochemical reaction.Keywords: bacteriorhodopsin, free-electron laser, retinal isomerization mechanism, time-resolved crystallography
Procedia PDF Downloads 2481475 Predicting Loss of Containment in Surface Pipeline using Computational Fluid Dynamics and Supervised Machine Learning Model to Improve Process Safety in Oil and Gas Operations
Authors: Muhammmad Riandhy Anindika Yudhy, Harry Patria, Ramadhani Santoso
Abstract:
Loss of containment is the primary hazard that process safety management is concerned within the oil and gas industry. Escalation to more serious consequences all begins with the loss of containment, starting with oil and gas release from leakage or spillage from primary containment resulting in pool fire, jet fire and even explosion when reacted with various ignition sources in the operations. Therefore, the heart of process safety management is avoiding loss of containment and mitigating its impact through the implementation of safeguards. The most effective safeguard for the case is an early detection system to alert Operations to take action prior to a potential case of loss of containment. The detection system value increases when applied to a long surface pipeline that is naturally difficult to monitor at all times and is exposed to multiple causes of loss of containment, from natural corrosion to illegal tapping. Based on prior researches and studies, detecting loss of containment accurately in the surface pipeline is difficult. The trade-off between cost-effectiveness and high accuracy has been the main issue when selecting the traditional detection method. The current best-performing method, Real-Time Transient Model (RTTM), requires analysis of closely positioned pressure, flow and temperature (PVT) points in the pipeline to be accurate. Having multiple adjacent PVT sensors along the pipeline is expensive, hence generally not a viable alternative from an economic standpoint.A conceptual approach to combine mathematical modeling using computational fluid dynamics and a supervised machine learning model has shown promising results to predict leakage in the pipeline. Mathematical modeling is used to generate simulation data where this data is used to train the leak detection and localization models. Mathematical models and simulation software have also been shown to provide comparable results with experimental data with very high levels of accuracy. While the supervised machine learning model requires a large training dataset for the development of accurate models, mathematical modeling has been shown to be able to generate the required datasets to justify the application of data analytics for the development of model-based leak detection systems for petroleum pipelines. This paper presents a review of key leak detection strategies for oil and gas pipelines, with a specific focus on crude oil applications, and presents the opportunities for the use of data analytics tools and mathematical modeling for the development of robust real-time leak detection and localization system for surface pipelines. A case study is also presented.Keywords: pipeline, leakage, detection, AI
Procedia PDF Downloads 1911474 Study the Sloshing Phenomenon in the Tank Filled Partially with Liquid Using Computational Fluid Dynamics (CFD) Simulation
Authors: Amit Kumar, Jaikumar V., Pradeep A. G., Shivakumar Bhavi
Abstract:
Amit Kumar, Jaikumar V, Pradeep AG, Shivakumar Bhavi Reducing sloshing is one of the major challenges in industries where transporting of liquid is involved. The present study investigates the sloshing effect for different liquid levels of 50% of the tank capacity. CFD simulation for two different baffle configurations has been carried out using a time-based multiphase Volume of fluid (VOF) scheme. Baffles were introduced to examine the sloshing effect inside the tank. Results were compared against the baseline case to assess the effectiveness of baffles; maximum liquid height over the period of the simulation was considered as the parameter for measuring the sloshing effect inside the tank. It was found that the addition of baffles reduced the sloshing effect inside the tank as compared to the baseline model.Keywords: CFD, sloshing, VOF, multiphase
Procedia PDF Downloads 1921473 Variable-Fidelity Surrogate Modelling with Kriging
Authors: Selvakumar Ulaganathan, Ivo Couckuyt, Francesco Ferranti, Tom Dhaene, Eric Laermans
Abstract:
Variable-fidelity surrogate modelling offers an efficient way to approximate function data available in multiple degrees of accuracy each with varying computational cost. In this paper, a Kriging-based variable-fidelity surrogate modelling approach is introduced to approximate such deterministic data. Initially, individual Kriging surrogate models, which are enhanced with gradient data of different degrees of accuracy, are constructed. Then these Gradient enhanced Kriging surrogate models are strategically coupled using a recursive CoKriging formulation to provide an accurate surrogate model for the highest fidelity data. While, intuitively, gradient data is useful to enhance the accuracy of surrogate models, the primary motivation behind this work is to investigate if it is also worthwhile incorporating gradient data of varying degrees of accuracy.Keywords: Kriging, CoKriging, Surrogate modelling, Variable- fidelity modelling, Gradients
Procedia PDF Downloads 5581472 Experimental and CFD of Desgined Small Wind Turbine
Authors: Tarek A. Mekail, Walid M. A. Elmagid
Abstract:
Many researches have concentrated on improving the aerodynamic performance of wind turbine blade through testing and theoretical studies. A small wind turbine blade is designed, fabricated and tested. The power performance of small horizontal axis wind turbines is simulated in details using Computational Fluid Dynamic (CFD). The three-dimensional CFD models are presented using ANSYS-CFX v13 software for predicting the performance of a small horizontal axis wind turbine. The simulation results are compared with the experimental data measured from a small wind turbine model, which designed according to a vehicle-based test system. The analysis of wake effect and aerodynamic of the blade can be carried out when the rotational effect was simulated. Finally, comparison between experimental, numerical and analytical performance has been done. The comparison is fairly good.Keywords: small wind turbine, CFD of wind turbine, CFD, performance of wind turbine, test of small wind turbine, wind turbine aerodynamic, 3D model
Procedia PDF Downloads 5421471 Further Analysis of Global Robust Stability of Neural Networks with Multiple Time Delays
Authors: Sabri Arik
Abstract:
In this paper, we study the global asymptotic robust stability of delayed neural networks with norm-bounded uncertainties. By employing the Lyapunov stability theory and Homeomorphic mapping theorem, we derive some new types of sufficient conditions ensuring the existence, uniqueness and global asymptotic stability of the equilibrium point for the class of neural networks with discrete time delays under parameter uncertainties and with respect to continuous and slopebounded activation functions. An important aspect of our results is their low computational complexity as the reported results can be verified by checking some properties symmetric matrices associated with the uncertainty sets of network parameters. The obtained results are shown to be generalization of some of the previously published corresponding results. Some comparative numerical examples are also constructed to compare our results with some closely related existing literature results.Keywords: neural networks, delayed systems, lyapunov functionals, stability analysis
Procedia PDF Downloads 5271470 Efficient Pre-Processing of Single-Cell Assay for Transposase Accessible Chromatin with High-Throughput Sequencing Data
Authors: Fan Gao, Lior Pachter
Abstract:
The primary tool currently used to pre-process 10X Chromium single-cell ATAC-seq data is Cell Ranger, which can take very long to run on standard datasets. To facilitate rapid pre-processing that enables reproducible workflows, we present a suite of tools called scATAK for pre-processing single-cell ATAC-seq data that is 15 to 18 times faster than Cell Ranger on mouse and human samples. Our tool can also calculate chromatin interaction potential matrices, and generate open chromatin signal and interaction traces for cell groups. We use scATAK tool to explore the chromatin regulatory landscape of a healthy adult human brain and unveil cell-type specific features, and show that it provides a convenient and computational efficient approach for pre-processing single-cell ATAC-seq data.Keywords: single-cell, ATAC-seq, bioinformatics, open chromatin landscape, chromatin interactome
Procedia PDF Downloads 1551469 Analysis of Moving Loads on Bridges Using Surrogate Models
Authors: Susmita Panda, Arnab Banerjee, Ajinkya Baxy, Bappaditya Manna
Abstract:
The design of short to medium-span high-speed bridges in critical locations is an essential aspect of vehicle-bridge interaction. Due to dynamic interaction between moving load and bridge, mathematical models or finite element modeling computations become time-consuming. Thus, to reduce the computational effort, a universal approximator using an artificial neural network (ANN) has been used to evaluate the dynamic response of the bridge. The data set generation and training of surrogate models have been conducted over the results obtained from mathematical modeling. Further, the robustness of the surrogate model has been investigated, which showed an error percentage of less than 10% with conventional methods. Additionally, the dependency of the dynamic response of the bridge on various load and bridge parameters has been highlighted through a parametric study.Keywords: artificial neural network, mode superposition method, moving load analysis, surrogate models
Procedia PDF Downloads 1001468 Additive Manufacturing of Overhangs: From Temporary Supports to Self-Support
Authors: Paulo Mendonca, Nzar Faiq Naqeshbandi
Abstract:
The objective of this study is to propose an interactive design environment that outlines the underlying computational framework to reach self-supporting overhangs. The research demonstrates the digital printability of overhangs taking into consideration factors related to the geometry design, the material used, the applied support, and the printing set-up of slicing and the extruder inclination. Parametric design tools can contribute to the design phase, form-finding, and stability optimization of self-supporting structures while printing in order to hold the components in place until they are sufficiently advanced to support themselves. The challenge is to ensure the stability of the printed parts in the critical inclinations during the whole fabrication process. Facilitating the identification of parameterization will allow to predict and optimize the process. Later, in the light of the previous findings, some guidelines of simulations and physical tests are given to be conducted for estimating the structural and functional performance.Keywords: additive manufacturing, overhangs, self-support overhangs, printability, parametric tools
Procedia PDF Downloads 1221467 Radar Signal Detection Using Neural Networks in Log-Normal Clutter for Multiple Targets Situations
Authors: Boudemagh Naime
Abstract:
Automatic radar detection requires some methods of adapting to variations in the background clutter in order to control their false alarm rate. The problem becomes more complicated in non-Gaussian environment. In fact, the conventional approach in real time applications requires a complex statistical modeling and much computational operations. To overcome these constraints, we propose another approach based on artificial neural network (ANN-CMLD-CFAR) using a Back Propagation (BP) training algorithm. The considered environment follows a log-normal distribution in the presence of multiple Rayleigh-targets. To evaluate the performances of the considered detector, several situations, such as scale parameter and the number of interferes targets, have been investigated. The simulation results show that the ANN-CMLD-CFAR processor outperforms the conventional statistical one.Keywords: radat detection, ANN-CMLD-CFAR, log-normal clutter, statistical modelling
Procedia PDF Downloads 3641466 The Implementation of Secton Method for Finding the Root of Interpolation Function
Authors: Nur Rokhman
Abstract:
A mathematical function gives relationship between the variables composing the function. Interpolation can be viewed as a process of finding mathematical function which goes through some specified points. There are many interpolation methods, namely: Lagrange method, Newton method, Spline method etc. For some specific condition, such as, big amount of interpolation points, the interpolation function can not be written explicitly. This such function consist of computational steps. The solution of equations involving the interpolation function is a problem of solution of non linear equation. Newton method will not work on the interpolation function, for the derivative of the interpolation function cannot be written explicitly. This paper shows the use of Secton method to determine the numerical solution of the function involving the interpolation function. The experiment shows the fact that Secton method works better than Newton method in finding the root of Lagrange interpolation function.Keywords: Secton method, interpolation, non linear function, numerical solution
Procedia PDF Downloads 3791465 An Efficient Algorithm for Global Alignment of Protein-Protein Interaction Networks
Authors: Duc Dong Do, Ngoc Ha Tran, Thanh Hai Dang, Cao Cuong Dang, Xuan Huan Hoang
Abstract:
Global aligning two protein-protein interaction networks is an essentially important task in bioinformatics/computational biology field of study. It is a challenging and widely studied research topic in recent years. Accurately aligned networks allow us to identify functional modules of proteins and/ororthologous proteins from which unknown functions of a protein can be inferred. We here introduce a novel efficient heuristic global network alignment algorithm called FASTAn, including two phases: the first to construct an initial alignment and the second to improve such alignment by exerting a local optimization repeated procedure. The experimental results demonstrated that FASTAn outperformed the state-of-the-art global network alignment algorithm namely SPINAL in terms of both commonly used objective scores and the run-time.Keywords: FASTAn, Heuristic algorithm, biological network alignment, protein-protein interaction networks
Procedia PDF Downloads 6041464 Non-Invasive Data Extraction from Machine Display Units Using Video Analytics
Authors: Ravneet Kaur, Joydeep Acharya, Sudhanshu Gaur
Abstract:
Artificial Intelligence (AI) has the potential to transform manufacturing by improving shop floor processes such as production, maintenance and quality. However, industrial datasets are notoriously difficult to extract in a real-time, streaming fashion thus, negating potential AI benefits. The main example is some specialized industrial controllers that are operated by custom software which complicates the process of connecting them to an Information Technology (IT) based data acquisition network. Security concerns may also limit direct physical access to these controllers for data acquisition. To connect the Operational Technology (OT) data stored in these controllers to an AI application in a secure, reliable and available way, we propose a novel Industrial IoT (IIoT) solution in this paper. In this solution, we demonstrate how video cameras can be installed in a factory shop floor to continuously obtain images of the controller HMIs. We propose image pre-processing to segment the HMI into regions of streaming data and regions of fixed meta-data. We then evaluate the performance of multiple Optical Character Recognition (OCR) technologies such as Tesseract and Google vision to recognize the streaming data and test it for typical factory HMIs and realistic lighting conditions. Finally, we use the meta-data to match the OCR output with the temporal, domain-dependent context of the data to improve the accuracy of the output. Our IIoT solution enables reliable and efficient data extraction which will improve the performance of subsequent AI applications.Keywords: human machine interface, industrial internet of things, internet of things, optical character recognition, video analytics
Procedia PDF Downloads 1091463 Numerical Study of Pressure Losses of Turbulence Drilling Fluid Flow in the Oil Wellbore
Authors: Alireza Mehdizadeh, Ghanbarali Sheikhzadeh
Abstract:
In this paper the pressure loss of drilling fluid flow in the annulus is investigated. On this purpose the domains between two concentric and two eccentric cylinders are considered as computational domains. In this research foam is used as drilling fluid. Firstly simulation results for laminar flow and non Newtonian fluid and different density like 100, 200, 300 kg/m3 and different inner cylinder rotational velocity like 100, 200, 300 RPM is presented. These results are compared and matched with references results. The power law and Herschel Bulkly methods are used for non Newtonian fluid modeling. After that computations are repeated with turbulence flow considering. K- Model is used for turbulence modeling. Results show that in laminar flow Herschel bulkly model has best result in comparison with power law model. And pressure loss in turbulence flow is higher than laminar flow.Keywords: simulation, concentric cylinders, drilling, non Newtonian
Procedia PDF Downloads 5661462 Automatic LV Segmentation with K-means Clustering and Graph Searching on Cardiac MRI
Authors: Hae-Yeoun Lee
Abstract:
Quantification of cardiac function is performed by calculating blood volume and ejection fraction in routine clinical practice. However, these works have been performed by manual contouring,which requires computational costs and varies on the observer. In this paper, an automatic left ventricle segmentation algorithm on cardiac magnetic resonance images (MRI) is presented. Using knowledge on cardiac MRI, a K-mean clustering technique is applied to segment blood region on a coil-sensitivity corrected image. Then, a graph searching technique is used to correct segmentation errors from coil distortion and noises. Finally, blood volume and ejection fraction are calculated. Using cardiac MRI from 15 subjects, the presented algorithm is tested and compared with manual contouring by experts to show outstanding performance.Keywords: cardiac MRI, graph searching, left ventricle segmentation, K-means clustering
Procedia PDF Downloads 3991461 Visual Improvement Outcome of Pars Plana Vitrectomy Combined Endofragmentation and Secondary IOL Implantation for Dropped Nucleus After Cataract Surgery : A Case Report
Authors: Saut Samuel Simamora
Abstract:
PURPOSE: Nucleus drop is one of the most feared and severe complications of modern cataract surgery. The lens material may drop through iatrogenic breaks of the posterior capsule. The incidence of the nucleus as the complication of phacoemulsification increases concomitant to the increased frequency of phacoemulsification. Pars plana vitrectomy (PPV) followed by endofragmentation and secondary intraocular lens (IOL) implantation is the choice of management procedure. This case report aims to present the outcome of PPV for the treatment dropped nucleus after cataract surgery METHODS: A 65 year old female patient came to Vitreoretina department with chief complaints blurry vision in her left eye after phacoemulsification one month before. Ophthalmological examination revealed visual acuity of the right eye (VA RE) was 6/15, and the left eye (VA LE) was hand movement. The intraocular pressure (IOP) on the right eye was 18 mmHg, and on the left eye was 59 mmHg. On her left eye, there were aphakic, dropped lens nucleus and secondary glaucoma.RESULTS: The patient got antiglaucoma agent until her IOP was decreased. She underwent pars plana vitrectomy to remove dropped nucleus and iris fixated IOL. One week post operative evaluation revealed VA LE was 6/7.5 and iris fixated IOL in proper position. CONCLUSIONS: Nucleus drop generally occurs in phacoemulsification cataract surgery techniques. Retained lens nucleus or fragments in the vitreous may cause severe intraocular inflammation leading to secondary glaucoma. The proper and good management for retained lens fragments in nucleus drop give excellent outcome to patient.Keywords: secondary glaucoma, complication of phacoemulsification, nucleus drop, pars plana vitrectomy
Procedia PDF Downloads 791460 Medical Diagnosis of Retinal Diseases Using Artificial Intelligence Deep Learning Models
Authors: Ethan James
Abstract:
Over one billion people worldwide suffer from some level of vision loss or blindness as a result of progressive retinal diseases. Many patients, particularly in developing areas, are incorrectly diagnosed or undiagnosed whatsoever due to unconventional diagnostic tools and screening methods. Artificial intelligence (AI) based on deep learning (DL) convolutional neural networks (CNN) have recently gained a high interest in ophthalmology for its computer-imaging diagnosis, disease prognosis, and risk assessment. Optical coherence tomography (OCT) is a popular imaging technique used to capture high-resolution cross-sections of retinas. In ophthalmology, DL has been applied to fundus photographs, optical coherence tomography, and visual fields, achieving robust classification performance in the detection of various retinal diseases including macular degeneration, diabetic retinopathy, and retinitis pigmentosa. However, there is no complete diagnostic model to analyze these retinal images that provide a diagnostic accuracy above 90%. Thus, the purpose of this project was to develop an AI model that utilizes machine learning techniques to automatically diagnose specific retinal diseases from OCT scans. The algorithm consists of neural network architecture that was trained from a dataset of over 20,000 real-world OCT images to train the robust model to utilize residual neural networks with cyclic pooling. This DL model can ultimately aid ophthalmologists in diagnosing patients with these retinal diseases more quickly and more accurately, therefore facilitating earlier treatment, which results in improved post-treatment outcomes.Keywords: artificial intelligence, deep learning, imaging, medical devices, ophthalmic devices, ophthalmology, retina
Procedia PDF Downloads 181