Search results for: deep approaches
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5565

Search results for: deep approaches

5385 A Deep-Learning Based Prediction of Pancreatic Adenocarcinoma with Electronic Health Records from the State of Maine

Authors: Xiaodong Li, Peng Gao, Chao-Jung Huang, Shiying Hao, Xuefeng B. Ling, Yongxia Han, Yaqi Zhang, Le Zheng, Chengyin Ye, Modi Liu, Minjie Xia, Changlin Fu, Bo Jin, Karl G. Sylvester, Eric Widen

Abstract:

Predicting the risk of Pancreatic Adenocarcinoma (PA) in advance can benefit the quality of care and potentially reduce population mortality and morbidity. The aim of this study was to develop and prospectively validate a risk prediction model to identify patients at risk of new incident PA as early as 3 months before the onset of PA in a statewide, general population in Maine. The PA prediction model was developed using Deep Neural Networks, a deep learning algorithm, with a 2-year electronic-health-record (EHR) cohort. Prospective results showed that our model identified 54.35% of all inpatient episodes of PA, and 91.20% of all PA that required subsequent chemoradiotherapy, with a lead-time of up to 3 months and a true alert of 67.62%. The risk assessment tool has attained an improved discriminative ability. It can be immediately deployed to the health system to provide automatic early warnings to adults at risk of PA. It has potential to identify personalized risk factors to facilitate customized PA interventions.

Keywords: cancer prediction, deep learning, electronic health records, pancreatic adenocarcinoma

Procedia PDF Downloads 122
5384 Introduction to Techno-Sectoral Innovation System Modeling and Functions Formulating

Authors: S. M. Azad, H. Ghodsi Pour, F. Roshannafasa

Abstract:

In recent years ‘technology management and policymaking’ is one of the most important problems in management science. In this field, different generations of innovation and technology management are presented which the earliest one is Innovation System (IS) approach. In a general classification, innovation systems are divided in to 4 approaches: Technical, sectoral, regional, and national. There are many researches in relation to each of these approaches in different academic fields. Every approach has some benefits. If two or more approaches hybrid, their benefits would be combined. In addition, according to the sectoral structure of the governance model in Iran, in many sectors such as information technology, the combination of three other approaches with sectoral approach is essential. Hence, in this paper, combining two IS approaches (technical and sectoral) and using system dynamics, a generic model is presented for a sample of software industry. As a complimentary point, this article is introducing a new hybrid approach called Techno-Sectoral Innovation System. This TSIS model is accomplished by Changing concepts of the ‘functions’ which came from Technological IS literature and using them into sectoral system as measurable indicators.

Keywords: innovation system, technology, techno-sectoral system, functional indicators, system dynamics

Procedia PDF Downloads 411
5383 Investigating the Factors Affecting Generalization of Deep Learning Models for Plant Disease Detection

Authors: Praveen S. Muthukumarana, Achala C. Aponso

Abstract:

A large percentage of global crop harvest is lost due to crop diseases. Timely identification and treatment of crop diseases is difficult in many developing nations due to insufficient trained professionals in the field of agriculture. Many crop diseases can be accurately diagnosed by visual symptoms. In the past decade, deep learning has been successfully utilized in domains such as healthcare but adoption in agriculture for plant disease detection is rare. The literature shows that models trained with popular datasets such as PlantVillage does not generalize well on real world images. This paper attempts to find out how to make plant disease identification models that generalize well with real world images.

Keywords: agriculture, convolutional neural network, deep learning, plant disease classification, plant disease detection, plant disease diagnosis

Procedia PDF Downloads 104
5382 Research on Key Technologies on Initial Installation of Ultra-Deep-Water Dynamic Umbilical

Authors: Weiwei Xie, Yichao Li

Abstract:

The initial installation of the umbilical can affect the subsequent installation process and final installation. Meanwhile, the design of both ends of the ultra-deep water dynamic umbilical (UDWDU), as well as the design of the surface unit and the subsea production system connected by UDWDU,], varies in different oil and gas fields. To optimize the installation process of UDWDU, on the basis of the summary and analysis of the surface-end and the subsea-end design of UDWDU and the mainstream construction resources, the method of initial installation from the surface unit side or the subsea production system side of UDWDU is studied, and each initiation installation method is pointed out if some difficulties that may be encountered.

Keywords: dynamic umbilical, ultra-deep-water, initial installation, installation process

Procedia PDF Downloads 113
5381 Identification of Breast Anomalies Based on Deep Convolutional Neural Networks and K-Nearest Neighbors

Authors: Ayyaz Hussain, Tariq Sadad

Abstract:

Breast cancer (BC) is one of the widespread ailments among females globally. The early prognosis of BC can decrease the mortality rate. Exact findings of benign tumors can avoid unnecessary biopsies and further treatments of patients under investigation. However, due to variations in images, it is a tough job to isolate cancerous cases from normal and benign ones. The machine learning technique is widely employed in the classification of BC pattern and prognosis. In this research, a deep convolution neural network (DCNN) called AlexNet architecture is employed to get more discriminative features from breast tissues. To achieve higher accuracy, K-nearest neighbor (KNN) classifiers are employed as a substitute for the softmax layer in deep learning. The proposed model is tested on a widely used breast image database called MIAS dataset for experimental purposes and achieved 99% accuracy.

Keywords: breast cancer, DCNN, KNN, mammography

Procedia PDF Downloads 105
5380 A Conv-Long Short-term Memory Deep Learning Model for Traffic Flow Prediction

Authors: Ali Reza Sattarzadeh, Ronny J. Kutadinata, Pubudu N. Pathirana, Van Thanh Huynh

Abstract:

Traffic congestion has become a severe worldwide problem, affecting everyday life, fuel consumption, time, and air pollution. The primary causes of these issues are inadequate transportation infrastructure, poor traffic signal management, and rising population. Traffic flow forecasting is one of the essential and effective methods in urban congestion and traffic management, which has attracted the attention of researchers. With the development of technology, undeniable progress has been achieved in existing methods. However, there is a possibility of improvement in the extraction of temporal and spatial features to determine the importance of traffic flow sequences and extraction features. In the proposed model, we implement the convolutional neural network (CNN) and long short-term memory (LSTM) deep learning models for mining nonlinear correlations and their effectiveness in increasing the accuracy of traffic flow prediction in the real dataset. According to the experiments, the results indicate that implementing Conv-LSTM networks increases the productivity and accuracy of deep learning models for traffic flow prediction.

Keywords: deep learning algorithms, intelligent transportation systems, spatiotemporal features, traffic flow prediction

Procedia PDF Downloads 124
5379 Novel Synthesis of Metal Oxide Nanoparticles from Type IV Deep Eutectic Solvents

Authors: Lorenzo Gontrani, Marilena Carbone, Domenica Tommasa Donia, Elvira Maria Bauer, Pietro Tagliatesta

Abstract:

One of the fields where DES shows remarkable added values is the synthesis Of inorganic materials, in particular nanoparticles. In this field, the higher- ent and highly-tunable nano-homogeneities of DES structure give origin to a marked templating effect, a precious role that has led to the recent bloom of a vast number of studies exploiting these new synthesis media to prepare Nanomaterials and composite structures of various kinds. In this contribution, the most recent developments in the field will be reviewed, and some ex-citing examples of novel metal oxide nanoparticles syntheses using non-toxic type-IV Deep Eutectic Solvents will be described. The prepared materials possess nanometric dimensions and show flower-like shapes. The use of the pre- pared nanoparticles as fluorescent materials for the detection of various contaminants is under development.

Keywords: metal deep eutectic solvents, nanoparticles, inorganic synthesis, type IV DES, lamellar

Procedia PDF Downloads 99
5378 A Study of Agile Based Approaches to Improve Software Quality

Authors: Gurmeet Kaur

Abstract:

Agile software development methods are being recognized as popular, and efficient approach to the development of software system that has a short delivery period with high quality also that meets customer requirements with zero defect. In agile software development, quality means quality of code where in the quality is maintained through the use of methods or approaches like refactoring, test driven development, behavior driven development, acceptance test driven development, and demand driven development. Software quality is measured in term of metrics such as the number of defects during development of software. Usage of above mentioned methods or approaches, reduces the possibilities of defects in developed software, and hence improve quality. This paper focuses on study of agile based quality methods or approaches for software development that ensures improved quality of software as well as reduced cost, and customer satisfaction.

Keywords: ATDD, BDD, DDD, TDD

Procedia PDF Downloads 126
5377 Facial Emotion Recognition with Convolutional Neural Network Based Architecture

Authors: Koray U. Erbas

Abstract:

Neural networks are appealing for many applications since they are able to learn complex non-linear relationships between input and output data. As the number of neurons and layers in a neural network increase, it is possible to represent more complex relationships with automatically extracted features. Nowadays Deep Neural Networks (DNNs) are widely used in Computer Vision problems such as; classification, object detection, segmentation image editing etc. In this work, Facial Emotion Recognition task is performed by proposed Convolutional Neural Network (CNN)-based DNN architecture using FER2013 Dataset. Moreover, the effects of different hyperparameters (activation function, kernel size, initializer, batch size and network size) are investigated and ablation study results for Pooling Layer, Dropout and Batch Normalization are presented.

Keywords: convolutional neural network, deep learning, deep learning based FER, facial emotion recognition

Procedia PDF Downloads 226
5376 Distributed System Computing Resource Scheduling Algorithm Based on Deep Reinforcement Learning

Authors: Yitao Lei, Xingxiang Zhai, Burra Venkata Durga Kumar

Abstract:

As the quantity and complexity of computing in large-scale software systems increase, distributed system computing becomes increasingly important. The distributed system realizes high-performance computing by collaboration between different computing resources. If there are no efficient resource scheduling resources, the abuse of distributed computing may cause resource waste and high costs. However, resource scheduling is usually an NP-hard problem, so we cannot find a general solution. However, some optimization algorithms exist like genetic algorithm, ant colony optimization, etc. The large scale of distributed systems makes this traditional optimization algorithm challenging to work with. Heuristic and machine learning algorithms are usually applied in this situation to ease the computing load. As a result, we do a review of traditional resource scheduling optimization algorithms and try to introduce a deep reinforcement learning method that utilizes the perceptual ability of neural networks and the decision-making ability of reinforcement learning. Using the machine learning method, we try to find important factors that influence the performance of distributed system computing and help the distributed system do an efficient computing resource scheduling. This paper surveys the application of deep reinforcement learning on distributed system computing resource scheduling proposes a deep reinforcement learning method that uses a recurrent neural network to optimize the resource scheduling, and proposes the challenges and improvement directions for DRL-based resource scheduling algorithms.

Keywords: resource scheduling, deep reinforcement learning, distributed system, artificial intelligence

Procedia PDF Downloads 80
5375 Effect of Depth on the Distribution of Zooplankton in Wushishi Lake Minna, Niger State, Nigeria

Authors: Adamu Zubairu Mohammed, Fransis Oforum Arimoro, Salihu Maikudi Ibrahim, Y. I. Auta, T. I. Arowosegbe, Y. Abdullahi

Abstract:

The present study was conducted to evaluate the effect of depth on the distribution of zooplankton and some physicochemical parameters in Tungan Kawo Lake (Wushishi dam). Water and zooplankton samples were collected from the surface, 3.0 meters deep and 6.0 meters deep, for a period of 24 hours for six months. Standard procedures were adopted for the determination of physicochemical parameters. Results have shown significant differences in the pH, DO, BOD Hardness, Na, and Mg. A total of 1764 zooplankton were recorded, comprising 35 species, with cladocera having 18 species (58%), 14 species of copepoda (41%), 3 species of diptera (1.0%). Results show that more of the zooplankton were recorded in the 3.0 meters-deep region compared to the two other depts and a significant difference was observed in the distribution of Ceriodaphnia dubia, Daphnia laevis, and Leptodiaptomus coloradensis. Though the most abundant zooplankton was recorded in the 3.0 meters deep, Leptodiaptomus coloradesnsis, which was observed in the 6.0 meters deep as the most individual observed, this was followed by Daphnia laevis. Canonical correspondence analysis between physicochemical parameters and the zooplankton indicated a good relationship in the Lake. Ceriodaphnia dubia was found to have a good association with oxygen, sodium, and potassium, while Daphnia laevis and Leptodiaptomus coloradensis are in good relationship with magnesium and phosphorus. It was generally observed that this depth does not have much influence on the distribution of zooplankton in Wushishi Lake.

Keywords: zooplankton, standard procedures, canonical correspondence analysis, Wushishi, canonical, physicochemical parameter

Procedia PDF Downloads 53
5374 Development of Web-Based Iceberg Detection Using Deep Learning

Authors: A. Kavya Sri, K. Sai Vineela, R. Vanitha, S. Rohith

Abstract:

Large pieces of ice that break from the glaciers are known as icebergs. The threat that icebergs pose to navigation, production of offshore oil and gas services, and underwater pipelines makes their detection crucial. In this project, an automated iceberg tracking method using deep learning techniques and satellite images of icebergs is to be developed. With a temporal resolution of 12 days and a spatial resolution of 20 m, Sentinel-1 (SAR) images can be used to track iceberg drift over the Southern Ocean. In contrast to multispectral images, SAR images are used for analysis in meteorological conditions. This project develops a web-based graphical user interface to detect and track icebergs using sentinel-1 images. To track the movement of the icebergs by using temporal images based on their latitude and longitude values and by comparing the center and area of all detected icebergs. Testing the accuracy is done by precision and recall measures.

Keywords: synthetic aperture radar (SAR), icebergs, deep learning, spatial resolution, temporal resolution

Procedia PDF Downloads 55
5373 Deep Supervision Based-Unet to Detect Buildings Changes from VHR Aerial Imagery

Authors: Shimaa Holail, Tamer Saleh, Xiongwu Xiao

Abstract:

Building change detection (BCD) from satellite imagery is an essential topic in urbanization monitoring, agricultural land management, and updating geospatial databases. Recently, methods for detecting changes based on deep learning have made significant progress and impressive results. However, it has the problem of being insensitive to changes in buildings with complex spectral differences, and the features being extracted are not discriminatory enough, resulting in incomplete buildings and irregular boundaries. To overcome these problems, we propose a dual Siamese network based on the Unet model with the addition of a deep supervision strategy (DS) in this paper. This network consists of a backbone (encoder) based on ImageNet pre-training, a fusion block, and feature pyramid networks (FPN) to enhance the step-by-step information of the changing regions and obtain a more accurate BCD map. To train the proposed method, we created a new dataset (EGY-BCD) of high-resolution and multi-temporal aerial images captured over New Cairo in Egypt to detect building changes for this purpose. The experimental results showed that the proposed method is effective and performs well with the EGY-BCD dataset regarding the overall accuracy, F1-score, and mIoU, which were 91.6 %, 80.1 %, and 73.5 %, respectively.

Keywords: building change detection, deep supervision, semantic segmentation, EGY-BCD dataset

Procedia PDF Downloads 71
5372 Computer Aided Analysis of Breast Based Diagnostic Problems from Mammograms Using Image Processing and Deep Learning Methods

Authors: Ali Berkan Ural

Abstract:

This paper presents the analysis, evaluation, and pre-diagnosis of early stage breast based diagnostic problems (breast cancer, nodulesorlumps) by Computer Aided Diagnosing (CAD) system from mammogram radiological images. According to the statistics, the time factor is crucial to discover the disease in the patient (especially in women) as possible as early and fast. In the study, a new algorithm is developed using advanced image processing and deep learning method to detect and classify the problem at earlystagewithmoreaccuracy. This system first works with image processing methods (Image acquisition, Noiseremoval, Region Growing Segmentation, Morphological Operations, Breast BorderExtraction, Advanced Segmentation, ObtainingRegion Of Interests (ROIs), etc.) and segments the area of interest of the breast and then analyzes these partly obtained area for cancer detection/lumps in order to diagnosis the disease. After segmentation, with using the Spectrogramimages, 5 different deep learning based methods (specified Convolutional Neural Network (CNN) basedAlexNet, ResNet50, VGG16, DenseNet, Xception) are applied to classify the breast based problems.

Keywords: computer aided diagnosis, breast cancer, region growing, segmentation, deep learning

Procedia PDF Downloads 62
5371 Visual Analytics of Higher Order Information for Trajectory Datasets

Authors: Ye Wang, Ickjai Lee

Abstract:

Due to the widespread of mobile sensing, there is a strong need to handle trails of moving objects, trajectories. This paper proposes three visual analytic approaches for higher order information of trajectory data sets based on the higher order Voronoi diagram data structure. Proposed approaches reveal geometrical information, topological, and directional information. Experimental results demonstrate the applicability and usefulness of proposed three approaches.

Keywords: visual analytics, higher order information, trajectory datasets, spatio-temporal data

Procedia PDF Downloads 380
5370 User Modeling from the Perspective of Improvement in Search Results: A Survey of the State of the Art

Authors: Samira Karimi-Mansoub, Rahem Abri

Abstract:

Currently, users expect high quality and personalized information from search results. To satisfy user’s needs, personalized approaches to web search have been proposed. These approaches can provide the most appropriate answer for user’s needs by using user context and incorporating information about query provided by combining search technologies. To carry out personalized web search, there is a need to make different techniques on whole of user search process. There are the number of possible deployment of personalized approaches such as personalized web search, personalized recommendation, personalized summarization and filtering systems and etc. but the common feature of all approaches in various domains is that user modeling is utilized to provide personalized information from the Web. So the most important work in personalized approaches is user model mining. User modeling applications and technologies can be used in various domains depending on how the user collected information may be extracted. In addition to, the used techniques to create user model is also different in each of these applications. Since in the previous studies, there was not a complete survey in this field, our purpose is to present a survey on applications and techniques of user modeling from the viewpoint of improvement in search results by considering the existing literature and researches.

Keywords: filtering systems, personalized web search, user modeling, user search behavior

Procedia PDF Downloads 245
5369 Comparison Study between Deep Mixed Columns and Encased Sand Column for Soft Clay Soil in Egypt

Authors: Walid El Kamash

Abstract:

Sand columns (or granular piles) can be employed as soil strengthening for flexible constructions such as road embankments, oil storage tanks in addition to multistory structures. The challenge of embedding the sand columns in soft soil is that the surrounding soft soil cannot avail the enough confinement stress in order to keep the form of the sand column. Therefore, the sand columns which were installed in such soil will lose their ability to perform needed load-bearing capacity. The encasement, besides increasing the strength and stiffness of the sand column, prevents the lateral squeezing of sands when the column is installed even in extremely soft soils, thus enabling quicker and more economical installation. This paper investigates the improvement in load capacity of the sand column by encasement through a comprehensive parametric study using the 3-D finite difference analysis for the soft clay of soil in Egypt. Moreover, the study was extended to include a comparison study between encased sand column and Deep Mixed columns (DM). The study showed that confining the sand by geosynthetic resulted in an increment of shear strength. That result paid the attention to use encased sand stone rather than deep mixed columns due to relative high permeability of the first material.

Keywords: encased sand column, Deep mixed column, numerical analysis, improving soft soil

Procedia PDF Downloads 345
5368 Utilizing Federated Learning for Accurate Prediction of COVID-19 from CT Scan Images

Authors: Jinil Patel, Sarthak Patel, Sarthak Thakkar, Deepti Saraswat

Abstract:

Recently, the COVID-19 outbreak has spread across the world, leading the World Health Organization to classify it as a global pandemic. To save the patient’s life, the COVID-19 symptoms have to be identified. But using an AI (Artificial Intelligence) model to identify COVID-19 symptoms within the allotted time was challenging. The RT-PCR test was found to be inadequate in determining the COVID status of a patient. To determine if the patient has COVID-19 or not, a Computed Tomography Scan (CT scan) of patient is a better alternative. It will be challenging to compile and store all the data from various hospitals on the server, though. Federated learning, therefore, aids in resolving this problem. Certain deep learning models help to classify Covid-19. This paper will have detailed work of certain deep learning models like VGG19, ResNet50, MobileNEtv2, and Deep Learning Aggregation (DLA) along with maintaining privacy with encryption.

Keywords: federated learning, COVID-19, CT-scan, homomorphic encryption, ResNet50, VGG-19, MobileNetv2, DLA

Procedia PDF Downloads 42
5367 Automatic Identification of Aquatic Insects Based on Deep Learning and Computer Vision

Authors: Predrag Simović, Katarina Stojanović, Milena Radenković, Dimitrija Savić Zdravković, Aleksandar Milosavljević, Bratislav Predić, Milenka Božanić, Ana Petrović, Djuradj Milošević

Abstract:

Mayflies (Ephemeroptera), stoneflies (Plecoptera), and caddisflies (Trichoptera) (collectively referred to as EPT) are key participants in most freshwater habitats and often exhibit high diversity. Moreover, their presence and relative abundance are used in freshwater ecological and biomonitoring studies. Current methods for freshwater ecosystem biomonitoring follow a traditional approach of taxa monitoring based on morphological characters, which is time-consuming, and often generates data sets with low taxonomic resolution and unverifiable identification precision. To assist in solving identification problems and contribute to the knowledge of the distribution of many species, there was a need to develop alternative approaches in macroinvertebrate sample identification. Here, we establish an automatic machine-based identification approach for EPT taxa (Insect) using deep Convolutional Neural Networks (CNNs) and computer vision to increase the efficiency and taxonomic resolution in biomonitoring. The 5 550 specimens were collected from freshwater ecosystems of Serbia, and the deep model was built upon 90 EPT taxa. The protocol for obtaining images included the following stages: taxonomic identification by human experts and DNA barcoding validation, mounting the larvae, and photographing the dorsal side using a stereomicroscope and camera (16 650 individuals). The most informative image regions (the dorsal segments of individuals) for the decision-making process in the deep learning model were visualized using Gradient Weighted Class Activation Mapping (Grad-CAM). After training the artificial neural network, a CNN model was then built that was able to classify the 90 EPT taxa into their respective taxonomic categories automatically with 98.7%. Our approach offers a straightforward and efficient solution for routine monitoring programs, focusing on key biotic descriptors, such as EPT taxa. In addition, this application provides a streamlined solution that not only saves time, reduces equipment and expert requirements but also significantly enhances reliability and information content. The identification of the EPT larvae is difficult because of the variation of morphological features even within a single genus or the close resemblance of several species, and therefore, future research should focus on increasing the number of entities (species) in the model.

Keywords: convolutional neural networks, DNA barcoding, EPT taxa, biomonitoring

Procedia PDF Downloads 43
5366 A Student Centered Learning Environment in Engineering Education: Design and a Longitudinal Study of Impact

Authors: Tom O'Mahony

Abstract:

This article considers the design of a student-centered learning environment in engineering education. The learning environment integrates a number of components, including project-based learning, collaborative learning, two-stage assignments, active learning lectures, and a flipped-classroom. Together these elements place the individual learner and their learning at the center of the environment by focusing on understanding, enhancing relevance, applying learning, obtaining rich feedback, making choices, and taking responsibility. The evolution of this environment from 2014 to the present day is outlined. The impact of this environment on learners and their learning is evaluated via student questionnaires that consist of both open and closed-ended questions. The closed questions indicate that students found the learning environment to be really interesting and enjoyable (rated as 4.7 on a 5 point scale) and encouraged students to adopt a deep approach towards studying the course materials (rated as 4.0 on a 5 point scale). A content analysis of the open-ended questions provides evidence that the project, active learning lectures, and flipped classroom all contribute to the success of this environment. Furthermore, this analysis indicates that the two-stage assessment process, in which feedback is provided between a draft and final assignment, is the key component and the dominant theme. A limitation of the study is the small class size (less than 20 learners per year), but, to some degree, this is compensated for by the longitudinal nature of the study.

Keywords: deep approaches, formative assessment, project-based learning, student-centered learning

Procedia PDF Downloads 81
5365 Refined Edge Detection Network

Authors: Omar Elharrouss, Youssef Hmamouche, Assia Kamal Idrissi, Btissam El Khamlichi, Amal El Fallah-Seghrouchni

Abstract:

Edge detection is represented as one of the most challenging tasks in computer vision, due to the complexity of detecting the edges or boundaries in real-world images that contains objects of different types and scales like trees, building as well as various backgrounds. Edge detection is represented also as a key task for many computer vision applications. Using a set of backbones as well as attention modules, deep-learning-based methods improved the detection of edges compared with the traditional methods like Sobel and Canny. However, images of complex scenes still represent a challenge for these methods. Also, the detected edges using the existing approaches suffer from non-refined results while the image output contains many erroneous edges. To overcome this, n this paper, by using the mechanism of residual learning, a refined edge detection network is proposed (RED-Net). By maintaining the high resolution of edges during the training process, and conserving the resolution of the edge image during the network stage, we make the pooling outputs at each stage connected with the output of the previous layer. Also, after each layer, we use an affined batch normalization layer as an erosion operation for the homogeneous region in the image. The proposed methods are evaluated using the most challenging datasets including BSDS500, NYUD, and Multicue. The obtained results outperform the designed edge detection networks in terms of performance metrics and quality of output images.

Keywords: edge detection, convolutional neural networks, deep learning, scale-representation, backbone

Procedia PDF Downloads 65
5364 The Content-Based Classroom: Perspectives on Integrating Language and Content

Authors: Mourad Ben Bennani

Abstract:

Views of language and language learning have undergone a tremendous change over the last decades. Language is no longer seen as a set of structured rules. It is rather viewed as a tool of interaction and communication. This shift in views has resulted in change in viewing language learning, which gave birth to various approaches and methodologies of language teaching. Two of these approaches are content-based instruction and content and language integrated learning (CLIL). These are similar approaches which integrate content and foreign/second language learning through various methodologies and models as a result of different implementations around the world. This presentation deals with sociocultural view of CBI and CLIL. It also defines language and content as vital components of CBI and CLIL. Next it reviews the origins of CBI and the continuum perspectives and CLIL definitions and models featured in the literature. Finally it summarizes current aspects around research in program evaluation with a focus on the benefits and challenges of these innovative approaches for second language teaching.

Keywords: CBI, CLIL, CBI continuum, CLIL models

Procedia PDF Downloads 385
5363 A Deep Learning Based Integrated Model For Spatial Flood Prediction

Authors: Vinayaka Gude Divya Sampath

Abstract:

The research introduces an integrated prediction model to assess the susceptibility of roads in a future flooding event. The model consists of deep learning algorithm for forecasting gauge height data and Flood Inundation Mapper (FIM) for spatial flooding. An optimal architecture for Long short-term memory network (LSTM) was identified for the gauge located on Tangipahoa River at Robert, LA. Dropout was applied to the model to evaluate the uncertainty associated with the predictions. The estimates are then used along with FIM to identify the spatial flooding. Further geoprocessing in ArcGIS provides the susceptibility values for different roads. The model was validated based on the devastating flood of August 2016. The paper discusses the challenges for generalization the methodology for other locations and also for various types of flooding. The developed model can be used by the transportation department and other emergency response organizations for effective disaster management.

Keywords: deep learning, disaster management, flood prediction, urban flooding

Procedia PDF Downloads 110
5362 Single-Camera Basketball Tracker through Pose and Semantic Feature Fusion

Authors: Adrià Arbués-Sangüesa, Coloma Ballester, Gloria Haro

Abstract:

Tracking sports players is a widely challenging scenario, specially in single-feed videos recorded in tight courts, where cluttering and occlusions cannot be avoided. This paper presents an analysis of several geometric and semantic visual features to detect and track basketball players. An ablation study is carried out and then used to remark that a robust tracker can be built with Deep Learning features, without the need of extracting contextual ones, such as proximity or color similarity, nor applying camera stabilization techniques. The presented tracker consists of: (1) a detection step, which uses a pretrained deep learning model to estimate the players pose, followed by (2) a tracking step, which leverages pose and semantic information from the output of a convolutional layer in a VGG network. Its performance is analyzed in terms of MOTA over a basketball dataset with more than 10k instances.

Keywords: basketball, deep learning, feature extraction, single-camera, tracking

Procedia PDF Downloads 109
5361 DLtrace: Toward Understanding and Testing Deep Learning Information Flow in Deep Learning-Based Android Apps

Authors: Jie Zhang, Qianyu Guo, Tieyi Zhang, Zhiyong Feng, Xiaohong Li

Abstract:

With the widespread popularity of mobile devices and the development of artificial intelligence (AI), deep learning (DL) has been extensively applied in Android apps. Compared with traditional Android apps (traditional apps), deep learning based Android apps (DL-based apps) need to use more third-party application programming interfaces (APIs) to complete complex DL inference tasks. However, existing methods (e.g., FlowDroid) for detecting sensitive information leakage in Android apps cannot be directly used to detect DL-based apps as they are difficult to detect third-party APIs. To solve this problem, we design DLtrace; a new static information flow analysis tool that can effectively recognize third-party APIs. With our proposed trace and detection algorithms, DLtrace can also efficiently detect privacy leaks caused by sensitive APIs in DL-based apps. Moreover, using DLtrace, we summarize the non-sequential characteristics of DL inference tasks in DL-based apps and the specific functionalities provided by DL models for such apps. We propose two formal definitions to deal with the common polymorphism and anonymous inner-class problems in the Android static analyzer. We conducted an empirical assessment with DLtrace on 208 popular DL-based apps in the wild and found that 26.0% of the apps suffered from sensitive information leakage. Furthermore, DLtrace has a more robust performance than FlowDroid in detecting and identifying third-party APIs. The experimental results demonstrate that DLtrace expands FlowDroid in understanding DL-based apps and detecting security issues therein.

Keywords: mobile computing, deep learning apps, sensitive information, static analysis

Procedia PDF Downloads 124
5360 Turning Points in the Development of Translator Training in the West from the 1980s to the Present

Authors: B. Sayaheen

Abstract:

The translator’s competence is one of the topics that has received a great deal of research in the field of translation studies because such competencies are still debatable and not yet agreed upon. Besides, scholars tackle this topic from different points of view. Approaches to teaching these competencies have gone through some developments. This paper aims at investigating these developments, exploring the major turning points and shifts in the developments of teaching methods in translator training. The significance of these turning points and the external or internal causes will also be discussed. Based on the past and present status of teaching approaches in translator training, this paper tries to predict the future of these approaches. This paper is mainly concerned with developments of teaching approaches in the West since the 1980s to the present. The reason behind choosing this specific period is not because translator training started in the 1980s but because most criticism of the teacher-centered approach started at that time. The implications of this research stem from the fact that it identifies the turning points and the causes that led teachers to adopt student-centered approaches rather than teacher-centered approaches and then to incorporate technology and the Internet in translator training. These reasons were classified as external or internal reasons. Translation programs in the West and in other cultures can benefit from this study. Translation programs in the West can notice that teaching translation is geared toward incorporating more technologies. If these programs already use technology and the Internet to teach translation, they might benefit from the assumed future direction of teaching translation. On the other hand, some non-Western countries, and to be specific some professors, are still applying the teacher-centered approach. Moreover, these programs should include technology and the Internet in their teaching approaches to meet the drastic changes in the translation process, which seems to rely more on software and technologies to accomplish the translator’s tasks. Finally, translator training has borrowed many of its approaches from other disciplines, mainly language teaching. The teaching approaches in translator training have gone through some developments, from teacher-centered to student-centered and then toward the integration of technologies and the Internet. Both internal and external causes have played a crucial role in these developments. These borrowed approaches should be comprehensively evaluated in order to see if they achieve the goals of translator training. Such evaluation may lead us to come up with new teaching approaches developed specifically for translator training. While considering these methods and designing new approaches, we need to keep an eye on the future needs of the market.

Keywords: turning points, developments, translator training, market, The West

Procedia PDF Downloads 85
5359 How to Guide Students from Surface to Deep Learning: Applied Philosophy in Management Education

Authors: Lihong Wu, Raymond Young

Abstract:

The ability to learn is one of the most critical skills in the information age. However, many students do not have a clear understanding of what learning is, what they are learning, and why they are learning. Many students study simply to pass rather than to learn something useful for their career and their life. They have a misconception about learning and a wrong attitude towards learning. This research explores student attitudes to study in management education and explores how to intercede to lead students from shallow to deeper modes of learning.

Keywords: knowledge, surface learning, deep learning, education

Procedia PDF Downloads 466
5358 Upconversion Nanomaterials for Applications in Life Sciences and Medicine

Authors: Yong Zhang

Abstract:

Light has proven to be useful in a wide range of biomedical applications such as fluorescence imaging, photoacoustic imaging, optogenetics, photodynamic therapy, photothermal therapy, and light controlled drug/gene delivery. Taking photodynamic therapy (PDT) as an example, PDT has been proven clinically effective in early lung cancer, bladder cancer, head, and neck cancer and is the primary treatment for skin cancer as well. However, clinical use of PDT is severely constrained by the low penetration depth of visible light through thick tissue, limiting its use to target regions only a few millimeters deep. One way to enhance the range is to use invisible near-infrared (NIR) light within the optical window (700–1100nm) for biological tissues, extending the depth up to 1cm with no observable damage to the intervening tissue. We have demonstrated use of NIR-to-visible upconversion fluorescent nanoparticles (UCNPs), emitting visible fluorescence when excited by a NIR light at 980nm, as a nanotransducer for PDT to convert deep tissue-penetrating NIR light to visible light suitable for activating photosensitizers. The unique optical properties of UCNPs enable the upconversion wavelength to be tuned and matched to the activation absorption wavelength of the photosensitizer. At depths beyond 1cm, however, tissue remains inaccessible to light even within the NIR window, and this critical depth limitation renders existing phototherapy ineffective against most deep-seated cancers. We have demonstrated some new treatment modalities for deep-seated cancers based on UCNP hydrogel implants and miniaturized, wirelessly powered optoelectronic devices for light delivery to deep tissues.

Keywords: upconversion, fluorescent, nanoparticle, bioimaging, photodynamic therapy

Procedia PDF Downloads 128
5357 Enhancer: An Effective Transformer Architecture for Single Image Super Resolution

Authors: Pitigalage Chamath Chandira Peiris

Abstract:

A widely researched domain in the field of image processing in recent times has been single image super-resolution, which tries to restore a high-resolution image from a single low-resolution image. Many more single image super-resolution efforts have been completed utilizing equally traditional and deep learning methodologies, as well as a variety of other methodologies. Deep learning-based super-resolution methods, in particular, have received significant interest. As of now, the most advanced image restoration approaches are based on convolutional neural networks; nevertheless, only a few efforts have been performed using Transformers, which have demonstrated excellent performance on high-level vision tasks. The effectiveness of CNN-based algorithms in image super-resolution has been impressive. However, these methods cannot completely capture the non-local features of the data. Enhancer is a simple yet powerful Transformer-based approach for enhancing the resolution of images. A method for single image super-resolution was developed in this study, which utilized an efficient and effective transformer design. This proposed architecture makes use of a locally enhanced window transformer block to alleviate the enormous computational load associated with non-overlapping window-based self-attention. Additionally, it incorporates depth-wise convolution in the feed-forward network to enhance its ability to capture local context. This study is assessed by comparing the results obtained for popular datasets to those obtained by other techniques in the domain.

Keywords: single image super resolution, computer vision, vision transformers, image restoration

Procedia PDF Downloads 74
5356 Evaluation of Modern Natural Language Processing Techniques via Measuring a Company's Public Perception

Authors: Burak Oksuzoglu, Savas Yildirim, Ferhat Kutlu

Abstract:

Opinion mining (OM) is one of the natural language processing (NLP) problems to determine the polarity of opinions, mostly represented on a positive-neutral-negative axis. The data for OM is usually collected from various social media platforms. In an era where social media has considerable control over companies’ futures, it’s worth understanding social media and taking actions accordingly. OM comes to the fore here as the scale of the discussion about companies increases, and it becomes unfeasible to gauge opinion on individual levels. Thus, the companies opt to automize this process by applying machine learning (ML) approaches to their data. For the last two decades, OM or sentiment analysis (SA) has been mainly performed by applying ML classification algorithms such as support vector machines (SVM) and Naïve Bayes to a bag of n-gram representations of textual data. With the advent of deep learning and its apparent success in NLP, traditional methods have become obsolete. Transfer learning paradigm that has been commonly used in computer vision (CV) problems started to shape NLP approaches and language models (LM) lately. This gave a sudden rise to the usage of the pretrained language model (PTM), which contains language representations that are obtained by training it on the large datasets using self-supervised learning objectives. The PTMs are further fine-tuned by a specialized downstream task dataset to produce efficient models for various NLP tasks such as OM, NER (Named-Entity Recognition), Question Answering (QA), and so forth. In this study, the traditional and modern NLP approaches have been evaluated for OM by using a sizable corpus belonging to a large private company containing about 76,000 comments in Turkish: SVM with a bag of n-grams, and two chosen pre-trained models, multilingual universal sentence encoder (MUSE) and bidirectional encoder representations from transformers (BERT). The MUSE model is a multilingual model that supports 16 languages, including Turkish, and it is based on convolutional neural networks. The BERT is a monolingual model in our case and transformers-based neural networks. It uses a masked language model and next sentence prediction tasks that allow the bidirectional training of the transformers. During the training phase of the architecture, pre-processing operations such as morphological parsing, stemming, and spelling correction was not used since the experiments showed that their contribution to the model performance was found insignificant even though Turkish is a highly agglutinative and inflective language. The results show that usage of deep learning methods with pre-trained models and fine-tuning achieve about 11% improvement over SVM for OM. The BERT model achieved around 94% prediction accuracy while the MUSE model achieved around 88% and SVM did around 83%. The MUSE multilingual model shows better results than SVM, but it still performs worse than the monolingual BERT model.

Keywords: BERT, MUSE, opinion mining, pretrained language model, SVM, Turkish

Procedia PDF Downloads 112