Search results for: artificial neural network approach
17362 A Radiomics Approach to Predict the Evolution of Prostate Imaging Reporting and Data System Score 3/5 Prostate Areas in Multiparametric Magnetic Resonance
Authors: Natascha C. D'Amico, Enzo Grossi, Giovanni Valbusa, Ala Malasevschi, Gianpiero Cardone, Sergio Papa
Abstract:
Purpose: To characterize, through a radiomic approach, the nature of areas classified PI-RADS (Prostate Imaging Reporting and Data System) 3/5, recognized in multiparametric prostate magnetic resonance with T2-weighted (T2w), diffusion and perfusion sequences with paramagnetic contrast. Methods and Materials: 24 cases undergoing multiparametric prostate MR and biopsy were admitted to this pilot study. Clinical outcome of the PI-RADS 3/5 was found through biopsy, finding 8 malignant tumours. The analysed images were acquired with a Philips achieva 1.5T machine with a CE- T2-weighted sequence in the axial plane. Semi-automatic tumour segmentation was carried out on MR images using 3DSlicer image analysis software. 45 shape-based, intensity-based and texture-based features were extracted and represented the input for preprocessing. An evolutionary algorithm (a TWIST system based on KNN algorithm) was used to subdivide the dataset into training and testing set and select features yielding the maximal amount of information. After this pre-processing 20 input variables were selected and different machine learning systems were used to develop a predictive model based on a training testing crossover procedure. Results: The best machine learning system (three-layers feed-forward neural network) obtained a global accuracy of 90% ( 80 % sensitivity and 100% specificity ) with a ROC of 0.82. Conclusion: Machine learning systems coupled with radiomics show a promising potential in distinguishing benign from malign tumours in PI-RADS 3/5 areas.Keywords: machine learning, MR prostate, PI-Rads 3, radiomics
Procedia PDF Downloads 18817361 The Effect of Artificial Intelligence on Accounting and Finance
Authors: Evrime Fawzy Ishak Gadelsayed
Abstract:
This paper presents resource intake accounting as an inventive manner to cope with control accounting, which concentrates on administrators as the crucial customers of the information and offers satisfactory statistics of conventional control accounting. This machine underscores that the association's asset motivates prices; as a consequence, in costing frameworks, the emphasis ought to be on assets and their usage. Resource consumption accounting consolidates two costing methodologies, action-based totally and the German cost accounting approach called GPK. This methodology, however, is a danger to managers when making the management accounting undertaking operational. The motive for this article is to clarify the concept of resource intake accounting, its elements and highlights and use of this approach in associations. Inside the first area, we present useful resource consumption accounting, the basis, reasons for its improvement, and the issues that are faced beyond costing frameworks. At that point, we deliver the requirements and presumptions of this approach; ultimately, we depict the execution of this approach in associations and its preferences over other costing techniques.Keywords: financial statement fraud, forensic accounting, fraud prevention and detection, auditing, audit expectation gap, corporate governance resource consumption accounting, management accounting, action based method, German cost accounting method
Procedia PDF Downloads 717360 Detection of Alzheimer's Protein on Nano Designed Polymer Surfaces in Water and Artificial Saliva
Authors: Sevde Altuntas, Fatih Buyukserin
Abstract:
Alzheimer’s disease is responsible for irreversible neural damage of brain parts. One of the disease markers is Amyloid-β 1-42 protein that accumulates in the brain in the form plaques. The basic problem for detection of the protein is the low amount of protein that cannot be detected properly in body liquids such as blood, saliva or urine. To solve this problem, tests like ELISA or PCR are proposed which are expensive, require specialized personnel and can contain complex protocols. Therefore, Surface-enhanced Raman Spectroscopy (SERS) a good candidate for detection of Amyloid-β 1-42 protein. Because the spectroscopic technique can potentially allow even single molecule detection from liquid and solid surfaces. Besides SERS signal can be improved by using nanopattern surface and also is specific to molecules. In this context, our study proposes to fabricate diagnostic test models that utilize Au-coated nanopatterned polycarbonate (PC) surfaces modified with Thioflavin - T to detect low concentrations of Amyloid-β 1-42 protein in water and artificial saliva medium by the enhancement of protein SERS signal. The nanopatterned PC surface that was used to enhance SERS signal was fabricated by using Anodic Alumina Membranes (AAM) as a template. It is possible to produce AAMs with different column structures and varying thicknesses depending on voltage and anodization time. After fabrication process, the pore diameter of AAMs can be arranged with dilute acid solution treatment. In this study, two different columns structures were prepared. After a surface modification to decrease their surface energy, AAMs were treated with PC solution. Following the solvent evaporation, nanopatterned PC films with tunable pillared structures were peeled off from the membrane surface. The PC film was then modified with Au and Thioflavin-T for the detection of Amyloid-β 1-42 protein. The protein detection studies were conducted first in water via this biosensor platform. Same measurements were conducted in artificial saliva to detect the presence of Amyloid Amyloid-β 1-42 protein. SEM, SERS and contact angle measurements were carried out for the characterization of different surfaces and further demonstration of the protein attachment. SERS enhancement factor calculations were also completed via experimental results. As a result, our research group fabricated diagnostic test models that utilize Au-coated nanopatterned polycarbonate (PC) surfaces modified with Thioflavin-T to detect low concentrations of Alzheimer’s Amiloid – β protein in water and artificial saliva medium. This work was supported by The Scientific and Technological Research Council of Turkey (TUBITAK) Grant No: 214Z167.Keywords: alzheimer, anodic aluminum oxide, nanotopography, surface enhanced Raman spectroscopy
Procedia PDF Downloads 29117359 A Network of Nouns and Their Features :A Neurocomputational Study
Authors: Skiker Kaoutar, Mounir Maouene
Abstract:
Neuroimaging studies indicate that a large fronto-parieto-temporal network support nouns and their features, with some areas store semantic knowledge (visual, auditory, olfactory, gustatory,…), other areas store lexical representation and other areas are implicated in general semantic processing. However, it is not well understood how this fronto-parieto-temporal network can be modulated by different semantic tasks and different semantic relations between nouns. In this study, we combine a behavioral semantic network, functional MRI studies involving object’s related nouns and brain network studies to explain how different semantic tasks and different semantic relations between nouns can modulate the activity within the brain network of nouns and their features. We first describe how nouns and their features form a large scale brain network. For this end, we examine the connectivities between areas recruited during the processing of nouns to know which configurations of interaction areas are possible. We can thus identify if, for example, brain areas that store semantic knowledge communicate via functional/structural links with areas that store lexical representations. Second, we examine how this network is modulated by different semantic tasks involving nouns and finally, we examine how category specific activation may result from the semantic relations among nouns. The results indicate that brain network of nouns and their features is highly modulated and flexible by different semantic tasks and semantic relations. At the end, this study can be used as a guide to help neurosientifics to interpret the pattern of fMRI activations detected in the semantic processing of nouns. Specifically; this study can help to interpret the category specific activations observed extensively in a large number of neuroimaging studies and clinical studies.Keywords: nouns, features, network, category specificity
Procedia PDF Downloads 52117358 Proposed Anticipating Learning Classifier System for Cloud Intrusion Detection (ALCS-CID)
Authors: Wafa' Slaibi Alsharafat
Abstract:
Cloud computing is a modern approach in network environment. According to increased number of network users and online systems, there is a need to help these systems to be away from unauthorized resource access and detect any attempts for privacy contravention. For that purpose, Intrusion Detection System is an effective security mechanism to detect any attempts of attacks for cloud resources and their information. In this paper, Cloud Intrusion Detection System has been proposed in term of reducing or eliminating any attacks. This model concerns about achieving high detection rate after conducting a set of experiments using benchmarks dataset called KDD'99.Keywords: IDS, cloud computing, anticipating classifier system, intrusion detection
Procedia PDF Downloads 47417357 Understanding Cognitive Fatigue From FMRI Scans With Self-supervised Learning
Authors: Ashish Jaiswal, Ashwin Ramesh Babu, Mohammad Zaki Zadeh, Fillia Makedon, Glenn Wylie
Abstract:
Functional magnetic resonance imaging (fMRI) is a neuroimaging technique that records neural activations in the brain by capturing the blood oxygen level in different regions based on the task performed by a subject. Given fMRI data, the problem of predicting the state of cognitive fatigue in a person has not been investigated to its full extent. This paper proposes tackling this issue as a multi-class classification problem by dividing the state of cognitive fatigue into six different levels, ranging from no-fatigue to extreme fatigue conditions. We built a spatio-temporal model that uses convolutional neural networks (CNN) for spatial feature extraction and a long short-term memory (LSTM) network for temporal modeling of 4D fMRI scans. We also applied a self-supervised method called MoCo (Momentum Contrast) to pre-train our model on a public dataset BOLD5000 and fine-tuned it on our labeled dataset to predict cognitive fatigue. Our novel dataset contains fMRI scans from Traumatic Brain Injury (TBI) patients and healthy controls (HCs) while performing a series of N-back cognitive tasks. This method establishes a state-of-the-art technique to analyze cognitive fatigue from fMRI data and beats previous approaches to solve this problem.Keywords: fMRI, brain imaging, deep learning, self-supervised learning, contrastive learning, cognitive fatigue
Procedia PDF Downloads 18917356 Accounting for Downtime Effects in Resilience-Based Highway Network Restoration Scheduling
Authors: Zhenyu Zhang, Hsi-Hsien Wei
Abstract:
Highway networks play a vital role in post-disaster recovery for disaster-damaged areas. Damaged bridges in such networks can disrupt the recovery activities by impeding the transportation of people, cargo, and reconstruction resources. Therefore, rapid restoration of damaged bridges is of paramount importance to long-term disaster recovery. In the post-disaster recovery phase, the key to restoration scheduling for a highway network is prioritization of bridge-repair tasks. Resilience is widely used as a measure of the ability to recover with which a network can return to its pre-disaster level of functionality. In practice, highways will be temporarily blocked during the downtime of bridge restoration, leading to the decrease of highway-network functionality. The failure to take downtime effects into account can lead to overestimation of network resilience. Additionally, post-disaster recovery of highway networks is generally divided into emergency bridge repair (EBR) in the response phase and long-term bridge repair (LBR) in the recovery phase, and both of EBR and LBR are different in terms of restoration objectives, restoration duration, budget, etc. Distinguish these two phases are important to precisely quantify highway network resilience and generate suitable restoration schedules for highway networks in the recovery phase. To address the above issues, this study proposes a novel resilience quantification method for the optimization of long-term bridge repair schedules (LBRS) taking into account the impact of EBR activities and restoration downtime on a highway network’s functionality. A time-dependent integer program with recursive functions is formulated for optimally scheduling LBR activities. Moreover, since uncertainty always exists in the LBRS problem, this paper extends the optimization model from the deterministic case to the stochastic case. A hybrid genetic algorithm that integrates a heuristic approach into a traditional genetic algorithm to accelerate the evolution process is developed. The proposed methods are tested using data from the 2008 Wenchuan earthquake, based on a regional highway network in Sichuan, China, consisting of 168 highway bridges on 36 highways connecting 25 cities/towns. The results show that, in this case, neglecting the bridge restoration downtime can lead to approximately 15% overestimation of highway network resilience. Moreover, accounting for the impact of EBR on network functionality can help to generate a more specific and reasonable LBRS. The theoretical and practical values are as follows. First, the proposed network recovery curve contributes to comprehensive quantification of highway network resilience by accounting for the impact of both restoration downtime and EBR activities on the recovery curves. Moreover, this study can improve the highway network resilience from the organizational dimension by providing bridge managers with optimal LBR strategies.Keywords: disaster management, highway network, long-term bridge repair schedule, resilience, restoration downtime
Procedia PDF Downloads 15017355 Distraction from Pain: An fMRI Study on the Role of Age-Related Changes in Executive Functions
Authors: Katharina M. Rischer, Angelika Dierolf, Ana M. Gonzalez-Roldan, Pedro Montoya, Fernand Anton, Marian van der Meulen
Abstract:
Even though age has been associated with increased and prolonged episodes of pain, little is known about potential age-related changes in the ˈtop-downˈ modulation of pain, such as cognitive distraction from pain. The analgesic effects of distraction result from competition for attentional resources in the prefrontal cortex (PFC), a region that is also involved in executive functions. Given that the PFC shows pronounced age-related atrophy, distraction may be less effective in reducing pain in older compared to younger adults. The aim of this study was to investigate the influence of aging on task-related analgesia and the underpinning neural mechanisms, with a focus on the role of executive functions in distraction from pain. In a first session, 64 participants (32 young adults: 26.69 ± 4.14 years; 32 older adults: 68.28 ± 7.00 years) completed a battery of neuropsychological tests. In a second session, participants underwent a pain distraction paradigm, while fMRI images were acquired. In this paradigm, participants completed a low (0-back) and a high (2-back) load condition of a working memory task while receiving either warm or painful thermal stimuli to their lower arm. To control for age-related differences in sensitivity to pain and perceived task difficulty, stimulus intensity, and task speed were individually calibrated. Results indicate that both age groups showed significantly reduced activity in a network of regions involved in pain processing when completing the high load distraction task; however, young adults showed a larger neural distraction effect in different parts of the insula and the thalamus. Moreover, better executive functions, in particular inhibitory control abilities, were associated with a larger behavioral and neural distraction effect. These findings clearly demonstrate that top-down control of pain is affected in older age, and could explain the higher vulnerability for older adults to develop chronic pain. Moreover, our findings suggest that the assessment of executive functions may be a useful tool for predicting the efficacy of cognitive pain modulation strategies in older adults.Keywords: executive functions, cognitive pain modulation, fMRI, PFC
Procedia PDF Downloads 14417354 Data Collection Techniques for Robotics to Identify the Facial Expressions of Traumatic Brain Injured Patients
Authors: Chaudhary Muhammad Aqdus Ilyas, Matthias Rehm, Kamal Nasrollahi, Thomas B. Moeslund
Abstract:
This paper presents the investigation of data collection procedures, associated with robots when placed with traumatic brain injured (TBI) patients for rehabilitation purposes through facial expression and mood analysis. Rehabilitation after TBI is very crucial due to nature of injury and variation in recovery time. It is advantageous to analyze these emotional signals in a contactless manner, due to the non-supportive behavior of patients, limited muscle movements and increase in negative emotional expressions. This work aims at the development of framework where robots can recognize TBI emotions through facial expressions to perform rehabilitation tasks by physical, cognitive or interactive activities. The result of these studies shows that with customized data collection strategies, proposed framework identify facial and emotional expressions more accurately that can be utilized in enhancing recovery treatment and social interaction in robotic context.Keywords: computer vision, convolution neural network- long short term memory network (CNN-LSTM), facial expression and mood recognition, multimodal (RGB-thermal) analysis, rehabilitation, robots, traumatic brain injured patients
Procedia PDF Downloads 15517353 Implementation of Social Network Analysis to Analyze the Dependency between Construction Bid Packages
Authors: Kawalpreet Kaur, Panagiotis Mitropoulos
Abstract:
The division of the project scope into work packages is the most important step in the preconstruction phase of construction projects. The work division determines the scope and complexity of each bid package, resulting in dependencies between project participants performing these work packages. The coordination between project participants is necessary because of these dependencies. Excessive dependencies between the bid packages create coordination difficulties, leading to delays, added costs, and contractual friction among project participants. However, the literature on construction provides limited knowledge regarding work structuring approaches, issues, and challenges. Manufacturing industry literature provides a systematic approach to defining the project scope into work packages, and the implementation of social network analysis (SNA) in manufacturing is an effective approach to defining and analyzing the divided scope of work at the dependencies level. This paper presents a case study of implementing a similar approach using SNA in construction bid packages. The study uses SNA to analyze the scope of bid packages and determine the dependency between scope elements. The method successfully identifies the bid package with the maximum interaction with other trade contractors and the scope elements that are crucial for project performance. The analysis provided graphical and quantitative information on bid package dependencies. The study can be helpful in performing an analysis to determine the dependencies between bid packages and their scope elements and how these scope elements are critical for project performance. The study illustrates the potential use of SNA as a systematic approach to analyzing bid package dependencies in construction projects, which can guide the division of crucial scope elements to minimize negative impacts on project performance.Keywords: work structuring, bid packages, work breakdown, project participants
Procedia PDF Downloads 7917352 Customized Design of Amorphous Solids by Generative Deep Learning
Authors: Yinghui Shang, Ziqing Zhou, Rong Han, Hang Wang, Xiaodi Liu, Yong Yang
Abstract:
The design of advanced amorphous solids, such as metallic glasses, with targeted properties through artificial intelligence signifies a paradigmatic shift in physical metallurgy and materials technology. Here, we developed a machine-learning architecture that facilitates the generation of metallic glasses with targeted multifunctional properties. Our architecture integrates the state-of-the-art unsupervised generative adversarial network model with supervised models, allowing the incorporation of general prior knowledge derived from thousands of data points across a vast range of alloy compositions, into the creation of data points for a specific type of composition, which overcame the common issue of data scarcity typically encountered in the design of a given type of metallic glasses. Using our generative model, we have successfully designed copper-based metallic glasses, which display exceptionally high hardness or a remarkably low modulus. Notably, our architecture can not only explore uncharted regions in the targeted compositional space but also permits self-improvement after experimentally validated data points are added to the initial dataset for subsequent cycles of data generation, hence paving the way for the customized design of amorphous solids without human intervention.Keywords: metallic glass, artificial intelligence, mechanical property, automated generation
Procedia PDF Downloads 5617351 Towards End-To-End Disease Prediction from Raw Metagenomic Data
Authors: Maxence Queyrel, Edi Prifti, Alexandre Templier, Jean-Daniel Zucker
Abstract:
Analysis of the human microbiome using metagenomic sequencing data has demonstrated high ability in discriminating various human diseases. Raw metagenomic sequencing data require multiple complex and computationally heavy bioinformatics steps prior to data analysis. Such data contain millions of short sequences read from the fragmented DNA sequences and stored as fastq files. Conventional processing pipelines consist in multiple steps including quality control, filtering, alignment of sequences against genomic catalogs (genes, species, taxonomic levels, functional pathways, etc.). These pipelines are complex to use, time consuming and rely on a large number of parameters that often provide variability and impact the estimation of the microbiome elements. Training Deep Neural Networks directly from raw sequencing data is a promising approach to bypass some of the challenges associated with mainstream bioinformatics pipelines. Most of these methods use the concept of word and sentence embeddings that create a meaningful and numerical representation of DNA sequences, while extracting features and reducing the dimensionality of the data. In this paper we present an end-to-end approach that classifies patients into disease groups directly from raw metagenomic reads: metagenome2vec. This approach is composed of four steps (i) generating a vocabulary of k-mers and learning their numerical embeddings; (ii) learning DNA sequence (read) embeddings; (iii) identifying the genome from which the sequence is most likely to come and (iv) training a multiple instance learning classifier which predicts the phenotype based on the vector representation of the raw data. An attention mechanism is applied in the network so that the model can be interpreted, assigning a weight to the influence of the prediction for each genome. Using two public real-life data-sets as well a simulated one, we demonstrated that this original approach reaches high performance, comparable with the state-of-the-art methods applied directly on processed data though mainstream bioinformatics workflows. These results are encouraging for this proof of concept work. We believe that with further dedication, the DNN models have the potential to surpass mainstream bioinformatics workflows in disease classification tasks.Keywords: deep learning, disease prediction, end-to-end machine learning, metagenomics, multiple instance learning, precision medicine
Procedia PDF Downloads 12517350 Flushing Model for Artificial Islands in the Persian Gulf
Authors: Sawsan Eissa, Momen Gharib, Omnia Kabbany
Abstract:
A flushing numerical study has been performed for intended artificial islands on the Persian Gulf coast in Abu Dhabi, UAE. The island masterplan was tested for flushing using the DELFT 3D hydrodynamic model, and it was found that its residence time exceeds the acceptable PIANC flushing Criteria. Therefore, a number of mitigation measures were applied and tested one by one using the flushing model. Namely, changing the location of the entrance opening, dredging, removing part of the mangrove existing in the near vicinity to create a channel, removing the mangrove altogether, using culverts of different numbers and locations, and pumping at selected points. The pumping option gave the best solution, but it was disregarded due to high capital and running costs. Therefore, it opted for a combination of other solutions, including removing mangroves, introducing culverts, and adjusting island boundaries and types of protection.Keywords: hydrodynamics, flushing, delft 3d, Persian Gulf, artificial islands.
Procedia PDF Downloads 6117349 Wired Network Services in Mobile Phones
Authors: Subhash Reddy
Abstract:
Mobile communication in today’s world means a lot to the human kind, through this many deals are made and others are broken, within seconds. That is because of our sophisticated methods of transporting the data at very high speeds and to very long distances, within no time. That is also because we kept on changing the method of serving the connections as the no of connections kept on increasing, that has led to many methods like TDMA, CDMA, and FDMA, etc. in wireless communications. And also the areas, where the connections are provided are also divided into CELLS, which are the basic blocks for cellular communications. Along with the wireless network, providing a wired network in mobile phones would serve as a very good alternative and would divert the extra traffic of a cell, so that a CELL which is providing wireless network can operate more efficiently.Keywords: CDMA, FDMA, TDMA, CELL
Procedia PDF Downloads 48617348 Measurement and Analysis of Building Penetration Loss for Mobile Networks in Tripoli Area
Authors: Tammam A. Benmusa, Mohamed A. Shlibek, Rawad M. Swesi
Abstract:
The investigation of Buildings Penetration Loss (BPL) of radio signal is getting more and more important. It plays an important role in calculating the indoor coverage for wireless communication networks. In this paper, the theory behind BPL and its mechanisms have been reviewed. The operating frequency, coverage area type, climate condition, time of measurement, and other factors affecting the values of BPL have been discussed. The practical part of this work was conducting 4000 measurements of BPL in different areas in the Libyan capital, Tripoli, to get empirical model for this loss. The measurements were taken for 2 different types of wireless communication networks; mobile telephone network (for Almadar company), which operates at 900 MHz and WiMAX network (LTT company) which operates at 2500 MHz. The results for each network were summarized and presented in several graphs. The graphs are showing how the BPL affected by: time of measurement, morphology (type of area), and climatic environment.Keywords: building penetration loss, wireless network, mobile network, link budget, indoor network performance
Procedia PDF Downloads 38417347 Governance in the Age of Artificial intelligence and E- Government
Authors: Mernoosh Abouzari, Shahrokh Sahraei
Abstract:
Electronic government is a way for governments to use new technology that provides people with the necessary facilities for proper access to government information and services, improving the quality of services and providing broad opportunities to participate in democratic processes and institutions. That leads to providing the possibility of easy use of information technology in order to distribute government services to the customer without holidays, which increases people's satisfaction and participation in political and economic activities. The expansion of e-government services and its movement towards intelligentization has the ability to re-establish the relationship between the government and citizens and the elements and components of the government. Electronic government is the result of the use of information and communication technology (ICT), which by implementing it at the government level, in terms of the efficiency and effectiveness of government systems and the way of providing services, tremendous commercial changes are created, which brings people's satisfaction at the wide level will follow. The main level of electronic government services has become objectified today with the presence of artificial intelligence systems, which recent advances in artificial intelligence represent a revolution in the use of machines to support predictive decision-making and Classification of data. With the use of deep learning tools, artificial intelligence can mean a significant improvement in the delivery of services to citizens and uplift the work of public service professionals while also inspiring a new generation of technocrats to enter government. This smart revolution may put aside some functions of the government, change its components, and concepts such as governance, policymaking or democracy will change in front of artificial intelligence technology, and the top-down position in governance may face serious changes, and If governments delay in using artificial intelligence, the balance of power will change and private companies will monopolize everything with their pioneering in this field, and the world order will also depend on rich multinational companies and in fact, Algorithmic systems will become the ruling systems of the world. It can be said that currently, the revolution in information technology and biotechnology has been started by engineers, large economic companies, and scientists who are rarely aware of the political complexities of their decisions and certainly do not represent anyone. Therefore, it seems that if liberalism, nationalism, or any other religion wants to organize the world of 2050, it should not only rationalize the concept of artificial intelligence and complex data algorithm but also mix them in a new and meaningful narrative. Therefore, the changes caused by artificial intelligence in the political and economic order will lead to a major change in the way all countries deal with the phenomenon of digital globalization. In this paper, while debating the role and performance of e-government, we will discuss the efficiency and application of artificial intelligence in e-government, and we will consider the developments resulting from it in the new world and the concepts of governance.Keywords: electronic government, artificial intelligence, information and communication technology., system
Procedia PDF Downloads 9417346 Detecting Manipulated Media Using Deep Capsule Network
Authors: Joseph Uzuazomaro Oju
Abstract:
The ease at which manipulated media can be created, and the increasing difficulty in identifying fake media makes it a great threat. Most of the applications used for the creation of these high-quality fake videos and images are built with deep learning. Hence, the use of deep learning in creating a detection mechanism cannot be overemphasized. Any successful fake media that is being detected before it reached the populace will save people from the self-doubt of either a content is genuine or fake and will ensure the credibility of videos and images. The methodology introduced in this paper approaches the manipulated media detection challenge using a combo of VGG-19 and a deep capsule network. In the case of videos, they are converted into frames, which, in turn, are resized and cropped to the face region. These preprocessed images/videos are fed to the VGG-19 network to extract the latent features. The extracted latent features are inputted into a deep capsule network enhanced with a 3D -convolution dynamic routing agreement. The 3D –convolution dynamic routing agreement algorithm helps to reduce the linkages between capsules networks. Thereby limiting the poor learning shortcoming of multiple capsule network layers. The resultant output from the deep capsule network will indicate a media to be either genuine or fake.Keywords: deep capsule network, dynamic routing, fake media detection, manipulated media
Procedia PDF Downloads 13417345 Under the 'Umbrella' Project: A Volunteer-Mentoring Approach for Socially Disadvantaged University Students
Authors: Evridiki Zachopoulou, Vasilis Grammatikopoulos, Michail Vitoulis, Athanasios Gregoriadis
Abstract:
In the last ten years, the recent economic crisis in Greece has decreased the financial ability and strength of several families when it comes to supporting their children’s studies. As a result, the number of students who are significantly delaying or even dropping out of their university studies is constantly increasing. The students who are at greater risk for academic failure are those who are facing various problems and social disadvantages, like health problems, special needs, family poverty or unemployment, single-parent students, immigrant students, etc. The ‘Umbrella’ project is a volunteer-based initiative to tackle this problem at International Hellenic University. The main purpose of the project is to provide support to disadvantaged students at a socio-emotional, academic, and practical level in order to help them complete their undergraduate studies. More specifically, the ‘Umbrella’ project has the following goals: (a) to develop a consulting-supporting network based on volunteering senior students, called ‘i-mentors’. (b) to train the volunteering i-mentors and create a systematic and consistent support procedure for students at-risk, (c), to develop a service that, parallel to the i-mentor network will be ensuring opportunities for at-risk students to find a job, (d) to support students who are coping with accessibility difficulties, (e) to secure the sustainability of the ‘Umbrella’ project after the completion of the funding of the project. The innovation of the Umbrella project is in its holistic-person-centered approach that will be providing individualized support -via the i-mentors network- to any disadvantaged student that will come ‘under the Umbrella.’Keywords: peer mentoring, student support, socially disadvantaged students, volunteerism in higher education
Procedia PDF Downloads 23417344 Use Cloud-Based Watson Deep Learning Platform to Train Models Faster and More Accurate
Authors: Susan Diamond
Abstract:
Machine Learning workloads have traditionally been run in high-performance computing (HPC) environments, where users log in to dedicated machines and utilize the attached GPUs to run training jobs on huge datasets. Training of large neural network models is very resource intensive, and even after exploiting parallelism and accelerators such as GPUs, a single training job can still take days. Consequently, the cost of hardware is a barrier to entry. Even when upfront cost is not a concern, the lead time to set up such an HPC environment takes months from acquiring hardware to set up the hardware with the right set of firmware, software installed and configured. Furthermore, scalability is hard to achieve in a rigid traditional lab environment. Therefore, it is slow to react to the dynamic change in the artificial intelligent industry. Watson Deep Learning as a service, a cloud-based deep learning platform that mitigates the long lead time and high upfront investment in hardware. It enables robust and scalable sharing of resources among the teams in an organization. It is designed for on-demand cloud environments. Providing a similar user experience in a multi-tenant cloud environment comes with its own unique challenges regarding fault tolerance, performance, and security. Watson Deep Learning as a service tackles these challenges and present a deep learning stack for the cloud environments in a secure, scalable and fault-tolerant manner. It supports a wide range of deep-learning frameworks such as Tensorflow, PyTorch, Caffe, Torch, Theano, and MXNet etc. These frameworks reduce the effort and skillset required to design, train, and use deep learning models. Deep Learning as a service is used at IBM by AI researchers in areas including machine translation, computer vision, and healthcare.Keywords: deep learning, machine learning, cognitive computing, model training
Procedia PDF Downloads 20917343 Supporting Densification through the Planning and Implementation of Road Infrastructure in the South African Context
Authors: K. Govender, M. Sinclair
Abstract:
This paper demonstrates a proof of concept whereby shorter trips and land use densification can be promoted through an alternative approach to planning and implementation of road infrastructure in the South African context. It briefly discusses how the development of the Compact City concept relies on a combination of promoting shorter trips and densification through a change in focus in road infrastructure provision. The methodology developed in this paper uses a traffic model to test the impact of synthesized deterrence functions on congestion locations in the road network through the assignment of traffic on the study network. The results from this study demonstrate that intelligent planning of road infrastructure can indeed promote reduced urban sprawl, increased residential density and mixed-use areas which are supported by an efficient public transport system; and reduced dependence on the freeway network with a fixed road infrastructure budget. The study has resonance for all cities where urban sprawl is seemingly unstoppable.Keywords: compact cities, densification, road infrastructure planning, transportation modelling
Procedia PDF Downloads 17817342 Particle Filter State Estimation Algorithm Based on Improved Artificial Bee Colony Algorithm
Authors: Guangyuan Zhao, Nan Huang, Xuesong Han, Xu Huang
Abstract:
In order to solve the problem of sample dilution in the traditional particle filter algorithm and achieve accurate state estimation in a nonlinear system, a particle filter method based on an improved artificial bee colony (ABC) algorithm was proposed. The algorithm simulated the process of bee foraging and optimization and made the high likelihood region of the backward probability of particles moving to improve the rationality of particle distribution. The opposition-based learning (OBL) strategy is introduced to optimize the initial population of the artificial bee colony algorithm. The convergence factor is introduced into the neighborhood search strategy to limit the search range and improve the convergence speed. Finally, the crossover and mutation operations of the genetic algorithm are introduced into the search mechanism of the following bee, which makes the algorithm jump out of the local extreme value quickly and continue to search the global extreme value to improve its optimization ability. The simulation results show that the improved method can improve the estimation accuracy of particle filters, ensure the diversity of particles, and improve the rationality of particle distribution.Keywords: particle filter, impoverishment, state estimation, artificial bee colony algorithm
Procedia PDF Downloads 15217341 Increase in the Persistence of Various Invaded Multiplex Metacommunities Induced by Heterogeneity of Motifs
Authors: Dweepabiswa Bagchi, D. V. Senthilkumar
Abstract:
Numerous studies have typically demonstrated the devastation of invasions on an isolated ecosystem or, at most, a network of dispersively coupled similar ecosystem patches. Using such a simplistic 2-D network model, one can only consider dispersal coupling and inter-species trophic interactions. However, in a realistic ecosystem, numerous species co-exist and interact trophically and non-trophically in groups of 2 or more. Even different types of dispersal can introduce complexity in an ecological network. Therefore, a more accurate representation of actual ecosystems (or ecological networks) is a complex network consisting of motifs formed by two or more interacting species. Here, the apropos structure of the network should be multiplex or multi-layered. Motifs between different patches or species should be identical within the same layer and vary from one layer to another. This study investigates three distinct ecological multiplex networks facing invasion from one or more external species. This work determines and quantifies the criteria for the increased extinction risk of these networks. The dynamical states of the network with high extinction risk, i.e., the danger states, and those with low extinction risk, i.e., the resistive network states, are both subsequently identified. The analysis done in this study further quantifies the persistence of the entire network corresponding to simultaneous changes in the strength of invasive dispersal and higher-order trophic and non-trophic interactions. This study also demonstrates that the ecosystems enjoy an inherent advantage against invasions due to their multiplex network structure.Keywords: increased ecosystem persistence, invasion on ecosystems, multiplex networks, non-trophic interactions
Procedia PDF Downloads 6417340 Corporate Digital Responsibility in Construction Engineering-Construction 4.0: Ethical Guidelines for Digitization and Artificial Intelligence
Authors: Weber-Lewerenz Bianca
Abstract:
Digitization is developing fast and has become a powerful tool for digital planning, construction, and operations. Its transformation bears high potentials for companies, is critical for success, and thus, requires responsible handling. This study provides an assessment of calls made in the sustainable development goals by the United Nations (SDGs), White Papers on AI by international institutions, EU-Commission and German Government requesting for the consideration and protection of values and fundamental rights, the careful demarcation between machine (artificial) and human intelligence and the careful use of such technologies. The study discusses digitization and the impacts of artificial intelligence (AI) in construction engineering from an ethical perspective by generating data via conducting case studies and interviewing experts as part of the qualitative method. This research evaluates critically opportunities and risks revolving around corporate digital responsibility (CDR) in the construction industry. To the author's knowledge, no study has set out to investigate how CDR in construction could be conceptualized, especially in relation to the digitization and AI, to mitigate digital transformation both in large, medium-sized, and small companies. No study addressed the key research question: Where can CDR be allocated, how shall its adequate ethical framework be designed to support digital innovations in order to make full use of the potentials of digitization and AI? Now is the right timing for constructive approaches and apply ethics-by-design in order to develop and implement a safe and efficient AI. This represents the first study in construction engineering applying a holistic, interdisciplinary, inclusive approach to provide guidelines for orientation, examine benefits of AI and define ethical principles as the key driver for success, resources-cost-time efficiency, and sustainability using digital technologies and AI in construction engineering to enhance digital transformation. Innovative corporate organizations starting new business models are more likely to succeed than those dominated by conservative, traditional attitudes.Keywords: construction engineering, digitization, digital transformation, artificial intelligence, ethics, corporate digital responsibility, digital innovation
Procedia PDF Downloads 25017339 Performance Evaluation of Contemporary Classifiers for Automatic Detection of Epileptic EEG
Authors: K. E. Ch. Vidyasagar, M. Moghavvemi, T. S. S. T. Prabhat
Abstract:
Epilepsy is a global problem, and with seizures eluding even the smartest of diagnoses a requirement for automatic detection of the same using electroencephalogram (EEG) would have a huge impact in diagnosis of the disorder. Among a multitude of methods for automatic epilepsy detection, one should find the best method out, based on accuracy, for classification. This paper reasons out, and rationalizes, the best methods for classification. Accuracy is based on the classifier, and thus this paper discusses classifiers like quadratic discriminant analysis (QDA), classification and regression tree (CART), support vector machine (SVM), naive Bayes classifier (NBC), linear discriminant analysis (LDA), K-nearest neighbor (KNN) and artificial neural networks (ANN). Results show that ANN is the most accurate of all the above stated classifiers with 97.7% accuracy, 97.25% specificity and 98.28% sensitivity in its merit. This is followed closely by SVM with 1% variation in result. These results would certainly help researchers choose the best classifier for detection of epilepsy.Keywords: classification, seizure, KNN, SVM, LDA, ANN, epilepsy
Procedia PDF Downloads 52017338 Wireless Sensor Networks Optimization by Using 2-Stage Algorithm Based on Imperialist Competitive Algorithm
Authors: Hamid R. Lashgarian Azad, Seyed N. Shetab Boushehri
Abstract:
Wireless sensor networks (WSN) have become progressively popular due to their wide range of applications. Wireless Sensor Network is made of numerous tiny sensor nodes that are battery-powered. It is a very significant problem to maximize the lifetime of wireless sensor networks. In this paper, we propose a two-stage protocol based on an imperialist competitive algorithm (2S-ICA) to solve a sensor network optimization problem. The energy of the sensors can be greatly reduced and the lifetime of the network reduced by long communication distances between the sensors and the sink. We can minimize the overall communication distance considerably, thereby extending the lifetime of the network lifetime through connecting sensors into a series of independent clusters using 2SICA. Comparison results of the proposed protocol and LEACH protocol, which is common to solving WSN problems, show that our protocol has a better performance in terms of improving network life and increasing the number of transmitted data.Keywords: wireless sensor network, imperialist competitive algorithm, LEACH protocol, k-means clustering
Procedia PDF Downloads 10317337 Application of Wireless Sensor Networks: A Survey in Thailand
Authors: Sathapath Kilaso
Abstract:
Nowadays, Today, wireless sensor networks are an important technology that works with Internet of Things. It is receiving various data from many sensor. Then sent to processing or storing. By wireless network or through the Internet. The devices around us are intelligent, can receiving/transmitting and processing data and communicating through the system. There are many applications of wireless sensor networks, such as smart city, smart farm, environmental management, weather. This article will explore the use of wireless sensor networks in Thailand and collect data from Thai Thesis database in 2012-2017. How to Implementing Wireless Sensor Network Technology. Advantage from this study To know the usage wireless technology in many fields. This will be beneficial for future research. In this study was found the most widely used wireless sensor network in agriculture field. Especially for smart farms. And the second is the adoption of the environment. Such as weather stations and water inspection.Keywords: wireless sensor network, smart city, survey, Adhoc Network
Procedia PDF Downloads 20717336 A Sectional Control Method to Decrease the Accumulated Survey Error of Tunnel Installation Control Network
Authors: Yinggang Guo, Zongchun Li
Abstract:
In order to decrease the accumulated survey error of tunnel installation control network of particle accelerator, a sectional control method is proposed. Firstly, the accumulation rule of positional error with the length of the control network is obtained by simulation calculation according to the shape of the tunnel installation-control-network. Then, the RMS of horizontal positional precision of tunnel backbone control network is taken as the threshold. When the accumulated error is bigger than the threshold, the tunnel installation control network should be divided into subsections reasonably. On each segment, the middle survey station is taken as the datum for independent adjustment calculation. Finally, by taking the backbone control points as faint datums, the weighted partial parameters adjustment is performed with the adjustment results of each segment and the coordinates of backbone control points. The subsections are jointed and unified into the global coordinate system in the adjustment process. An installation control network of the linac with a length of 1.6 km is simulated. The RMS of positional deviation of the proposed method is 2.583 mm, and the RMS of the difference of positional deviation between adjacent points reaches 0.035 mm. Experimental results show that the proposed sectional control method can not only effectively decrease the accumulated survey error but also guarantee the relative positional precision of the installation control network. So it can be applied in the data processing of tunnel installation control networks, especially for large particle accelerators.Keywords: alignment, tunnel installation control network, accumulated survey error, sectional control method, datum
Procedia PDF Downloads 19117335 Towards Long-Range Pixels Connection for Context-Aware Semantic Segmentation
Authors: Muhammad Zubair Khan, Yugyung Lee
Abstract:
Deep learning has recently achieved enormous response in semantic image segmentation. The previously developed U-Net inspired architectures operate with continuous stride and pooling operations, leading to spatial data loss. Also, the methods lack establishing long-term pixels connection to preserve context knowledge and reduce spatial loss in prediction. This article developed encoder-decoder architecture with bi-directional LSTM embedded in long skip-connections and densely connected convolution blocks. The network non-linearly combines the feature maps across encoder-decoder paths for finding dependency and correlation between image pixels. Additionally, the densely connected convolutional blocks are kept in the final encoding layer to reuse features and prevent redundant data sharing. The method applied batch-normalization for reducing internal covariate shift in data distributions. The empirical evidence shows a promising response to our method compared with other semantic segmentation techniques.Keywords: deep learning, semantic segmentation, image analysis, pixels connection, convolution neural network
Procedia PDF Downloads 10317334 Comparative Study of Properties of Iranian Historical Gardens by Focusing on Climate
Authors: Malihe Ahmadi
Abstract:
Nowadays, stress, tension and neural problems are among the most important concerns of the present age. The environment plays key role on improving mental health and reducing stress of citizens. Establishing balance and appropriate relationship between city and natural environment is of the most important approaches of present century. Type of approach and logical planning for urban green spaces as one of the basic sections of integration with nature, not only plays key role on quality and efficiency of comprehensive urban planning; but also it increases the system of distributing social activities and happiness and lively property of urban environments that leads to permanent urban development. The main purpose of recovering urban identity is considering culture, history and human life style in past. This is a documentary-library research that evaluates the historical properties of Iranian gardens in compliance with climate condition. Results of this research reveal that in addition to following Iranian gardens from common principles of land lot, structure of flowers and plants, water, specific buildings during different ages, the role of climate at different urban areas is among the basics of determining method of designing green spaces and different buildings located at diverse areas i.e. Iranian gardens are a space for merging natural and artificial elements that has inseparable connection with semantic principles and guarantees different functions. Some of the necessities of designing present urban gardens are including: recognition and recreation.Keywords: historical gardens, climate, properties of Iranian gardens, Iran
Procedia PDF Downloads 39717333 Improved Artificial Bee Colony Algorithm for Non-Convex Economic Power Dispatch Problem
Authors: Badr M. Alshammari, T. Guesmi
Abstract:
This study presents a modified version of the artificial bee colony (ABC) algorithm by including a local search technique for solving the non-convex economic power dispatch problem. The local search step is incorporated at the end of each iteration. Total system losses, valve-point loading effects and prohibited operating zones have been incorporated in the problem formulation. Thus, the problem becomes highly nonlinear and with discontinuous objective function. The proposed technique is validated using an IEEE benchmark system with ten thermal units. Simulation results demonstrate that the proposed optimization algorithm has better convergence characteristics in comparison with the original ABC algorithm.Keywords: economic power dispatch, artificial bee colony, valve-point loading effects, prohibited operating zones
Procedia PDF Downloads 258