Search results for: discrete feature
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2187

Search results for: discrete feature

837 Four-Way Coupled CFD-Dem Simulation of Concrete Pipe Flow Using a Non-Newtonian Rheological Model: Investigating the Simulation of Lubrication Layer Formation and Plug Flow Zones

Authors: Tooran Tavangar, Masoud Hosseinpoor, Jeffrey S. Marshall, Ammar Yahia, Kamal Henri Khayat

Abstract:

In this study, a four-way coupled CFD-DEM methodology was used to simulate the behavior of concrete pipe flow. Fresh concrete, characterized as a biphasic suspension, features aggregates comprising the solid-suspended phase with diverse particle-size distributions (PSD) within a non-Newtonian cement paste/mortar matrix forming the liquid phase. The fluid phase was simulated using CFD, while the aggregates were modeled using DEM. Interaction forces between the fluid and solid particles were considered through CFD-DEM computations. To capture the viscoelastic characteristics of the suspending fluid, a bi-viscous approach was adopted, incorporating a critical shear rate proportional to the yield stress of the mortar. In total, three diphasic suspensions were simulated, each featuring distinct particle size distributions and a concentration of 10% for five subclasses of spherical particles ranging from 1 to 17 mm in a suspending fluid. The adopted bi-viscous approach successfully simulated both un-sheared (plug flow) and sheared zones. Furthermore, shear-induced particle migration (SIPM) was assessed by examining coefficients of variation in particle concentration across the pipe. These SIPM values were then compared with results obtained using CFD-DEM under the Newtonian assumption. The study highlighted the crucial role of yield stress in the mortar phase, revealing that lower yield stress values can lead to increased flow rates and higher SIPM across the pipe.

Keywords: computational fluid dynamics, concrete pumping, coupled CFD-DEM, discrete element method, plug flow, shear-induced particle migration.

Procedia PDF Downloads 58
836 Crack Initiation Assessment during Fracture of Heat Treated Duplex Stainless Steels

Authors: Faraj Ahmed E. Alhegagi, Anagia M. Khamkam Mohamed, Bassam F. Alhajaji

Abstract:

Duplex stainless steels (DSS) are widely employed in industry for apparatus working with sea water in petroleum, refineries and in chemical plants. Fracture of DSS takes place by cleavage of the ferrite phase and the austenite phase ductile tear off. Pop-in is an important feature takes place during fracture of DSS. The procedure of Pop-ins assessment plays an important role in fracture toughness studies. In present work, Zeron100 DSS specimens were heat treated at different temperatures, cooled and pulled to failure to assess the pop-ins criterion in crack initiation prediction. The outcome results were compared to the British Standard (BS 7448) and the ASTEM standard (E1290) for Crack-Tip Opening Displacement (CTOD) fracture toughness measurement. Pop-in took place during specimens loading specially for those specimens heat treated at higher temperatures. The standard BS7448 was followed to check specimen validity for fractured toughness assessment by direct determination of KIC. In most cases, specimens were invalid for KIC measurement. The two procedures were equivalent only when single pop-ins were assessed. A considerable contrast in fracture toughness value between was observed where multiple pop-ins were assessed.

Keywords: fracture toughness, stainless steels, pop ins, crack assessment

Procedia PDF Downloads 117
835 Deep Learning Based, End-to-End Metaphor Detection in Greek with Recurrent and Convolutional Neural Networks

Authors: Konstantinos Perifanos, Eirini Florou, Dionysis Goutsos

Abstract:

This paper presents and benchmarks a number of end-to-end Deep Learning based models for metaphor detection in Greek. We combine Convolutional Neural Networks and Recurrent Neural Networks with representation learning to bear on the metaphor detection problem for the Greek language. The models presented achieve exceptional accuracy scores, significantly improving the previous state-of-the-art results, which had already achieved accuracy 0.82. Furthermore, no special preprocessing, feature engineering or linguistic knowledge is used in this work. The methods presented achieve accuracy of 0.92 and F-score 0.92 with Convolutional Neural Networks (CNNs) and bidirectional Long Short Term Memory networks (LSTMs). Comparable results of 0.91 accuracy and 0.91 F-score are also achieved with bidirectional Gated Recurrent Units (GRUs) and Convolutional Recurrent Neural Nets (CRNNs). The models are trained and evaluated only on the basis of training tuples, the related sentences and their labels. The outcome is a state-of-the-art collection of metaphor detection models, trained on limited labelled resources, which can be extended to other languages and similar tasks.

Keywords: metaphor detection, deep learning, representation learning, embeddings

Procedia PDF Downloads 145
834 Effect of Financing Sources on Firm Performance: A Study of Indian Private Limited Small and Medium Enterprises

Authors: Denila Jinny Arulraj, Thillai Rajan Annamalai

Abstract:

This paper aims to study the relationship between funding sources and firm performance of Indian private limited SMEs using cross-sectional data obtained from a nation-wide census. A unique feature of the study is that it analyses firms that use only one form of external funding. Employing Propensity Score Matching, we find that obtaining any form of external finance has a negative influence on equivalents of profit margin and return on assets and a negative influence on asset turnover of small firms. But, the impact of institutional sources of funding on small enterprises is found to be lesser than that of non-institutional sources of funding. External/institutional sources of funding have a less negative impact on the profit margin for medium enterprises and have no significant influence on other measures of performance. The contribution of this research is the discovery of institutional sources wielding a lesser influence on performance measures considered. It is also found that institutional sources can benefit small enterprises more than medium enterprises.

Keywords: external finance, institutional finance, non-institutional finance, performance, India, SME

Procedia PDF Downloads 266
833 Exploring the 1,3-Dipolar Cycloaddition Reaction between Nitrilimine and 6-Methyl-4,5-dihydropyridazin-3(2h)-one through MEDT and Molecular Docking Analysis

Authors: Zineb Ouahdi

Abstract:

Spirocyclic compound derivatives, with their unique heterocyclic motifs, serve as a continual source of inspiration in the pursuit of developing potential therapeutic agents. These compounds are diverse in their chemical structures; some have fully saturated skeletons, while others are partially unsaturated. Nevertheless, these compounds share a characteristic feature with natural products - the presence of at least one heteroatom in one of their rings. The inclusion of a C = O dipolarophile in pyridazinones imparts an exciting aspect for 1,3-dipolar cycloaddition reactions, the focal point of our study. Our research has involved a detailed theoretical investigation of the reaction between ethyl (Z)-2-bromo-2-(2-(p-tolyl)hydrazono)acetate and 6-methyl-4,5-dihydropyridazine-3(2H)-one. This has been accomplished using the DFT/B3LYP/6-31g(d,p) method, intending to illuminate the chemical pathway of this reaction. The chemical reactivity theories we used for this purpose included FMO, TS, and local and global indices derived from conceptual DFT. The theoretical framework outlined in this study allowed us to propose a reaction mechanism for cycloaddition reactions. It also enabled the identification of the potential activities of the analyzed compounds (P1, P2, P3, P4, P5, and P6) against the major protease of the coronavirus disease (COVID-19). This was achieved using various computational tools, including AutoDock Tools, Autodock Vina, Autodock 4, and PYRX.

Keywords: MEDT, pyridazin, cycloaddition, FMO, DFT, docking

Procedia PDF Downloads 92
832 Noninvasive Brain-Machine Interface to Control Both Mecha TE Robotic Hands Using Emotiv EEG Neuroheadset

Authors: Adrienne Kline, Jaydip Desai

Abstract:

Electroencephalogram (EEG) is a noninvasive technique that registers signals originating from the firing of neurons in the brain. The Emotiv EEG Neuroheadset is a consumer product comprised of 14 EEG channels and was used to record the reactions of the neurons within the brain to two forms of stimuli in 10 participants. These stimuli consisted of auditory and visual formats that provided directions of ‘right’ or ‘left.’ Participants were instructed to raise their right or left arm in accordance with the instruction given. A scenario in OpenViBE was generated to both stimulate the participants while recording their data. In OpenViBE, the Graz Motor BCI Stimulator algorithm was configured to govern the duration and number of visual stimuli. Utilizing EEGLAB under the cross platform MATLAB®, the electrodes most stimulated during the study were defined. Data outputs from EEGLAB were analyzed using IBM SPSS Statistics® Version 20. This aided in determining the electrodes to use in the development of a brain-machine interface (BMI) using real-time EEG signals from the Emotiv EEG Neuroheadset. Signal processing and feature extraction were accomplished via the Simulink® signal processing toolbox. An Arduino™ Duemilanove microcontroller was used to link the Emotiv EEG Neuroheadset and the right and left Mecha TE™ Hands.

Keywords: brain-machine interface, EEGLAB, emotiv EEG neuroheadset, OpenViBE, simulink

Procedia PDF Downloads 492
831 Automatic Tagging and Accuracy in Assamese Text Data

Authors: Chayanika Hazarika Bordoloi

Abstract:

This paper is an attempt to work on a highly inflectional language called Assamese. This is also one of the national languages of India and very little has been achieved in terms of computational research. Building a language processing tool for a natural language is not very smooth as the standard and language representation change at various levels. This paper presents inflectional suffixes of Assamese verbs and how the statistical tools, along with linguistic features, can improve the tagging accuracy. Conditional random fields (CRF tool) was used to automatically tag and train the text data; however, accuracy was improved after linguistic featured were fed into the training data. Assamese is a highly inflectional language; hence, it is challenging to standardizing its morphology. Inflectional suffixes are used as a feature of the text data. In order to analyze the inflections of Assamese word forms, a list of suffixes is prepared. This list comprises suffixes, comprising of all possible suffixes that various categories can take is prepared. Assamese words can be classified into inflected classes (noun, pronoun, adjective and verb) and un-inflected classes (adverb and particle). The corpus used for this morphological analysis has huge tokens. The corpus is a mixed corpus and it has given satisfactory accuracy. The accuracy rate of the tagger has gradually improved with the modified training data.

Keywords: CRF, morphology, tagging, tagset

Procedia PDF Downloads 186
830 From Waste to Wealth: A Future Paradigm for Plastic Management Using Blockchain Technology

Authors: Jim Shi, Jasmine Chang, Nesreen El-Rayes

Abstract:

The world has been experiencing a steadily increasing trend in both the production and consumption of plastic. The global consumer revolution should not have been possible without plastic, thanks to its salient feature of inexpensiveness and durability. But, as a two-edged sword, its durable quality has returned to haunt and even jeopardized us. That exacerbating the plastic crisis has attracted various global initiatives and actions. Simultaneously, firms are eager to adopt new technology as they witness and perceive more potential and merit of Industry 4.0 technologies. For example, Blockchain technology (BCT) is drawing the attention of numerous stakeholders because of its wide range of outstanding features that promise to enhance supply chain operations. However, from a research perspective, most of the literature addresses the plastic crisis from either environmental or social perspectives. In contrast, analysis from the data science perspective and technology is relatively scarce. To this end, this study aims to fill this gap and cover the plastic crisis from a holistic view of environmental, social, technological, and business perspectives. In particular, we propose a mathematical model to examine the inclusion of BCT to enhance and improve the efficiency on the upstream and the downstream sides of the plastic value, where the whole value chain is coordinated systematically, and its interoperability can be optimized. Consequently, the Environmental, Social, and Governance (ESG) goal and Circular Economics (CE) sustainability can be maximized.

Keywords: blockchain technology, plastic, circular economy, sustainability

Procedia PDF Downloads 75
829 Improving Axial-Attention Network via Cross-Channel Weight Sharing

Authors: Nazmul Shahadat, Anthony S. Maida

Abstract:

In recent years, hypercomplex inspired neural networks improved deep CNN architectures due to their ability to share weights across input channels and thus improve cohesiveness of representations within the layers. The work described herein studies the effect of replacing existing layers in an Axial Attention ResNet with their quaternion variants that use cross-channel weight sharing to assess the effect on image classification. We expect the quaternion enhancements to produce improved feature maps with more interlinked representations. We experiment with the stem of the network, the bottleneck layer, and the fully connected backend by replacing them with quaternion versions. These modifications lead to novel architectures which yield improved accuracy performance on the ImageNet300k classification dataset. Our baseline networks for comparison were the original real-valued ResNet, the original quaternion-valued ResNet, and the Axial Attention ResNet. Since improvement was observed regardless of which part of the network was modified, there is a promise that this technique may be generally useful in improving classification accuracy for a large class of networks.

Keywords: axial attention, representational networks, weight sharing, cross-channel correlations, quaternion-enhanced axial attention, deep networks

Procedia PDF Downloads 76
828 Analysis of Vocal Fold Vibrations from High-Speed Digital Images Based on Dynamic Time Warping

Authors: A. I. A. Rahman, Sh-Hussain Salleh, K. Ahmad, K. Anuar

Abstract:

Analysis of vocal fold vibration is essential for understanding the mechanism of voice production and for improving clinical assessment of voice disorders. This paper presents a Dynamic Time Warping (DTW) based approach to analyze and objectively classify vocal fold vibration patterns. The proposed technique was designed and implemented on a Glottal Area Waveform (GAW) extracted from high-speed laryngeal images by delineating the glottal edges for each image frame. Feature extraction from the GAW was performed using Linear Predictive Coding (LPC). Several types of voice reference templates from simulations of clear, breathy, fry, pressed and hyperfunctional voice productions were used. The patterns of the reference templates were first verified using the analytical signal generated through Hilbert transformation of the GAW. Samples from normal speakers’ voice recordings were then used to evaluate and test the effectiveness of this approach. The classification of the voice patterns using the technique of LPC and DTW gave the accuracy of 81%.

Keywords: dynamic time warping, glottal area waveform, linear predictive coding, high-speed laryngeal images, Hilbert transform

Procedia PDF Downloads 235
827 The Communication Library DIALOG for iFDAQ of the COMPASS Experiment

Authors: Y. Bai, M. Bodlak, V. Frolov, S. Huber, V. Jary, I. Konorov, D. Levit, J. Novy, D. Steffen, O. Subrt, M. Virius

Abstract:

Modern experiments in high energy physics impose great demands on the reliability, the efficiency, and the data rate of Data Acquisition Systems (DAQ). This contribution focuses on the development and deployment of the new communication library DIALOG for the intelligent, FPGA-based Data Acquisition System (iFDAQ) of the COMPASS experiment at CERN. The iFDAQ utilizing a hardware event builder is designed to be able to readout data at the maximum rate of the experiment. The DIALOG library is a communication system both for distributed and mixed environments, it provides a network transparent inter-process communication layer. Using the high-performance and modern C++ framework Qt and its Qt Network API, the DIALOG library presents an alternative to the previously used DIM library. The DIALOG library was fully incorporated to all processes in the iFDAQ during the run 2016. From the software point of view, it might be considered as a significant improvement of iFDAQ in comparison with the previous run. To extend the possibilities of debugging, the online monitoring of communication among processes via DIALOG GUI is a desirable feature. In the paper, we present the DIALOG library from several insights and discuss it in a detailed way. Moreover, the efficiency measurement and comparison with the DIM library with respect to the iFDAQ requirements is provided.

Keywords: data acquisition system, DIALOG library, DIM library, FPGA, Qt framework, TCP/IP

Procedia PDF Downloads 315
826 Interplay of Power Management at Core and Server Level

Authors: Jörg Lenhardt, Wolfram Schiffmann, Jörg Keller

Abstract:

While the feature sizes of recent Complementary Metal Oxid Semiconductor (CMOS) devices decrease the influence of static power prevails their energy consumption. Thus, power savings that benefit from Dynamic Frequency and Voltage Scaling (DVFS) are diminishing and temporal shutdown of cores or other microchip components become more worthwhile. A consequence of powering off unused parts of a chip is that the relative difference between idle and fully loaded power consumption is increased. That means, future chips and whole server systems gain more power saving potential through power-aware load balancing, whereas in former times this power saving approach had only limited effect, and thus, was not widely adopted. While powering off complete servers was used to save energy, it will be superfluous in many cases when cores can be powered down. An important advantage that comes with that is a largely reduced time to respond to increased computational demand. We include the above developments in a server power model and quantify the advantage. Our conclusion is that strategies from datacenters when to power off server systems might be used in the future on core level, while load balancing mechanisms previously used at core level might be used in the future at server level.

Keywords: power efficiency, static power consumption, dynamic power consumption, CMOS

Procedia PDF Downloads 213
825 Investigation of Possible Precancerous Viral Markers in Dental Follicles of Asymptomatic Impacted Teeth

Authors: Serap Keskin Tunç, Cennet Neslihan Eroğlu, Sevinç Şahin, Selda Seçkin

Abstract:

It has been suggested that various viruses may play a role in the pathogenesis of cancerous oral lesions in the literature. The aim of this study was to investigate the presence of both possible precancerous viral markers (HPV, HHV8, HSV1, HSV2, and EBV), and p53 and Ki-67 in the dental follicles of asymptomatic impacted teeth. A hundred healthy volunteers, older than 18 years old, included in the study. Dental follicles of extracted impacted teeth were excised and fixated in 10% formaldehyde. Histopathological and immunohistochemical examinations using HPV (containing HPV 8 and HPV 11), p16 (containing HPV 16), HHV8, HSV1, HSV2, EBV, p53 and Ki-67 antibodies were carried out. Also, the immunohistochemical results were correlated with the clinicopathological feature by Chi-square test statistically No dysplasia or neoplasm was observed. 62% of the cases were positive for p16, 32% were positive for EBV, 26% were positive for HSV1, immunohistochemically. All cases were immunonegative for HPV, HSV2, and HHV8. There was statistically significant correlation between overexpression of p53 with both EBV and p16 positivity (p<0.05). Direct correlation between higher expression of Ki-67 between EBV immunopositivity was detected (p<0.05). Thus, these viruses may be suggested to show trophism to the dental follicles acting as a reservoir. In conclusion, all dental follicles of extracted impacted teeth should be examined histopathologically in order to detect and prevent possible viral oncogenesis.

Keywords: dental follicles, Ki67, p53, precancerous markers viral markers

Procedia PDF Downloads 283
824 Towards Long-Range Pixels Connection for Context-Aware Semantic Segmentation

Authors: Muhammad Zubair Khan, Yugyung Lee

Abstract:

Deep learning has recently achieved enormous response in semantic image segmentation. The previously developed U-Net inspired architectures operate with continuous stride and pooling operations, leading to spatial data loss. Also, the methods lack establishing long-term pixels connection to preserve context knowledge and reduce spatial loss in prediction. This article developed encoder-decoder architecture with bi-directional LSTM embedded in long skip-connections and densely connected convolution blocks. The network non-linearly combines the feature maps across encoder-decoder paths for finding dependency and correlation between image pixels. Additionally, the densely connected convolutional blocks are kept in the final encoding layer to reuse features and prevent redundant data sharing. The method applied batch-normalization for reducing internal covariate shift in data distributions. The empirical evidence shows a promising response to our method compared with other semantic segmentation techniques.

Keywords: deep learning, semantic segmentation, image analysis, pixels connection, convolution neural network

Procedia PDF Downloads 96
823 Skin Manifestations in Children With Inborn Errors of Immunity in a Tertiary Care Hospital in Iran

Authors: Zahra Salehi Shahrbabaki, Zahra Chavoshzadeh, Fahimeh Abdollahimajd, Samin Sharafian, Tolue Mahdavi, Mahnaz Jamee

Abstract:

Background: Inborn errors of immunity (IEIs) are monogenic diseases of the immune the system with broad clinical manifestations. Despite the increasing genetic advancements, the diagnosis of IEIs still leans on clinical diagnosis. Dermatologic manifestations are observed in a large number of IEI patients and can lead to proper approach, prompt intervention and improved prognosis. Methods: This cross-sectional study was carried out between 2018 and 2020 on IEIs at a Children's tertiary care center in Tehran, Iran. Demographic details (including age, sex, and parental consanguinity), age at onset of symptoms and family history of IEI with were recorded. Results :212 patients were included. Cutaneous findings were reported in (95 ,44.8%) patients. and 61 of 95 (64.2%) reported skin lesions as the first clinical presentation. Skin infection (69, 72.6%) was the most frequent cutaneous manifestation, followed by an eczematous rash (24, 25 %). Conclusions: Skin manifestations are common feature in IEI patients and can be readily recognizable by healthcare providers. This study tried to provide information on prognostic consequences.

Keywords: primary immuno deficiency, inborn errror of metabolism, skin manifestation, skin infection

Procedia PDF Downloads 89
822 Normalized P-Laplacian: From Stochastic Game to Image Processing

Authors: Abderrahim Elmoataz

Abstract:

More and more contemporary applications involve data in the form of functions defined on irregular and topologically complicated domains (images, meshs, points clouds, networks, etc). Such data are not organized as familiar digital signals and images sampled on regular lattices. However, they can be conveniently represented as graphs where each vertex represents measured data and each edge represents a relationship (connectivity or certain affinities or interaction) between two vertices. Processing and analyzing these types of data is a major challenge for both image and machine learning communities. Hence, it is very important to transfer to graphs and networks many of the mathematical tools which were initially developed on usual Euclidean spaces and proven to be efficient for many inverse problems and applications dealing with usual image and signal domains. Historically, the main tools for the study of graphs or networks come from combinatorial and graph theory. In recent years there has been an increasing interest in the investigation of one of the major mathematical tools for signal and image analysis, which are Partial Differential Equations (PDEs) variational methods on graphs. The normalized p-laplacian operator has been recently introduced to model a stochastic game called tug-of-war-game with noise. Part interest of this class of operators arises from the fact that it includes, as particular case, the infinity Laplacian, the mean curvature operator and the traditionnal Laplacian operators which was extensiveley used to models and to solve problems in image processing. The purpose of this paper is to introduce and to study a new class of normalized p-Laplacian on graphs. The introduction is based on the extension of p-harmonious function introduced in as discrete approximation for both infinity Laplacian and p-Laplacian equations. Finally, we propose to use these operators as a framework for solving many inverse problems in image processing.

Keywords: normalized p-laplacian, image processing, stochastic game, inverse problems

Procedia PDF Downloads 506
821 Federalism, a System of Government: Comparative Study of Australia and Canada

Authors: Rana Tajammal Rashid

Abstract:

Federalism is a political system in which government power and responsibility are divided between a federal legislature and units of the state or provincial legislatures. This system provides the structure for the states having large territory and through that can manage the state affairs and administration easily. Many of the largest countries in the world are federations, like; The United States, Canada, India, Pakistan South Africa, Argentina, and Australia. Every large democratic nation has a federal system of government. This study will explore the feature and good governance of two developed countries Canada and Australia. This study will be helpful to the developing countries like Pakistan, India which have a federal form of structure to run the affairs of the state. In the federal system of Pakistan there are lot of issues and conflicts with the provinces with a comparative study of these two developed countries, i.e., Australia and Canada, our policy and decision maker political actors will understand in which way a state will successfully manage the issues related to federalism. This study will also provide the help to the students of comparative politics that how to analysis the different political system of the developed countries of the world.

Keywords: federalism, features of federalism, types of federalism, history of federalism, Australian federalism, Canadian federalism, federalism developments, executives, federal and provincial autonomy legislative, judicial

Procedia PDF Downloads 278
820 Deep Supervision Based-Unet to Detect Buildings Changes from VHR Aerial Imagery

Authors: Shimaa Holail, Tamer Saleh, Xiongwu Xiao

Abstract:

Building change detection (BCD) from satellite imagery is an essential topic in urbanization monitoring, agricultural land management, and updating geospatial databases. Recently, methods for detecting changes based on deep learning have made significant progress and impressive results. However, it has the problem of being insensitive to changes in buildings with complex spectral differences, and the features being extracted are not discriminatory enough, resulting in incomplete buildings and irregular boundaries. To overcome these problems, we propose a dual Siamese network based on the Unet model with the addition of a deep supervision strategy (DS) in this paper. This network consists of a backbone (encoder) based on ImageNet pre-training, a fusion block, and feature pyramid networks (FPN) to enhance the step-by-step information of the changing regions and obtain a more accurate BCD map. To train the proposed method, we created a new dataset (EGY-BCD) of high-resolution and multi-temporal aerial images captured over New Cairo in Egypt to detect building changes for this purpose. The experimental results showed that the proposed method is effective and performs well with the EGY-BCD dataset regarding the overall accuracy, F1-score, and mIoU, which were 91.6 %, 80.1 %, and 73.5 %, respectively.

Keywords: building change detection, deep supervision, semantic segmentation, EGY-BCD dataset

Procedia PDF Downloads 108
819 Baseline Study on Human Trafficking Crimes: A Case Study of Mapping Human Trafficking Crimes in East Java Province, Indonesia

Authors: Ni Komang Desy Arya Pinatih

Abstract:

Transnational crime is a crime with 'unique' feature because the activities benefit the lack of state monitoring on the borders so dealing with it cannot be based on conventional engagement but also need joint operation with other countries. On the other hand with the flow of globalization and the growth of information technology and transportation, states become more vulnerable to transnational crime threats especially human trafficking. This paper would examine transnational crime activities, especially human trafficking in Indonesia. With the case study on the mapping of human trafficking crime in East Java province, Indonesia, this paper would try to analyze how the difference in human trafficking crime trends at the national and sub-national levels. The findings of this research were first, there is difference in human trafficking crime trends whereas at the national level the trend is rising, while at sub-national (province) level the trend is declining. Second, regarding the decline of human trafficking number, it’s interesting to see how the method to decrease human trafficking crime in East Jawa Province in order to reduce transnational crime accounts in the region. These things are hopefully becoming a model for transnational crimes engagement in other regions to reduce human trafficking numbers as much as possible.

Keywords: transnational crime, human trafficking, southeast Asia, anticipation model on transnational crimes

Procedia PDF Downloads 296
818 An Integrative Computational Pipeline for Detection of Tumor Epitopes in Cancer Patients

Authors: Tanushree Jaitly, Shailendra Gupta, Leila Taher, Gerold Schuler, Julio Vera

Abstract:

Genomics-based personalized medicine is a promising approach to fight aggressive tumors based on patient's specific tumor mutation and expression profiles. A remarkable case is, dendritic cell-based immunotherapy, in which tumor epitopes targeting patient's specific mutations are used to design a vaccine that helps in stimulating cytotoxic T cell mediated anticancer immunity. Here we present a computational pipeline for epitope-based personalized cancer vaccines using patient-specific haplotype and cancer mutation profiles. In the workflow proposed, we analyze Whole Exome Sequencing and RNA Sequencing patient data to detect patient-specific mutations and their expression level. Epitopes including the tumor mutations are computationally predicted using patient's haplotype and filtered based on their expression level, binding affinity, and immunogenicity. We calculate binding energy for each filtered major histocompatibility complex (MHC)-peptide complex using docking studies, and use this feature to select good epitope candidates further.

Keywords: cancer immunotherapy, epitope prediction, NGS data, personalized medicine

Procedia PDF Downloads 244
817 Reliability Analysis of Variable Stiffness Composite Laminate Structures

Authors: A. Sohouli, A. Suleman

Abstract:

This study focuses on reliability analysis of variable stiffness composite laminate structures to investigate the potential structural improvement compared to conventional (straight fibers) composite laminate structures. A computational framework was developed which it consists of a deterministic design step and reliability analysis. The optimization part is Discrete Material Optimization (DMO) and the reliability of the structure is computed by Monte Carlo Simulation (MCS) after using Stochastic Response Surface Method (SRSM). The design driver in deterministic optimization is the maximum stiffness, while optimization method concerns certain manufacturing constraints to attain industrial relevance. These manufacturing constraints are the change of orientation between adjacent patches cannot be too large and the maximum number of successive plies of a particular fiber orientation should not be too high. Variable stiffness composites may be manufactured by Automated Fiber Machines (AFP) which provides consistent quality with good production rates. However, laps and gaps are the most important challenges to steer fibers that effect on the performance of the structures. In this study, the optimal curved fiber paths at each layer of composites are designed in the first step by DMO, and then the reliability analysis is applied to investigate the sensitivity of the structure with different standard deviations compared to the straight fiber angle composites. The random variables are material properties and loads on the structures. The results show that the variable stiffness composite laminate structures are much more reliable, even for high standard deviation of material properties, than the conventional composite laminate structures. The reason is that the variable stiffness composite laminates allow tailoring stiffness and provide the possibility of adjusting stress and strain distribution favorably in the structures.

Keywords: material optimization, Monte Carlo simulation, reliability analysis, response surface method, variable stiffness composite structures

Procedia PDF Downloads 514
816 Analysis Of Fine Motor Skills in Chronic Neurodegenerative Models of Huntington’s Disease and Amyotrophic Lateral Sclerosis

Authors: T. Heikkinen, J. Oksman, T. Bragge, A. Nurmi, O. Kontkanen, T. Ahtoniemi

Abstract:

Motor impairment is an inherent phenotypic feature of several chronic neurodegenerative diseases, and pharmacological therapies aimed to counterbalance the motor disability have a great market potential. Animal models of chronic neurodegenerative diseases display a number deteriorating motor phenotype during the disease progression. There is a wide array of behavioral tools to evaluate motor functions in rodents. However, currently existing methods to study motor functions in rodents are often limited to evaluate gross motor functions only at advanced stages of the disease phenotype. The most commonly applied traditional motor assays used in CNS rodent models, lack the sensitivity to capture fine motor impairments or improvements. Fine motor skill characterization in rodents provides a more sensitive tool to capture more subtle motor dysfunctions and therapeutic effects. Importantly, similar approach, kinematic movement analysis, is also used in clinic, and applied both in diagnosis and determination of therapeutic response to pharmacological interventions. The aim of this study was to apply kinematic gait analysis, a novel and automated high precision movement analysis system, to characterize phenotypic deficits in three different chronic neurodegenerative animal models, a transgenic mouse model (SOD1 G93A) for amyotrophic lateral sclerosis (ALS), and R6/2 and Q175KI mouse models for Huntington’s disease (HD). The readouts from walking behavior included gait properties with kinematic data, and body movement trajectories including analysis of various points of interest such as movement and position of landmarks in the torso, tail and joints. Mice (transgenic and wild-type) from each model were analyzed for the fine motor kinematic properties at young ages, prior to the age when gross motor deficits are clearly pronounced. Fine motor kinematic Evaluation was continued in the same animals until clear motor dysfunction with conventional motor assays was evident. Time course analysis revealed clear fine motor skill impairments in each transgenic model earlier than what is seen with conventional gross motor tests. Motor changes were quantitatively analyzed for up to ~80 parameters, and the largest data sets of HD models were further processed with principal component analysis (PCA) to transform the pool of individual parameters into a smaller and focused set of mutually uncorrelated gait parameters showing strong genotype difference. Kinematic fine motor analysis of transgenic animal models described in this presentation show that this method isa sensitive, objective and fully automated tool that allows earlier and more sensitive detection of progressive neuromuscular and CNS disease phenotypes. As a result of the analysis a comprehensive set of fine motor parameters for each model is created, and these parameters provide better understanding of the disease progression and enhanced sensitivity of this assay for therapeutic testing compared to classical motor behavior tests. In SOD1 G93A, R6/2, and Q175KI mice, the alterations in gait were evident already several weeks earlier than with traditional gross motor assays. Kinematic testing can be applied to a wider set of motor readouts beyond gait in order to study whole body movement patterns such as with relation to joints and various body parts longitudinally, providing a sophisticated and translatable method for disseminating motor components in rodent disease models and evaluating therapeutic interventions.

Keywords: Gait analysis, kinematic, motor impairment, inherent feature

Procedia PDF Downloads 352
815 Violence Detection and Tracking on Moving Surveillance Video Using Machine Learning Approach

Authors: Abe Degale D., Cheng Jian

Abstract:

When creating automated video surveillance systems, violent action recognition is crucial. In recent years, hand-crafted feature detectors have been the primary method for achieving violence detection, such as the recognition of fighting activity. Researchers have also looked into learning-based representational models. On benchmark datasets created especially for the detection of violent sequences in sports and movies, these methods produced good accuracy results. The Hockey dataset's videos with surveillance camera motion present challenges for these algorithms for learning discriminating features. Image recognition and human activity detection challenges have shown success with deep representation-based methods. For the purpose of detecting violent images and identifying aggressive human behaviours, this research suggested a deep representation-based model using the transfer learning idea. The results show that the suggested approach outperforms state-of-the-art accuracy levels by learning the most discriminating features, attaining 99.34% and 99.98% accuracy levels on the Hockey and Movies datasets, respectively.

Keywords: violence detection, faster RCNN, transfer learning and, surveillance video

Procedia PDF Downloads 93
814 Effect of Hybrid Fibers on Mechanical Properties in Autoclaved Aerated Concrete

Authors: B. Vijay Antony Raj, Umarani Gunasekaran, R. Thiru Kumara Raja Vallaban

Abstract:

Fibrous autoclaved aerated concrete (FAAC) is concrete containing fibrous material in it which helps to increase its structural integrity when compared to that of convention autoclaved aerated concrete (CAAC). These short discrete fibers are uniformly distributed and randomly oriented, which enhances the bond strength within the aerated concrete matrix. Conventional red-clay bricks create larger impact to the environment due to red soil depletion and it also consumes large amount to time for construction. Whereas, AAC are larger in size, lighter in weight and it is environmentally friendly in nature and hence it is a viable replacement for red-clay bricks. Internal micro cracks and corner cracks are the only disadvantages of conventional autoclaved aerated concrete, to resolve this particular issue it is preferable to make use of fibers in it.These fibers are bonded together within the matrix and they induce the aerated concrete to withstand considerable stresses, especially during the post cracking stage. Hence, FAAC has the capability of enhancing the mechanical properties and energy absorption capacity of CAAC. In this research work, individual fibers like glass, nylon, polyester and polypropylene are used they generally reduce the brittle fracture of AAC.To study the fibre’s surface topography and composition, SEM analysis is performed and then to determine the composition of a specimen as a whole as well as the composition of individual components EDAX mapping is carried out and then an experimental approach was performed to determine the effect of hybrid (multiple) fibres at various dosage (0.5%, 1%, 1.5%) and curing temperature of 180-2000 C is maintained to determine the mechanical properties of autoclaved aerated concrete. As an analytical part, the outcome experimental results is compared with fuzzy logic using MATLAB.

Keywords: fiberous AAC, crack control, energy absorption, mechanical properies, SEM, EDAX, MATLAB

Procedia PDF Downloads 265
813 Radon and Thoron Determination in Natural Ancient Mine Using Nuclear Track Detectors: Radiation Dose Assessment

Authors: L. Oufni, M. Amrane, R. Rabi

Abstract:

Radon (and thoron) is a naturally occurring radioactive noble gas, having variable distribution in the geological environment. The exposure of human beings to ionizing radiation from natural sources is a continuing and inescapable feature of life on earth. Radon, thoron and their short-lived decay products in the atmosphere are the most important contributors to human exposure from natural sources. The aim of this study is to determine alpha-and beta-activities per unit volume of air due to radon (222Rn), thoron (220Rn) and their progenies in the air of ancient mine of Aouli in which there is no working activity is situated at approximately 25 km north of the city of Midelt (Morocco), by using LR-115 type II and CR-39 solid state nuclear track detectors (SSNTDs). Equilibrium factors between radon and its daughters and between thoron and its progeny were evaluated in the studied atmospheres. The committed equivalent doses due to the 218Po and 214Po radon short-lived progeny were evaluated in different tissues of the respiratory tract of the visitors of the considered ancient mine. The visitors in these mines spent a good amount of time. It was essential to let the staff know about these values and take the needed steps to prevent any health complications.

Keywords: radon, thoron, concentration, exposure dose, SSNTD, mine

Procedia PDF Downloads 532
812 Data Science-Based Key Factor Analysis and Risk Prediction of Diabetic

Authors: Fei Gao, Rodolfo C. Raga Jr.

Abstract:

This research proposal will ascertain the major risk factors for diabetes and to design a predictive model for risk assessment. The project aims to improve diabetes early detection and management by utilizing data science techniques, which may improve patient outcomes and healthcare efficiency. The phase relation values of each attribute were used to analyze and choose the attributes that might influence the examiner's survival probability using Diabetes Health Indicators Dataset from Kaggle’s data as the research data. We compare and evaluate eight machine learning algorithms. Our investigation begins with comprehensive data preprocessing, including feature engineering and dimensionality reduction, aimed at enhancing data quality. The dataset, comprising health indicators and medical data, serves as a foundation for training and testing these algorithms. A rigorous cross-validation process is applied, and we assess their performance using five key metrics like accuracy, precision, recall, F1-score, and area under the receiver operating characteristic curve (AUC-ROC). After analyzing the data characteristics, investigate their impact on the likelihood of diabetes and develop corresponding risk indicators.

Keywords: diabetes, risk factors, predictive model, risk assessment, data science techniques, early detection, data analysis, Kaggle

Procedia PDF Downloads 66
811 Pharmacophore-Based Modeling of a Series of Human Glutaminyl Cyclase Inhibitors to Identify Lead Molecules by Virtual Screening, Molecular Docking and Molecular Dynamics Simulation Study

Authors: Ankur Chaudhuri, Sibani Sen Chakraborty

Abstract:

In human, glutaminyl cyclase activity is highly abundant in neuronal and secretory tissues and is preferentially restricted to hypothalamus and pituitary. The N-terminal modification of β-amyloids (Aβs) peptides by the generation of a pyro-glutamyl (pGlu) modified Aβs (pE-Aβs) is an important process in the initiation of the formation of neurotoxic plaques in Alzheimer’s disease (AD). This process is catalyzed by glutaminyl cyclase (QC). The expression of QC is characteristically up-regulated in the early stage of AD, and the hallmark of the inhibition of QC is the prevention of the formation of pE-Aβs and plaques. A computer-aided drug design (CADD) process was employed to give an idea for the designing of potentially active compounds to understand the inhibitory potency against human glutaminyl cyclase (QC). This work elaborates the ligand-based and structure-based pharmacophore exploration of glutaminyl cyclase (QC) by using the known inhibitors. Three dimensional (3D) quantitative structure-activity relationship (QSAR) methods were applied to 154 compounds with known IC50 values. All the inhibitors were divided into two sets, training-set, and test-sets. Generally, training-set was used to build the quantitative pharmacophore model based on the principle of structural diversity, whereas the test-set was employed to evaluate the predictive ability of the pharmacophore hypotheses. A chemical feature-based pharmacophore model was generated from the known 92 training-set compounds by HypoGen module implemented in Discovery Studio 2017 R2 software package. The best hypothesis was selected (Hypo1) based upon the highest correlation coefficient (0.8906), lowest total cost (463.72), and the lowest root mean square deviation (2.24Å) values. The highest correlation coefficient value indicates greater predictive activity of the hypothesis, whereas the lower root mean square deviation signifies a small deviation of experimental activity from the predicted one. The best pharmacophore model (Hypo1) of the candidate inhibitors predicted comprised four features: two hydrogen bond acceptor, one hydrogen bond donor, and one hydrophobic feature. The Hypo1 was validated by several parameters such as test set activity prediction, cost analysis, Fischer's randomization test, leave-one-out method, and heat map of ligand profiler. The predicted features were then used for virtual screening of potential compounds from NCI, ASINEX, Maybridge and Chembridge databases. More than seven million compounds were used for this purpose. The hit compounds were filtered by drug-likeness and pharmacokinetics properties. The selective hits were docked to the high-resolution three-dimensional structure of the target protein glutaminyl cyclase (PDB ID: 2AFU/2AFW) to filter these hits further. To validate the molecular docking results, the most active compound from the dataset was selected as a reference molecule. From the density functional theory (DFT) study, ten molecules were selected based on their highest HOMO (highest occupied molecular orbitals) energy and the lowest bandgap values. Molecular dynamics simulations with explicit solvation systems of the final ten hit compounds revealed that a large number of non-covalent interactions were formed with the binding site of the human glutaminyl cyclase. It was suggested that the hit compounds reported in this study could help in future designing of potent inhibitors as leads against human glutaminyl cyclase.

Keywords: glutaminyl cyclase, hit lead, pharmacophore model, simulation

Procedia PDF Downloads 128
810 Effect of Large English Studies Classes on Linguistic Achievement and Classroom Discourse at Junior Secondary Level in Yobe State

Authors: Clifford Irikefe Gbeyonron

Abstract:

Applied linguists concur that there is low-level achievement in English language use among Nigerian secondary school students. One of the factors that exacerbate this is classroom feature of which large class size is obvious. This study investigated the impact of large classes on learning English as a second language (ESL) at junior secondary school (JSS) in Yobe State. To achieve this, Solomon four-group experimental design was used. 382 subjects were divided into four groups and taught ESL for thirteen weeks. 356 subjects wrote the post-test. Data from the systematic observation and post-test were analyzed via chi square and ANOVA. Results indicated that learners in large classes (LLC) attain lower linguistic progress than learners in small classes (LSC). Furthermore, LSC have more chances to access teacher evaluation and participate actively in classroom discourse than LLC. In consequence, large classes have adverse effects on learning ESL in Yobe State. This is inimical to English language education given that each learner of ESL has their individual peculiarity within each class. It is recommended that strategies that prioritize individualization, grouping, use of language teaching aides, and theorization of innovative models in respect of large classes be considered.

Keywords: large classes, achievement, classroom discourse

Procedia PDF Downloads 402
809 Sea-Land Segmentation Method Based on the Transformer with Enhanced Edge Supervision

Authors: Lianzhong Zhang, Chao Huang

Abstract:

Sea-land segmentation is a basic step in many tasks such as sea surface monitoring and ship detection. The existing sea-land segmentation algorithms have poor segmentation accuracy, and the parameter adjustments are cumbersome and difficult to meet actual needs. Also, the current sea-land segmentation adopts traditional deep learning models that use Convolutional Neural Networks (CNN). At present, the transformer architecture has achieved great success in the field of natural images, but its application in the field of radar images is less studied. Therefore, this paper proposes a sea-land segmentation method based on the transformer architecture to strengthen edge supervision. It uses a self-attention mechanism with a gating strategy to better learn relative position bias. Meanwhile, an additional edge supervision branch is introduced. The decoder stage allows the feature information of the two branches to interact, thereby improving the edge precision of the sea-land segmentation. Based on the Gaofen-3 satellite image dataset, the experimental results show that the method proposed in this paper can effectively improve the accuracy of sea-land segmentation, especially the accuracy of sea-land edges. The mean IoU (Intersection over Union), edge precision, overall precision, and F1 scores respectively reach 96.36%, 84.54%, 99.74%, and 98.05%, which are superior to those of the mainstream segmentation models and have high practical application values.

Keywords: SAR, sea-land segmentation, deep learning, transformer

Procedia PDF Downloads 168
808 Human Action Recognition Using Variational Bayesian HMM with Dirichlet Process Mixture of Gaussian Wishart Emission Model

Authors: Wanhyun Cho, Soonja Kang, Sangkyoon Kim, Soonyoung Park

Abstract:

In this paper, we present the human action recognition method using the variational Bayesian HMM with the Dirichlet process mixture (DPM) of the Gaussian-Wishart emission model (GWEM). First, we define the Bayesian HMM based on the Dirichlet process, which allows an infinite number of Gaussian-Wishart components to support continuous emission observations. Second, we have considered an efficient variational Bayesian inference method that can be applied to drive the posterior distribution of hidden variables and model parameters for the proposed model based on training data. And then we have derived the predictive distribution that may be used to classify new action. Third, the paper proposes a process of extracting appropriate spatial-temporal feature vectors that can be used to recognize a wide range of human behaviors from input video image. Finally, we have conducted experiments that can evaluate the performance of the proposed method. The experimental results show that the method presented is more efficient with human action recognition than existing methods.

Keywords: human action recognition, Bayesian HMM, Dirichlet process mixture model, Gaussian-Wishart emission model, Variational Bayesian inference, prior distribution and approximate posterior distribution, KTH dataset

Procedia PDF Downloads 345