Search results for: geometric network
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5232

Search results for: geometric network

3642 Opinion Mining and Sentiment Analysis on DEFT

Authors: Najiba Ouled Omar, Azza Harbaoui, Henda Ben Ghezala

Abstract:

Current research practices sentiment analysis with a focus on social networks, DEfi Fouille de Texte (DEFT) (Text Mining Challenge) evaluation campaign focuses on opinion mining and sentiment analysis on social networks, especially social network Twitter. It aims to confront the systems produced by several teams from public and private research laboratories. DEFT offers participants the opportunity to work on regularly renewed themes and proposes to work on opinion mining in several editions. The purpose of this article is to scrutinize and analyze the works relating to opinions mining and sentiment analysis in the Twitter social network realized by DEFT. It examines the tasks proposed by the organizers of the challenge and the methods used by the participants.

Keywords: opinion mining, sentiment analysis, emotion, polarity, annotation, OSEE, figurative language, DEFT, Twitter, Tweet

Procedia PDF Downloads 124
3641 A Study of Human Communication in an Internet Community

Authors: Andrew Laghos

Abstract:

The Internet is a big part of our everyday lives. People can now access the internet from a variety of places including home, college, and work. Many airports, hotels, restaurants and cafeterias, provide free wireless internet to their visitors. Using technologies like computers, tablets, and mobile phones, we spend a lot of our time online getting entertained, getting informed, and communicating with each other. This study deals with the latter part, namely, human communication through the Internet. People can communicate with each other using social media, social network sites (SNS), e-mail, messengers, chatrooms, and so on. By connecting with each other they form virtual communities. Regarding SNS, types of connections that can be studied include friendships and cliques. Analyzing these connections is important to help us understand online user behavior. The method of Social Network Analysis (SNA) was used on a case study, and results revealed the existence of some useful patterns of interactivity between the participants. The study ends with implications of the results and ideas for future research.

Keywords: human communication, internet communities, online user behavior, psychology

Procedia PDF Downloads 490
3640 An Implementation of Meshless Method for Modeling an Elastoplasticity Coupled to Damage

Authors: Sendi Zohra, Belhadjsalah Hedi, Labergere Carl, Saanouni Khemais

Abstract:

The modeling of mechanical problems including both material and geometric nonlinearities with Finite Element Method (FEM) remains challenging. Meshless methods offer special properties to get rid of well-known drawbacks of the FEM. The main objective of Meshless Methods is to eliminate the difficulty of meshing and remeshing the entire structure by simply insertion or deletion of nodes, and alleviate other problems associated with the FEM, such as element distortion, locking and others. In this study, a robust numerical implementation of an Element Free Galerkin Method for an elastoplastic coupled to damage problem is presented. Several results issued from the numerical simulations by a DynamicExplicit resolution scheme are analyzed and critically compared with Element Finite Method results. Finally, different numerical examples are carried out to demonstrate the efficiency of this method.

Keywords: damage, dynamic explicit, elastoplasticity, isotropic hardening, meshless

Procedia PDF Downloads 279
3639 Security Issues in Long Term Evolution-Based Vehicle-To-Everything Communication Networks

Authors: Mujahid Muhammad, Paul Kearney, Adel Aneiba

Abstract:

The ability for vehicles to communicate with other vehicles (V2V), the physical (V2I) and network (V2N) infrastructures, pedestrians (V2P), etc. – collectively known as V2X (Vehicle to Everything) – will enable a broad and growing set of applications and services within the intelligent transport domain for improving road safety, alleviate traffic congestion and support autonomous driving. The telecommunication research and industry communities and standardization bodies (notably 3GPP) has finally approved in Release 14, cellular communications connectivity to support V2X communication (known as LTE – V2X). LTE – V2X system will combine simultaneous connectivity across existing LTE network infrastructures via LTE-Uu interface and direct device-to-device (D2D) communications. In order for V2X services to function effectively, a robust security mechanism is needed to ensure legal and safe interaction among authenticated V2X entities in the LTE-based V2X architecture. The characteristics of vehicular networks, and the nature of most V2X applications, which involve human safety makes it significant to protect V2X messages from attacks that can result in catastrophically wrong decisions/actions include ones affecting road safety. Attack vectors include impersonation attacks, modification, masquerading, replay, MiM attacks, and Sybil attacks. In this paper, we focus our attention on LTE-based V2X security and access control mechanisms. The current LTE-A security framework provides its own access authentication scheme, the AKA protocol for mutual authentication and other essential cryptographic operations between UEs and the network. V2N systems can leverage this protocol to achieve mutual authentication between vehicles and the mobile core network. However, this protocol experiences technical challenges, such as high signaling overhead, lack of synchronization, handover delay and potential control plane signaling overloads, as well as privacy preservation issues, which cannot satisfy the adequate security requirements for majority of LTE-based V2X services. This paper examines these challenges and points to possible ways by which they can be addressed. One possible solution, is the implementation of the distributed peer-to-peer LTE security mechanism based on the Bitcoin/Namecoin framework, to allow for security operations with minimal overhead cost, which is desirable for V2X services. The proposed architecture can ensure fast, secure and robust V2X services under LTE network while meeting V2X security requirements.

Keywords: authentication, long term evolution, security, vehicle-to-everything

Procedia PDF Downloads 157
3638 Linac Quality Controls Using An Electronic Portal Imaging Device

Authors: Domingo Planes Meseguer, Raffaele Danilo Esposito, Maria Del Pilar Dorado Rodriguez

Abstract:

Monthly quality control checks for a Radiation Therapy Linac may be performed is a simple and efficient way once they have been standardized and protocolized. On the other hand this checks, in spite of being imperatives, require a not negligible execution times in terms of machine time and operators time. Besides it must be taken into account the amount of disposable material which may be needed together with the use of commercial software for their performing. With the aim of optimizing and standardizing mechanical-geometric checks and multi leaves collimator checks, we decided to implement a protocol which makes use of the Electronic Portal Imaging Device (EPID) available on our Linacs. The user is step by step guided by the software during the whole procedure. Acquired images are automatically analyzed by our programs all of them written using only free software.

Keywords: quality control checks, linac, radiation oncology, medical physics, free software

Procedia PDF Downloads 189
3637 Convolution Neural Network Based on Hypnogram of Sleep Stages to Predict Dosages and Types of Hypnotic Drugs for Insomnia

Authors: Chi Wu, Dean Wu, Wen-Te Liu, Cheng-Yu Tsai, Shin-Mei Hsu, Yin-Tzu Lin, Ru-Yin Yang

Abstract:

Background: The results of previous studies compared the benefits and risks of receiving insomnia medication. However, the effects between hypnotic drugs used and enhancement of sleep quality were still unclear. Objective: The aim of this study is to establish a prediction model for hypnotic drugs' dosage used for insomnia subjects and associated the relationship between sleep stage ratio change and drug types. Methodologies: According to American Academy of Sleep Medicine (AASM) guideline, sleep stages were classified and transformed to hypnogram via the polysomnography (PSG) in a hospital in New Taipei City (Taiwan). The subjects with diagnosis for insomnia without receiving hypnotic drugs treatment were be set as the comparison group. Conversely, hypnotic drugs dosage within the past three months was obtained from the clinical registration for each subject. Furthermore, the collecting subjects were divided into two groups for training and testing. After training convolution neuron network (CNN) to predict types of hypnotics used and dosages are taken, the test group was used to evaluate the accuracy of classification. Results: We recruited 76 subjects in this study, who had been done PSG for transforming hypnogram from their sleep stages. The accuracy of dosages obtained from confusion matrix on the test group by CNN is 81.94%, and accuracy of hypnotic drug types used is 74.22%. Moreover, the subjects with high ratio of wake stage were correctly classified as requiring medical treatment. Conclusion: CNN with hypnogram was potentially used for adjusting the dosage of hypnotic drugs and providing subjects to pre-screening the types of hypnotic drugs taken.

Keywords: convolution neuron network, hypnotic drugs, insomnia, polysomnography

Procedia PDF Downloads 181
3636 A Survey of Field Programmable Gate Array-Based Convolutional Neural Network Accelerators

Authors: Wei Zhang

Abstract:

With the rapid development of deep learning, neural network and deep learning algorithms play a significant role in various practical applications. Due to the high accuracy and good performance, Convolutional Neural Networks (CNNs) especially have become a research hot spot in the past few years. However, the size of the networks becomes increasingly large scale due to the demands of the practical applications, which poses a significant challenge to construct a high-performance implementation of deep learning neural networks. Meanwhile, many of these application scenarios also have strict requirements on the performance and low-power consumption of hardware devices. Therefore, it is particularly critical to choose a moderate computing platform for hardware acceleration of CNNs. This article aimed to survey the recent advance in Field Programmable Gate Array (FPGA)-based acceleration of CNNs. Various designs and implementations of the accelerator based on FPGA under different devices and network models are overviewed, and the versions of Graphic Processing Units (GPUs), Application Specific Integrated Circuits (ASICs) and Digital Signal Processors (DSPs) are compared to present our own critical analysis and comments. Finally, we give a discussion on different perspectives of these acceleration and optimization methods on FPGA platforms to further explore the opportunities and challenges for future research. More helpfully, we give a prospect for future development of the FPGA-based accelerator.

Keywords: deep learning, field programmable gate array, FPGA, hardware accelerator, convolutional neural networks, CNN

Procedia PDF Downloads 113
3635 Multi-Sender MAC Protocol Based on Temporal Reuse in Underwater Acoustic Networks

Authors: Dongwon Lee, Sunmyeng Kim

Abstract:

Underwater acoustic networks (UANs) have become a very active research area in recent years. Compared with wireless networks, UANs are characterized by the limited bandwidth, long propagation delay and high channel dynamic in acoustic modems, which pose challenges to the design of medium access control (MAC) protocol. The characteristics severely affect network performance. In this paper, we study a MS-MAC (Multi-Sender MAC) protocol in order to improve network performance. The proposed protocol exploits temporal reuse by learning the propagation delays to neighboring nodes. A source node locally calculates the transmission schedules of its neighboring nodes and itself based on the propagation delays to avoid collisions. Performance evaluation is conducted using simulation, and confirms that the proposed protocol significantly outperforms the previous protocol in terms of throughput.

Keywords: acoustic channel, MAC, temporal reuse, UAN

Procedia PDF Downloads 338
3634 Next Generation Radiation Risk Assessment and Prediction Tools Generation Applying AI-Machine (Deep) Learning Algorithms

Authors: Selim M. Khan

Abstract:

Indoor air quality is strongly influenced by the presence of radioactive radon (222Rn) gas. Indeed, exposure to high 222Rn concentrations is unequivocally linked to DNA damage and lung cancer and is a worsening issue in North American and European built environments, having increased over time within newer housing stocks as a function of as yet unclear variables. Indoor air radon concentration can be influenced by a wide range of environmental, structural, and behavioral factors. As some of these factors are quantitative while others are qualitative, no single statistical model can determine indoor radon level precisely while simultaneously considering all these variables across a complex and highly diverse dataset. The ability of AI- machine (deep) learning to simultaneously analyze multiple quantitative and qualitative features makes it suitable to predict radon with a high degree of precision. Using Canadian and Swedish long-term indoor air radon exposure data, we are using artificial deep neural network models with random weights and polynomial statistical models in MATLAB to assess and predict radon health risk to human as a function of geospatial, human behavioral, and built environmental metrics. Our initial artificial neural network with random weights model run by sigmoid activation tested different combinations of variables and showed the highest prediction accuracy (>96%) within the reasonable iterations. Here, we present details of these emerging methods and discuss strengths and weaknesses compared to the traditional artificial neural network and statistical methods commonly used to predict indoor air quality in different countries. We propose an artificial deep neural network with random weights as a highly effective method for assessing and predicting indoor radon.

Keywords: radon, radiation protection, lung cancer, aI-machine deep learnng, risk assessment, risk prediction, Europe, North America

Procedia PDF Downloads 87
3633 Impact of Transportation on the Economic Growth of Nigeria

Authors: E. O. E. Nnadi

Abstract:

Transportation is a critical factor in the economic growth and development of any nation, region or state. Good transportation network supports every sector of the economy like the manufacturing, transportation and encourages investors thereby affect the overall economic prosperity. The paper evaluates the impact of transportation on the economic growth of Nigeria using south eastern states as a case study. The choice of the case study is its importance as the commercial and industrial nerve of the country. About 200 respondents who are of different professions such as dealers in goods, transporters, contractors, consultants, bankers were selected and a set of questionnaire were administered to using the systematic sampling technique in the five states of the region. Descriptive statistics and relative importance index (RII) technique was employed for the analysis of the data gathered. The findings of the analysis reveal that Nigeria has the least effective ratio per population in Africa of 949.91 km/Person. Conclusion was drawn to improve road network in the area and the country as a whole to enhance the economic activities of the people.

Keywords: economic growth, south-east, transportation, transportation cost, Nigeria

Procedia PDF Downloads 264
3632 Coal Mining Safety Monitoring Using Wsn

Authors: Somdatta Saha

Abstract:

The main purpose was to provide an implementable design scenario for underground coal mines using wireless sensor networks (WSNs). The main reason being that given the intricacies in the physical structure of a coal mine, only low power WSN nodes can produce accurate surveillance and accident detection data. The work mainly concentrated on designing and simulating various alternate scenarios for a typical mine and comparing them based on the obtained results to arrive at a final design. In the Era of embedded technology, the Zigbee protocols are used in more and more applications. Because of the rapid development of sensors, microcontrollers, and network technology, a reliable technological condition has been provided for our automatic real-time monitoring of coal mine. The underground system collects temperature, humidity and methane values of coal mine through sensor nodes in the mine; it also collects the number of personnel inside the mine with the help of an IR sensor, and then transmits the data to information processing terminal based on ARM.

Keywords: ARM, embedded board, wireless sensor network (Zigbee)

Procedia PDF Downloads 334
3631 A Study of Adult Lifelong Learning Consulting and Service System in Taiwan

Authors: Wan Jen Chang

Abstract:

Back ground: Taiwan's current adult lifelong learning services have expanded from vocational training to universal lifelong learning. However, both the professional knowledge training of learning guidance and consulting services and the provision of adult online learning consulting service systems still need to be established. Purpose: The purposes of this study are as follows: 1. Analyze the professional training mechanism for cultivating adult lifelong learning consultation and coaching; 2. Explore the feasibility of constructing a system that uses network technology to provide adult learning consultation services. Research design: This study conducts a literature analysis of counseling and coaching policy reports on lifelong learning in European countries and the United States. There are two focus discussions were conducted with 15 lifelong learning scholars, experts and practitioners as research subjects. The following two topics were discussed and suggested: 1. The current situation, needs and professional ability training mechanism of "Adult Lifelong Learning Consulting and Services"; 2. Strategies for establishing an "Adult Lifelong Learning Consulting and Service internet System". Conclusion: 1.Based on adult lifelong learning consulting and service needs, plan a professional knowledge training and certification system.2.Adult lifelong learning consulting and service professional knowledge and skills training should include the use of network technology to provide consulting service skills.3.To establish an adult lifelong learning consultation and service system, the Ministry of Education should promulgate policies and measures at the central level and entrust local governments or private organizations to implement them.4.The adult lifelong learning consulting and service system can combine the national qualifications framework, private sector and NPO to expand learning consulting service partners.

Keywords: adult lifelong learning, profesional knowledge, consulting and service, network system

Procedia PDF Downloads 59
3630 Presenting a Job Scheduling Algorithm Based on Learning Automata in Computational Grid

Authors: Roshanak Khodabakhsh Jolfaei, Javad Akbari Torkestani

Abstract:

As a cooperative environment for problem-solving, it is necessary that grids develop efficient job scheduling patterns with regard to their goals, domains and structure. Since the Grid environments facilitate distributed calculations, job scheduling appears in the form of a critical problem for the management of Grid sources that influences severely on the efficiency for the whole Grid environment. Due to the existence of some specifications such as sources dynamicity and conditions of the network in Grid, some algorithm should be presented to be adjustable and scalable with increasing the network growth. For this purpose, in this paper a job scheduling algorithm has been presented on the basis of learning automata in computational Grid which the performance of its results were compared with FPSO algorithm (Fuzzy Particle Swarm Optimization algorithm) and GJS algorithm (Grid Job Scheduling algorithm). The obtained numerical results indicated the superiority of suggested algorithm in comparison with FPSO and GJS. In addition, the obtained results classified FPSO and GJS in the second and third position respectively after the mentioned algorithm.

Keywords: computational grid, job scheduling, learning automata, dynamic scheduling

Procedia PDF Downloads 327
3629 A Numerical Hybrid Finite Element Model for Lattice Structures Using 3D/Beam Elements

Authors: Ahmadali Tahmasebimoradi, Chetra Mang, Xavier Lorang

Abstract:

Thanks to the additive manufacturing process, lattice structures are replacing the traditional structures in aeronautical and automobile industries. In order to evaluate the mechanical response of the lattice structures, one has to resort to numerical techniques. Ansys is a globally well-known and trusted commercial software that allows us to model the lattice structures and analyze their mechanical responses using either solid or beam elements. In this software, a script may be used to systematically generate the lattice structures for any size. On the one hand, solid elements allow us to correctly model the contact between the substrates (the supports of the lattice structure) and the lattice structure, the local plasticity, and the junctions of the microbeams. However, their computational cost increases rapidly with the size of the lattice structure. On the other hand, although beam elements reduce the computational cost drastically, it doesn’t correctly model the contact between the lattice structures and the substrates nor the junctions of the microbeams. Also, the notion of local plasticity is not valid anymore. Moreover, the deformed shape of the lattice structure doesn’t correspond to the deformed shape of the lattice structure using 3D solid elements. In this work, motivated by the pros and cons of the 3D and beam models, a numerically hybrid model is presented for the lattice structures to reduce the computational cost of the simulations while avoiding the aforementioned drawbacks of the beam elements. This approach consists of the utilization of solid elements for the junctions and beam elements for the microbeams connecting the corresponding junctions to each other. When the global response of the structure is linear, the results from the hybrid models are in good agreement with the ones from the 3D models for body-centered cubic with z-struts (BCCZ) and body-centered cubic without z-struts (BCC) lattice structures. However, the hybrid models have difficulty to converge when the effect of large deformation and local plasticity are considerable in the BCCZ structures. Furthermore, the effect of the junction’s size of the hybrid models on the results is investigated. For BCCZ lattice structures, the results are not affected by the junction’s size. This is also valid for BCC lattice structures as long as the ratio of the junction’s size to the diameter of the microbeams is greater than 2. The hybrid model can take into account the geometric defects. As a demonstration, the point clouds of two lattice structures are parametrized in a platform called LATANA (LATtice ANAlysis) developed by IRT-SystemX. In this process, for each microbeam of the lattice structures, an ellipse is fitted to capture the effect of shape variation and roughness. Each ellipse is represented by three parameters; semi-major axis, semi-minor axis, and angle of rotation. Having the parameters of the ellipses, the lattice structures are constructed in Spaceclaim (ANSYS) using the geometrical hybrid approach. The results show a negligible discrepancy between the hybrid and 3D models, while the computational cost of the hybrid model is lower than the computational cost of the 3D model.

Keywords: additive manufacturing, Ansys, geometric defects, hybrid finite element model, lattice structure

Procedia PDF Downloads 107
3628 Cascaded Neural Network for Internal Temperature Forecasting in Induction Motor

Authors: Hidir S. Nogay

Abstract:

In this study, two systems were created to predict interior temperature in induction motor. One of them consisted of a simple ANN model which has two layers, ten input parameters and one output parameter. The other one consisted of eight ANN models connected each other as cascaded. Cascaded ANN system has 17 inputs. Main reason of cascaded system being used in this study is to accomplish more accurate estimation by increasing inputs in the ANN system. Cascaded ANN system is compared with simple conventional ANN model to prove mentioned advantages. Dataset was obtained from experimental applications. Small part of the dataset was used to obtain more understandable graphs. Number of data is 329. 30% of the data was used for testing and validation. Test data and validation data were determined for each ANN model separately and reliability of each model was tested. As a result of this study, it has been understood that the cascaded ANN system produced more accurate estimates than conventional ANN model.

Keywords: cascaded neural network, internal temperature, inverter, three-phase induction motor

Procedia PDF Downloads 335
3627 Genome-Wide Expression Profiling of Cicer arietinum Heavy Metal Toxicity

Authors: B. S. Yadav, A. Mani, S. Srivastava

Abstract:

Chickpea (Cicer arietinum L.) is an annual, self-pollinating, diploid (2n = 2x = 16) pulse crop that ranks second in world legume production after common bean (Phaseolus vulgaris). ICC 4958 flowers approximately 39 days after sowing under peninsular Indian conditions and the crop matures in less than 90 days in rained environments. The estimated collective yield losses due to abiotic stresses (6.4 million t) have been significantly higher than for biotic stresses (4.8 million t). Most legumes are known to be salt sensitive, and therefore, it is becoming increasingly important to produce cultivars tolerant to high-salinity in addition to other abiotic and biotic stresses for sustainable chickpea production. Our aim was to identify the genes that are involved in the defence mechanism against heavy metal toxicity in chickpea and establish the biological network of heavy metal toxicity in chickpea. ICC4958 variety of chick pea was taken and grown in normal condition and 150µM concentration of different heavy metal salt like CdCl₂, K₂Cr2O₇, NaAsO₂. At 15th day leave samples were collected and stored in RNA Later solution microarray was performed for checking out differential gene expression pattern. Our studies revealed that 111 common genes that involved in defense mechanism were up regulated and 41 genes were commonly down regulated during treatment of 150µM concentration of CdCl₂, K₂Cr₂O₇, and NaAsO₂. Biological network study shows that the genes which are differentially expressed are highly connected and having high betweenness and centrality.

Keywords: abiotic stress, biological network, chickpea, microarray

Procedia PDF Downloads 183
3626 Comprehensive Studies on the Aerodynamic Characteristics of Subsonic Scarf Inlets

Authors: M. Jegannath, V. Akshaya, B. Arunkumar, G. Lakshmi Soundharya, V. Thenmozhi, S. Varun, V. R. S. Kumar

Abstract:

For scarf inlet design, the primary variable of interest is the circumferential extent over which the extended lower lip is formed. In this paper, an attempt has been made to optimize the aerodynamic shape of a subsonic scarf inlet with aerodynamically shaped center-body with a particular value of the circumferential extent. The parametric analytical studies have been carried out using a Spalart-Allmaras turbulence model. From our preliminary studies, we concluded that for a particular value of circumferential extent, there will be an exact shape of the center-body with certain geometric orientation for the existence of an aerodynamically efficient scarf inlet for modern aircraft engines. This numerical study is a pointer towards for the design optimization of scarf inlets for modern aircraft engines.

Keywords: aerodynamics of scarf inlets, inlet design, modern aircraft inlets, subsonic scarf inlet

Procedia PDF Downloads 298
3625 Control of Belts for Classification of Geometric Figures by Artificial Vision

Authors: Juan Sebastian Huertas Piedrahita, Jaime Arturo Lopez Duque, Eduardo Luis Perez Londoño, Julián S. Rodríguez

Abstract:

The process of generating computer vision is called artificial vision. The artificial vision is a branch of artificial intelligence that allows the obtaining, processing, and analysis of any type of information especially the ones obtained through digital images. Actually the artificial vision is used in manufacturing areas for quality control and production, as these processes can be realized through counting algorithms, positioning, and recognition of objects that can be measured by a single camera (or more). On the other hand, the companies use assembly lines formed by conveyor systems with actuators on them for moving pieces from one location to another in their production. These devices must be previously programmed for their good performance and must have a programmed logic routine. Nowadays the production is the main target of every industry, quality, and the fast elaboration of the different stages and processes in the chain of production of any product or service being offered. The principal base of this project is to program a computer that recognizes geometric figures (circle, square, and triangle) through a camera, each one with a different color and link it with a group of conveyor systems to organize the mentioned figures in cubicles, which differ from one another also by having different colors. This project bases on artificial vision, therefore the methodology needed to develop this project must be strict, this one is detailed below: 1. Methodology: 1.1 The software used in this project is QT Creator which is linked with Open CV libraries. Together, these tools perform to realize the respective program to identify colors and forms directly from the camera to the computer. 1.2 Imagery acquisition: To start using the libraries of Open CV is necessary to acquire images, which can be captured by a computer’s web camera or a different specialized camera. 1.3 The recognition of RGB colors is realized by code, crossing the matrices of the captured images and comparing pixels, identifying the primary colors which are red, green, and blue. 1.4 To detect forms it is necessary to realize the segmentation of the images, so the first step is converting the image from RGB to grayscale, to work with the dark tones of the image, then the image is binarized which means having the figure of the image in a white tone with a black background. Finally, we find the contours of the figure in the image to detect the quantity of edges to identify which figure it is. 1.5 After the color and figure have been identified, the program links with the conveyor systems, which through the actuators will classify the figures in their respective cubicles. Conclusions: The Open CV library is a useful tool for projects in which an interface between a computer and the environment is required since the camera obtains external characteristics and realizes any process. With the program for this project any type of assembly line can be optimized because images from the environment can be obtained and the process would be more accurate.

Keywords: artificial intelligence, artificial vision, binarized, grayscale, images, RGB

Procedia PDF Downloads 370
3624 Time Efficient Color Coding for Structured-Light 3D Scanner

Authors: Po-Hao Huang, Pei-Ju Chiang

Abstract:

The structured light 3D scanner is commonly used for measuring the 3D shape of an object. Through projecting designed light patterns on the object, deformed patterns can be obtained and used for the geometric shape reconstruction. At present, Gray code is the most reliable and commonly used light pattern in the structured light 3D scanner. However, the trade-off between scanning efficiency and accuracy is a long-standing and challenging problem. The design of light patterns plays a significant role in the scanning efficiency and accuracy. Thereby, we proposed a novel encoding method integrating color information and Gray-code to improve the scanning efficiency. We will demonstrate that with the proposed method, the scanning time can be reduced to approximate half of the one needed by Gray-code without reduction of precision.

Keywords: gray-code, structured light scanner, 3D shape acquisition, 3D reconstruction

Procedia PDF Downloads 449
3623 Attention-based Adaptive Convolution with Progressive Learning in Speech Enhancement

Authors: Tian Lan, Yixiang Wang, Wenxin Tai, Yilan Lyu, Zufeng Wu

Abstract:

The monaural speech enhancement task in the time-frequencydomain has a myriad of approaches, with the stacked con-volutional neural network (CNN) demonstrating superiorability in feature extraction and selection. However, usingstacked single convolutions method limits feature represen-tation capability and generalization ability. In order to solvethe aforementioned problem, we propose an attention-basedadaptive convolutional network that integrates the multi-scale convolutional operations into a operation-specific blockvia input dependent attention to adapt to complex auditoryscenes. In addition, we introduce a two-stage progressivelearning method to enlarge the receptive field without a dra-matic increase in computation burden. We conduct a series ofexperiments based on the TIMIT corpus, and the experimen-tal results prove that our proposed model is better than thestate-of-art models on all metrics.

Keywords: speech enhancement, adaptive convolu-tion, progressive learning, time-frequency domain

Procedia PDF Downloads 112
3622 Estimation of Endogenous Brain Noise from Brain Response to Flickering Visual Stimulation Magnetoencephalography Visual Perception Speed

Authors: Alexander N. Pisarchik, Parth Chholak

Abstract:

Intrinsic brain noise was estimated via magneto-encephalograms (MEG) recorded during perception of flickering visual stimuli with frequencies of 6.67 and 8.57 Hz. First, we measured the mean phase difference between the flicker signal and steady-state event-related field (SSERF) in the occipital area where the brain response at the flicker frequencies and their harmonics appeared in the power spectrum. Then, we calculated the probability distribution of the phase fluctuations in the regions of frequency locking and computed its kurtosis. Since kurtosis is a measure of the distribution’s sharpness, we suppose that inverse kurtosis is related to intrinsic brain noise. In our experiments, the kurtosis value varied among subjects from K = 3 to K = 5 for 6.67 Hz and from 2.6 to 4 for 8.57 Hz. The majority of subjects demonstrated leptokurtic kurtosis (K < 3), i.e., the distribution tails approached zero more slowly than Gaussian. In addition, we found a strong correlation between kurtosis and brain complexity measured as the correlation dimension, so that the MEGs of subjects with higher kurtosis exhibited lower complexity. The obtained results are discussed in the framework of nonlinear dynamics and complex network theories. Specifically, in a network of coupled oscillators, phase synchronization is mainly determined by two antagonistic factors, noise, and the coupling strength. While noise worsens phase synchronization, the coupling improves it. If we assume that each neuron and each synapse contribute to brain noise, the larger neuronal network should have stronger noise, and therefore phase synchronization should be worse, that results in smaller kurtosis. The described method for brain noise estimation can be useful for diagnostics of some brain pathologies associated with abnormal brain noise.

Keywords: brain, flickering, magnetoencephalography, MEG, visual perception, perception time

Procedia PDF Downloads 134
3621 The Friendship Network Stability of Preschool Children during One Pedagogical Season

Authors: Yili Wang, Jarmo Kinos, Tuire Palonen, Tarja-Riitta Hurme

Abstract:

This longitudinal study aims to examine how five- and six-year-old children’s peer relationships are formed and fostered during one preschool year in a southwestern Finnish preschool. All 16 kindergarteners participated in the study (at dyad level N=240; i.e., 16 x 15 relationships among the children). The children were divided into four daily groups, based on the table order during the daily routines, and four intervention groups, based on the teachers’ pedagogical plan. During the intervention, one iPad was given to each group in order to stimulate interaction among peers and, thus, enable the children to form new peer relationships. In the data gathering, sociometric nomination techniques were used to investigate the nature (i.e., stability and mutuality) of the peer relationships. The data was collected five times during the year to see what kind of peer relationship changes occurred at the dyad level and the group level, i.e., in establishing and losing friendship ties among the children. Social network analyses were used to analyze the data. The results indicate that the children’s preference for gender segregation was strong compared to age preference and intervention. In all, the number of reciprocal friendship ties and the mutual absence of friendship ties increased towards the end of the year, whereas the number of unilateral friendship ties decreased. This indicates that children’s nominations narrow down; thus, the group structure becomes more crystalized. Instead of extending their friendship networks, children seek stable and mutual relationships with their peers in their middle childhood years. The intervention only had a slightly negative influence on children’s peer relationships.

Keywords: intervention study, peer relationship, preschool education, social network analysis, sociometric ratings

Procedia PDF Downloads 256
3620 Generalized Rough Sets Applied to Graphs Related to Urban Problems

Authors: Mihai Rebenciuc, Simona Mihaela Bibic

Abstract:

Branch of modern mathematics, graphs represent instruments for optimization and solving practical applications in various fields such as economic networks, engineering, network optimization, the geometry of social action, generally, complex systems including contemporary urban problems (path or transport efficiencies, biourbanism, & c.). In this paper is studied the interconnection of some urban network, which can lead to a simulation problem of a digraph through another digraph. The simulation is made univoc or more general multivoc. The concepts of fragment and atom are very useful in the study of connectivity in the digraph that is simulation - including an alternative evaluation of k- connectivity. Rough set approach in (bi)digraph which is proposed in premier in this paper contribute to improved significantly the evaluation of k-connectivity. This rough set approach is based on generalized rough sets - basic facts are presented in this paper.

Keywords: (bi)digraphs, rough set theory, systems of interacting agents, complex systems

Procedia PDF Downloads 232
3619 Fake Accounts Detection in Twitter Based on Minimum Weighted Feature Set

Authors: Ahmed ElAzab, Amira M. Idrees, Mahmoud A. Mahmoud, Hesham Hefny

Abstract:

Social networking sites such as Twitter and Facebook attracts over 500 million users across the world, for those users, their social life, even their practical life, has become interrelated. Their interaction with social networking has affected their life forever. Accordingly, social networking sites have become among the main channels that are responsible for vast dissemination of different kinds of information during real time events. This popularity in Social networking has led to different problems including the possibility of exposing incorrect information to their users through fake accounts which results to the spread of malicious content during life events. This situation can result to a huge damage in the real world to the society in general including citizens, business entities, and others. In this paper, we present a classification method for detecting fake accounts on Twitter. The study determines the minimized set of the main factors that influence the detection of the fake accounts on Twitter, then the determined factors have been applied using different classification techniques, a comparison of the results for these techniques has been performed and the most accurate algorithm is selected according to the accuracy of the results. The study has been compared with different recent research in the same area, this comparison has proved the accuracy of the proposed study. We claim that this study can be continuously applied on Twitter social network to automatically detect the fake accounts, moreover, the study can be applied on different Social network sites such as Facebook with minor changes according to the nature of the social network which are discussed in this paper.

Keywords: fake accounts detection, classification algorithms, twitter accounts analysis, features based techniques

Procedia PDF Downloads 390
3618 Railway Ballast Volumes Automated Estimation Based on LiDAR Data

Authors: Bahar Salavati Vie Le Sage, Ismaïl Ben Hariz, Flavien Viguier, Sirine Noura Kahil, Audrey Jacquin, Maxime Convert

Abstract:

The ballast layer plays a key role in railroad maintenance and the geometry of the track structure. Ballast also holds the track in place as the trains roll over it. Track ballast is packed between the sleepers and on the sides of railway tracks. An imbalance in ballast volume on the tracks can lead to safety issues as well as a quick degradation of the overall quality of the railway segment. If there is a lack of ballast in the track bed during the summer, there is a risk that the rails will expand and buckle slightly due to the high temperatures. Furthermore, the knowledge of the ballast quantities that will be excavated during renewal works is important for efficient ballast management. The volume of excavated ballast per meter of track can be calculated based on excavation depth, excavation width, volume of track skeleton (sleeper and rail) and sleeper spacing. Since 2012, SNCF has been collecting 3D points cloud data covering its entire railway network by using 3D laser scanning technology (LiDAR). This vast amount of data represents a modelization of the entire railway infrastructure, allowing to conduct various simulations for maintenance purposes. This paper aims to present an automated method for ballast volume estimation based on the processing of LiDAR data. The estimation of abnormal volumes in ballast on the tracks is performed by analyzing the cross-section of the track. Further, since the amount of ballast required varies depending on the track configuration, the knowledge of the ballast profile is required. Prior to track rehabilitation, excess ballast is often present in the ballast shoulders. Based on 3D laser scans, a Digital Terrain Model (DTM) was generated and automatic extraction of the ballast profiles from this data is carried out. The surplus in ballast is then estimated by performing a comparison between this ballast profile obtained empirically, and a geometric modelization of the theoretical ballast profile thresholds as dictated by maintenance standards. Ideally, this excess should be removed prior to renewal works and recycled to optimize the output of the ballast renewal machine. Based on these parameters, an application has been developed to allow the automatic measurement of ballast profiles. We evaluated the method on a 108 kilometers segment of railroad LiDAR scans, and the results show that the proposed algorithm detects ballast surplus that amounts to values close to the total quantities of spoil ballast excavated.

Keywords: ballast, railroad, LiDAR , cloud point, track ballast, 3D point

Procedia PDF Downloads 94
3617 Structural Identification for Layered Composite Structures through a Wave and Finite Element Methodology

Authors: Rilwan Kayode Apalowo, Dimitrios Chronopoulos

Abstract:

An approach for identifying the geometric and material characteristics of layered composite structures through an inverse wave and finite element methodology is proposed. These characteristics are obtained through multi-frequency single shot measurements. However, it is established that the frequency regime of the measurements does not matter, meaning that both ultrasonic and structural dynamics frequency spectra can be employed. Taking advantage of a full FE (finite elements) description of the periodic composite, the scheme is able to account for arbitrarily complex structures. In order to demonstrate the robustness of the presented scheme, it is applied to a sandwich composite panel and results are compared with that of experimental characterization techniques. Excellent agreement is obtained with the experimental measurements.

Keywords: structural identification, non-destructive evaluation, finite elements, wave propagation, layered structures, ultrasound

Procedia PDF Downloads 127
3616 Scheduling in a Single-Stage, Multi-Item Compatible Process Using Multiple Arc Network Model

Authors: Bokkasam Sasidhar, Ibrahim Aljasser

Abstract:

The problem of finding optimal schedules for each equipment in a production process is considered, which consists of a single stage of manufacturing and which can handle different types of products, where changeover for handling one type of product to the other type incurs certain costs. The machine capacity is determined by the upper limit for the quantity that can be processed for each of the products in a set up. The changeover costs increase with the number of set ups and hence to minimize the costs associated with the product changeover, the planning should be such that similar types of products should be processed successively so that the total number of changeovers and in turn the associated set up costs are minimized. The problem of cost minimization is equivalent to the problem of minimizing the number of set ups or equivalently maximizing the capacity utilization in between every set up or maximizing the total capacity utilization. Further, the production is usually planned against customers’ orders, and generally different customers’ orders are assigned one of the two priorities – “normal” or “priority” order. The problem of production planning in such a situation can be formulated into a Multiple Arc Network (MAN) model and can be solved sequentially using the algorithm for maximizing flow along a MAN and the algorithm for maximizing flow along a MAN with priority arcs. The model aims to provide optimal production schedule with an objective of maximizing capacity utilization, so that the customer-wise delivery schedules are fulfilled, keeping in view the customer priorities. Algorithms have been presented for solving the MAN formulation of the production planning with customer priorities. The application of the model is demonstrated through numerical examples.

Keywords: scheduling, maximal flow problem, multiple arc network model, optimization

Procedia PDF Downloads 393
3615 Incorporating Lexical-Semantic Knowledge into Convolutional Neural Network Framework for Pediatric Disease Diagnosis

Authors: Xiaocong Liu, Huazhen Wang, Ting He, Xiaozheng Li, Weihan Zhang, Jian Chen

Abstract:

The utilization of electronic medical record (EMR) data to establish the disease diagnosis model has become an important research content of biomedical informatics. Deep learning can automatically extract features from the massive data, which brings about breakthroughs in the study of EMR data. The challenge is that deep learning lacks semantic knowledge, which leads to impracticability in medical science. This research proposes a method of incorporating lexical-semantic knowledge from abundant entities into a convolutional neural network (CNN) framework for pediatric disease diagnosis. Firstly, medical terms are vectorized into Lexical Semantic Vectors (LSV), which are concatenated with the embedded word vectors of word2vec to enrich the feature representation. Secondly, the semantic distribution of medical terms serves as Semantic Decision Guide (SDG) for the optimization of deep learning models. The study evaluate the performance of LSV-SDG-CNN model on four kinds of Chinese EMR datasets. Additionally, CNN, LSV-CNN, and SDG-CNN are designed as baseline models for comparison. The experimental results show that LSV-SDG-CNN model outperforms baseline models on four kinds of Chinese EMR datasets. The best configuration of the model yielded an F1 score of 86.20%. The results clearly demonstrate that CNN has been effectively guided and optimized by lexical-semantic knowledge, and LSV-SDG-CNN model improves the disease classification accuracy with a clear margin.

Keywords: convolutional neural network, electronic medical record, feature representation, lexical semantics, semantic decision

Procedia PDF Downloads 120
3614 Network Conditioning and Transfer Learning for Peripheral Nerve Segmentation in Ultrasound Images

Authors: Harold Mauricio Díaz-Vargas, Cristian Alfonso Jimenez-Castaño, David Augusto Cárdenas-Peña, Guillermo Alberto Ortiz-Gómez, Alvaro Angel Orozco-Gutierrez

Abstract:

Precise identification of the nerves is a crucial task performed by anesthesiologists for an effective Peripheral Nerve Blocking (PNB). Now, anesthesiologists use ultrasound imaging equipment to guide the PNB and detect nervous structures. However, visual identification of the nerves from ultrasound images is difficult, even for trained specialists, due to artifacts and low contrast. The recent advances in deep learning make neural networks a potential tool for accurate nerve segmentation systems, so addressing the above issues from raw data. The most widely spread U-Net network yields pixel-by-pixel segmentation by encoding the input image and decoding the attained feature vector into a semantic image. This work proposes a conditioning approach and encoder pre-training to enhance the nerve segmentation of traditional U-Nets. Conditioning is achieved by the one-hot encoding of the kind of target nerve a the network input, while the pre-training considers five well-known deep networks for image classification. The proposed approach is tested in a collection of 619 US images, where the best C-UNet architecture yields an 81% Dice coefficient, outperforming the 74% of the best traditional U-Net. Results prove that pre-trained models with the conditional approach outperform their equivalent baseline by supporting learning new features and enriching the discriminant capability of the tested networks.

Keywords: nerve segmentation, U-Net, deep learning, ultrasound imaging, peripheral nerve blocking

Procedia PDF Downloads 91
3613 Power Aware Modified I-LEACH Protocol Using Fuzzy IF Then Rules

Authors: Gagandeep Singh, Navdeep Singh

Abstract:

Due to limited battery of sensor nodes, so energy efficiency found to be main constraint in WSN. Therefore the main focus of the present work is to find the ways to minimize the energy consumption problem and will results; enhancement in the network stability period and life time. Many researchers have proposed different kind of the protocols to enhance the network lifetime further. This paper has evaluated the issues which have been neglected in the field of the WSNs. WSNs are composed of multiple unattended ultra-small, limited-power sensor nodes. Sensor nodes are deployed randomly in the area of interest. Sensor nodes have limited processing, wireless communication and power resource capabilities Sensor nodes send sensed data to sink or Base Station (BS). I-LEACH gives adaptive clustering mechanism which very efficiently deals with energy conservations. This paper ends up with the shortcomings of various adaptive clustering based WSNs protocols.

Keywords: WSN, I-Leach, MATLAB, sensor

Procedia PDF Downloads 262