Search results for: graph attention neural network
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9485

Search results for: graph attention neural network

7415 Enhancement of Environmental Security by the Application of Wireless Sensor Network in Nigeria

Authors: Ahmadu Girgiri, Lawan Gana Ali, Mamman M. Baba

Abstract:

Environmental security clearly articulates the perfections and developments of various communities around the world irrespective of the region, culture, religion or social inclination. Although, the present state of insecurity has become serious issue devastating the peace, unity, stability and progress of man and his physical environment particularly in developing countries. Recently, measure of security and it management in Nigeria has been a bottle-neck to the effectiveness and advancement of various sectors that include; business, education, social relations, politics and above all an economy. Several measures have been considered on mitigating environment insecurity such as surveillance, demarcation, security personnel empowerment and the likes, but still the issue remains disturbing. In this paper, we present the application of new technology that contributes to the improvement of security surveillance known as “Wireless Sensor Network (WSN)”. The system is new, smart and emerging technology that provides monitoring, detection and aggregation of information using sensor nodes and wireless network. WSN detects, monitors and stores information or activities in the deployed area such as schools, environment, business centers, public squares, industries, and outskirts and transmit to end users. This will reduce the cost of security funding and eases security surveillance depending on the nature and the requirement of the deployment.

Keywords: application, environment, insecurity, sensor, wireless sensor network

Procedia PDF Downloads 263
7414 Performance Evaluation of Routing Protocols for Video Conference over MPLS VPN Network

Authors: Abdullah Al Mamun, Tarek R. Sheltami

Abstract:

Video conferencing is a highly demanding facility now a days in order to its real time characteristics, but faster communication is the prior requirement of this technology. Multi Protocol Label Switching (MPLS) IP Virtual Private Network (VPN) address this problem and it is able to make a communication faster than others techniques. However, this paper studies the performance comparison of video traffic between two routing protocols namely the Enhanced Interior Gateway Protocol(EIGRP) and Open Shortest Path First (OSPF). The combination of traditional routing and MPLS improve the forwarding mechanism, scalability and overall network performance. We will use GNS3 and OPNET Modeler 14.5 to simulate many different scenarios and metrics such as delay, jitter and mean opinion score (MOS) value are measured. The simulation result will show that OSPF and BGP-MPLS VPN offers best performance for video conferencing application.

Keywords: OSPF, BGP, EIGRP, MPLS, Video conference, Provider router, edge router, layer3 VPN

Procedia PDF Downloads 331
7413 STATCOM’s Contribution to the Improvement of Voltage Plan and Power Flow in an Electrical Transmission Network

Authors: M. Adjabi, A. Amiar, P. O. Logerais

Abstract:

Flexible Alternative Current Systems Transmission (FACTS) are used since nearly four decades and present very good dynamic performances. The purpose of this work is to study the behavior of a system where Static Compensator (STATCOM) is located at the midpoint of a transmission line which is the idea of the project functioning in disturbed modes with various levels of load. The studied model and starting from the analysis of various alternatives will lead to the checking of the aptitude of the STATCOM to maintain the voltage plan and to improve the power flow in electro-energetic system which is the east region of Algerian 400 kV transmission network. The steady state performance of STATCOM’s controller is analyzed through computer simulations with Matlab/Simulink program. The simulation results have demonstrated that STATCOM can be effectively applied in power transmission systems to solve the problems of poor dynamic performance and voltage regulation.

Keywords: STATCOM, reactive power, power flow, voltage plan, Algerian network

Procedia PDF Downloads 569
7412 STATCOM's Contribution to the Improvement of Voltage Plan and Power Flow in an Electrical Transmission Network

Authors: M. Adjabi, A. Amiar, P. O. Logerais

Abstract:

Flexible Alternative Current Systems Transmission (FACTS) are used since nearly four decades and present very good dynamic performances. The purpose of this work is to study the behavior of a system where Static Compensator (STATCOM) is located at the midpoint of a transmission line which is the idea of the project functioning in disturbed modes with various levels of load. The studied model and starting from the analysis of various alternatives will lead to the checking of the aptitude of the STATCOM to maintain the voltage plan and to improve the power flow in electro-energetic system which is the east region of Algerian 400 kV transmission network. The steady state performance of STATCOM’s controller is analyzed through computer simulations with Matlab/Simulink program. The simulation results have demonstrated that STATCOM can be effectively applied in power transmission systems to solve the problems of poor dynamic performance and voltage regulation.

Keywords: STATCOM, reactive power, power flow, voltage plan, Algerian network

Procedia PDF Downloads 600
7411 Wireless Sensor Anomaly Detection Using Soft Computing

Authors: Mouhammd Alkasassbeh, Alaa Lasasmeh

Abstract:

We live in an era of rapid development as a result of significant scientific growth. Like other technologies, wireless sensor networks (WSNs) are playing one of the main roles. Based on WSNs, ZigBee adds many features to devices, such as minimum cost and power consumption, and increasing the range and connect ability of sensor nodes. ZigBee technology has come to be used in various fields, including science, engineering, and networks, and even in medicinal aspects of intelligence building. In this work, we generated two main datasets, the first being based on tree topology and the second on star topology. The datasets were evaluated by three machine learning (ML) algorithms: J48, meta.j48 and multilayer perceptron (MLP). Each topology was classified into normal and abnormal (attack) network traffic. The dataset used in our work contained simulated data from network simulation 2 (NS2). In each database, the Bayesian network meta.j48 classifier achieved the highest accuracy level among other classifiers, of 99.7% and 99.2% respectively.

Keywords: IDS, Machine learning, WSN, ZigBee technology

Procedia PDF Downloads 543
7410 Light, Restorativeness and Performance in the Workplace: A Pilot Study

Authors: D. Scarpanti, M. Brondino, M. Pasini

Abstract:

Background: the present study explores the role of light and restorativeness on work. According with the Attention Restoration Theory (ART) and a Model of Work Environment, the main idea is that some features of environment, i.e., lighting, influences the direct attention, and so, the performance. Restorativeness refers to the presence/absence level of all the characteristics of physical environment that help to regenerate direct attention. Specifically, lighting can affect level of fascination and attention in one hand; and in other hand promotes several biological functions via pineal gland. Different reviews on this topic show controversial results. In order to bring light on this topic, the hypotheses of this study are that lighting can affect the construct of restorativeness and, in the second time, the restorativeness can affect the performance. Method: the participants are 30 workers of a mechatronic company in the North Italy. Every subject answered to a questionnaire valuing their subjective perceptions of environment in a different way: some objective features of environment, like lighting, temperature and air quality; some subjective perceptions of this environment; finally, the participants answered about their perceived performance. The main attention is on the features of light and his components: visual comfort, general preferences and pleasantness; and the dimensions of the construct of restorativeness; fascination, coherence and being away. The construct of performance per se is conceptualized in three level: individual, team membership and organizational membership; and in three different components: proficiency, adaptability, and proactivity, for a total of 9 subcomponents. Findings: path analysis showed that some characteristics of lighting respectively affected the dimension of fascination; and, as expected, the dimension of fascination affected work performance. Conclusions: The present study is a first pilot step of a wide research. These first results can be summarized with the statement that lighting and restorativeness contribute to explain work performance variability: in details perceptions of visual comfort, satisfaction and pleasantness, and fascination respectively. Results related to fascination are particularly interesting because fascination is conceptualized as the opposite of the construct of direct attention. The main idea is, in order to regenerate attentional capacity, it’s necessary to provide a lacking of attention (fascination). The sample size did not permit to test simultaneously the role of the perceived characteristics of light to see how they differently contribute to predict fascination of the work environment. However, the results highlighted the important role that light could have in predicting restorativeness dimensions and probably with a larger sample we could find larger effects also on work performance. Furthermore, longitudinal data will contribute to better analyze the causal model along time. Applicative implications: the present pilot study highlights the relevant role of lighting and perceived restorativeness in the work environment and the importance to focus attention on light features and the restorative characteristics in the design of work environments.

Keywords: lighting, performance, restorativeness, workplace

Procedia PDF Downloads 154
7409 Detecting and Thwarting Interest Flooding Attack in Information Centric Network

Authors: Vimala Rani P, Narasimha Malikarjunan, Mercy Shalinie S

Abstract:

Data Networking was brought forth as an instantiation of information-centric networking. The attackers can send a colossal number of spoofs to take hold of the Pending Interest Table (PIT) named an Interest Flooding attack (IFA) since the in- interests are recorded in the PITs of the intermediate routers until they receive corresponding Data Packets are go beyond the time limit. These attacks can be detrimental to network performance. PIT expiration rate or the Interest satisfaction rate, which cannot differentiate the IFA from attacks, is the criterion Traditional IFA detection techniques are concerned with. Threshold values can casually affect Threshold-based traditional methods. This article proposes an accurate IFA detection mechanism based on a Multiple Feature-based Extreme Learning Machine (MF-ELM). Accuracy of the attack detection can be increased by presenting the entropy of Internet names, Interest satisfaction rate and PIT usage as features extracted in the MF-ELM classifier. Furthermore, we deploy a queue-based hostile Interest prefix mitigation mechanism. The inference of this real-time test bed is that the mechanism can help the network to resist IFA with higher accuracy and efficiency.

Keywords: information-centric network, pending interest table, interest flooding attack, MF-ELM classifier, queue-based mitigation strategy

Procedia PDF Downloads 205
7408 Using the Weakest Precondition to Achieve Self-Stabilization in Critical Networks

Authors: Antonio Pizzarello, Oris Friesen

Abstract:

Networks, such as the electric power grid, must demonstrate exemplary performance and integrity. Integrity depends on the quality of both the system design model and the deployed software. Integrity of the deployed software is key, for both the original versions and the many that occur throughout numerous maintenance activity. Current software engineering technology and practice do not produce adequate integrity. Distributed systems utilize networks where each node is an independent computer system. The connections between them is realized via a network that is normally redundantly connected to guarantee the presence of a path between two nodes in the case of failure of some branch. Furthermore, at each node, there is software which may fail. Self-stabilizing protocols are usually present that recognize failure in the network and perform a repair action that will bring the node back to a correct state. These protocols first introduced by E. W. Dijkstra are currently present in almost all Ethernets. Super stabilization protocols capable of reacting to a change in the network topology due to the removal or addition of a branch in the network are less common but are theoretically defined and available. This paper describes how to use the Software Integrity Assessment (SIA) methodology to analyze self-stabilizing software. SIA is based on the UNITY formalism for parallel and distributed programming, which allows the analysis of code for verifying the progress property p leads-to q that describes the progress of all computations starting in a state satisfying p to a state satisfying q via the execution of one or more system modules. As opposed to demonstrably inadequate test and evaluation methods SIA allows the analysis and verification of any network self-stabilizing software as well as any other software that is designed to recover from failure without external intervention of maintenance personnel. The model to be analyzed is obtained by automatic translation of the system code to a transition system that is based on the use of the weakest precondition.

Keywords: network, power grid, self-stabilization, software integrity assessment, UNITY, weakest precondition

Procedia PDF Downloads 223
7407 Evaluation of Routing Protocols in Mobile Adhoc Networks

Authors: Anu Malhotra

Abstract:

An Ad-hoc network is one that is an autonomous, self configuring network made up of mobile nodes connected via wireless links. Ad-hoc networks often consist of nodes, mobile hosts (MH) or mobile stations (MS, also serving as routers) connected by wireless links. Different routing protocols are used for data transmission in between the nodes in an adhoc network. In this paper two protocols (OLSR and AODV) are analyzed on the basis of two parameters i.e. time delay and throughput with different data rates. On the basis of these analysis, we observed that with same data rate, AODV protocol is having more time delay than the OLSR protocol whereas throughput for the OLSR protocol is less compared to the AODV protocol.

Keywords: routing adhoc, mobile hosts, mobile stations, OLSR protocol, AODV protocol

Procedia PDF Downloads 506
7406 Influence Analysis of Macroeconomic Parameters on Real Estate Price Variation in Taipei, Taiwan

Authors: Li Li, Kai-Hsuan Chu

Abstract:

It is well known that the real estate price depends on a lot of factors. Each house current value is dependent on the location, room number, transportation, living convenience, year and surrounding environments. Although, there are different experienced models for housing agent to estimate the price, it is a case by case study without overall dynamic variation investigation. However, many economic parameters may more or less influence the real estate price variation. Here, the influences of most macroeconomic parameters on real estate price are investigated individually based on least-square scheme and grey correlation strategy. Then those parameters are classified into leading indices, simultaneous indices and laggard indices. In addition, the leading time period is evaluated based on least square method. The important leading and simultaneous indices can be used to establish an artificial intelligent neural network model for real estate price variation prediction. The real estate price variation of Taipei, Taiwan during 2005 ~ 2017 are chosen for this research data analysis and validation. The results show that the proposed method has reasonable prediction function for real estate business reference.

Keywords: real estate price, least-square, grey correlation, macroeconomics

Procedia PDF Downloads 198
7405 Academic Influence of Social Network Sites on the Collegiate Performance of Technical College Students

Authors: Jameson McFarlane, Thorne J. McFarlane, Leon Bernard

Abstract:

Social network sites (SNS) is an emerging phenomenon that is here to stay. The popularity and the ubiquity of the SNS technology are undeniable. Because most SNS are free and easy to use people from all walks of life and from almost any age are attracted to that technology. College age students are by far the largest segment of the population using SNS. Since most SNS have been adapted for mobile devices, not only do you find students using this technology in their study, while working on labs or on projects, a substantial number of students have been found to use SNS even while listening to lectures. This study found that SNS use has a significant negative impact on the grade point average of college students particularly in the first semester. However, this negative impact is greatly diminished by the end of the third semester partly because the students have adjusted satisfactorily to the challenges of college or because they have learned how to adequately manage their time. It was established that the kinds of activities the students are engaged in during the SNS use are the leading factor affecting academic performance. Of those activities, using SNS during a lecture or while studying is the foremost contributing factor to lower academic performance. This is due to “cognitive” or “information” bottleneck, a condition in which the students find it very difficult to multitask or to switch between resources leading to inefficiency in information retention and thus, educational performance.

Keywords: social network sites, social network analysis, regression coefficient, psychological engagement

Procedia PDF Downloads 179
7404 Subjective Time as a Marker of the Present Consciousness

Authors: Anastasiya Paltarzhitskaya

Abstract:

Subjective time plays an important role in consciousness processes and self-awareness at the moment. The concept of intrinsic neural timescales (INT) explains the difference in perceiving various time intervals. The capacity to experience the present builds on the fundamental properties of temporal cognition. The challenge that both philosophy and neuroscience try to answer is how the brain differentiates the present from the past and future. In our work, we analyze papers which describe mechanisms involved in the perception of ‘present’ and ‘non-present’, i.e., future and past moments. Taking into account that we perceive time intervals even during rest or relaxation, we suppose that the default-mode network activity can code time features, including the present moment. We can compare some results of time perceptual studies, where brain activity was shown in states with different flows of time, including resting states and during “mental time travel”. According to the concept of mental traveling, we employ a range of scenarios which demand episodic memory. However, some papers show that the hippocampal region does not activate during time traveling. It is a controversial result that is further complicated by the phenomenological aspect that includes a holistic set of information about the individual’s past and future.

Keywords: temporal consciousness, time perception, memory, present

Procedia PDF Downloads 76
7403 VANETs Geographic Routing Protocols: A survey

Authors: Ramin Karimi

Abstract:

One of common highly mobile wireless ad hoc networks is Vehicular Ad Hoc Networks. Hence routing in vehicular ad hoc network (VANET) has attracted much attention during the last few years. VANET is characterized by its high mobility of nodes and specific topology patterns. Moreover these networks encounter a significant loss rate and a very short duration of communication. In vehicular ad hoc networks, one of challenging is routing of data due to high speed mobility and changing topology of vehicles. Geographic routing protocols are becoming popular due to advancement and availability of GPS devices. Delay Tolerant Networks (DTNs) are a class of networks that enable communication where connectivity issues like sparse connectivity, intermittent connectivity; high latency, long delay, high error rates, asymmetric data rate, and even no end-to-end connectivity exist. In this paper, we review the existing Geographic Routing Protocols for VANETs and also provide a qualitative comparison of them.

Keywords: vehicular ad hoc networks, mobility, geographic routing, delay tolerant networks

Procedia PDF Downloads 520
7402 Using Social Network Analysis for Cyber Threat Intelligence

Authors: Vasileios Anastopoulos

Abstract:

Cyber threat intelligence assists organizations in understanding the threats they face and helps them make educated decisions on preparing their defenses. Sharing of threat intelligence and threat information is increasingly leveraged by organizations and enterprises, and various software solutions are already available, with the open-source malware information sharing platform (MISP) being a popular one. In this work, a methodology for the production of cyber threat intelligence using the threat information stored in MISP is proposed. The methodology leverages the discipline of social network analysis and the diamond model, a model used for intrusion analysis, to produce cyber threat intelligence. The workings are demonstrated with a case study on a production MISP instance of a real organization. The paper concluded with a discussion on the proposed methodology and possible directions for further research.

Keywords: cyber threat intelligence, diamond model, malware information sharing platform, social network analysis

Procedia PDF Downloads 178
7401 Effect of Variable Fluxes on Optimal Flux Distribution in a Metabolic Network

Authors: Ehsan Motamedian

Abstract:

Finding all optimal flux distributions of a metabolic model is an important challenge in systems biology. In this paper, a new algorithm is introduced to identify all alternate optimal solutions of a large scale metabolic network. The algorithm reduces the model to decrease computations for finding optimal solutions. The algorithm was implemented on the Escherichia coli metabolic model to find all optimal solutions for lactate and acetate production. There were more optimal flux distributions when acetate production was optimized. The model was reduced from 1076 to 80 variable fluxes for lactate while it was reduced to 91 variable fluxes for acetate. These 11 more variable fluxes resulted in about three times more optimal flux distributions. Variable fluxes were from 12 various metabolic pathways and most of them belonged to nucleotide salvage and extra cellular transport pathways.

Keywords: flux variability, metabolic network, mixed-integer linear programming, multiple optimal solutions

Procedia PDF Downloads 434
7400 Social Networks in Business: The Complex Concept of Wasta and the Impact of Islam on the Perception of This Practice

Authors: Sa'ad Ali

Abstract:

This study explores wasta as an example of a social network and how it impacts business practice in the Arab Middle East, drawing links with social network impact in different regions of the world. In doing so, particular attention will be paid to the socio-economic and cultural influences on business practice. In exploring relationships in business, concepts such as social network analysis, social capital and group identity are used to explore the different forms of social networks and how they influence business decisions and practices in the regions and countries where they prevail. The use of social networks to achieve objectives is known as guanxi in China, wasta in the Arab Middle East and blat in ex-Soviet countries. Wasta can be defined as favouritism based on tribal and family affiliation and is a widespread practice that has a substantial impact on political, social and business interactions in the Arab Middle East. Within the business context, it is used in several ways, such as to secure a job or promotion or to cut through bureaucracy in government interactions. The little research available is fragmented, and most studies reveal a negative attitude towards its usage in business. Paradoxically, while wasta is widely practised, people from the Arab Middle East often deny its influence. Moreover, despite the regular exhibition of a negative opinion on the practice of wasta, it can also be a source of great pride. This paper addresses this paradox by conducting a positional literature review, exploring the current literature on wasta and identifying how the identified paradox can be explained. The findings highlight how wasta, to a large extent, has been treated as an umbrella concept, whilst it is a highly complex practice which has evolved from intermediary wasta to intercessory wasta and therefore from bonding social capital relationships to more bridging social capital relationships. In addition, the research found that Islam, as the predominant religion in the region and the main source of ethical guidance for the majority of people from the region, plays a substantial role in this paradox. Specifically, it is submitted that wasta can be viewed positively in Islam when it is practised to aid others without breaking Islamic ethical guidelines, whilst it can be viewed negatively when it is used in contradiction with the teachings of Islam. As such, the unique contribution to knowledge of this study is that it ties together the fragmented literature on wasta, highlighting and helping us understand its complexity. In addition, it sheds light on the role of Islam in wasta practices, aiding our understanding of the paradoxical nature of the practice.

Keywords: Islamic ethics, social capital, social networks, Wasta

Procedia PDF Downloads 146
7399 On the Limits of Board Diversity: Impact of Network Effect on Director Appointments

Authors: Vijay Marisetty, Poonam Singh

Abstract:

Research on the effect of director's network connections on investor welfare is inconclusive. Some studies suggest that directors' connections are beneficial, in terms of, improving earnings information, firms valuation for new investors. On the other hand, adverse effects of directorial networks are also reported, in terms of higher earnings management, options back dating fraud, reduction in firm performance, lower board monitoring. From regulatory perspective, the role of directorial networks on corporate welfare is crucial. Cognizant of the possible ill effects associated with directorial networks, large investors, for better representation on the boards, are building their own database of prospective directors who are highly qualified, however, sourced from outside the highly connected directorial labor market. For instance, following Dodd-Frank Reform Act, California Public Employees' Retirement Systems (CalPERs) has initiated a database for registering aspiring and highly qualified directors to nominate them for board seats (proxy access). Our paper stems from this background and tries to explore the chances of outside directors getting directorships who lack established network connections. The paper is able to identify such aspiring directors' information by accessing a unique Indian data sourced from an online portal that aims to match the supply of registered aspirants with the growing demand for outside directors in India. The online portal's tie-up with stock exchanges ensures firms to access the new pool of directors. Such direct access to the background details of aspiring directors over a period of 10 years, allows us to examine the chances of aspiring directors without corporate network, to enter directorial network. Using this resume data of 16105 aspiring corporate directors in India, who have no prior board experience in the directorial labor market, the paper analyses the entry dynamics in corporate directors' labor market. The database also allows us to investigate the value of corporate network by comparing non-network new entrants with incumbent networked directors. The study develops measures of network centrality and network degree based on merit, i.e. network of individuals belonging to elite educational institutions, like Indian Institute of Management (IIM) or Indian Institute of Technology (IIT) and based on job or company, i.e. network of individuals serving in the same company. The paper then measures the impact of these networks on the appointment of first time directors and subsequent appointment of directors. The paper reports the following main results: 1. The likelihood of becoming a corporate director, without corporate network strength, is only 1 out 100 aspirants. This is inspite of comparable educational background and similar duration of corporate experience; 2. Aspiring non-network directors' elite educational ties help them to secure directorships. However, for post-board appointments, their newly acquired corporate network strength overtakes as their main determinant for subsequent board appointments and compensation. The results thus highlight the limitations in increasing board diversity.

Keywords: aspiring corporate directors, board diversity, director labor market, director networks

Procedia PDF Downloads 312
7398 High-Capacity Image Steganography using Wavelet-based Fusion on Deep Convolutional Neural Networks

Authors: Amal Khalifa, Nicolas Vana Santos

Abstract:

Steganography has been known for centuries as an efficient approach for covert communication. Due to its popularity and ease of access, image steganography has attracted researchers to find secure techniques for hiding information within an innocent looking cover image. In this research, we propose a novel deep-learning approach to digital image steganography. The proposed method, DeepWaveletFusion, uses convolutional neural networks (CNN) to hide a secret image into a cover image of the same size. Two CNNs are trained back-to-back to merge the Discrete Wavelet Transform (DWT) of both colored images and eventually be able to blindly extract the hidden image. Based on two different image similarity metrics, a weighted gain function is used to guide the learning process and maximize the quality of the retrieved secret image and yet maintaining acceptable imperceptibility. Experimental results verified the high recoverability of DeepWaveletFusion which outperformed similar deep-learning-based methods.

Keywords: deep learning, steganography, image, discrete wavelet transform, fusion

Procedia PDF Downloads 90
7397 Performance Analysis of N-Tier Grid Protocol for Resource Constrained Wireless Sensor Networks

Authors: Jai Prakash Prasad, Suresh Chandra Mohan

Abstract:

Modern wireless sensor networks (WSN) consist of small size, low cost devices which are networked through tight wireless communications. WSN fundamentally offers cooperation, coordination among sensor networks. Potential applications of wireless sensor networks are in healthcare, natural disaster prediction, data security, environmental monitoring, home appliances, entertainment etc. The design, development and deployment of WSN based on application requirements. The WSN design performance is optimized to improve network lifetime. The sensor node resources constrain such as energy and bandwidth imposes the limitation on efficient resource utilization and sensor node management. The proposed N-Tier GRID routing protocol focuses on the design of energy efficient large scale wireless sensor network for improved performance than the existing protocol.

Keywords: energy efficient, network lifetime, sensor networks, wireless communication

Procedia PDF Downloads 469
7396 Improving Subjective Bias Detection Using Bidirectional Encoder Representations from Transformers and Bidirectional Long Short-Term Memory

Authors: Ebipatei Victoria Tunyan, T. A. Cao, Cheol Young Ock

Abstract:

Detecting subjectively biased statements is a vital task. This is because this kind of bias, when present in the text or other forms of information dissemination media such as news, social media, scientific texts, and encyclopedias, can weaken trust in the information and stir conflicts amongst consumers. Subjective bias detection is also critical for many Natural Language Processing (NLP) tasks like sentiment analysis, opinion identification, and bias neutralization. Having a system that can adequately detect subjectivity in text will boost research in the above-mentioned areas significantly. It can also come in handy for platforms like Wikipedia, where the use of neutral language is of importance. The goal of this work is to identify the subjectively biased language in text on a sentence level. With machine learning, we can solve complex AI problems, making it a good fit for the problem of subjective bias detection. A key step in this approach is to train a classifier based on BERT (Bidirectional Encoder Representations from Transformers) as upstream model. BERT by itself can be used as a classifier; however, in this study, we use BERT as data preprocessor as well as an embedding generator for a Bi-LSTM (Bidirectional Long Short-Term Memory) network incorporated with attention mechanism. This approach produces a deeper and better classifier. We evaluate the effectiveness of our model using the Wiki Neutrality Corpus (WNC), which was compiled from Wikipedia edits that removed various biased instances from sentences as a benchmark dataset, with which we also compare our model to existing approaches. Experimental analysis indicates an improved performance, as our model achieved state-of-the-art accuracy in detecting subjective bias. This study focuses on the English language, but the model can be fine-tuned to accommodate other languages.

Keywords: subjective bias detection, machine learning, BERT–BiLSTM–Attention, text classification, natural language processing

Procedia PDF Downloads 130
7395 Optimizing Production Yield Through Process Parameter Tuning Using Deep Learning Models: A Case Study in Precision Manufacturing

Authors: Tolulope Aremu

Abstract:

This paper is based on the idea of using deep learning methodology for optimizing production yield by tuning a few key process parameters in a manufacturing environment. The study was explicitly on how to maximize production yield and minimize operational costs by utilizing advanced neural network models, specifically Long Short-Term Memory and Convolutional Neural Networks. These models were implemented using Python-based frameworks—TensorFlow and Keras. The targets of the research are the precision molding processes in which temperature ranges between 150°C and 220°C, the pressure ranges between 5 and 15 bar, and the material flow rate ranges between 10 and 50 kg/h, which are critical parameters that have a great effect on yield. A dataset of 1 million production cycles has been considered for five continuous years, where detailed logs are present showing the exact setting of parameters and yield output. The LSTM model would model time-dependent trends in production data, while CNN analyzed the spatial correlations between parameters. Models are designed in a supervised learning manner. For the model's loss, an MSE loss function is used, optimized through the Adam optimizer. After running a total of 100 training epochs, 95% accuracy was achieved by the models recommending optimal parameter configurations. Results indicated that with the use of RSM and DOE traditional methods, there was an increase in production yield of 12%. Besides, the error margin was reduced by 8%, hence consistent quality products from the deep learning models. The monetary value was annually around $2.5 million, the cost saved from material waste, energy consumption, and equipment wear resulting from the implementation of optimized process parameters. This system was deployed in an industrial production environment with the help of a hybrid cloud system: Microsoft Azure, for data storage, and the training and deployment of their models were performed on Google Cloud AI. The functionality of real-time monitoring of the process and automatic tuning of parameters depends on cloud infrastructure. To put it into perspective, deep learning models, especially those employing LSTM and CNN, optimize the production yield by fine-tuning process parameters. Future research will consider reinforcement learning with a view to achieving further enhancement of system autonomy and scalability across various manufacturing sectors.

Keywords: production yield optimization, deep learning, tuning of process parameters, LSTM, CNN, precision manufacturing, TensorFlow, Keras, cloud infrastructure, cost saving

Procedia PDF Downloads 31
7394 Predicting Open Chromatin Regions in Cell-Free DNA Whole Genome Sequencing Data by Correlation Clustering  

Authors: Fahimeh Palizban, Farshad Noravesh, Amir Hossein Saeidian, Mahya Mehrmohamadi

Abstract:

In the recent decade, the emergence of liquid biopsy has significantly improved cancer monitoring and detection. Dying cells, including those originating from tumors, shed their DNA into the blood and contribute to a pool of circulating fragments called cell-free DNA. Accordingly, identifying the tissue origin of these DNA fragments from the plasma can result in more accurate and fast disease diagnosis and precise treatment protocols. Open chromatin regions are important epigenetic features of DNA that reflect cell types of origin. Profiling these features by DNase-seq, ATAC-seq, and histone ChIP-seq provides insights into tissue-specific and disease-specific regulatory mechanisms. There have been several studies in the area of cancer liquid biopsy that integrate distinct genomic and epigenomic features for early cancer detection along with tissue of origin detection. However, multimodal analysis requires several types of experiments to cover the genomic and epigenomic aspects of a single sample, which will lead to a huge amount of cost and time. To overcome these limitations, the idea of predicting OCRs from WGS is of particular importance. In this regard, we proposed a computational approach to target the prediction of open chromatin regions as an important epigenetic feature from cell-free DNA whole genome sequence data. To fulfill this objective, local sequencing depth will be fed to our proposed algorithm and the prediction of the most probable open chromatin regions from whole genome sequencing data can be carried out. Our method integrates the signal processing method with sequencing depth data and includes count normalization, Discrete Fourie Transform conversion, graph construction, graph cut optimization by linear programming, and clustering. To validate the proposed method, we compared the output of the clustering (open chromatin region+, open chromatin region-) with previously validated open chromatin regions related to human blood samples of the ATAC-DB database. The percentage of overlap between predicted open chromatin regions and the experimentally validated regions obtained by ATAC-seq in ATAC-DB is greater than 67%, which indicates meaningful prediction. As it is evident, OCRs are mostly located in the transcription start sites (TSS) of the genes. In this regard, we compared the concordance between the predicted OCRs and the human genes TSS regions obtained from refTSS and it showed proper accordance around 52.04% and ~78% with all and the housekeeping genes, respectively. Accurately detecting open chromatin regions from plasma cell-free DNA-seq data is a very challenging computational problem due to the existence of several confounding factors, such as technical and biological variations. Although this approach is in its infancy, there has already been an attempt to apply it, which leads to a tool named OCRDetector with some restrictions like the need for highly depth cfDNA WGS data, prior information about OCRs distribution, and considering multiple features. However, we implemented a graph signal clustering based on a single depth feature in an unsupervised learning manner that resulted in faster performance and decent accuracy. Overall, we tried to investigate the epigenomic pattern of a cell-free DNA sample from a new computational perspective that can be used along with other tools to investigate genetic and epigenetic aspects of a single whole genome sequencing data for efficient liquid biopsy-related analysis.

Keywords: open chromatin regions, cancer, cell-free DNA, epigenomics, graph signal processing, correlation clustering

Procedia PDF Downloads 150
7393 Hormone Replacement Therapy (HRT) and Its Impact on the All-Cause Mortality of UK Women: A Matched Cohort Study 1984-2017

Authors: Nurunnahar Akter, Elena Kulinskaya, Nicholas Steel, Ilyas Bakbergenuly

Abstract:

Although Hormone Replacement Therapy (HRT) is an effective treatment in ameliorating menopausal symptoms, it has mixed effects on different health outcomes, increasing, for instance, the risk of breast cancer. Because of this, many symptomatic women are left untreated. Untreated menopausal symptoms may result in other health issues, which eventually put an extra burden and costs to the health care system. All-cause mortality analysis may explain the net benefits and risks of the HRT therapy. However, it received far less attention in HRT studies. This study investigated the impact of HRT on all-cause mortality using electronically recorded primary care data from The Health Improvement Network (THIN) that broadly represents the female population in the United Kingdom (UK). The study entry date for this study was the record of the first HRT prescription from 1984, and patients were followed up until death or transfer to another GP practice or study end date, which was January 2017. 112,354 HRT users (cases) were matched with 245,320 non-users by age at HRT initiation and general practice (GP). The hazards of all-cause mortality associated with HRT were estimated by a parametric Weibull-Cox model adjusting for a wide range of important medical, lifestyle, and socio-demographic factors. The multilevel multiple imputation techniques were used to deal with missing data. This study found that during 32 years of follow-up, combined HRT reduced the hazard ratio (HR) of all-cause mortality by 9% (HR: 0.91; 95% Confidence Interval, 0.88-0.94) in women of age between 46 to 65 at first treatment compared to the non-users of the same age. Age-specific mortality analyses found that combined HRT decreased mortality by 13% (HR: 0.87; 95% CI, 0.82-0.92), 12% (HR: 0.88; 95% CI, 0.82-0.93), and 8% (HR: 0.92; 95% CI, 0.85-0.98), in 51 to 55, 56 to 60, and 61 to 65 age group at first treatment, respectively. There was no association between estrogen-only HRT and women’s all-cause mortality. The findings from this study may help to inform the choices of women at menopause and to further educate the clinicians and resource planners.

Keywords: hormone replacement therapy, multiple imputations, primary care data, the health improvement network (THIN)

Procedia PDF Downloads 170
7392 Computational Characterization of Electronic Charge Transfer in Interfacial Phospholipid-Water Layers

Authors: Samira Baghbanbari, A. B. P. Lever, Payam S. Shabestari, Donald Weaver

Abstract:

Existing signal transmission models, although undoubtedly useful, have proven insufficient to explain the full complexity of information transfer within the central nervous system. The development of transformative models will necessitate a more comprehensive understanding of neuronal lipid membrane electrophysiology. Pursuant to this goal, the role of highly organized interfacial phospholipid-water layers emerges as a promising case study. A series of phospholipids in neural-glial gap junction interfaces as well as cholesterol molecules have been computationally modelled using high-performance density functional theory (DFT) calculations. Subsequent 'charge decomposition analysis' calculations have revealed a net transfer of charge from phospholipid orbitals through the organized interfacial water layer before ultimately finding its way to cholesterol acceptor molecules. The specific pathway of charge transfer from phospholipid via water layers towards cholesterol has been mapped in detail. Cholesterol is an essential membrane component that is overrepresented in neuronal membranes as compared to other mammalian cells; given this relative abundance, its apparent role as an electronic acceptor may prove to be a relevant factor in further signal transmission studies of the central nervous system. The timescales over which this electronic charge transfer occurs have also been evaluated by utilizing a system design that systematically increases the number of water molecules separating lipids and cholesterol. Memory loss through hydrogen-bonded networks in water can occur at femtosecond timescales, whereas existing action potential-based models are limited to micro or nanosecond scales. As such, the development of future models that attempt to explain faster timescale signal transmission in the central nervous system may benefit from our work, which provides additional information regarding fast timescale energy transfer mechanisms occurring through interfacial water. The study possesses a dataset that includes six distinct phospholipids and a collection of cholesterol. Ten optimized geometric characteristics (features) were employed to conduct binary classification through an artificial neural network (ANN), differentiating cholesterol from the various phospholipids. This stems from our understanding that all lipids within the first group function as electronic charge donors, while cholesterol serves as an electronic charge acceptor.

Keywords: charge transfer, signal transmission, phospholipids, water layers, ANN

Procedia PDF Downloads 72
7391 Ensemble of Deep CNN Architecture for Classifying the Source and Quality of Teff Cereal

Authors: Belayneh Matebie, Michael Melese

Abstract:

The study focuses on addressing the challenges in classifying and ensuring the quality of Eragrostis Teff, a small and round grain that is the smallest cereal grain. Employing a traditional classification method is challenging because of its small size and the similarity of its environmental characteristics. To overcome this, this study employs a machine learning approach to develop a source and quality classification system for Teff cereal. Data is collected from various production areas in the Amhara regions, considering two types of cereal (high and low quality) across eight classes. A total of 5,920 images are collected, with 740 images for each class. Image enhancement techniques, including scaling, data augmentation, histogram equalization, and noise removal, are applied to preprocess the data. Convolutional Neural Network (CNN) is then used to extract relevant features and reduce dimensionality. The dataset is split into 80% for training and 20% for testing. Different classifiers, including FVGG16, FINCV3, QSCTC, EMQSCTC, SVM, and RF, are employed for classification, achieving accuracy rates ranging from 86.91% to 97.72%. The ensemble of FVGG16, FINCV3, and QSCTC using the Max-Voting approach outperforms individual algorithms.

Keywords: Teff, ensemble learning, max-voting, CNN, SVM, RF

Procedia PDF Downloads 53
7390 Application of Data Mining Techniques for Tourism Knowledge Discovery

Authors: Teklu Urgessa, Wookjae Maeng, Joong Seek Lee

Abstract:

Application of five implementations of three data mining classification techniques was experimented for extracting important insights from tourism data. The aim was to find out the best performing algorithm among the compared ones for tourism knowledge discovery. Knowledge discovery process from data was used as a process model. 10-fold cross validation method is used for testing purpose. Various data preprocessing activities were performed to get the final dataset for model building. Classification models of the selected algorithms were built with different scenarios on the preprocessed dataset. The outperformed algorithm tourism dataset was Random Forest (76%) before applying information gain based attribute selection and J48 (C4.5) (75%) after selection of top relevant attributes to the class (target) attribute. In terms of time for model building, attribute selection improves the efficiency of all algorithms. Artificial Neural Network (multilayer perceptron) showed the highest improvement (90%). The rules extracted from the decision tree model are presented, which showed intricate, non-trivial knowledge/insight that would otherwise not be discovered by simple statistical analysis with mediocre accuracy of the machine using classification algorithms.

Keywords: classification algorithms, data mining, knowledge discovery, tourism

Procedia PDF Downloads 295
7389 Global Voltage Harmonic Index for Measuring Harmonic Situation of Power Grids: A Focus on Power Transformers

Authors: Alireza Zabihi, Saeed Peyghami, Hossein Mokhtari

Abstract:

With the increasing deployment of renewable power plants, such as solar and wind, it is crucial to measure the harmonic situation of the grid. This paper proposes a global voltage harmonic index to measure the harmonic situation of the power grid with a focus on power transformers. The power electronics systems used to connect these plants to the network can introduce harmonics, leading to increased losses, reduced efficiency, false operation of protective relays, and equipment damage due to harmonic intensifications. The proposed index considers the losses caused by harmonics in power transformers which are of great importance and value to the network, providing a comprehensive measure of the harmonic situation of the grid. The effectiveness of the proposed index is evaluated on a real-world distribution network, and the results demonstrate its ability to identify the harmonic situation of the network, particularly in relation to power transformers. The proposed index provides a comprehensive measure of the harmonic situation of the grid, taking into account the losses caused by harmonics in power transformers. The proposed index has the potential to support power companies in optimizing their power systems and to guide researchers in developing effective mitigation strategies for harmonics in the power grid.

Keywords: global voltage harmonic index, harmonics, power grid, power quality, power transformers, renewable energy

Procedia PDF Downloads 127
7388 Developing Model for Fuel Consumption Optimization in Aviation Industry

Authors: Somesh Kumar Sharma, Sunanad Gupta

Abstract:

The contribution of aviation to society and economy is undisputedly significant. The aviation industry drives economic and social progress by contributing prominently to tourism, commerce and improved quality of life. Identifying the amount of fuel consumed by an aircraft while moving in both airspace and ground networks is critical to air transport economics. Aviation fuel is a major operating cost parameter of the aviation industry and at the same time it is prone to various constraints. This article aims to develop a model for fuel consumption of aviation product. The paper tailors the information for the fuel consumption optimization in terms of information development, information evaluation and information refinement. The information is evaluated and refined using statistical package R and Factor Analysis which is further validated with neural networking. The study explores three primary dimensions which are finally summarized into 23 influencing variables in contrast to 96 variables available in literature. The 23 variables explored in this study should be considered as highly influencing variables for fuel consumption which will contribute significantly towards fuel optimization.

Keywords: fuel consumption, civil aviation industry, neural networking, optimization

Procedia PDF Downloads 340
7387 Multi-Scale Urban Spatial Evolution Analysis Based on Space Syntax: A Case Study in Modern Yangzhou, China

Authors: Dai Zhimei, Hua Chen

Abstract:

The exploration of urban spatial evolution is an important part of urban development research. Therefore, the evolutionary modern Yangzhou urban spatial texture was taken as the research object, and Spatial Syntax was used as the main research tool, this paper explored Yangzhou spatial evolution law and its driving factors from the urban street network scale, district scale and street scale. The study has concluded that at the urban scale, Yangzhou urban spatial evolution is the result of a variety of causes, including physical and geographical condition, policy and planning factors, and traffic conditions, and the evolution of space also has an impact on social, economic, environmental and cultural factors. At the district and street scales, changes in space will have a profound influence on the history of the city and the activities of people. At the end of the article, the matters needing attention during the evolution of urban space were summarized.

Keywords: block, space syntax and methodology, street, urban space, Yangzhou

Procedia PDF Downloads 181
7386 Machine Learning Techniques in Bank Credit Analysis

Authors: Fernanda M. Assef, Maria Teresinha A. Steiner

Abstract:

The aim of this paper is to compare and discuss better classifier algorithm options for credit risk assessment by applying different Machine Learning techniques. Using records from a Brazilian financial institution, this study uses a database of 5,432 companies that are clients of the bank, where 2,600 clients are classified as non-defaulters, 1,551 are classified as defaulters and 1,281 are temporarily defaulters, meaning that the clients are overdue on their payments for up 180 days. For each case, a total of 15 attributes was considered for a one-against-all assessment using four different techniques: Artificial Neural Networks Multilayer Perceptron (ANN-MLP), Artificial Neural Networks Radial Basis Functions (ANN-RBF), Logistic Regression (LR) and finally Support Vector Machines (SVM). For each method, different parameters were analyzed in order to obtain different results when the best of each technique was compared. Initially the data were coded in thermometer code (numerical attributes) or dummy coding (for nominal attributes). The methods were then evaluated for each parameter and the best result of each technique was compared in terms of accuracy, false positives, false negatives, true positives and true negatives. This comparison showed that the best method, in terms of accuracy, was ANN-RBF (79.20% for non-defaulter classification, 97.74% for defaulters and 75.37% for the temporarily defaulter classification). However, the best accuracy does not always represent the best technique. For instance, on the classification of temporarily defaulters, this technique, in terms of false positives, was surpassed by SVM, which had the lowest rate (0.07%) of false positive classifications. All these intrinsic details are discussed considering the results found, and an overview of what was presented is shown in the conclusion of this study.

Keywords: artificial neural networks (ANNs), classifier algorithms, credit risk assessment, logistic regression, machine Learning, support vector machines

Procedia PDF Downloads 103