Search results for: Artificial Neural Network (ANNs)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6647

Search results for: Artificial Neural Network (ANNs)

5387 Singularization: A Technique for Protecting Neural Networks

Authors: Robert Poenaru, Mihail Pleşa

Abstract:

In this work, a solution that addresses the protection of pre-trained neural networks is developed: Singularization. This method involves applying permutations to the weight matrices of a pre-trained model, introducing a form of structured noise that obscures the original model’s architecture. These permutations make it difficult for an attacker to reconstruct the original model, even if the permuted weights are obtained. Experimental benchmarks indicate that the application of singularization has a profound impact on model performance, often degrading it to the point where retraining from scratch becomes necessary to recover functionality, which is particularly effective for securing intellectual property in neural networks. Moreover, unlike other approaches, singularization is lightweight and computationally efficient, which makes it well suited for resource-constrained environments. Our experiments also demonstrate that this technique performs efficiently in various image classification tasks, highlighting its broad applicability and practicality in real-world scenarios.

Keywords: machine learning, ANE, CNN, security

Procedia PDF Downloads 14
5386 Mapping Network Connection of Personality Traits and Psychiatric Symptoms in Chinese Adolescents

Authors: Yichao Lv, Minmin Cai, Yanqiang Tao, Xinyuan Zou, Chao Zhang, Xiangping Liu

Abstract:

Objective: This study aims to explore the network structure of personality traits and mental health and identify key factors for effective intervention strategies. Methods: All participants (N = 6,067; 3,368 females) underwent the Eysenck Personality Scale (EPQ) to measure personality traits and the Symptom Self-rating Scale (SCL-90) to measure psychiatric symptoms. Using the mean value of the SCL-90 total score plus one standard deviation as the cutoff, 854 participants (14.08%; 528 females) were categorized as individuals exhibiting potential psychological symptoms and were included in the follow-up network analysis. The structure and bridge centrality of the network for dimensions of EPQ and SCL-90 were estimated. Results: Between the EPQ and SCL-90, psychoticism (P), extraversion (E), and neuroticism (N) showed the strongest positive correlations with somatization (Som), interpersonal sensitivity (IS), and hostility (Hos), respectively. Extraversion (E), somatization (Som), and anxiety (Anx) were identified as the most important bridge factors influencing the overall network. Conclusions: This study explored the network structure and complex connections between mental health and personality traits from a network perspective, providing potential targets for intervening in adolescent personality traits and mental health.

Keywords: EPQ, SCL-90, Chinese adolescents, network analysis

Procedia PDF Downloads 48
5385 Three Issues for Integrating Artificial Intelligence into Legal Reasoning

Authors: Fausto Morais

Abstract:

Artificial intelligence has been widely used in law. Programs are able to classify suits, to identify decision-making patterns, to predict outcomes, and to formalize legal arguments as well. In Brazil, the artificial intelligence victor has been classifying cases to supreme court’s standards. When those programs act doing those tasks, they simulate some kind of legal decision and legal arguments, raising doubts about how artificial intelligence can be integrated into legal reasoning. Taking this into account, the following three issues are identified; the problem of hypernormatization, the argument of legal anthropocentrism, and the artificial legal principles. Hypernormatization can be seen in the Brazilian legal context in the Supreme Court’s usage of the Victor program. This program generated efficiency and consistency. On the other hand, there is a feasible risk of over standardizing factual and normative legal features. Then legal clerks and programmers should work together to develop an adequate way to model legal language into computational code. If this is possible, intelligent programs may enact legal decisions in easy cases automatically cases, and, in this picture, the legal anthropocentrism argument takes place. Such an argument argues that just humans beings should enact legal decisions. This is so because human beings have a conscience, free will, and self unity. In spite of that, it is possible to argue against the anthropocentrism argument and to show how intelligent programs may work overcoming human beings' problems like misleading cognition, emotions, and lack of memory. In this way, intelligent machines could be able to pass legal decisions automatically by classification, as Victor in Brazil does, because they are binding by legal patterns and should not deviate from them. Notwithstanding, artificial intelligent programs can be helpful beyond easy cases. In hard cases, they are able to identify legal standards and legal arguments by using machine learning. For that, a dataset of legal decisions regarding a particular matter must be available, which is a reality in Brazilian Judiciary. Doing such procedure, artificial intelligent programs can support a human decision in hard cases, providing legal standards and arguments based on empirical evidence. Those legal features claim an argumentative weight in legal reasoning and should serve as references for judges when they must decide to maintain or overcome a legal standard.

Keywords: artificial intelligence, artificial legal principles, hypernormatization, legal anthropocentrism argument, legal reasoning

Procedia PDF Downloads 145
5384 Comparison of Two Neural Networks To Model Margarine Age And Predict Shelf-Life Using Matlab

Authors: Phakamani Xaba, Robert Huberts, Bilainu Oboirien

Abstract:

The present study was aimed at developing & comparing two neural-network-based predictive models to predict shelf-life/product age of South African margarine using free fatty acid (FFA), water droplet size (D3.3), water droplet distribution (e-sigma), moisture content, peroxide value (PV), anisidine valve (AnV) and total oxidation (totox) value as input variables to the model. Brick margarine products which had varying ages ranging from fresh i.e. week 0 to week 47 were sourced. The brick margarine products which had been stored at 10 & 25 °C and were characterized. JMP and MATLAB models to predict shelf-life/ margarine age were developed and their performances were compared. The key performance indicators to evaluate the model performances were correlation coefficient (CC), root mean square error (RMSE), and mean absolute percentage error (MAPE) relative to the actual data. The MATLAB-developed model showed a better performance in all three performance indicators. The correlation coefficient of the MATLAB model was 99.86% versus 99.74% for the JMP model, the RMSE was 0.720 compared to 1.005 and the MAPE was 7.4% compared to 8.571%. The MATLAB model was selected to be the most accurate, and then, the number of hidden neurons/ nodes was optimized to develop a single predictive model. The optimized MATLAB with 10 neurons showed a better performance compared to the models with 1 & 5 hidden neurons. The developed models can be used by margarine manufacturers, food research institutions, researchers etc, to predict shelf-life/ margarine product age, optimize addition of antioxidants, extend shelf-life of products and proactively troubleshoot for problems related to changes which have an impact on shelf-life of margarine without conducting expensive trials.

Keywords: margarine shelf-life, predictive modelling, neural networks, oil oxidation

Procedia PDF Downloads 197
5383 Recent Developments in Artificial Intelligence and Information Communications Technology

Authors: Dolapo Adeyemo

Abstract:

Technology can be designed specifically for geriatrics and persons with disabilities or ICT accessibility solutions. Both solutions stand to benefit from advances in Artificial intelligence, which are computer systems that perform tasks that require human intelligence. Tasks such as decision making, visual perception, speech recognition, and even language translation are useful in both situation and will provide significant benefits to people with temporarily or permanent disabilities. This research’s goal is to review innovations focused on the use of artificial intelligence that bridges the accessibility gap in technology from a user-centered perspective. A mixed method approach that utilized a comprehensive review of academic literature on the subject combined with semi structure interviews of users, developers, and technology product owners. The internet of things and artificial intelligence technology is creating new opportunities in the assistive technology space and proving accessibility to existing technology. Device now more adaptable to the needs of the user by learning the behavior of users as they interact with the internet. Accessibility to devices have witnessed significant enhancements that continue to benefit people with disabilities. Examples of other advances identified are prosthetic limbs like robotic arms supported by artificial intelligence, route planning software for the visually impaired, and decision support tools for people with disabilities and even clinicians that provide care.

Keywords: ICT, IOT, accessibility solutions, universal design

Procedia PDF Downloads 87
5382 Analysis of Biomarkers Intractable Epileptogenic Brain Networks with Independent Component Analysis and Deep Learning Algorithms: A Comprehensive Framework for Scalable Seizure Prediction with Unimodal Neuroimaging Data in Pediatric Patients

Authors: Bliss Singhal

Abstract:

Epilepsy is a prevalent neurological disorder affecting approximately 50 million individuals worldwide and 1.2 million Americans. There exist millions of pediatric patients with intractable epilepsy, a condition in which seizures fail to come under control. The occurrence of seizures can result in physical injury, disorientation, unconsciousness, and additional symptoms that could impede children's ability to participate in everyday tasks. Predicting seizures can help parents and healthcare providers take precautions, prevent risky situations, and mentally prepare children to minimize anxiety and nervousness associated with the uncertainty of a seizure. This research proposes a comprehensive framework to predict seizures in pediatric patients by evaluating machine learning algorithms on unimodal neuroimaging data consisting of electroencephalogram signals. The bandpass filtering and independent component analysis proved to be effective in reducing the noise and artifacts from the dataset. Various machine learning algorithms’ performance is evaluated on important metrics such as accuracy, precision, specificity, sensitivity, F1 score and MCC. The results show that the deep learning algorithms are more successful in predicting seizures than logistic Regression, and k nearest neighbors. The recurrent neural network (RNN) gave the highest precision and F1 Score, long short-term memory (LSTM) outperformed RNN in accuracy and convolutional neural network (CNN) resulted in the highest Specificity. This research has significant implications for healthcare providers in proactively managing seizure occurrence in pediatric patients, potentially transforming clinical practices, and improving pediatric care.

Keywords: intractable epilepsy, seizure, deep learning, prediction, electroencephalogram channels

Procedia PDF Downloads 84
5381 The Possible Application of Artificial Intelligence in Hungarian Court Practice

Authors: László Schmidt

Abstract:

In the context of artificial intelligence, we need to pay primary and particular attention to ethical principles not only in the design process but also during the application process. According to the European Commission's Ethical Guidelines, AI must have three main characteristics: it must be legal, ethical and stabil. We must never lose sight of the ethical principles because we risk that this new technology will not help democratic decision-making under the rule of law, but will, on the contrary, destroy it. The rapid spread and use of artificial intelligence poses an enormous challenge to both lawmaking and law enforcement. On legislation because AI permeates many areas of our daily lives that the legislator must regulate. We can see how challenging it is to regulate e.g., selfdriving cars/taxis/vans etc. Not to mention, more recently, cryptocurrencies and Chat GPT, the use of which also requires legislative intervention, from copyright to scientific use and even law of succession. Artificial intelligence also poses an extraordinary challenge to law enforcement. In criminal cases, police and prosecutors can make great use of AI in investigations, e.g. in forensics, DNA samples, reconstruction, identification, etc. But it can also be of great help in the detection of crimes committed in cyberspace. In criminal or civil court proceedings, AI can also play a major role in the evaluation of evidence and proof. For example, a photo or video or audio recording could be immediately revealed as genuine or fake. Likewise, the authenticity or falsification of a document could be determined much more quickly and cheaply than with current procedure (expert witnesses). Neither the current Hungarian Civil Procedure Act nor the Criminal Procedure Act allows the use of artificial intelligence in the evidentiary process. However, this should be changed. To use this technology in court proceedings would be very useful. The procedures would be faster, simpler, and therefore cheaper. Artificial intelligence could also replace much of the work of expert witnesses. Its introduction into judicial procedures would certainly be justified, but with due respect for human rights, the right to a fair trial and other democratic and rule of law guarantees.

Keywords: artificial intelligence, judiciary, Hungarian, court practice

Procedia PDF Downloads 79
5380 NSBS: Design of a Network Storage Backup System

Authors: Xinyan Zhang, Zhipeng Tan, Shan Fan

Abstract:

The first layer of defense against data loss is the backup data. This paper implements an agent-based network backup system used the backup, server-storage and server-backup agent these tripartite construction, and we realize the snapshot and hierarchical index in the NSBS. It realizes the control command and data flow separation, balances the system load, thereby improving the efficiency of the system backup and recovery. The test results show the agent-based network backup system can effectively improve the task-based concurrency, reasonably allocate network bandwidth, the system backup performance loss costs smaller and improves data recovery efficiency by 20%.

Keywords: agent, network backup system, three architecture model, NSBS

Procedia PDF Downloads 459
5379 Machine Learning in Agriculture: A Brief Review

Authors: Aishi Kundu, Elhan Raza

Abstract:

"Necessity is the mother of invention" - Rapid increase in the global human population has directed the agricultural domain toward machine learning. The basic need of human beings is considered to be food which can be satisfied through farming. Farming is one of the major revenue generators for the Indian economy. Agriculture is not only considered a source of employment but also fulfils humans’ basic needs. So, agriculture is considered to be the source of employment and a pillar of the economy in developing countries like India. This paper provides a brief review of the progress made in implementing Machine Learning in the agricultural sector. Accurate predictions are necessary at the right time to boost production and to aid the timely and systematic distribution of agricultural commodities to make their availability in the market faster and more effective. This paper includes a thorough analysis of various machine learning algorithms applied in different aspects of agriculture (crop management, soil management, water management, yield tracking, livestock management, etc.).Due to climate changes, crop production is affected. Machine learning can analyse the changing patterns and come up with a suitable approach to minimize loss and maximize yield. Machine Learning algorithms/ models (regression, support vector machines, bayesian models, artificial neural networks, decision trees, etc.) are used in smart agriculture to analyze and predict specific outcomes which can be vital in increasing the productivity of the Agricultural Food Industry. It is to demonstrate vividly agricultural works under machine learning to sensor data. Machine Learning is the ongoing technology benefitting farmers to improve gains in agriculture and minimize losses. This paper discusses how the irrigation and farming management systems evolve in real-time efficiently. Artificial Intelligence (AI) enabled programs to emerge with rich apprehension for the support of farmers with an immense examination of data.

Keywords: machine Learning, artificial intelligence, crop management, precision farming, smart farming, pre-harvesting, harvesting, post-harvesting

Procedia PDF Downloads 105
5378 Improving Fingerprinting-Based Localization System Using Generative Artificial Intelligence

Authors: Getaneh Berie Tarekegn

Abstract:

A precise localization system is crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarming, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. The most common method for providing continuous positioning services in outdoor environments is by using a global navigation satellite system (GNSS). Due to nonline-of-sight, multipath, and weather conditions, GNSS systems do not perform well in dense urban, urban, and suburban areas.This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a novel semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. We also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 39 cm, and more than 90% of the errors are less than 82 cm. That is, numerical results proved that, in comparison to traditional methods, the proposed SRCLoc method can significantly improve positioning performance and reduce radio map construction costs.

Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine

Procedia PDF Downloads 71
5377 Suggestion for Malware Detection Agent Considering Network Environment

Authors: Ji-Hoon Hong, Dong-Hee Kim, Nam-Uk Kim, Tai-Myoung Chung

Abstract:

Smartphone users are increasing rapidly. Accordingly, many companies are running BYOD (Bring Your Own Device: Policies to bring private-smartphones to the company) policy to increase work efficiency. However, smartphones are always under the threat of malware, thus the company network that is connected smartphone is exposed to serious risks. Most smartphone malware detection techniques are to perform an independent detection (perform the detection of a single target application). In this paper, we analyzed a variety of intrusion detection techniques. Based on the results of analysis propose an agent using the network IDS.

Keywords: android malware detection, software-defined network, interaction environment, android malware detection, software-defined network, interaction environment

Procedia PDF Downloads 434
5376 Reliability Improvement of Power System Networks Using Adaptive Genetic Algorithm

Authors: Alireza Alesaadi

Abstract:

Reliability analysis is a powerful method for determining the weak points of the electrical networks. In designing of electrical network, it is tried to design the most reliable network with minimal system shutting down, but it is usually associated with increasing the cost. In this paper, using adaptive genetic algorithm, a method was presented that provides the most reliable system with a certain economical cost. Finally, the proposed method is applied to a sample network and results will be analyzed.

Keywords: reliability, adaptive genetic algorithm, electrical network, communication engineering

Procedia PDF Downloads 510
5375 GIS-Based Topographical Network for Minimum “Exertion” Routing

Authors: Katherine Carl Payne, Moshe Dror

Abstract:

The problem of minimum cost routing has been extensively explored in a variety of contexts. While there is a prevalence of routing applications based on least distance, time, and related attributes, exertion-based routing has remained relatively unexplored. In particular, the network structures traditionally used to construct minimum cost paths are not suited to representing exertion or finding paths of least exertion based on road gradient. In this paper, we introduce a topographical network or “topograph” that enables minimum cost routing based on the exertion metric on each arc in a given road network as it is related to changes in road gradient. We describe an algorithm for topograph construction and present the implementation of the topograph on a road network of the state of California with ~22 million nodes.

Keywords: topograph, RPE, routing, GIS

Procedia PDF Downloads 546
5374 The Effect of Newspaper Reporting on COVID-19 Vaccine Hesitancy: A Randomised Controlled Trial

Authors: Anna Rinaldi, Pierfrancesco Dellino

Abstract:

COVID-19 vaccine hesitancy can be observed at different rates in different countries. In June 2021, 1,068 people were surveyed in France and Italy to inquire about individual potential acceptance, focusing on time preferences in a risk-return framework: having the vaccination today, in a month, and in 3 months; perceived risks of vaccination and COVID-19; and expected benefit of the vaccine. A randomized controlled trial was conducted to understand how everyday stimuli like fact-based news about vaccines impact an audience's acceptance of vaccination. The main experiment involved two groups of participants and two different articles about vaccine-related thrombosis taken from two Italian newspapers. One article used a more abstract description and language, and the other used a more anecdotal description and concrete language; each group read only one of these articles. Two other groups were assigned categorization tasks; one was asked to complete a concrete categorization task, and the other an abstract categorization task. Individual preferences for vaccination were found to be variable and unstable over time, and individual choices of accepting, refusing, or delaying could be affected by the way news is written. In order to understand these dynamic preferences, the present work proposes a new model based on seven categories of human behaviors that were validated by a neural network. A treatment effect was observed: participants who read the articles shifted to vaccine hesitancy categories more than participants assigned to other treatments and control. Furthermore, there was a significant gender effect, showing that the type of language leading to a lower hesitancy rate for men is correlated with a higher hesitancy rate for women and vice versa. This outcome should be taken into consideration for an appropriate gender-based communication campaign aimed at achieving herd immunity. The trial was registered at ClinicalTrials.gov NCT05582564 (17/10/2022).

Keywords: vaccine hesitancy, risk elicitation, neural network, covid19

Procedia PDF Downloads 85
5373 Impact of Climate Change on Sea Level Rise along the Coastline of Mumbai City, India

Authors: Chakraborty Sudipta, A. R. Kambekar, Sarma Arnab

Abstract:

Sea-level rise being one of the most important impacts of anthropogenic induced climate change resulting from global warming and melting of icebergs at Arctic and Antarctic, the investigations done by various researchers both on Indian Coast and elsewhere during the last decade has been reviewed in this paper. The paper aims to ascertain the propensity of consistency of different suggested methods to predict the near-accurate future sea level rise along the coast of Mumbai. Case studies at East Coast, Southern Tip and West and South West coast of India have been reviewed. Coastal Vulnerability Index of several important international places has been compared, which matched with Intergovernmental Panel on Climate Change forecasts. The application of Geographic Information System mapping, use of remote sensing technology, both Multi Spectral Scanner and Thematic Mapping data from Landsat classified through Iterative Self-Organizing Data Analysis Technique for arriving at high, moderate and low Coastal Vulnerability Index at various important coastal cities have been observed. Instead of data driven, hindcast based forecast for Significant Wave Height, additional impact of sea level rise has been suggested. Efficacy and limitations of numerical methods vis-à-vis Artificial Neural Network has been assessed, importance of Root Mean Square error on numerical results is mentioned. Comparing between various computerized methods on forecast results obtained from MIKE 21 has been opined to be more reliable than Delft 3D model.

Keywords: climate change, Coastal Vulnerability Index, global warming, sea level rise

Procedia PDF Downloads 132
5372 Network Word Discovery Framework Based on Sentence Semantic Vector Similarity

Authors: Ganfeng Yu, Yuefeng Ma, Shanliang Yang

Abstract:

The word discovery is a key problem in text information retrieval technology. Methods in new word discovery tend to be closely related to words because they generally obtain new word results by analyzing words. With the popularity of social networks, individual netizens and online self-media have generated various network texts for the convenience of online life, including network words that are far from standard Chinese expression. How detect network words is one of the important goals in the field of text information retrieval today. In this paper, we integrate the word embedding model and clustering methods to propose a network word discovery framework based on sentence semantic similarity (S³-NWD) to detect network words effectively from the corpus. This framework constructs sentence semantic vectors through a distributed representation model, uses the similarity of sentence semantic vectors to determine the semantic relationship between sentences, and finally realizes network word discovery by the meaning of semantic replacement between sentences. The experiment verifies that the framework not only completes the rapid discovery of network words but also realizes the standard word meaning of the discovery of network words, which reflects the effectiveness of our work.

Keywords: text information retrieval, natural language processing, new word discovery, information extraction

Procedia PDF Downloads 95
5371 Artificial Intelligence Approach to Manage Human Resources Information System Process in the Construction Industry

Authors: Ahmed Emad Ahmed

Abstract:

This paper aims to address the concept of human resources information systems (HRIS) and how to link it to new technologies such as artificial intelligence (AI) to be implemented in two human resources processes. A literature view has been collected to cover the main points related to HRIS, AI, and BC. A study case has been presented by generating a random HRIS to apply some AI operations to it. Then, an algorithm was applied to the database to complete some human resources processes, including training and performance appraisal, using a pre-trained AI model. After that, outputs and results have been presented and discussed briefly. Finally, a conclusion has been introduced to show the ability of new technologies such as AI and ML to be applied to the human resources management processes.

Keywords: human resources new technologies, HR artificial intelligence, HRIS AI models, construction AI HRIS

Procedia PDF Downloads 170
5370 GAILoc: Improving Fingerprinting-Based Localization System Using Generative Artificial Intelligence

Authors: Getaneh Berie Tarekegn

Abstract:

A precise localization system is crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarming, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. The most common method for providing continuous positioning services in outdoor environments is by using a global navigation satellite system (GNSS). Due to nonline-of-sight, multipath, and weather conditions, GNSS systems do not perform well in dense urban, urban, and suburban areas.This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a novel semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. We also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 39 cm, and more than 90% of the errors are less than 82 cm. That is, numerical results proved that, in comparison to traditional methods, the proposed SRCLoc method can significantly improve positioning performance and reduce radio map construction costs.

Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine

Procedia PDF Downloads 75
5369 Optimized Dynamic Bayesian Networks and Neural Verifier Test Applied to On-Line Isolated Characters Recognition

Authors: Redouane Tlemsani, Redouane, Belkacem Kouninef, Abdelkader Benyettou

Abstract:

In this paper, our system is a Markovien system which we can see it like a Dynamic Bayesian Networks. One of the major interests of these systems resides in the complete training of the models (topology and parameters) starting from training data. The Bayesian Networks are representing models of dubious knowledge on complex phenomena. They are a union between the theory of probability and the graph theory in order to give effective tools to represent a joined probability distribution on a set of random variables. The representation of knowledge bases on description, by graphs, relations of causality existing between the variables defining the field of study. The theory of Dynamic Bayesian Networks is a generalization of the Bayesians networks to the dynamic processes. Our objective amounts finding the better structure which represents the relationships (dependencies) between the variables of a dynamic bayesian network. In applications in pattern recognition, one will carry out the fixing of the structure which obliges us to admit some strong assumptions (for example independence between some variables).

Keywords: Arabic on line character recognition, dynamic Bayesian network, pattern recognition, networks

Procedia PDF Downloads 618
5368 Artificial Habitat Mapping in Adriatic Sea

Authors: Annalisa Gaetani, Anna Nora Tassetti, Gianna Fabi

Abstract:

The hydroacoustic technology is an efficient tool to study the sea environment: the most recent advancement in artificial habitat mapping involves acoustic systems to investigate fish abundance, distribution and behavior in specific areas. Along with a detailed high-coverage bathymetric mapping of the seabed, the high-frequency Multibeam Echosounder (MBES) offers the potential of detecting fine-scale distribution of fish aggregation, combining its ability to detect at the same time the seafloor and the water column. Surveying fish schools distribution around artificial structures, MBES allows to evaluate how their presence modifies the biological natural habitat overtime in terms of fish attraction and abundance. In the last years, artificial habitat mapping experiences have been carried out by CNR-ISMAR in the Adriatic sea: fish assemblages aggregating at offshore gas platforms and artificial reefs have been systematically monitored employing different kinds of methodologies. This work focuses on two case studies: a gas extraction platform founded at 80 meters of depth in the central Adriatic sea, 30 miles far from the coast of Ancona, and the concrete and steel artificial reef of Senigallia, deployed by CNR-ISMAR about 1.2 miles offshore at a depth of 11.2 m . Relating the MBES data (metrical dimensions of fish assemblages, shape, depth, density etc.) with the results coming from other methodologies, such as experimental fishing surveys and underwater video camera, it has been possible to investigate the biological assemblage attracted by artificial structures hypothesizing which species populate the investigated area and their spatial dislocation from these artificial structures. Processing MBES bathymetric and water column data, 3D virtual scenes of the artificial habitats have been created, receiving an intuitive-looking depiction of their state and allowing overtime to evaluate their change in terms of dimensional characteristics and depth fish schools’ disposition. These MBES surveys play a leading part in the general multi-year programs carried out by CNR-ISMAR with the aim to assess potential biological changes linked to human activities on.

Keywords: artificial habitat mapping, fish assemblages, hydroacustic technology, multibeam echosounder

Procedia PDF Downloads 260
5367 Optimization of Reliability and Communicability of a Random Two-Dimensional Point Patterns Using Delaunay Triangulation

Authors: Sopheak Sorn, Kwok Yip Szeto

Abstract:

Reliability is one of the important measures of how well the system meets its design objective, and mathematically is the probability that a complex system will perform satisfactorily. When the system is described by a network of N components (nodes) and their L connection (links), the reliability of the system becomes a network design problem that is an NP-hard combinatorial optimization problem. In this paper, we address the network design problem for a random point set’s pattern in two dimensions. We make use of a Voronoi construction with each cell containing exactly one point in the point pattern and compute the reliability of the Voronoi’s dual, i.e. the Delaunay graph. We further investigate the communicability of the Delaunay network. We find that there is a positive correlation and a negative correlation between the homogeneity of a Delaunay's degree distribution with its reliability and its communicability respectively. Based on the correlations, we alter the communicability and the reliability by performing random edge flips, which preserve the number of links and nodes in the network but can increase the communicability in a Delaunay network at the cost of its reliability. This transformation is later used to optimize a Delaunay network with the optimum geometric mean between communicability and reliability. We also discuss the importance of the edge flips in the evolution of real soap froth in two dimensions.

Keywords: Communicability, Delaunay triangulation, Edge Flip, Reliability, Two dimensional network, Voronio

Procedia PDF Downloads 419
5366 A New Method to Reduce 5G Application Layer Payload Size

Authors: Gui Yang Wu, Bo Wang, Xin Wang

Abstract:

Nowadays, 5G service-based interface architecture uses text-based payload like JSON to transfer business data between network functions, which has obvious advantages as internet services but causes unnecessarily larger traffic. In this paper, a new 5G application payload size reduction method is presented to provides the mechanism to negotiate about new capability between network functions when network communication starts up and how 5G application data are reduced according to negotiated information with peer network function. Without losing the advantages of 5G text-based payload, this method demonstrates an excellent result on application payload size reduction and does not increase the usage quota of computing resource. Implementation of this method does not impact any standards or specifications and not change any encoding or decoding functionality too. In a real 5G network, this method will contribute to network efficiency and eventually save considerable computing resources.

Keywords: 5G, JSON, payload size, service-based interface

Procedia PDF Downloads 184
5365 Thermal Network Model for a Large Scale AC Induction Motor

Authors: Sushil Kumar, M. Dakshina Murty

Abstract:

Thermal network modelling has proven to be important tool for thermal analysis of electrical machine. This article investigates numerical thermal network model and experimental performance of a large-scale AC motor. Experimental temperatures were measured using RTD in the stator which have been compared with the numerical data. Thermal network modelling fairly predicts the temperature of various components inside the large-scale AC motor. Results of stator winding temperature is compared with experimental results which are in close agreement with accuracy of 6-10%. This method of predicting hot spots within AC motors can be readily used by the motor designers for estimating the thermal hot spots of the machine.

Keywords: AC motor, thermal network, heat transfer, modelling

Procedia PDF Downloads 326
5364 The Usage of Artificial Intelligence in Instagram

Authors: Alanod Alqasim, Yasmine Iskandarani, Sita Algethami, Jawaher alzughaiby

Abstract:

This study focuses on the usage of AI (Artificial Intelligence) systems and features on the Instagram application and how it influences user experience and satisfaction. The aim is to evaluate the techniques and current capabilities, restrictions, and potential future directions of AI in an Instagram application. Following a concise explanation of the core concepts underlying AI usage on Instagram. To answer this question, 19 randomly selected users were asked to complete a 9-question survey on their experience and satisfaction with the app's features (Filters, user preferences, translation tool) and authenticity. The results revealed that there were three prevalent allegations. These declarations include that Instagram has an extremely attractive user interface; secondly, Instagram creates a strong sense of community; and lastly, Instagram has an important influence on mental health.

Keywords: AI (Artificial Intelligence), instagram, features, satisfaction, experience

Procedia PDF Downloads 82
5363 Artificial Intelligence and Cybernetics in Bertrand Russell’s Philosophy

Authors: Djoudi Ali

Abstract:

In this article, we shall expose some of the more interesting interactions of philosophy and cybernetics, some philosophical issues arising in cybernetic systems, and some questions in philosophy of our daily life related to the artificial intelligence. Many of these are fruitfully explored in the article..This article will shed light also on the importance of science and technology in our life and what are the main problems of misusing the latest technologies known under artificial intelligence and cybernatics acoording to Bertrand Russell’s point of view; then to analyse his project of reforms inculding science progress risks , the article show also the whole aspect of the impact of technology on peace , nature and on individual daily behavior, we shall discuss all issues and defies imposing by this new era , The article will invest in showing what Russell will suggest to eliminate or to slow down the dangers of these changes and what are the main solutions to protect the indiviual’s rights and responsiblities In this article, We followed a different methodology, like analysis method and sometimes the historical or descriptive method, without forgetting criticizing some conclusions when it is logically needed In the end, we mentioned what is supposed to be solutions suggested by Bertrand Russell that should be taken into considerations during the next decades and how to protect our ennvironement and the human being of any risk of disappearing

Keywords: artificial intelligence, technology, cybernetics, sience

Procedia PDF Downloads 125
5362 One-Step Time Series Predictions with Recurrent Neural Networks

Authors: Vaidehi Iyer, Konstantin Borozdin

Abstract:

Time series prediction problems have many important practical applications, but are notoriously difficult for statistical modeling. Recently, machine learning methods have been attracted significant interest as a practical tool applied to a variety of problems, even though developments in this field tend to be semi-empirical. This paper explores application of Long Short Term Memory based Recurrent Neural Networks to the one-step prediction of time series for both trend and stochastic components. Two types of data are analyzed - daily stock prices, that are often considered to be a typical example of a random walk, - and weather patterns dominated by seasonal variations. Results from both analyses are compared, and reinforced learning framework is used to select more efficient between Recurrent Neural Networks and more traditional auto regression methods. It is shown that both methods are able to follow long-term trends and seasonal variations closely, but have difficulties with reproducing day-to-day variability. Future research directions and potential real world applications are briefly discussed.

Keywords: long short term memory, prediction methods, recurrent neural networks, reinforcement learning

Procedia PDF Downloads 229
5361 Intelligent Indoor Localization Using WLAN Fingerprinting

Authors: Gideon C. Joseph

Abstract:

The ability to localize mobile devices is quite important, as some applications may require location information of these devices to operate or deliver better services to the users. Although there are several ways of acquiring location data of mobile devices, the WLAN fingerprinting approach has been considered in this work. This approach uses the Received Signal Strength Indicator (RSSI) measurement as a function of the position of the mobile device. RSSI is a quantitative technique of describing the radio frequency power carried by a signal. RSSI may be used to determine RF link quality and is very useful in dense traffic scenarios where interference is of major concern, for example, indoor environments. This research aims to design a system that can predict the location of a mobile device, when supplied with the mobile’s RSSIs. The developed system takes as input the RSSIs relating to the mobile device, and outputs parameters that describe the location of the device such as the longitude, latitude, floor, and building. The relationship between the Received Signal Strengths (RSSs) of mobile devices and their corresponding locations is meant to be modelled; hence, subsequent locations of mobile devices can be predicted using the developed model. It is obvious that describing mathematical relationships between the RSSIs measurements and localization parameters is one option to modelling the problem, but the complexity of such an approach is a serious turn-off. In contrast, we propose an intelligent system that can learn the mapping of such RSSIs measurements to the localization parameters to be predicted. The system is capable of upgrading its performance as more experiential knowledge is acquired. The most appealing consideration to using such a system for this task is that complicated mathematical analysis and theoretical frameworks are excluded or not needed; the intelligent system on its own learns the underlying relationship in the supplied data (RSSI levels) that corresponds to the localization parameters. These localization parameters to be predicted are of two different tasks: Longitude and latitude of mobile devices are real values (regression problem), while the floor and building of the mobile devices are of integer values or categorical (classification problem). This research work presents artificial neural network based intelligent systems to model the relationship between the RSSIs predictors and the mobile device localization parameters. The designed systems were trained and validated on the collected WLAN fingerprint database. The trained networks were then tested with another supplied database to obtain the performance of trained systems on achieved Mean Absolute Error (MAE) and error rates for the regression and classification tasks involved therein.

Keywords: indoor localization, WLAN fingerprinting, neural networks, classification, regression

Procedia PDF Downloads 347
5360 The Use of Artificial Intelligence to Harmonization in the Lawmaking Process

Authors: Supriyadi, Andi Intan Purnamasari, Aminuddin Kasim, Sulbadana, Mohammad Reza

Abstract:

The development of the Industrial Revolution Era 4.0 brought a significant influence in the administration of countries in all parts of the world, including Indonesia, not only in the administration and economic sectors but the ways and methods of forming laws should also be adjusted. Until now, the process of making laws carried out by the Parliament with the Government still uses the classical method. The law-making process still uses manual methods, such as typing harmonization of regulations, so that it is not uncommon for errors to occur, such as writing errors, copying articles and so on, things that require a high level of accuracy and relying on inventory and harmonization carried out manually by humans. However, this method often creates several problems due to errors and inaccuracies on the part of officers who harmonize laws after discussion and approval; this has a very serious impact on the system of law formation in Indonesia. The use of artificial intelligence in the process of forming laws seems to be justified and becomes the answer in order to minimize the disharmony of various laws and regulations. This research is normative research using the Legislative Approach and the Conceptual Approach. This research focuses on the question of how to use Artificial Intelligence for Harmonization in the Lawmaking Process.

Keywords: artificial intelligence, harmonization, laws, intelligence

Procedia PDF Downloads 162
5359 Refined Edge Detection Network

Authors: Omar Elharrouss, Youssef Hmamouche, Assia Kamal Idrissi, Btissam El Khamlichi, Amal El Fallah-Seghrouchni

Abstract:

Edge detection is represented as one of the most challenging tasks in computer vision, due to the complexity of detecting the edges or boundaries in real-world images that contains objects of different types and scales like trees, building as well as various backgrounds. Edge detection is represented also as a key task for many computer vision applications. Using a set of backbones as well as attention modules, deep-learning-based methods improved the detection of edges compared with the traditional methods like Sobel and Canny. However, images of complex scenes still represent a challenge for these methods. Also, the detected edges using the existing approaches suffer from non-refined results while the image output contains many erroneous edges. To overcome this, n this paper, by using the mechanism of residual learning, a refined edge detection network is proposed (RED-Net). By maintaining the high resolution of edges during the training process, and conserving the resolution of the edge image during the network stage, we make the pooling outputs at each stage connected with the output of the previous layer. Also, after each layer, we use an affined batch normalization layer as an erosion operation for the homogeneous region in the image. The proposed methods are evaluated using the most challenging datasets including BSDS500, NYUD, and Multicue. The obtained results outperform the designed edge detection networks in terms of performance metrics and quality of output images.

Keywords: edge detection, convolutional neural networks, deep learning, scale-representation, backbone

Procedia PDF Downloads 102
5358 Identification of Soft Faults in Branched Wire Networks by Distributed Reflectometry and Multi-Objective Genetic Algorithm

Authors: Soumaya Sallem, Marc Olivas

Abstract:

This contribution presents a method for detecting, locating, and characterizing soft faults in a complex wired network. The proposed method is based on multi-carrier reflectometry MCTDR (Multi-Carrier Time Domain Reflectometry) combined with a multi-objective genetic algorithm. In order to ensure complete network coverage and eliminate diagnosis ambiguities, the MCTDR test signal is injected at several points on the network, and the data is merged between different reflectometers (sensors) distributed on the network. An adapted multi-objective genetic algorithm is used to merge data in order to obtain more accurate faults location and characterization. The proposed method performances are evaluated from numerical and experimental results.

Keywords: wired network, reflectometry, network distributed diagnosis, multi-objective genetic algorithm

Procedia PDF Downloads 195