Search results for: random graph
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2443

Search results for: random graph

1933 Fraud Detection in Credit Cards with Machine Learning

Authors: Anjali Chouksey, Riya Nimje, Jahanvi Saraf

Abstract:

Online transactions have increased dramatically in this new ‘social-distancing’ era. With online transactions, Fraud in online payments has also increased significantly. Frauds are a significant problem in various industries like insurance companies, baking, etc. These frauds include leaking sensitive information related to the credit card, which can be easily misused. Due to the government also pushing online transactions, E-commerce is on a boom. But due to increasing frauds in online payments, these E-commerce industries are suffering a great loss of trust from their customers. These companies are finding credit card fraud to be a big problem. People have started using online payment options and thus are becoming easy targets of credit card fraud. In this research paper, we will be discussing machine learning algorithms. We have used a decision tree, XGBOOST, k-nearest neighbour, logistic-regression, random forest, and SVM on a dataset in which there are transactions done online mode using credit cards. We will test all these algorithms for detecting fraud cases using the confusion matrix, F1 score, and calculating the accuracy score for each model to identify which algorithm can be used in detecting frauds.

Keywords: machine learning, fraud detection, artificial intelligence, decision tree, k nearest neighbour, random forest, XGBOOST, logistic regression, support vector machine

Procedia PDF Downloads 133
1932 KPI and Tool for the Evaluation of Competency in Warehouse Management for Furniture Business

Authors: Kritchakhris Na-Wattanaprasert

Abstract:

The objective of this research is to design and develop a prototype of a key performance indicator system this is suitable for warehouse management in a case study and use requirement. In this study, we design a prototype of key performance indicator system (KPI) for warehouse case study of furniture business by methodology in step of identify scope of the research and study related papers, gather necessary data and users requirement, develop key performance indicator base on balance scorecard, design pro and database for key performance indicator, coding the program and set relationship of database and finally testing and debugging each module. This study use Balance Scorecard (BSC) for selecting and grouping key performance indicator. The system developed by using Microsoft SQL Server 2010 is used to create the system database. In regard to visual-programming language, Microsoft Visual C# 2010 is chosen as the graphic user interface development tool. This system consists of six main menus: menu login, menu main data, menu financial perspective, menu customer perspective, menu internal, and menu learning and growth perspective. Each menu consists of key performance indicator form. Each form contains a data import section, a data input section, a data searches – edit section, and a report section. The system generates outputs in 5 main reports, the KPI detail reports, KPI summary report, KPI graph report, benchmarking summary report and benchmarking graph report. The user will select the condition of the report and period time. As the system has been developed and tested, discovers that it is one of the ways to judging the extent to warehouse objectives had been achieved. Moreover, it encourages the warehouse functional proceed with more efficiency. In order to be useful propose for other industries, can adjust this system appropriately. To increase the usefulness of the key performance indicator system, the recommendations for further development are as follows: -The warehouse should review the target value and set the better suitable target periodically under the situation fluctuated in the future. -The warehouse should review the key performance indicators and set the better suitable key performance indicators periodically under the situation fluctuated in the future for increasing competitiveness and take advantage of new opportunities.

Keywords: key performance indicator, warehouse management, warehouse operation, logistics management

Procedia PDF Downloads 417
1931 Rejection Sensitivity and Romantic Relationships: A Systematic Review and Meta-Analysis

Authors: Mandira Mishra, Mark Allen

Abstract:

This meta-analysis explored whether rejection sensitivity relates to facets of romantic relationships. A comprehensive literature search identified 60 studies (147 effect sizes; 16,955 participants) that met inclusion criteria. Data were analysed using inverse-variance weighted random effects meta-analysis. Mean effect sizes from 21 meta-analyses provided evidence that more rejection sensitive individuals report lower levels of relationship satisfaction and relationship closeness, lower levels of perceived partner satisfaction, a greater likelihood of intimate partner violence (perpetration and victimization), higher levels of relationship concerns and relationship conflict, and higher levels of jealousy and self-silencing behaviours. There was also some evidence that rejection sensitive individuals are more likely to engage in risky sexual behaviour and are more prone to sexual compulsivity. There was no evidence of publication bias and various levels of heterogeneity in computed averages. Random effects meta-regression identified participant age and sex as important moderators of pooled mean effects. These findings provide a foundation for the theoretical development of rejection sensitivity in romantic relationships and should be of interest to relationship and marriage counsellors and other relationship professionals.

Keywords: intimate partner violence, relationship satisfaction, commitment, sexual orientation, risky sexual behaviour

Procedia PDF Downloads 68
1930 Deconstructing Local Area Networks Using MaatPeace

Authors: Gerald Todd

Abstract:

Recent advances in random epistemologies and ubiquitous theory have paved the way for web services. Given the current status of linear-time communication, cyberinformaticians compellingly desire the exploration of link-level acknowledgements. In order to realize this purpose, we concentrate our efforts on disconfirming that DHTs and model checking are mostly incompatible.

Keywords: LAN, cyberinformatics, model checking, communication

Procedia PDF Downloads 382
1929 Effectiveness of Cranberry Ingesting for Prevention of Urinary Tract Infection: A Systematic Review and Meta-Analysis of Randomized Controlled Trials

Authors: Yu-Chieh Huang, Pei-Shih Chen, Tao-Hsin Tung

Abstract:

Background: Urinary tract infection is the most common bacterial infection to our best knowledge. Objective: This study is to investigate whether cranberry ingesting could improve the urinary tract infection. Methods: We searched the PubMed and Cochrane Library for relevant randomized controlled trials without language limitations between 9 March 1994 and June 30, 2017, with a priori defined inclusion and exclusion criteria. The search terms included (cranberry OR Vaccinium macrocarpon OR Vaccinium oxy-coccus OR Vaccinium microcarpum OR Vaccinium erythrocarpum OR Vaccinium) AND (urinary tract infection OR bacteriuria OR pyuria) AND (effect OR effective-ness OR efficacy) AND (random OR randomized). Results: There were 26 studies met the selection criteria included among 4709 eligible participants. We analyzed all trials in meta-analysis. The random-effects pooled risk ratio (RR) for the group using cranberry versus using placebo was 0.75; 95%CI[0.63, 0.880]; p-value=0.0002) and heterogeneity was 56%. Furthermore, we divided the subjects into different subgroup to analysis. Ingesting cranberry seemed to be more effective in some subgroups, including the patients with recurrent UTI (RR, 0.71; 95%CI[0.54,0.93]; p-value=0.002) (I²= 65%) and female population (RR, 0.73, 95%CI[0.58,0.92]; p-value=0.002) (I²= 59%). The prevention effect was not different between cranberry and trimethoprim (RR, 1.25, 95%CI[0.67, 2.33]; p-value=0.49) (I²= 68%). No matter the forms of cranberry were capsules or juice, the efficacy was useful. Conclusions: It is showed that cranberry ingesting is usefully associated with prevention UTI. There are more effective in prevention of UTI in some groups.

Keywords: cranberry, effectiveness, prevention, urinary tract infect

Procedia PDF Downloads 373
1928 Thermal Analysis of Automobile Radiator Using Nanofluids

Authors: S. Sumanth, Babu Rao Ponangi, K. N. Seetharamu

Abstract:

As the technology is emerging day by day, there is a need for some better methodology which will enhance the performance of radiator. Nanofluid is the one area which has promised the enhancement of the radiator performance. Currently, nanofluid has got a well effective solution for enhancing the performance of the automobile radiators. Suspending the nano sized particle in the base fluid, which has got better thermal conductivity value when compared to a base fluid, is preferably considered for nanofluid. In the current work, at first mathematical formulation has been carried out, which will govern the performance of the radiator. Current work is justified by plotting the graph for different parameters. Current work justifies the enhancement of radiator performance using nanofluid.

Keywords: nanofluid, radiator performance, graphene, gamma aluminium oxide (γ-Al2O3), titanium dioxide (TiO2)

Procedia PDF Downloads 233
1927 Gaussian Probability Density for Forest Fire Detection Using Satellite Imagery

Authors: S. Benkraouda, Z. Djelloul-Khedda, B. Yagoubi

Abstract:

we present a method for early detection of forest fires from a thermal infrared satellite image, using the image matrix of the probability of belonging. The principle of the method is to compare a theoretical mathematical model to an experimental model. We considered that each line of the image matrix, as an embodiment of a non-stationary random process. Since the distribution of pixels in the satellite image is statistically dependent, we divided these lines into small stationary and ergodic intervals to characterize the image by an adequate mathematical model. A standard deviation was chosen to generate random variables, so each interval behaves naturally like white Gaussian noise. The latter has been selected as the mathematical model that represents a set of very majority pixels, which we can be considered as the image background. Before modeling the image, we made a few pretreatments, then the parameters of the theoretical Gaussian model were extracted from the modeled image, these settings will be used to calculate the probability of each interval of the modeled image to belong to the theoretical Gaussian model. The high intensities pixels are regarded as foreign elements to it, so they will have a low probability, and the pixels that belong to the background image will have a high probability. Finally, we did present the reverse of the matrix of probabilities of these intervals for a better fire detection.

Keywords: forest fire, forest fire detection, satellite image, normal distribution, theoretical gaussian model, thermal infrared matrix image

Procedia PDF Downloads 124
1926 Solution-Processed Threshold Switching Selectors Based on Highly Flexible, Transparent and Scratchable Silver Nanowires Conductive Films

Authors: Peiyuan Guan, Tao Wan, Dewei Chu

Abstract:

With the flash memory approaching its physical limit, the emerging resistive random-access memory (RRAM) has been considered as one of the most promising candidates for the next-generation non-volatile memory. One selector-one resistor configuration has shown the most promising way to resolve the crosstalk issue without affecting the scalability and high-density integration of the RRAM array. By comparison with other candidates of selectors (such as diodes and nonlinear devices), threshold switching selectors dominated by formation/spontaneous rupture of fragile conductive filaments have been proved to possess low voltages, high selectivity, and ultra-low current leakage. However, the flexibility and transparency of selectors are barely mentioned. Therefore, it is a matter of urgency to develop a selector with highly flexible and transparent properties to assist the application of RRAM for a diversity of memory devices. In this work, threshold switching selectors were designed using a facilely solution-processed fabrication on AgNWs@PDMS composite films, which show high flexibility, transparency and scratch resistance. As-fabricated threshold switching selectors also have revealed relatively high selectivity (~107), low operating voltages (Vth < 1 V) and good switching performance.

Keywords: flexible and transparent, resistive random-access memory, silver nanowires, threshold switching selector

Procedia PDF Downloads 115
1925 Talent-to-Vec: Using Network Graphs to Validate Models with Data Sparsity

Authors: Shaan Khosla, Jon Krohn

Abstract:

In a recruiting context, machine learning models are valuable for recommendations: to predict the best candidates for a vacancy, to match the best vacancies for a candidate, and compile a set of similar candidates for any given candidate. While useful to create these models, validating their accuracy in a recommendation context is difficult due to a sparsity of data. In this report, we use network graph data to generate useful representations for candidates and vacancies. We use candidates and vacancies as network nodes and designate a bi-directional link between them based on the candidate interviewing for the vacancy. After using node2vec, the embeddings are used to construct a validation dataset with a ranked order, which will help validate new recommender systems.

Keywords: AI, machine learning, NLP, recruiting

Procedia PDF Downloads 73
1924 Mechanical-Reliability Coupling for a Bearing Capacity Assessment of Shallow Foundations

Authors: Amal Hentati, Mbarka Selmi, Tarek Kormi, Julien Baroth, Barthelemy Harthong

Abstract:

The impact of uncertainties on the performance assessment of shallow foundations is often significant. The need of the geotechnical engineers to a more objective and rigorous description of soil variations permitting to quantify these uncertainties and to incorporate them into calculation methods led to the development of reliability approaches. In this context, a mechanical-reliability coupling was developed in this paper, using a program coded in Matlab and the finite element software Abaqus, for the bearing capacity assessment of shallow foundations. The reliability analysis, based on the finite element method, assumed both soil cohesion and friction angle as uncertain parameters characterized by normal or lognormal probability distributions. The inherent spatial variability of both soil properties was, then, taken into account using 1D stationary random fields. The application of the proposed methodology to a shallow foundation subjected to a centered vertical loading permitted to highlight the proposed process interest. Findings proved the insufficiency of the conventional approach to predict the foundation failure and a high sensitivity of the ultimate loads to the soil properties uncertainties, mainly those related to the friction angle, was noted. Moreover, an asymmetry of both displacement and velocity fields was obtained.

Keywords: mechanical-reliability coupling, finite element method, shallow foundation, random fields, spatial variability

Procedia PDF Downloads 645
1923 The Effect of Intimate Partner Violence on Child Abuse in South Korea: Focused on the Moderating Effects of Patriarchal Attitude and Informal Social Control

Authors: Hye Lin Yang, Clifton R. Emery

Abstract:

Purpose: The purpose of this study is to examine the effects of intimate partner violence on child abuse, whether patriarchal attitude and informal social control moderate the relationship between intimate partner violence and child abuse. This study was conducted with data from The Seoul Families and Neighborhoods Study (SFNS). The SFNS is a representative random probability 3-stage cluster sample of 541 cohabiting couples in Seoul, South Korea collected in 2012. To verify research models, Random effect analysis were used. All analyses were performed using the Stata program. Results: Crucial findings are the following. First, intimate partner violence showed a significantly positive relationship with Child abuse. Second, there are significant moderating effects of informal social control on intimate partner violence - child abuse. Third, there are significant moderating effects of patriarchal attitude on intimate partner violence - child abuse. In other words, Patriarchal attitude is a significant risk factor of child abuse and informal social control is a significant Protection factor of child abuse. Based on results, the policy and practical implications for preventing child abuse, promoting informal social control were discussed.

Keywords: Intimate partner violence, child abuse, informal social control, patriarchal attitude

Procedia PDF Downloads 281
1922 Nonlinear Analysis in Investigating the Complexity of Neurophysiological Data during Reflex Behavior

Authors: Juliana A. Knocikova

Abstract:

Methods of nonlinear signal analysis are based on finding that random behavior can arise in deterministic nonlinear systems with a few degrees of freedom. Considering the dynamical systems, entropy is usually understood as a rate of information production. Changes in temporal dynamics of physiological data are indicating evolving of system in time, thus a level of new signal pattern generation. During last decades, many algorithms were introduced to assess some patterns of physiological responses to external stimulus. However, the reflex responses are usually characterized by short periods of time. This characteristic represents a great limitation for usual methods of nonlinear analysis. To solve the problems of short recordings, parameter of approximate entropy has been introduced as a measure of system complexity. Low value of this parameter is reflecting regularity and predictability in analyzed time series. On the other side, increasing of this parameter means unpredictability and a random behavior, hence a higher system complexity. Reduced neurophysiological data complexity has been observed repeatedly when analyzing electroneurogram and electromyogram activities during defence reflex responses. Quantitative phrenic neurogram changes are also obvious during severe hypoxia, as well as during airway reflex episodes. Concluding, the approximate entropy parameter serves as a convenient tool for analysis of reflex behavior characterized by short lasting time series.

Keywords: approximate entropy, neurophysiological data, nonlinear dynamics, reflex

Procedia PDF Downloads 289
1921 Parkinson’s Disease Detection Analysis through Machine Learning Approaches

Authors: Muhtasim Shafi Kader, Fizar Ahmed, Annesha Acharjee

Abstract:

Machine learning and data mining are crucial in health care, as well as medical information and detection. Machine learning approaches are now being utilized to improve awareness of a variety of critical health issues, including diabetes detection, neuron cell tumor diagnosis, COVID 19 identification, and so on. Parkinson’s disease is basically a disease for our senior citizens in Bangladesh. Parkinson's Disease indications often seem progressive and get worst with time. People got affected trouble walking and communicating with the condition advances. Patients can also have psychological and social vagaries, nap problems, hopelessness, reminiscence loss, and weariness. Parkinson's disease can happen in both men and women. Though men are affected by the illness at a proportion that is around partial of them are women. In this research, we have to get out the accurate ML algorithm to find out the disease with a predictable dataset and the model of the following machine learning classifiers. Therefore, nine ML classifiers are secondhand to portion study to use machine learning approaches like as follows, Naive Bayes, Adaptive Boosting, Bagging Classifier, Decision Tree Classifier, Random Forest classifier, XBG Classifier, K Nearest Neighbor Classifier, Support Vector Machine Classifier, and Gradient Boosting Classifier are used.

Keywords: naive bayes, adaptive boosting, bagging classifier, decision tree classifier, random forest classifier, XBG classifier, k nearest neighbor classifier, support vector classifier, gradient boosting classifier

Procedia PDF Downloads 117
1920 The Relationship between the Use of Social Networks with Executive Functions and Academic Performance in High School Students in Tehran

Authors: Esmail Sadipour

Abstract:

The use of social networks is increasing day by day in all societies. The purpose of this research was to know the relationship between the use of social networks (Instagram, WhatsApp, and Telegram) with executive functions and academic performance in first-year female high school students. This research was applied in terms of purpose, quantitative in terms of data type, and correlational in terms of technique. The population of this research consisted of all female high school students in the first year of district 2 of Tehran. Using Green's formula, the sample size of 150 people was determined and selected by cluster random method. In this way, from all 17 high schools in district 2 of Tehran, 5 high schools were selected by a simple random method and then one class was selected from each high school, and a total of 155 students were selected. To measure the use of social networks, a researcher-made questionnaire was used, the Barclay test (2012) was used for executive functions, and last semester's GPA was used for academic performance. Pearson's correlation coefficient and multivariate regression were used to analyze the data. The results showed that there is a negative relationship between the amount of use of social networks and self-control, self-motivation and time self-management. In other words, the more the use of social networks, the fewer executive functions of students, self-control, self-motivation, and self-management of their time. Also, with the increase in the use of social networks, the academic performance of students has decreased.

Keywords: social networks, executive function, academic performance, working memory

Procedia PDF Downloads 71
1919 Probabilistic Slope Stability Analysis of Excavation Induced Landslides Using Hermite Polynomial Chaos

Authors: Schadrack Mwizerwa

Abstract:

The characterization and prediction of landslides are crucial for assessing geological hazards and mitigating risks to infrastructure and communities. This research aims to develop a probabilistic framework for analyzing excavation-induced landslides, which is fundamental for assessing geological hazards and mitigating risks to infrastructure and communities. The study uses Hermite polynomial chaos, a non-stationary random process, to analyze the stability of a slope and characterize the failure probability of a real landslide induced by highway construction excavation. The correlation within the data is captured using the Karhunen-Loève (KL) expansion theory, and the finite element method is used to analyze the slope's stability. The research contributes to the field of landslide characterization by employing advanced random field approaches, providing valuable insights into the complex nature of landslide behavior and the effectiveness of advanced probabilistic models for risk assessment and management. The data collected from the Baiyuzui landslide, induced by highway construction, is used as an illustrative example. The findings highlight the importance of considering the probabilistic nature of landslides and provide valuable insights into the complex behavior of such hazards.

Keywords: Hermite polynomial chaos, Karhunen-Loeve, slope stability, probabilistic analysis

Procedia PDF Downloads 62
1918 K-Means Clustering-Based Infinite Feature Selection Method

Authors: Seyyedeh Faezeh Hassani Ziabari, Sadegh Eskandari, Maziar Salahi

Abstract:

Infinite Feature Selection (IFS) algorithm is an efficient feature selection algorithm that selects a subset of features of all sizes (including infinity). In this paper, we present an improved version of it, called clustering IFS (CIFS), by clustering the dataset in advance. To do so, first, we apply the K-means algorithm to cluster the dataset, then we apply IFS. In the CIFS method, the spatial and temporal complexities are reduced compared to the IFS method. Experimental results on 6 datasets show the superiority of CIFS compared to IFS in terms of accuracy, running time, and memory consumption.

Keywords: feature selection, infinite feature selection, clustering, graph

Procedia PDF Downloads 108
1917 The Analysis of Split Graphs in Social Networks Based on the k-Cardinality Assignment Problem

Authors: Ivan Belik

Abstract:

In terms of social networks split graphs correspond to the variety of interpersonal and intergroup relations. In this paper we analyse the interaction between the cliques (socially strong and trusty groups) and the independent sets (fragmented and non-connected groups of people) as the basic components of any split graph. Based on the Semi-Lagrangean relaxation for the k-cardinality assignment problem we show the way of how to minimize the socially risky interactions between the cliques and the independent sets within the social network.

Keywords: cliques, independent sets, k-cardinality assignment, social networks, split graphs

Procedia PDF Downloads 305
1916 Land Use/Land Cover Mapping Using Landsat 8 and Sentinel-2 in a Mediterranean Landscape

Authors: Moschos Vogiatzis, K. Perakis

Abstract:

Spatial-explicit and up-to-date land use/land cover information is fundamental for spatial planning, land management, sustainable development, and sound decision-making. In the last decade, many satellite-derived land cover products at different spatial, spectral, and temporal resolutions have been developed, such as the European Copernicus Land Cover product. However, more efficient and detailed information for land use/land cover is required at the regional or local scale. A typical Mediterranean basin with a complex landscape comprised of various forest types, crops, artificial surfaces, and wetlands was selected to test and develop our approach. In this study, we investigate the improvement of Copernicus Land Cover product (CLC2018) using Landsat 8 and Sentinel-2 pixel-based classification based on all available existing geospatial data (Forest Maps, LPIS, Natura2000 habitats, cadastral parcels, etc.). We examined and compared the performance of the Random Forest classifier for land use/land cover mapping. In total, 10 land use/land cover categories were recognized in Landsat 8 and 11 in Sentinel-2A. A comparison of the overall classification accuracies for 2018 shows that Landsat 8 classification accuracy was slightly higher than Sentinel-2A (82,99% vs. 80,30%). We concluded that the main land use/land cover types of CLC2018, even within a heterogeneous area, can be successfully mapped and updated according to CLC nomenclature. Future research should be oriented toward integrating spatiotemporal information from seasonal bands and spectral indexes in the classification process.

Keywords: classification, land use/land cover, mapping, random forest

Procedia PDF Downloads 110
1915 Predicting Options Prices Using Machine Learning

Authors: Krishang Surapaneni

Abstract:

The goal of this project is to determine how to predict important aspects of options, including the ask price. We want to compare different machine learning models to learn the best model and the best hyperparameters for that model for this purpose and data set. Option pricing is a relatively new field, and it can be very complicated and intimidating, especially to inexperienced people, so we want to create a machine learning model that can predict important aspects of an option stock, which can aid in future research. We tested multiple different models and experimented with hyperparameter tuning, trying to find some of the best parameters for a machine-learning model. We tested three different models: a Random Forest Regressor, a linear regressor, and an MLP (multi-layer perceptron) regressor. The most important feature in this experiment is the ask price; this is what we were trying to predict. In the field of stock pricing prediction, there is a large potential for error, so we are unable to determine the accuracy of the models based on if they predict the pricing perfectly. Due to this factor, we determined the accuracy of the model by finding the average percentage difference between the predicted and actual values. We tested the accuracy of the machine learning models by comparing the actual results in the testing data and the predictions made by the models. The linear regression model performed worst, with an average percentage error of 17.46%. The MLP regressor had an average percentage error of 11.45%, and the random forest regressor had an average percentage error of 7.42%

Keywords: finance, linear regression model, machine learning model, neural network, stock price

Procedia PDF Downloads 62
1914 High Voltage Magnetic Pulse Generation Using Capacitor Discharge Technique

Authors: Mohamed Adel Abdallah

Abstract:

A high voltage magnetic pulse is designed by applying an electrical pulse to the coil. Capacitor banks are developed to generate a pulse current. Switching circuit consisting of DPDT switches, thyristor, and triggering circuit is built and tested. The coil current is measured using a Hall-effect current sensor. The magnetic pulse created is measured and tabulated in the graph. Simulation using FEMM is done to compare the results obtained between experiment and simulation. This technology can be applied to area such as medical equipment, measuring instrument, and military equipment.

Keywords: high voltage, magnetic pulse, capacitor discharge, coil

Procedia PDF Downloads 661
1913 Implementation of an Associative Memory Using a Restricted Hopfield Network

Authors: Tet H. Yeap

Abstract:

An analog restricted Hopfield Network is presented in this paper. It consists of two layers of nodes, visible and hidden nodes, connected by directional weighted paths forming a bipartite graph with no intralayer connection. An energy or Lyapunov function was derived to show that the proposed network will converge to stable states. By introducing hidden nodes, the proposed network can be trained to store patterns and has increased memory capacity. Training to be an associative memory, simulation results show that the associative memory performs better than a classical Hopfield network by being able to perform better memory recall when the input is noisy.

Keywords: restricted Hopfield network, Lyapunov function, simultaneous perturbation stochastic approximation

Procedia PDF Downloads 113
1912 Robust Stabilization against Unknown Consensus Network

Authors: Myung-Gon Yoon, Jung-Ho Moon, Tae Kwon Ha

Abstract:

This paper considers a robust stabilization problem of a single agent in a multi-agent consensus system composed of identical agents, when the network topology of the system is completely unknown. It is shown that the transfer function of an agent in a consensus system can be described as a multiplicative perturbation of the isolated agent transfer function in frequency domain. Applying known robust stabilization results, we present sufficient conditions for a robust stabilization of an agent against unknown network topology.

Keywords: single agent control, multi-agent system, transfer function, graph angle

Procedia PDF Downloads 434
1911 Ontology-Based Systemizing of the Science Information Devoted to Waste Utilizing by Methanogenesis

Authors: Ye. Shapovalov, V. Shapovalov, O. Stryzhak, A. Salyuk

Abstract:

Over the past decades, amount of scientific information has been growing exponentially. It became more complicated to process and systemize this amount of data. The approach to systematization of scientific information on the production of biogas based on the ontological IT platform “T.O.D.O.S.” has been developed. It has been proposed to select semantic characteristics of each work for their further introduction into the IT platform “T.O.D.O.S.”. An ontological graph with a ranking function for previous scientific research and for a system of selection of microorganisms has been worked out. These systems provide high performance of information management of scientific information.

Keywords: ontology-based analysis, analysis of scientific data, methanogenesis, microorganism hierarchy, 'T.O.D.O.S.'

Procedia PDF Downloads 142
1910 A Machine Learning Approach for Earthquake Prediction in Various Zones Based on Solar Activity

Authors: Viacheslav Shkuratskyy, Aminu Bello Usman, Michael O’Dea, Saifur Rahman Sabuj

Abstract:

This paper examines relationships between solar activity and earthquakes; it applied machine learning techniques: K-nearest neighbour, support vector regression, random forest regression, and long short-term memory network. Data from the SILSO World Data Center, the NOAA National Center, the GOES satellite, NASA OMNIWeb, and the United States Geological Survey were used for the experiment. The 23rd and 24th solar cycles, daily sunspot number, solar wind velocity, proton density, and proton temperature were all included in the dataset. The study also examined sunspots, solar wind, and solar flares, which all reflect solar activity and earthquake frequency distribution by magnitude and depth. The findings showed that the long short-term memory network model predicts earthquakes more correctly than the other models applied in the study, and solar activity is more likely to affect earthquakes of lower magnitude and shallow depth than earthquakes of magnitude 5.5 or larger with intermediate depth and deep depth.

Keywords: k-nearest neighbour, support vector regression, random forest regression, long short-term memory network, earthquakes, solar activity, sunspot number, solar wind, solar flares

Procedia PDF Downloads 58
1909 Digital Forgery Detection by Signal Noise Inconsistency

Authors: Bo Liu, Chi-Man Pun

Abstract:

A novel technique for digital forgery detection by signal noise inconsistency is proposed in this paper. The forged area spliced from the other picture contains some features which may be inconsistent with the rest part of the image. Noise pattern and the level is a possible factor to reveal such inconsistency. To detect such noise discrepancies, the test picture is initially segmented into small pieces. The noise pattern and level of each segment are then estimated by using various filters. The noise features constructed in this step are utilized in energy-based graph cut to expose forged area in the final step. Experimental results show that our method provides a good illustration of regions with noise inconsistency in various scenarios.

Keywords: forgery detection, splicing forgery, noise estimation, noise

Procedia PDF Downloads 442
1908 A Comparative Study of Sampling-Based Uncertainty Propagation with First Order Error Analysis and Percentile-Based Optimization

Authors: M. Gulam Kibria, Shourav Ahmed, Kais Zaman

Abstract:

In system analysis, the information on the uncertain input variables cause uncertainty in the system responses. Different probabilistic approaches for uncertainty representation and propagation in such cases exist in the literature. Different uncertainty representation approaches result in different outputs. Some of the approaches might result in a better estimation of system response than the other approaches. The NASA Langley Multidisciplinary Uncertainty Quantification Challenge (MUQC) has posed challenges about uncertainty quantification. Subproblem A, the uncertainty characterization subproblem, of the challenge posed is addressed in this study. In this subproblem, the challenge is to gather knowledge about unknown model inputs which have inherent aleatory and epistemic uncertainties in them with responses (output) of the given computational model. We use two different methodologies to approach the problem. In the first methodology we use sampling-based uncertainty propagation with first order error analysis. In the other approach we place emphasis on the use of Percentile-Based Optimization (PBO). The NASA Langley MUQC’s subproblem A is developed in such a way that both aleatory and epistemic uncertainties need to be managed. The challenge problem classifies each uncertain parameter as belonging to one the following three types: (i) An aleatory uncertainty modeled as a random variable. It has a fixed functional form and known coefficients. This uncertainty cannot be reduced. (ii) An epistemic uncertainty modeled as a fixed but poorly known physical quantity that lies within a given interval. This uncertainty is reducible. (iii) A parameter might be aleatory but sufficient data might not be available to adequately model it as a single random variable. For example, the parameters of a normal variable, e.g., the mean and standard deviation, might not be precisely known but could be assumed to lie within some intervals. It results in a distributional p-box having the physical parameter with an aleatory uncertainty, but the parameters prescribing its mathematical model are subjected to epistemic uncertainties. Each of the parameters of the random variable is an unknown element of a known interval. This uncertainty is reducible. From the study, it is observed that due to practical limitations or computational expense, the sampling is not exhaustive in sampling-based methodology. That is why the sampling-based methodology has high probability of underestimating the output bounds. Therefore, an optimization-based strategy to convert uncertainty described by interval data into a probabilistic framework is necessary. This is achieved in this study by using PBO.

Keywords: aleatory uncertainty, epistemic uncertainty, first order error analysis, uncertainty quantification, percentile-based optimization

Procedia PDF Downloads 224
1907 A Methodology for the Synthesis of Multi-Processors

Authors: Hamid Yasinian

Abstract:

Random epistemologies and hash tables have garnered minimal interest from both security experts and experts in the last several years. In fact, few information theorists would disagree with the evaluation of expert systems. In our research, we discover how flip-flop gates can be applied to the study of superpages. Though such a hypothesis at first glance seems perverse, it is derived from known results.

Keywords: synthesis, multi-processors, interactive model, moor’s law

Procedia PDF Downloads 417
1906 Peak Frequencies in the Collective Membrane Potential of a Hindmarsh-Rose Small-World Neural Network

Authors: Sun Zhe, Ruggero Micheletto

Abstract:

As discussed extensively in many studies, noise in neural networks have an important role in the functioning and time evolution of the system. The mechanism by which noise induce stochastic resonance enhancing and influencing certain operations is not clarified nor is the mechanism of information storage and coding. With the present research we want to study the role of noise, especially focusing on the frequency peaks in a three variable Hindmarsh−Rose Small−World network. We investigated the behaviour of the network to external noises. We demonstrate that a variation of signal to noise ratio of about 10 dB induces an increase in membrane potential signal of about 15%, averaged over the whole network. We also considered the integral of the whole membrane potential as a paradigm of internal noise, the one generated by the brain network. We showed that this internal noise is attenuated with the size of the network or with the number of random connections. By means of Fourier analysis we found that it has distinct peaks of frequencies, moreover, we showed that increasing the size of the network introducing more neurons, reduced the maximum frequencies generated by the network, whereas the increase in the number of random connections (determined by the small-world probability p) led to a trend toward higher frequencies. This study may give clues on how networks utilize noise to alter the collective behaviour of the system in their operations.

Keywords: neural networks, stochastic processes, small-world networks, discrete Fourier analysis

Procedia PDF Downloads 277
1905 Quantum Inspired Security on a Mobile Phone

Authors: Yu Qin, Wanjiaman Li

Abstract:

The widespread use of mobile electronic devices increases the complexities of mobile security. This thesis aims to provide a secure communication environment for smartphone users. Some research proves that the one-time pad is one of the securest encryption methods, and that the key distribution problem can be solved by using the QKD (quantum key distribution). The objective of this project is to design an Android APP (application) to exchange several random keys between mobile phones. Inspired by QKD, the developed APP uses the quick response (QR) code as a carrier to dispatch large amounts of one-time keys. After evaluating the performance of APP, it allows the mobile phone to capture and decode 1800 bytes of random data in 600ms. The continuous scanning mode of APP is designed to improve the overall transmission performance and user experience, and the maximum transmission rate of this mode is around 2200 bytes/s. The omnidirectional readability and error correction capability of QR code gives it a better real-life application, and the features of adequate storage capacity and quick response optimize overall transmission efficiency. The security of this APP is guaranteed since QR code is exchanged face-to-face, eliminating the risk of being eavesdropped. Also, the id of QR code is the only message that would be transmitted through the whole communication. The experimental results show this project can achieve superior transmission performance, and the correlation between the transmission rate of the system and several parameters, such as the QR code size, has been analyzed. In addition, some existing technologies and the main findings in the context of the project are summarized and critically compared in detail.

Keywords: one-time pad, QKD (quantum key distribution), QR code, application

Procedia PDF Downloads 131
1904 A Reduced Distributed Sate Space for Modular Petri Nets

Authors: Sawsen Khlifa, Chiheb AMeur Abid, Belhassan Zouari

Abstract:

Modular verification approaches have been widely attempted to cope with the well known state explosion problem. This paper deals with the modular verification of modular Petri nets. We propose a reduced version for the modular state space of a given modular Petri net. The new structure allows the creation of smaller modular graphs. Each one draws the behavior of the corresponding module and outlines some global information. Hence, this version helps to overcome the explosion problem and to use less memory space. In this condensed structure, the verification of some generic properties concerning one module is limited to the exploration of its associated graph.

Keywords: distributed systems, modular verification, petri nets, state space explosition

Procedia PDF Downloads 95