Search results for: nullity of graphs
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 331

Search results for: nullity of graphs

271 Measurement and Analysis of Building Penetration Loss for Mobile Networks in Tripoli Area

Authors: Tammam A. Benmusa, Mohamed A. Shlibek, Rawad M. Swesi

Abstract:

The investigation of Buildings Penetration Loss (BPL) of radio signal is getting more and more important. It plays an important role in calculating the indoor coverage for wireless communication networks. In this paper, the theory behind BPL and its mechanisms have been reviewed. The operating frequency, coverage area type, climate condition, time of measurement, and other factors affecting the values of BPL have been discussed. The practical part of this work was conducting 4000 measurements of BPL in different areas in the Libyan capital, Tripoli, to get empirical model for this loss. The measurements were taken for 2 different types of wireless communication networks; mobile telephone network (for Almadar company), which operates at 900 MHz and WiMAX network (LTT company) which operates at 2500 MHz. The results for each network were summarized and presented in several graphs. The graphs are showing how the BPL affected by: time of measurement, morphology (type of area), and climatic environment.

Keywords: building penetration loss, wireless network, mobile network, link budget, indoor network performance

Procedia PDF Downloads 384
270 CMOS Positive and Negative Resistors Based on Complementary Regulated Cascode Topology with Cross-Coupled Regulated Transistors

Authors: Kittipong Tripetch, Nobuhiko Nakano

Abstract:

Two types of floating active resistors based on a complementary regulated cascode topology with cross-coupled regulated transistors are presented in this paper. The first topology is a high swing complementary regulated cascode active resistor. The second topology is a complementary common gate with a regulated cross coupled transistor. The small-signal input resistances of the floating resistors are derived. Three graphs of the input current versus the input voltage for different aspect ratios are designed and plotted using the Cadence Spectre 0.18-µm Rohm Semiconductor process. The total harmonic distortion graphs are plotted for three different aspect ratios with different input-voltage amplitudes and different input frequencies. From the simulation results, it is observed that a resistance of approximately 8.52 MΩ can be obtained from supply voltage at  ±0.9 V.

Keywords: floating active resistor, complementary common gate, complementary regulated cascode, current mirror

Procedia PDF Downloads 259
269 Some Codes for Variants in Graphs

Authors: Sofia Ait Bouazza

Abstract:

We consider the problem of finding a minimum identifying code in a graph. This problem was initially introduced in 1998 and has been since fundamentally connected to a wide range of applications (fault diagnosis, location detection …). Suppose we have a building into which we need to place fire alarms. Suppose each alarm is designed so that it can detect any fire that starts either in the room in which it is located or in any room that shares a doorway with the room. We want to detect any fire that may occur or use the alarms which are sounding to not only to not only detect any fire but be able to tell exactly where the fire is located in the building. For reasons of cost, we want to use as few alarms as necessary. The first problem involves finding a minimum domination set of a graph. If the alarms are three state alarms capable of distinguishing between a fire in the same room as the alarm and a fire in an adjacent room, we are trying to find a minimum locating domination set. If the alarms are two state alarms that can only sound if there is a fire somewhere nearby, we are looking for a differentiating domination set of a graph. These three areas are the subject of much active research; we primarily focus on the third problem. An identifying code of a graph G is a dominating set C such that every vertex x of G is distinguished from other vertices by the set of vertices in C that are at distance at most r≥1 from x. When only vertices out of the code are asked to be identified, we get the related concept of a locating dominating set. The problem of finding an identifying code (resp a locating dominating code) of minimum size is a NP-hard problem, even when the input graph belongs to a number of specific graph classes. Therefore, we study this problem in some restricted classes of undirected graphs like split graph, line graph and path in a directed graph. Then we present some results on the identifying code by giving an exact value of upper total locating domination and a total 2-identifying code in directed and undirected graph. Moreover we determine exact values of locating dominating code and edge identifying code of thin headless spider and locating dominating code of complete suns.

Keywords: identiying codes, locating dominating set, split graphs, thin headless spider

Procedia PDF Downloads 478
268 Visualizing the Commercial Activity of a City by Analyzing the Data Information in Layers

Authors: Taras Agryzkov, Jose L. Oliver, Leandro Tortosa, Jose Vicent

Abstract:

This paper aims to demonstrate how network models can be used to understand and to deal with some aspects of urban complexity. As it is well known, the Theory of Architecture and Urbanism has been using for decades’ intellectual tools based on the ‘sciences of complexity’ as a strategy to propose theoretical approaches about cities and about architecture. In this sense, it is possible to find a vast literature in which for instance network theory is used as an instrument to understand very diverse questions about cities: from their commercial activity to their heritage condition. The contribution of this research consists in adding one step of complexity to this process: instead of working with one single primal graph as it is usually done, we will show how new network models arise from the consideration of two different primal graphs interacting in two layers. When we model an urban network through a mathematical structure like a graph, the city is usually represented by a set of nodes and edges that reproduce its topology, with the data generated or extracted from the city embedded in it. All this information is normally displayed in a single layer. Here, we propose to separate the information in two layers so that we can evaluate the interaction between them. Besides, both layers may be composed of structures that do not have to coincide: from this bi-layer system, groups of interactions emerge, suggesting reflections and in consequence, possible actions.

Keywords: graphs, mathematics, networks, urban studies

Procedia PDF Downloads 180
267 Stock Market Prediction Using Convolutional Neural Network That Learns from a Graph

Authors: Mo-Se Lee, Cheol-Hwi Ahn, Kee-Young Kwahk, Hyunchul Ahn

Abstract:

Over the past decade, deep learning has been in spotlight among various machine learning algorithms. In particular, CNN (Convolutional Neural Network), which is known as effective solution for recognizing and classifying images, has been popularly applied to classification and prediction problems in various fields. In this study, we try to apply CNN to stock market prediction, one of the most challenging tasks in the machine learning research. In specific, we propose to apply CNN as the binary classifier that predicts stock market direction (up or down) by using a graph as its input. That is, our proposal is to build a machine learning algorithm that mimics a person who looks at the graph and predicts whether the trend will go up or down. Our proposed model consists of four steps. In the first step, it divides the dataset into 5 days, 10 days, 15 days, and 20 days. And then, it creates graphs for each interval in step 2. In the next step, CNN classifiers are trained using the graphs generated in the previous step. In step 4, it optimizes the hyper parameters of the trained model by using the validation dataset. To validate our model, we will apply it to the prediction of KOSPI200 for 1,986 days in eight years (from 2009 to 2016). The experimental dataset will include 14 technical indicators such as CCI, Momentum, ROC and daily closing price of KOSPI200 of Korean stock market.

Keywords: convolutional neural network, deep learning, Korean stock market, stock market prediction

Procedia PDF Downloads 425
266 Investigation of the Role of Friction in Reducing Pedestrian Injuries in Accidents at Intersections

Authors: Seyed Abbas Tabatabaei, Afshin Ghanbarzadeh, Mehdi Abidizadeh

Abstract:

Nowadays the subject of road traffic accidents and the high social and economic costs due to them is the most fundamental problem that experts and providers of transport and traffic brought to a challenge. One of the most effective measures is to enhance the skid resistance of road surface. This research aims to study the intersection of one case in Ahwaz and the effect of increasing the skid resistance in reducing pedestrian injuries in accidents at intersections. In this research the device was developed to measure the coefficient of friction and tried the rules and practices of it have a high similarity with the Locked Wheel Trailer. This device includes a steel frame, wheels, hydration systems, and force gauge. The output of the device is that the force gauge registers. By investigate this data and applying the relationships relative surface coefficient of friction is obtained. Friction coefficient data for the current state and the state of the new pavement are obtained and plotted on the graphs based on the graphs we can compare the two situations and speed at the moment of collision between the two modes are compared. The results show that increasing the coefficient of friction to what extent can be effective on the severity and number of accidents.

Keywords: intersection, coefficient of friction, skid resistance, locked wheels, accident, pedestrian

Procedia PDF Downloads 328
265 Kinetic Study of the Esterification of Unsaturated Fatty Acids from Salmon Oil (Salmosalar L.)

Authors: André Luis Lima de Oliveira, Vera Lúcia Viana do Nascimento, Victória Maura Silva Bermudez, Mauricio Nunes Kleinberg, João Carlos da Costa Assunção, José Osvaldo Beserra Carioca

Abstract:

The objective of this study was to synthesize a triglyceride with high content of unsaturated fatty acids from salmon oil (Salmo salar L.) by esterification with glycerol catalyzed dealuminized zeolite. A kinetic study was conducted to determine the reaction order and the activation energy. A statistical study was conducted to determine optimal reaction conditions. Initially, the crude oil was refined salmon physically and chemically. The crude oil was hydrolyzed and unsaturated free fatty acids were separated by urea complexation method. An experimental project to verify the parameters (temperature, glycerin and catalyst) with the greatest impact on the reaction was developed. In experiments aliquots were taken at predetermined times to measure the amount of free fatty acids. Pareto, surface, contour and hub graphs were used to determine the factors that maximized the reaction. According to the graphs the best reaction conditions were: temperature 80 ° C, the proportion glycerine/oil 5: 1 and 1% of catalyst. The kinetic data showed that the system was compatible with a second-order reaction. After analyzing the rate constant versus temperature charts a value of 85.31 kJ/mol was obtained for the reaction activation energy.

Keywords: esterification, kinect, oil, salmon

Procedia PDF Downloads 521
264 Stress Concentration Trend for Combined Loading Conditions

Authors: Aderet M. Pantierer, Shmuel Pantierer, Raphael Cordina, Yougashwar Budhoo

Abstract:

Stress concentration occurs when there is an abrupt change in geometry, a mechanical part under loading. These changes in geometry can include holes, notches, or cracks within the component. The modifications create larger stress within the part. This maximum stress is difficult to determine, as it is directly at the point of the minimum area. Strain gauges have yet to be developed to analyze stresses at such minute areas. Therefore, a stress concentration factor must be utilized. The stress concentration factor is a dimensionless parameter calculated solely on the geometry of a part. The factor is multiplied by the nominal, or average, stress of the component, which can be found analytically or experimentally. Stress concentration graphs exist for common loading conditions and geometrical configurations to aid in the determination of the maximum stress a part can withstand. These graphs were developed from historical data yielded from experimentation. This project seeks to verify a stress concentration graph for combined loading conditions. The aforementioned graph was developed using CATIA Finite Element Analysis software. The results of this analysis will be validated through further testing. The 3D modeled parts will be subjected to further finite element analysis using Patran-Nastran software. The finite element models will then be verified by testing physical specimen using a tensile testing machine. Once the data is validated, the unique stress concentration graph will be submitted for publication so it can aid engineers in future projects.

Keywords: stress concentration, finite element analysis, finite element models, combined loading

Procedia PDF Downloads 443
263 Mathematical Toolbox for editing Equations and Geometrical Diagrams and Graphs

Authors: Ayola D. N. Jayamaha, Gihan V. Dias, Surangika Ranathunga

Abstract:

Currently there are lot of educational tools designed for mathematics. Open source software such as GeoGebra and Octave are bulky in their architectural structure. In addition, there is MathLab software, which facilitates much more than what we ask for. Many of the computer aided online grading and assessment tools require integrating editors to their software. However, there are not exist suitable editors that cater for all their needs in editing equations and geometrical diagrams and graphs. Some of the existing software for editing equations is Alfred’s Equation Editor, Codecogs, DragMath, Maple, MathDox, MathJax, MathMagic, MathFlow, Math-o-mir, Microsoft Equation Editor, MiraiMath, OpenOffice, WIRIS Editor and MyScript. Some of them are commercial, open source, supports handwriting recognition, mobile apps, renders MathML/LaTeX, Flash / Web based and javascript display engines. Some of the diagram editors are GeoKone.NET, Tabulae, Cinderella 1.4, MyScript, Dia, Draw2D touch, Gliffy, GeoGebra, Flowchart, Jgraph, JointJS, J painter Online diagram editor and 2D sketcher. All these software are open source except for MyScript and can be used for editing mathematical diagrams. However, they do not fully cater the needs of a typical computer aided assessment tool or Educational Platform for Mathematics. This solution provides a Web based, lightweight, easy to implement and integrate solution of an html5 canvas that renders on all of the modern web browsers. The scope of the project is an editor that covers equations and mathematical diagrams and drawings on the O/L Mathematical Exam Papers in Sri Lanka. Using the tool the students can enter any equation to the system which can be on an online remote learning platform. The users can also create and edit geometrical drawings, graphs and do geometrical constructions that require only Compass and Ruler from the Editing Interface provided by the Software. The special feature of this software is the geometrical constructions. It allows the users to create geometrical constructions such as angle bisectors, perpendicular lines, angles of 600 and perpendicular bisectors. The tool correctly imitates the functioning of rulers and compasses to create the required geometrical construction. Therefore, the users are able to do geometrical drawings on the computer successfully and we have a digital format of the geometrical drawing for further processing. Secondly, we can create and edit Venn Diagrams, color them and label them. In addition, the students can draw probability tree diagrams and compound probability outcome grids. They can label and mark regions within the grids. Thirdly, students can draw graphs (1st order and 2nd order). They can mark points on a graph paper and the system connects the dots to draw the graph. Further students are able to draw standard shapes such as circles and rectangles by selecting points on a grid or entering the parametric values.

Keywords: geometrical drawings, html5 canvas, mathematical equations, toolbox

Procedia PDF Downloads 376
262 Study of Storms on the Javits Center Green Roof

Authors: Alexander Cho, Harsho Sanyal, Joseph Cataldo

Abstract:

A quantitative analysis of the different variables on both the South and North green roofs of the Jacob K. Javits Convention Center was taken to find mathematical relationships between net radiation and evapotranspiration (ET), average outside temperature, and the lysimeter weight. Groups of datasets were analyzed, and the relationships were plotted on linear and semi-log graphs to find consistent relationships. Antecedent conditions for each rainstorm were also recorded and plotted against the volumetric water difference within the lysimeter. The first relation was the inverse parabolic relationship between the lysimeter weight and the net radiation and ET. The peaks and valleys of the lysimeter weight corresponded to valleys and peaks in the net radiation and ET respectively, with the 8/22/15 and 1/22/16 datasets showing this trend. The U-shaped and inverse U-shaped plots of the two variables coincided, indicating an inverse relationship between the two variables. Cross variable relationships were examined through graphs with lysimeter weight as the dependent variable on the y-axis. 10 out of 16 of the plots of lysimeter weight vs. outside temperature plots had R² values > 0.9. Antecedent conditions were also recorded for rainstorms, categorized by the amount of precipitation accumulating during the storm. Plotted against the change in the volumetric water weight difference within the lysimeter, a logarithmic regression was found with large R² values. The datasets were compared using the Mann Whitney U-test to see if the datasets were statistically different, using a significance level of 5%; all datasets compared showed a U test statistic value, proving the null hypothesis of the datasets being different from being true.

Keywords: green roof, green infrastructure, Javits Center, evapotranspiration, net radiation, lysimeter

Procedia PDF Downloads 114
261 Code Embedding for Software Vulnerability Discovery Based on Semantic Information

Authors: Joseph Gear, Yue Xu, Ernest Foo, Praveen Gauravaran, Zahra Jadidi, Leonie Simpson

Abstract:

Deep learning methods have been seeing an increasing application to the long-standing security research goal of automatic vulnerability detection for source code. Attention, however, must still be paid to the task of producing vector representations for source code (code embeddings) as input for these deep learning models. Graphical representations of code, most predominantly Abstract Syntax Trees and Code Property Graphs, have received some use in this task of late; however, for very large graphs representing very large code snip- pets, learning becomes prohibitively computationally expensive. This expense may be reduced by intelligently pruning this input to only vulnerability-relevant information; however, little research in this area has been performed. Additionally, most existing work comprehends code based solely on the structure of the graph at the expense of the information contained by the node in the graph. This paper proposes Semantic-enhanced Code Embedding for Vulnerability Discovery (SCEVD), a deep learning model which uses semantic-based feature selection for its vulnerability classification model. It uses information from the nodes as well as the structure of the code graph in order to select features which are most indicative of the presence or absence of vulnerabilities. This model is implemented and experimentally tested using the SARD Juliet vulnerability test suite to determine its efficacy. It is able to improve on existing code graph feature selection methods, as demonstrated by its improved ability to discover vulnerabilities.

Keywords: code representation, deep learning, source code semantics, vulnerability discovery

Procedia PDF Downloads 158
260 Estimation of the Mean of the Selected Population

Authors: Kalu Ram Meena, Aditi Kar Gangopadhyay, Satrajit Mandal

Abstract:

Two normal populations with different means and same variance are considered, where the variances are known. The population with the smaller sample mean is selected. Various estimators are constructed for the mean of the selected normal population. Finally, they are compared with respect to the bias and MSE risks by the method of Monte-Carlo simulation and their performances are analysed with the help of graphs.

Keywords: estimation after selection, Brewster-Zidek technique, estimators, selected populations

Procedia PDF Downloads 512
259 On the Existence of Homotopic Mapping Between Knowledge Graphs and Graph Embeddings

Authors: Jude K. Safo

Abstract:

Knowledge Graphs KG) and their relation to Graph Embeddings (GE) represent a unique data structure in the landscape of machine learning (relative to image, text and acoustic data). Unlike the latter, GEs are the only data structure sufficient for representing hierarchically dense, semantic information needed for use-cases like supply chain data and protein folding where the search space exceeds the limits traditional search methods (e.g. page-rank, Dijkstra, etc.). While GEs are effective for compressing low rank tensor data, at scale, they begin to introduce a new problem of ’data retreival’ which we observe in Large Language Models. Notable attempts by transE, TransR and other prominent industry standards have shown a peak performance just north of 57% on WN18 and FB15K benchmarks, insufficient practical industry applications. They’re also limited, in scope, to next node/link predictions. Traditional linear methods like Tucker, CP, PARAFAC and CANDECOMP quickly hit memory limits on tensors exceeding 6.4 million nodes. This paper outlines a topological framework for linear mapping between concepts in KG space and GE space that preserve cardinality. Most importantly we introduce a traceable framework for composing dense linguistic strcutures. We demonstrate performance on WN18 benchmark this model hits. This model does not rely on Large Langauge Models (LLM) though the applications are certainy relevant here as well.

Keywords: representation theory, large language models, graph embeddings, applied algebraic topology, applied knot theory, combinatorics

Procedia PDF Downloads 68
258 Topological Language for Classifying Linear Chord Diagrams via Intersection Graphs

Authors: Michela Quadrini

Abstract:

Chord diagrams occur in mathematics, from the study of RNA to knot theory. They are widely used in theory of knots and links for studying the finite type invariants, whereas in molecular biology one important motivation to study chord diagrams is to deal with the problem of RNA structure prediction. An RNA molecule is a linear polymer, referred to as the backbone, that consists of four types of nucleotides. Each nucleotide is represented by a point, whereas each chord of the diagram stands for one interaction for Watson-Crick base pairs between two nonconsecutive nucleotides. A chord diagram is an oriented circle with a set of n pairs of distinct points, considered up to orientation preserving diffeomorphisms of the circle. A linear chord diagram (LCD) is a special kind of graph obtained cutting the oriented circle of a chord diagram. It consists of a line segment, called its backbone, to which are attached a number of chords with distinct endpoints. There is a natural fattening on any linear chord diagram; the backbone lies on the real axis, while all the chords are in the upper half-plane. Each linear chord diagram has a natural genus of its associated surface. To each chord diagram and linear chord diagram, it is possible to associate the intersection graph. It consists of a graph whose vertices correspond to the chords of the diagram, whereas the chord intersections are represented by a connection between the vertices. Such intersection graph carries a lot of information about the diagram. Our goal is to define an LCD equivalence class in terms of identity of intersection graphs, from which many chord diagram invariants depend. For studying these invariants, we introduce a new representation of Linear Chord Diagrams based on a set of appropriate topological operators that permits to model LCD in terms of the relations among chords. Such set is composed of: crossing, nesting, and concatenations. The crossing operator is able to generate the whole space of linear chord diagrams, and a multiple context free grammar able to uniquely generate each LDC starting from a linear chord diagram adding a chord for each production of the grammar is defined. In other words, it allows to associate a unique algebraic term to each linear chord diagram, while the remaining operators allow to rewrite the term throughout a set of appropriate rewriting rules. Such rules define an LCD equivalence class in terms of the identity of intersection graphs. Starting from a modelled RNA molecule and the linear chord, some authors proposed a topological classification and folding. Our LCD equivalence class could contribute to the RNA folding problem leading to the definition of an algorithm that calculates the free energy of the molecule more accurately respect to the existing ones. Such LCD equivalence class could be useful to obtain a more accurate estimate of link between the crossing number and the topological genus and to study the relation among other invariants.

Keywords: chord diagrams, linear chord diagram, equivalence class, topological language

Procedia PDF Downloads 201
257 Effects of Duct Geometry, Thickness and Types of Liners on Transmission Loss for Absorptive Silencers

Authors: M. Kashfi, K. Jahani

Abstract:

Sound attenuation in absorptive silencers has been analyzed in this paper. The structure of such devices is as follows. When the rigid duct of an expansion chamber has been lined by a packed absorptive material under a perforated membrane, incident sound waves will be dissipated by the absorptive liners. This kind of silencer, usually are applicable for medium to high frequency ranges. Several conditions for different absorptive materials, variety in their thicknesses, and different shapes of the expansion chambers have been studied in this paper. Also, graphs of sound attenuation have been compared between empty expansion chamber and duct of silencer with applying liner. Plane waves have been assumed in inlet and outlet regions of the silencer. Presented results that have been achieved by applying finite element method (FEM), have shown the dependence of the sound attenuation spectrum to flow resistivity and the thicknesses of the absorptive materials, and geometries of the cross section (configuration of the silencer). As flow resistivity and thickness of absorptive materials increase, sound attenuation improves. In this paper, diagrams of the transmission loss (TL) for absorptive silencers in five different cross sections (rectangle, circle, ellipse, square, and rounded rectangle as the main geometry) have been presented. Also, TL graphs for silencers using different absorptive material (glass wool, wood fiber, and kind of spongy materials) as liner with three different thicknesses of 5 mm, 15 mm, and 30 mm for glass wool liner have been exhibited. At first, the effect of substances of the absorptive materials with the specific flow resistivity and densities on the TL spectrum, then the effect of the thicknesses of the glass wool, and at last the efficacy of the shape of the cross section of the silencer have been investigated.

Keywords: transmission loss, absorptive material, flow resistivity, thickness, frequency

Procedia PDF Downloads 250
256 Probabilistic Graphical Model for the Web

Authors: M. Nekri, A. Khelladi

Abstract:

The world wide web network is a network with a complex topology, the main properties of which are the distribution of degrees in power law, A low clustering coefficient and a weak average distance. Modeling the web as a graph allows locating the information in little time and consequently offering a help in the construction of the research engine. Here, we present a model based on the already existing probabilistic graphs with all the aforesaid characteristics. This work will consist in studying the web in order to know its structuring thus it will enable us to modelize it more easily and propose a possible algorithm for its exploration.

Keywords: clustering coefficient, preferential attachment, small world, web community

Procedia PDF Downloads 272
255 Identifying Coloring in Graphs with Twins

Authors: Souad Slimani, Sylvain Gravier, Simon Schmidt

Abstract:

Recently, several vertex identifying notions were introduced (identifying coloring, lid-coloring,...); these notions were inspired by identifying codes. All of them, as well as original identifying code, is based on separating two vertices according to some conditions on their closed neighborhood. Therefore, twins can not be identified. So most of known results focus on twin-free graph. Here, we show how twins can modify optimal value of vertex-identifying parameters for identifying coloring and locally identifying coloring.

Keywords: identifying coloring, locally identifying coloring, twins, separating

Procedia PDF Downloads 148
254 Comparative Review of Models for Forecasting Permanent Deformation in Unbound Granular Materials

Authors: Shamsulhaq Amin

Abstract:

Unbound granular materials (UGMs) are pivotal in ensuring long-term quality, especially in the layers under the surface of flexible pavements and other constructions. This study seeks to better understand the behavior of the UGMs by looking at popular models for predicting lasting deformation under various levels of stresses and load cycles. These models focus on variables such as the number of load cycles, stress levels, and features specific to materials and were evaluated on the basis of their ability to accurately predict outcomes. The study showed that these factors play a crucial role in how well the models work. Therefore, the research highlights the need to look at a wide range of stress situations to more accurately predict how much the UGMs bend or shift. The research looked at important factors, like how permanent deformation relates to the number of times a load is applied, how quickly this phenomenon happens, and the shakedown effect, in two different types of UGMs: granite and limestone. A detailed study was done over 100,000 load cycles, which provided deep insights into how these materials behave. In this study, a number of factors, such as the level of stress applied, the number of load cycles, the density of the material, and the moisture present were seen as the main factors affecting permanent deformation. It is vital to fully understand these elements for better designing pavements that last long and handle wear and tear. A series of laboratory tests were performed to evaluate the mechanical properties of materials and acquire model parameters. The testing included gradation tests, CBR tests, and Repeated load triaxial tests. The repeated load triaxial tests were crucial for studying the significant components that affect deformation. This test involved applying various stress levels to estimate model parameters. In addition, certain model parameters were established by regression analysis, and optimization was conducted to improve outcomes. Afterward, the material parameters that were acquired were used to construct graphs for each model. The graphs were subsequently compared to the outcomes obtained from the repeated load triaxial testing. Additionally, the models were evaluated to determine if they demonstrated the two inherent deformation behaviors of materials when subjected to repetitive load: the initial phase, post-compaction, and the second phase volumetric changes. In this study, using log-log graphs was key to making the complex data easier to understand. This method made the analysis clearer and helped make the findings easier to interpret, adding both precision and depth to the research. This research provides important insight into picking the right models for predicting how these materials will act under expected stress and load conditions. Moreover, it offers crucial information regarding the effect of load cycle and permanent deformation as well as the shakedown effect on granite and limestone UGMs.

Keywords: permanent deformation, unbound granular materials, load cycles, stress level

Procedia PDF Downloads 38
253 Pairwise Relative Primality of Integers and Independent Sets of Graphs

Authors: Jerry Hu

Abstract:

Let G = (V, E) with V = {1, 2, ..., k} be a graph, the k positive integers a₁, a₂, ..., ak are G-wise relatively prime if (aᵢ, aⱼ ) = 1 for {i, j} ∈ E. We use an inductive approach to give an asymptotic formula for the number of k-tuples of integers that are G-wise relatively prime. An exact formula is obtained for the probability that k positive integers are G-wise relatively prime. As a corollary, we also provide an exact formula for the probability that k positive integers have exactly r relatively prime pairs.

Keywords: graph, independent set, G-wise relatively prime, probability

Procedia PDF Downloads 92
252 The Use of Network Theory in Heritage Cities

Authors: J. L. Oliver, T. Agryzkov, L. Tortosa, J. Vicent, J. Santacruz

Abstract:

This paper aims to demonstrate how the use of Network Theory can be applied to a very interesting and complex urban situation: The parts of a city which may have some patrimonial value, but because of their lack of relevant architectural elements, they are not considered to be historic in a conventional sense. In this paper, we use the suburb of La Villaflora in the city of Quito, Ecuador as our case study. We first propose a system of indicators as a tool to characterize and quantify the historic value of a geographic area. Then, we apply these indicators to the suburb of La Villaflora and use Network Theory to understand and propose actions.

Keywords: graphs, mathematics, networks, urban studies

Procedia PDF Downloads 369
251 A Graph Theoretic Algorithm for Bandwidth Improvement in Computer Networks

Authors: Mehmet Karaata

Abstract:

Given two distinct vertices (nodes) source s and target t of a graph G = (V, E), the two node-disjoint paths problem is to identify two node-disjoint paths between s ∈ V and t ∈ V . Two paths are node-disjoint if they have no common intermediate vertices. In this paper, we present an algorithm with O(m)-time complexity for finding two node-disjoint paths between s and t in arbitrary graphs where m is the number of edges. The proposed algorithm has a wide range of applications in ensuring reliability and security of sensor, mobile and fixed communication networks.

Keywords: disjoint paths, distributed systems, fault-tolerance, network routing, security

Procedia PDF Downloads 442
250 Topological Indices of Some Graph Operations

Authors: U. Mary

Abstract:

Let be a graph with a finite, nonempty set of objects called vertices together with a set of unordered pairs of distinct vertices of called edges. The vertex set is denoted by and the edge set by. Given two graphs and the wiener index of, wiener index for the splitting graph of a graph, the first Zagreb index of and its splitting graph, the 3-steiner wiener index of, the 3-steiner wiener index of a special graph are explored in this paper.

Keywords: complementary prism graph, first Zagreb index, neighborhood corona graph, steiner distance, splitting graph, steiner wiener index, wiener index

Procedia PDF Downloads 570
249 Regional Low Gravity Anomalies Influencing High Concentrations of Heavy Minerals on Placer Deposits

Authors: T. B. Karu Jayasundara

Abstract:

Regions of low gravity and gravity anomalies both influence heavy mineral concentrations on placer deposits. Economically imported heavy minerals are likely to have higher levels of deposition in low gravity regions of placer deposits. This can be found in coastal regions of Southern Asia, particularly in Sri Lanka and Peninsula India and areas located in the lowest gravity region of the world. The area about 70 kilometers of the east coast of Sri Lanka is covered by a high percentage of ilmenite deposits, and the southwest coast of the island consists of Monazite placer deposit. These deposits are one of the largest placer deposits in the world. In India, the heavy mineral industry has a good market. On the other hand, based on the coastal placer deposits recorded, the high gravity region located around Papua New Guinea, has no such heavy mineral deposits. In low gravity regions, with the help of other depositional environmental factors, the grains have more time and space to float in the sea, this helps bring high concentrations of heavy mineral deposits to the coast. The effect of low and high gravity can be demonstrated by using heavy mineral separation devices.  The Wilfley heavy mineral separating table is one of these; it is extensively used in industries and in laboratories for heavy mineral separation. The horizontally oscillating Wilfley table helps to separate heavy and light mineral grains in to deferent fractions, with the use of water. In this experiment, the low and high angle of the Wilfley table are representing low and high gravity respectively. A sample mixture of grain size <0.85 mm of heavy and light mineral grains has been used for this experiment. The high and low angle of the table was 60 and 20 respectively for this experiment. The separated fractions from the table are again separated into heavy and light minerals, with the use of heavy liquid, which consists of a specific gravity of 2.85. The fractions of separated heavy and light minerals have been used for drawing the two-dimensional graphs. The graphs show that the low gravity stage has a high percentage of heavy minerals collected in the upper area of the table than in the high gravity stage. The results of the experiment can be used for the comparison of regional low gravity and high gravity levels of heavy minerals. If there are any heavy mineral deposits in the high gravity regions, these deposits will take place far away from the coast, within the continental shelf.

Keywords: anomaly, gravity, influence, mineral

Procedia PDF Downloads 199
248 Developing a Structured Example Space for Finding the Collision Points of Functions and Their Inverse

Authors: M. Saeed, A. Shahidzadeh

Abstract:

Interaction between teachers and learners requires applying a set of samples (examples) which helps to create coordination between the goals and methods. The main result and achievement and application of samples (examples) are that they can bring the teacher and learner to a shared understanding of the concept. mathematical concepts, and also one of the challenging issues in the discussion of the function is to find the collision points of functions of and, regarding that the example space of teachers is different in this issue, this paper aims to present an example space including several problems of the secondary school with the help of intuition and drawing various graphs of functions of and for more familiarity of teachers.

Keywords: inverse function, educational example, Mathematic example, example space

Procedia PDF Downloads 179
247 Existence and Construction of Maximal Rectangular Duals

Authors: Krishnendra Shekhawat

Abstract:

Given a graph G = (V, E), a rectangular dual of G represents the vertices of G by a set of interior-disjoint rectangles such that two rectangles touch if and only if there is an edge between the two corresponding vertices in G. Rectangular duals do not exist for every graph, so we can define maximal rectangular duals. A maximal rectangular dual is a rectangular dual of a graph G such that there exists no graph G ′ with a rectangular dual where G is a subgraph of G ′. In this paper, we enumerate all maximal rectangular duals (or, to be precise, the corresponding planar graphs) up to six nodes and presents a necessary condition for the existence of a rectangular dual. This work allegedly has applications in integrated circuit design and architectural floor plans.

Keywords: adjacency, degree sequence, dual graph, rectangular dual

Procedia PDF Downloads 266
246 Survey Paper on Graph Coloring Problem and Its Application

Authors: Prateek Chharia, Biswa Bhusan Ghosh

Abstract:

Graph coloring is one of the prominent concepts in graph coloring. It can be defined as a coloring of the various regions of the graph such that all the constraints are fulfilled. In this paper various graphs coloring approaches like greedy coloring, Heuristic search for maximum independent set and graph coloring using edge table is described. Graph coloring can be used in various real time applications like student time tabling generation, Sudoku as a graph coloring problem, GSM phone network.

Keywords: graph coloring, greedy coloring, heuristic search, edge table, sudoku as a graph coloring problem

Procedia PDF Downloads 539
245 A Study of Families of Bistar and Corona Product of Graph: Reverse Topological Indices

Authors: Gowtham Kalkere Jayanna, Mohamad Nazri Husin

Abstract:

Graph theory, chemistry, and technology are all combined in cheminformatics. The structure and physiochemical properties of organic substances are linked using some useful graph invariants and the corresponding molecular graph. In this paper, we study specific reverse topological indices such as the reverse sum-connectivity index, the reverse Zagreb index, the reverse arithmetic-geometric, and the geometric-arithmetic, the reverse Sombor, the reverse Nirmala indices for the bistar graphs B (n: m) and the corona product Kₘ∘Kₙ', where Kₙ' Represent the complement of a complete graph Kₙ.

Keywords: reverse topological indices, bistar graph, the corona product, graph

Procedia PDF Downloads 96
244 Methodology for the Analysis of Energy Efficiency in Pneumatics Systems

Authors: Mario Lupaca, Karol Munoz, Victor De Negri

Abstract:

The present article presents a methodology for the improvement of the energy efficiency in pneumatic systems through the restoring of air. In this way, three techniques of expansion of a cylinder are identified: Expansion using the air of the compressor (conventional), restoring the air (efficient), and combining the air of the compressor and the restored air (hybrid). The methodology starts with the development of the GRAFCET of the system so that it can be decided whether to expand the cylinder in a conventional, efficient, or hybrid way. The methodology can be applied to any case. Finally, graphs of comparison between the three methods of expansion with certain cylinder strokes and workloads are presented, to facilitate the subsequent selection of one system or another.

Keywords: energetic, efficiency, GRAFCET, methodology, pneumatic

Procedia PDF Downloads 310
243 Talent-to-Vec: Using Network Graphs to Validate Models with Data Sparsity

Authors: Shaan Khosla, Jon Krohn

Abstract:

In a recruiting context, machine learning models are valuable for recommendations: to predict the best candidates for a vacancy, to match the best vacancies for a candidate, and compile a set of similar candidates for any given candidate. While useful to create these models, validating their accuracy in a recommendation context is difficult due to a sparsity of data. In this report, we use network graph data to generate useful representations for candidates and vacancies. We use candidates and vacancies as network nodes and designate a bi-directional link between them based on the candidate interviewing for the vacancy. After using node2vec, the embeddings are used to construct a validation dataset with a ranked order, which will help validate new recommender systems.

Keywords: AI, machine learning, NLP, recruiting

Procedia PDF Downloads 84
242 A Graph SEIR Cellular Automata Based Model to Study the Spreading of a Transmittable Disease

Authors: Natasha Sharma, Kulbhushan Agnihotri

Abstract:

Cellular Automata are discrete dynamical systems which are based on local character and spatial disparateness of the spreading process. These factors are generally neglected by traditional models based on differential equations for epidemic spread. The aim of this work is to introduce an SEIR model based on cellular automata on graphs to imitate epidemic spreading. Distinctively, it is an SEIR-type model where the population is divided into susceptible, exposed, infected and recovered individuals. The results obtained from simulations are in accordance with the spreading behavior of a real time epidemics.

Keywords: cellular automata, epidemic spread, graph, susceptible

Procedia PDF Downloads 459