Search results for: graph ordering
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 580

Search results for: graph ordering

130 Sustainability Modelling and Sustainability Evaluation of a Mechanical System in a Concurrent Engineering Environment: A Digraph and Matrix Approach

Authors: Anand Ankush, Wani Mohammed Farooq

Abstract:

A procedure based on digraph and matrix method is developed for modelling and evaluation of sustainability of Mechanical System in a concurrent engineering environment.The sustainability parameters of a Mechanical System are identified and are called sustainability attributes. Consideration of attributes and their interrelations is rudiment in modeling and evaluation of sustainability index. Sustainability attributes of a Mechanical System are modelled in termsof sustainability digraph. The graph is represented by a one-to-one matrix for sustainability expression which is based on sustainability attributes. A variable sustainability relationship permanent matrix is defined to develop sustainability expression(VPF-t) which is also useful in comparing two systems in a concurrent environment. The sustainability index of Mechanical System is obtained from permanent of matrix by substituting the numerical values of attributes and their interrelations. A higher value of index implies better sustainability of system.The ideal value of index is obtained from matrix expression which is useful in assessing relative sustainability of a Mechanical System in a concurrent engineering environment. The procedure is not only useful for evaluation of sustainability of a Mechanical System at conceptual design stage but can also be used for design and development of systems at system design stage. A step-by-step procedure for evaluation of sustainability index is also suggested and is illustrated by means of an example.

Keywords: digraph, matrix method, mechanical system, sustainability

Procedia PDF Downloads 358
129 Study on the Incidence of Chikungunya Infection in Swat Region

Authors: Nasib Zaman, Maneesha Kour, Muhammad Rizwan, Fazal Akbar

Abstract:

Abstract: Chikungunya fever is a re-emerging rapidly spreading mosquito-borne disease cause by Aedes albopictus and Aedes aegypti mosquito vectors. Currently, it is affecting millions of people globally. Objective: This study's main objective was to find the incidence of chikungunya fever in the Swat region and the factors associated with the spread of this infection. Method: This study was carried out in different areas of Swat. Blood samples and data were collected from selected patients, and a questionnaire was filled for each patient. 3-5ml of the specimen was taken from the patient's vein and serum, or plasma was separated by centrifugation. Chikungunya tests were performed for IgG and IgM antibodies. The data was analyzed by SPSS and Graph Paid Prism 5. Results: A total of 169 patients were included in this study, out of which 103 (60.9%) having age less than 30 years were positive for chikungunya infection and 66 (39.1%) having more than 30 years were negative for this infection. Only 1 (0.6%) were positive for both IgG and IgM antibody. About 15 (8.9%) patients have diagnosed with positive IgG antibodies, and 25 (26.6%) patients were positive for IgM positive antibodies. The infection rate was significantly higher in males compared to females 71 (59.6%) vs. 14 (38%) P value=0.088, OR=1.7. Conclusion: This study concludes clinical knowledge and awareness that are necessary for a diagnosis of chikungunya infection properly. Therefore it is important to educate people for the eradication of this infection. Recommendation: This study also recommends investigating the other risk factors associated with this infection.

Keywords: Chikungunya, risk factor, Incidence, antibodies, mosquito

Procedia PDF Downloads 123
128 Microwave Assisted Thermal Cracking of Castor Oil Zeolite ZSM-5 as Catalyst for Biofuel Production

Authors: Ghazi Faisal Najmuldeen, Ali Abdul Rahman–Al Ezzi, Tharmathas A/L Alagappan

Abstract:

The aim of this investigation was to produce biofuel from castor oil through microwave assisted thermal cracking with zeolite ZSM-5 as catalyst. The obtained results showed that microwave assisted thermal cracking of castor oil with Zeolite ZSM-5 as catalyst generates products consisting of alcohol, methyl esters and fatty acids. The products obtained from this experimental procedure by the cracking of castor oil are components of biodiesel. Samples of cracked castor oil containing 1, 3 and 5wt % catalyst was analyzed, however, only the sample containing the 5wt % catalyst showed significant presence of condensate. FTIR and GCMS studies show that the condensate obtained is an unsaturated fatty acid, is 9, 12-octadecadienoic acid, suitable for biofuel use. 9, 12-octadecadienoic acid is an unsaturated fatty acid with a molecular weight of 280.445 g/mol. Characterization of the sample demonstrates that functional group for the products from the three samples display a similar peak in the FTIR graph analysis at 1700 cm-1 and 3600 cm-1. The result obtained from GCMS shows that there are 16 peaks obtained from the sample. The compound with the highest peak area is 9, 12-octadecadienoic acid with a retention time of 9.941 and 24.65 peak areas. All these compounds are organic material and can be characterized as biofuel and biodiesel.

Keywords: castor oil, biofuel, biodiesel, thermal cracking, microwave

Procedia PDF Downloads 230
127 An Adaptive Dimensionality Reduction Approach for Hyperspectral Imagery Semantic Interpretation

Authors: Akrem Sellami, Imed Riadh Farah, Basel Solaiman

Abstract:

With the development of HyperSpectral Imagery (HSI) technology, the spectral resolution of HSI became denser, which resulted in large number of spectral bands, high correlation between neighboring, and high data redundancy. However, the semantic interpretation is a challenging task for HSI analysis due to the high dimensionality and the high correlation of the different spectral bands. In fact, this work presents a dimensionality reduction approach that allows to overcome the different issues improving the semantic interpretation of HSI. Therefore, in order to preserve the spatial information, the Tensor Locality Preserving Projection (TLPP) has been applied to transform the original HSI. In the second step, knowledge has been extracted based on the adjacency graph to describe the different pixels. Based on the transformation matrix using TLPP, a weighted matrix has been constructed to rank the different spectral bands based on their contribution score. Thus, the relevant bands have been adaptively selected based on the weighted matrix. The performance of the presented approach has been validated by implementing several experiments, and the obtained results demonstrate the efficiency of this approach compared to various existing dimensionality reduction techniques. Also, according to the experimental results, we can conclude that this approach can adaptively select the relevant spectral improving the semantic interpretation of HSI.

Keywords: band selection, dimensionality reduction, feature extraction, hyperspectral imagery, semantic interpretation

Procedia PDF Downloads 347
126 Numerical Solution of Manning's Equation in Rectangular Channels

Authors: Abdulrahman Abdulrahman

Abstract:

When the Manning equation is used, a unique value of normal depth in the uniform flow exists for a given channel geometry, discharge, roughness, and slope. Depending on the value of normal depth relative to the critical depth, the flow type (supercritical or subcritical) for a given characteristic of channel conditions is determined whether or not flow is uniform. There is no general solution of Manning's equation for determining the flow depth for a given flow rate, because the area of cross section and the hydraulic radius produce a complicated function of depth. The familiar solution of normal depth for a rectangular channel involves 1) a trial-and-error solution; 2) constructing a non-dimensional graph; 3) preparing tables involving non-dimensional parameters. Author in this paper has derived semi-analytical solution to Manning's equation for determining the flow depth given the flow rate in rectangular open channel. The solution was derived by expressing Manning's equation in non-dimensional form, then expanding this form using Maclaurin's series. In order to simplify the solution, terms containing power up to 4 have been considered. The resulted equation is a quartic equation with a standard form, where its solution was obtained by resolving this into two quadratic factors. The proposed solution for Manning's equation is valid over a large range of parameters, and its maximum error is within -1.586%.

Keywords: channel design, civil engineering, hydraulic engineering, open channel flow, Manning's equation, normal depth, uniform flow

Procedia PDF Downloads 212
125 Isolate-Specific Variations among Clinical Isolates of Brucella Identified by Whole-Genome Sequencing, Bioinformatics and Comparative Genomics

Authors: Abu S. Mustafa, Mohammad W. Khan, Faraz Shaheed Khan, Nazima Habibi

Abstract:

Brucellosis is a zoonotic disease of worldwide prevalence. There are at least four species and several strains of Brucella that cause human disease. Brucella genomes have very limited variation across strains, which hinder strain identification using classical molecular techniques, including PCR and 16 S rDNA sequencing. The aim of this study was to perform whole genome sequencing of clinical isolates of Brucella and perform bioinformatics and comparative genomics analyses to determine the existence of genetic differences across the isolates of a single Brucella species and strain. The draft sequence data were generated from 15 clinical isolates of Brucella melitensis (biovar 2 strain 63/9) using MiSeq next generation sequencing platform. The generated reads were used for further assembly and analysis. All the analysis was performed using Bioinformatics work station (8 core i7 processor, 8GB RAM with Bio-Linux operating system). FastQC was used to determine the quality of reads and low quality reads were trimmed or eliminated using Fastx_trimmer. Assembly was done by using Velvet and ABySS softwares. The ordering of assembled contigs was performed by Mauve. An online server RAST was employed to annotate the contigs assembly. Annotated genomes were compared using Mauve and ACT tools. The QC score for DNA sequence data, generated by MiSeq, was higher than 30 for 80% of reads with more than 100x coverage, which suggested that data could be utilized for further analysis. However when analyzed by FastQC, quality of four reads was not good enough for creating a complete genome draft so remaining 11 samples were used for further analysis. The comparative genome analyses showed that despite sharing same gene sets, single nucleotide polymorphisms and insertions/deletions existed across different genomes, which provided a variable extent of diversity to these bacteria. In conclusion, the next generation sequencing, bioinformatics, and comparative genome analysis can be utilized to find variations (point mutations, insertions and deletions) across different genomes of Brucella within a single strain. This information could be useful in surveillance and epidemiological studies supported by Kuwait University Research Sector grants MI04/15 and SRUL02/13.

Keywords: brucella, bioinformatics, comparative genomics, whole genome sequencing

Procedia PDF Downloads 375
124 Exploring the Role of Humorous Dialogues in Advertisements of Pakistani Network Companies: Analysis of Discourses through Multi-Modal Critical Approach

Authors: Jane E. Alam Solangi

Abstract:

The contribution of the study is to explore the important part of humorous dialogues in cellular network advertisements. This promotes the message of valuable construction and promotion of network companies in Pakistan that employ different and broad techniques to give promotion to selling products. It merely instigates the consumers to buy it. The results of the study after analysis of its collected data gives a vision that advertisers of network advertisements use humorous dialogues as a significant device to the greater level. The source of entertainment in the advertisement is accompanied by the texts and humorous discourses to influence buying decisions of the consumers. Therefore, it tends to neutralize personal and social based values. The earlier contribution of scholars presented that the technical employment of humorous devices leads to the successful market of the relevant products. In order to analyze the humorous discourse devices, the approach of multi-modality of Fairclough (1989) is used. It is accompanied by the framework of Kress and van Leeuwen’s (1996). It analyzes the visual graph of the grammar. The overall findings in the study verified the role of humorous devices in the captivation of consumers’ decision to buy the product that interests them. Therefore, the role of humor acts as a breaker of the monotonous rhythm of advertisements.

Keywords: advertisements, devices, humorous, multi-modality, networks, Pakistan

Procedia PDF Downloads 100
123 Particle Size Analysis of Itagunmodi Southwestern Nigeria Alluvial Gold Ore Sample by Gaudin Schumann Method

Authors: Olaniyi Awe, Adelana R. Adetunji, Abraham Adeleke

Abstract:

Mining of alluvial gold ore by artisanal miners has been going on for decades at Itagunmodi, Southwestern Nigeria. In order to optimize the traditional panning gravity separation method commonly used in the area, a mineral particle size analysis study is critical. This study analyzed alluvial gold ore samples collected at identified five different locations in the area with a view to determine the ore particle size distributions. 500g measured of as-received alluvial gold ore sample was introduced into the uppermost sieve of an electrical sieve shaker consisting of sieves arranged in the order of decreasing nominal apertures of 5600μm, 3350μm, 2800μm, 355μm, 250μm, 125μm and 90μm, and operated for 20 minutes. The amount of material retained on each sieve was measured and tabulated for analysis. A screen analysis graph using the Gaudin Schuman method was drawn for each of the screen tests on the alluvial samples. The study showed that the percentages of fine particle size -125+90 μm fraction were 45.00%, 36.00%, 39.60%, 43.00% and 36.80% for the selected samples. These primary ore characteristic results provide reference data for the alluvial gold ore processing method selection, process performance measurement and optimization.

Keywords: alluvial gold ore, sieve shaker, particle size, Gaudin Schumann

Procedia PDF Downloads 52
122 Prototyping the Problem Oriented Medical Record for Connected Health Based on TypeGraphQL

Authors: Sabah Mohammed, Jinan Fiaidhi, Darien Sawyer

Abstract:

Data integration of health through connected services can save lives in the event of a medical emergency or provide efficient and effective interventions for the benefit of the patients through the integration of bedside and bench side clinical research. Such integration will support all wind of change in healthcare by being predictive, pre-emptive, personalized, problem-oriented and participatory. Prototyping a healthcare system that enables data integration has been a big challenge for healthcare for a long time. However, an innovative solution started to emerge by focusing on problem lists where everything can connect the problem list forming a growing graph. This notion was introduced by Dr. Lawrence Weed in early 70’s, but the enabling technologies weren’t mature enough to provide a successful implementation prototype. In this article, we are describing our efforts in prototyping Dr. Lawrence Weed's problem-oriented medical record (POMR) and his patient case schema (SOAP) to shape a prototype for connected health. For this, we are using the TypeGraphQL API and our enterprise-based QL4POMR to describe a Web-Based gateway for healthcare services connectivity. Our prototype has reported success in connecting to the HL7 FHIR medical record and the OpenTarget biomedical repositories.

Keywords: connected health, problem-oriented healthcare record, SOAP, QL4POMR, typegraphQL

Procedia PDF Downloads 95
121 On a Transient Magnetohydrodynamics Heat Transfer Within Radiative Porous Channel Due to Convective Boundary Condition

Authors: Bashiru Abdullahi, Isah Bala Yabo, Ibrahim Yakubu Seini

Abstract:

In this paper, the steady/transient MHD heat transfer within radiative porous channel due to convective boundary conditions is considered. The solution of the steady-state and that of the transient version were conveyed by Perturbation and Finite difference methods respectively. The heat transfer mechanism of the present work ascertains the influence of Biot number〖(B〗_i1), magnetizing parameter (M), radiation parameter(R), temperature difference, suction/injection(S) Grashof number (Gr) and time (t) on velocity (u), temperature(θ), skin friction(τ), and Nusselt number (Nu). The results established were discussed with the help of a line graph. It was found that the velocity, temperature, and skin friction decay with increasing suction/injection and magnetizing parameters while the Nusselt number upsurges with suction/injection at y = 0 and falls at y =1. The steady-state solution was in perfect agreement with the transient version for a significant value of time t. It is interesting to report that the Biot number has a cogent influence consequently, as its values upsurge the result of the present work slant the extended literature.

Keywords: heat transfer, thermal radiation, porous channel, MHD, transient, convective boundary condition

Procedia PDF Downloads 117
120 Composition Dependence of Ni 2p Core Level Shift in Fe1-xNix Alloys

Authors: Shakti S. Acharya, V. R. R. Medicherla, Rajeev Rawat, Komal Bapna, Deepnarayan Biswas, Khadija Ali, K. Maiti

Abstract:

The discovery of invar effect in 35% Ni concentration Fe1-xNix alloy has stimulated enormous experimental and theoretical research. Elemental Fe and low Ni concentration Fe1-xNix alloys which possess body centred cubic (bcc) crystal structure at ambient temperature and pressure transform to hexagonally close packed (hcp) phase at around 13 GPa. Magnetic order was found to be absent at 11K for Fe92Ni8 alloy when subjected to a high pressure of 26 GPa. The density functional theoretical calculations predicted substantial hyperfine magnetic fields, but were not observed in Mossbaur spectroscopy. The bulk modulus of fcc Fe1-xNix alloys with Ni concentration more than 35%, is found to be independent of pressure. The magnetic moment of Fe is also found be almost same in these alloys from 4 to 10 GPa pressure. Fe1-xNix alloys exhibit a complex microstructure which is formed by a series of complex phase transformations like martensitic transformation, spinodal decomposition, ordering, mono-tectoid reaction, eutectoid reaction at temperatures below 400°C. Despite the existence of several theoretical models the field is still in its infancy lacking full knowledge about the anomalous properties exhibited by these alloys. Fe1-xNix alloys have been prepared by arc melting the high purity constituent metals in argon ambient. These alloys have annealed at around 3000C in vacuum sealed quartz tube for two days to make the samples homogeneous. These alloys have been structurally characterized by x-ray diffraction and were found to exhibit a transition from bcc to fcc for x > 0.3. Ni 2p core levels of the alloys have been measured using high resolution (0.45 eV) x-ray photoelectron spectroscopy. Ni 2p core level shifts to lower binding energy with respect to that of pure Ni metal giving rise to negative core level shifts (CLSs). Measured CLSs exhibit a linear dependence in fcc region (x > 0.3) and were found to deviate slightly in bcc region (x < 0.3). ESCA potential model fails correlate CLSs with site potentials or charges in metallic alloys. CLSs in these alloys occur mainly due to shift in valence bands with composition due to intra atomic charge redistribution.

Keywords: arc melting, core level shift, ESCA potential model, valence band

Procedia PDF Downloads 375
119 Adaptive Analysis of Housing Policies in Development Programming After 1970s (Case Study: Kermanshah City in the Western Iran)

Authors: Zeinab. Shahrokhifar, Abolfazl Meshkini, Seyed Ali. Alavi

Abstract:

Considering the different dimensions of deprivation, housing supply is noted as a basic requirement in Iran after 1979 (coming to work of the new government). The government had built the constitution and obliged to meet this need in the form of five-year development programs in Iran’s provinces. This study focused on the adaptive analysis of housing policies in these five development programs in Kermanshah province located in western Iran. Our research is divided into two different analytical sections. In the first section, we collected the documentary information using approved plans and field studies. In the second section, a questionnaire was prepared and designed for the elite community (30) to support the documentary analysis. The results showed that various projects adopted in the form of strategic plans and implemented the policies included both quantitative and qualitative housing in Kermanshah province after 1979. The quality of housing, from the first to the fifth development plans has improved the situation in the housing indicators. The quantity of housing units for households has also been implemented through various policies that has desired results. The sequences of housing policies and plans do not overlap in the five development programs. According to the radar graph, the development programs overlapped in some policies, which shows the continuation of the previous policies, but this overlap is not perfect.

Keywords: law enforcement policy, housing policy, development programs, housing indicators, the city of Kermanshah

Procedia PDF Downloads 67
118 Network and Sentiment Analysis of U.S. Congressional Tweets

Authors: Chaitanya Kanakamedala, Hansa Pradhan, Carter Gilbert

Abstract:

Social media platforms, such as Twitter, are excellent datasets for understanding human interactions and sentiments. This report explores social dynamics among US Congressional members through a network analysis applied to a dataset of tweets spanning 2008 to 2017 from the ’US Congressional Tweets Dataset’. In this report, we preform network analysis where connections between users (edges) are established based on a similarity threshold: two tweets are connected if the tweets they post are similar. By utilizing the Natural Language Toolkit (NLTK) and NetworkX, we quantified tweet similarity and constructed a graph comprising various interconnected components. Each component represents a cluster of users with closely aligned content. We then preform sentiment analysis on each cluster to explore the prevalent emotions and opinions within these groups. Our findings reveal that despite the initial expectation of distinct ideological divisions typically aligning with party lines, the analysis exposed a high degree of topical convergence across tweets from different political affiliations. The analysis preformed in this report not only highlights the potential of social media as a tool for political communication but also suggests a complex layer of interaction that transcends traditional partisan boundaries, reflecting a complicated landscape of politics in the digital age.

Keywords: natural language processing, sentiment analysis, centrality analysis, topic modeling

Procedia PDF Downloads 24
117 Aggregation Scheduling Algorithms in Wireless Sensor Networks

Authors: Min Kyung An

Abstract:

In Wireless Sensor Networks which consist of tiny wireless sensor nodes with limited battery power, one of the most fundamental applications is data aggregation which collects nearby environmental conditions and aggregates the data to a designated destination, called a sink node. Important issues concerning the data aggregation are time efficiency and energy consumption due to its limited energy, and therefore, the related problem, named Minimum Latency Aggregation Scheduling (MLAS), has been the focus of many researchers. Its objective is to compute the minimum latency schedule, that is, to compute a schedule with the minimum number of timeslots, such that the sink node can receive the aggregated data from all the other nodes without any collision or interference. For the problem, the two interference models, the graph model and the more realistic physical interference model known as Signal-to-Interference-Noise-Ratio (SINR), have been adopted with different power models, uniform-power and non-uniform power (with power control or without power control), and different antenna models, omni-directional antenna and directional antenna models. In this survey article, as the problem has proven to be NP-hard, we present and compare several state-of-the-art approximation algorithms in various models on the basis of latency as its performance measure.

Keywords: data aggregation, convergecast, gathering, approximation, interference, omni-directional, directional

Procedia PDF Downloads 225
116 A Decision-Support Tool for Humanitarian Distribution Planners in the Face of Congestion at Security Checkpoints: A Real-World Case Study

Authors: Mohanad Rezeq, Tarik Aouam, Frederik Gailly

Abstract:

In times of armed conflicts, various security checkpoints are placed by authorities to control the flow of merchandise into and within areas of conflict. The flow of humanitarian trucks that is added to the regular flow of commercial trucks, together with the complex security procedures, creates congestion and long waiting times at the security checkpoints. This causes distribution costs to increase and shortages of relief aid to the affected people to occur. Our research proposes a decision-support tool to assist planners and policymakers in building efficient plans for the distribution of relief aid, taking into account congestion at security checkpoints. The proposed tool is built around a multi-item humanitarian distribution planning model based on multi-phase design science methodology that has as its objective to minimize distribution and back ordering costs subject to capacity constraints that reflect congestion effects using nonlinear clearing functions. Using the 2014 Gaza War as a case study, we illustrate the application of the proposed tool, model the underlying relief-aid humanitarian supply chain, estimate clearing functions at different security checkpoints, and conduct computational experiments. The decision support tool generated a shipment plan that was compared to two benchmarks in terms of total distribution cost, average lead time and work in progress (WIP) at security checkpoints, and average inventory and backorders at distribution centers. The first benchmark is the shipment plan generated by the fixed capacity model, and the second is the actual shipment plan implemented by the planners during the armed conflict. According to our findings, modeling and optimizing supply chain flows reduce total distribution costs, average truck wait times at security checkpoints, and average backorders when compared to the executed plan and the fixed-capacity model. Finally, scenario analysis concludes that increasing capacity at security checkpoints can lower total operations costs by reducing the average lead time.

Keywords: humanitarian distribution planning, relief-aid distribution, congestion, clearing functions

Procedia PDF Downloads 80
115 Application of Supervised Deep Learning-based Machine Learning to Manage Smart Homes

Authors: Ahmed Al-Adaileh

Abstract:

Renewable energy sources, domestic storage systems, controllable loads and machine learning technologies will be key components of future smart homes management systems. An energy management scheme that uses a Deep Learning (DL) approach to support the smart home management systems, which consist of a standalone photovoltaic system, storage unit, heating ventilation air-conditioning system and a set of conventional and smart appliances, is presented. The objective of the proposed scheme is to apply DL-based machine learning to predict various running parameters within a smart home's environment to achieve maximum comfort levels for occupants, reduced electricity bills, and less dependency on the public grid. The problem is using Reinforcement learning, where decisions are taken based on applying the Continuous-time Markov Decision Process. The main contribution of this research is the proposed framework that applies DL to enhance the system's supervised dataset to offer unlimited chances to effectively support smart home systems. A case study involving a set of conventional and smart appliances with dedicated processing units in an inhabited building can demonstrate the validity of the proposed framework. A visualization graph can show "before" and "after" results.

Keywords: smart homes systems, machine learning, deep learning, Markov Decision Process

Procedia PDF Downloads 190
114 Design of an Ensemble Learning Behavior Anomaly Detection Framework

Authors: Abdoulaye Diop, Nahid Emad, Thierry Winter, Mohamed Hilia

Abstract:

Data assets protection is a crucial issue in the cybersecurity field. Companies use logical access control tools to vault their information assets and protect them against external threats, but they lack solutions to counter insider threats. Nowadays, insider threats are the most significant concern of security analysts. They are mainly individuals with legitimate access to companies information systems, which use their rights with malicious intents. In several fields, behavior anomaly detection is the method used by cyber specialists to counter the threats of user malicious activities effectively. In this paper, we present the step toward the construction of a user and entity behavior analysis framework by proposing a behavior anomaly detection model. This model combines machine learning classification techniques and graph-based methods, relying on linear algebra and parallel computing techniques. We show the utility of an ensemble learning approach in this context. We present some detection methods tests results on an representative access control dataset. The use of some explored classifiers gives results up to 99% of accuracy.

Keywords: cybersecurity, data protection, access control, insider threat, user behavior analysis, ensemble learning, high performance computing

Procedia PDF Downloads 122
113 Leveraging the Power of Dual Spatial-Temporal Data Scheme for Traffic Prediction

Authors: Yang Zhou, Heli Sun, Jianbin Huang, Jizhong Zhao, Shaojie Qiao

Abstract:

Traffic prediction is a fundamental problem in urban environment, facilitating the smart management of various businesses, such as taxi dispatching, bike relocation, and stampede alert. Most earlier methods rely on identifying the intrinsic spatial-temporal correlation to forecast. However, the complex nature of this problem entails a more sophisticated solution that can simultaneously capture the mutual influence of both adjacent and far-flung areas, with the information of time-dimension also incorporated seamlessly. To tackle this difficulty, we propose a new multi-phase architecture, DSTDS (Dual Spatial-Temporal Data Scheme for traffic prediction), that aims to reveal the underlying relationship that determines future traffic trend. First, a graph-based neural network with an attention mechanism is devised to obtain the static features of the road network. Then, a multi-granularity recurrent neural network is built in conjunction with the knowledge from a grid-based model. Subsequently, the preceding output is fed into a spatial-temporal super-resolution module. With this 3-phase structure, we carry out extensive experiments on several real-world datasets to demonstrate the effectiveness of our approach, which surpasses several state-of-the-art methods.

Keywords: traffic prediction, spatial-temporal, recurrent neural network, dual data scheme

Procedia PDF Downloads 111
112 Optimization of Hate Speech and Abusive Language Detection on Indonesian-language Twitter using Genetic Algorithms

Authors: Rikson Gultom

Abstract:

Hate Speech and Abusive language on social media is difficult to detect, usually, it is detected after it becomes viral in cyberspace, of course, it is too late for prevention. An early detection system that has a fairly good accuracy is needed so that it can reduce conflicts that occur in society caused by postings on social media that attack individuals, groups, and governments in Indonesia. The purpose of this study is to find an early detection model on Twitter social media using machine learning that has high accuracy from several machine learning methods studied. In this study, the support vector machine (SVM), Naïve Bayes (NB), and Random Forest Decision Tree (RFDT) methods were compared with the Support Vector machine with genetic algorithm (SVM-GA), Nave Bayes with genetic algorithm (NB-GA), and Random Forest Decision Tree with Genetic Algorithm (RFDT-GA). The study produced a comparison table for the accuracy of the hate speech and abusive language detection model, and presented it in the form of a graph of the accuracy of the six algorithms developed based on the Indonesian-language Twitter dataset, and concluded the best model with the highest accuracy.

Keywords: abusive language, hate speech, machine learning, optimization, social media

Procedia PDF Downloads 125
111 Methylprednisolone Injection Did Not Inhibit Anti-Hbs Response Following Hepatitis B Vaccination in Mice

Authors: P. O. Ughachukwu, P. O. Okonkwo, P. C. Unekwe, J. O. Ogamba

Abstract:

Background: The prevalence of hepatitis B viral infection is high worldwide with liver cirrhosis and hepatocellular carcinoma as important complications. Cases of poor antibody response to hepatitis B vaccination abound. Immunosuppression, especially from glucocorticoids, is often cited as a cause of poor antibody response and there are documented evidences of irrational administration of glucocorticoids to children and adults. The study was, therefore, designed to find out if administration of glucocorticoids affects immune response to vaccination against hepatitis B in mice. Methods: Mice of both sexes were randomly divided into 2 groups. Daily intramuscular methylprednisolone injections, (15 mg kg-1), were given to the test group while sterile deionized water (0.1ml) was given to control mice for 30 days. On day 6 all mice were given 2 μg (0.1ml) hepatitis B vaccine and a booster dose on day 27. On day 34, blood samples were collected and analyzed for anti-HBs titres using enzyme-linked immunosorbent assay (ELISA). Statistical analysis was done using Graph Pad Prism 5.0 and the results taken as statistically significant at p value < 0.05. Results: There were positive serum anti-HBs responses in all mice groups but the differences in titres were not statistically significant. Conclusions: At the dosages and length of exposure used in this study, methylprednisolone injection did not significantly inhibit anti-HBs response in mice following immunization against hepatitis B virus. By extrapolation, methylprednisolone, when used in the usual clinical doses and duration of therapy, is not likely to inhibit immune response to hepatitis B vaccinations in man.

Keywords: anti-HBs, hepatitis B vaccine, immune response, methylprednisolone, mice

Procedia PDF Downloads 320
110 ParkedGuard: An Efficient and Accurate Parked Domain Detection System Using Graphical Locality Analysis and Coarse-To-Fine Strategy

Authors: Chia-Min Lai, Wan-Ching Lin, Hahn-Ming Lee, Ching-Hao Mao

Abstract:

As world wild internet has non-stop developments, making profit by lending registered domain names emerges as a new business in recent years. Unfortunately, the larger the market scale of domain lending service becomes, the riskier that there exist malicious behaviors or malwares hiding behind parked domains will be. Also, previous work for differentiating parked domain suffers two main defects: 1) too much data-collecting effort and CPU latency needed for features engineering and 2) ineffectiveness when detecting parked domains containing external links that are usually abused by hackers, e.g., drive-by download attack. Aiming for alleviating above defects without sacrificing practical usability, this paper proposes ParkedGuard as an efficient and accurate parked domain detector. Several scripting behavioral features were analyzed, while those with special statistical significance are adopted in ParkedGuard to make feature engineering much more cost-efficient. On the other hand, finding memberships between external links and parked domains was modeled as a graph mining problem, and a coarse-to-fine strategy was elaborately designed by leverage the graphical locality such that ParkedGuard outperforms the state-of-the-art in terms of both recall and precision rates.

Keywords: coarse-to-fine strategy, domain parking service, graphical locality analysis, parked domain

Procedia PDF Downloads 404
109 Investigating the Regulation System of the Synchronous Motor Excitation Mode Serving as a Reactive Power Source

Authors: Baghdasaryan Marinka, Ulikyan Azatuhi

Abstract:

The efficient usage of the compensation abilities of the electrical drive synchronous motors used in production processes can essentially improve the technical and economic indices of the process.  Reducing the flows of the reactive electrical energy due to the compensation of reactive power allows to significantly reduce the load losses of power in the electrical networks. As a result of analyzing the scientific works devoted to the issues of regulating the excitation of the synchronous motors, the need for comprehensive investigation and estimation of the excitation mode has been substantiated. By means of the obtained transmission functions, in the Simulink environment of the software package MATLAB, the transition processes of the excitation mode have been studied. As a result of obtaining and estimating the graph of the Nyquist plot and the transient process, the necessity of developing the Proportional-Integral-Derivative (PID) regulator has been justified. The transient processes of the system of the PID regulator have been investigated, and the amplitude–phase characteristics of the system have been estimated. The analysis of the obtained results has shown that the regulation indices of the developed system have been improved. The developed system can be successfully applied for regulating the excitation voltage of different-power synchronous motors, operating with a changing load, ensuring a value of the power coefficient close to 1.

Keywords: transition process, synchronous motor, excitation mode, regulator, reactive power

Procedia PDF Downloads 228
108 The Use of Layered Neural Networks for Classifying Hierarchical Scientific Fields of Study

Authors: Colin Smith, Linsey S Passarella

Abstract:

Due to the proliferation and decentralized nature of academic publication, no widely accepted scheme exists for organizing papers by their scientific field of study (FoS) to the author’s best knowledge. While many academic journals require author provided keywords for papers, these keywords range wildly in scope and are not consistent across papers, journals, or field domains, necessitating alternative approaches to paper classification. Past attempts to perform field-of-study (FoS) classification on scientific texts have largely used a-hierarchical FoS schemas or ignored the schema’s inherently hierarchical structure, e.g. by compressing the structure into a single layer for multi-label classification. In this paper, we introduce an application of a Layered Neural Network (LNN) to the problem of performing supervised hierarchical classification of scientific fields of study (FoS) on research papers. In this approach, paper embeddings from a pretrained language model are fed into a top-down LNN. Beginning with a single neural network (NN) for the highest layer of the class hierarchy, each node uses a separate local NN to classify the subsequent subfield child node(s) for an input embedding of concatenated paper titles and abstracts. We compare our LNN-FOS method to other recent machine learning methods using the Microsoft Academic Graph (MAG) FoS hierarchy and find that the LNN-FOS offers increased classification accuracy at each FoS hierarchical level.

Keywords: hierarchical classification, layer neural network, scientific field of study, scientific taxonomy

Procedia PDF Downloads 128
107 Traffic Prediction with Raw Data Utilization and Context Building

Authors: Zhou Yang, Heli Sun, Jianbin Huang, Jizhong Zhao, Shaojie Qiao

Abstract:

Traffic prediction is essential in a multitude of ways in modern urban life. The researchers of earlier work in this domain carry out the investigation chiefly with two major focuses: (1) the accurate forecast of future values in multiple time series and (2) knowledge extraction from spatial-temporal correlations. However, two key considerations for traffic prediction are often missed: the completeness of raw data and the full context of the prediction timestamp. Concentrating on the two drawbacks of earlier work, we devise an approach that can address these issues in a two-phase framework. First, we utilize the raw trajectories to a greater extent through building a VLA table and data compression. We obtain the intra-trajectory features with graph-based encoding and the intertrajectory ones with a grid-based model and the technique of back projection that restore their surrounding high-resolution spatial-temporal environment. To the best of our knowledge, we are the first to study direct feature extraction from raw trajectories for traffic prediction and attempt the use of raw data with the least degree of reduction. In the prediction phase, we provide a broader context for the prediction timestamp by taking into account the information that are around it in the training dataset. Extensive experiments on several well-known datasets have verified the effectiveness of our solution that combines the strength of raw trajectory data and prediction context. In terms of performance, our approach surpasses several state-of-the-art methods for traffic prediction.

Keywords: traffic prediction, raw data utilization, context building, data reduction

Procedia PDF Downloads 125
106 Ontology Expansion via Synthetic Dataset Generation and Transformer-Based Concept Extraction

Authors: Andrey Khalov

Abstract:

The rapid proliferation of unstructured data in IT infrastructure management demands innovative approaches for extracting actionable knowledge. This paper presents a framework for ontology-based knowledge extraction that combines relational graph neural networks (R-GNN) with large language models (LLMs). The proposed method leverages the DOLCE framework as the foundational ontology, extending it with concepts from ITSMO for domain-specific applications in IT service management and outsourcing. A key component of this research is the use of transformer-based models, such as DeBERTa-v3-large, for automatic entity and relationship extraction from unstructured texts. Furthermore, the paper explores how transfer learning techniques can be applied to fine-tune large language models (LLaMA) for using to generate synthetic datasets to improve precision in BERT-based entity recognition and ontology alignment. The resulting IT Ontology (ITO) serves as a comprehensive knowledge base that integrates domain-specific insights from ITIL processes, enabling more efficient decision-making. Experimental results demonstrate significant improvements in knowledge extraction and relationship mapping, offering a cutting-edge solution for enhancing cognitive computing in IT service environments.

Keywords: ontology expansion, synthetic dataset, transformer fine-tuning, concept extraction, DOLCE, BERT, taxonomy, LLM, NER

Procedia PDF Downloads 0
105 Hot Corrosion and Oxidation Degradation Mechanism of Turbine Materials in a Water Vapor Environment at a Higher Temperature

Authors: Mairaj Ahmad, L. Paglia, F. Marra, V. Genova, G. Pulci

Abstract:

This study employed Rene N4 and FSX 414 superalloys, which are used in numerous turbine engine components due of their high strength, outstanding fatigue, creep, thermal, and corrosion-resistant properties. An in-depth examination of corrosion mechanisms with vapor present at high temperature is necessary given the industrial trend toward introducing increasing amounts of hydrogen into combustion chambers in order to boost power generation and minimize pollution in contrast to conventional fuels. These superalloys were oxidized in recent tests for 500, 1000, 2000, 3000 and 4000 hours at 982±5°C temperatures with a steady airflow at a flow rate of 10L/min and 1.5 bar pressure. These superalloys were also examined for wet corrosion for 500, 1000, 2000, 3000, and 4000 hours in a combination of air and water vapor flowing at a 10L/min rate. Weight gain, X-ray diffraction (XRD), scanning electron microscopy (SEM), and energy dispersive x-ray spectroscopy (EDS) were used to assess the oxidation and heat corrosion resistance capabilities of these alloys before and after 500, 1000, and 2000 hours. The oxidation/corrosion processes that accompany the formation of these oxide scales are shown in the graph of mass gain vs time. In both dry and wet oxidation, oxides like Al2O3, TiO2, NiCo2O4, Ni3Al, Ni3Ti, Cr2O3, MnCr2O4, CoCr2O4, and certain volatile compounds notably CrO2(OH)2, Cr(OH)3, Fe(OH)2, and Si(OH)4 are formed.

Keywords: hot corrosion, oxidation, turbine materials, high temperature corrosion, super alloys

Procedia PDF Downloads 83
104 Geochemistry Identification of Volcanic Rocks Product of Krakatau Volcano Eruption for Katastropis Mitigation Planning

Authors: Agil Gemilang Ramadhan, Novian Triandanu

Abstract:

Since 1929, the first appearance in sea level, Anak Krakatau volcano growth relatively quickly. During the 80 years up to 2010 has reached the height of 320 meter above sea level. The possibility of catastrophic explosive eruption could happen again if the chemical composition of rocks from the eruption changed from alkaline magma into acid magma. Until now Anak Krakatau volcanic activity is still quite active as evidenced by the frequency of eruptions that produced ash sized pyroclastic deposits - bomb. Purpose of this study was to identify changes in the percentage of rock geochemistry any results eruption of Anak Krakatau volcano to see consistency change the percentage content of silica in the magma that affect the type of volcanic eruptions. Results from this study will be produced in the form of a diagram the data changes the chemical composition of rocks of Anak Krakatau volcano. Changes in the composition of any silica eruption are illustrated in a graph. If the increase in the percentage of silica is happening consistently and it is assumed to increase in the time scale of a few percent, then to achieve silica content of 68 % (acid composition) that will produce an explosive eruption will know the approximate time. All aspects of the factors driving the increased threat of danger to the public should be taken into account. Catastrophic eruption katatropis mitigation can be planned early so that when these disasters happen later, casualties can be minimized.

Keywords: Krakatau volcano, rock geochemistry, catastrophic eruption, mitigation

Procedia PDF Downloads 278
103 Frontier Dynamic Tracking in the Field of Urban Plant and Habitat Research: Data Visualization and Analysis Based on Journal Literature

Authors: Shao Qi

Abstract:

The article uses the CiteSpace knowledge graph analysis tool to sort and visualize the journal literature on urban plants and habitats in the Web of Science and China National Knowledge Infrastructure databases. Based on a comprehensive interpretation of the visualization results of various data sources and the description of the intrinsic relationship between high-frequency keywords using knowledge mapping, the research hotspots, processes and evolution trends in this field are analyzed. Relevant case studies are also conducted for the hotspot contents to explore the means of landscape intervention and synthesize the understanding of research theories. The results show that (1) from 1999 to 2022, the research direction of urban plants and habitats gradually changed from focusing on plant and animal extinction and biological invasion to the field of human urban habitat creation, ecological restoration, and ecosystem services. (2) The results of keyword emergence and keyword growth trend analysis show that habitat creation research has shown a rapid and stable growth trend since 2017, and ecological restoration has gained long-term sustained attention since 2004. The hotspots of future research on urban plants and habitats in China may focus on habitat creation and ecological restoration.

Keywords: research trends, visual analysis, habitat creation, ecological restoration

Procedia PDF Downloads 58
102 A User-Directed Approach to Optimization via Metaprogramming

Authors: Eashan Hatti

Abstract:

In software development, programmers often must make a choice between high-level programming and high-performance programs. High-level programming encourages the use of complex, pervasive abstractions. However, the use of these abstractions degrades performance-high performance demands that programs be low-level. In a compiler, the optimizer attempts to let the user have both. The optimizer takes high-level, abstract code as an input and produces low-level, performant code as an output. However, there is a problem with having the optimizer be a built-in part of the compiler. Domain-specific abstractions implemented as libraries are common in high-level languages. As a language’s library ecosystem grows, so does the number of abstractions that programmers will use. If these abstractions are to be performant, the optimizer must be extended with new optimizations to target them, or these abstractions must rely on existing general-purpose optimizations. The latter is often not as effective as needed. The former presents too significant of an effort for the compiler developers, as they are the only ones who can extend the language with new optimizations. Thus, the language becomes more high-level, yet the optimizer – and, in turn, program performance – falls behind. Programmers are again confronted with a choice between high-level programming and high-performance programs. To investigate a potential solution to this problem, we developed Peridot, a prototype programming language. Peridot’s main contribution is that it enables library developers to easily extend the language with new optimizations themselves. This allows the optimization workload to be taken off the compiler developers’ hands and given to a much larger set of people who can specialize in each problem domain. Because of this, optimizations can be much more effective while also being much more numerous. To enable this, Peridot supports metaprogramming designed for implementing program transformations. The language is split into two fragments or “levels”, one for metaprogramming, the other for high-level general-purpose programming. The metaprogramming level supports logic programming. Peridot’s key idea is that optimizations are simply implemented as metaprograms. The meta level supports several specific features which make it particularly suited to implementing optimizers. For instance, metaprograms can automatically deduce equalities between the programs they are optimizing via unification, deal with variable binding declaratively via higher-order abstract syntax, and avoid the phase-ordering problem via non-determinism. We have found that this design centered around logic programming makes optimizers concise and easy to write compared to their equivalents in functional or imperative languages. Overall, implementing Peridot has shown that its design is a viable solution to the problem of writing code which is both high-level and performant.

Keywords: optimization, metaprogramming, logic programming, abstraction

Procedia PDF Downloads 83
101 Contemporary Army Prints for Women’s Wear Kurti

Authors: Shaleni Bajpai, Nancy Stephan

Abstract:

Various designs of women’s kurtis with different styles, motifs and prints were available in market but none of the kurtis was found in army print. Mostly army prints are used for men’s wear like jackets, trousers, caps, bags. The main colours available in military prints were beige, parrot green, red, dark blue, light blue, orange, bottle green, pink and the original military green colour. As the original camouflage is banned in civil wears so the different variety and colours were used in this study to popularize army prints in women’s wear. The aim of this project was to construct different styles of women kurti’s with various colours of different military prints. Mood board, inspiration and colour board was prepared to design the kurtis. The fabric used for construction was army printed poplin and crepe. The designing and construction of kurti’s were divided into two categories such as - casual and party wear. Casual wear had simple silhouette like a-line, high-low and waist coat style whereas party wear included princess line, panelled and bandhani style. Structured questionnaire was prepared to assess the acceptance of newly designed kurtis with respect to colour combination, overall appearance and cost. Purposively sampling method was adopted for selection of respondents. Opinion was taken from 100 women of various age groups. The result and analysis was presented through graph and percentage. Kurtis in army print of both the categories were appreciated by the respondents.

Keywords: army, kurti, casual wear, party wear

Procedia PDF Downloads 299