Search results for: approximate graph matching
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1245

Search results for: approximate graph matching

345 Confidence Intervals for Process Capability Indices for Autocorrelated Data

Authors: Jane A. Luke

Abstract:

Persistent pressure passed on to manufacturers from escalating consumer expectations and the ever growing global competitiveness have produced a rapidly increasing interest in the development of various manufacturing strategy models. Academic and industrial circles are taking keen interest in the field of manufacturing strategy. Many manufacturing strategies are currently centered on the traditional concepts of focused manufacturing capabilities such as quality, cost, dependability and innovation. Process capability indices was conducted assuming that the process under study is in statistical control and independent observations are generated over time. However, in practice, it is very common to come across processes which, due to their inherent natures, generate autocorrelated observations. The degree of autocorrelation affects the behavior of patterns on control charts. Even, small levels of autocorrelation between successive observations can have considerable effects on the statistical properties of conventional control charts. When observations are autocorrelated the classical control charts exhibit nonrandom patterns and lack of control. Many authors have considered the effect of autocorrelation on the performance of statistical process control charts. In this paper, the effect of autocorrelation on confidence intervals for different PCIs was included. Stationary Gaussian processes is explained. Effect of autocorrelation on PCIs is described in detail. Confidence intervals for Cp and Cpk are constructed for PCIs when data are both independent and autocorrelated. Confidence intervals for Cp and Cpk are computed. Approximate lower confidence limits for various Cpk are computed assuming AR(1) model for the data. Simulation studies and industrial examples are considered to demonstrate the results.

Keywords: autocorrelation, AR(1) model, Bissell’s approximation, confidence intervals, statistical process control, specification limits, stationary Gaussian processes

Procedia PDF Downloads 362
344 Frontier Dynamic Tracking in the Field of Urban Plant and Habitat Research: Data Visualization and Analysis Based on Journal Literature

Authors: Shao Qi

Abstract:

The article uses the CiteSpace knowledge graph analysis tool to sort and visualize the journal literature on urban plants and habitats in the Web of Science and China National Knowledge Infrastructure databases. Based on a comprehensive interpretation of the visualization results of various data sources and the description of the intrinsic relationship between high-frequency keywords using knowledge mapping, the research hotspots, processes and evolution trends in this field are analyzed. Relevant case studies are also conducted for the hotspot contents to explore the means of landscape intervention and synthesize the understanding of research theories. The results show that (1) from 1999 to 2022, the research direction of urban plants and habitats gradually changed from focusing on plant and animal extinction and biological invasion to the field of human urban habitat creation, ecological restoration, and ecosystem services. (2) The results of keyword emergence and keyword growth trend analysis show that habitat creation research has shown a rapid and stable growth trend since 2017, and ecological restoration has gained long-term sustained attention since 2004. The hotspots of future research on urban plants and habitats in China may focus on habitat creation and ecological restoration.

Keywords: research trends, visual analysis, habitat creation, ecological restoration

Procedia PDF Downloads 44
343 Anomaly Detection in Financial Markets Using Tucker Decomposition

Authors: Salma Krafessi

Abstract:

The financial markets have a multifaceted, intricate environment, and enormous volumes of data are produced every day. To find investment possibilities, possible fraudulent activity, and market oddities, accurate anomaly identification in this data is essential. Conventional methods for detecting anomalies frequently fail to capture the complex organization of financial data. In order to improve the identification of abnormalities in financial time series data, this study presents Tucker Decomposition as a reliable multi-way analysis approach. We start by gathering closing prices for the S&P 500 index across a number of decades. The information is converted to a three-dimensional tensor format, which contains internal characteristics and temporal sequences in a sliding window structure. The tensor is then broken down using Tucker Decomposition into a core tensor and matching factor matrices, allowing latent patterns and relationships in the data to be captured. A possible sign of abnormalities is the reconstruction error from Tucker's Decomposition. We are able to identify large deviations that indicate unusual behavior by setting a statistical threshold. A thorough examination that contrasts the Tucker-based method with traditional anomaly detection approaches validates our methodology. The outcomes demonstrate the superiority of Tucker's Decomposition in identifying intricate and subtle abnormalities that are otherwise missed. This work opens the door for more research into multi-way data analysis approaches across a range of disciplines and emphasizes the value of tensor-based methods in financial analysis.

Keywords: tucker decomposition, financial markets, financial engineering, artificial intelligence, decomposition models

Procedia PDF Downloads 28
342 Contemporary Army Prints for Women’s Wear Kurti

Authors: Shaleni Bajpai, Nancy Stephan

Abstract:

Various designs of women’s kurtis with different styles, motifs and prints were available in market but none of the kurtis was found in army print. Mostly army prints are used for men’s wear like jackets, trousers, caps, bags. The main colours available in military prints were beige, parrot green, red, dark blue, light blue, orange, bottle green, pink and the original military green colour. As the original camouflage is banned in civil wears so the different variety and colours were used in this study to popularize army prints in women’s wear. The aim of this project was to construct different styles of women kurti’s with various colours of different military prints. Mood board, inspiration and colour board was prepared to design the kurtis. The fabric used for construction was army printed poplin and crepe. The designing and construction of kurti’s were divided into two categories such as - casual and party wear. Casual wear had simple silhouette like a-line, high-low and waist coat style whereas party wear included princess line, panelled and bandhani style. Structured questionnaire was prepared to assess the acceptance of newly designed kurtis with respect to colour combination, overall appearance and cost. Purposively sampling method was adopted for selection of respondents. Opinion was taken from 100 women of various age groups. The result and analysis was presented through graph and percentage. Kurtis in army print of both the categories were appreciated by the respondents.

Keywords: army, kurti, casual wear, party wear

Procedia PDF Downloads 277
341 Observation of the Flow Behavior for a Rising Droplet in a Mini-Slot

Authors: H. Soltani, J. Hadfield, M. Redmond, D. S. Nobes

Abstract:

The passage of oil droplets through a vertical mini-slot were investigated in this study. Oil-in-water emulsion can undergo coalescence of finer oil droplets forming droplets of a size that need to be considered individually. This occurs in a number of industrial processes and has important consequences at a scale where both body and surfaces forces are relevant. In the study, two droplet diameters of smaller than the slot width and a relatively larger diameter where the oil droplet can interact directly with the slot wall were generated. To monitor fluid motion, a particle shadow velocimetry (PSV) imaging technique was used to study fluid flow motion inside and around a single oil droplet rising in a net co-flow. The droplet was a transparent canola oil and the surrounding working fluid was glycerol, adjusted to allow a matching of refractive index between the two fluids. Particles seeded in both fluids were observed with the PSV system allowing the capture of the velocity field both within the droplet and in the surrounds. The effect of droplet size on the droplet internal circulation was observed. Part of the study was related the potential generation of flow structures, such as von Karman vortex shedding already observed in rising droplets in infinite reservoirs and their interaction with the mini-channel. Results show that two counter-rotating vortices exist inside the droplets as they pass through slot. The vorticity map analysis shows that the droplet of relatively larger size has a stronger internal circulation.

Keywords: rising droplet, rectangular orifice, particle shadow velocimetry, match refractive index

Procedia PDF Downloads 154
340 ADP Approach to Evaluate the Blood Supply Network of Ontario

Authors: Usama Abdulwahab, Mohammed Wahab

Abstract:

This paper presents the application of uncapacitated facility location problems (UFLP) and 1-median problems to support decision making in blood supply chain networks. A plethora of factors make blood supply-chain networks a complex, yet vital problem for the regional blood bank. These factors are rapidly increasing demand; criticality of the product; strict storage and handling requirements; and the vastness of the theater of operations. As in the UFLP, facilities can be opened at any of $m$ predefined locations with given fixed costs. Clients have to be allocated to the open facilities. In classical location models, the allocation cost is the distance between a client and an open facility. In this model, the costs are the allocation cost, transportation costs, and inventory costs. In order to address this problem the median algorithm is used to analyze inventory, evaluate supply chain status, monitor performance metrics at different levels of granularity, and detect potential problems and opportunities for improvement. The Euclidean distance data for some Ontario cities (demand nodes) are used to test the developed algorithm. Sitation software, lagrangian relaxation algorithm, and branch and bound heuristics are used to solve this model. Computational experiments confirm the efficiency of the proposed approach. Compared to the existing modeling and solution methods, the median algorithm approach not only provides a more general modeling framework but also leads to efficient solution times in general.

Keywords: approximate dynamic programming, facility location, perishable product, inventory model, blood platelet, P-median problem

Procedia PDF Downloads 485
339 MIMIC: A Multi Input Micro-Influencers Classifier

Authors: Simone Leonardi, Luca Ardito

Abstract:

Micro-influencers are effective elements in the marketing strategies of companies and institutions because of their capability to create an hyper-engaged audience around a specific topic of interest. In recent years, many scientific approaches and commercial tools have handled the task of detecting this type of social media users. These strategies adopt solutions ranging from rule based machine learning models to deep neural networks and graph analysis on text, images, and account information. This work compares the existing solutions and proposes an ensemble method to generalize them with different input data and social media platforms. The deployed solution combines deep learning models on unstructured data with statistical machine learning models on structured data. We retrieve both social media accounts information and multimedia posts on Twitter and Instagram. These data are mapped into feature vectors for an eXtreme Gradient Boosting (XGBoost) classifier. Sixty different topics have been analyzed to build a rule based gold standard dataset and to compare the performances of our approach against baseline classifiers. We prove the effectiveness of our work by comparing the accuracy, precision, recall, and f1 score of our model with different configurations and architectures. We obtained an accuracy of 0.91 with our best performing model.

Keywords: deep learning, gradient boosting, image processing, micro-influencers, NLP, social media

Procedia PDF Downloads 151
338 Computational Identification of Signalling Pathways in Protein Interaction Networks

Authors: Angela U. Makolo, Temitayo A. Olagunju

Abstract:

The knowledge of signaling pathways is central to understanding the biological mechanisms of organisms since it has been identified that in eukaryotic organisms, the number of signaling pathways determines the number of ways the organism will react to external stimuli. Signaling pathways are studied using protein interaction networks constructed from protein-protein interaction data obtained using high throughput experimental procedures. However, these high throughput methods are known to produce very high rates of false positive and negative interactions. In order to construct a useful protein interaction network from this noisy data, computational methods are applied to validate the protein-protein interactions. In this study, a computational technique to identify signaling pathways from a protein interaction network constructed using validated protein-protein interaction data was designed. A weighted interaction graph of the Saccharomyces cerevisiae (Baker’s Yeast) organism using the proteins as the nodes and interactions between them as edges was constructed. The weights were obtained using Bayesian probabilistic network to estimate the posterior probability of interaction between two proteins given the gene expression measurement as biological evidence. Only interactions above a threshold were accepted for the network model. A pathway was formalized as a simple path in the interaction network from a starting protein and an ending protein of interest. We were able to identify some pathway segments, one of which is a segment of the pathway that signals the start of the process of meiosis in S. cerevisiae.

Keywords: Bayesian networks, protein interaction networks, Saccharomyces cerevisiae, signalling pathways

Procedia PDF Downloads 513
337 The Difference of Learning Outcomes in Reading Comprehension between Text and Film as The Media in Indonesian Language for Foreign Speaker in Intermediate Level

Authors: Siti Ayu Ningsih

Abstract:

This study aims to find the differences outcomes in learning reading comprehension with text and film as media on Indonesian Language for foreign speaker (BIPA) learning at intermediate level. By using quantitative and qualitative research methods, the respondent of this study is a single respondent from D'Royal Morocco Integrative Islamic School in grade nine from secondary level. Quantitative method used to calculate the learning outcomes that have been given the appropriate action cycle, whereas qualitative method used to translate the findings derived from quantitative methods to be described. The technique used in this study is the observation techniques and testing work. Based on the research, it is known that the use of the text media is more effective than the film for intermediate level of Indonesian Language for foreign speaker learner. This is because, when using film the learner does not have enough time to take note the difficult vocabulary and don't have enough time to look for the meaning of the vocabulary from the dictionary. While the use of media texts shows the better effectiveness because it does not require additional time to take note the difficult words. For the words that are difficult or strange, the learner can immediately find its meaning from the dictionary. The presence of the text is also very helpful for Indonesian Language for foreign speaker learner to find the answers according to the questions more easily. By matching the vocabulary of the question into the text references.

Keywords: Indonesian language for foreign speaker, learning outcome, media, reading comprehension

Procedia PDF Downloads 179
336 Collaboration of UNFPA and USAID to Mobilize Domestic Government Resources for Contraceptive Procurement in Madagascar

Authors: Josiane Yaguibou, Ngoy Kishimba, Issiaka v. Coulibaly, Sabrina Pestilli, Falinirina Razanalison, Hantanirina Andremanisa

Abstract:

Background: In recent years, Madagascar has faced a significant reduction in donors’ financial resources for the purchase of contraceptive products to meet the family planning needs of the population. In order to ensure the sustainability of the family planning program in the current context, UNFPA Madagascar engaged in a series of initiatives with the ultimate scope of identifying sustainable financing mechanisms for the program. Program intervention: UNFPA Madagascar established a strict collaboration with USAID to engage in a series of joint advocacy and resource mobilization activities with the government. The following initiatives were conducted: (i) Organization of a high-level Round Table to engage the government; (ii) Support to the government in renewing the FP2030 Commitments; (iii) Signature of the Country Compact 2022-2024; (iv) Allocation of government funds in 2022 and 2023 of over 829,222 USD; (v) Obtaining a Matching Fund of 1.5 million USD from UNFPA to encourage the government to allocate resources for the purchase of contraceptive products. Program Implications: The collaboration and the joint advocacy made it possible to (i) have budgetary allocations from the government to purchase products in 2022 and 2023 with a significant reduction in financing gaps; (ii) to convince the government to seek additional financing from partners such as the World Bank which granted more than 8 million USD for the purchase of products; (iii) reduce stock shortages from more than 30% to 15%.

Keywords: UNFPA, USAID, collaboration, contraceptives

Procedia PDF Downloads 42
335 The Determination of the Phosphorous Solubility in the Iron by the Function of the Other Components

Authors: Andras Dezső, Peter Baumli, George Kaptay

Abstract:

The phosphorous is the important components in the steels, because it makes the changing of the mechanical properties and possibly modifying the structure. The phosphorous can be create the Fe3P compounds, what is segregated in the ferrite grain boundary in the intervals of the nano-, or microscale. This intermetallic compound is decreasing the mechanical properties, for example it makes the blue brittleness which means that the brittle created by the segregated particles at 200 ... 300°C. This work describes the phosphide solubility by the other components effect. We make calculations for the Ni, Mo, Cu, S, V, C, Si, Mn, and the Cr elements by the Thermo-Calc software. We predict the effects by approximate functions. The binary Fe-P system has a solubility line, which has a determinating equation. The result is below: lnwo = -3,439 – 1.903/T where the w0 means the weight percent of the maximum soluted concentration of the phosphorous, and the T is the temperature in Kelvin. The equation show that the P more soluble element when the temperature increasing. The nickel, molybdenum, vanadium, silicon, manganese, and the chromium make dependence to the maximum soluted concentration. These functions are more dependent by the elements concentration, which are lower when we put these elements in our steels. The copper, sulphur and carbon do not make effect to the phosphorous solubility. We predict that all of cases the maximum solubility concentration increases when the temperature more and more high. Between 473K and 673 K, in the phase diagram, these systems contain mostly two or three phase eutectoid, and the singe phase, ferritic intervals. In the eutectoid areas the ferrite, the iron-phosphide, and the metal (III)-phospide are in the equilibrium. In these modelling we predicted that which elements are good for avoid the phosphide segregation or not. These datas are important when we make or choose the steels, where the phosphide segregation stopping our possibilities.

Keywords: phosphorous, steel, segregation, thermo-calc software

Procedia PDF Downloads 603
334 Graph-Based Semantical Extractive Text Analysis

Authors: Mina Samizadeh

Abstract:

In the past few decades, there has been an explosion in the amount of available data produced from various sources with different topics. The availability of this enormous data necessitates us to adopt effective computational tools to explore the data. This leads to an intense growing interest in the research community to develop computational methods focused on processing this text data. A line of study focused on condensing the text so that we are able to get a higher level of understanding in a shorter time. The two important tasks to do this are keyword extraction and text summarization. In keyword extraction, we are interested in finding the key important words from a text. This makes us familiar with the general topic of a text. In text summarization, we are interested in producing a short-length text which includes important information about the document. The TextRank algorithm, an unsupervised learning method that is an extension of the PageRank (algorithm which is the base algorithm of Google search engine for searching pages and ranking them), has shown its efficacy in large-scale text mining, especially for text summarization and keyword extraction. This algorithm can automatically extract the important parts of a text (keywords or sentences) and declare them as a result. However, this algorithm neglects the semantic similarity between the different parts. In this work, we improved the results of the TextRank algorithm by incorporating the semantic similarity between parts of the text. Aside from keyword extraction and text summarization, we develop a topic clustering algorithm based on our framework, which can be used individually or as a part of generating the summary to overcome coverage problems.

Keywords: keyword extraction, n-gram extraction, text summarization, topic clustering, semantic analysis

Procedia PDF Downloads 46
333 Development of a Plug-In Hybrid Powertrain System with Double Continuously Variable Transmissions

Authors: Cheng-Chi Yu, Chi-Shiun Chiou

Abstract:

This study developed a plug-in hybrid powertrain system which consisted of two continuous variable transmissions. By matching between the engine, motor, generator, and dual continuous variable transmissions, this integrated power system can take advantages of the components. The hybrid vehicle can be driven by the internal combustion engine, or electric motor alone, or by these two power sources together when the vehicle is driven in hard acceleration or high load. The energy management of this integrated hybrid system controls the power systems based on rule-based control strategy to achieve better fuel economy. When the vehicle driving power demand is low, the internal combustion engine is operating in the low efficiency region, so the internal combustion engine is shut down, and the vehicle is driven by motor only. When the vehicle driving power demand is high, internal combustion engine would operate in the high efficiency region; then the vehicle could be driven by internal combustion engine. This strategy would operate internal combustion engine only in optimal efficiency region to improve the fuel economy. In this research, the vehicle simulation model was built in MATLAB/ Simulink environment. The analysis results showed that the power coupled efficiency of the hybrid powertrain system with dual continuous variable transmissions was better than that of the Honda hybrid system on the market.

Keywords: plug-in hybrid power system, fuel economy, performance, continuously variable transmission

Procedia PDF Downloads 265
332 The Role of Attachment and Dyadic Coping in Shaping Relational Intimacy

Authors: Anna Wendolowska, Dorota Czyzowska

Abstract:

An intimate relationship is a significant factor that influences romantic partners’ well-being. In the face of stress, avoidant partners often employ a defense-against-intimacy strategy, leading to reduced relationship satisfaction, intimacy, interdependence, and longevity. Dyadic coping can buffer the negative effects of stress on relational satisfaction. Emotional competence mediates the relationship between insecure attachment and intimacy. In the current study, the link between attachment, different forms of dyadic coping, and various aspects of relationship satisfaction was examined. Both partners completed the attachment style questionnaire, the well matching couple questionnaire, and the dyadic coping inventory. The data was analyzed using the actor–partner interdependence model. The results highlighted a negative association between insecure-avoidant attachment style and intimacy. The actor effects of avoidant attachment on relational intimacy for women and for men were significant, whilst the partner effects for both spouses were not significant. The emotion-focused common dyadic coping moderated the relationship between avoidance of attachment and the partner's sense of intimacy. After controlling for the emotion-focused common dyadic coping, the actor effect of attachment on intimacy for men was slightly weaker, and the actor effect for women turned out to be insignificant. The emotion-focused common dyadic coping weakened the negative association between insecure attachment and relational intimacy. The impact of adult attachment and dyadic coping significantly contributes to subjective relational well-being.

Keywords: adult attachment, dyadic coping, relational intimacy, relationship satisfaction

Procedia PDF Downloads 133
331 Clarifying the Possible Symptomatic Pathway of Comorbid Depression, Anxiety, and Stress Among Adolescents Exposed to Childhood Trauma: Insight from the Network Approach

Authors: Xinyuan Zou, Qihui Tang, Shujian Wang, Yulin Huang, Jie Gui, Xiangping Liu, Gang Liu, Yanqiang Tao

Abstract:

Childhood trauma can have a long-lasting influence on individuals and contribute to mental disorders, including depression and anxiety. The current study aimed to explore the symptomatic and developmental patterns of depression, anxiety, and stress among adolescents who have suffered from childhood trauma. A total of 3,598 college students (female = 1,617 (44.94%), Mean Age = 19.68, SD Age = 1.35) in China completed the Childhood Trauma Questionnaire (CTQ) and the Depression, Anxiety, and Stress Scales (DASS-21), and 2,337 participants met the selection standard based on the cut-off scores of the CTQ. The symptomatic network and directed acyclic graph (DAG) network approaches were used. The results revealed that males reported experiencing significantly more physical abuse, physical neglect, emotional neglect, and sexual abuse compared to females. However, females scored significantly higher than males on all items of DASS-21, except for “Worthless”. No significant difference between the two genders was observed in the network structure and global strength. Meanwhile, among all participants, “Down-hearted” and “Agitated” appeared to be the most interconnected symptoms, the bridge symptoms in the symptom network, as well as the most vital symptoms in the DAG network. Apart from that, “No-relax” also served as the most prominent symptom in the DAG network. The results suggested that intervention targeted at assisting adolescents in developing more adaptive coping strategies with stress and regulating emotion could benefit the alleviation of comorbid depression, anxiety, and stress.

Keywords: symptom network, childhood trauma, depression, anxiety, stress

Procedia PDF Downloads 30
330 Cognitive Model of Analogy Based on Operation of the Brain Cells: Glial, Axons and Neurons

Authors: Ozgu Hafizoglu

Abstract:

Analogy is an essential tool of human cognition that enables connecting diffuse and diverse systems with attributional, deep structural, casual relations that are essential to learning, to innovation in artificial worlds, and to discovery in science. Cognitive Model of Analogy (CMA) leads and creates information pattern transfer within and between domains and disciplines in science. This paper demonstrates the Cognitive Model of Analogy (CMA) as an evolutionary approach to scientific research. The model puts forward the challenges of deep uncertainty about the future, emphasizing the need for flexibility of the system in order to enable reasoning methodology to adapt to changing conditions. In this paper, the model of analogical reasoning is created based on brain cells, their fractal, and operational forms within the system itself. Visualization techniques are used to show correspondences. Distinct phases of the problem-solving processes are divided thusly: encoding, mapping, inference, and response. The system is revealed relevant to brain activation considering each of these phases with an emphasis on achieving a better visualization of the brain cells: glial cells, axons, axon terminals, and neurons, relative to matching conditions of analogical reasoning and relational information. It’s found that encoding, mapping, inference, and response processes in four-term analogical reasoning are corresponding with the fractal and operational forms of brain cells: glial, axons, and neurons.

Keywords: analogy, analogical reasoning, cognitive model, brain and glials

Procedia PDF Downloads 163
329 Maximum Likelihood Estimation Methods on a Two-Parameter Rayleigh Distribution under Progressive Type-Ii Censoring

Authors: Daniel Fundi Murithi

Abstract:

Data from economic, social, clinical, and industrial studies are in some way incomplete or incorrect due to censoring. Such data may have adverse effects if used in the estimation problem. We propose the use of Maximum Likelihood Estimation (MLE) under a progressive type-II censoring scheme to remedy this problem. In particular, maximum likelihood estimates (MLEs) for the location (µ) and scale (λ) parameters of two Parameter Rayleigh distribution are realized under a progressive type-II censoring scheme using the Expectation-Maximization (EM) and the Newton-Raphson (NR) algorithms. These algorithms are used comparatively because they iteratively produce satisfactory results in the estimation problem. The progressively type-II censoring scheme is used because it allows the removal of test units before the termination of the experiment. Approximate asymptotic variances and confidence intervals for the location and scale parameters are derived/constructed. The efficiency of EM and the NR algorithms is compared given root mean squared error (RMSE), bias, and the coverage rate. The simulation study showed that in most sets of simulation cases, the estimates obtained using the Expectation-maximization algorithm had small biases, small variances, narrower/small confidence intervals width, and small root of mean squared error compared to those generated via the Newton-Raphson (NR) algorithm. Further, the analysis of a real-life data set (data from simple experimental trials) showed that the Expectation-Maximization (EM) algorithm performs better compared to Newton-Raphson (NR) algorithm in all simulation cases under the progressive type-II censoring scheme.

Keywords: expectation-maximization algorithm, maximum likelihood estimation, Newton-Raphson method, two-parameter Rayleigh distribution, progressive type-II censoring

Procedia PDF Downloads 133
328 Multi-Dimension Threat Situation Assessment Based on Network Security Attributes

Authors: Yang Yu, Jian Wang, Jiqiang Liu, Lei Han, Xudong He, Shaohua Lv

Abstract:

As the increasing network attacks become more and more complex, network situation assessment based on log analysis cannot meet the requirements to ensure network security because of the low quality of logs and alerts. This paper addresses the lack of consideration of security attributes of hosts and attacks in the network. Identity and effectiveness of Distributed Denial of Service (DDoS) are hard to be proved in risk assessment based on alerts and flow matching. This paper proposes a multi-dimension threat situation assessment method based on network security attributes. First, the paper offers an improved Common Vulnerability Scoring System (CVSS) calculation, which includes confident risk, integrity risk, availability risk and a weighted risk. Second, the paper introduces deterioration rate of properties collected by sensors in hosts and network, which aimed at assessing the time and level of DDoS attacks. Third, the paper introduces distribution of asset value in security attributes considering features of attacks and network, which aimed at assessing and show the whole situation. Experiments demonstrate that the approach reflects effectiveness and level of DDoS attacks, and the result can show the primary threat in network and security requirement of network. Through comparison and analysis, the method reflects more in security requirement and security risk situation than traditional methods based on alert and flow analyzing.

Keywords: DDoS evaluation, improved CVSS, network security attribute, threat situation assessment

Procedia PDF Downloads 184
327 Discrimination and Classification of Vestibular Neuritis Using Combined Fisher and Support Vector Machine Model

Authors: Amine Ben Slama, Aymen Mouelhi, Sondes Manoubi, Chiraz Mbarek, Hedi Trabelsi, Mounir Sayadi, Farhat Fnaiech

Abstract:

Vertigo is a sensation of feeling off balance; the cause of this symptom is very difficult to interpret and needs a complementary exam. Generally, vertigo is caused by an ear problem. Some of the most common causes include: benign paroxysmal positional vertigo (BPPV), Meniere's disease and vestibular neuritis (VN). In clinical practice, different tests of videonystagmographic (VNG) technique are used to detect the presence of vestibular neuritis (VN). The topographical diagnosis of this disease presents a large diversity in its characteristics that confirm a mixture of problems for usual etiological analysis methods. In this study, a vestibular neuritis analysis method is proposed with videonystagmography (VNG) applications using an estimation of pupil movements in the case of an uncontrolled motion to obtain an efficient and reliable diagnosis results. First, an estimation of the pupil displacement vectors using with Hough Transform (HT) is performed to approximate the location of pupil region. Then, temporal and frequency features are computed from the rotation angle variation of the pupil motion. Finally, optimized features are selected using Fisher criterion evaluation for discrimination and classification of the VN disease.Experimental results are analyzed using two categories: normal and pathologic. By classifying the reduced features using the Support Vector Machine (SVM), 94% is achieved as classification accuracy. Compared to recent studies, the proposed expert system is extremely helpful and highly effective to resolve the problem of VNG analysis and provide an accurate diagnostic for medical devices.

Keywords: nystagmus, vestibular neuritis, videonystagmographic system, VNG, Fisher criterion, support vector machine, SVM

Procedia PDF Downloads 120
326 A Single Stage Cleft Rhinoplasty Technique for Primary Unilateral Cleft Lip and Palate 'The Gujrat Technique'

Authors: Diaa Othman, Muhammad Adil Khan, Muhammad Riaz

Abstract:

Without an early intervention to correct the unilateral complete cleft lip and palate deformity, nasal architecture can progress to an exaggerated cleft nose deformity. We present the results of a modified unilateral cleft rhinoplasty procedure ‘the Gujrat technique’ to correct this deformity. Ninety pediatric and adult patients with non-syndromic unilateral cleft lip underwent primary and secondary composite cleft rhinoplasty using the Gujrat technique as a single stage operation over a 10-year period. The technique involved an open rhinoplasty with Tennison lip repair, and employed a combination of three autologous cartilage grafts, seven cartilage-molding sutures and a prolene mesh graft for alar base support. Post-operative evaluation of nasal symmetry was undertaken using the validated computer program ‘SymNose’. Functional outcome and patient satisfaction were assessed using the NOSE scale and ROE (rhinoplasty outcome evaluation) questionnaires. The single group study design used the non-parametric matching pairs Wilcoxon Sign test (p < 0.001), and showed overall good to excellent functional and aesthetic outcomes, including nasal projection and tip definition, and higher scores of the digital SymNose grading system. Objective assessment of the Gujrat cleft rhinoplasty technique demonstrates its aesthetic appeal and functional versatility. Overall it is a simple and reproducible technique, with no significant complications.

Keywords: cleft lip and palate, congenital rhinoplasty, nasal deformity, secondary rhinoplasty

Procedia PDF Downloads 180
325 A Robust Spatial Feature Extraction Method for Facial Expression Recognition

Authors: H. G. C. P. Dinesh, G. Tharshini, M. P. B. Ekanayake, G. M. R. I. Godaliyadda

Abstract:

This paper presents a new spatial feature extraction method based on principle component analysis (PCA) and Fisher Discernment Analysis (FDA) for facial expression recognition. It not only extracts reliable features for classification, but also reduces the feature space dimensions of pattern samples. In this method, first each gray scale image is considered in its entirety as the measurement matrix. Then, principle components (PCs) of row vectors of this matrix and variance of these row vectors along PCs are estimated. Therefore, this method would ensure the preservation of spatial information of the facial image. Afterwards, by incorporating the spectral information of the eigen-filters derived from the PCs, a feature vector was constructed, for a given image. Finally, FDA was used to define a set of basis in a reduced dimension subspace such that the optimal clustering is achieved. The method of FDA defines an inter-class scatter matrix and intra-class scatter matrix to enhance the compactness of each cluster while maximizing the distance between cluster marginal points. In order to matching the test image with the training set, a cosine similarity based Bayesian classification was used. The proposed method was tested on the Cohn-Kanade database and JAFFE database. It was observed that the proposed method which incorporates spatial information to construct an optimal feature space outperforms the standard PCA and FDA based methods.

Keywords: facial expression recognition, principle component analysis (PCA), fisher discernment analysis (FDA), eigen-filter, cosine similarity, bayesian classifier, f-measure

Procedia PDF Downloads 406
324 Effect of Ausubel's Advance Organizer Model to Enhancing Meta-Cognition of Students at Secondary Level

Authors: Qaisara Parveen, M. Imran Yousuf

Abstract:

The purpose of this study was to find the effectiveness of the use of advance organizer model for enhancing meta-cognition of students in the subject of science. It was hypothesized that the students of experimental group taught through advance organizer model would show the better cognition than the students of control group taught through traditional teaching. The population of the study consisted of all secondary school students studying in government high school located in Rawalpindi. The sample of the study consisted of 50 students of 9th class of humanities group. The sample was selected on the basis of their pretest scores through matching, and the groups were randomly assigned for the treatment. The experimental group was taught through advance organizer model while the control group was taught through traditional teaching. The self-developed achievement test was used for the purpose of pretest and posttest. After collecting the pre-test score and post-test score, the data was analyzed and interpreted by use of descriptive statistics as mean and standard deviation and inferential statistics t-test. The findings indicate that students taught using advance organizers had a higher level of meta-cognition as compared to control group. Further, meta cognition level of boys was found higher than that of girls students. This study also revealed the fact that though the students at different meta-cognition level approached learning situations in a different manner, Advance organizer model is far superior to Traditional method of teaching.

Keywords: descriptive, experimental, humanities, meta-cognition, statistics, science

Procedia PDF Downloads 287
323 Environment Management Practices at Oil and Natural Gas Corporation Hazira Gas Processing Complex

Authors: Ashish Agarwal, Vaibhav Singh

Abstract:

Harmful emissions from oil and gas processing facilities have long remained a matter of concern for governments and environmentalists throughout the world. This paper analyses Oil and Natural Gas Corporation (ONGC) gas processing plant in Hazira, Gujarat, India. It is the largest gas-processing complex in the country designed to process 41MMSCMD sour natural gas & associated sour condensate. The complex, sprawling over an area of approximate 705 hectares is the mother plant for almost all industries at Hazira and enroute Hazira Bijapur Jagdishpur pipeline. Various sources of pollution from each unit starting from Gas Terminal to Dew Point Depression unit and Caustic Wash unit along the processing chain were examined with the help of different emission data obtained from ONGC. Pollution discharged to the environment was classified into Water, Air, Hazardous Waste and Solid (Non-Hazardous) Waste so as to analyze each one of them efficiently. To protect air environment, Sulphur recovery unit along with automatic ambient air quality monitoring stations, automatic stack monitoring stations among numerous practices were adopted. To protect water environment different effluent treatment plants were used with due emphasis on aquaculture of the nearby area. Hazira plant has obtained the authorization for handling and disposal of five types of hazardous waste. Most of the hazardous waste were sold to authorized recyclers and the rest was given to Gujarat Pollution Control Board authorized vendors. Non-Hazardous waste was also handled with an overall objective of zero negative impact on the environment. The effect of methods adopted is evident from emission data of the plant which was found to be well under Gujarat Pollution Control Board limits.

Keywords: sulphur recovery unit, effluent treatment plant, hazardous waste, sour gas

Procedia PDF Downloads 200
322 An Adaptive Back-Propagation Network and Kalman Filter Based Multi-Sensor Fusion Method for Train Location System

Authors: Yu-ding Du, Qi-lian Bao, Nassim Bessaad, Lin Liu

Abstract:

The Global Navigation Satellite System (GNSS) is regarded as an effective approach for the purpose of replacing the large amount used track-side balises in modern train localization systems. This paper describes a method based on the data fusion of a GNSS receiver sensor and an odometer sensor that can significantly improve the positioning accuracy. A digital track map is needed as another sensor to project two-dimensional GNSS position to one-dimensional along-track distance due to the fact that the train’s position can only be constrained on the track. A model trained by BP neural network is used to estimate the trend positioning error which is related to the specific location and proximate processing of the digital track map. Considering that in some conditions the satellite signal failure will lead to the increase of GNSS positioning error, a detection step for GNSS signal is applied. An adaptive weighted fusion algorithm is presented to reduce the standard deviation of train speed measurement. Finally an Extended Kalman Filter (EKF) is used for the fusion of the projected 1-D GNSS positioning data and the 1-D train speed data to get the estimate position. Experimental results suggest that the proposed method performs well, which can reduce positioning error notably.

Keywords: multi-sensor data fusion, train positioning, GNSS, odometer, digital track map, map matching, BP neural network, adaptive weighted fusion, Kalman filter

Procedia PDF Downloads 227
321 Finite Element Analysis of a Modular Brushless Wound Rotor Synchronous Machine

Authors: H. T. Le Luong, C. Hénaux, F. Messine, G. Bueno-Mariani, S. Mollov, N. Voyer

Abstract:

This paper presents a comparative study of different modular brushless wound rotor synchronous machine (MB-WRSM). The goal of the study is to highlight the structure which offers the best fault tolerant capability and the highest output performances. The fundamental winding factor is calculated by using the method based on EMF phasors as a significant criterion to select the preferred number of phases, stator slots, and poles. With the limited number of poles for a small machine (3.67kW/7000rpm), 15 different machines for preferred phase/slot/pole combinations are analyzed using two-dimensional (2-D) finite element method and compared according to three criteria: torque density, torque ripple and efficiency. The 7phase/7slot/6pole machine is chosen with the best compromise of high torque density, small torque ripple (3.89%) and high nominal efficiency (95%). This machine is then compared with a reference design surface permanent magnet synchronous machine (SPMSM). In conclusion, this paper provides an electromagnetic analysis of a new brushless wound-rotor synchronous machine using multiphase non-overlapping fractional slot double layer winding. The simulation results are discussed and demonstrate that the MB-WRSM presents interesting performance features, with overall performance closely matching that of an equivalent SPMSM.

Keywords: finite element method (FEM), machine performance, modular wound rotor synchronous machine, non-overlapping concentrated winding

Procedia PDF Downloads 268
320 Q-Map: Clinical Concept Mining from Clinical Documents

Authors: Sheikh Shams Azam, Manoj Raju, Venkatesh Pagidimarri, Vamsi Kasivajjala

Abstract:

Over the past decade, there has been a steep rise in the data-driven analysis in major areas of medicine, such as clinical decision support system, survival analysis, patient similarity analysis, image analytics etc. Most of the data in the field are well-structured and available in numerical or categorical formats which can be used for experiments directly. But on the opposite end of the spectrum, there exists a wide expanse of data that is intractable for direct analysis owing to its unstructured nature which can be found in the form of discharge summaries, clinical notes, procedural notes which are in human written narrative format and neither have any relational model nor any standard grammatical structure. An important step in the utilization of these texts for such studies is to transform and process the data to retrieve structured information from the haystack of irrelevant data using information retrieval and data mining techniques. To address this problem, the authors present Q-Map in this paper, which is a simple yet robust system that can sift through massive datasets with unregulated formats to retrieve structured information aggressively and efficiently. It is backed by an effective mining technique which is based on a string matching algorithm that is indexed on curated knowledge sources, that is both fast and configurable. The authors also briefly examine its comparative performance with MetaMap, one of the most reputed tools for medical concepts retrieval and present the advantages the former displays over the latter.

Keywords: information retrieval, unified medical language system, syntax based analysis, natural language processing, medical informatics

Procedia PDF Downloads 110
319 An Application of Path Planning Algorithms for Autonomous Inspection of Buried Pipes with Swarm Robots

Authors: Richard Molyneux, Christopher Parrott, Kirill Horoshenkov

Abstract:

This paper aims to demonstrate how various algorithms can be implemented within swarms of autonomous robots to provide continuous inspection within underground pipeline networks. Current methods of fault detection within pipes are costly, time consuming and inefficient. As such, solutions tend toward a more reactive approach, repairing faults, as opposed to proactively seeking leaks and blockages. The paper presents an efficient inspection method, showing that autonomous swarm robotics is a viable way of monitoring underground infrastructure. Tailored adaptations of various Vehicle Routing Problems (VRP) and path-planning algorithms provide a customised inspection procedure for complicated networks of underground pipes. The performance of multiple algorithms is compared to determine their effectiveness and feasibility. Notable inspirations come from ant colonies and stigmergy, graph theory, the k-Chinese Postman Problem ( -CPP) and traffic theory. Unlike most swarm behaviours which rely on fast communication between agents, underground pipe networks are a highly challenging communication environment with extremely limited communication ranges. This is due to the extreme variability in the pipe conditions and relatively high attenuation of acoustic and radio waves with which robots would usually communicate. This paper illustrates how to optimise the inspection process and how to increase the frequency with which the robots pass each other, without compromising the routes they are able to take to cover the whole network.

Keywords: autonomous inspection, buried pipes, stigmergy, swarm intelligence, vehicle routing problem

Procedia PDF Downloads 138
318 Data Mining Spatial: Unsupervised Classification of Geographic Data

Authors: Chahrazed Zouaoui

Abstract:

In recent years, the volume of geospatial information is increasing due to the evolution of communication technologies and information, this information is presented often by geographic information systems (GIS) and stored on of spatial databases (BDS). The classical data mining revealed a weakness in knowledge extraction at these enormous amounts of data due to the particularity of these spatial entities, which are characterized by the interdependence between them (1st law of geography). This gave rise to spatial data mining. Spatial data mining is a process of analyzing geographic data, which allows the extraction of knowledge and spatial relationships from geospatial data, including methods of this process we distinguish the monothematic and thematic, geo- Clustering is one of the main tasks of spatial data mining, which is registered in the part of the monothematic method. It includes geo-spatial entities similar in the same class and it affects more dissimilar to the different classes. In other words, maximize intra-class similarity and minimize inter similarity classes. Taking account of the particularity of geo-spatial data. Two approaches to geo-clustering exist, the dynamic processing of data involves applying algorithms designed for the direct treatment of spatial data, and the approach based on the spatial data pre-processing, which consists of applying clustering algorithms classic pre-processed data (by integration of spatial relationships). This approach (based on pre-treatment) is quite complex in different cases, so the search for approximate solutions involves the use of approximation algorithms, including the algorithms we are interested in dedicated approaches (clustering methods for partitioning and methods for density) and approaching bees (biomimetic approach), our study is proposed to design very significant to this problem, using different algorithms for automatically detecting geo-spatial neighborhood in order to implement the method of geo- clustering by pre-treatment, and the application of the bees algorithm to this problem for the first time in the field of geo-spatial.

Keywords: mining, GIS, geo-clustering, neighborhood

Procedia PDF Downloads 359
317 Utilizing Fiber-Based Modeling to Explore the Presence of a Soft Storey in Masonry-Infilled Reinforced Concrete Structures

Authors: Akram Khelaifia, Salah Guettala, Nesreddine Djafar Henni, Rachid Chebili

Abstract:

Recent seismic events have underscored the significant influence of masonry infill walls on the resilience of structures. The irregular positioning of these walls exacerbates their adverse effects, resulting in substantial material and human losses. Research and post-earthquake evaluations emphasize the necessity of considering infill walls in both the design and assessment phases. This study delves into the presence of soft stories in reinforced concrete structures with infill walls. Employing an approximate method relying on pushover analysis results, fiber-section-based macro-modeling is utilized to simulate the behavior of infill walls. The findings shed light on the presence of soft first stories, revealing a notable 240% enhancement in resistance for weak column—strong beam-designed frames due to infill walls. Conversely, the effect is more moderate at 38% for strong column—weak beam-designed frames. Interestingly, the uniform distribution of infill walls throughout the structure's height does not influence soft-story emergence in the same seismic zone, irrespective of column-beam strength. In regions with low seismic intensity, infill walls dissipate energy, resulting in consistent seismic behavior regardless of column configuration. Despite column strength, structures with open-ground stories remain vulnerable to soft first-story emergence, underscoring the crucial role of infill walls in reinforced concrete structural design.

Keywords: masonry infill walls, soft Storey, pushover analysis, fiber section, macro-modeling

Procedia PDF Downloads 33
316 Spatial Integration at the Room-Level of 'Sequina' Slum Area in Alexandria, Egypt

Authors: Ali Essam El Shazly

Abstract:

The slum survey of 'Sequina' area in Alexandria details the building rooms of twenty-building samples according to the integral measure of space syntax. The essence of room organization sets the most integrative 'visitor' domain between the 'inhabitant' wings of less integrated 'parent' than the 'children' structure with visual ring of 'balcony' space. Despite the collective real relative asymmetry of 'pheno-type' aggregation, the relative asymmetry of individual layouts reveals 'geno-type' structure of spatial diversity. The multifunction of rooms optimizes the integral structure of graph and visibility merge, which contrasts with the deep tailing structure of distinctive social domains. The most integrative layout inverts the geno-type into freed rooms of shallow 'inhabitant' domain against the off-centered 'visitor' space, while the most segregated layout further restricts the pheno-type through isolated 'visitor' from 'inhabitant' domains across the 'staircase' public domain. The catalyst 'kitchen & living' spaces demonstrate multi-structural dimensions among the various social domains. The former ranges from most exposed central integrity to the most hidden 'motherhood' territories. The latter, however, mostly integrates at centrality or at the further ringy 'childern' domain. The study concludes social structure of spatial integrity for redevelopment, which is determined through the micro-level survey of rooms with integral dimensions.

Keywords: Alexandria, Sequina slum, spatial integration, space syntax

Procedia PDF Downloads 408