Search results for: chemical graph theory
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9114

Search results for: chemical graph theory

8934 A Study of Chinese-specific Terms in Government Work Report(2017-2019) from the Perspective of Relevance Theory

Authors: Shi Jiaxin

Abstract:

The Government Work Report is an essential form of document in the government of the People’s Republic of China. It covers all aspects of Chinese society and reflects China’s development strategy and trend. There are countless special terms in Government Work Report. Only by understanding Chinese-specific terms can we understand the content of the Government Work Report. Only by accurately translating the Chinese-specific terms can people come from all across the world know the Chinese government work report and understand China. Relevance theory is a popular theory of cognitive pragmatics. Relevance Translation Theory, which is closely related to Relevance Theory, has crucial and major guiding significance for the translation of Chinese-specific. Through studying Relevance Theory and researching the translation techniques, strategies and applications in the process of translating Chinese-specific terms from the perspective of Relevance Theory, we can understand the meaning and connotation of Chinese-specific terms, then solve various problems in the process of C-E translation, and strengthen our translation ability.

Keywords: government work report, Chinese-specific terms, relevance theory, translation

Procedia PDF Downloads 123
8933 Value from Environmental and Cultural Perspectives or Two Sides of the Same Coin

Authors: Vilem Paril, Dominika Tothova

Abstract:

This paper discusses the value theory in cultural heritage and the value theory in environmental economics. Two economic views of the value theory are compared within the field of cultural heritage maintenance and within the field of the environment. The main aims are to find common features in these two differently structured theories under the layer of differently defined terms as well as really differing features of these two approaches, to clear the confusion which stems from different terminology as in fact these terms capture the same aspects of reality and to show possible inspiration these two perspectives can offer one another. Another aim is to present these two value systems in one value framework. First, important moments of the value theory from the economic perspective are presented, leading to the marginal revolution of (not only) the Austrian School. Then the theory of value within cultural heritage and environmental economics are explored. Finally, individual approaches are compared and their potential mutual inspiration searched for.

Keywords: cultural heritage, environmental economics, existence value, value theory

Procedia PDF Downloads 285
8932 Modeling Aggregation of Insoluble Phase in Reactors

Authors: A. Brener, B. Ismailov, G. Berdalieva

Abstract:

In the paper we submit the modification of kinetic Smoluchowski equation for binary aggregation applying to systems with chemical reactions of first and second orders in which the main product is insoluble. The goal of this work is to create theoretical foundation and engineering procedures for calculating the chemical apparatuses in the conditions of joint course of chemical reactions and processes of aggregation of insoluble dispersed phases which are formed in working zones of the reactor.

Keywords: binary aggregation, clusters, chemical reactions, insoluble phases

Procedia PDF Downloads 269
8931 Sampling Effects on Secondary Voltage Control of Microgrids Based on Network of Multiagent

Authors: M. J. Park, S. H. Lee, C. H. Lee, O. M. Kwon

Abstract:

This paper studies a secondary voltage control framework of the microgrids based on the consensus for a communication network of multiagent. The proposed control is designed by the communication network with one-way links. The communication network is modeled by a directed graph. At this time, the concept of sampling is considered as the communication constraint among each distributed generator in the microgrids. To analyze the sampling effects on the secondary voltage control of the microgrids, by using Lyapunov theory and some mathematical techniques, the sufficient condition for such problem will be established regarding linear matrix inequality (LMI). Finally, some simulation results are given to illustrate the necessity of the consideration of the sampling effects on the secondary voltage control of the microgrids.

Keywords: microgrids, secondary control, multiagent, sampling, LMI

Procedia PDF Downloads 297
8930 Investigating the Neural Heterogeneity of Developmental Dyscalculia

Authors: Fengjuan Wang, Azilawati Jamaludin

Abstract:

Developmental Dyscalculia (DD) is defined as a particular learning difficulty with continuous challenges in learning requisite math skills that cannot be explained by intellectual disability or educational deprivation. Recent studies have increasingly recognized that DD is a heterogeneous, instead of monolithic, learning disorder with not only cognitive and behavioral deficits but so too neural dysfunction. In recent years, neuroimaging studies employed group comparison to explore the neural underpinnings of DD, which contradicted the heterogenous nature of DD and may obfuscate critical individual differences. This research aimed to investigate the neural heterogeneity of DD using case studies with functional near-infrared spectroscopy (fNIRS). A total of 54 aged 6-7 years old of children participated in this study, comprising two comprehensive cognitive assessments, an 8-minute resting state, and an 8-minute one-digit addition task. Nine children met the criteria of DD and scored at or below 85 (i.e., the 16th percentile) on the Mathematics or Math Fluency subtest of the Wechsler Individual Achievement Test, Third Edition (WIAT-III) (both subtest scores were 90 and below). The remaining 45 children formed the typically developing (TD) group. Resting-state data and brain activation in the inferior frontal gyrus (IFG), superior frontal gyrus (SFG), and intraparietal sulcus (IPS) were collected for comparison between each case and the TD group. Graph theory was used to analyze the brain network under the resting state. This theory represents the brain network as a set of nodes--brain regions—and edges—pairwise interactions across areas to reveal the architectural organizations of the nervous network. Next, a single-case methodology developed by Crawford et al. in 2010 was used to compare each case’s brain network indicators and brain activation against 45 TD children’s average data. Results showed that three out of the nine DD children displayed significant deviation from TD children’s brain indicators. Case 1 had inefficient nodal network properties. Case 2 showed inefficient brain network properties and weaker activation in the IFG and IPS areas. Case 3 displayed inefficient brain network properties with no differences in activation patterns. As a rise above, the present study was able to distill differences in architectural organizations and brain activation of DD vis-à-vis TD children using fNIRS and single-case methodology. Although DD is regarded as a heterogeneous learning difficulty, it is noted that all three cases showed lower nodal efficiency in the brain network, which may be one of the neural sources of DD. Importantly, although the current “brain norm” established for the 45 children is tentative, the results from this study provide insights not only for future work in “developmental brain norm” with reliable brain indicators but so too the viability of single-case methodology, which could be used to detect differential brain indicators of DD children for early detection and interventions.

Keywords: brain activation, brain network, case study, developmental dyscalculia, functional near-infrared spectroscopy, graph theory, neural heterogeneity

Procedia PDF Downloads 27
8929 Power Iteration Clustering Based on Deflation Technique on Large Scale Graphs

Authors: Taysir Soliman

Abstract:

One of the current popular clustering techniques is Spectral Clustering (SC) because of its advantages over conventional approaches such as hierarchical clustering, k-means, etc. and other techniques as well. However, one of the disadvantages of SC is the time consuming process because it requires computing the eigenvectors. In the past to overcome this disadvantage, a number of attempts have been proposed such as the Power Iteration Clustering (PIC) technique, which is one of versions from SC; some of PIC advantages are: 1) its scalability and efficiency, 2) finding one pseudo-eigenvectors instead of computing eigenvectors, and 3) linear combination of the eigenvectors in linear time. However, its worst disadvantage is an inter-class collision problem because it used only one pseudo-eigenvectors which is not enough. Previous researchers developed Deflation-based Power Iteration Clustering (DPIC) to overcome problems of PIC technique on inter-class collision with the same efficiency of PIC. In this paper, we developed Parallel DPIC (PDPIC) to improve the time and memory complexity which is run on apache spark framework using sparse matrix. To test the performance of PDPIC, we compared it to SC, ESCG, ESCALG algorithms on four small graph benchmark datasets and nine large graph benchmark datasets, where PDPIC proved higher accuracy and better time consuming than other compared algorithms.

Keywords: spectral clustering, power iteration clustering, deflation-based power iteration clustering, Apache spark, large graph

Procedia PDF Downloads 145
8928 Memetic Algorithm for Solving the One-To-One Shortest Path Problem

Authors: Omar Dib, Alexandre Caminada, Marie-Ange Manier

Abstract:

The purpose of this study is to introduce a novel approach to solve the one-to-one shortest path problem. A directed connected graph is assumed in which all edges’ weights are positive. Our method is based on a memetic algorithm in which we combine a genetic algorithm (GA) and a variable neighborhood search method (VNS). We compare our approximate method with two exact algorithms Dijkstra and Integer Programming (IP). We made experimentations using random generated, complete and real graph instances. In most case studies, numerical results show that our method outperforms exact methods with 5% average gap to the optimality. Our algorithm’s average speed is 20-times faster than Dijkstra and more than 1000-times compared to IP. The details of the experimental results are also discussed and presented in the paper.

Keywords: shortest path problem, Dijkstra’s algorithm, integer programming, memetic algorithm

Procedia PDF Downloads 430
8927 Exploring the Nature and Meaning of Theory in the Field of Neuroeducation Studies

Authors: Ali Nouri

Abstract:

Neuroeducation is one of the most exciting research fields which is continually evolving. However, there is a need to develop its theoretical bases in connection to practice. The present paper is a starting attempt in this regard to provide a space from which to think about neuroeducational theory and invoke more investigation in this area. Accordingly, a comprehensive theory of neuroeducation could be defined as grouping or clustering of concepts and propositions that describe and explain the nature of human learning to provide valid interpretations and implications useful for educational practice in relation to philosophical aspects or values. Whereas it should be originated from the philosophical foundations of the field and explain its normative significance, it needs to be testable in terms of rigorous evidence to fundamentally advance contemporary educational policy and practice. There is thus pragmatically a need to include a course on neuroeducational theory into the curriculum of the field. In addition, there is a need to articulate and disseminate considerable discussion over the subject within professional journals and academic societies.

Keywords: neuroeducation studies, neuroeducational theory, theory building, neuroeducation research

Procedia PDF Downloads 419
8926 Towards Establishing a Universal Theory of Project Management

Authors: Divine Kwaku Ahadzie

Abstract:

Project management (PM) as a concept has evolved from the early 20th Century into a recognized academic and professional discipline, and indications are that it has come to stay in the 21st Century as a world-wide paradigm shift for managing successful construction projects. However, notwithstanding the strong inroads that PM has made in legitimizing its academic and professional status in construction management practice, the underlining philosophies are still based on cases and conventional practices. An important theoretical issue yet to be addressed is the lack of a universal theory that offers philosophical legitimacy for the PM concept as a uniquely specialized management concept. Here, it is hypothesized that the law of entropy, the theory of uncertainties and the theory of risk management offer plausible explanations for addressing the lacuna of what constitute PM theory. The theoretical bases of these plausible underlying theories are argued and attempts made to establish the functional relationships that exist between these theories and the PM concept. The paper then draws on data related to the success and/or failure of a number of construction projects to validate the theory.

Keywords: concepts, construction, project management, universal theory

Procedia PDF Downloads 294
8925 Policy Compliance in Information Security

Authors: R. Manjula, Kaustav Bagchi, Sushant Ramesh, Anush Baskaran

Abstract:

In the past century, the emergence of information technology has had a significant positive impact on human life. While companies tend to be more involved in the completion of projects, the turn of the century has seen importance being given to investment in information security policies. These policies are essential to protect important data from adversaries, and thus following these policies has become one of the most important attributes revolving around information security models. In this research, we have focussed on the factors affecting information security policy compliance in two models : The theory of planned behaviour and the integration of the social bond theory and the involvement theory into a single model. Finally, we have given a proposal of where these theories would be successful.

Keywords: information technology, information security, involvement theory, policies, social bond theory

Procedia PDF Downloads 334
8924 Damage in Cementitious Materials Exposed to Sodium Chloride Solution and Thermal Cycling: The Effect of Using Supplementary Cementitious Materials

Authors: Fadi Althoey, Yaghoob Farnam

Abstract:

Sodium chloride (NaCl) can interact with the tricalcium aluminate (C3A) and its hydrates in concrete matrix. This interaction can result in formation of a harmful chemical phase as the temperature changes. It is thought that this chemical phase is embroiled in the premature concrete deterioration in the cold regions. This work examines the potential formation of the harmful chemical phase in various pastes prepared by using different types of ordinary portland cement (OPC) and supplementary cementitious materials (SCMs). The quantification of the chemical phase was done by using a low temperature differential scanning calorimetry. The results showed that the chemical phase formation can be reduced by using Type V cement (low content of C3A). The use of SCMs showed different behaviors on the formation of the chemical phase. Slag and Class F fly ash can reduce the chemical phase by the dilution of cement whereas silica fume can reduce the amount of the chemical phase by dilution and pozzolanic activates. Interestingly, the use of Class C fly ash has a negative effect on concrete exposed to NaCl through increasing the formation of the chemical phase.

Keywords: concrete, damage, chemcial phase, NaCl, SCMs

Procedia PDF Downloads 102
8923 Encapsulation of Volatile Citronella Essential oil by Coacervation: Efficiency and Release Kinetic Study

Authors: Rafeqah Raslan, Mastura AbdManaf, Junaidah Jai, Istikamah Subuki, Ana Najwa Mustapa

Abstract:

The volatile citronella essential oil was encapsulated by simple coacervation and complex coacervation using gum Arabic and gelatin as wall material. Glutaraldehyde was used in the methodology as crosslinking agent. The citronella standard calibration graph was developed with R2 equal to 0.9523 for the accurate determination of encapsulation efficiency and release study. The release kinetic was analyzed based on Fick’s law of diffusion for polymeric system and linear graph of log fraction release over log time was constructed to determine the release rate constant, k and diffusion coefficient, n. Both coacervation methods in the present study produce encapsulation efficiency around 94%. The capsules morphology analysis supported the release kinetic mechanisms of produced capsules for both coacervation process.

Keywords: simple coacervation, complex coacervation, encapsulation efficiency, release kinetic study

Procedia PDF Downloads 281
8922 Principles of Teaching for Successful Intelligence

Authors: Shabnam

Abstract:

The purpose of this study was to see importance of successful intelligence in education which can enhance achievement. There are a number of researches which have tried to apply psychological theories of education and many researches emphasized the role of thinking and intelligence. While going through the various researches, it was found that many students could learn more effectively than they do, if they were taught in a way that better matched their patterns of abilities. Attempts to apply psychological theories to education can falter on the translation of the theory into educational practice. Often, this translation is not clear. Therefore, when a program does not succeed, it is not clear whether the lack of success was due to the inadequacy of the theory or the inadequacy of the implementation of the theory. A set of basic principles for translating a theory into practice can help clarify just what an educational implementation should (and should not) look like. Sternberg’s theory of successful intelligence; analytical, creative and practical intelligence provides a way to create such a match. The results suggest that theory of successful intelligence provides successful interventions in classrooms and provides a proven model for gifted education. This article presents principles for translating a triarchic theory of successful intelligence into educational practice.

Keywords: successful intelligence, analytical, creative and practical intelligence, achievement, success, resilience

Procedia PDF Downloads 551
8921 Stress Concentration Trend for Combined Loading Conditions

Authors: Aderet M. Pantierer, Shmuel Pantierer, Raphael Cordina, Yougashwar Budhoo

Abstract:

Stress concentration occurs when there is an abrupt change in geometry, a mechanical part under loading. These changes in geometry can include holes, notches, or cracks within the component. The modifications create larger stress within the part. This maximum stress is difficult to determine, as it is directly at the point of the minimum area. Strain gauges have yet to be developed to analyze stresses at such minute areas. Therefore, a stress concentration factor must be utilized. The stress concentration factor is a dimensionless parameter calculated solely on the geometry of a part. The factor is multiplied by the nominal, or average, stress of the component, which can be found analytically or experimentally. Stress concentration graphs exist for common loading conditions and geometrical configurations to aid in the determination of the maximum stress a part can withstand. These graphs were developed from historical data yielded from experimentation. This project seeks to verify a stress concentration graph for combined loading conditions. The aforementioned graph was developed using CATIA Finite Element Analysis software. The results of this analysis will be validated through further testing. The 3D modeled parts will be subjected to further finite element analysis using Patran-Nastran software. The finite element models will then be verified by testing physical specimen using a tensile testing machine. Once the data is validated, the unique stress concentration graph will be submitted for publication so it can aid engineers in future projects.

Keywords: stress concentration, finite element analysis, finite element models, combined loading

Procedia PDF Downloads 387
8920 Farmers’ Awareness and Behavior of Chemical Pesticide Uses in Suan Luang Sub-District Municipality, Ampawa, Samut Songkram, Thailand

Authors: Paiboon Jeamponk, Tikamporn Thipsaeng

Abstract:

This paper is aimed to investigate farmers’ level of awareness and behavior of chemical pesticide uses, by using a case study of Suan Luang Sub- District Municipality, Ampawa, Samut Songkram Province. Questionnaire was employed in this study with the farmers from 46 households to explore their level of awareness in chemical pesticide uses, while interview and observation were adopted in exploring their behavior of chemical pesticide uses. The findings reflected the farmers’ high level of awareness in chemical pesticide uses in the hazardous effects of the chemical to human and environmental health, while their behavior of chemical pesticide uses explained their awareness paid to the right way of using pesticides, for instance reading the direction on the label, keeping children and animals away from the area of pesticide mixing, covering body with clothes and wearing hat and mask, no smoking, eating or drinking during pesticide spray or standing in windward direction.

Keywords: awareness, behavior, pesticide, farmers

Procedia PDF Downloads 389
8919 Human Posture Estimation Based on Multiple Viewpoints

Authors: Jiahe Liu, HongyangYu, Feng Qian, Miao Luo

Abstract:

This study aimed to address the problem of improving the confidence of key points by fusing multi-view information, thereby estimating human posture more accurately. We first obtained multi-view image information and then used the MvP algorithm to fuse this multi-view information together to obtain a set of high-confidence human key points. We used these as the input for the Spatio-Temporal Graph Convolution (ST-GCN). ST-GCN is a deep learning model used for processing spatio-temporal data, which can effectively capture spatio-temporal relationships in video sequences. By using the MvP algorithm to fuse multi-view information and inputting it into the spatio-temporal graph convolution model, this study provides an effective method to improve the accuracy of human posture estimation and provides strong support for further research and application in related fields.

Keywords: multi-view, pose estimation, ST-GCN, joint fusion

Procedia PDF Downloads 24
8918 The Problem of Now in Special Relativity Theory

Authors: Mogens Frank Mikkelsen

Abstract:

Special Relativity Theory (SRT) includes only one characteristic of light, the speed is equal to all observers, and by excluding other relevant characteristics of light, the common interpretation of SRT should be regarded as merely an approximative theory. By rethinking the iconic double light cones, a revised version of SRT can be developed. The revised concept of light cones acknowledges an asymmetry of past and future light cones and introduced a concept of the extended past to explain the predictions as something other than the future. Combining this with the concept of photon-paired events, led to the inference that Special Relativity theory can support the existence of Now.

Keywords: relativity, light cone, Minkowski, time

Procedia PDF Downloads 42
8917 A New Developed Formula to Determine the Shear Buckling Stress in Welded Aluminum Plate Girders

Authors: Badr Alsulami, Ahmed S. Elamary

Abstract:

This paper summarizes and presents main results of an in-depth numerical analysis dealing with the shear buckling resistance of aluminum plate girders. The studies conducted have permitted the development of a simple design expression to determine the critical shear buckling stress in aluminum web panels. This expression takes into account the effects of reduction of strength in aluminum alloys due to the welding process. Ultimate shear resistance (USR) of plate girders can be obtained theoretically using Cardiff theory or Hӧglund’s theory. USR of aluminum alloy plate girders predicted theoretically using BS8118 appear inconsistent when compared with test data. Theoretical predictions based on Hӧglund’s theory, are more realistic. Cardiff theory proposed to predict the USR of steel plate girders only. Welded aluminum alloy plate girders studied experimentally by others; the USR resulted from tests are reviewed. Comparison between the test results with the values obtained from Hӧglund’s theory, BS8118 design method, and Cardiff theory performed theoretically. Finally, a new equation based on Cardiff tension-field theory proposed to predict theoretically the USR of aluminum plate girders.

Keywords: shear resistance, aluminum, Cardiff theory, Hӧglund's theory, plate girder

Procedia PDF Downloads 373
8916 A Combinatorial Representation for the Invariant Measure of Diffusion Processes on Metric Graphs

Authors: Michele Aleandri, Matteo Colangeli, Davide Gabrielli

Abstract:

We study a generalization to a continuous setting of the classical Markov chain tree theorem. In particular, we consider an irreducible diffusion process on a metric graph. The unique invariant measure has an atomic component on the vertices and an absolutely continuous part on the edges. We show that the corresponding density at x can be represented by a normalized superposition of the weights associated to metric arborescences oriented toward the point x. A metric arborescence is a metric tree oriented towards its root. The weight of each oriented metric arborescence is obtained by the product of the exponential of integrals of the form ∫a/b², where b is the drift and σ² is the diffusion coefficient, along the oriented edges, for a weight for each node determined by the local orientation of the arborescence around the node and for the inverse of the diffusion coefficient at x. The metric arborescences are obtained by cutting the original metric graph along some edges.

Keywords: diffusion processes, metric graphs, invariant measure, reversibility

Procedia PDF Downloads 124
8915 Two-Level Graph Causality to Detect and Predict Random Cyber-Attacks

Authors: Van Trieu, Shouhuai Xu, Yusheng Feng

Abstract:

Tracking attack trajectories can be difficult, with limited information about the nature of the attack. Even more difficult as attack information is collected by Intrusion Detection Systems (IDSs) due to the current IDSs having some limitations in identifying malicious and anomalous traffic. Moreover, IDSs only point out the suspicious events but do not show how the events relate to each other or which event possibly cause the other event to happen. Because of this, it is important to investigate new methods capable of performing the tracking of attack trajectories task quickly with less attack information and dependency on IDSs, in order to prioritize actions during incident responses. This paper proposes a two-level graph causality framework for tracking attack trajectories in internet networks by leveraging observable malicious behaviors to detect what is the most probable attack events that can cause another event to occur in the system. Technically, given the time series of malicious events, the framework extracts events with useful features, such as attack time and port number, to apply to the conditional independent tests to detect the relationship between attack events. Using the academic datasets collected by IDSs, experimental results show that the framework can quickly detect the causal pairs that offer meaningful insights into the nature of the internet network, given only reasonable restrictions on network size and structure. Without the framework’s guidance, these insights would not be able to discover by the existing tools, such as IDSs. It would cost expert human analysts a significant time if possible. The computational results from the proposed two-level graph network model reveal the obvious pattern and trends. In fact, more than 85% of causal pairs have the average time difference between the causal and effect events in both computed and observed data within 5 minutes. This result can be used as a preventive measure against future attacks. Although the forecast may be short, from 0.24 seconds to 5 minutes, it is long enough to be used to design a prevention protocol to block those attacks.

Keywords: causality, multilevel graph, cyber-attacks, prediction

Procedia PDF Downloads 124
8914 Computation of Natural Logarithm Using Abstract Chemical Reaction Networks

Authors: Iuliia Zarubiieva, Joyun Tseng, Vishwesh Kulkarni

Abstract:

Recent researches has focused on nucleic acids as a substrate for designing biomolecular circuits for in situ monitoring and control. A common approach is to express them by a set of idealised abstract chemical reaction networks (ACRNs). Here, we present new results on how abstract chemical reactions, viz., catalysis, annihilation and degradation, can be used to implement circuit that accurately computes logarithm function using the method of Arithmetic-Geometric Mean (AGM), which has not been previously used in conjunction with ACRNs.

Keywords: chemical reaction networks, ratio computation, stability, robustness

Procedia PDF Downloads 125
8913 Efficient Heuristic Algorithm to Speed Up Graphcut in Gpu for Image Stitching

Authors: Tai Nguyen, Minh Bui, Huong Ninh, Tu Nguyen, Hai Tran

Abstract:

GraphCut algorithm has been widely utilized to solve various types of computer vision problems. Its expensive computational cost encouraged many researchers to improve the speed of the algorithm. Recent works proposed schemes that work on parallel computing platforms such as CUDA. However, the problem of low convergence speed prevents the usage of GraphCut for real time applications. In this paper, we propose global suppression heuristic to boost the conver-gence process of the algorithm. A parallel implementation of GraphCut algorithm on CUDA designed for the image stitching problem is introduced. Our method achieves up to 3× time boost on the graph of size 80 × 480 compared to the best sequential GraphCut algorithm while achieving satisfactory stitched images, suitable for panorama applications. Our source code will be soon available for further research.

Keywords: CUDA, graph cut, image stitching, texture synthesis, maxflow/mincut algorithm

Procedia PDF Downloads 81
8912 Beyond Geometry: The Importance of Surface Properties in Space Syntax Research

Authors: Christoph Opperer

Abstract:

Space syntax is a theory and method for analyzing the spatial layout of buildings and urban environments to understand how they can influence patterns of human movement, social interaction, and behavior. While direct visibility is a key factor in space syntax research, important visual information such as light, color, texture, etc., are typically not considered, even though psychological studies have shown a strong correlation to the human perceptual experience within physical space – with light and color, for example, playing a crucial role in shaping the perception of spaciousness. Furthermore, these surface properties are often the visual features that are most salient and responsible for drawing attention to certain elements within the environment. This paper explores the potential of integrating these factors into general space syntax methods and visibility-based analysis of space, particularly for architectural spatial layouts. To this end, we use a combination of geometric (isovist) and topological (visibility graph) approaches together with image-based methods, allowing a comprehensive exploration of the relationship between spatial geometry, visual aesthetics, and human experience. Custom-coded ray-tracing techniques are employed to generate spherical panorama images, encoding three-dimensional spatial data in the form of two-dimensional images. These images are then processed through computer vision algorithms to generate saliency-maps, which serve as a visual representation of areas most likely to attract human attention based on their visual properties. The maps are subsequently used to weight the vertices of isovists and the visibility graph, placing greater emphasis on areas with high saliency. Compared to traditional methods, our weighted visibility analysis introduces an additional layer of information density by assigning different weights or importance levels to various aspects within the field of view. This extends general space syntax measures to provide a more nuanced understanding of visibility patterns that better reflect the dynamics of human attention and perception. Furthermore, by drawing parallels to traditional isovist and VGA analysis, our weighted approach emphasizes a crucial distinction, which has been pointed out by Ervin and Steinitz: the difference between what is possible to see and what is likely to be seen. Therefore, this paper emphasizes the importance of including surface properties in visibility-based analysis to gain deeper insights into how people interact with their surroundings and to establish a stronger connection with human attention and perception.

Keywords: space syntax, visibility analysis, isovist, visibility graph, visual features, human perception, saliency detection, raytracing, spherical images

Procedia PDF Downloads 31
8911 Chemical Analysis of Available Portland Cement in Libyan Market Using X-Ray Fluorescence

Authors: M. A. Elbagermia, A. I. Alajtala, M. Alkerzab

Abstract:

This study compares the quality of different brands of Portland Cement (PC) available in Libyan market. The amounts of chemical constituents like SiO2, Al2O3, Fe2O3, CaO, MgO, SO3, and Lime Saturation Factor (LSF) were determined in accordance with Libyan (L.S.S) and Amrican (A.S.S) Standard Specifications. All the cement studies were found to be good for concrete work especially where no special property is required. The chemical and mineralogical analyses for studied clinker samples show that the dominant phases composition are C3S and C2S while the C3A and C4AF are less abundant.

Keywords: Portland cement, chemical composition, Libyan market, X-Ray fluorescence

Procedia PDF Downloads 311
8910 PhD Research Design and Descriptive Theory: Theoretical Framework for Development of Integrated Management System

Authors: Samuel Quashie

Abstract:

The importance of theory for PhD construction management research cannot be underestimated, as it requires a sound theoretical basis. Theory efficiency reduces errors in the research problem, solving it by building upon current theory. Provides a structure for examination, enables the efficient development of the construction management field and to it practical real world problems. The aim is to develop the theoretical framework for the application of descriptive theory within the PhD research design To apply the proposed theoretical framework using the case of the topic of ‘integrated management system,’ classifying the phenomena into categories, explore the association between the category–defining attributes and the outcome observed. Forming categorization based upon attributes of phenomena (framework and typologies), and statement of association (models). Predicting (deductive process) and confirming (inductive process). The descriptive theory is important and provides a structure for examination, enables the efficient development of construction management field and to it practical real world problems. In conclusion, the work done in management presents fertile ground for research and theory development.

Keywords: descriptive theory, PhD research design, theoretical framework, construction management

Procedia PDF Downloads 376
8909 Study on Robot Trajectory Planning by Robot End-Effector Using Dual Curvature Theory of the Ruled Surface

Authors: Y. S. Oh, P. Abhishesh, B. S. Ryuh

Abstract:

This paper presents the method of trajectory planning by the robot end-effector which accounts for more accurate and smooth differential geometry of the ruled surface generated by tool line fixed with end-effector based on the methods of curvature theory of ruled surface and the dual curvature theory, and focuses on the underlying relation to unite them for enhancing the efficiency for trajectory planning. Robot motion can be represented as motion properties of the ruled surface generated by trajectory of the Tool Center Point (TCP). The linear and angular properties of the six degree-of-freedom motion of end-effector are computed using the explicit formulas and functions from curvature theory and dual curvature theory. This paper explains the complete dualization of ruled surface and shows that the linear and angular motion applied using the method of dual curvature theory is more accurate and less complex.

Keywords: dual curvature theory, robot end effector, ruled surface, TCP (Tool Center Point)

Procedia PDF Downloads 327
8908 Codifying the Creative Self: Conflicts of Theory and Content in Creative Writing

Authors: Danielle L. Iamarino

Abstract:

This paper explores the embattled territory of academic creative writing—and most focally, the use of critical theory in the teaching and structuring of creative practice. It places creative writing in contemporary social, cultural, and otherwise anthropological contexts, and evaluates conventional creative writing pedagogies based on how well they serve the updated needs of increasingly diverse student congregations. With continued emphasis on student-centered learning, this paper compares theoretical to practical applications of discipline-specific knowledge, examining and critiquing theory in terms of its relevance, accessibility, and whether or not it is both actionable and beneficial in the creative writing classroom.

Keywords: creative writing, literary theory, content, pedagogy, workshop, teaching

Procedia PDF Downloads 295
8907 Research Analysis in Eclectic Theory (Kaboudan and Sfandiar)

Authors: Farideh Alizadeh, Mohd Nasir Hashi

Abstract:

Present research investigates eclecticism in Iranian theatre on the basis of eclectic theory. Eclectic theatre is a new theory in postmodernism. The theory appeared during 60th – 70th century in some theatres such as “Conference of the Birds”. Special theatrical forms have been developed in many geographical- cultural areas of the world and are indigenous to that area. These forms, as compared with original forms, are considered to be traditional while being comprehensive, the form is considered to be national. Kaboudan and Sfandiar theatre has been influenced by elements of traditional form of Iran.

Keywords: eclectic theatre, theatrical forms, tradition, play

Procedia PDF Downloads 354
8906 Detailed Quantum Circuit Design and Evaluation of Grover's Algorithm for the Bounded Degree Traveling Salesman Problem Using the Q# Language

Authors: Wenjun Hou, Marek Perkowski

Abstract:

The Traveling Salesman problem is famous in computing and graph theory. In short, it asks for the Hamiltonian cycle of the least total weight in a given graph with N nodes. All variations on this problem, such as those with K-bounded-degree nodes, are classified as NP-complete in classical computing. Although several papers propose theoretical high-level designs of quantum algorithms for the Traveling Salesman Problem, no quantum circuit implementation of these algorithms has been created up to our best knowledge. In contrast to previous papers, the goal of this paper is not to optimize some abstract complexity measures based on the number of oracle iterations, but to be able to evaluate the real circuit and time costs of the quantum computer. Using the emerging quantum programming language Q# developed by Microsoft, which runs quantum circuits in a quantum computer simulation, an implementation of the bounded-degree problem and its respective quantum circuit were created. To apply Grover’s algorithm to this problem, a quantum oracle was designed, evaluating the cost of a particular set of edges in the graph as well as its validity as a Hamiltonian cycle. Repeating the Grover algorithm with an oracle that finds successively lower cost each time allows to transform the decision problem to an optimization problem, finding the minimum cost of Hamiltonian cycles. N log₂ K qubits are put into an equiprobablistic superposition by applying the Hadamard gate on each qubit. Within these N log₂ K qubits, the method uses an encoding in which every node is mapped to a set of its encoded edges. The oracle consists of several blocks of circuits: a custom-written edge weight adder, node index calculator, uniqueness checker, and comparator, which were all created using only quantum Toffoli gates, including its special forms, which are Feynman and Pauli X. The oracle begins by using the edge encodings specified by the qubits to calculate each node that this path visits and adding up the edge weights along the way. Next, the oracle uses the calculated nodes from the previous step and check that all the nodes are unique. Finally, the oracle checks that the calculated cost is less than the previously-calculated cost. By performing the oracle an optimal number of times, a correct answer can be generated with very high probability. The oracle of the Grover Algorithm is modified using the recalculated minimum cost value, and this procedure is repeated until the cost cannot be further reduced. This algorithm and circuit design have been verified, using several datasets, to generate correct outputs.

Keywords: quantum computing, quantum circuit optimization, quantum algorithms, hybrid quantum algorithms, quantum programming, Grover’s algorithm, traveling salesman problem, bounded-degree TSP, minimal cost, Q# language

Procedia PDF Downloads 142
8905 Assessment of Potential Chemical Exposure to Betamethasone Valerate and Clobetasol Propionate in Pharmaceutical Manufacturing Laboratories

Authors: Nadeen Felemban, Hamsa Banjer, Rabaah Jaafari

Abstract:

One of the most common hazards in the pharmaceutical industry is the chemical hazard, which can cause harm or develop occupational health diseases/illnesses due to chronic exposures to hazardous substances. Therefore, a chemical agent management system is required, including hazard identification, risk assessment, controls for specific hazards and inspections, to keep your workplace healthy and safe. However, routine management monitoring is also required to verify the effectiveness of the control measures. Moreover, Betamethasone Valerate and Clobetasol Propionate are some of the APIs (Active Pharmaceutical Ingredients) with highly hazardous classification-Occupational Hazard Category (OHC 4), which requires a full containment (ECA-D) during handling to avoid chemical exposure. According to Safety Data Sheet, those chemicals are reproductive toxicants (reprotoxicant H360D), which may affect female workers’ health and cause fatal damage to an unborn child, or impair fertility. In this study, qualitative (chemical Risk assessment-qCRA) was conducted to assess the chemical exposure during handling of Betamethasone Valerate and Clobetasol Propionate in pharmaceutical laboratories. The outcomes of qCRA identified that there is a risk of potential chemical exposure (risk rating 8 Amber risk). Therefore, immediate actions were taken to ensure interim controls (according to the Hierarchy of controls) are in place and in use to minimize the risk of chemical exposure. No open handlings should be done out of the Steroid Glove Box Isolator (SGB) with the required Personal Protective Equipment (PPEs). The PPEs include coverall, nitrile hand gloves, safety shoes and powered air-purifying respirators (PAPR). Furthermore, a quantitative assessment (personal air sampling) was conducted to verify the effectiveness of the engineering controls (SGB Isolator) and to confirm if there is chemical exposure, as indicated earlier by qCRA. Three personal air samples were collected using an air sampling pump and filter (IOM2 filters, 25mm glass fiber media). The collected samples were analyzed by HPLC in the BV lab, and the measured concentrations were reported in (ug/m3) with reference to Occupation Exposure Limits, 8hr OELs (8hr TWA) for each analytic. The analytical results are needed in 8hr TWA (8hr Time-weighted Average) to be analyzed using Bayesian statistics (IHDataAnalyst). The results of the Bayesian Likelihood Graph indicate (category 0), which means Exposures are de "minimus," trivial, or non-existent Employees have little to no exposure. Also, these results indicate that the 3 samplings are representative samplings with very low variations (SD=0.0014). In conclusion, the engineering controls were effective in protecting the operators from such exposure. However, routine chemical monitoring is required every 3 years unless there is a change in the processor type of chemicals. Also, frequent management monitoring (daily, weekly, and monthly) is required to ensure the control measures are in place and in use. Furthermore, a Similar Exposure Group (SEG) was identified in this activity and included in the annual health surveillance for health monitoring.

Keywords: occupational health and safety, risk assessment, chemical exposure, hierarchy of control, reproductive

Procedia PDF Downloads 142