Search results for: spectral domain
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2404

Search results for: spectral domain

574 Optimal Parameters of Two-Color Ionizing Laser Pulses for Terahertz Generation

Authors: I. D. Laryushin, V. A. Kostin, A. A. Silaev, N. V. Vvedenskii

Abstract:

Generation of broadband intense terahertz (THz) radiation attracts reasonable interest due to various applications, such as the THz time-domain spectroscopy, the probing and control of various ultrafast processes, the THz imaging with subwavelength resolution, and many others. One of the most promising methods for generating powerful and broadband terahertz pulses is based on focusing two-color femtosecond ionizing laser pulses in gases, including ambient air. For this method, the amplitudes of terahertz pulses are determined by the free-electron current density remaining in a formed plasma after the passage of the laser pulse. The excitation of this residual current density can be treated as multi-wave mixing: Аn effective generation of terahertz radiation is possible only when the frequency ratio of one-color components in the two-color pulse is close to irreducible rational fraction a/b with small odd sum a + b. This work focuses on the optimal parameters (polarizations and intensities) of laser components for the strongest THz generation. The optimal values of parameters are found numerically and analytically with the use of semiclassical approach for calculating the residual current density. For frequency ratios close to a/(a ± 1) with natural a, the strongest THz generation is shown to take place when the both laser components have circular polarizations and equal intensities. For this optimal case, an analytical formula for the residual current density was derived. For the frequency ratios such as 2/5, the two-color ionizing pulses with circularly polarized components practically do not excite the residual current density. However, the optimal parameters correspond generally to specific elliptical (not linear) polarizations of the components and intensity ratios close to unity.

Keywords: broadband terahertz radiation, ionization, laser plasma, ultrashort two-color pulses

Procedia PDF Downloads 191
573 Cartographic Depiction and Visualization of Wetlands Changes in the North-Western States of India

Authors: Bansal Ashwani

Abstract:

Cartographic depiction and visualization of wetland changes is an important tool to map spatial-temporal information about the wetland dynamics effectively and to comprehend the response of these water bodies in maintaining the groundwater and surrounding ecosystem. This is true for the states of North Western India, i.e., J&K, Himachal, Punjab, and Haryana that are bestowed upon with several natural wetlands in the flood plains or on the courses of its rivers. Thus, the present study documents, analyses and reconstructs the lost wetlands, which existed in the flood plains of the major river basins of these states, i.e., Chenab, Jhelum, Satluj, Beas, Ravi, and Ghagar, in the beginning of the 20th century. To achieve the objective, the study has used multi-temporal datasets since the 1960s using high to medium resolution satellite datasets, e.g., Corona (1960s/70s), Landsat (1990s-2017) and Sentinel (2017). The Sentinel (2017) satellite image has been used for making the wetland inventory owing to its comparatively higher spatial resolution with multi-spectral bands. In addition, historical records, repeated photographs, historical maps, field observations including geomorphological evidence were also used. The water index techniques, i.e., band rationing, normalized difference water index (NDWI), modified NDWI (MNDWI) have been compared and used to map the wetlands. The wetland types found in the north-western states have been categorized under 19 classes suggested by Space Application Centre, India. These enable the researcher to provide with the wetlands inventory and a series of cartographic representation that includes overlaying multiple temporal wetlands extent vectors. A preliminary result shows the general state of wetland shrinkage since the 1960s with varying area shrinkage rate from one wetland to another. In addition, it is observed that majority of wetlands have not been documented so far and even do not have names. Moreover, the purpose is to emphasize their elimination in addition to establishing a baseline dataset that can be a tool for wetland planning and management. Finally, the applicability of cartographic depiction and visualization, historical map sources, repeated photographs and remote sensing data for reconstruction of long term wetlands fluctuations, especially in the northern part of India, will be addressed.

Keywords: cartographic depiction and visualization, wetland changes, NDWI/MDWI, geomorphological evidence and remote sensing

Procedia PDF Downloads 245
572 The Development of Space-Time and Space-Number Associations: The Role of Non-Symbolic vs. Symbolic Representations

Authors: Letizia Maria Drammis, Maria Antonella Brandimonte

Abstract:

The idea that people use space representations to think about time and number received support from several lines of research. However, how these representations develop in children and then shape space-time and space-number mappings is still a debated issue. In the present study, 40 children (20 pre-schoolers and 20 elementary-school children) performed 4 main tasks, which required the use of more concrete (non-symbolic) or more abstract (symbolic) space-time and space-number associations. In the non-symbolic conditions, children were required to order pictures of everyday-life events occurring in a specific temporal order (Temporal sequences) and of quantities varying in numerosity (Numerical sequences). In the symbolic conditions, they were asked to perform the typical time-to-position and number-to-position tasks by mapping time-related words and numbers onto lines. Results showed that children performed reliably better in the non-symbolic Time conditions than the symbolic Time conditions, independently of age, whereas only pre-schoolers performed worse in the Number-to-position task (symbolic) as compared to the Numerical sequence (non-symbolic) task. In addition, only older children mapped time-related words onto space following the typical left-right orientation, pre-schoolers’ performance being somewhat mixed. In contrast, mapping numbers onto space showed a clear left-right orientation, independently of age. Overall, these results indicate a cross-domain difference in the way younger and older children process time and number, with time-related tasks being more difficult than number-related tasks only when space-time tasks require symbolic representations.

Keywords: space-time associations, space-number associations, orientation, children

Procedia PDF Downloads 320
571 A Structured Mechanism for Identifying Political Influencers on Social Media Platforms: Top 10 Saudi Political Twitter Users

Authors: Ahmad Alsolami, Darren Mundy, Manuel Hernandez-Perez

Abstract:

Social media networks, such as Twitter, offer the perfect opportunity to either positively or negatively affect political attitudes on large audiences. The existence of influential users who have developed a reputation for their knowledge and experience of specific topics is a major factor contributing to this impact. Therefore, knowledge of the mechanisms to identify influential users on social media is vital for understanding their effect on their audience. The concept of the influential user is related to the concept of opinion leaders' to indicate that ideas first flow from mass media to opinion leaders and then to the rest of the population. Hence, the objective of this research was to provide reliable and accurate structural mechanisms to identify influential users, which could be applied to different platforms, places, and subjects. Twitter was selected as the platform of interest, and Saudi Arabia as the context for the investigation. These were selected because Saudi Arabia has a large number of Twitter users, some of whom are considerably active in setting agendas and disseminating ideas. The study considered the scientific methods that have been used to identify public opinion leaders before, utilizing metrics software on Twitter. The key findings propose multiple novel metrics to compare Twitter influencers, including the number of followers, social authority and the use of political hashtags, and four secondary filtering measures. Thus, using ratio and percentage calculations to classify the most influential users, Twitter accounts were filtered, analyzed and included. The structured approach is used as a mechanism to explore the top ten influencers on Twitter from the political domain in Saudi Arabia.

Keywords: Twitter, influencers, structured mechanism, Saudi Arabia

Procedia PDF Downloads 100
570 A Phenomenological Approach to Computational Modeling of Analogy

Authors: José Eduardo García-Mendiola

Abstract:

In this work, a phenomenological approach to computational modeling of analogy processing is carried out. The paper goes through the consideration of the structure of the analogy, based on the possibility of sustaining the genesis of its elements regarding Husserl's genetic theory of association. Among particular processes which take place in order to get analogical inferences, there is one which arises crucial for enabling efficient base cases retrieval through long-term memory, namely analogical transference grounded on familiarity. In general, it has been argued that analogical reasoning is a way by which a conscious agent tries to determine or define a certain scope of objects and relationships between them using previous knowledge of other familiar domain of objects and relations. However, looking for a complete description of analogy process, a deeper consideration of phenomenological nature is required in so far, its simulation by computational programs is aimed. Also, one would get an idea of how complex it would be to have a fully computational account of the analogy elements. In fact, familiarity is not a result of a mere chain of repetitions of objects or events but generated insofar as the object/attribute or event in question is integrable inside a certain context that is taking shape as functionalities and functional approaches or perspectives of the object are being defined. Its familiarity is generated not by the identification of its parts or objective determinations as if they were isolated from those functionalities and approaches. Rather, at the core of such a familiarity between entities of different kinds lays the way they are functionally encoded. So, and hoping to make deeper inroads towards these topics, this essay allows us to consider that cognitive-computational perspectives can visualize, from the phenomenological projection of the analogy process reviewing achievements already obtained as well as exploration of new theoretical-experimental configurations towards implementation of analogy models in specific as well as in general purpose machines.

Keywords: analogy, association, encoding, retrieval

Procedia PDF Downloads 97
569 Text Analysis to Support Structuring and Modelling a Public Policy Problem-Outline of an Algorithm to Extract Inferences from Textual Data

Authors: Claudia Ehrentraut, Osama Ibrahim, Hercules Dalianis

Abstract:

Policy making situations are real-world problems that exhibit complexity in that they are composed of many interrelated problems and issues. To be effective, policies must holistically address the complexity of the situation rather than propose solutions to single problems. Formulating and understanding the situation and its complex dynamics, therefore, is a key to finding holistic solutions. Analysis of text based information on the policy problem, using Natural Language Processing (NLP) and Text analysis techniques, can support modelling of public policy problem situations in a more objective way based on domain experts knowledge and scientific evidence. The objective behind this study is to support modelling of public policy problem situations, using text analysis of verbal descriptions of the problem. We propose a formal methodology for analysis of qualitative data from multiple information sources on a policy problem to construct a causal diagram of the problem. The analysis process aims at identifying key variables, linking them by cause-effect relationships and mapping that structure into a graphical representation that is adequate for designing action alternatives, i.e., policy options. This study describes the outline of an algorithm used to automate the initial step of a larger methodological approach, which is so far done manually. In this initial step, inferences about key variables and their interrelationships are extracted from textual data to support a better problem structuring. A small prototype for this step is also presented.

Keywords: public policy, problem structuring, qualitative analysis, natural language processing, algorithm, inference extraction

Procedia PDF Downloads 571
568 Body Fluids Identification by Raman Spectroscopy and Matrix-Assisted Laser Desorption/Ionization Time-of-Flight Mass Spectrometry

Authors: Huixia Shi, Can Hu, Jun Zhu, Hongling Guo, Haiyan Li, Hongyan Du

Abstract:

The identification of human body fluids during forensic investigations is a critical step to determine key details, and present strong evidence to testify criminal in a case. With the popularity of DNA and improved detection technology, the potential question must be revolved that whether the suspect’s DNA derived from saliva or semen, menstrual or peripheral blood, how to identify the red substance or aged blood traces on the spot is blood; How to determine who contribute the right one in mixed stains. In recent years, molecular approaches have been developing increasingly on mRNA, miRNA, DNA methylation and microbial markers, but appear expensive, time-consuming, and destructive disadvantages. Physicochemical methods are utilized frequently such us scanning electron microscopy/energy spectroscopy and X-ray fluorescence and so on, but results only showing one or two characteristics of body fluid itself and that out of working in unknown or mixed body fluid stains. This paper focuses on using chemistry methods Raman spectroscopy and matrix-assisted laser desorption/ionization time-of-flight mass spectrometry to discriminate species of peripheral blood, menstrual blood, semen, saliva, vaginal secretions, urine or sweat. Firstly, non-destructive, confirmatory, convenient and fast Raman spectroscopy method combined with more accurate matrix-assisted laser desorption/ionization time-of-flight mass spectrometry method can totally distinguish one from other body fluids. Secondly, 11 spectral signatures and specific metabolic molecules have been obtained by analysis results after 70 samples detected. Thirdly, Raman results showed peripheral and menstrual blood, saliva and vaginal have highly similar spectroscopic features. Advanced statistical analysis of the multiple Raman spectra must be requested to classify one to another. On the other hand, it seems that the lactic acid can differentiate peripheral and menstrual blood detected by matrix-assisted laser desorption/ionization time-of-flight mass spectrometry, but that is not a specific metabolic molecule, more sensitivity ones will be analyzed in a forward study. These results demonstrate the great potential of the developed chemistry methods for forensic applications, although more work is needed for method validation.

Keywords: body fluids, identification, Raman spectroscopy, matrix-assisted laser desorption/ionization time-of-flight mass spectrometry

Procedia PDF Downloads 117
567 Time-dependent Association between Recreational Cannabinoid Use and Memory Performance in Healthy Adults: A Neuroimaging Study of Human Connectome Project

Authors: Kamyar Moradi

Abstract:

Background: There is mixed evidence regarding the association between recreational cannabinoid use and memory performance. One of the major reasons for the present controversy is different cannabinoid use-related covariates that influence the cognitive status of an individual. Adjustment of these confounding variables provides accurate insight into the real effects of cannabinoid use on memory status. In this study, we sought to investigate the association between recent recreational cannabinoid use and memory performance while correcting the model for other possible covariates such as demographic characteristics and duration, and amount of cannabinoid use. Methods: Cannabinoid users were assigned to two groups based on the results of THC urine drug screen test (THC+ group: n = 110, THC- group: n = 410). THC urine drug screen test has a high sensitivity and specificity in detecting cannabinoid use in the last 3-4 weeks. The memory domain of NIH Toolbox battery and brain MRI volumetric measures were compared between the groups while adjusting for confounding variables. Results: After Benjamini-Hochberg p-value correction, the performance in all of the measured memory outcomes, including vocabulary comprehension, episodic memory, executive function/cognitive flexibility, processing speed, reading skill, working memory, and fluid cognition, were significantly weaker in THC+ group (p values less than 0.05). Also, volume of gray matter, left supramarginal, right precuneus, right inferior/middle temporal, right hippocampus, left entorhinal, and right pars orbitalis regions were significantly smaller in THC+ group. Conclusions: this study provides evidence regarding the acute effect of recreational cannabis use on memory performance. Further studies are warranted to confirm the results.

Keywords: brain MRI, cannabis, memory, recreational use, THC urine test

Procedia PDF Downloads 173
566 Formal Ontology of Quality Space. Location, Subordination and Determination

Authors: Claudio Calosi, Damiano Costa, Paolo Natali

Abstract:

Determination is the relation that holds between certain kinds of properties, determinables – such as “being colored”, and others, determinates – such as “being red”. Subordination is the relation that holds between genus properties – such as “being an animal”, and others, species properties – such as “being human”'. It is widely held that Determination and Subordination share important similarities, yet also crucial differences. But what grounds such similarities and differences? This question is hardly ever addressed. The present paper provides the first step towards filling this gap in the literature. It argues that a locational theory of instantiation, roughly the view that to have a property is to occupy a location in quality space, holds the key for such an answer. More precisely, it argues that both principles of Determination and Subordination are just examples of more general principles of location. Consider Determination. The principle that everything that has a determinate has a determinable boils down to the claim that everything that has a precise location in quality space is in quality space – an eminently reasonable principle. The principle that nothing can have two determinates (at the same level of determination) boils down to the principle that nothing can be “multilocated” in quality space. In effect, the following provides a “translation table” between principles of location and determination: LOCATION DETERMINATION Functionality At Most One Determination Focus At Most One Determination & Requisite Determination* Exactness Requisite Determination* Super-Exactness Requisite Determination Exactitude Requisite Determination Converse-Exactness Determinable Inehritance This grounds the similarity between Determination and Subordination. What about the differences? The paper argues that the differences boil down to the mereological structure of the regions that are occupied in quality space, in particular whether they are simple or complex. The key technical detail is that Determination and Subordination induce a “set-theoretic rooted tree” structure over the domain of properties. Interestingly, the analysis also provides a possible justification for the Aristotelian claim that being is not a genus property – an argument that the paper develops in some detail.

Keywords: determinables/determinates, genus/species, location, Aristotle on being is not a genus

Procedia PDF Downloads 58
565 Stress-Strain Relation for Human Trabecular Bone Based on Nanoindentation Measurements

Authors: Marek Pawlikowski, Krzysztof Jankowski, Konstanty Skalski, Anna Makuch

Abstract:

Nanoindentation or depth-sensing indentation (DSI) technique has proven to be very useful to measure mechanical properties of various tissues at a micro-scale. Bone tissue, both trabecular and cortical one, is one of the most commonly tested tissues by means of DSI. Most often such tests on bone samples are carried out to compare the mechanical properties of lamellar and interlamellar bone, osteonal bone as well as compact and cancellous bone. In the paper, a relation between stress and strain for human trabecular bone is presented. The relation is based on the results of nanoindentation tests. The formulation of a constitutive model for human trabecular bone is based on nanoindentation tests. In the study, the approach proposed by Olivier-Pharr is adapted. The tests were carried out on samples of trabecular tissue extracted from human femoral heads. The heads were harvested during surgeries of artificial hip joint implantation. Before samples preparation, the heads were kept in 95% alcohol in temperature 4 Celsius degrees. The cubic samples cut out of the heads were stored in the same conditions. The dimensions of the specimens were 25 mm x 25 mm x 20 mm. The number of 20 samples have been tested. The age range of donors was between 56 and 83 years old. The tests were conducted with the indenter spherical tip of the diameter 0.200 mm. The maximum load was P = 500 mN and the loading rate 500 mN/min. The data obtained from the DSI tests allows one only to determine bone behoviour in terms of nanoindentation force vs. nanoindentation depth. However, it is more interesting and useful to know the characteristics of trabecular bone in the stress-strain domain. This allows one to simulate trabecular bone behaviour in a more realistic way. The stress-strain curves obtained in the study show relation between the age and the mechanical behaviour of trabecular bone. It was also observed that the bone matrix of trabecular tissue indicates an ability of energy absorption.

Keywords: constitutive model, mechanical behaviour, nanoindentation, trabecular bone

Procedia PDF Downloads 199
564 Inter-Annual Variations of Sea Surface Temperature in the Arabian Sea

Authors: K. S. Sreejith, C. Shaji

Abstract:

Though both Arabian Sea and its counterpart Bay of Bengal is forced primarily by the semi-annually reversing monsoons, the spatio-temporal variations of surface waters is very strong in the Arabian Sea as compared to the Bay of Bengal. This study focuses on the inter-annual variability of Sea Surface Temperature (SST) in the Arabian Sea by analysing ERSST dataset which covers 152 years of SST (January 1854 to December 2002) based on the ICOADS in situ observations. To capture the dominant SST oscillations and to understand the inter-annual SST variations at various local regions of the Arabian Sea, wavelet analysis was performed on this long time-series SST dataset. This tool is advantageous over other signal analysing tools like Fourier analysis, based on the fact that it unfolds a time-series data (signal) both in frequency and time domain. This technique makes it easier to determine dominant modes of variability and explain how those modes vary in time. The analysis revealed that pentadal SST oscillations predominate at most of the analysed local regions in the Arabian Sea. From the time information of wavelet analysis, it was interpreted that these cold and warm events of large amplitude occurred during the periods 1870-1890, 1890-1910, 1930-1950, 1980-1990 and 1990-2005. SST oscillations with peaks having period of ~ 2-4 years was found to be significant in the central and eastern regions of Arabian Sea. This indicates that the inter-annual SST variation in the Indian Ocean is affected by the El Niño-Southern Oscillation (ENSO) and Indian Ocean Dipole (IOD) events.

Keywords: Arabian Sea, ICOADS, inter-annual variation, pentadal oscillation, SST, wavelet analysis

Procedia PDF Downloads 262
563 Genome-Wide Identification and Characterization of MLO Family Genes in Pumpkin (Cucurbita maxima Duch.)

Authors: Khin Thanda Win, Chunying Zhang, Sanghyeob Lee

Abstract:

Mildew resistance locus o (Mlo), a plant-specific gene family with seven-transmembrane (TM), plays an important role in plant resistance to powdery mildew (PM). PM caused by Podosphaera xanthii is a widespread plant disease and probably represents the major fungal threat for many Cucurbits. The recent Cucurbita maxima genome sequence data provides an opportunity to identify and characterize the MLO gene family in this species. Total twenty genes (designated CmaMLO1 through CmaMLO20) have been identified by using an in silico cloning method with the MLO gene sequences of Cucumis sativus, Cucumis melo, Citrullus lanatus and Cucurbita pepo as probes. These CmaMLOs were evenly distributed on 15 chromosomes of 20 C. maxima chromosomes without any obvious clustering. Multiple sequence alignment showed that the common structural features of MLO gene family, such as TM domains, a calmodulin-binding domain and 30 important amino acid residues for MLO function, were well conserved. Phylogenetic analysis of the CmaMLO genes and other plant species reveals seven different clades (I through VII) and only clade IV is specific to monocots (rice, barley, and wheat). Phylogenetic and structural analyses provided preliminary evidence that five genes belonged to clade V could be the susceptibility genes which may play the importance role in PM resistance. This study is the first comprehensive report on MLO genes in C. maxima to our knowledge. These findings will facilitate the functional analysis of the MLOs related to PM susceptibility and are valuable resources for the development of disease resistance in pumpkin.

Keywords: Mildew resistance locus o (Mlo), powdery mildew, phylogenetic relationship, susceptibility genes

Procedia PDF Downloads 168
562 Multiscale Hub: An Open-Source Framework for Practical Atomistic-To-Continuum Coupling

Authors: Masoud Safdari, Jacob Fish

Abstract:

Despite vast amount of existing theoretical knowledge, the implementation of a universal multiscale modeling, analysis, and simulation software framework remains challenging. Existing multiscale software and solutions are often domain-specific, closed-source and mandate a high-level of experience and skills in both multiscale analysis and programming. Furthermore, tools currently existing for Atomistic-to-Continuum (AtC) multiscaling are developed with the assumptions such as accessibility of high-performance computing facilities to the users. These issues mentioned plus many other challenges have reduced the adoption of multiscale in academia and especially industry. In the current work, we introduce Multiscale Hub (MsHub), an effort towards making AtC more accessible through cloud services. As a joint effort between academia and industry, MsHub provides a universal web-enabled framework for practical multiscaling. Developed on top of universally acclaimed scientific programming language Python, the package currently provides an open-source, comprehensive, easy-to-use framework for AtC coupling. MsHub offers an easy to use interface to prominent molecular dynamics and multiphysics continuum mechanics packages such as LAMMPS and MFEM (a free, lightweight, scalable C++ library for finite element methods). In this work, we first report on the design philosophy of MsHub, challenges identified and issues faced regarding its implementation. MsHub takes the advantage of a comprehensive set of tools and algorithms developed for AtC that can be used for a variety of governing physics. We then briefly report key AtC algorithms implemented in MsHub. Finally, we conclude with a few examples illustrating the capabilities of the package and its future directions.

Keywords: atomistic, continuum, coupling, multiscale

Procedia PDF Downloads 162
561 Implementation of Free-Field Boundary Condition for 2D Site Response Analysis in OpenSees

Authors: M. Eskandarighadi, C. R. McGann

Abstract:

It is observed from past experiences of earthquakes that local site conditions can significantly affect the strong ground motion characteristics experience at the site. One-dimensional seismic site response analysis is the most common approach for investigating site response. This approach assumes that soil is homogeneous and infinitely extended in the horizontal direction. Therefore, tying side boundaries together is one way to model this behavior, as the wave passage is assumed to be only vertical. However, 1D analysis cannot capture the 2D nature of wave propagation, soil heterogeneity, and 2D soil profile with features such as inclined layer boundaries. In contrast, 2D seismic site response modeling can consider all of the mentioned factors to better understand local site effects on strong ground motions. 2D wave propagation and considering that the soil profile on the two sides of the model may not be identical clarifies the importance of a boundary condition on each side that can minimize the unwanted reflections from the edges of the model and input appropriate loading conditions. Ideally, the model size should be sufficiently large to minimize the wave reflection, however, due to computational limitations, increasing the model size is impractical in some cases. Another approach is to employ free-field boundary conditions that take into account the free-field motion that would exist far from the model domain and apply this to the sides of the model. This research focuses on implementing free-field boundary conditions in OpenSees for 2D site response analysisComparisons are made between 1D models and 2D models with various boundary conditions, and details and limitations of the developed free-field boundary modeling approach are discussed.

Keywords: boundary condition, free-field, opensees, site response analysis, wave propagation

Procedia PDF Downloads 130
560 Development of Medical Intelligent Process Model Using Ontology Based Technique

Authors: Emmanuel Chibuogu Asogwa, Tochukwu Sunday Belonwu

Abstract:

An urgent demand for creative solutions has been created by the rapid expansion of medical knowledge, the complexity of patient care, and the requirement for more precise decision-making. As a solution to this problem, the creation of a Medical Intelligent Process Model (MIPM) utilizing ontology-based appears as a promising way to overcome this obstacle and unleash the full potential of healthcare systems. The development of a Medical Intelligent Process Model (MIPM) using ontology-based techniques is motivated by a lack of quick access to relevant medical information and advanced tools for treatment planning and clinical decision-making, which ontology-based techniques can provide. The aim of this work is to develop a structured and knowledge-driven framework that leverages ontology, a formal representation of domain knowledge, to enhance various aspects of healthcare. Object-Oriented Analysis and Design Methodology (OOADM) were adopted in the design of the system as we desired to build a usable and evolvable application. For effective implementation of this work, we used the following materials/methods/tools: the medical dataset for the test of our model in this work was obtained from Kaggle. The ontology-based technique was used with Confusion Matrix, MySQL, Python, Hypertext Markup Language (HTML), Hypertext Preprocessor (PHP), Cascaded Style Sheet (CSS), JavaScript, Dreamweaver, and Fireworks. According to test results on the new system using Confusion Matrix, both the accuracy and overall effectiveness of the medical intelligent process significantly improved by 20% compared to the previous system. Therefore, using the model is recommended for healthcare professionals.

Keywords: ontology-based, model, database, OOADM, healthcare

Procedia PDF Downloads 60
559 Facile Fabrication of TiO₂NT/Fe₂O₃@Ag₂CO₃ Nanocomposite and Its Highly Efficient Visible Light Photocatalytic and Antibacterial Activity

Authors: Amal A. Al-Kahlawy, Heba H. El-Maghrabi

Abstract:

Due to the increasing need to environment protection in real time need to energize new materials are under extensive investigations. Between others, TiO2 nanotubes (TNTs) nanocomposite with iron oxide and silver carbonate, are promising alternatives as high-efficiency visible light photocatalyst due to their unique properties and their superior charge transport properties. Our efforts in this domain aim the construction of novel nanocomposite of TiO2NT/Fe2O3@Ag2CO3. The structure, surface morphology, chemical composition and optical properties were characterized by X-ray diffraction (XRD), Raman, Fourier-transform infrared spectroscopy (FTIR), scanning electron microscopy (SEM), energy dispersive X-ray spectrometer (EDS), transmission electron microscopy (TEM), selected area electron diffraction (SAED) and UV–vis diffuse reflectance spectroscopy (DRS). XRD results confirm the interaction of TiO2-NT with iron oxide. This novel nanocomposite shows remarkably enhanced performance for phenol compounds photodegradation. The experimental data shows a promising photocatalytic activity. In particular, a maximum value of 450 mg/g was removed within 60 min at solar light irradiation with degradation efficiency of 99.5%. The high photocatalytic activity of the nanocomposite is found to be related to the increased adsorption toward chemical species, enhanced light absorption and efficient charge separation and transfer. Finally, the designed TiO2NT/Fe2O3@Ag2CO3 nanocomposite has a great degree of sustainability and could has a potential application for the industrial treatment of wastewater containing toxic organic materials.

Keywords: nanocomposite, photocatalyst, solar energy, titanium dioxide nanotubes

Procedia PDF Downloads 224
558 Awareness among Medical Students and Faculty about Integration of Artifical Intelligence Literacy in Medical Curriculum

Authors: Fatima Faraz

Abstract:

BACKGROUND: While Artificial intelligence (AI) provides new opportunities across a wide variety of industries, healthcare is no exception. AI can lead to advancements in how the healthcare system functions and improves the quality of patient care. Developing countries like Pakistan are lagging in the implementation of AI-based solutions in healthcare. This demands increased knowledge and AI literacy among health care professionals. OBJECTIVES: To assess the level of awareness among medical students and faculty about AI in preparation for teaching AI basics and data science applications in clinical practice in an integrated medical curriculum. METHODS: An online 15-question semi-structured questionnaire, previously tested and validated, was delivered among participants through convenience sampling. The questionnaire composed of 3 parts: participant’s background knowledge, AI awareness, and attitudes toward AI applications in medicine. RESULTS: A total of 182 students and 39 faculty members from Rawalpindi Medical University, Pakistan, participated in the study. Only 26% of students and 46.2% of faculty members responded that they were aware of AI topics in clinical medicine. The major source of AI knowledge was social media (35.7%) for students and professional talks and colleagues (43.6%) for faculty members. 23.5% of participants answered that they personally had a basic understanding of AI. Students and faculty (60.1%) were interested in AI in patient care and teaching domain. These findings parallel similar published AI survey results. CONCLUSION: This survey concludes interest among students and faculty in AI developments and technology applications in healthcare. Further studies are required in order to correctly fit AI in the integrated modular curriculum of medical education.

Keywords: medical education, data science, artificial intelligence, curriculum

Procedia PDF Downloads 85
557 Exploring the Role of Building Information Modeling for Delivering Successful Construction Projects

Authors: Muhammad Abu Bakar Tariq

Abstract:

Construction industry plays a crucial role in the progress of societies and economies. Furthermore, construction projects have social as well as economic implications, thus, their success/failure have wider impacts. However, the industry is lagging behind in terms of efficiency and productivity. Building Information Modeling (BIM) is recognized as a revolutionary development in Architecture, Engineering and Construction (AEC) industry. There are numerous interest groups around the world providing definitions of BIM, proponents describing its advantages and opponents identifying challenges/barriers regarding adoption of BIM. This research is aimed at to determine what actually BIM is, along with its potential role in delivering successful construction projects. The methodology is critical analysis of secondary data sources i.e. information present in public domain, which include peer reviewed journal articles, industry and government reports, conference papers, books, case studies etc. It is discovered that clash detection and visualization are two major advantages of BIM. Clash detection option identifies clashes among structural, architectural and MEP designs before construction actually commences, which subsequently saves time as well as cost and ensures quality during execution phase of a project. Visualization is a powerful tool that facilitates in rapid decision-making in addition to communication and coordination among stakeholders throughout project’s life cycle. By eliminating inconsistencies that consume time besides cost during actual construction, improving collaboration among stakeholders throughout project’s life cycle, BIM can play a positive role to achieve efficiency and productivity that consequently deliver successful construction projects.

Keywords: building information modeling, clash detection, construction project success, visualization

Procedia PDF Downloads 239
556 Effects of Stokes Shift and Purcell Enhancement in Fluorescence Assisted Radiative Cooling

Authors: Xue Ma, Yang Fu, Dangyuan Lei

Abstract:

Passive daytime radiative cooling is an emerging technology which has attracted worldwide attention in recent years due to its huge potential in cooling buildings without the use of electricity. Various coating materials with different optical properties have been developed to improve the daytime radiative cooling performance. However, commercial cooling coatings comprising functional fillers with optical bandgaps within the solar spectral range suffers from severe intrinsic absorption, limiting their cooling performance. Fortunately, it has recently been demonstrated that introducing fluorescent materials into polymeric coatings can covert the absorbed sunlight to fluorescent emissions and hence increase the effective solar reflectance and cooling performance. In this paper, we experimentally investigate the key factors for fluorescence-assisted radiative cooling with TiO2-based white coatings. The surrounding TiO2 nanoparticles, which enable spatial and temporal light confinement through multiple Mie scattering, lead to Purcell enhancement of phosphors in the coating. Photoluminescence lifetimes of two phosphors (BaMgAl10O17:Eu2+ and (Sr, Ba)SiO4:Eu2+) exhibit significant reduction of ~61% and ~23%, indicating Purcell factors of 2.6 and 1.3, respectively. Moreover, smaller Stokes shifts of the phosphors are preferred to further diminish solar absorption. Field test of fluorescent cooling coatings demonstrate an improvement of ~4% solar reflectance for the BaMgAl10O17:Eu2+-based fluorescent cooling coating. However, to maximize solar reflectance, a white appearance is introduced based on multiple Mie scattering by the broad size distribution of fillers, which is visually pressurized and aesthetically bored. Besides, most colored pigments absorb visible light significantly and convert it to non-radiative thermal energy, offsetting the cooling effect. Therefore, current colored cooling coatings are facing the compromise between color saturation and cooling effect. To solve this problem, we introduced colored fluorescent materials into white coating based on SiO2 microspheres as a top layer, covering a white cooling coating based on TiO2. Compared with the colored pigments, fluorescent materials could re-emit the absorbed light, reducing the solar absorption introduced by coloration. Our work investigated the scattering properties of SiO2 dielectric spheres with different diameters and detailly discussed their impact on the PL properties of phosphors, paving the way for colored fluorescent-assisted cooling coting to application and industrialization.

Keywords: solar reflection, infrared emissivity, mie scattering, photoluminescent emission, radiative cooling

Procedia PDF Downloads 71
555 Nondestructive Inspection of Reagents under High Attenuated Cardboard Box Using Injection-Seeded THz-Wave Parametric Generator

Authors: Shin Yoneda, Mikiya Kato, Kosuke Murate, Kodo Kawase

Abstract:

In recent years, there have been numerous attempts to smuggle narcotic drugs and chemicals by concealing them in international mail. Combatting this requires a non-destructive technique that can identify such illicit substances in mail. Terahertz (THz) waves can pass through a wide variety of materials, and many chemicals show specific frequency-dependent absorption, known as a spectral fingerprint, in the THz range. Therefore, it is reasonable to investigate non-destructive mail inspection techniques that use THz waves. For this reason, in this work, we tried to identify reagents under high attenuation shielding materials using injection-seeded THz-wave parametric generator (is-TPG). Our THz spectroscopic imaging system using is-TPG consisted of two non-linear crystals for emission and detection of THz waves. A micro-chip Nd:YAG laser and a continuous wave tunable external cavity diode laser were used as the pump and seed source, respectively. The pump beam and seed beam were injected to the LiNbO₃ crystal satisfying the noncollinear phase matching condition in order to generate high power THz-wave. The emitted THz wave was irradiated to the sample which was raster scanned by the x-z stage while changing the frequencies, and we obtained multispectral images. Then the transmitted THz wave was focused onto another crystal for detection and up-converted to the near infrared detection beam based on nonlinear optical parametric effects, wherein the detection beam intensity was measured using an infrared pyroelectric detector. It was difficult to identify reagents in a cardboard box because of high noise levels. In this work, we introduce improvements for noise reduction and image clarification, and the intensity of the near infrared detection beam was converted correctly to the intensity of the THz wave. A Gaussian spatial filter is also introduced for a clearer THz image. Through these improvements, we succeeded in identification of reagents hidden in a 42-mm thick cardboard box filled with several obstacles, which attenuate 56 dB at 1.3 THz, by improving analysis methods. Using this system, THz spectroscopic imaging was possible for saccharides and may also be applied to cases where illicit drugs are hidden in the box, and multiple reagents are mixed together. Moreover, THz spectroscopic imaging can be achieved through even thicker obstacles by introducing an NIR detector with higher sensitivity.

Keywords: nondestructive inspection, principal component analysis, terahertz parametric source, THz spectroscopic imaging

Procedia PDF Downloads 161
554 Hand Symbol Recognition Using Canny Edge Algorithm and Convolutional Neural Network

Authors: Harshit Mittal, Neeraj Garg

Abstract:

Hand symbol recognition is a pivotal component in the domain of computer vision, with far-reaching applications spanning sign language interpretation, human-computer interaction, and accessibility. This research paper discusses the approach with the integration of the Canny Edge algorithm and convolutional neural network. The significance of this study lies in its potential to enhance communication and accessibility for individuals with hearing impairments or those engaged in gesture-based interactions with technology. In the experiment mentioned, the data is manually collected by the authors from the webcam using Python codes, to increase the dataset augmentation, is applied to original images, which makes the model more compatible and advanced. Further, the dataset of about 6000 coloured images distributed equally in 5 classes (i.e., 1, 2, 3, 4, 5) are pre-processed first to gray images and then by the Canny Edge algorithm with threshold 1 and 2 as 150 each. After successful data building, this data is trained on the Convolutional Neural Network model, giving accuracy: 0.97834, precision: 0.97841, recall: 0.9783, and F1 score: 0.97832. For user purposes, a block of codes is built in Python to enable a window for hand symbol recognition. This research, at its core, seeks to advance the field of computer vision by providing an advanced perspective on hand sign recognition. By leveraging the capabilities of the Canny Edge algorithm and convolutional neural network, this study contributes to the ongoing efforts to create more accurate, efficient, and accessible solutions for individuals with diverse communication needs.

Keywords: hand symbol recognition, computer vision, Canny edge algorithm, convolutional neural network

Procedia PDF Downloads 41
553 Possible Mechanism of DM2 Development in OSA Patients Mediated via Rev-Erb-Alpha and NPAS2 Proteins

Authors: Filip Franciszek Karuga, Szymon Turkiewicz, Marta Ditmer, Marcin Sochal, Piotr Białasiewicz, Agata Gabryelska

Abstract:

Circadian rhythm, an internal coordinator of physiological processes is composed of a set of semi-autonomous clocks. Clocks are regulated through the expression of circadian clock genes which form feedback loops, creating an oscillator. The primary loop consists of activators: CLOCK, BMAL1 and repressors: CRY, PER. CLOCK can be substituted by the Neuronal PAS Domain Protein 2 (NPAS2). Orphan nuclear receptor (REV-ERB-α) is a component of the secondary major loop, modulating the expression of BMAL1. Circadian clocks might be disrupted by the obstructive sleep apnea (OSA), which has also been associated with type II diabetes mellitus (DM2). Interestingly, studies suggest that dysregulation of NPAS2 and REV-ERB-α might contribute to the pathophysiology of DM2 as well. The goal of our study was to examine the role of NPAS2 and REV-ERB-α in DM2 in OSA patients. After examination of the clinical data, all participants underwent polysomnography (PSG) to assess their apnea-hypopnea index (AHI). Based on the acquired data participants were assigned to one of 3 groups: OSA (AHI>30, no DM2; n=17 for NPAS2 and 34 for REV-ERB-α), DM2 (AHI>30 + DM2; n=7 for NPAS2 and 15 for REV-ERB-α) and control group (AHI<5, no DM2; n=16 for NPAS2 and 31 for REV-ERB-α). ELISA immunoassay was performed to assess the serum protein level of REV-ERB-α and NPAS2. The only statistically significant difference between groups was observed in NPAS2 protein level (p=0.037). Post-hoc analysis showed significant differences between the OSA and the control group (p=0.017). AHI and NPAS2 level was significantly correlated (r=-0.478, p=0.002) in all groups. A significant correlation was observed between the REV-ERB-α level and sleep efficiency (r=0.617, p=0.005) as well as sleep maintenance efficiency (r=0.645, p=0.003) in the OSA group. We conclude, that NPAS2 is associated with OSA severity and might contribute to metabolic sequelae of this disease. REV-ERB-α on the other hand can influence sleep continuity and efficiency.

Keywords: OSA, diabetes mellitus, endocrinology, chronobiology

Procedia PDF Downloads 141
552 Recent Developments in the Application of Deep Learning to Stock Market Prediction

Authors: Shraddha Jain Sharma, Ratnalata Gupta

Abstract:

Predicting stock movements in the financial market is both difficult and rewarding. Analysts and academics are increasingly using advanced approaches such as machine learning techniques to anticipate stock price patterns, thanks to the expanding capacity of computing and the recent advent of graphics processing units and tensor processing units. Stock market prediction is a type of time series prediction that is incredibly difficult to do since stock prices are influenced by a variety of financial, socioeconomic, and political factors. Furthermore, even minor mistakes in stock market price forecasts can result in significant losses for companies that employ the findings of stock market price prediction for financial analysis and investment. Soft computing techniques are increasingly being employed for stock market prediction due to their better accuracy than traditional statistical methodologies. The proposed research looks at the need for soft computing techniques in stock market prediction, the numerous soft computing approaches that are important to the field, past work in the area with their prominent features, and the significant problems or issue domain that the area involves. For constructing a predictive model, the major focus is on neural networks and fuzzy logic. The stock market is extremely unpredictable, and it is unquestionably tough to correctly predict based on certain characteristics. This study provides a complete overview of the numerous strategies investigated for high accuracy prediction, with a focus on the most important characteristics.

Keywords: stock market prediction, artificial intelligence, artificial neural networks, fuzzy logic, accuracy, deep learning, machine learning, stock price, trading volume

Procedia PDF Downloads 67
551 Cyber Security and Risk Assessment of the e-Banking Services

Authors: Aisha F. Bushager

Abstract:

Today we are more exposed than ever to cyber threats and attacks at personal, community, organizational, national, and international levels. More aspects of our lives are operating on computer networks simply because we are living in the fifth domain, which is called the Cyberspace. One of the most sensitive areas that are vulnerable to cyber threats and attacks is the Electronic Banking (e-Banking) area, where the banking sector is providing online banking services to its clients. To be able to obtain the clients trust and encourage them to practice e-Banking, also, to maintain the services provided by the banks and ensure safety, cyber security and risks control should be given a high priority in the e-banking area. The aim of the study is to carry out risk assessment on the e-banking services and determine the cyber threats, cyber attacks, and vulnerabilities that are facing the e-banking area specifically in the Kingdom of Bahrain. To collect relevant data, structured interviews were taken place with e-banking experts in different banks. Then, collected data where used as in input to the risk management framework provided by the National Institute of Standards and Technology (NIST), which was the model used in the study to assess the risks associated with e-banking services. The findings of the study showed that the cyber threats are commonly human errors, technical software or hardware failure, and hackers, on the other hand, the most common attacks facing the e-banking sector were phishing, malware attacks, and denial-of-service. The risks associated with the e-banking services were around the moderate level, however, more controls and countermeasures must be applied to maintain the moderate level of risks. The results of the study will help banks discover their vulnerabilities and maintain their online services, in addition, it will enhance the cyber security and contribute to the management and control of risks that are facing the e-banking sector.

Keywords: cyber security, e-banking, risk assessment, threats identification

Procedia PDF Downloads 333
550 The Relevance of Shared Cultural Leadership in the Survival of the Language and of the Francophone Culture in a Minority Language Environment

Authors: Lyne Chantal Boudreau, Claudine Auger, Arline Laforest

Abstract:

As an English-speaking country, Canada faces challenges in French-language education. During both editions of a provincial congress on education planned and conducted under shared cultural leadership, three organizers created a Francophone space where, for the first time in the province of New Brunswick (the only officially bilingual province in Canada), a group of stakeholders from the school, post-secondary and community sectors have succeeded in contributing to reflections on specific topics by sharing winning practices to meet the challenges of learning in a minority Francophone environment. Shared cultural leadership is a hybrid between theories of leadership styles in minority communities and theories of shared leadership. Through shared cultural leadership, the goal is simply to guide leadership and to set up all minority leaderships in minority context through shared leadership. This leadership style requires leaders to transition from a hierarchical to a horizontal approach, that is, to an approach where each individual is at the same level. In this exploratory research, it has been demonstrated that shared leadership exercised under the T-learning model best fosters the mobilization of all partners in advancing in-depth knowledge in a particular field while simultaneously allowing learning of the elements related to the domain in question. This session will present how it is possible to mobilize the whole community through leaders who continually develop their knowledge and skills in their specific field but also in related fields. Leaders in this style of management associated to shared cultural leadership acquire the ability to consider solutions to problems from a holistic perspective and to develop a collective power derived from the leadership of each and everyone in a space where all are rallied to promote the ultimate advancement of society.

Keywords: education, minority context, shared leadership, t-leaning

Procedia PDF Downloads 229
549 Multidimensional Inequality and Deprivation Among Tribal Communities of Andhra Pradesh, India

Authors: Sanjay Sinha, Mohd Umair Khan

Abstract:

The level of income inequality in India has been worrisome as the World Inequality Report termed it as a “poor and unequal country, with an affluent elite”. As important as income is to understand inequality and deprivation, it is just one dimension. But the historical roots and current realities of inequality and deprivation in India lies in many of the non-income dimensions such as housing, nutrition, education, agency, sense of inclusion etc. which are often ignored, especially in solution-oriented research. The level of inequality and deprivation among the tribal is one such case. There is a corpus of literature establishing that the tribal communities in India are disadvantageous on various grounds. Given their rural geography, issues of access and quality of basic facilities such as education and healthcare are often unaddressed. COVID-19 has further exacerbated this challenge and climate change will make it even more worrying. With this background, a succinct measurement tool at the village level is necessary to design short to medium-term actions with reference to risk mitigation for tribal communities. This research paper examines the level of inequality and deprivation among the tribal communities in the rural areas of Andhra Pradesh state of India using a Multidimensional Inequality and Deprivation Index based on the Alkire-Foster methodology. The methodology is theoretically grounded in the capability approach propounded by Amartya Sen, emphasizing on achieving the “beings and doings” (functionings) an individual reason to value. In the index, the authors have five domains, including Livelihood, Food Security, Education, Health and Housing and these domains are divided into sixteen indicators. This assessment is followed by domain-wise short-term and long-term solutions.

Keywords: Andhra Pradesh, Alkire-Foster methodology, deprivation, inequality, multidimensionality, poverty, tribal

Procedia PDF Downloads 139
548 Fast Bayesian Inference of Multivariate Block-Nearest Neighbor Gaussian Process (NNGP) Models for Large Data

Authors: Carlos Gonzales, Zaida Quiroz, Marcos Prates

Abstract:

Several spatial variables collected at the same location that share a common spatial distribution can be modeled simultaneously through a multivariate geostatistical model that takes into account the correlation between these variables and the spatial autocorrelation. The main goal of this model is to perform spatial prediction of these variables in the region of study. Here we focus on a geostatistical multivariate formulation that relies on sharing common spatial random effect terms. In particular, the first response variable can be modeled by a mean that incorporates a shared random spatial effect, while the other response variables depend on this shared spatial term, in addition to specific random spatial effects. Each spatial random effect is defined through a Gaussian process with a valid covariance function, but in order to improve the computational efficiency when the data are large, each Gaussian process is approximated to a Gaussian random Markov field (GRMF), specifically to the block nearest neighbor Gaussian process (Block-NNGP). This approach involves dividing the spatial domain into several dependent blocks under certain constraints, where the cross blocks allow capturing the spatial dependence on a large scale, while each individual block captures the spatial dependence on a smaller scale. The multivariate geostatistical model belongs to the class of Latent Gaussian Models; thus, to achieve fast Bayesian inference, it is used the integrated nested Laplace approximation (INLA) method. The good performance of the proposed model is shown through simulations and applications for massive data.

Keywords: Block-NNGP, geostatistics, gaussian process, GRMF, INLA, multivariate models.

Procedia PDF Downloads 79
547 Study of Mechanical Properties of Large Scale Flexible Silicon Solar Modules on the Various Substrates

Authors: M. Maleczek, Leszek Bogdan, Kazimierz Drabczyk, Agnieszka Iwan

Abstract:

Crystalline silicon (Si) solar cells are the main product in the market among the various photovoltaic technologies concerning such advantages as: material richness, high carrier mobilities, broad spectral absorption range and established technology. However, photovoltaic technology on the stiff substrates are heavier, more fragile and less cost-effective than devices on the flexible substrates to be applied in special applications. The main goal of our work was to incorporate silicon solar cells into various fabric, without any change of the electrical and mechanical parameters of devices. This work is realized for the GEKON project (No. GEKON2/O4/268473/23/2016) sponsored by The National Centre for Research and Development and The National Fund for Environmental Protection and Water Management. In our work, the polyamide or polyester fabrics were used as a flexible substrate in the created devices. Applied fabrics differ in tensile and tear strength. All investigated polyamide fabrics are resistant to weathering and UV, while polyester ones is resistant to ozone, water and ageing. The examined fabrics are tight at 100 cm water per 2 hours. In our work, commercial silicon solar cells with the size 156 × 156 mm were cut into nine parts (called single solar cells) by diamond saw and laser. Gap and edge after cutting of solar cells were checked by transmission electron microscope (TEM) to study morphology and quality of the prepared single solar cells. Modules with the size of 160 × 70 cm (containing about 80 single solar cells) were created and investigated by electrical and mechanical methods. Weight of constructed module is about 1.9 kg. Three types of solar cell architectures such as: -fabric/EVA/Si solar cell/EVA/film for lamination, -backsheet PET/EVA/Si solar cell/EVA/film for lamination, -fabric/EVA/Si solar cell/EVA/tempered glass, were investigated taking into consideration type of fabric and lamination process together with the size of solar cells. In investigated devices EVA, it is ethylene-vinyl acetate, while PET - polyethylene terephthalate. Depend on the lamination process and compatibility of textile with solar cell an efficiency of investigated flexible silicon solar cells was in the range of 9.44-16.64 %. Multi folding and unfolding of flexible module has no impact on its efficiency as was detected by Instron equipment. Power (P) of constructed solar module is 30 W, while voltage about 36 V. Finally, solar panel contains five modules with the polyamide fabric and tempered glass will be produced commercially for different applications (dual use).

Keywords: flexible devices, mechanical properties, silicon solar cells, textiles

Procedia PDF Downloads 160
546 Faster Pedestrian Recognition Using Deformable Part Models

Authors: Alessandro Preziosi, Antonio Prioletti, Luca Castangia

Abstract:

Deformable part models achieve high precision in pedestrian recognition, but all publicly available implementations are too slow for real-time applications. We implemented a deformable part model algorithm fast enough for real-time use by exploiting information about the camera position and orientation. This implementation is both faster and more precise than alternative DPM implementations. These results are obtained by computing convolutions in the frequency domain and using lookup tables to speed up feature computation. This approach is almost an order of magnitude faster than the reference DPM implementation, with no loss in precision. Knowing the position of the camera with respect to horizon it is also possible prune many hypotheses based on their size and location. The range of acceptable sizes and positions is set by looking at the statistical distribution of bounding boxes in labelled images. With this approach it is not needed to compute the entire feature pyramid: for example higher resolution features are only needed near the horizon. This results in an increase in mean average precision of 5% and an increase in speed by a factor of two. Furthermore, to reduce misdetections involving small pedestrians near the horizon, input images are supersampled near the horizon. Supersampling the image at 1.5 times the original scale, results in an increase in precision of about 4%. The implementation was tested against the public KITTI dataset, obtaining an 8% improvement in mean average precision over the best performing DPM-based method. By allowing for a small loss in precision computational time can be easily brought down to our target of 100ms per image, reaching a solution that is faster and still more precise than all publicly available DPM implementations.

Keywords: autonomous vehicles, deformable part model, dpm, pedestrian detection, real time

Procedia PDF Downloads 260
545 Understanding Cyber Kill Chains: Optimal Allocation of Monitoring Resources Using Cooperative Game Theory

Authors: Roy. H. A. Lindelauf

Abstract:

Cyberattacks are complex processes consisting of multiple interwoven tasks conducted by a set of agents. Interdictions and defenses against such attacks often rely on cyber kill chain (CKC) models. A CKC is a framework that tries to capture the actions taken by a cyber attacker. There exists a growing body of literature on CKCs. Most of this work either a) describes the CKC with respect to one or more specific cyberattacks or b) discusses the tools and technologies used by the attacker at each stage of the CKC. Defenders, facing scarce resources, have to decide where to allocate their resources given the CKC and partial knowledge on the tools and techniques attackers use. In this presentation CKCs are analyzed through the lens of covert projects, i.e., interrelated tasks that have to be conducted by agents (human and/or computer) with the aim of going undetected. Various aspects of covert project models have been studied abundantly in the operations research and game theory domain, think of resource-limited interdiction actions that maximally delay completion times of a weapons project for instance. This presentation has investigated both cooperative and non-cooperative game theoretic covert project models and elucidated their relation to CKC modelling. To view a CKC as a covert project each step in the CKC is broken down into tasks and there are players of which each one is capable of executing a subset of the tasks. Additionally, task inter-dependencies are represented by a schedule. Using multi-glove cooperative games it is shown how a defender can optimize the allocation of his scarce resources (what, where and how to monitor) against an attacker scheduling a CKC. This study presents and compares several cooperative game theoretic solution concepts as metrics for assigning resources to the monitoring of agents.

Keywords: cyber defense, cyber kill chain, game theory, information warfare techniques

Procedia PDF Downloads 126