Search results for: 3D object reconstruction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1804

Search results for: 3D object reconstruction

1414 Monocular 3D Person Tracking AIA Demographic Classification and Projective Image Processing

Authors: McClain Thiel

Abstract:

Object detection and localization has historically required two or more sensors due to the loss of information from 3D to 2D space, however, most surveillance systems currently in use in the real world only have one sensor per location. Generally, this consists of a single low-resolution camera positioned above the area under observation (mall, jewelry store, traffic camera). This is not sufficient for robust 3D tracking for applications such as security or more recent relevance, contract tracing. This paper proposes a lightweight system for 3D person tracking that requires no additional hardware, based on compressed object detection convolutional-nets, facial landmark detection, and projective geometry. This approach involves classifying the target into a demographic category and then making assumptions about the relative locations of facial landmarks from the demographic information, and from there using simple projective geometry and known constants to find the target's location in 3D space. Preliminary testing, although severely lacking, suggests reasonable success in 3D tracking under ideal conditions.

Keywords: monocular distancing, computer vision, facial analysis, 3D localization

Procedia PDF Downloads 139
1413 Turkish Airlines' 85th Anniversary Commercial: An Analysis of the Institutional Identity of a Brand in Terms of Glocalization

Authors: Samil Ozcan

Abstract:

Airlines companies target different customer segments in consideration of pricing, service quality, flight network, etc. and their brand positioning accords with the marketization strategies developed in the same direction. The object of this study, Turkish Airlines, has many peculiarities regarding its brand positioning as compared to its rivals in the sector. In the first place, it appeals to a global customer group because of its Star Alliance membership and its broad flight network with 315 destination points. The second group in its customer segmentation includes domestic customers. For this group, the company follows a marketing strategy that plays to local culture and accentuates the image of Turkishness as an emotional allurement. The advertisements and publicity projects designed in this regard put little emphasis on the service quality the company offers to its clients; it addresses the emotions of the consumers rather than individual benefits and relies on the historical memory of the nation and shared cultural values. This study examines the publicity work which aims at the second segment customer group focusing on Turkish Airlines’ 85th Anniversary Commercial through a symbolic meaning analysis approach. The commercial presents six stories with undertones of nationalism in its theme. Nationalism is not just the product of collective interests based on reason but a result of patriotism in the sense of loyalty to state and nation and love of ethnic belonging. While nationalism refers to concrete notions such as blood tie, common ancestor, shared history, it is not the actuality of these notions that it draws its real strength but the emotions invested in them. The myths of origin, the idea of common homeland, boundary definitions, and symbolic acculturation have instrumental importance in the development of these commonalities. The commercial offers concrete examples for an analysis of Connor’s definition of nationalism based on emotions. Turning points in the history of the Turkish Republic and the historical mission Turkish Airlines undertook in these moments are narrated in six stories in the commercial with a highly emotional theme. These emotions, in general, depend on collective memory generated by national consciousness. Collective memory is not simply remembering the past. It is constructed through the reconstruction and reinterpretation of the past in the present moment. This study inquires the motivations behind the nationalist emotions generated within the collective memory by engaging with the commercial released for the 85th anniversary of Turkish Airlines as the object of analysis. Symbols and myths can be read as key concepts that reveal the relation between 'identity and memory'. Because myths and symbols do not merely reflect on collective memory, they reconstruct it as well. In this sense, the theme of the commercial defines the image of Turkishness with virtues such as self-sacrifice, helpfulness, humanity, and courage through a process of meaning creation based on symbolic mythologizations like flag and homeland. These virtues go beyond describing the image of Turkishness and become an instrument that defines and gives meaning to Turkish identity.

Keywords: collective memory, emotions, identity, nationalism

Procedia PDF Downloads 153
1412 Pilot Study of Determining the Impact of Surface Subsidence at The Intersection of Cave Mining with the Surface Using an Electrical Impedance Tomography

Authors: Ariungerel Jargal

Abstract:

: Cave mining is a bulk underground mining method, which allows large low-grade deposits to be mined underground. This method involves undermining the orebody to make it collapse under its own weight into a series of chambers from which the ore extracted. It is a useful technique to extend the life of large deposits previously mined by open pits, and it is a method increasingly proposed for new mines around the world. We plan to conduct a feasibility study using Electrical impedance tomography (EIT) technology to show how much subsidence there is at the intersection with the cave mining surface. EIT is an imaging technique which uses electrical measurements at electrodes attached on the body surface to yield a cross-sectional image of conductivity changes within the object. EIT has been developed in several different applications areas as a simpler, cheaper alternative to many other imaging methods. A low frequency current is injected between pairs of electrodes while voltage measurements are collected at all other electrode pairs. In the difference EIT, images are reconstructed of the change in conductivity distribution (σ) between the acquisition of the two sets of measurements. Image reconstruction in EIT requires the solution of an ill-conditioned nonlinear inverse problem on noisy data, typically requiring make simpler assumptions or regularization. It is noted that the ratio of current to voltage represents a complex value according to Ohm’s law, and that it is theoretically possible to re-express EIT. The results of the experiment were presented on the simulation, and it was concluded that it is possible to conduct further real experiments. Drill a certain number of holes in the top wall of the cave to attach the electrodes, flow a current through them, and measure and acquire the potential through these electrodes. Appropriate values should be selected depending on the distance between the holes, the frequency and duration of the measurements, the surface characteristics and the size of the study area using an EIT device.

Keywords: impedance tomography, cave mining, soil, EIT device

Procedia PDF Downloads 126
1411 Modeling of Building a Conceptual Scheme for Multimodal Freight Transportation Information System

Authors: Gia Surguladze, Nino Topuria, Lily Petriashvili, Giorgi Surguladze

Abstract:

Modeling of building processes of a multimodal freight transportation support information system is discussed based on modern CASE technologies. Functional efficiencies of ports in the eastern part of the Black Sea are analyzed taking into account their ecological, seasonal, resource usage parameters. By resources, we mean capacities of berths, cranes, automotive transport, as well as work crews and neighbouring airports. For the purpose of designing database of computer support system for Managerial (Logistics) function, using Object-Role Modeling (ORM) tool (NORMA – Natural ORM Architecture) is proposed, after which Entity Relationship Model (ERM) is generated in automated process. The software is developed based on Process-Oriented and Service-Oriented architecture, in Visual Studio.NET environment.

Keywords: seaport resources, business-processes, multimodal transportation, CASE technology, object-role model, entity relationship model, SOA

Procedia PDF Downloads 430
1410 A Psychoanalytic Lens: Unmasked Layers of the Self among Post-Graduate Psychology Students in Surviving the COVID-19 Lockdown

Authors: Sharon Sibanda, Benny Motileng

Abstract:

The World Health Organisation (WHO) identified the Sars-Cov-2 (COVID-19) as a pandemic on the 12ᵗʰ of March 2020, with South Africa recording its first case on the 5ᵗʰ of March 2020. The rapidly spreading virus led the South African government to implement one of the strictest nationwide lockdowns globally, resulting in the closing down of all institutions of higher learning effective March 18ᵗʰ 2020. Thus, this qualitative study primarily aimed to explore whether post-graduate psychology students were in a state of a depleted or cohesive self, post the psychological isolation of COVID-19 risk-adjusted level 5 lockdown. Semi-structured interviews from a qualitative interpretive approach comprising N=6 psychology post-graduate students facilitated a rich understanding of their intra-psychic experiences of the self. Thematic analysis of data gathered from the interviews illuminated how students were forced into the self by the emotional isolation of hard lockdown, with the emergence of core psychic conflict often defended against through external self-object experiences. The findings also suggest that lockdown stripped off this sample of psychology post-graduate students’ defensive escape from the inner self through external self-object distractions. The external self was stripped to the core of the internal self by the isolation of hard lockdown, thereby uncovering the psychic function of roles and defenses amalgamated throughout modern cultural consciousness that dictates self-functioning. The study suggests modelling reflexivity skills in the integration of internal and external self-experience dynamics as part of a training model for continued personal and professional development for psychology students.

Keywords: COVID-19, fragmentation, self-object experience, true/false self

Procedia PDF Downloads 58
1409 Palaeo-Environmental Reconstruction of the Wet Zone of Sri Lanka: A Zooarchaeological Perspective

Authors: Kalangi Rodrigo

Abstract:

Sri Lanka has been known as an island which has a diverse variety of prehistoric occupation among ecological zones. Defining the paleoecology of the past societies has been an archaeological thought developed in the 1960s. It is mainly concerned with the reconstruction from available geological and biological evidence of past biota, populations, communities, landscapes, environments, and ecosystems. Sri Lanka has dealt with this subject, and considerable research has been already undertaken. The fossil and material record of Sri Lanka’s Wet Zone tropical forests continues from c. 38,000–34,000 ybp. This early and persistent human fossil, technical, and cultural florescence, as well as a collection of well-preserved tropical-forest rock shelters with associated 'on-site' palaeoenvironmental records, makes Sri Lanka a central and unusual case study to determine the extent and strength of early human tropical forest encounters. Excavations carried out in prehistoric caves in the low country wet zone has shown that in the last 50,000 years, the temperature in the lowland rainforests has not exceeded 5°C. When taking Batadombalena alone, the entire seven layers have yielded an uninterrupted occupation of Acavus sp and Canerium zeylanicum, a plant that grows in the middle of the rainforest. Acavus, which is highly sensitive to rainforest ecosystems, has been well documented in many of the lowland caves, confirming that the wetland rainforest environment has remained intact at least for the last 50,000 years. If the dry and arid conditions in the upper hills regions affected the wet zone, the Tufted Gray Lunger (semnopithecus priam), must also meet with the prehistoric caves in the wet zone thrown over dry climate. However, the bones in the low country wet zone do not find any of the fragments belonging to Turfed Gray Lunger, and prehistoric human consumption is bestowed with purple-faced leaf monkey (Trachypithecus vetulus) and Toque Macaque (Macaca Sinica). The skeletal remains of Lyriocephalus scutatus, a full-time resident in rain forests, have also been recorded among lowland caves. But, in zoological terms, these remains may be the remains of the Barking deer (Muntiacus muntjak), which is currently found in the wet zone. For further investigations, the mtDNA test of genetic diversity (Bottleneck effect) and pollen study from lowland caves should determine whether the wet zone climate has persisted over the last 50,000 years, or whether the dry weather affected in the mountainous region has invaded the wet zone.

Keywords: paleoecology, prehistory, zooarchaeology, reconstruction, palaeo-climate

Procedia PDF Downloads 140
1408 Automated Transformation of 3D Point Cloud to BIM Model: Leveraging Algorithmic Modeling for Efficient Reconstruction

Authors: Radul Shishkov, Orlin Davchev

Abstract:

The digital era has revolutionized architectural practices, with building information modeling (BIM) emerging as a pivotal tool for architects, engineers, and construction professionals. However, the transition from traditional methods to BIM-centric approaches poses significant challenges, particularly in the context of existing structures. This research introduces a technical approach to bridge this gap through the development of algorithms that facilitate the automated transformation of 3D point cloud data into detailed BIM models. The core of this research lies in the application of algorithmic modeling and computational design methods to interpret and reconstruct point cloud data -a collection of data points in space, typically produced by 3D scanners- into comprehensive BIM models. This process involves complex stages of data cleaning, feature extraction, and geometric reconstruction, which are traditionally time-consuming and prone to human error. By automating these stages, our approach significantly enhances the efficiency and accuracy of creating BIM models for existing buildings. The proposed algorithms are designed to identify key architectural elements within point clouds, such as walls, windows, doors, and other structural components, and to translate these elements into their corresponding BIM representations. This includes the integration of parametric modeling techniques to ensure that the generated BIM models are not only geometrically accurate but also embedded with essential architectural and structural information. Our methodology has been tested on several real-world case studies, demonstrating its capability to handle diverse architectural styles and complexities. The results showcase a substantial reduction in time and resources required for BIM model generation while maintaining high levels of accuracy and detail. This research contributes significantly to the field of architectural technology by providing a scalable and efficient solution for the integration of existing structures into the BIM framework. It paves the way for more seamless and integrated workflows in renovation and heritage conservation projects, where the accuracy of existing conditions plays a critical role. The implications of this study extend beyond architectural practices, offering potential benefits in urban planning, facility management, and historic preservation.

Keywords: BIM, 3D point cloud, algorithmic modeling, computational design, architectural reconstruction

Procedia PDF Downloads 63
1407 Digital Forensic Exploration Framework for Email and Instant Messaging Applications

Authors: T. Manesh, Abdalla A. Alameen, M. Mohemmed Sha, A. Mohamed Mustaq Ahmed

Abstract:

Email and instant messaging applications are foremost and extensively used electronic communication methods in this era of information explosion. These applications are generally used for exchange of information using several frontend applications from various service providers by its users. Almost all such communications are now secured using SSL or TLS security over HTTP communication. At the same time, it is also noted that cyber criminals and terrorists have started exchanging information using these methods. Since communication is encrypted end-to-end, tracing significant forensic details and actual content of messages are found to be unattended and severe challenges by available forensic tools. These challenges seriously affect in procuring substantial evidences against such criminals from their working environments. This paper presents a vibrant forensic exploration and architectural framework which not only decrypts any communication or network session but also reconstructs actual message contents of email as well as instant messaging applications. The framework can be effectively used in proxy servers and individual computers and it aims to perform forensic reconstruction followed by analysis of webmail and ICQ messaging applications. This forensic framework exhibits a versatile nature as it is equipped with high speed packet capturing hardware, a well-designed packet manipulating algorithm. It regenerates message contents over regular as well as SSL encrypted SMTP, POP3 and IMAP protocols and catalyzes forensic presentation procedure for prosecution of cyber criminals by producing solid evidences of their actual communication as per court of law of specific countries.

Keywords: forensics, network sessions, packet reconstruction, packet reordering

Procedia PDF Downloads 344
1406 Pre-Service Mathematics Teachers’ Mental Construction in Solving Equations and Inequalities Using ACE Teaching Cycle

Authors: Abera Kotu, Girma Tesema, Mitiku Tadesse

Abstract:

This study investigated ACE supported instruction and pre-service mathematics teachers’ mental construction in solving equations and inequalities. A mixed approach with concurrent parallel design was employed. It was conducted on two intact groups of regular first-year pre-service mathematics teachers at Fiche College of Teachers’ Education in which one group was assigned as an intervention group and the other group as a comparison group using the lottery method. There were 33 participants in the intervention and 32 participants in the comparison. Six pre-service mathematics teachers were selected for interview using purposive sampling based on pre-test results. An instruction supported with ACE cycle was given to the intervention group for two weeks duration of time. Written tasks, interviews, and observations were used to collect data. Data collected from written tasks were analyzed quantitatively using independent samples t-test and effect size. Data collected from interviews and observations were analyzed narratively. The findings of the study uncovered that ACE-supported instruction has a moderate effect on Pre-service Mathematics Teachers’ levels of conceptualizations of action, process, object, ad schema. Moreover, the ACE supported group out scored and performed better than the usual traditional method supported groups across the levels of conceptualization. The majority of pre-service mathematics teachers’ levels of conceptualizations were at action and process levels and their levels of conceptualization were linked with genetic decomposition more at action and object levels than object and schema. The use of ACE supported instruction is recommended to improve pre-service mathematics teachers’ mental construction.

Keywords: ACE teaching cycle, APOS theory, mental construction, genetic composition

Procedia PDF Downloads 16
1405 An Investigation of Aluminum Foil-Epoxy Laminated Composites for Rapid Tooling Applications

Authors: Kevlin Govender, Anthony Walker, Glen Bright

Abstract:

Mass customization is an area of increased importance and the development of rapid tooling applications is pivotal to the success of mass customization. This paper presents a laminated object manufacturing (LOM) process for rapid tooling. The process is termed 3D metal laminate printing and utilizes domestic-grade aluminum foil and epoxy for layered manufacturing. A detailed explanation of the process is presented to produce complex metal laminated composite parts. Aluminum-epoxy composite specimens were manufactured from 0.016mm aluminum and subjected to tensile tests to determine the mechanical properties of the manufactured composite in relation to solid metal specimens. The fracture zone of the specimens was analyzed under scanning electron microscopy (SEM) in order to characterize the fracture mode and study the interfacial bonding of the manufactured laminate specimens.

Keywords: 3D metal laminate printer, aluminum-epoxy composite, laminated object manufacturing, rapid tooling

Procedia PDF Downloads 290
1404 Comparative Analysis of Automation Testing Tools

Authors: Amit Bhanushali

Abstract:

In the ever-changing landscape of software development, automated software testing has emerged as a critical component of the Software Development Life Cycle (SDLC). This research undertakes a comparative study of three major automated testing tools -UFT, Selenium, and RPA- evaluating them on usability, maintenance, and effectiveness. Leveraging existing JAVA-based applications as test cases, the study aims to guide testers in selecting the optimal tool for specific applications. By exploring key features such as source and licensing, testing expenses, object repositories, usability, and language support, the research provides practical insights into UFT, Selenium, and RPA. Acknowledging the pivotal role of these tools in streamlining testing processes amid time constraints and resource limitations, the study assists professionals in making informed choices aligned with their organizational needs.

Keywords: software testing tools, software development lifecycle (SDLC), test automation frameworks, automated software, JAVA-based, UFT, selenium and RPA (robotic process automation), source and licensing, object repository

Procedia PDF Downloads 97
1403 Cross-Cultural Study of Stroop Interference among Juvenile Delinquents

Authors: Tanusree Moitra, Garga Chatterjee, Diganta Mukherjee, Anjali Ghosh

Abstract:

Stroop task is considered to be an important measure of selective attention. However, the color – word Stroop task cannot be administered to the illiterate population. Some of the participants in the present study are illiterate, therefore, object – color Stroop task was used among male juvenile delinquents of India and Bangladesh citizenship (IC & BC), housed in delinquent home in India. The purpose of the study is to test the hypothesis that over - selective attention is present among juvenile delinquents across both the countries. Eighty juvenile delinquents and matched control of 12 – 18 years (50 IC juvenile delinquents, 30 BC juvenile delinquents and 50 Indian control) were shown 24 familiar objects in both typical (e.g. a red apple) and atypical (e.g. a blue apple) color. Repeated – measure factorial ANOVA was used and it was found that all the three groups have taken longer response time in the atypical condition compared to the typical condition. However, contrary to the over - selective attention hypothesis, both groups of juvenile delinquents displayed higher Stroop interference in comparison to the matched control group. The findings of the study can be explained on the basis of anxiety score. IC and BC juvenile delinquents have high anxiety score compared to the control group which indicates that increased anxiety is correlated with the interference produced by the atypical color object stimuli when compared with the typical object stimuli. Funding acknowledgement: Authors acknowledge Department of Science and Technology, Government of India for financial support to the first author of the paper vide Reference no. SR/CSRI/PDF -01/2013 under Cognitive Science Research Initiative (CSRI) to carry out this work.

Keywords: Bangladesh, India, male juvenile delinquent, objects - color Stroop task

Procedia PDF Downloads 343
1402 Immobilized Iron Oxide Nanoparticles for Stem Cell Reconstruction in Magnetic Particle Imaging

Authors: Kolja Them, Johannes Salamon, Harald Ittrich, Michael Kaul, Tobias Knopp

Abstract:

Superparamagnetic iron oxide nanoparticles (SPIONs) are nanoscale magnets which can be biologically functionalized for biomedical applications. Stem cell therapies to repair damaged tissue, magnetic fluid hyperthermia for cancer therapy and targeted drug delivery based on SPIONs are prominent examples where the visualization of a preferably low concentrated SPION distribution is essential. In 2005 a new method for tomographic SPION imaging has been introduced. The method named magnetic particle imaging (MPI) takes advantage of the nanoparticles magnetization change caused by an oscillating, external magnetic field and allows to directly image the time-dependent nanoparticle distribution. The SPION magnetization can be changed by the electron spin dynamics as well as by a mechanical rotation of the nanoparticle. In this work different calibration methods in MPI are investigated for image reconstruction of magnetically labeled stem cells. It is shown that a calibration using rotationally immobilized SPIONs provides a higher quality of stem cell images with fewer artifacts than a calibration using mobile SPIONs. The enhancement of the image quality and the reduction of artifacts enables the localization and identification of a smaller number of magnetically labeled stem cells. This is important for future medical applications where low concentrations of functionalized SPIONs interacting with biological matter have to be localized.

Keywords: biomedical imaging, iron oxide nanoparticles, magnetic particle imaging, stem cell imaging

Procedia PDF Downloads 464
1401 Separating Permanent and Induced Magnetic Signature: A Simple Approach

Authors: O. J. G. Somsen, G. P. M. Wagemakers

Abstract:

Magnetic signature detection provides sensitive detection of metal objects, especially in the natural environment. Our group is developing a tabletop setup for magnetic signatures of various small and model objects. A particular issue is the separation of permanent and induced magnetization. While the latter depends only on the composition and shape of the object, the former also depends on the magnetization history. With common deperming techniques, a significant permanent signature may still remain, which confuses measurements of the induced component. We investigate a basic technique of separating the two. Measurements were done by moving the object along an aluminum rail while the three field components are recorded by a detector attached near the center. This is done first with the rail parallel to the Earth magnetic field and then with anti-parallel orientation. The reversal changes the sign of the induced- but not the permanent magnetization so that the two can be separated. Our preliminary results on a small iron block show excellent reproducibility. A considerable permanent magnetization was indeed present, resulting in a complex asymmetric signature. After separation, a much more symmetric induced signature was obtained that can be studied in detail and compared with theoretical calculations.

Keywords: magnetic signature, data analysis, magnetization, deperming techniques

Procedia PDF Downloads 451
1400 The Potential of Southern Malang as Geotourism Site: The Distribution of Geodiversity and Geotrek in Southern Malang, Indonesia

Authors: Arda Bagus M, Yehezkiel Festian P, Budianto Santoso

Abstract:

The Tourism Area of Southern Malang is administratively located in the Regency of Malang, East Java Province, Indonesia and geographically is in a position between 112o17' - 112o57' E dan 7o44' - 8o26' S. Southern Malang consists of several sub-districts that directly borders with the Indian Ocean, such as Donomulyo, Bantur, Gedangan, Sumbermanjing, Tirto Yudo, and Ampel Gading. This area has a high geotourism potential because of the existence of geodiversity such as beaches, waterfalls, caves, and karst area. However, to the best of the authors’ knowledge, there is still no systematic data that informs the geotourism potentials to the public. The aim of this research is to complete the lack of data and then arrange it systematically so it can be used for both tourism and research purposes. Research methods such as field observation, literature study, and depth interview to local people have been implemented. Aspects reviewed by visiting the field are accommodation, transportation, and the feasibility of a place to be geotourism object. The primary data was taken in Sumbermanjing, Gedangan, Bantur, and Donomulyo sub-district. A literature study is needed to determine the regional geology of Southern Malang and as a comparison to new data obtained in the field. The results of the literature study show that southern Malang consists of three formations: Wonosari Formation, Mandalaka Formation, and River-swamps Sediment Formation with the age range of Oligocene to Quaternary. Depth interviews have been conducted by involving local people with the aim of knowing cultural-history in the research area. From this research, the geotourism object distribution map has been made. The map also includes Geotrek and basic geological information of each object. The results of this research can support the development of geotourism in Southern Malang.

Keywords: geodiversity, geology, geotourism, geotrek, southern Malang

Procedia PDF Downloads 175
1399 Image Classification with Localization Using Convolutional Neural Networks

Authors: Bhuyain Mobarok Hossain

Abstract:

Image classification and localization research is currently an important strategy in the field of computer vision. The evolution and advancement of deep learning and convolutional neural networks (CNN) have greatly improved the capabilities of object detection and image-based classification. Target detection is important to research in the field of computer vision, especially in video surveillance systems. To solve this problem, we will be applying a convolutional neural network of multiple scales at multiple locations in the image in one sliding window. Most translation networks move away from the bounding box around the area of interest. In contrast to this architecture, we consider the problem to be a classification problem where each pixel of the image is a separate section. Image classification is the method of predicting an individual category or specifying by a shoal of data points. Image classification is a part of the classification problem, including any labels throughout the image. The image can be classified as a day or night shot. Or, likewise, images of cars and motorbikes will be automatically placed in their collection. The deep learning of image classification generally includes convolutional layers; the invention of it is referred to as a convolutional neural network (CNN).

Keywords: image classification, object detection, localization, particle filter

Procedia PDF Downloads 305
1398 The Dependence of the Liquid Application on the Coverage of the Sprayed Objects in Terms of the Characteristics of the Sprayed Object during Spraying

Authors: Beata Cieniawska, Deta Łuczycka, Katarzyna Dereń

Abstract:

When assessing the quality of the spraying procedure, three indicators are used: uneven distribution of precipitation of liquid sprayed, degree of coverage of sprayed surfaces, and deposition of liquid spraying However, there is a lack of information on the relationship between the quality parameters of the procedure. Therefore, the research was carried out at the Institute of Agricultural Engineering of Wrocław University of Environmental and Life Sciences. The aim of the study was to determine the relationship between the degree of coverage of sprayed surfaces and the deposition of liquid in the aspect of the parametric characteristics of the protected plant using selected single and double stream nozzles. Experiments were conducted under laboratory conditions. The carrier of nozzles acted as an independent self-propelled sprayer used for spraying, whereas the parametric characteristics of plants were determined using artificial plants as the ratio of the vertical projection surface and the horizontal projection surface. The results and their analysis showed a strong and very strong correlation between the analyzed parameters in terms of the characteristics of the sprayed object.

Keywords: degree of coverage, deposition of liquid, nozzle, spraying

Procedia PDF Downloads 334
1397 Silymarin Reverses Scopolamine-Induced Memory Deficit in Object Recognition Test in Rats: A Behavioral, Biochemical, Histopathological and Immunohistochemical Study

Authors: Salma A. El-Marasy, Reham M. Abd-Elsalam, Omar A. Ahmed-Farid

Abstract:

Dementia is characterized by impairments in memory and other cognitive abilities. This study aims to elucidate the possible ameliorative effect of silymarin on scopolamine-induced dementia using the object recognition test (ORT). The study was extended to demonstrate the role of cholinergic activity, oxidative stress, neuroinflammation, brain neurotransmitters and histopathological changes in the anti-amnestic effect of silymarin in demented rats. Wistar rats were pretreated with silymarin (200, 400, 800 mg/kg) or donepezil (10 mg/kg) orally for 14 consecutive days. Dementia was induced after the last drug administration by a single intraperitoneal dose of scopolamine (16 mg/kg). Then behavioral, biochemical, histopathological, and immunohistochemical analyses were then performed. Rats pretreated with silymarin counteracted scopolamine-induced non-spatial working memory impairment in the ORT and decreased acetylcholinesterase (AChE) activity, reduced malondialdehyde (MDA), elevated reduced glutathione (GSH), restored gamma-aminobutyric acid (GABA) and dopamine (DA) contents in the cortical and hippocampal brain homogenates. Silymarin dose-dependently reversed scopolamine-induced histopathological changes. Immunohistochemical analysis showed that silymarin dose-dependently mitigated protein expression of a glial fibrillary acidic protein (GFAP) and nuclear factor kappa-B (NF-κB) in the brain cortex and hippocampus. All these effects of silymarin were similar to that of the standard anti-amnestic drug, donepezil. This study reveals that the ameliorative effect of silymarin on scopolamine-induced dementia in rats using the ORT maybe in part mediated by, enhancement of cholinergic activity, anti-oxidant and anti-inflammatory activities as well as mitigation in brain neurotransmitters and histopathological changes.

Keywords: dementia, donepezil, object recognition test, rats, silymarin, scopolamine

Procedia PDF Downloads 138
1396 Reconstruction of Visual Stimuli Using Stable Diffusion with Text Conditioning

Authors: ShyamKrishna Kirithivasan, Shreyas Battula, Aditi Soori, Richa Ramesh, Ramamoorthy Srinath

Abstract:

The human brain, among the most complex and mysterious aspects of the body, harbors vast potential for extensive exploration. Unraveling these enigmas, especially within neural perception and cognition, delves into the realm of neural decoding. Harnessing advancements in generative AI, particularly in Visual Computing, seeks to elucidate how the brain comprehends visual stimuli observed by humans. The paper endeavors to reconstruct human-perceived visual stimuli using Functional Magnetic Resonance Imaging (fMRI). This fMRI data is then processed through pre-trained deep-learning models to recreate the stimuli. Introducing a new architecture named LatentNeuroNet, the aim is to achieve the utmost semantic fidelity in stimuli reconstruction. The approach employs a Latent Diffusion Model (LDM) - Stable Diffusion v1.5, emphasizing semantic accuracy and generating superior quality outputs. This addresses the limitations of prior methods, such as GANs, known for poor semantic performance and inherent instability. Text conditioning within the LDM's denoising process is handled by extracting text from the brain's ventral visual cortex region. This extracted text undergoes processing through a Bootstrapping Language-Image Pre-training (BLIP) encoder before it is injected into the denoising process. In conclusion, a successful architecture is developed that reconstructs the visual stimuli perceived and finally, this research provides us with enough evidence to identify the most influential regions of the brain responsible for cognition and perception.

Keywords: BLIP, fMRI, latent diffusion model, neural perception.

Procedia PDF Downloads 68
1395 Application of Pedicled Perforator Flaps in Large Cavities of the Breast

Authors: Neerja Gupta

Abstract:

Objective-Reconstruction of large cavities of the breast without contralateral symmetrisation Background- Reconstruction of breast includes a wide spectrum of procedures from displacement to regional and distant flaps. The pedicled Perforator flaps cover a wide spectrum of reconstruction surgery for all quadrants of the breast, especially in patients with comorbidities. These axial flaps singly or adjunct are based on a near constant perforator vessel, a ratio of 2:1 at its entry in a flap is good to maintain vascularity. The perforators of lateral chest wall viz LICAP, LTAP have overlapping perfurosomes without clear demarcation. LTAP is localized in the narrow zone between the lateral breast fold and anterior axillary line,2.5-3.8cm from the fold. MICAP are localized at 1-2 cm from sternum. Being 1-2mm in diameter, a Single perforator is good to maintain the flap. LICAP has a dominant perforator in 6th-11th spaces, while LTAP has higher placed dominant perforators in 4th and 5th spaces. Methodology-Six consecutive patients who underwent reconstruction of the breast with pedicled perforator flaps were retrospectively analysed. Selections of the flap was done based on the size and locations of the tumour, anticipated volume loss, willingness to undergo contralateral symmetrisation, cosmetic expectations, and finances available.3 patients underwent vertical LTAP, the distal limit of the flap being the inframammary crease. 3 patients underwent MICAP, oriented along the axis of rib, the distal limit being the anterior axillary line. Preoperative identification was done using a unidirectional hand held doppler. The flap was raised caudal to cranial, the pivot point of rotation being the vessel entry into the skin. The donor area is determined by the skin pinch. Flap harvest time was 20-25 minutes. Intra operative vascularity was assessed with dermal bleed. The patient immediate pre, post-operative and follow up pics were compared independently by two breast surgeons. Patients were given a breast Q questionnaire (licensed) for scoring. Results-The median age of six patients was 46. Each patient had a hospital stay of 24 hours. None of the patients was willing for contralateral symmetrisation. The specimen dimensions were from 8x6.8x4 cm to 19x16x9 cm. The breast volume reconstructed range was 30 percent to 45 percent. All wide excision had free margins on frozen. The mean flap dimensions were 12x5x4.5 cm. One LTAP underwent marginal necrosis and delayed wound healing due to seroma. Three patients were phyllodes, of which one was borderline, and 2 were benign on final histopathology. All other 3 patients were invasive ductal cancer and have completed their radiation. The median follow up is 7 months the satisfaction scores at median follow of 7 months are 90 for physical wellbeing and 85 for surgical results. Surgeons scored fair to good in Harvard score. Conclusion- Pedicled perforator flaps are a valuable option for 3/8th volume of breast defects. LTAP is preferred for tumours at the Central, upper, and outer quadrants of the breast and MICAP for the inner and lower quadrant. The vascularity of the flap is dependent on the angiosomalterritories; adequate venous and cavity drainage.

Keywords: breast, oncoplasty, pedicled, perforator

Procedia PDF Downloads 187
1394 A Data Science Pipeline for Algorithmic Trading: A Comparative Study in Applications to Finance and Cryptoeconomics

Authors: Luyao Zhang, Tianyu Wu, Jiayi Li, Carlos-Gustavo Salas-Flores, Saad Lahrichi

Abstract:

Recent advances in AI have made algorithmic trading a central role in finance. However, current research and applications are disconnected information islands. We propose a generally applicable pipeline for designing, programming, and evaluating algorithmic trading of stock and crypto tokens. Moreover, we provide comparative case studies for four conventional algorithms, including moving average crossover, volume-weighted average price, sentiment analysis, and statistical arbitrage. Our study offers a systematic way to program and compare different trading strategies. Moreover, we implement our algorithms by object-oriented programming in Python3, which serves as open-source software for future academic research and applications.

Keywords: algorithmic trading, AI for finance, fintech, machine learning, moving average crossover, volume weighted average price, sentiment analysis, statistical arbitrage, pair trading, object-oriented programming, python3

Procedia PDF Downloads 144
1393 The Change of Urban Land Use/Cover Using Object Based Approach for Southern Bali

Authors: I. Gusti A. A. Rai Asmiwyati, Robert J. Corner, Ashraf M. Dewan

Abstract:

Change on land use/cover (LULC) dominantly affects spatial structure and function. It can have such impacts by disrupting social culture practice and disturbing physical elements. Thus, it has become essential to understand of the dynamics in time and space of LULC as it can be used as a critical input for developing sustainable LULC. This study was an attempt to map and monitor the LULC change in Bali Indonesia from 2003 to 2013. Using object based classification to improve the accuracy, and change detection, multi temporal land use/cover data were extracted from a set of ASTER satellite image. The overall accuracies of the classification maps of 2003 and 2013 were 86.99% and 80.36%, respectively. Built up area and paddy field were the dominant type of land use/cover in both years. Patch increase dominantly in 2003 illustrated the rapid paddy field fragmentation and the huge occurring transformation. This approach is new for the case of diverse urban features of Bali that has been growing fast and increased the classification accuracy than the manual pixel based classification.

Keywords: land use/cover, urban, Bali, ASTER

Procedia PDF Downloads 540
1392 Understanding the Impact of Spatial Light Distribution on Object Identification in Low Vision: A Pilot Psychophysical Study

Authors: Alexandre Faure, Yoko Mizokami, éRic Dinet

Abstract:

These recent years, the potential of light in assisting visually impaired people in their indoor mobility has been demonstrated by different studies. Implementing smart lighting systems for selective visual enhancement, especially designed for low-vision people, is an approach that breaks with the existing visual aids. The appearance of the surface of an object is significantly influenced by the lighting conditions and the constituent materials of the objects. Appearance of objects may appear to be different from expectation. Therefore, lighting conditions lead to an important part of accurate material recognition. The main objective of this work was to investigate the effect of the spatial distribution of light on object identification in the context of low vision. The purpose was to determine whether and what specific lighting approaches should be preferred for visually impaired people. A psychophysical experiment was designed to study the ability of individuals to identify the smallest cube of a pair under different lighting diffusion conditions. Participants were divided into two distinct groups: a reference group of observers with normal or corrected-to-normal visual acuity and a test group, in which observers were required to wear visual impairment simulation glasses. All participants were presented with pairs of cubes in a "miniature room" and were instructed to estimate the relative size of the two cubes. The miniature room replicates real-life settings, adorned with decorations and separated from external light sources by black curtains. The correlated color temperature was set to 6000 K, and the horizontal illuminance at the object level at approximately 240 lux. The objects presented for comparison consisted of 11 white cubes and 11 black cubes of different sizes manufactured with a 3D printer. Participants were seated 60 cm away from the objects. Two different levels of light diffuseness were implemented. After receiving instructions, participants were asked to judge whether the two presented cubes were the same size or if one was smaller. They provided one of five possible answers: "Left one is smaller," "Left one is smaller but unsure," "Same size," "Right one is smaller," or "Right one is smaller but unsure.". The method of constant stimuli was used, presenting stimulus pairs in a random order to prevent learning and expectation biases. Each pair consisted of a comparison stimulus and a reference cube. A psychometric function was constructed to link stimulus value with the frequency of correct detection, aiming to determine the 50% correct detection threshold. Collected data were analyzed through graphs illustrating participants' responses to stimuli, with accuracy increasing as the size difference between cubes grew. Statistical analyses, including 2-way ANOVA tests, showed that light diffuseness had no significant impact on the difference threshold, whereas object color had a significant influence in low vision scenarios. The first results and trends derived from this pilot experiment clearly and strongly suggest that future investigations could explore extreme diffusion conditions to comprehensively assess the impact of diffusion on object identification. For example, the first findings related to light diffuseness may be attributed to the range of manipulation, emphasizing the need to explore how other lighting-related factors interact with diffuseness.

Keywords: Lighting, Low Vision, Visual Aid, Object Identification, Psychophysical Experiment

Procedia PDF Downloads 64
1391 High Resolution Satellite Imagery and Lidar Data for Object-Based Tree Species Classification in Quebec, Canada

Authors: Bilel Chalghaf, Mathieu Varin

Abstract:

Forest characterization in Quebec, Canada, is usually assessed based on photo-interpretation at the stand level. For species identification, this often results in a lack of precision. Very high spatial resolution imagery, such as DigitalGlobe, and Light Detection and Ranging (LiDAR), have the potential to overcome the limitations of aerial imagery. To date, few studies have used that data to map a large number of species at the tree level using machine learning techniques. The main objective of this study is to map 11 individual high tree species ( > 17m) at the tree level using an object-based approach in the broadleaf forest of Kenauk Nature, Quebec. For the individual tree crown segmentation, three canopy-height models (CHMs) from LiDAR data were assessed: 1) the original, 2) a filtered, and 3) a corrected model. The corrected CHM gave the best accuracy and was then coupled with imagery to refine tree species crown identification. When compared with photo-interpretation, 90% of the objects represented a single species. For modeling, 313 variables were derived from 16-band WorldView-3 imagery and LiDAR data, using radiance, reflectance, pixel, and object-based calculation techniques. Variable selection procedures were employed to reduce their number from 313 to 16, using only 11 bands to aid reproducibility. For classification, a global approach using all 11 species was compared to a semi-hierarchical hybrid classification approach at two levels: (1) tree type (broadleaf/conifer) and (2) individual broadleaf (five) and conifer (six) species. Five different model techniques were used: (1) support vector machine (SVM), (2) classification and regression tree (CART), (3) random forest (RF), (4) k-nearest neighbors (k-NN), and (5) linear discriminant analysis (LDA). Each model was tuned separately for all approaches and levels. For the global approach, the best model was the SVM using eight variables (overall accuracy (OA): 80%, Kappa: 0.77). With the semi-hierarchical hybrid approach, at the tree type level, the best model was the k-NN using six variables (OA: 100% and Kappa: 1.00). At the level of identifying broadleaf and conifer species, the best model was the SVM, with OA of 80% and 97% and Kappa values of 0.74 and 0.97, respectively, using seven variables for both models. This paper demonstrates that a hybrid classification approach gives better results and that using 16-band WorldView-3 with LiDAR data leads to more precise predictions for tree segmentation and classification, especially when the number of tree species is large.

Keywords: tree species, object-based, classification, multispectral, machine learning, WorldView-3, LiDAR

Procedia PDF Downloads 133
1390 Towards Expanding the Use of the Online Judge UnitJudge for Java Programming Exercises and Web Development Practices in Computer Science Education

Authors: Iván García-Magariño, Javier Bravo-Agapito, Marta López-Fernández

Abstract:

Online judges have proven their utility in partial auto-evaluation of programming short exercises in the last decades. UnitJudge online judge has the advantage of facilitating the evaluation of separate units to provide more segregate and meaningful feedback to students in complex exercises and practices. This paper discusses the use of UnitUdge in advanced Java object-oriented programming exercises and web development practices. This later usage has been proposed by means of the Selenium Java library and classes to provide the web address. Consequently, UnitJudge is an online judge system that can be applied in several subjects, and therefore, many other students would take advantage of self-testing their exercises. This paper presents the experiments with a Java programming exercise for learning Java object-oriented classes with a generic type. Considering 10 students who voluntarily used UnitJudge, 80% successfully learned this concept, passing the judge exercise with correct results.

Keywords: online judges, programming skills, computer science education, auto-evaluation

Procedia PDF Downloads 100
1389 Relativity in Toddlers' Understanding of the Physical World as Key to Misconceptions in the Science Classroom

Authors: Michael Hast

Abstract:

Within their first year, infants can differentiate between objects based on their weight. By at least 5 years children hold consistent weight-related misconceptions about the physical world, such as that heavy things fall faster than lighter ones because of their weight. Such misconceptions are seen as a challenge for science education since they are often highly resistant to change through instruction. Understanding the time point of emergence of such ideas could, therefore, be crucial for early science pedagogy. The paper thus discusses two studies that jointly address the issue by examining young children’s search behaviour in hidden displacement tasks under consideration of relative object weight. In both studies, they were tested with a heavy or a light ball, and they either had information about one of the balls only or both. In Study 1, 88 toddlers aged 2 to 3½ years watched a ball being dropped into a curved tube and were then allowed to search for the ball in three locations – one straight beneath the tube entrance, one where the curved tube lead to, and one that corresponded to neither of the previous outcomes. Success and failure at the task were not impacted by weight of the balls alone in any particular way. However, from around 3 years onwards, relative lightness, gained through having tactile experience of both balls beforehand, enhanced search success. Conversely, relative heaviness increased search errors such that children increasingly searched in the location immediately beneath the tube entry – known as the gravity bias. In Study 2, 60 toddlers aged 2, 2½ and 3 years watched a ball roll down a ramp and behind a screen with four doors, with a barrier placed along the ramp after one of four doors. Toddlers were allowed to open the doors to find the ball. While search accuracy generally increased with age, relative weight did not play a role in 2-year-olds’ search behaviour. Relative lightness improved 2½-year-olds’ searches. At 3 years, both relative lightness and relative heaviness had a significant impact, with the former improving search accuracy and the latter reducing it. Taken together, both studies suggest that between 2 and 3 years of age, relative object weight is increasingly taken into consideration in navigating naïve physical concepts. In particular, it appears to contribute to the early emergence of misconceptions relating to object weight. This insight from developmental psychology research may have consequences for early science education and related pedagogy towards early conceptual change.

Keywords: conceptual development, early science education, intuitive physics, misconceptions, object weight

Procedia PDF Downloads 190
1388 A Framework for Automated Nuclear Waste Classification

Authors: Seonaid Hume, Gordon Dobie, Graeme West

Abstract:

Detecting and localizing radioactive sources is a necessity for safe and secure decommissioning of nuclear facilities. An important aspect for the management of the sort-and-segregation process is establishing the spatial distributions and quantities of the waste radionuclides, their type, corresponding activity, and ultimately classification for disposal. The data received from surveys directly informs decommissioning plans, on-site incident management strategies, the approach needed for a new cell, as well as protecting the workforce and the public. Manual classification of nuclear waste from a nuclear cell is time-consuming, expensive, and requires significant expertise to make the classification judgment call. Also, in-cell decommissioning is still in its relative infancy, and few techniques are well-developed. As with any repetitive and routine tasks, there is the opportunity to improve the task of classifying nuclear waste using autonomous systems. Hence, this paper proposes a new framework for the automatic classification of nuclear waste. This framework consists of five main stages; 3D spatial mapping and object detection, object classification, radiological mapping, source localisation based on gathered evidence and finally, waste classification. The first stage of the framework, 3D visual mapping, involves object detection from point cloud data. A review of related applications in other industries is provided, and recommendations for approaches for waste classification are made. Object detection focusses initially on cylindrical objects since pipework is significant in nuclear cells and indeed any industrial site. The approach can be extended to other commonly occurring primitives such as spheres and cubes. This is in preparation of stage two, characterizing the point cloud data and estimating the dimensions, material, degradation, and mass of the objects detected in order to feature match them to an inventory of possible items found in that nuclear cell. Many items in nuclear cells are one-offs, have limited or poor drawings available, or have been modified since installation, and have complex interiors, which often and inadvertently pose difficulties when accessing certain zones and identifying waste remotely. Hence, this may require expert input to feature match objects. The third stage, radiological mapping, is similar in order to facilitate the characterization of the nuclear cell in terms of radiation fields, including the type of radiation, activity, and location within the nuclear cell. The fourth stage of the framework takes the visual map for stage 1, the object characterization from stage 2, and radiation map from stage 3 and fuses them together, providing a more detailed scene of the nuclear cell by identifying the location of radioactive materials in three dimensions. The last stage involves combining the evidence from the fused data sets to reveal the classification of the waste in Bq/kg, thus enabling better decision making and monitoring for in-cell decommissioning. The presentation of the framework is supported by representative case study data drawn from an application in decommissioning from a UK nuclear facility. This framework utilises recent advancements of the detection and mapping capabilities of complex radiation fields in three dimensions to make the process of classifying nuclear waste faster, more reliable, cost-effective and safer.

Keywords: nuclear decommissioning, radiation detection, object detection, waste classification

Procedia PDF Downloads 200
1387 Second-Order Complex Systems: Case Studies of Autonomy and Free Will

Authors: Eric Sanchis

Abstract:

Although there does not exist a definitive consensus on a precise definition of a complex system, it is generally considered that a system is complex by nature. The presented work illustrates a different point of view: a system becomes complex only with regard to the question posed to it, i.e., with regard to the problem which has to be solved. A complex system is a couple (question, object). Because the number of questions posed to a given object can be potentially substantial, complexity does not present a uniform face. Two types of complex systems are clearly identified: first-order complex systems and second-order complex systems. First-order complex systems physically exist. They are well-known because they have been studied by the scientific community for a long time. In second-order complex systems, complexity results from the system composition and its articulation that are partially unknown. For some of these systems, there is no evidence of their existence. Vagueness is the keyword characterizing this kind of systems. Autonomy and free will, two mental productions of the human cognitive system, can be identified as second-order complex systems. A classification based on the properties structure makes it possible to discriminate complex properties from the others and to model this kind of second order complex systems. The final outcome is an implementable synthetic property that distinguishes the solid aspects of the actual property from those that are uncertain.

Keywords: autonomy, free will, synthetic property, vaporous complex systems

Procedia PDF Downloads 205
1386 Audio-Visual Co-Data Processing Pipeline

Authors: Rita Chattopadhyay, Vivek Anand Thoutam

Abstract:

Speech is the most acceptable means of communication where we can quickly exchange our feelings and thoughts. Quite often, people can communicate orally but cannot interact or work with computers or devices. It’s easy and quick to give speech commands than typing commands to computers. In the same way, it’s easy listening to audio played from a device than extract output from computers or devices. Especially with Robotics being an emerging market with applications in warehouses, the hospitality industry, consumer electronics, assistive technology, etc., speech-based human-machine interaction is emerging as a lucrative feature for robot manufacturers. Considering this factor, the objective of this paper is to design the “Audio-Visual Co-Data Processing Pipeline.” This pipeline is an integrated version of Automatic speech recognition, a Natural language model for text understanding, object detection, and text-to-speech modules. There are many Deep Learning models for each type of the modules mentioned above, but OpenVINO Model Zoo models are used because the OpenVINO toolkit covers both computer vision and non-computer vision workloads across Intel hardware and maximizes performance, and accelerates application development. A speech command is given as input that has information about target objects to be detected and start and end times to extract the required interval from the video. Speech is converted to text using the Automatic speech recognition QuartzNet model. The summary is extracted from text using a natural language model Generative Pre-Trained Transformer-3 (GPT-3). Based on the summary, essential frames from the video are extracted, and the You Only Look Once (YOLO) object detection model detects You Only Look Once (YOLO) objects on these extracted frames. Frame numbers that have target objects (specified objects in the speech command) are saved as text. Finally, this text (frame numbers) is converted to speech using text to speech model and will be played from the device. This project is developed for 80 You Only Look Once (YOLO) labels, and the user can extract frames based on only one or two target labels. This pipeline can be extended for more than two target labels easily by making appropriate changes in the object detection module. This project is developed for four different speech command formats by including sample examples in the prompt used by Generative Pre-Trained Transformer-3 (GPT-3) model. Based on user preference, one can come up with a new speech command format by including some examples of the respective format in the prompt used by the Generative Pre-Trained Transformer-3 (GPT-3) model. This pipeline can be used in many projects like human-machine interface, human-robot interaction, and surveillance through speech commands. All object detection projects can be upgraded using this pipeline so that one can give speech commands and output is played from the device.

Keywords: OpenVINO, automatic speech recognition, natural language processing, object detection, text to speech

Procedia PDF Downloads 80
1385 Structural Invertibility and Optimal Sensor Node Placement for Error and Input Reconstruction in Dynamic Systems

Authors: Maik Kschischo, Dominik Kahl, Philipp Wendland, Andreas Weber

Abstract:

Understanding and modelling of real-world complex dynamic systems in biology, engineering and other fields is often made difficult by incomplete knowledge about the interactions between systems states and by unknown disturbances to the system. In fact, most real-world dynamic networks are open systems receiving unknown inputs from their environment. To understand a system and to estimate the state dynamics, these inputs need to be reconstructed from output measurements. Reconstructing the input of a dynamic system from its measured outputs is an ill-posed problem if only a limited number of states is directly measurable. A first requirement for solving this problem is the invertibility of the input-output map. In our work, we exploit the fact that invertibility of a dynamic system is a structural property, which depends only on the network topology. Therefore, it is possible to check for invertibility using a structural invertibility algorithm which counts the number of node disjoint paths linking inputs and outputs. The algorithm is efficient enough, even for large networks up to a million nodes. To understand structural features influencing the invertibility of a complex dynamic network, we analyze synthetic and real networks using the structural invertibility algorithm. We find that invertibility largely depends on the degree distribution and that dense random networks are easier to invert than sparse inhomogeneous networks. We show that real networks are often very difficult to invert unless the sensor nodes are carefully chosen. To overcome this problem, we present a sensor node placement algorithm to achieve invertibility with a minimum set of measured states. This greedy algorithm is very fast and also guaranteed to find an optimal sensor node-set if it exists. Our results provide a practical approach to experimental design for open, dynamic systems. Since invertibility is a necessary condition for unknown input observers and data assimilation filters to work, it can be used as a preprocessing step to check, whether these input reconstruction algorithms can be successful. If not, we can suggest additional measurements providing sufficient information for input reconstruction. Invertibility is also important for systems design and model building. Dynamic models are always incomplete, and synthetic systems act in an environment, where they receive inputs or even attack signals from their exterior. Being able to monitor these inputs is an important design requirement, which can be achieved by our algorithms for invertibility analysis and sensor node placement.

Keywords: data-driven dynamic systems, inversion of dynamic systems, observability, experimental design, sensor node placement

Procedia PDF Downloads 150