Search results for: demonstration wildfire detection and action from space
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9576

Search results for: demonstration wildfire detection and action from space

8796 Zonal and Sequential Extraction Design for Large Flat Space to Achieve Perpetual Tenability

Authors: Mingjun Xu, Man Pun Wan

Abstract:

This study proposed an effective smoke control strategy for the large flat space with a low ceiling to achieve the requirement of perpetual tenability. For the large flat space with a low ceiling, the depth of the smoke reservoir is very shallow, and it is difficult to perpetually constrain the smoke within a limited space. A series of numerical tests were conducted to determine the smoke strategy. A zonal design i.e., the fire zone and two adjacent zones was proposed and validated to be effective in controlling smoke. Once a fire happens in a compartment space, the Engineered Smoke Control (ESC) system will be activated in three zones i.e., the fire zone, in which the fire happened, and two adjacent zones. The smoke can be perpetually constrained within the three smoke zones. To further improve the extraction efficiency, sequential activation of the ESC system within the 3 zones turned out to be more efficient than simultaneous activation. Additionally, the proposed zonal and sequential extraction design can reduce the mechanical extraction flow rate by up to 40.7 % as compared to the conventional method, which is much more economical than that of the conventional method.

Keywords: performance-based design, perpetual tenability, smoke control, fire plume

Procedia PDF Downloads 74
8795 Enhancement Method of Network Traffic Anomaly Detection Model Based on Adversarial Training With Category Tags

Authors: Zhang Shuqi, Liu Dan

Abstract:

For the problems in intelligent network anomaly traffic detection models, such as low detection accuracy caused by the lack of training samples, poor effect with small sample attack detection, a classification model enhancement method, F-ACGAN(Flow Auxiliary Classifier Generative Adversarial Network) which introduces generative adversarial network and adversarial training, is proposed to solve these problems. Generating adversarial data with category labels could enhance the training effect and improve classification accuracy and model robustness. FACGAN consists of three steps: feature preprocess, which includes data type conversion, dimensionality reduction and normalization, etc.; A generative adversarial network model with feature learning ability is designed, and the sample generation effect of the model is improved through adversarial iterations between generator and discriminator. The adversarial disturbance factor of the gradient direction of the classification model is added to improve the diversity and antagonism of generated data and to promote the model to learn from adversarial classification features. The experiment of constructing a classification model with the UNSW-NB15 dataset shows that with the enhancement of FACGAN on the basic model, the classification accuracy has improved by 8.09%, and the score of F1 has improved by 6.94%.

Keywords: data imbalance, GAN, ACGAN, anomaly detection, adversarial training, data augmentation

Procedia PDF Downloads 106
8794 A Machine Learning Approach for Detecting and Locating Hardware Trojans

Authors: Kaiwen Zheng, Wanting Zhou, Nan Tang, Lei Li, Yuanhang He

Abstract:

The integrated circuit industry has become a cornerstone of the information society, finding widespread application in areas such as industry, communication, medicine, and aerospace. However, with the increasing complexity of integrated circuits, Hardware Trojans (HTs) implanted by attackers have become a significant threat to their security. In this paper, we proposed a hardware trojan detection method for large-scale circuits. As HTs introduce physical characteristic changes such as structure, area, and power consumption as additional redundant circuits, we proposed a machine-learning-based hardware trojan detection method based on the physical characteristics of gate-level netlists. This method transforms the hardware trojan detection problem into a machine-learning binary classification problem based on physical characteristics, greatly improving detection speed. To address the problem of imbalanced data, where the number of pure circuit samples is far less than that of HTs circuit samples, we used the SMOTETomek algorithm to expand the dataset and further improve the performance of the classifier. We used three machine learning algorithms, K-Nearest Neighbors, Random Forest, and Support Vector Machine, to train and validate benchmark circuits on Trust-Hub, and all achieved good results. In our case studies based on AES encryption circuits provided by trust-hub, the test results showed the effectiveness of the proposed method. To further validate the method’s effectiveness for detecting variant HTs, we designed variant HTs using open-source HTs. The proposed method can guarantee robust detection accuracy in the millisecond level detection time for IC, and FPGA design flows and has good detection performance for library variant HTs.

Keywords: hardware trojans, physical properties, machine learning, hardware security

Procedia PDF Downloads 148
8793 Analytical Modeling of Drain Current for DNA Biomolecule Detection in Double-Gate Tunnel Field-Effect Transistor Biosensor

Authors: Ashwani Kumar

Abstract:

Abstract- This study presents an analytical modeling approach for analyzing the drain current behavior in Tunnel Field-Effect Transistor (TFET) biosensors used for the detection of DNA biomolecules. The proposed model focuses on elucidating the relationship between the drain current and the presence of DNA biomolecules, taking into account the impact of various device parameters and biomolecule characteristics. Through comprehensive analysis, the model offers insights into the underlying mechanisms governing the sensing performance of TFET biosensors, aiding in the optimization of device design and operation. A non-local tunneling model is incorporated with other essential models to accurately trace the simulation and modeled data. An experimental validation of the model is provided, demonstrating its efficacy in accurately predicting the drain current response to DNA biomolecule detection. The sensitivity attained from the analytical model is compared and contrasted with the ongoing research work in this area.

Keywords: biosensor, double-gate TFET, DNA detection, drain current modeling, sensitivity

Procedia PDF Downloads 58
8792 Labview-Based System for Fiber Links Events Detection

Authors: Bo Liu, Qingshan Kong, Weiqing Huang

Abstract:

With the rapid development of modern communication, diagnosing the fiber-optic quality and faults in real-time is widely focused. In this paper, a Labview-based system is proposed for fiber-optic faults detection. The wavelet threshold denoising method combined with Empirical Mode Decomposition (EMD) is applied to denoise the optical time domain reflectometer (OTDR) signal. Then the method based on Gabor representation is used to detect events. Experimental measurements show that signal to noise ratio (SNR) of the OTDR signal is improved by 1.34dB on average, compared with using the wavelet threshold denosing method. The proposed system has a high score in event detection capability and accuracy. The maximum detectable fiber length of the proposed Labview-based system can be 65km.

Keywords: empirical mode decomposition, events detection, Gabor transform, optical time domain reflectometer, wavelet threshold denoising

Procedia PDF Downloads 123
8791 Analyzing Extended Reality Technologies for Human Space Exploration

Authors: Morgan Kuligowski, Marientina Gotsis

Abstract:

Extended reality (XR) technologies share an intertwined history with spaceflight and innovation. New advancements in XR technologies offer expanding possibilities to advance the future of human space exploration with increased crew autonomy. This paper seeks to identify implementation gaps between existing and proposed XR space applications to inform future mission planning. A review of virtual reality, augmented reality, and mixed reality technologies implemented aboard the International Space Station revealed a total of 16 flown investigations. A secondary set of ground-tested XR human spaceflight applications were systematically retrieved from literature sources. The two sets of XR technologies, those flown and those existing in the literature were analyzed to characterize application domains and device types. Comparisons between these groups revealed untapped application areas for XR to support crew psychological health, in-flight training, and extravehicular operations on future flights. To fill these roles, integrating XR technologies with advancements in biometric sensors and machine learning tools is expected to transform crew capabilities.

Keywords: augmented reality, extended reality, international space station, mixed reality, virtual reality

Procedia PDF Downloads 216
8790 Indicator-Immobilized, Cellulose Based Optical Sensing Membrane for the Detection of Heavy Metal Ions

Authors: Nisha Dhariwal, Anupama Sharma

Abstract:

The synthesis of cellulose nanofibrils quaternized with 3‐chloro‐2‐hydroxypropyltrimethylammonium chloride (CHPTAC) in NaOH/urea aqueous solution has been reported. Xylenol Orange (XO) has been used as an indicator for selective detection of Sn (II) ions, by its immobilization on quaternized cellulose membrane. The effects of pH, reagent concentration and reaction time on the immobilization of XO have also been studied. The linear response, limit of detection, and interference of other metal ions have also been studied and no significant interference has been observed. The optical chemical sensor displayed good durability and short response time with negligible leaching of the reagent.

Keywords: cellulose, chemical sensor, heavy metal ions, indicator immobilization

Procedia PDF Downloads 301
8789 Impact of the Action Antropic in the Desertification of Steppe in Algeria

Authors: Kadi-Hanifi Halima

Abstract:

Stipa tenacissima is a plant with a big ecological value (against desertification) and economical stake (paper industry). It is important by its pastoral value due to the inflorescence. It occupied large areas between the Tellian atlas and the Saharian atlas, at the present, these areas of alfa have regressed a lot. This regression is estimated at 1% per year. The principal cause is a human responsibility. The drought is just an aggravating circumstance. The eradication of such a kind of species will have serious consequences upon the equilibrium of all the steppic ecosystem. Thus, we have thought necessary and urgent to know the alfa ecosystem, under all its aspects (climatic, floristic, and edaphic), this diagnostic could direct the fight actions against desertification

Keywords: desertification, anthropic action, soils, Stipa tenacissima

Procedia PDF Downloads 313
8788 Surface Hole Defect Detection of Rolled Sheets Based on Pixel Classification Approach

Authors: Samira Taleb, Sakina Aoun, Slimane Ziani, Zoheir Mentouri, Adel Boudiaf

Abstract:

Rolling is a pressure treatment technique that modifies the shape of steel ingots or billets between rotating rollers. During this process, defects may form on the surface of the rolled sheets and are likely to affect the performance and quality of the finished product. In our study, we developed a method for detecting surface hole defects using a pixel classification approach. This work includes several steps. First, we performed image preprocessing to delimit areas with and without hole defects on the sheet image. Then, we developed the histograms of each area to generate the gray level membership intervals of the pixels that characterize each area. As we noticed an intersection between the characteristics of the gray level intervals of the images of the two areas, we finally performed a learning step based on a series of detection tests to refine the membership intervals of each area, and to choose the defect detection criterion in order to optimize the recognition of the surface hole.

Keywords: classification, defect, surface, detection, hole

Procedia PDF Downloads 24
8787 Comparison of Leeway Space Predictions Using Moyers and Tanaka-Johnston Upper Jaw and Lower Jaw on Batak Tribe Between Male and Female in Elementary School Students in Medan City, Sumatera Utara, Indonesia: A Cross-sectional Study

Authors: Hilda Fitria Lubis, Erna Sulistyawati

Abstract:

Objective: The study aims to compare Leeway space averages between Moyers and Tanaka-Johnston's analysis of elementary school students from the Batak tribe in Medan City. Material and Methods: The study involved 106 students from the Batak tribe elementary school in Medan, comprising 53 male and 53 female students. The samples obtained were then printed on both jaws to obtain a working model, and the mesiodistal width of the four permanent biting teeth of the lower jaw and the amount of space available on the canine-premolar region, as well as the predicted mesiodistal number of the canine-premolar on the Moyers probability table with a 75% degree of confidence and the Tanaka-Johnston formula. Results: Using Moyers analysis, students at Batak Elementary School in Medan City have an average Leeway space value of 2 mm on the upper jaw and 2.78 mm on the lower jaw. The average Leeway spatial value using Tanaka-Johnston analysis in the Batak tribe in elementary school in Medan City is 1.33 mm on the top jaw and 2.39 mm on the bottom jaw. Conclusion: According to Moyers and Tanaka-Johnsnton's analysis of both the upper and lower jaws in elementary school students of the Batak tribe in Medan City, there is a significant difference between Leeway's average space.

Keywords: leeways space, batak tribe, genders, diagnosis

Procedia PDF Downloads 35
8786 Minimizing the Impact of Covariate Detection Limit in Logistic Regression

Authors: Shahadut Hossain, Jacek Wesolowski, Zahirul Hoque

Abstract:

In many epidemiological and environmental studies covariate measurements are subject to the detection limit. In most applications, covariate measurements are usually truncated from below which is known as left-truncation. Because the measuring device, which we use to measure the covariate, fails to detect values falling below the certain threshold. In regression analyses, it causes inflated bias and inaccurate mean squared error (MSE) to the estimators. This paper suggests a response-based regression calibration method to correct the deleterious impact introduced by the covariate detection limit in the estimators of the parameters of simple logistic regression model. Compared to the maximum likelihood method, the proposed method is computationally simpler, and hence easier to implement. It is robust to the violation of distributional assumption about the covariate of interest. In producing correct inference, the performance of the proposed method compared to the other competing methods has been investigated through extensive simulations. A real-life application of the method is also shown using data from a population-based case-control study of non-Hodgkin lymphoma.

Keywords: environmental exposure, detection limit, left truncation, bias, ad-hoc substitution

Procedia PDF Downloads 238
8785 Hybrid Anomaly Detection Using Decision Tree and Support Vector Machine

Authors: Elham Serkani, Hossein Gharaee Garakani, Naser Mohammadzadeh, Elaheh Vaezpour

Abstract:

Intrusion detection systems (IDS) are the main components of network security. These systems analyze the network events for intrusion detection. The design of an IDS is through the training of normal traffic data or attack. The methods of machine learning are the best ways to design IDSs. In the method presented in this article, the pruning algorithm of C5.0 decision tree is being used to reduce the features of traffic data used and training IDS by the least square vector algorithm (LS-SVM). Then, the remaining features are arranged according to the predictor importance criterion. The least important features are eliminated in the order. The remaining features of this stage, which have created the highest level of accuracy in LS-SVM, are selected as the final features. The features obtained, compared to other similar articles which have examined the selected features in the least squared support vector machine model, are better in the accuracy, true positive rate, and false positive. The results are tested by the UNSW-NB15 dataset.

Keywords: decision tree, feature selection, intrusion detection system, support vector machine

Procedia PDF Downloads 266
8784 The Relationship between Representational Conflicts, Generalization, and Encoding Requirements in an Instance Memory Network

Authors: Mathew Wakefield, Matthew Mitchell, Lisa Wise, Christopher McCarthy

Abstract:

The properties of memory representations in artificial neural networks have cognitive implications. Distributed representations that encode instances as a pattern of activity across layers of nodes afford memory compression and enforce the selection of a single point in instance space. These encoding schemes also appear to distort the representational space, as well as trading off the ability to validate that input information is within the bounds of past experience. In contrast, a localist representation which encodes some meaningful information into individual nodes in a network layer affords less memory compression while retaining the integrity of the representational space. This allows the validity of an input to be determined. The validity (or familiarity) of input along with the capacity of localist representation for multiple instance selections affords a memory sampling approach that dynamically balances the bias-variance trade-off. When the input is familiar, bias may be high by referring only to the most similar instances in memory. When the input is less familiar, variance can be increased by referring to more instances that capture a broader range of features. Using this approach in a localist instance memory network, an experiment demonstrates a relationship between representational conflict, generalization performance, and memorization demand. Relatively small sampling ranges produce the best performance on a classic machine learning dataset of visual objects. Combining memory validity with conflict detection produces a reliable confidence judgement that can separate responses with high and low error rates. Confidence can also be used to signal the need for supervisory input. Using this judgement, the need for supervised learning as well as memory encoding can be substantially reduced with only a trivial detriment to classification performance.

Keywords: artificial neural networks, representation, memory, conflict monitoring, confidence

Procedia PDF Downloads 129
8783 Developing an Accurate AI Algorithm for Histopathologic Cancer Detection

Authors: Leah Ning

Abstract:

This paper discusses the development of a machine learning algorithm that accurately detects metastatic breast cancer (cancer has spread elsewhere from its origin part) in selected images that come from pathology scans of lymph node sections. Being able to develop an accurate artificial intelligence (AI) algorithm would help significantly in breast cancer diagnosis since manual examination of lymph node scans is both tedious and oftentimes highly subjective. The usage of AI in the diagnosis process provides a much more straightforward, reliable, and efficient method for medical professionals and would enable faster diagnosis and, therefore, more immediate treatment. The overall approach used was to train a convolution neural network (CNN) based on a set of pathology scan data and use the trained model to binarily classify if a new scan were benign or malignant, outputting a 0 or a 1, respectively. The final model’s prediction accuracy is very high, with 100% for the train set and over 70% for the test set. Being able to have such high accuracy using an AI model is monumental in regard to medical pathology and cancer detection. Having AI as a new tool capable of quick detection will significantly help medical professionals and patients suffering from cancer.

Keywords: breast cancer detection, AI, machine learning, algorithm

Procedia PDF Downloads 92
8782 Coding and Decoding versus Space Diversity for ‎Rayleigh Fading Radio Frequency Channels ‎

Authors: Ahmed Mahmoud Ahmed Abouelmagd

Abstract:

The diversity is the usual remedy of the transmitted signal level variations (Fading phenomena) in radio frequency channels. Diversity techniques utilize two or more copies of a signal and combine those signals to combat fading. The basic concept of diversity is to transmit the signal via several independent diversity branches to get independent signal replicas via time – frequency - space - and polarization diversity domains. Coding and decoding processes can be an alternative remedy for fading phenomena, it cannot increase the channel capacity, but it can improve the error performance. In this paper we propose the use of replication decoding with BCH code class, and Viterbi decoding algorithm with convolution coding; as examples of coding and decoding processes. The results are compared to those obtained from two optimized selection space diversity techniques. The performance of Rayleigh fading channel, as the model considered for radio frequency channels, is evaluated for each case. The evaluation results show that the coding and decoding approaches, especially the BCH coding approach with replication decoding scheme, give better performance compared to that of selection space diversity optimization approaches. Also, an approach for combining the coding and decoding diversity as well as the space diversity is considered, the main disadvantage of this approach is its complexity but it yields good performance results.

Keywords: Rayleigh fading, diversity, BCH codes, Replication decoding, ‎convolution coding, viterbi decoding, space diversity

Procedia PDF Downloads 443
8781 Collision Detection Algorithm Based on Data Parallelism

Authors: Zhen Peng, Baifeng Wu

Abstract:

Modern computing technology enters the era of parallel computing with the trend of sustainable and scalable parallelism. Single Instruction Multiple Data (SIMD) is an important way to go along with the trend. It is able to gather more and more computing ability by increasing the number of processor cores without the need of modifying the program. Meanwhile, in the field of scientific computing and engineering design, many computation intensive applications are facing the challenge of increasingly large amount of data. Data parallel computing will be an important way to further improve the performance of these applications. In this paper, we take the accurate collision detection in building information modeling as an example. We demonstrate a model for constructing a data parallel algorithm. According to the model, a complex object is decomposed into the sets of simple objects; collision detection among complex objects is converted into those among simple objects. The resulting algorithm is a typical SIMD algorithm, and its advantages in parallelism and scalability is unparalleled in respect to the traditional algorithms.

Keywords: data parallelism, collision detection, single instruction multiple data, building information modeling, continuous scalability

Procedia PDF Downloads 291
8780 Design and Development of Automatic Onion Harvester

Authors: P. Revathi, T. Mrunalini, K. Padma Priya, P. Ramya, R. Saranya

Abstract:

During the tough times of covid, those people who were hospitalized found it difficult to always convey what they wanted to or needed to the attendee. Sometimes the attendees might also not be there. In that case, the patients can use simple hand gestures to control electrical appliances (like its set it for a zero watts bulb)and three other gestures for voice note intimation. In this AI-based hand recognition project, NodeMCU is used for the control action of the relay, and it is connected to the firebase for storing the value in the cloud and is interfaced with the python code via raspberry pi. For three hand gestures, a voice clip is added for intimation to the attendee. This is done with the help of Google’s text to speech and the inbuilt audio file option in the raspberry pi 4. All the 5 gestures will be detected when shown with their hands via a webcam which is placed for gesture detection. A personal computer is used for displaying the gestures and for running the code in the raspberry pi imager.

Keywords: onion harvesting, automatic pluging, camera, raspberry pi

Procedia PDF Downloads 199
8779 Challenges in the Material and Action-Resistance Factor Design for Embedded Retaining Wall Limit State Analysis

Authors: Kreso Ivandic, Filip Dodigovic, Damir Stuhec

Abstract:

The paper deals with the proposed 'Material' and 'Action-resistance factor' design methods in designing the embedded retaining walls. The parametric analysis of evaluating the differences of the output values mutually and compared with classic approach computation was performed. There is a challenge with the criteria for choosing the proposed calculation design methods in Eurocode 7 with respect to current technical regulations and regular engineering practice. The basic criterion for applying a particular design method is to ensure minimum an equal degree of reliability in relation to the current practice. The procedure of combining the relevant partial coefficients according to design methods was carried out. The use of mentioned partial coefficients should result in the same level of safety, regardless of load combinations, material characteristics and problem geometry. This proposed approach of the partial coefficients related to the material and/or action-resistance should aimed at building a bridge between calculations used so far and pure probability analysis. The measure to compare the results was to determine an equivalent safety factor for each analysis. The results show a visible wide span of equivalent values of the classic safety factors.

Keywords: action-resistance factor design, classic approach, embedded retaining wall, Eurocode 7, limit states, material factor design

Procedia PDF Downloads 231
8778 The Effect of Object Presentation on Action Memory in School-Aged Children

Authors: Farzaneh Badinlou, Reza Kormi-Nouri, Monika Knopf

Abstract:

Enacted tasks are typically remembered better than when the same task materials are only verbally encoded, a robust finding referred to as the enactment effect. It has been assumed that enactment effect is independent of object presence but the size of enactment effect can be increased by providing objects at study phase in adults. To clarify the issues in children, free recall and cued recall performance of action phrases with or without using real objects were compared in 410 school-aged children from four age groups (8, 10, 12 and 14 years old). In this study, subjects were instructed to learn a series of action phrases under three encoding conditions, participants listened to verbal action phrases (VTs), performed the phrases (SPTs: subject-performed tasks), and observed the experimenter perform the phrases (EPTs: experimenter-performed tasks). Then, free recall and cued recall memory tests were administrated. The results revealed that the real object compared with imaginary objects improved recall performance in SPTs and EPTs, but more so in VTs. It was also found that the object presence was not necessary for the occurrence of the enactment effect but it was changed the size of enactment effect in all age groups. The size of enactment effect was more pronounced for imaginary objects than the real object in both free recall and cued recall memory tests in children. It was discussed that SPTs and EPTs deferentially facilitate item-specific and relation information processing and providing the objects can moderate the processing underlying the encoding conditions.

Keywords: action memory, enactment effect, item-specific processing, object, relational processing, school-aged children

Procedia PDF Downloads 240
8777 Optimizing Machine Learning Through Python Based Image Processing Techniques

Authors: Srinidhi. A, Naveed Ahmed, Twinkle Hareendran, Vriksha Prakash

Abstract:

This work reviews some of the advanced image processing techniques for deep learning applications. Object detection by template matching, image denoising, edge detection, and super-resolution modelling are but a few of the tasks. The paper looks in into great detail, given that such tasks are crucial preprocessing steps that increase the quality and usability of image datasets in subsequent deep learning tasks. We review some of the methods for the assessment of image quality, more specifically sharpness, which is crucial to ensure a robust performance of models. Further, we will discuss the development of deep learning models specific to facial emotion detection, age classification, and gender classification, which essentially includes the preprocessing techniques interrelated with model performance. Conclusions from this study pinpoint the best practices in the preparation of image datasets, targeting the best trade-off between computational efficiency and retaining important image features critical for effective training of deep learning models.

Keywords: image processing, machine learning applications, template matching, emotion detection

Procedia PDF Downloads 20
8776 Self-Organizing Maps for Credit Card Fraud Detection

Authors: ChunYi Peng, Wei Hsuan CHeng, Shyh Kuang Ueng

Abstract:

This study focuses on the application of self-organizing maps (SOM) technology in analyzing credit card transaction data, aiming to enhance the accuracy and efficiency of fraud detection. Som, as an artificial neural network, is particularly suited for pattern recognition and data classification, making it highly effective for the complex and variable nature of credit card transaction data. By analyzing transaction characteristics with SOM, the research identifies abnormal transaction patterns that could indicate potentially fraudulent activities. Moreover, this study has developed a specialized visualization tool to intuitively present the relationships between SOM analysis outcomes and transaction data, aiding financial institution personnel in quickly identifying and responding to potential fraud, thereby reducing financial losses. Additionally, the research explores the integration of SOM technology with composite intelligent system technologies (including finite state machines, fuzzy logic, and decision trees) to further improve fraud detection accuracy. This multimodal approach provides a comprehensive perspective for identifying and understanding various types of fraud within credit card transactions. In summary, by integrating SOM technology with visualization tools and composite intelligent system technologies, this research offers a more effective method of fraud detection for the financial industry, not only enhancing detection accuracy but also deepening the overall understanding of fraudulent activities.

Keywords: self-organizing map technology, fraud detection, information visualization, data analysis, composite intelligent system technologies, decision support technologies

Procedia PDF Downloads 60
8775 On the Representation of Actuator Faults Diagnosis and Systems Invertibility

Authors: F. Sallem, B. Dahhou, A. Kamoun

Abstract:

In this work, the main problem considered is the detection and the isolation of the actuator fault. A new formulation of the linear system is generated to obtain the conditions of the actuator fault diagnosis. The proposed method is based on the representation of the actuator as a subsystem connected with the process system in cascade manner. The designed formulation is generated to obtain the conditions of the actuator fault detection and isolation. Detectability conditions are expressed in terms of the invertibility notions. An example and a comparative analysis with the classic formulation illustrate the performances of such approach for simple actuator fault diagnosis by using the linear model of nuclear reactor.

Keywords: actuator fault, Fault detection, left invertibility, nuclear reactor, observability, parameter intervals, system inversion

Procedia PDF Downloads 406
8774 A Procedure for Post-Earthquake Damage Estimation Based on Detection of High-Frequency Transients

Authors: Aleksandar Zhelyazkov, Daniele Zonta, Helmut Wenzel, Peter Furtner

Abstract:

In the current research structural health monitoring is considered for addressing the critical issue of post-earthquake damage detection. A non-standard approach for damage detection via acoustic emission is presented - acoustic emissions are monitored in the low frequency range (up to 120 Hz). Such emissions are termed high-frequency transients. Further a damage indicator defined as the Time-Ratio Damage Indicator is introduced. The indicator relies on time-instance measurements of damage initiation and deformation peaks. Based on the time-instance measurements a procedure for estimation of the maximum drift ratio is proposed. Monitoring data is used from a shaking-table test of a full-scale reinforced concrete bridge pier. Damage of the experimental column is successfully detected and the proposed damage indicator is calculated.

Keywords: acoustic emission, damage detection, shaking table test, structural health monitoring

Procedia PDF Downloads 233
8773 Influence and Depiction of Power in an Urban Space

Authors: Kalpeshkumar Patel, Nikita Manvi

Abstract:

The paper is an attempt to understand the influence and depiction of power in an urban space by throwing light across a few examples across the architectural timeline. Power has been the medium through which ideologies function, as witnessed across the timeline. The center to understand this ideology is to apprehend how power is formed, captured, owned, traded, and distorted. Every urban space has power embedded in it, either for the people who are imposing it or for the public who are receiving it. The most fundamental question in the issue of power is who – who will judge, whose tastes will matter and whose interests are being served. Power is expressed and reinforced by regular means, a boundary and gates, a parade route, a dominant landmark, play of shape or scale in elevation, ceremonial axis, boulevards and avenues, the vista, bilateral symmetry, or regular order. Even if people accept the psychological efficacy of these forms, the way they perceive them may vary depending on the subject. They are cold devices of power used to make some people submit to others. Yet it is also true that these symbolic forms are attractive because they speak to the deep emotions of people. They do indeed give us a sense of security, stability and continuity, awe and pride. The Urban Space for mass assembly is an idea that continues to seduce dictators and democracies. It is a tradition as old as an agora and as manipulative as Baroque Rome.

Keywords: urban space, aggrandization, city planning, landscape, supremacy, democratic

Procedia PDF Downloads 130
8772 Quantum Localization of Vibrational Mirror in Cavity Optomechanics

Authors: Madiha Tariq, Hena Rabbani

Abstract:

Recently, cavity-optomechanics becomes an extensive research field that has manipulated the mechanical effects of light for coupling of the optical field with other physical objects specifically with regards to dynamical localization. We investigate the dynamical localization (both in momentum and position space) for a vibrational mirror in a Fabry-Pérot cavity driven by a single mode optical field and a transverse probe field. The weak probe field phenomenon results in classical chaos in phase space and spatio temporal dynamics in position |ψ(x)²| and momentum space |ψ(p)²| versus time show quantum localization in both momentum and position space. Also, we discuss the parametric dependencies of dynamical localization for a designated set of parameters to be experimentally feasible. Our work opens an avenue to manipulate the other optical phenomena and applicability of proposed work can be prolonged to turn-able laser sources in the future.

Keywords: dynamical localization, cavity optomechanics, Hamiltonian chaos, probe field

Procedia PDF Downloads 150
8771 Effect of SCN5A Gene Mutation in Endocardial Cell

Authors: Helan Satish, M. Ramasubba Reddy

Abstract:

The simulation of an endocardial cell for gene mutation in the cardiac sodium ion channel NaV1.5, encoded by SCN5A gene, is discussed. The characterization of Brugada Syndrome by loss of function effect on SCN5A mutation due to L812Q mutant present in the DII-S4 transmembrane region of the NaV1.5 channel protein and its effect in an endocardial cell is studied. Ten Tusscher model of human ventricular action potential is modified to incorporate the changes contributed by L812Q mutant in the endocardial cells. Results show that BrS-associated SCN5A mutation causes reduction in the inward sodium current by modifications in the channel gating dynamics such as delayed activation, enhanced inactivation, and slowed recovery from inactivation in the endocardial cell. A decrease in the inward sodium current was also observed, which affects depolarization phase (Phase 0) that leads to reduction in the spike amplitude of the cardiac action potential.

Keywords: SCN5A gene mutation, sodium channel, Brugada syndrome, cardiac arrhythmia, action potential

Procedia PDF Downloads 126
8770 A Generalized Framework for Adaptive Machine Learning Deployments in Algorithmic Trading

Authors: Robert Caulk

Abstract:

A generalized framework for adaptive machine learning deployments in algorithmic trading is introduced, tested, and released as open-source code. The presented software aims to test the hypothesis that recent data contains enough information to form a probabilistically favorable short-term price prediction. Further, the framework contains various adaptive machine learning techniques that are geared toward generating profit during strong trends and minimizing losses during trend changes. Results demonstrate that this adaptive machine learning approach is capable of capturing trends and generating profit. The presentation also discusses the importance of defining the parameter space associated with the dynamic training data-set and using the parameter space to identify and remove outliers from prediction data points. Meanwhile, the generalized architecture enables common users to exploit the powerful machinery while focusing on high-level feature engineering and model testing. The presentation also highlights common strengths and weaknesses associated with the presented technique and presents a broad range of well-tested starting points for feature set construction, target setting, and statistical methods for enforcing risk management and maintaining probabilistically favorable entry and exit points. The presentation also describes the end-to-end data processing tools associated with FreqAI, including automatic data fetching, data aggregation, feature engineering, safe and robust data pre-processing, outlier detection, custom machine learning and statistical tools, data post-processing, and adaptive training backtest emulation, and deployment of adaptive training in live environments. Finally, the generalized user interface is also discussed in the presentation. Feature engineering is simplified so that users can seed their feature sets with common indicator libraries (e.g. TA-lib, pandas-ta). The user also feeds data expansion parameters to fill out a large feature set for the model, which can contain as many as 10,000+ features. The presentation describes the various object-oriented programming techniques employed to make FreqAI agnostic to third-party libraries and external data sources. In other words, the back-end is constructed in such a way that users can leverage a broad range of common regression libraries (Catboost, LightGBM, Sklearn, etc) as well as common Neural Network libraries (TensorFlow, PyTorch) without worrying about the logistical complexities associated with data handling and API interactions. The presentation finishes by drawing conclusions about the most important parameters associated with a live deployment of the adaptive learning framework and provides the road map for future development in FreqAI.

Keywords: machine learning, market trend detection, open-source, adaptive learning, parameter space exploration

Procedia PDF Downloads 89
8769 Self-Organizing Maps for Credit Card Fraud Detection and Visualization

Authors: Peng Chun-Yi, Chen Wei-Hsuan, Ueng Shyh-Kuang

Abstract:

This study focuses on the application of self-organizing maps (SOM) technology in analyzing credit card transaction data, aiming to enhance the accuracy and efficiency of fraud detection. Som, as an artificial neural network, is particularly suited for pattern recognition and data classification, making it highly effective for the complex and variable nature of credit card transaction data. By analyzing transaction characteristics with SOM, the research identifies abnormal transaction patterns that could indicate potentially fraudulent activities. Moreover, this study has developed a specialized visualization tool to intuitively present the relationships between SOM analysis outcomes and transaction data, aiding financial institution personnel in quickly identifying and responding to potential fraud, thereby reducing financial losses. Additionally, the research explores the integration of SOM technology with composite intelligent system technologies (including finite state machines, fuzzy logic, and decision trees) to further improve fraud detection accuracy. This multimodal approach provides a comprehensive perspective for identifying and understanding various types of fraud within credit card transactions. In summary, by integrating SOM technology with visualization tools and composite intelligent system technologies, this research offers a more effective method of fraud detection for the financial industry, not only enhancing detection accuracy but also deepening the overall understanding of fraudulent activities.

Keywords: self-organizing map technology, fraud detection, information visualization, data analysis, composite intelligent system technologies, decision support technologies

Procedia PDF Downloads 60
8768 The Use of Artificial Intelligence in the Context of a Space Traffic Management System: Legal Aspects

Authors: George Kyriakopoulos, Photini Pazartzis, Anthi Koskina, Crystalie Bourcha

Abstract:

The need for securing safe access to and return from outer space, as well as ensuring the viability of outer space operations, maintains vivid the debate over the promotion of organization of space traffic through a Space Traffic Management System (STM). The proliferation of outer space activities in recent years as well as the dynamic emergence of the private sector has gradually resulted in a diverse universe of actors operating in outer space. The said developments created an increased adverse impact on outer space sustainability as the case of the growing number of space debris clearly demonstrates. The above landscape sustains considerable threats to outer space environment and its operators that need to be addressed by a combination of scientific-technological measures and regulatory interventions. In this context, recourse to recent technological advancements and, in particular, to Artificial Intelligence (AI) and machine learning systems, could achieve exponential results in promoting space traffic management with respect to collision avoidance as well as launch and re-entry procedures/phases. New technologies can support the prospects of a successful space traffic management system at an international scale by enabling, inter alia, timely, accurate and analytical processing of large data sets and rapid decision-making, more precise space debris identification and tracking and overall minimization of collision risks and reduction of operational costs. What is more, a significant part of space activities (i.e. launch and/or re-entry phase) takes place in airspace rather than in outer space, hence the overall discussion also involves the highly developed, both technically and legally, international (and national) Air Traffic Management System (ATM). Nonetheless, from a regulatory perspective, the use of AI for the purposes of space traffic management puts forward implications that merit particular attention. Key issues in this regard include the delimitation of AI-based activities as space activities, the designation of the applicable legal regime (international space or air law, national law), the assessment of the nature and extent of international legal obligations regarding space traffic coordination, as well as the appropriate liability regime applicable to AI-based technologies when operating for space traffic coordination, taking into particular consideration the dense regulatory developments at EU level. In addition, the prospects of institutionalizing international cooperation and promoting an international governance system, together with the challenges of establishment of a comprehensive international STM regime are revisited in the light of intervention of AI technologies. This paper aims at examining regulatory implications advanced by the use of AI technology in the context of space traffic management operations and its key correlating concepts (SSA, space debris mitigation) drawing in particular on international and regional considerations in the field of STM (e.g. UNCOPUOS, International Academy of Astronautics, European Space Agency, among other actors), the promising advancements of the EU approach to AI regulation and, last but not least, national approaches regarding the use of AI in the context of space traffic management, in toto. Acknowledgment: The present work was co-funded by the European Union and Greek national funds through the Operational Program "Human Resources Development, Education and Lifelong Learning " (NSRF 2014-2020), under the call "Supporting Researchers with an Emphasis on Young Researchers – Cycle B" (MIS: 5048145).

Keywords: artificial intelligence, space traffic management, space situational awareness, space debris

Procedia PDF Downloads 261
8767 ΕSW01: A Methodology for Approaching the Design of Interior Spaces

Authors: Eirini Krasaki

Abstract:

This paper addresses the problem of designing spaces in a consistently changing environment. Space is considered as a totality of forces that coexist in the same place. Forces form the identity of space and characterize the entities that coexist within the same totality. Interior space is considered as a totality of forces which develop within an envelope. This research focuses on the formation of the tripole space-forces-totality and studies the relation of this tripole to the interior space. The point of departure for this investigation has been set the historic center of Athens, a city center where the majority of building mass is unused. The objective of the study is to connect the development of interior spaces to the alterations of the conceptions that form the built environment. The research focuses on Evripidou street, an axis around which expand both commercial and residential centers. Along Evripidou street, three case studies elaborate: a) In case study 01, Evripidou street is examined as a megastructure in which totalities of interior spaces develop. b) In case study 02, a particular group of entities (polykatoikia) that expand in Evripidou street is investigated. c) In case study 03, a particular group of entities (apartment) that derives from a specific envelope is investigated. Throughout the studies and comparisons of different scales, a design methodology that addresses the design of interior space in relation to the dynamics of the built environment is evolved.

Keywords: methodology, research by design, interior, envelope, dynamics

Procedia PDF Downloads 175