Search results for: computational neural networks
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5579

Search results for: computational neural networks

1349 Identification of Biological Pathways Causative for Breast Cancer Using Unsupervised Machine Learning

Authors: Karthik Mittal

Abstract:

This study performs an unsupervised machine learning analysis to find clusters of related SNPs which highlight biological pathways that are important for the biological mechanisms of breast cancer. Studying genetic variations in isolation is illogical because these genetic variations are known to modulate protein production and function; the downstream effects of these modifications on biological outcomes are highly interconnected. After extracting the SNPs and their effect on different types of breast cancer using the MRBase library, two unsupervised machine learning clustering algorithms were implemented on the genetic variants: a k-means clustering algorithm and a hierarchical clustering algorithm; furthermore, principal component analysis was executed to visually represent the data. These algorithms specifically used the SNP’s beta value on the three different types of breast cancer tested in this project (estrogen-receptor positive breast cancer, estrogen-receptor negative breast cancer, and breast cancer in general) to perform this clustering. Two significant genetic pathways validated the clustering produced by this project: the MAPK signaling pathway and the connection between the BRCA2 gene and the ESR1 gene. This study provides the first proof of concept showing the importance of unsupervised machine learning in interpreting GWAS summary statistics.

Keywords: breast cancer, computational biology, unsupervised machine learning, k-means, PCA

Procedia PDF Downloads 146
1348 Analysis of Sweat Evaporation and Heat Transfer on Skin Surface: A Pointwise Numerical Study

Authors: Utsav Swarnkar, Rabi Pathak, Rina Maiti

Abstract:

This study aims to investigate the thermoregulatory role of sweating by comprehensively analyzing the evaporation process and its thermal cooling impact on local skin temperature at various time intervals. Traditional experimental methods struggle to fully capture these intricate phenomena. Therefore, numerical simulations play a crucial role in assessing sweat production rates and associated thermal cooling. This research utilizes transient computational fluid dynamics (CFD) to enhance our understanding of the evaporative cooling process on human skin. We conducted a simulation employing the k-w SST turbulence model. This simulation includes a scenario where sweat evaporation occurs over the skin surface, and at particular time intervals, temperatures at different locations have been observed and its effect explained. During this study, sweat evaporation was monitored on the skin surface following the commencement of the simulation. Subsequent to the simulation, various observations were made regarding temperature fluctuations at specific points over time intervals. It was noted that points situated closer to the periphery of the droplets exhibited higher levels of heat transfer and lower temperatures, whereas points within the droplets displayed contrasting trends.

Keywords: CFD, sweat, evaporation, multiphase flow, local heat loss

Procedia PDF Downloads 66
1347 Applications for Additive Manufacturing Technology for Reducing the Weight of Body Parts of Gas Turbine Engines

Authors: Liubov Magerramova, Mikhail Petrov, Vladimir Isakov, Liana Shcherbinina, Suren Gukasyan, Daniil Povalyukhin, Olga Klimova-Korsmik, Darya Volosevich

Abstract:

Aircraft engines are developing along the path of increasing resource, strength, reliability, and safety. The building of gas turbine engine body parts is a complex design and technological task. Particularly complex in the design and manufacturing are the casings of the input stages of helicopter gearboxes and central drives of aircraft engines. Traditional technologies, such as precision casting or isothermal forging, are characterized by significant limitations in parts production. For parts like housing, additive technologies guarantee spatial freedom and limitless or flexible design. This article presents the results of computational and experimental studies. These investigations justify the applicability of additive technologies (AT) to reduce the weight of aircraft housing gearbox parts by up to 32%. This is possible due to geometrical optimization compared to the classical, less flexible manufacturing methods and as-casted aircraft parts with over-insured values of safety factors. Using an example of the body of the input stage of an aircraft gearbox, visualization of the layer-by-layer manufacturing of a part based on thermal deformation was demonstrated.

Keywords: additive technologies, gas turbine engines, topological optimization, synthesis process

Procedia PDF Downloads 116
1346 The Effect of Artificial Intelligence on Urbanism, Architecture and Environmental Conditions

Authors: Abanoub Rady Shaker Saleb

Abstract:

Nowadays, design and architecture are being affected and underwent change with the rapid advancements in technology, economics, politics, society and culture. Architecture has been transforming with the latest developments after the inclusion of computers into design. Integration of design into the computational environment has revolutionized the architecture and new perspectives in architecture have been gained. The history of architecture shows the various technological developments and changes in which the architecture has transformed with time. Therefore, the analysis of integration between technology and the history of the architectural process makes it possible to build a consensus on the idea of how architecture is to proceed. In this study, each period that occurs with the integration of technology into architecture is addressed within historical process. At the same time, changes in architecture via technology are identified as important milestones and predictions with regards to the future of architecture have been determined. Developments and changes in technology and the use of technology in architecture within years are analyzed in charts and graphs comparatively. The historical process of architecture and its transformation via technology are supported with detailed literature review and they are consolidated with the examination of focal points of 20th-century architecture under the titles; parametric design, genetic architecture, simulation, and biomimicry. It is concluded that with the historical research between past and present; the developments in architecture cannot keep up with the advancements in technology and recent developments in technology overshadow the architecture, even the technology decides the direction of architecture. As a result, a scenario is presented with regards to the reach of technology in the future of architecture and the role of the architect.

Keywords: design and development the information technology architecture, enterprise architecture, enterprise architecture design result, TOGAF architecture development method (ADM)

Procedia PDF Downloads 69
1345 Brain Networks and Mathematical Learning Processes of Children

Authors: Felicitas Pielsticker, Christoph Pielsticker, Ingo Witzke

Abstract:

Neurological findings provide foundational results for many different disciplines. In this article we want to discuss these with a special focus on mathematics education. The intention is to make neuroscience research useful for the description of cognitive mathematical learning processes. A key issue of mathematics education is that students often behave as if their mathematical knowledge is constructed in isolated compartments with respect to the specific context of the original learning situation; supporting students to link these compartments to form a coherent mathematical society of mind is a fundamental task not only for mathematics teachers. This aspect goes hand in hand with the question if there is such a thing as abstract general mathematical knowledge detached from concrete reality. Educational Neuroscience may give answers to the question why students develop their mathematical knowledge in isolated subjective domains of experience and if it is generally possible to think in abstract terms. To address these questions, we will provide examples from different fields of mathematics education e.g. students’ development and understanding of the general concept of variables or the mathematical notion of universal proofs. We want to discuss these aspects in the reflection of functional studies which elucidate the role of specific brain regions in mathematical learning processes. In doing this the paper addresses concept formation processes of students in the mathematics classroom and how to support them adequately considering the results of (educational) neuroscience.

Keywords: brain regions, concept formation processes in mathematics education, proofs, teaching-learning processes

Procedia PDF Downloads 149
1344 CFD Modelling and Thermal Performance Analysis of Ventilated Double Skin Roof Structure

Authors: A. O. Idris, J. Virgone, A. I. Ibrahim, D. David, E. Vergnault

Abstract:

In hot countries, the major challenge is the air conditioning. The increase in energy consumption by air conditioning stems from the need to live in more comfortable buildings, which is understandable. But in Djibouti, one of the countries with the most expensive electricity in the world, this need is exacerbated by an architecture that is inappropriate and unsuitable for climatic conditions. This paper discusses the design of the roof which is the surface receiving the most solar radiation. The roof determines the general behavior of the building. The study presents Computational Fluid Dynamics (CFD) modeling and analysis of the energy performance of a double skin ventilated roof. The particularity of this study is that it considers the climate of Djibouti characterized by hot and humid conditions in winter and very hot and humid in summer. Roof simulations are carried out using the Ansys Fluent software to characterize the flow and the heat transfer induced in the ventilated roof in steady state. This modeling is carried out by comparing the influence of several parameters such as the internal emissivity of the upper surface, the thickness of the insulation of the roof and the thickness of the ventilated channel on heat gain through the roof. The energy saving potential compared to the current construction in Djibouti is also presented.

Keywords: building, double skin roof, CFD, thermo-fluid analysis, energy saving, forced convection, natural convection

Procedia PDF Downloads 263
1343 An Integrated Modular Approach Based Simulation of Cold Heavy Oil Production

Authors: Hamidreza Sahaleh

Abstract:

In this paper, the authors display an incorporated secluded way to deal with quantitatively foresee volumetric sand generation and improved oil recuperation. This model is in light of blend hypothesis with erosion mechanics, in which multiphase hydrodynamics and geo-mechanics are coupled in a predictable way by means of principal unknowns, for example, saturation, pressure, porosity, and formation displacements. Foamy oil is demonstrated as a scattering of gas bubbles caught in the oil, where these gas air bubbles keep up a higher repository weight. A secluded methodology is then received to adequately exploit the current propelled standard supply and stress-strain codes. The model is actualized into three coordinated computational modules, i.e. erosion module, store module, and geo-mechanics module. The stress, stream and erosion mathematical statements are understood independently for every time addition, and the coupling terms (porosity, penetrability, plastic shear strain, and so on) are gone among them and iterated until certain union is accomplished on a period step premise. The framework is capable regarding its abilities, yet practical in terms of computer requirements and maintenance. Numerical results of field studies are displayed to show the capacities of the model. The impacts of foamy oil stream and sand generation are additionally inspected to exhibit their effect on the upgraded hydrocarbon recuperation.

Keywords: oil recuperation, erosion mechanics, foamy oil, erosion module.

Procedia PDF Downloads 268
1342 Resource Orchestration Based on Two-Sides Scheduling in Computing Network Control Sytems

Authors: Li Guo, Jianhong Wang, Dian Huang, Shengzhong Feng

Abstract:

Computing networks as a new network architecture has shown great promise in boosting the utilization of different resources, such as computing, caching, and communications. To maximise the efficiency of resource orchestration in computing network control systems (CNCSs), this work proposes a dynamic orchestration strategy of a different resource based on task requirements from computing power requestors (CPRs). Specifically, computing power providers (CPPs) in CNCSs could share information with each other through communication channels on the basis of blockchain technology, especially their current idle resources. This dynamic process is modeled as a cooperative game in which CPPs have the same target of maximising long-term rewards by improving the resource utilization ratio. Meanwhile, the task requirements from CPRs, including size, deadline, and calculation, are simultaneously considered in this paper. According to task requirements, the proposed orchestration strategy could schedule the best-fitting resource in CNCSs, achieving the maximum long-term rewards of CPPs and the best quality of experience (QoE) of CRRs at the same time. Based on the EdgeCloudSim simulation platform, the efficiency of the proposed strategy is achieved from both sides of CPRs and CPPs. Besides, experimental results show that the proposed strategy outperforms the other comparisons in all cases.

Keywords: computing network control systems, resource orchestration, dynamic scheduling, blockchain, cooperative game

Procedia PDF Downloads 114
1341 A Sectional Control Method to Decrease the Accumulated Survey Error of Tunnel Installation Control Network

Authors: Yinggang Guo, Zongchun Li

Abstract:

In order to decrease the accumulated survey error of tunnel installation control network of particle accelerator, a sectional control method is proposed. Firstly, the accumulation rule of positional error with the length of the control network is obtained by simulation calculation according to the shape of the tunnel installation-control-network. Then, the RMS of horizontal positional precision of tunnel backbone control network is taken as the threshold. When the accumulated error is bigger than the threshold, the tunnel installation control network should be divided into subsections reasonably. On each segment, the middle survey station is taken as the datum for independent adjustment calculation. Finally, by taking the backbone control points as faint datums, the weighted partial parameters adjustment is performed with the adjustment results of each segment and the coordinates of backbone control points. The subsections are jointed and unified into the global coordinate system in the adjustment process. An installation control network of the linac with a length of 1.6 km is simulated. The RMS of positional deviation of the proposed method is 2.583 mm, and the RMS of the difference of positional deviation between adjacent points reaches 0.035 mm. Experimental results show that the proposed sectional control method can not only effectively decrease the accumulated survey error but also guarantee the relative positional precision of the installation control network. So it can be applied in the data processing of tunnel installation control networks, especially for large particle accelerators.

Keywords: alignment, tunnel installation control network, accumulated survey error, sectional control method, datum

Procedia PDF Downloads 191
1340 Computational Study on the Crystal Structure, Electronic and Optical Properties of Perovskites a2bx6 for Photovoltaic Applications

Authors: Harmel Meriem

Abstract:

The optoelectronic properties and high power conversion efficiency make lead halide perovskites ideal material for solar cell applications. However, the toxic nature of lead and the instability of organic cation are the two key challenges in the emerging perovskite solar cells. To overcome these challenges, we present our study about finding potential alternatives to lead in the form of A2BX6 perovskite using the first principles DFT-based calculations. The highly accurate modified Becke Johnson (mBJ) and hybrid functional (HSE06) have been used to investigate the Main Document Click here to view linked References to optoelectronic and thermoelectric properties of A2PdBr6 (A = K, Rb, and Cs) perovskite. The results indicate that different A-cations in A2PdBr6 can significantly alter their electronic and optical properties. Calculated band structures indicate semiconducting nature, with band gap values of 1.84, 1.53, and 1.54 eV for K2PdBr6, Rb2PdBr6, and Cs2PdBr6, respectively. We find strong optical absorption in the visible region with small effective masses for A2PdBr6. The ideal band gap and optimum light absorption suggest Rb2PdBr6 and Cs2PdBr6 potential candidates for the light absorption layer in perovskite solar cells. Additionally.

Keywords: soler cell, double perovskite, optoelectronic properties, ab-inotio study

Procedia PDF Downloads 128
1339 Distributed Automation System Based Remote Monitoring of Power Quality Disturbance on LV Network

Authors: Emmanuel D. Buedi, K. O. Boateng, Griffith S. Klogo

Abstract:

Electrical distribution networks are prone to power quality disturbances originating from the complexity of the distribution network, mode of distribution (overhead or underground) and types of loads used by customers. Data on the types of disturbances present and frequency of occurrence is needed for economic evaluation and hence finding solution to the problem. Utility companies have resorted to using secondary power quality devices such as smart meters to help gather the required data. Even though this approach is easier to adopt, data gathered from these devices may not serve the required purpose, since the installation of these devices in the electrical network usually does not conform to available PQM placement methods. This paper presents a design of a PQM that is capable of integrating into an existing DAS infrastructure to take advantage of available placement methodologies. The monitoring component of the design is implemented and installed to monitor an existing LV network. Data from the monitor is analyzed and presented. A portion of the LV network of the Electricity Company of Ghana is modeled in MATLAB-Simulink and analyzed under various earth fault conditions. The results presented show the ability of the PQM to detect and analyze PQ disturbance such as voltage sag and overvoltage. By adopting a placement methodology and installing these nodes, utilities are assured of accurate and reliable information with respect to the quality of power delivered to consumers.

Keywords: power quality, remote monitoring, distributed automation system, economic evaluation, LV network

Procedia PDF Downloads 349
1338 Use of Machine Learning in Data Quality Assessment

Authors: Bruno Pinto Vieira, Marco Antonio Calijorne Soares, Armando Sérgio de Aguiar Filho

Abstract:

Nowadays, a massive amount of information has been produced by different data sources, including mobile devices and transactional systems. In this scenario, concerns arise on how to maintain or establish data quality, which is now treated as a product to be defined, measured, analyzed, and improved to meet consumers' needs, which is the one who uses these data in decision making and companies strategies. Information that reaches low levels of quality can lead to issues that can consume time and money, such as missed business opportunities, inadequate decisions, and bad risk management actions. The step of selecting, identifying, evaluating, and selecting data sources with significant quality according to the need has become a costly task for users since the sources do not provide information about their quality. Traditional data quality control methods are based on user experience or business rules limiting performance and slowing down the process with less than desirable accuracy. Using advanced machine learning algorithms, it is possible to take advantage of computational resources to overcome challenges and add value to companies and users. In this study, machine learning is applied to data quality analysis on different datasets, seeking to compare the performance of the techniques according to the dimensions of quality assessment. As a result, we could create a ranking of approaches used, besides a system that is able to carry out automatically, data quality assessment.

Keywords: machine learning, data quality, quality dimension, quality assessment

Procedia PDF Downloads 148
1337 Computational Analyses of Persian Walnut Genetic Data: Notes on Genetic Diversity and Cultivar Phylogeny

Authors: Masoud Sheidaei, Melica Tabasi, Fahimeh Koohdar, Mona Sheidaei

Abstract:

Juglans regia L. is an economically important species of edible nuts. Iran is known as a center of origin of genetically rich walnut germplasm and expected to be found a large diversity within Iranian walnut populations. A detailed population genetic of local populations is useful for developing an optimal strategy for in situ conservation and can assist the breeders in crop improvement programs. Different phylogenetic studies have been carried out in this genus, but none has been concerned with genetic changes associated with geographical divergence and the identification of adaptive SNPs. Therefore, we carried out the present study to identify discriminating ITS nucleotides among Juglans species and also reveal association between ITS SNPs and geographical variables. We used different computations approaches like DAPC, CCA, and RDA analyses for the above-mentioned tasks. We also performed population genetics analyses for population effective size changes associated with the species expansion. The results obtained suggest that latitudinal distribution has a more profound effect on the species genetic changes. Similarly, multiple analytical approaches utilized for the identification of both discriminating DNA nucleotides/ SNPs almost produced congruent results. The SNPs with different phylogenetic importance were also identified by using a parsimony approach.

Keywords: Persian walnut, adaptive SNPs, data analyses, genetic diversity

Procedia PDF Downloads 129
1336 The Mathematics of Fractal Art: Using a Derived Cubic Method and the Julia Programming Language to Make Fractal Zoom Videos

Authors: Darsh N. Patel, Eric Olson

Abstract:

Fractals can be found everywhere, whether it be the shape of a leaf or a system of blood vessels. Fractals are used to help study and understand different physical and mathematical processes; however, their artistic nature is also beautiful to simply explore. This project explores fractals generated by a cubically convergent extension to Newton's method. With this iteration as a starting point, a complex plane spanning from -2 to 2 is created with a color wheel mapped onto it. Next, the polynomial whose roots the fractal will generate from is established. From the Fundamental Theorem of Algebra, it is known that any polynomial has as many roots (counted by multiplicity) as its degree. When generating the fractals, each root will receive its own color. The complex plane can then be colored to indicate the basins of attraction that converge to each root. From a computational point of view, this project’s code identifies which points converge to which roots and then obtains fractal images. A zoom path into the fractal was implemented to easily visualize the self-similar structure. This path was obtained by selecting keyframes at different magnifications through which a path is then interpolated. Using parallel processing, many images were generated and condensed into a video. This project illustrates how practical techniques used for scientific visualization can also have an artistic side.

Keywords: fractals, cubic method, Julia programming language, basin of attraction

Procedia PDF Downloads 253
1335 A Comparative Soft Computing Approach to Supplier Performance Prediction Using GEP and ANN Models: An Automotive Case Study

Authors: Seyed Esmail Seyedi Bariran, Khairul Salleh Mohamed Sahari

Abstract:

In multi-echelon supply chain networks, optimal supplier selection significantly depends on the accuracy of suppliers’ performance prediction. Different methods of multi criteria decision making such as ANN, GA, Fuzzy, AHP, etc have been previously used to predict the supplier performance but the “black-box” characteristic of these methods is yet a major concern to be resolved. Therefore, the primary objective in this paper is to implement an artificial intelligence-based gene expression programming (GEP) model to compare the prediction accuracy with that of ANN. A full factorial design with %95 confidence interval is initially applied to determine the appropriate set of criteria for supplier performance evaluation. A test-train approach is then utilized for the ANN and GEP exclusively. The training results are used to find the optimal network architecture and the testing data will determine the prediction accuracy of each method based on measures of root mean square error (RMSE) and correlation coefficient (R2). The results of a case study conducted in Supplying Automotive Parts Co. (SAPCO) with more than 100 local and foreign supply chain members revealed that, in comparison with ANN, gene expression programming has a significant preference in predicting supplier performance by referring to the respective RMSE and R-squared values. Moreover, using GEP, a mathematical function was also derived to solve the issue of ANN black-box structure in modeling the performance prediction.

Keywords: Supplier Performance Prediction, ANN, GEP, Automotive, SAPCO

Procedia PDF Downloads 419
1334 Using T-Splines to Model Point Clouds from Terrestrial Laser Scanner

Authors: G. Kermarrec, J. Hartmann

Abstract:

Spline surfaces are a major representation of freeform surfaces in the computer-aided graphic industry and were recently introduced in the field of geodesy for processing point clouds from terrestrial laser scanner (TLS). The surface fitting consists of approximating a trustworthy mathematical surface to a large numbered 3D point cloud. The standard B-spline surfaces lack of local refinement due to the tensor-product construction. The consequences are oscillating geometry, particularly in the transition from low-to-high curvature parts for scattered point clouds with missing data. More economic alternatives in terms of parameters on how to handle point clouds with a huge amount of observations are the recently introduced T-splines. As long as the partition of unity is guaranteed, their computational complexity is low, and they are flexible. T-splines are implemented in a commercial package called Rhino, a 3D modeler which is widely used in computer aided design to create and animate NURBS objects. We have applied T-splines surface fitting to terrestrial laser scanner point clouds from a bridge under load and a sheet pile wall with noisy observations. We will highlight their potential for modelling details with high trustworthiness, paving the way for further applications in terms of deformation analysis.

Keywords: deformation analysis, surface modelling, terrestrial laser scanner, T-splines

Procedia PDF Downloads 140
1333 Hybrid Genetic Approach for Solving Economic Dispatch Problems with Valve-Point Effect

Authors: Mohamed I. Mahrous, Mohamed G. Ashmawy

Abstract:

Hybrid genetic algorithm (HGA) is proposed in this paper to determine the economic scheduling of electric power generation over a fixed time period under various system and operational constraints. The proposed technique can outperform conventional genetic algorithms (CGAs) in the sense that HGA make it possible to improve both the quality of the solution and reduce the computing expenses. In contrast, any carefully designed GA is only able to balance the exploration and the exploitation of the search effort, which means that an increase in the accuracy of a solution can only occure at the sacrifice of convergent speed, and vice visa. It is unlikely that both of them can be improved simultaneously. The proposed hybrid scheme is developed in such a way that a simple GA is acting as a base level search, which makes a quick decision to direct the search towards the optimal region, and a local search method (pattern search technique) is next employed to do the fine tuning. The aim of the strategy is to achieve the cost reduction within a reasonable computing time. The effectiveness of the proposed hybrid technique is verified on two real public electricity supply systems with 13 and 40 generator units respectively. The simulation results obtained with the HGA for the two real systems are very encouraging with regard to the computational expenses and the cost reduction of power generation.

Keywords: genetic algorithms, economic dispatch, pattern search

Procedia PDF Downloads 444
1332 Gas Network Noncooperative Game

Authors: Teresa Azevedo PerdicoúLis, Paulo Lopes Dos Santos

Abstract:

The conceptualisation of the problem of network optimisation as a noncooperative game sets up a holistic interactive approach that brings together different network features (e.g., com-pressor stations, sources, and pipelines, in the gas context) where the optimisation objectives are different, and a single optimisation procedure becomes possible without having to feed results from diverse software packages into each other. A mathematical model of this type, where independent entities take action, offers the ideal modularity and subsequent problem decomposition in view to design a decentralised algorithm to optimise the operation and management of the network. In a game framework, compressor stations and sources are under-stood as players which communicate through network connectivity constraints–the pipeline model. That is, in a scheme similar to tatonnementˆ, the players appoint their best settings and then interact to check for network feasibility. The devolved degree of network unfeasibility informs the players about the ’quality’ of their settings, and this two-phase iterative scheme is repeated until a global optimum is obtained. Due to network transients, its optimisation needs to be assessed at different points of the control interval. For this reason, the proposed approach to optimisation has two stages: (i) the first stage computes along the period of optimisation in order to fulfil the requirement just mentioned; (ii) the second stage is initialised with the solution found by the problem computed at the first stage, and computes in the end of the period of optimisation to rectify the solution found at the first stage. The liability of the proposed scheme is proven correct on an abstract prototype and three example networks.

Keywords: connectivity matrix, gas network optimisation, large-scale, noncooperative game, system decomposition

Procedia PDF Downloads 152
1331 Two-Dimensional Transition Metal Dichalcogenides for Photodetection and Biosensing

Authors: Mariam Badmus, Bothina Manasreh

Abstract:

Transition metal dichalcogenides (TMDs) have gained significant attention as two-dimensional (2D) materials due to their intrinsic band gaps and unique properties, which make them ideal candidates for electronic and photonic applications. Unlike graphene, which lacks a band gap, TMDs (MX₂, where M is a transition metal and X is a chalcogen such as sulfur, selenium, or tellurium) exhibit semiconductor behavior and can be exfoliated into monolayers, enhancing their properties. The properties of these materials are investigated using density functional theory, a quantum mechanical computational method to solve Schrodinger equation for many body problems to calculate electron density of the atoms involved on which the energy and properties of a system depend. They show promise for use in photodetectors, biosensors, memory devices, and other technologies in communications, health, and energy sectors. In particular, metallic TMDs, which lack an intrinsic band gap, benefit from doping with transition metals, this improves their electronic and optical properties. Doping monolayer TMDs yields more significant improvements than doping bulk materials. Notably, doping with metals such as vanadium enhances the magnetization of TMDs, expanding their potential applications in spintronics. This work highlights the effects of doping on TMDs and explores strategies for optimizing their performance for advanced technological applications.

Keywords: concentration, doping, magnetization, monolayer

Procedia PDF Downloads 11
1330 An Artificial Intelligence Framework to Forecast Air Quality

Authors: Richard Ren

Abstract:

Air pollution is a serious danger to international well-being and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.

Keywords: air quality prediction, air pollution, artificial intelligence, machine learning algorithms

Procedia PDF Downloads 127
1329 Implicit and Explicit Mechanisms of Emotional Contagion

Authors: Andres Pinilla Palacios, Ricardo Tamayo

Abstract:

Emotional contagion is characterized as an automatic tendency to synchronize behaviors that facilitate emotional convergence among humans. It might thus play a pivotal role to understand the dynamics of key social interactions. However, a few research has investigated its potential mechanisms. We suggest two complementary but independent processes that may underlie emotional contagion. The efficient contagion hypothesis, based on fast and implicit bottom-up processes, modulated by familiarity and spread of activation in the emotional associative networks of memory. Secondly, the emotional contrast hypothesis, based on slow and explicit top-down processes guided by deliberated appraisal and hypothesis-testing. In order to assess these two hypotheses, an experiment with 39 participants was conducted. In the first phase, participants were induced (between-groups) to an emotional state (positive, neutral or negative) using a standardized video taken from the FilmStim database. In the second phase, participants classified and rated (within-subject) the emotional state of 15 faces (5 for each emotional state) taken from the POFA database. In the third phase, all participants were returned to a baseline emotional state using the same neutral video used in the first phase. In a fourth phase, participants classified and rated a new set of 15 faces. The accuracy in the identification and rating of emotions was partially explained by the efficient contagion hypothesis, but the speed with which these judgments were made was partially explained by the emotional contrast hypothesis. However, results are ambiguous, so a follow-up experiment is proposed in which emotional expressions and activation of the sympathetic system will be measured using EMG and EDA respectively.

Keywords: electromyography, emotional contagion, emotional valence, identification of emotions, imitation

Procedia PDF Downloads 316
1328 High Pressure Multiphase Flow Experiments: The Impact of Pressure on Flow Patterns Using an X-Ray Tomography Visualisation System

Authors: Sandy Black, Calum McLaughlin, Alessandro Pranzitelli, Marc Laing

Abstract:

Multiphase flow structures of two-phase multicomponent fluids were experimentally investigated in a large diameter high-pressure pipeline up to 130 bar at TÜV SÜD’s National Engineering Laboratory Advanced Multiphase Facility. One of the main objectives of the experimental test campaign was to evaluate the impact of pressure on multiphase flow patterns as much of the existing information is based on low-pressure measurements. The experiments were performed in a horizontal and vertical orientation in both 4-inch and 6-inch pipework using nitrogen, ExxsolTM D140 oil, and a 6% aqueous solution of NaCl at incremental pressures from 10 bar to 130 bar. To visualise the detailed structure of the flow of the entire cross-section of the pipe, a fast response X-ray tomography system was used. A wide range of superficial velocities from 0.6 m/s to 24.0 m/s for gas and 0.04 m/s and 6.48 m/s for liquid was examined to evaluate different flow regimes. The results illustrated the suppression of instabilities between the gas and the liquid at the measurement location and that intermittent or slug flow was observed less frequently as the pressure was increased. CFD modellings of low and high-pressure simulations were able to successfully predict the likelihood of intermittent flow; however, further tuning is necessary to predict the slugging frequency. The dataset generated is unique as limited datasets exist above 100 bar and is of considerable value to multiphase flow specialists and numerical modellers.

Keywords: computational fluid dynamics, high pressure, multiphase, X-ray tomography

Procedia PDF Downloads 143
1327 A Multi-Science Study of Modern Synergetic War and Its Information Security Component

Authors: Alexander G. Yushchenko

Abstract:

From a multi-science point of view, we analyze threats to security resulting from globalization of international information space and information and communication aggression of Russia. A definition of Ruschism is formulated as an ideology supporting aggressive actions of modern Russia against the Euro-Atlantic community. Stages of the hybrid war Russia is leading against Ukraine are described, including the elements of subversive activity of the special services, the activation of the military phase and the gradual shift of the focus of confrontation to the realm of information and communication technologies. We reveal an emergence of a threat for democratic states resulting from the destabilizing impact of a target state’s mass media and social networks being exploited by Russian secret services under freedom-of-speech disguise. Thus, we underline the vulnerability of cyber- and information security of the network society in regard of hybrid war. We propose to define the latter a synergetic war. Our analysis is supported with a long-term qualitative monitoring of representation of top state officials on popular TV channels and Facebook. From the memetics point of view, we have detected a destructive psycho-information technology used by the Kremlin, a kind of information catastrophe, the essence of which is explained in detail. In the conclusion, a comprehensive plan for information protection of the public consciousness and mentality of Euro-Atlantic citizens from the aggression of the enemy is proposed.

Keywords: cyber and information security, hybrid war, psycho-information technology, synergetic war, Ruschism

Procedia PDF Downloads 134
1326 Census and Mapping of Oil Palms Over Satellite Dataset Using Deep Learning Model

Authors: Gholba Niranjan Dilip, Anil Kumar

Abstract:

Conduct of accurate reliable mapping of oil palm plantations and census of individual palm trees is a huge challenge. This study addresses this challenge and developed an optimized solution implemented deep learning techniques on remote sensing data. The oil palm is a very important tropical crop. To improve its productivity and land management, it is imperative to have accurate census over large areas. Since, manual census is costly and prone to approximations, a methodology for automated census using panchromatic images from Cartosat-2, SkySat and World View-3 satellites is demonstrated. It is selected two different study sites in Indonesia. The customized set of training data and ground-truth data are created for this study from Cartosat-2 images. The pre-trained model of Single Shot MultiBox Detector (SSD) Lite MobileNet V2 Convolutional Neural Network (CNN) from the TensorFlow Object Detection API is subjected to transfer learning on this customized dataset. The SSD model is able to generate the bounding boxes for each oil palm and also do the counting of palms with good accuracy on the panchromatic images. The detection yielded an F-Score of 83.16 % on seven different images. The detections are buffered and dissolved to generate polygons demarcating the boundaries of the oil palm plantations. This provided the area under the plantations and also gave maps of their location, thereby completing the automated census, with a fairly high accuracy (≈100%). The trained CNN was found competent enough to detect oil palm crowns from images obtained from multiple satellite sensors and of varying temporal vintage. It helped to estimate the increase in oil palm plantations from 2014 to 2021 in the study area. The study proved that high-resolution panchromatic satellite image can successfully be used to undertake census of oil palm plantations using CNNs.

Keywords: object detection, oil palm tree census, panchromatic images, single shot multibox detector

Procedia PDF Downloads 160
1325 Decision Support: How Explainable A.I. Can Improve Transparency and Trust with Human Users

Authors: Devon Brown, Liu Chunmei

Abstract:

This paper will present an analysis as part of the researchers dissertation topic focusing on the intersection of affective and analytical directed acyclic graphs (DAGs) in the context of Decision Support Systems (DSS). The researcher’s work involves analyzing decision theory models like Affective and Bayesian Decision theory models and how they could be implemented under an Affective Computing Framework using Information Fusion and Human-Centered Design. Additionally, the researcher is beginning research on an Affective-Analytic Decision Framework (AADF) model for their dissertation research and are looking to merge logic and analytic models with empathetic insights into affective DAGs. Data-collection efforts begin Fall 2024 and in preparation for the efforts this paper looks to analyze previous research in this area and introduce the AADF framework and propose conceptual models for consideration. For this paper, the research emphasis is placed on analyzing Bayesian networks and Markov models which offer probabilistic techniques during uncertainty in decision-making. Ideally, including affect into analytic models will ensure algorithms can increase user trust with algorithms by including emotional states and the user’s experience with the goal of developing emotionally intelligent A.I. systems that can start to navigate the complex fabric of human emotion during decision-making.

Keywords: decision support systems, explainable AI, HCAI techniques, affective-analytical decision framework

Procedia PDF Downloads 20
1324 Analytical Study: An M-Learning App Reflecting the Factors Affecting Student’s Adoption of M-Learning

Authors: Ahmad Khachan, Ahmet Ozmen

Abstract:

This study aims to introduce a mobile bite-sized learning concept, a mobile application with social networks motivation factors that will encourage students to practice critical thinking, improve analytical skills and learn knowledge sharing. We do not aim to propose another e-learning or distance learning based tool like Moodle and Edmodo; instead, we introduce a mobile learning tool called Interactive M-learning Application. The tool reconstructs and strengthens the bonds between educators and learners and provides a foundation for integrating mobile devices in education. The application allows learners to stay connected all the time, share ideas, ask questions and learn from each other. It is built on Android since the Android has the largest platform share in the world and is dominating the market with 74.45% share in 2018. We have chosen Google-Firebase server for hosting because of flexibility, ease of hosting and real time update capabilities. The proposed m-learning tool was offered to four groups of university students in different majors. An improvement in the relation between the students, the teachers and the academic institution was obvious. Student’s performance got much better added to better analytical and critical skills advancement and moreover a willingness to adopt mobile learning in class. We have also compared our app with another tool in the same class for clarity and reliability of the results. The student’s mobile devices were used in this experimental study for diversity of devices and platform versions.

Keywords: education, engineering, interactive software, undergraduate education

Procedia PDF Downloads 155
1323 A Machine Learning Pipeline for Real-Time Activity Detection on Low Computational Power Devices for Metaverse Applications

Authors: Amit Kumar, Amanpreet Chander, Ashish Sahani

Abstract:

This paper presents our recent work on real-time human activity detection based on the media pipe pipeline and machine learning algorithms. The proposed system can detect human activities, including running, jumping, squatting, bending to the left or right, and standing still. This is a robust solution for developing a yoga, dance, metaverse, and fitness application that checks for the correction of the pose without having any additional monitor like a personal trainer. MediaPipe solution offers an open-source cross-platform which utilizes a two-step detector-tracker ML pipeline for live detection of key landmarks on our body which can be used for motion data collection. The prediction of real-time poses uses a variety of machine learning techniques and different types of analysis. Without primarily relying on powerful desktop environments for inference, our method achieves real-time performance on the majority of contemporary mobile phones, desktops/laptops, Python, or even the web. Experimental results show that our method outperforms the existing method in terms of accuracy and real-time capability, achieving an accuracy of 99.92% on testing datasets.

Keywords: human activity detection, media pipe, machine learning, metaverse applications

Procedia PDF Downloads 179
1322 Emerging Cyber Threats and Cognitive Vulnerabilities: Cyberterrorism

Authors: Oludare Isaac Abiodun, Esther Omolara Abiodun

Abstract:

The purpose of this paper is to demonstrate that cyberterrorism is existing and poses a threat to computer security and national security. Nowadays, people have become excitedly dependent upon computers, phones, the Internet, and the Internet of things systems to share information, communicate, conduct a search, etc. However, these network systems are at risk from a different source that is known and unknown. These network systems risk being caused by some malicious individuals, groups, organizations, or governments, they take advantage of vulnerabilities in the computer system to hawk sensitive information from people, organizations, or governments. In doing so, they are engaging themselves in computer threats, crime, and terrorism, thereby making the use of computers insecure for others. The threat of cyberterrorism is of various forms and ranges from one country to another country. These threats include disrupting communications and information, stealing data, destroying data, leaking, and breaching data, interfering with messages and networks, and in some cases, demanding financial rewards for stolen data. Hence, this study identifies many ways that cyberterrorists utilize the Internet as a tool to advance their malicious mission, which negatively affects computer security and safety. One could identify causes for disparate anomaly behaviors and the theoretical, ideological, and current forms of the likelihood of cyberterrorism. Therefore, for a countermeasure, this paper proposes the use of previous and current computer security models as found in the literature to help in countering cyberterrorism

Keywords: cyberterrorism, computer security, information, internet, terrorism, threat, digital forensic solution

Procedia PDF Downloads 96
1321 Free Vibration of Axially Functionally Graded Simply Supported Beams Using Differential Transformation Method

Authors: A. Selmi

Abstract:

Free vibration analysis of homogenous and axially functionally graded simply supported beams within the context of Euler-Bernoulli beam theory is presented in this paper. The material properties of the beams are assumed to obey the linear law distribution. The effective elastic modulus of the composite was predicted by using the rule of mixture. Here, the complexities which appear in solving differential equation of transverse vibration of composite beams which limit the analytical solution to some special cases are overcome using a relatively new approach called the Differential Transformation Method. This technique is applied for solving differential equation of transverse vibration of axially functionally graded beams. Natural frequencies and corresponding normalized mode shapes are calculated for different Young’s modulus ratios. MATLAB code is designed to solve the transformed differential equation of the beam. Comparison of the present results with the exact solutions proves the effectiveness, the accuracy, the simplicity, and computational stability of the differential transformation method. The effect of the Young’s modulus ratio on the normalized natural frequencies and mode shapes is found to be very important.

Keywords: differential transformation method, functionally graded material, mode shape, natural frequency

Procedia PDF Downloads 309
1320 Thermal-Fluid Characteristics of Heating Element in Rotary Heat Exchanger in Accordance with Fouling Phenomena

Authors: Young Mun Lee, Seon Ho Kim, Seok Min Choi, JeongJu Kim, Seungyeong Choi, Hyung Hee Cho

Abstract:

To decrease sulfur oxide in the flue gas from coal power plant, a flue gas de-sulfurization facility is operated. In the reactor, a chemical reaction occurs with a temperature change of the gas so that sulfur oxide is removed and cleaned air is emitted. In this process, temperature change induces a serious problem which is a cold erosion of stack. To solve this problem, the rotary heat exchanger is managed before the stack. In the heat exchanger, a heating element is equipped to increase a heat transfer area. Heat transfer and pressure loss is a big issue to improve a performance. In this research, thermal-fluid characteristics of the heating element are analyzed by computational fluid dynamics. Fouling simulation is also conducted to calculate a performance of heating element. Numerical analysis is performed on the situation where plugging phenomenon has already occurred and existed in the inlet region of the heating element. As the pressure of the rear part of the plugging decreases suddenly and the flow velocity becomes slower, it is found that the flow is gathered from both sides as it develops in the flow direction, and it is confirmed that the pressure difference due to plugging is increased.

Keywords: heating element, plugging, rotary heat exchanger, thermal fluid characteristics

Procedia PDF Downloads 485