Search results for: computational intelligence
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3427

Search results for: computational intelligence

2617 Disparities in Language Competence and Conflict: The Moderating Role of Cultural Intelligence in Intercultural Interactions

Authors: Catherine Peyrols Wu

Abstract:

Intercultural interactions are becoming increasingly common in organizations and life. These interactions are often the stage of miscommunication and conflict. In management research, these problems are commonly attributed to cultural differences in values and interactional norms. As a result, the notion that intercultural competence can minimize these challenges is widely accepted. Cultural differences, however, are not the only source of a challenge during intercultural interactions. The need to rely on a lingua franca – or common language between people who have different mother tongues – is another important one. In theory, a lingua franca can improve communication and ease coordination. In practice however, disparities in people’s ability and confidence to communicate in the language can exacerbate tensions and generate inefficiencies. In this study, we draw on power theory to develop a model of disparities in language competence and conflict in a multicultural work context. Specifically, we hypothesized that differences in language competence between interaction partners would be positively related to conflict such that people would report greater conflict with partners who have more dissimilar levels of language competence and lesser conflict with partners with more similar levels of language competence. Furthermore, we proposed that cultural intelligence (CQ) an intercultural competence that denotes an individual’s capability to be effective in intercultural situations, would weaken the relationship between disparities in language competence and conflict such that people would report less conflict with partners who have more dissimilar levels of language competence when the interaction partner has high CQ and more conflict when the partner has low CQ. We tested this model with a sample of 135 undergraduate students working in multicultural teams for 13 weeks. We used a round-robin design to examine conflict in 646 dyads nested within 21 teams. Results of analyses using social relations modeling provided support for our hypotheses. Specifically, we found that in intercultural dyads with large disparities in language competence, partners with the lowest level of language competence would report higher levels of interpersonal conflict. However, this relationship disappeared when the partner with higher language competence was also high in CQ. These findings suggest that communication in a lingua franca can be a source of conflict in intercultural collaboration when partners differ in their level of language competence and that CQ can alleviate these effects during collaboration with partners who have relatively lower levels of language competence. Theoretically, this study underscores the benefits of CQ as a complement to language competence for intercultural effectiveness. Practically, these results further attest to the benefits of investing resources to develop language competence and CQ in employees engaged in multicultural work.

Keywords: cultural intelligence, intercultural interactions, language competence, multicultural teamwork

Procedia PDF Downloads 161
2616 Two Stage Assembly Flowshop Scheduling Problem Minimizing Total Tardiness

Authors: Ali Allahverdi, Harun Aydilek, Asiye Aydilek

Abstract:

The two stage assembly flowshop scheduling problem has lots of application in real life. To the best of our knowledge, the two stage assembly flowshop scheduling problem with total tardiness performance measure and separate setup times has not been addressed so far, and hence, it is addressed in this paper. Different dominance relations are developed and several algorithms are proposed. Extensive computational experiments are conducted to evaluate the proposed algorithms. The computational experiments have shown that one of the algorithms performs much better than the others. Moreover, the experiments have shown that the best performing algorithm performs much better than the best existing algorithm for the case of zero setup times in the literature. Therefore, the proposed best performing algorithm not only can be used for problems with separate setup times but also for the case of zero setup times.

Keywords: scheduling, assembly flowshop, total tardiness, algorithm

Procedia PDF Downloads 340
2615 Copyright Clearance for Artificial Intelligence Training Data: Challenges and Solutions

Authors: Erva Akin

Abstract:

– The use of copyrighted material for machine learning purposes is a challenging issue in the field of artificial intelligence (AI). While machine learning algorithms require large amounts of data to train and improve their accuracy and creativity, the use of copyrighted material without permission from the authors may infringe on their intellectual property rights. In order to overcome copyright legal hurdle against the data sharing, access and re-use of data, the use of copyrighted material for machine learning purposes may be considered permissible under certain circumstances. For example, if the copyright holder has given permission to use the data through a licensing agreement, then the use for machine learning purposes may be lawful. It is also argued that copying for non-expressive purposes that do not involve conveying expressive elements to the public, such as automated data extraction, should not be seen as infringing. The focus of such ‘copy-reliant technologies’ is on understanding language rules, styles, and syntax and no creative ideas are being used. However, the non-expressive use defense is within the framework of the fair use doctrine, which allows the use of copyrighted material for research or educational purposes. The questions arise because the fair use doctrine is not available in EU law, instead, the InfoSoc Directive provides for a rigid system of exclusive rights with a list of exceptions and limitations. One could only argue that non-expressive uses of copyrighted material for machine learning purposes do not constitute a ‘reproduction’ in the first place. Nevertheless, the use of machine learning with copyrighted material is difficult because EU copyright law applies to the mere use of the works. Two solutions can be proposed to address the problem of copyright clearance for AI training data. The first is to introduce a broad exception for text and data mining, either mandatorily or for commercial and scientific purposes, or to permit the reproduction of works for non-expressive purposes. The second is that copyright laws should permit the reproduction of works for non-expressive purposes, which opens the door to discussions regarding the transposition of the fair use principle from the US into EU law. Both solutions aim to provide more space for AI developers to operate and encourage greater freedom, which could lead to more rapid innovation in the field. The Data Governance Act presents a significant opportunity to advance these debates. Finally, issues concerning the balance of general public interests and legitimate private interests in machine learning training data must be addressed. In my opinion, it is crucial that robot-creation output should fall into the public domain. Machines depend on human creativity, innovation, and expression. To encourage technological advancement and innovation, freedom of expression and business operation must be prioritised.

Keywords: artificial intelligence, copyright, data governance, machine learning

Procedia PDF Downloads 80
2614 Enhancing Email Security: A Multi-Layered Defense Strategy Approach and an AI-Powered Model for Identifying and Mitigating Phishing Attacks

Authors: Anastasios Papathanasiou, George Liontos, Athanasios Katsouras, Vasiliki Liagkou, Euripides Glavas

Abstract:

Email remains a crucial communication tool due to its efficiency, accessibility and cost-effectiveness, enabling rapid information exchange across global networks. However, the global adoption of email has also made it a prime target for cyber threats, including phishing, malware and Business Email Compromise (BEC) attacks, which exploit its integral role in personal and professional realms in order to perform fraud and data breaches. To combat these threats, this research advocates for a multi-layered defense strategy incorporating advanced technological tools such as anti-spam and anti-malware software, machine learning algorithms and authentication protocols. Moreover, we developed an artificial intelligence model specifically designed to analyze email headers and assess their security status. This AI-driven model examines various components of email headers, such as "From" addresses, ‘Received’ paths and the integrity of SPF, DKIM and DMARC records. Upon analysis, it generates comprehensive reports that indicate whether an email is likely to be malicious or benign. This capability empowers users to identify potentially dangerous emails promptly, enhancing their ability to avoid phishing attacks, malware infections and other cyber threats.

Keywords: email security, artificial intelligence, header analysis, threat detection, phishing, DMARC, DKIM, SPF, ai model

Procedia PDF Downloads 44
2613 Elitist Self-Adaptive Step-Size Search in Optimum Sizing of Steel Structures

Authors: Oğuzhan Hasançebi, Saeid Kazemzadeh Azad

Abstract:

This paper covers application of an elitist selfadaptive
step-size search (ESASS) to optimum design of steel
skeletal structures. In the ESASS two approaches are considered for
improving the convergence accuracy as well as the computational
efficiency of the original technique namely the so called selfadaptive
step-size search (SASS). Firstly, an additional randomness
is incorporated into the sampling step of the technique to preserve
exploration capability of the algorithm during the optimization.
Moreover, an adaptive sampling scheme is introduced to improve the
quality of final solutions. Secondly, computational efficiency of the
technique is accelerated via avoiding unnecessary analyses during the
optimization process using an upper bound strategy. The numerical
results demonstrate the usefulness of the ESASS in the sizing
optimization problems of steel truss and frame structures.

Keywords: structural design optimization, optimal sizing, metaheuristics, self-adaptive step-size search, steel trusses, steel frames

Procedia PDF Downloads 365
2612 Application of Computational Chemistry for Searching Anticancer Derivatives of 2-Phenazinamines as Bcr-Abl Tyrosine Kinase Inhibitors

Authors: Gajanan M. Sonwane

Abstract:

The computational studies on 2-phenazinamines with their protein targets have been carried out to design compounds with potential anticancer activity. This strategy of designing compounds possessing selectivity over specific tyrosine kinase has been achieved through G-QSAR and molecular docking studies. The objective of this research has been to design newer 2-phenazinamine derivatives as Bcr-Abl tyrosine kinase inhibitors by G-QSAR, molecular docking studies followed by wet-lab studies along with evaluation of their anticancer potential. Computational chemistry was done by using VLife MDS 4.3 and Autodock 4.2 followed by wet-lab experiments for synthesizing 2-phenazinamine derivatives. The chemical structures of ligands in 2D were drawn by employing Chemdraw 2D Ultra 8.0 and were converted into 3D. These were optimized by using a semi-empirical method called MOPAC. The protein structure was retrieved from RCSC protein data bank as a PDB file. The binding interactions of protein and ligands were done by using PYMOL. The molecular properties of the designed compounds were predicted in silico by using Osiris property explorer. The parent compound 2-phenazinamine was synthesized by reduction of 2, 4-dinitro-N-phenyl-benzenamine in the presence of tin chloride followed by cyclization in the presence of nitrobenzene and magnesium sulfate. The derivatization at the amino function of 2-phenazinamine was performed by treating parent compound with various aldehydes in the presence of dicyclohexylcarbodiimide (DCC) and urea to afford 2-(2-chlorophenyl)-3-(phenazine-2-yl) thiazolidine-4-one. Synthesized 39 novel derivatives of 2-phenazinamine and performed antioxidant activity, anti antiproliferative on the bulb of onion and anticancer activity on cell line showing significant competition with marked blockbuster drug imatinib.

Keywords: computer-aided drug design, tyrosin kinases, anticancer, docking

Procedia PDF Downloads 131
2611 Hardware Implementation and Real-time Experimental Validation of a Direction of Arrival Estimation Algorithm

Authors: Nizar Tayem, AbuMuhammad Moinuddeen, Ahmed A. Hussain, Redha M. Radaydeh

Abstract:

This research paper introduces an approach for estimating the direction of arrival (DOA) of multiple RF noncoherent sources in a uniform linear array (ULA). The proposed method utilizes a Capon-like estimation algorithm and incorporates LU decomposition to enhance the accuracy of DOA estimation while significantly reducing computational complexity compared to existing methods like the Capon method. Notably, the proposed method does not require prior knowledge of the number of sources. To validate its effectiveness, the proposed method undergoes validation through both software simulations and practical experimentation on a prototype testbed constructed using a software-defined radio (SDR) platform and GNU Radio software. The results obtained from MATLAB simulations and real-time experiments provide compelling evidence of the proposed method's efficacy.

Keywords: DOA estimation, real-time validation, software defined radio, computational complexity, Capon's method, GNU radio

Procedia PDF Downloads 66
2610 A Comparative Study on Deep Learning Models for Pneumonia Detection

Authors: Hichem Sassi

Abstract:

Pneumonia, being a respiratory infection, has garnered global attention due to its rapid transmission and relatively high mortality rates. Timely detection and treatment play a crucial role in significantly reducing mortality associated with pneumonia. Presently, X-ray diagnosis stands out as a reasonably effective method. However, the manual scrutiny of a patient's X-ray chest radiograph by a proficient practitioner usually requires 5 to 15 minutes. In situations where cases are concentrated, this places immense pressure on clinicians for timely diagnosis. Relying solely on the visual acumen of imaging doctors proves to be inefficient, particularly given the low speed of manual analysis. Therefore, the integration of artificial intelligence into the clinical image diagnosis of pneumonia becomes imperative. Additionally, AI recognition is notably rapid, with convolutional neural networks (CNNs) demonstrating superior performance compared to human counterparts in image identification tasks. To conduct our study, we utilized a dataset comprising chest X-ray images obtained from Kaggle, encompassing a total of 5216 training images and 624 test images, categorized into two classes: normal and pneumonia. Employing five mainstream network algorithms, we undertook a comprehensive analysis to classify these diseases within the dataset, subsequently comparing the results. The integration of artificial intelligence, particularly through improved network architectures, stands as a transformative step towards more efficient and accurate clinical diagnoses across various medical domains.

Keywords: deep learning, computer vision, pneumonia, models, comparative study

Procedia PDF Downloads 53
2609 Synthetic Classicism: A Machine Learning Approach to the Recognition and Design of Circular Pavilions

Authors: Federico Garrido, Mostafa El Hayani, Ahmed Shams

Abstract:

The exploration of the potential of artificial intelligence (AI) in architecture is still embryonic, however, its latent capacity to change design disciplines is significant. 'Synthetic Classism' is a research project that questions the underlying aspects of classically organized architecture not just in aesthetic terms but also from a geometrical and morphological point of view, intending to generate new architectural information using historical examples as source material. The main aim of this paper is to explore the uses of artificial intelligence and machine learning algorithms in architectural design while creating a coherent narrative to be contained within a design process. The purpose is twofold: on one hand, to develop and train machine learning algorithms to produce architectural information of small pavilions and on the other, to synthesize new information from previous architectural drawings. These algorithms intend to 'interpret' graphical information from each pavilion and then generate new information from it. The procedure, once these algorithms are trained, is the following: parting from a line profile, a synthetic 'front view' of a pavilion is generated, then using it as a source material, an isometric view is created from it, and finally, a top view is produced. Thanks to GAN algorithms, it is also possible to generate Front and Isometric views without any graphical input as well. The final intention of the research is to produce isometric views out of historical information, such as the pavilions from Sebastiano Serlio, James Gibbs, or John Soane. The idea is to create and interpret new information not just in terms of historical reconstruction but also to explore AI as a novel tool in the narrative of a creative design process. This research also challenges the idea of the role of algorithmic design associated with efficiency or fitness while embracing the possibility of a creative collaboration between artificial intelligence and a human designer. Hence the double feature of this research, both analytical and creative, first by synthesizing images based on a given dataset and then by generating new architectural information from historical references. We find that the possibility of creatively understand and manipulate historic (and synthetic) information will be a key feature in future innovative design processes. Finally, the main question that we propose is whether an AI could be used not just to create an original and innovative group of simple buildings but also to explore the possibility of fostering a novel architectural sensibility grounded on the specificities on the architectural dataset, either historic, human-made or synthetic.

Keywords: architecture, central pavilions, classicism, machine learning

Procedia PDF Downloads 136
2608 The National Socialist and Communist Propaganda Activities in the Turkish Press during the World War II

Authors: Asuman Tezcan Mirer

Abstract:

This proposed paper discusses nationalist socialist and communist propaganda struggles in the Turkish press during World War II. The paper aspires to analyze how government agencies directed and organized the Turkish press to prevent the "5th column" from influencing public opinion. During the Second World War, one of the most emphasized issues was propaganda and how Turkish citizens would be protected from the effects of disinformation. Istanbul became a significant headquarters for belligerent countries' intelligence services, and these services were involved in gathering intelligence and disseminating propaganda. The main motive of national socialist propaganda was "anti-communism" in Turkey. Subsidizing certain magazines, controlling German companies' advertisements and paper trade, spreading rumors, printing propaganda brochures, and showing German propaganda films are some tactics that the nationalist socialists applied before and during the Second World War. On the other hand, the communists targeted Turkish racist/ultra-nationalist groups and their publications, which were influenced by the Nazi regime. They were also involved in distributing Marxist publications, printing brochures, and broadcasting radio programs. This study composes of three parts. The first part describes the nationalist socialist and communist propaganda activities in Turkey during the Second World War. The second part addresses the debates over propaganda among selected newspapers representing different ideologies. Finally, the last part analyzes the Turkish government's press policy. It explains why the government allowed ideological debates in the press despite its authoritarian press policy and "active neutrality" stance in the international arena.

Keywords: propaganda, press, 5th column, World War II, Turkey

Procedia PDF Downloads 95
2607 A Review of Attractor Neural Networks and Their Use in Cognitive Science

Authors: Makenzy Lee Gilbert

Abstract:

This literature review explores the role of attractor neural networks (ANNs) in modeling psychological processes in artificial and biological systems. By synthesizing research from dynamical systems theory, psychology, and computational neuroscience, the review provides an overview of the current understanding of ANN function in memory formation, reinforcement, retrieval, and forgetting. Key mathematical foundations, including dynamical systems theory and energy functions, are discussed to explain the behavior and stability of these networks. The review also examines empirical applications of ANNs in cognitive processes such as semantic memory and episodic recall, as well as highlighting the hippocampus's role in pattern separation and completion. The review addresses challenges like catastrophic forgetting and noise effects on memory retrieval. By identifying gaps between theoretical models and empirical findings, it highlights the interdisciplinary nature of ANN research and suggests future exploration areas.

Keywords: attractor neural networks, connectionism, computational modeling, cognitive neuroscience

Procedia PDF Downloads 15
2606 Integrated Free Space Optical Communication and Optical Sensor Network System with Artificial Intelligence Techniques

Authors: Yibeltal Chanie Manie, Zebider Asire Munyelet

Abstract:

5G and 6G technology offers enhanced quality of service with high data transmission rates, which necessitates the implementation of the Internet of Things (IoT) in 5G/6G architecture. In this paper, we proposed the integration of free space optical communication (FSO) with fiber sensor networks for IoT applications. Recently, free-space optical communications (FSO) are gaining popularity as an effective alternative technology to the limited availability of radio frequency (RF) spectrum. FSO is gaining popularity due to flexibility, high achievable optical bandwidth, and low power consumption in several applications of communications, such as disaster recovery, last-mile connectivity, drones, surveillance, backhaul, and satellite communications. Hence, high-speed FSO is an optimal choice for wireless networks to satisfy the full potential of 5G/6G technology, offering 100 Gbit/s or more speed in IoT applications. Moreover, machine learning must be integrated into the design, planning, and optimization of future optical wireless communication networks in order to actualize this vision of intelligent processing and operation. In addition, fiber sensors are important to achieve real-time, accurate, and smart monitoring in IoT applications. Moreover, we proposed deep learning techniques to estimate the strain changes and peak wavelength of multiple Fiber Bragg grating (FBG) sensors using only the spectrum of FBGs obtained from the real experiment.

Keywords: optical sensor, artificial Intelligence, Internet of Things, free-space optics

Procedia PDF Downloads 55
2605 A Look at the Quantum Theory of Atoms in Molecules from the Discrete Morse Theory

Authors: Dairo Jose Hernandez Paez

Abstract:

The quantum theory of atoms in molecules (QTAIM) allows us to obtain topological information on electronic density in quantum mechanical systems. The QTAIM starts by considering the electron density as a continuous mathematical object. On the other hand, the discretization of electron density is also a mathematical object, which, from discrete mathematics, would allow a new approach to its topological study. From this point of view, it is necessary to develop a series of steps that provide the theoretical support that guarantees its application. Some of the steps that we consider most important are mentioned below: (1) obtain good representations of the electron density through computational calculations, (2) design a methodology for the discretization of electron density, and construct the simplicial complex. (3) Make an analysis of the discrete vector field associating the simplicial complex. (4) Finally, in this research, we propose to use the discrete Morse theory as a mathematical tool to carry out studies of electron density topology.

Keywords: discrete mathematics, Discrete Morse theory, electronic density, computational calculations

Procedia PDF Downloads 96
2604 Parameters Optimization of the Laminated Composite Plate for Sound Transmission Problem

Authors: Yu T. Tsai, Jin H. Huang

Abstract:

In this paper, the specific sound transmission loss (TL) of the laminated composite plate (LCP) with different material properties in each layer is investigated. The numerical method to obtain the TL of the LCP is proposed by using elastic plate theory. The transfer matrix approach is novelty presented for computational efficiency in solving the numerous layers of dynamic stiffness matrix (D-matrix) of the LCP. Besides the numerical simulations for calculating the TL of the LCP, the material properties inverse method is presented for the design of a laminated composite plate analogous to a metallic plate with a specified TL. As a result, it demonstrates that the proposed computational algorithm exhibits high efficiency with a small number of iterations for achieving the goal. This method can be effectively employed to design and develop tailor-made materials for various applications.

Keywords: sound transmission loss, laminated composite plate, transfer matrix approach, inverse problem, elastic plate theory, material properties

Procedia PDF Downloads 378
2603 Computational Fluid Dynamics and Experimental Evaluation of Two Batch Type Electrocoagulation Stirred Tank Reactors Used in the Removal of Cr (VI) from Waste Water

Authors: Phanindra Prasad Thummala, Umran Tezcan Un

Abstract:

In this study, hydrodynamics analysis of two batch type electrocoagulation stirred tank reactors, used for the electrocoagulation treatment of Cr(VI) wastewater, was carried using computational fluid dynamics (CFD). The aim of the study was to evaluate the impact of mixing characteristics on overall performance of electrocoagulation reactor. The CFD simulations were performed using ANSYS FLUENT 14.4 software. The mixing performance of each reactor was evaluated by numerically modelling tracer dispersion in each reactor configuration. The uniformity in tracer dispersion was assumed when 90% of the ratio of the maximum to minimum concentration of the tracer was realized. In parallel, experimental evaluation of both the electrocoagulation reactors for removal of Cr(VI) from wastewater was also carried out. The results of CFD and experimental analysis clearly show that the reactor which can give higher uniformity in lesser time, will perform better as an electrocoagulation reactor for removal of Cr(VI) from wastewater.

Keywords: CFD, stirred tank reactors, electrocoagulation, Cr(VI) wastewater

Procedia PDF Downloads 458
2602 Process Data-Driven Representation of Abnormalities for Efficient Process Control

Authors: Hyun-Woo Cho

Abstract:

Unexpected operational events or abnormalities of industrial processes have a serious impact on the quality of final product of interest. In terms of statistical process control, fault detection and diagnosis of processes is one of the essential tasks needed to run the process safely. In this work, nonlinear representation of process measurement data is presented and evaluated using a simulation process. The effect of using different representation methods on the diagnosis performance is tested in terms of computational efficiency and data handling. The results have shown that the nonlinear representation technique produced more reliable diagnosis results and outperforms linear methods. The use of data filtering step improved computational speed and diagnosis performance for test data sets. The presented scheme is different from existing ones in that it attempts to extract the fault pattern in the reduced space, not in the original process variable space. Thus this scheme helps to reduce the sensitivity of empirical models to noise.

Keywords: fault diagnosis, nonlinear technique, process data, reduced spaces

Procedia PDF Downloads 244
2601 An Improved Mesh Deformation Method Based on Radial Basis Function

Authors: Xuan Zhou, Litian Zhang, Shuixiang Li

Abstract:

Mesh deformation using radial basis function interpolation method has been demonstrated to produce quality meshes with relatively little computational cost using a concise algorithm. However, it still suffers from the limited deformation ability, especially in large deformation. In this paper, a pre-displacement improvement is proposed to improve the problem that illegal meshes always appear near the moving inner boundaries owing to the large relative displacement of the nodes near inner boundaries. In this improvement, nodes near the inner boundaries are first associated to the near boundary nodes, and a pre-displacement based on the displacements of associated boundary nodes is added to the nodes near boundaries in order to make the displacement closer to the boundary deformation and improve the deformation capability. Several 2D and 3D numerical simulation cases have shown that the pre-displacement improvement for radial basis function (RBF) method significantly improves the mesh quality near inner boundaries and deformation capability, with little computational burden increasement.

Keywords: mesh deformation, mesh quality, background mesh, radial basis function

Procedia PDF Downloads 359
2600 Data Analytics in Energy Management

Authors: Sanjivrao Katakam, Thanumoorthi I., Antony Gerald, Ratan Kulkarni, Shaju Nair

Abstract:

With increasing energy costs and its impact on the business, sustainability today has evolved from a social expectation to an economic imperative. Therefore, finding methods to reduce cost has become a critical directive for Industry leaders. Effective energy management is the only way to cut costs. However, Energy Management has been a challenge because it requires a change in old habits and legacy systems followed for decades. Today exorbitant levels of energy and operational data is being captured and stored by Industries, but they are unable to convert these structured and unstructured data sets into meaningful business intelligence. It must be noted that for quick decisions, organizations must learn to cope with large volumes of operational data in different formats. Energy analytics not only helps in extracting inferences from these data sets, but also is instrumental in transformation from old approaches of energy management to new. This in turn assists in effective decision making for implementation. It is the requirement of organizations to have an established corporate strategy for reducing operational costs through visibility and optimization of energy usage. Energy analytics play a key role in optimization of operations. The paper describes how today energy data analytics is extensively used in different scenarios like reducing operational costs, predicting energy demands, optimizing network efficiency, asset maintenance, improving customer insights and device data insights. The paper also highlights how analytics helps transform insights obtained from energy data into sustainable solutions. The paper utilizes data from an array of segments such as retail, transportation, and water sectors.

Keywords: energy analytics, energy management, operational data, business intelligence, optimization

Procedia PDF Downloads 360
2599 Designing of Tooling Solution for Material Handling in Highly Automated Manufacturing System

Authors: Muhammad Umair, Yuri Nikolaev, Denis Artemov, Ighor Uzhinsky

Abstract:

A flexible manufacturing system is an integral part of a smart factory of industry 4.0 in which every machine is interconnected and works autonomously. Robots are in the process of replacing humans in every industrial sector. As the cyber-physical-system (CPS) and artificial intelligence (AI) are advancing, the manufacturing industry is getting more dependent on computers than human brains. This modernization has boosted the production with high quality and accuracy and shifted from classic production to smart manufacturing systems. However, material handling for such automated productions is a challenge and needs to be addressed with the best possible solution. Conventional clamping systems are designed for manual work and not suitable for highly automated production systems. Researchers and engineers are trying to find the most economical solution for loading/unloading and transportation workpieces from a warehouse to a machine shop for machining operations and back to the warehouse without human involvement. This work aims to propose an advanced multi-shape tooling solution for highly automated manufacturing systems. The currently obtained result shows that it could function well with automated guided vehicles (AGVs) and modern conveyor belts. The proposed solution is following requirements to be automation-friendly, universal for different part geometry and production operations. We used a bottom-up approach in this work, starting with studying different case scenarios and their limitations and finishing with the general solution.

Keywords: artificial intelligence, cyber physics system, Industry 4.0, material handling, smart factory, flexible manufacturing system

Procedia PDF Downloads 126
2598 Establishing a Computational Screening Framework to Identify Environmental Exposures Using Untargeted Gas-Chromatography High-Resolution Mass Spectrometry

Authors: Juni C. Kim, Anna R. Robuck, Douglas I. Walker

Abstract:

The human exposome, which includes chemical exposures over the lifetime and their effects, is now recognized as an important measure for understanding human health; however, the complexity of the data makes the identification of environmental chemicals challenging. The goal of our project was to establish a computational workflow for the improved identification of environmental pollutants containing chlorine or bromine. Using the “pattern. search” function available in the R package NonTarget, we wrote a multifunctional script that searches mass spectral clusters from untargeted gas-chromatography high-resolution mass spectrometry (GC-HRMS) for the presence of spectra consistent with chlorine and bromine-containing organic compounds. The “pattern. search” function was incorporated into a different function that allows the evaluation of clusters containing multiple analyte fragments, has multi-core support, and provides a simplified output identifying listing compounds containing chlorine and/or bromine. The new function was able to process 46,000 spectral clusters in under 8 seconds and identified over 150 potential halogenated spectra. We next applied our function to a deidentified dataset from patients diagnosed with primary biliary cholangitis (PBC), primary sclerosing cholangitis (PSC), and healthy controls. Twenty-two spectra corresponded to potential halogenated compounds in the PSC and PBC dataset, including six significantly different in PBC patients, while four differed in PSC patients. We have developed an improved algorithm for detecting halogenated compounds in GC-HRMS data, providing a strategy for prioritizing exposures in the study of human disease.

Keywords: exposome, metabolome, computational metabolomics, high-resolution mass spectrometry, exposure, pollutants

Procedia PDF Downloads 131
2597 Computational Analysis of Potential Inhibitors Selected Based on Structural Similarity for the Src SH2 Domain

Authors: W. P. Hu, J. V. Kumar, Jeffrey J. P. Tsai

Abstract:

The inhibition of SH2 domain regulated protein-protein interactions is an attractive target for developing an effective chemotherapeutic approach in the treatment of disease. Molecular simulation is a useful tool for developing new drugs and for studying molecular recognition. In this study, we searched potential drug compounds for the inhibition of SH2 domain by performing structural similarity search in PubChem Compound Database. A total of 37 compounds were screened from the database, and then we used the LibDock docking program to evaluate the inhibition effect. The best three compounds (AP22408, CID 71463546 and CID 9917321) were chosen for MD simulations after the LibDock docking. Our results show that the compound CID 9917321 can produce a more stable protein-ligand complex compared to other two currently known inhibitors of Src SH2 domain. The compound CID 9917321 may be useful for the inhibition of SH2 domain based on these computational results. Subsequently experiments are needed to verify the effect of compound CID 9917321 on the SH2 domain in the future studies.

Keywords: nonpeptide inhibitor, Src SH2 domain, LibDock, molecular dynamics simulation

Procedia PDF Downloads 264
2596 A Rapid Prototyping Tool for Suspended Biofilm Growth Media

Authors: Erifyli Tsagkari, Stephanie Connelly, Zhaowei Liu, Andrew McBride, William Sloan

Abstract:

Biofilms play an essential role in treating water in biofiltration systems. The biofilm morphology and function are inextricably linked to the hydrodynamics of flow through a filter, and yet engineers rarely explicitly engineer this interaction. We develop a system that links computer simulation and 3-D printing to optimize and rapidly prototype filter media to optimize biofilm function with the hypothesis that biofilm function is intimately linked to the flow passing through the filter. A computational model that numerically solves the incompressible time-dependent Navier Stokes equations coupled to a model for biofilm growth and function is developed. The model is imbedded in an optimization algorithm that allows the model domain to adapt until criteria on biofilm functioning are met. This is applied to optimize the shape of filter media in a simple flow channel to promote biofilm formation. The computer code links directly to a 3-D printer, and this allows us to prototype the design rapidly. Its validity is tested in flow visualization experiments and by microscopy. As proof of concept, the code was constrained to explore a small range of potential filter media, where the medium acts as an obstacle in the flow that sheds a von Karman vortex street that was found to enhance the deposition of bacteria on surfaces downstream. The flow visualization and microscopy in the 3-D printed realization of the flow channel validated the predictions of the model and hence its potential as a design tool. Overall, it is shown that the combination of our computational model and the 3-D printing can be effectively used as a design tool to prototype filter media to optimize biofilm formation.

Keywords: biofilm, biofilter, computational model, von karman vortices, 3-D printing.

Procedia PDF Downloads 137
2595 Computational Material Modeling for Mechanical Properties Prediction of Nanoscale Carbon Based Cementitious Materials

Authors: Maryam Kiani, Abdul Basit Kiani

Abstract:

At larger scales, the performance of cementitious materials is impacted by processes occurring at the nanometer scale. These materials boast intricate hierarchical structures with random features that span from the nanometer to millimeter scale. It is fascinating to observe how the nanoscale processes influence the overall behavior and characteristics of these materials. By delving into and manipulating these processes, scientists and engineers can unlock the potential to create more durable and sustainable infrastructure and construction materials. It's like unraveling a hidden tapestry of secrets that hold the key to building stronger and more resilient structures. The present work employs simulations as the computational modeling methodology to predict mechanical properties for carbon/silica based cementitious materials at the molecular/nano scale level. Studies focused on understanding the effect of higher mechanical properties of cementitious materials with carbon silica nanoparticles via Material Studio materials modeling.

Keywords: nanomaterials, SiO₂, carbon black, mechanical properties

Procedia PDF Downloads 132
2594 The Protection of Artificial Intelligence (AI)-Generated Creative Works Through Authorship: A Comparative Analysis Between the UK and Nigerian Copyright Experience to Determine Lessons to Be Learnt from the UK

Authors: Esther Ekundayo

Abstract:

The nature of AI-generated works makes it difficult to identify an author. Although, some scholars have suggested that all the players involved in its creation should be allocated authorship according to their respective contribution. From the programmer who creates and designs the AI to the investor who finances the AI and to the user of the AI who most likely ends up creating the work in question. While others suggested that this issue may be resolved by the UK computer-generated works (CGW) provision under Section 9(3) of the Copyright Designs and Patents Act 1988. However, under the UK and Nigerian copyright law, only human-created works are recognised. This is usually assessed based on their originality. This simply means that the work must have been created as a result of its author’s creative and intellectual abilities and not copied. Such works are literary, dramatic, musical and artistic works and are those that have recently been a topic of discussion with regards to generative artificial intelligence (Generative AI). Unlike Nigeria, the UK CDPA recognises computer-generated works and vests its authorship with the human who made the necessary arrangement for its creation . However, making necessary arrangement in the case of Nova Productions Ltd v Mazooma Games Ltd was interpreted similarly to the traditional authorship principle, which requires the skills of the creator to prove originality. Although, some recommend that computer-generated works complicates this issue, and AI-generated works should enter the public domain as authorship cannot be allocated to AI itself. Additionally, the UKIPO recognising these issues in line with the growing AI trend in a public consultation launched in the year 2022, considered whether computer-generated works should be protected at all and why. If not, whether a new right with a different scope and term of protection should be introduced. However, it concluded that the issue of computer-generated works would be revisited as AI was still in its early stages. Conversely, due to the recent developments in this area with regards to Generative AI systems such as ChatGPT, Midjourney, DALL-E and AIVA, amongst others, which can produce human-like copyright creations, it is therefore important to examine the relevant issues which have the possibility of altering traditional copyright principles as we know it. Considering that the UK and Nigeria are both common law jurisdictions but with slightly differing approaches to this area, this research, therefore, seeks to answer the following questions by comparative analysis: 1)Who is the author of an AI-generated work? 2)Is the UK’s CGW provision worthy of emulation by the Nigerian law? 3) Would a sui generis law be capable of protecting AI-generated works and its author under both jurisdictions? This research further examines the possible barriers to the implementation of the new law in Nigeria, such as limited technical expertise and lack of awareness by the policymakers, amongst others.

Keywords: authorship, artificial intelligence (AI), generative ai, computer-generated works, copyright, technology

Procedia PDF Downloads 81
2593 Impact of Research-Informed Teaching and Case-Based Teaching on Memory Retention and Recall in University Students

Authors: Durvi Yogesh Vagani

Abstract:

This research paper explores the effectiveness of Research-informed teaching and Case-based teaching in enhancing the retention and recall of memory during discussions among university students. Additionally, it investigates the impact of using Artificial Intelligence (AI) tools on the quality of research conducted by students and its correlation with better recollection. The study hypothesizes that Case-based teaching will lead to greater recall and storage of information. The research gap in the use of AI in educational settings, particularly with actual participants, is addressed by leveraging a multi-method approach. The hypothesis is that the use of AI, such as ChatGPT and Bard, would lead to better retention and recall of information. Before commencing the study, participants' attention levels and IQ were assessed using the Digit Span Test and the Wechsler Adult Intelligence Scale, respectively, to ensure comparability among participants. Subsequently, participants were divided into four conditions, each group receiving identical information presented in different formats based on their assigned condition. Following this, participants engaged in a group discussion on the given topic. Their responses were then evaluated against a checklist. Finally, participants completed a brief test to measure their recall ability after the discussion. Preliminary findings suggest that students who utilize AI tools for learning demonstrate improved grasping of information and are more likely to integrate relevant information into discussions compared to providing extraneous details. Furthermore, Case-based teaching fosters greater attention and recall during discussions, while Research-informed teaching leads to greater knowledge for application. By addressing the research gap in AI application in education, this study contributes to a deeper understanding of effective teaching methodologies and the role of technology in student learning outcomes. The implication of the present research is to tailor teaching methods based on the subject matter. Case-based teaching facilitates application-based teaching, and research-based teaching can be beneficial for theory-heavy topics. Integrating AI in education. Combining AI with research-based teaching may optimize instructional strategies and deepen learning experiences. This research suggests tailoring teaching methods in psychology based on subject matter. Case-based teaching suits practical subjects, facilitating application, while research-based teaching aids understanding of theory-heavy topics. Integrating AI in education could enhance learning outcomes, offering detailed information tailored to students' needs.

Keywords: artificial intelligence, attention, case-based teaching, memory recall, memory retention, research-informed teaching

Procedia PDF Downloads 10
2592 Artificial Intelligence-Generated Previews of Hyaluronic Acid-Based Treatments

Authors: Ciro Cursio, Giulia Cursio, Pio Luigi Cursio, Luigi Cursio

Abstract:

Communication between practitioner and patient is of the utmost importance in aesthetic medicine: as of today, images of previous treatments are the most common tool used by doctors to describe and anticipate future results for their patients. However, using photos of other people often reduces the engagement of the prospective patient and is further limited by the number and quality of pictures available to the practitioner. Pre-existing work solves this issue in two ways: 3D scanning of the area with manual editing of the 3D model by the doctor or automatic prediction of the treatment by warping the image with hand-written parameters. The first approach requires the manual intervention of the doctor, while the second approach always generates results that aren’t always realistic. Thus, in one case, there is significant manual work required by the doctor, and in the other case, the prediction looks artificial. We propose an AI-based algorithm that autonomously generates a realistic prediction of treatment results. For the purpose of this study, we focus on hyaluronic acid treatments in the facial area. Our approach takes into account the individual characteristics of each face, and furthermore, the prediction system allows the patient to decide which area of the face she wants to modify. We show that the predictions generated by our system are realistic: first, the quality of the generated images is on par with real images; second, the prediction matches the actual results obtained after the treatment is completed. In conclusion, the proposed approach provides a valid tool for doctors to show patients what they will look like before deciding on the treatment.

Keywords: prediction, hyaluronic acid, treatment, artificial intelligence

Procedia PDF Downloads 109
2591 Design and Implementation of Low-code Model-building Methods

Authors: Zhilin Wang, Zhihao Zheng, Linxin Liu

Abstract:

This study proposes a low-code model-building approach that aims to simplify the development and deployment of artificial intelligence (AI) models. With an intuitive way to drag and drop and connect components, users can easily build complex models and integrate multiple algorithms for training. After the training is completed, the system automatically generates a callable model service API. This method not only lowers the technical threshold of AI development and improves development efficiency but also enhances the flexibility of algorithm integration and simplifies the deployment process of models. The core strength of this method lies in its ease of use and efficiency. Users do not need to have a deep programming background and can complete the design and implementation of complex models with a simple drag-and-drop operation. This feature greatly expands the scope of AI technology, allowing more non-technical people to participate in the development of AI models. At the same time, the method performs well in algorithm integration, supporting many different types of algorithms to work together, which further improves the performance and applicability of the model. In the experimental part, we performed several performance tests on the method. The results show that compared with traditional model construction methods, this method can make more efficient use, save computing resources, and greatly shorten the model training time. In addition, the system-generated model service interface has been optimized for high availability and scalability, which can adapt to the needs of different application scenarios.

Keywords: low-code, model building, artificial intelligence, algorithm integration, model deployment

Procedia PDF Downloads 16
2590 Effective Method of Paneling for Source/Vortex/Doublet Panel Methods Using Conformal Mapping

Authors: K. C. R. Perera, B. M. Hapuwatte

Abstract:

This paper presents an effective method to divide panels for mesh-less methods of source, vortex and doublet panel methods. In this research study the physical domain of air-foils were transformed into computational domain of a circle using conformal mapping technique of Joukowsky transformation. Then the circle is divided into panels of equal length and the co-ordinates were remapped into physical domain of the air-foil. With this method the leading edge and the trailing edge of the air-foil is panelled with a high density of panels and the rest of the body is panelled with low density of panels. The high density of panels in the leading edge and the trailing edge will increase the accuracy of the solutions obtained from panel methods where the fluid flow at the leading and trailing edges are complex.

Keywords: conformal mapping, Joukowsky transformation, physical domain, computational domain

Procedia PDF Downloads 371
2589 Numerical and Experimental Study on Bed-Wall Heat Transfer in Conical Fluidized Bed Combustor

Authors: Ik–Tae Im, H. M. Abdelmotalib, M. A. Youssef, S. B. Young

Abstract:

In this study the flow characteristics and bed-to-wall heat transfer in a gas-solid conical fluidized bed combustor were investigated using both experimental and numerical methods. The computational fluid dynamic (CFD) simulations were carried out using a commercial software, Fluent V6.3. A two-fluid Eulerian-Eulerian model was applied in order to simulate the gas–solid flow and heat transfer in a conical sand-air bed with 30o con angle and 22 cm static bed height. Effect of different fluidizing number varying in the range of 1.5 - 2.3, drag models namely (Syamlal-O’Brien and Gidaspow), and friction viscosity on flow and bed-to-wall heat transfer were analyzed. Both bed pressure drop and heat transfer coefficient increased with increasing inlet gas velocity. The Gidaspow drag model showed a better agreement with experimental results than other drag model. The friction viscosity had no clear effect on both hydrodynamics and heat transfer.

Keywords: computational fluid dynamics, heat transfer coefficient, hydrodynamics, renewable energy

Procedia PDF Downloads 404
2588 A Geometrical Multiscale Approach to Blood Flow Simulation: Coupling 2-D Navier-Stokes and 0-D Lumped Parameter Models

Authors: Azadeh Jafari, Robert G. Owens

Abstract:

In this study, a geometrical multiscale approach which means coupling together the 2-D Navier-Stokes equations, constitutive equations and 0-D lumped parameter models is investigated. A multiscale approach, suggest a natural way of coupling detailed local models (in the flow domain) with coarser models able to describe the dynamics over a large part or even the whole cardiovascular system at acceptable computational cost. In this study we introduce a new velocity correction scheme to decouple the velocity computation from the pressure one. To evaluate the capability of our new scheme, a comparison between the results obtained with Neumann outflow boundary conditions on the velocity and Dirichlet outflow boundary conditions on the pressure and those obtained using coupling with the lumped parameter model has been performed. Comprehensive studies have been done based on the sensitivity of numerical scheme to the initial conditions, elasticity and number of spectral modes. Improvement of the computational algorithm with stable convergence has been demonstrated for at least moderate Weissenberg number. We comment on mathematical properties of the reduced model, its limitations in yielding realistic and accurate numerical simulations, and its contribution to a better understanding of microvascular blood flow. We discuss the sophistication and reliability of multiscale models for computing correct boundary conditions at the outflow boundaries of a section of the cardiovascular system of interest. In this respect the geometrical multiscale approach can be regarded as a new method for solving a class of biofluids problems, whose application goes significantly beyond the one addressed in this work.

Keywords: geometrical multiscale models, haemorheology model, coupled 2-D navier-stokes 0-D lumped parameter modeling, computational fluid dynamics

Procedia PDF Downloads 354