Search results for: artificial intelligence based optimization
30406 Optimal Production and Maintenance Policy for a Partially Observable Production System with Stochastic Demand
Authors: Leila Jafari, Viliam Makis
Abstract:
In this paper, the joint optimization of the economic manufacturing quantity (EMQ), safety stock level, and condition-based maintenance (CBM) is presented for a partially observable, deteriorating system subject to random failure. The demand is stochastic and it is described by a Poisson process. The stochastic model is developed and the optimization problem is formulated in the semi-Markov decision process framework. A modification of the policy iteration algorithm is developed to find the optimal policy. A numerical example is presented to compare the optimal policy with the policy considering zero safety stock.Keywords: condition-based maintenance, economic manufacturing quantity, safety stock, stochastic demand
Procedia PDF Downloads 47030405 Relation between Initial Stability of the Dental Implant and Bone-Implant Contact Level
Authors: Jui-Ting Hsu, Heng-Li Huang, Ming-Tzu Tsai, Kuo-Chih Su, Lih-Jyh Fuh
Abstract:
The objectives of this study were to measure the initial stability of the dental implant (ISQ and PTV) in the artificial foam bone block with three different quality levels. In addition, the 3D bone to implant contact percentage (BIC%) was measured based on the micro-computed tomography images. Furthermore, the relation between the initial stability of dental implant (ISQ and PTV) and BIC% were calculated. The experimental results indicated that enhanced the material property of the artificial foam bone increased the initial stability of the dental implant. The Pearson’s correlation coefficient between the BIC% and the two approaches (ISQ and PTV) were 0.652 and 0.745.Keywords: dental implant, implant stability quotient, peak insertion torque, bone-implant contact, micro-computed tomography
Procedia PDF Downloads 58430404 Hierarchical Optimization of Composite Deployable Bridge Treadway Using Particle Swarm Optimization
Authors: Ashraf Osman
Abstract:
Effective deployable bridges that are characterized by an increased capacity to weight ratio are recently needed for post-disaster rapid mobility and military operations. In deployable bridging, replacing metals as the fabricating material with advanced composite laminates as lighter alternatives with higher strength is highly advantageous. This article presents a hierarchical optimization strategy of a composite bridge treadway considering maximum strength design and bridge weight minimization. Shape optimization of a generic deployable bridge beam cross-section is performed to achieve better stress distribution over the bridge treadway hull. The developed cross-section weight is minimized up to reserving the margins of safety of the deployable bridging code provisions. Hence, the strength of composite bridge plates is maximized through varying the plies orientation. Different loading cases are considered of a tracked vehicle patch load. The orthotropic plate properties of a composite sandwich core are used to simulate the bridge deck structural behavior. Whereas, the failure analysis is conducted using Tsai-Wu failure criterion. The naturally inspired particle swarm optimization technique is used in this study. The proposed technique efficiently reduced the weight to capacity ratio of the developed bridge beam.Keywords: CFRP deployable bridges, disaster relief, military bridging, optimization of composites, particle swarm optimization
Procedia PDF Downloads 14330403 'Low Electronic Noise' Detector Technology in Computed Tomography
Authors: A. Ikhlef
Abstract:
Image noise in computed tomography, is mainly caused by the statistical noise, system noise reconstruction algorithm filters. Since last few years, low dose x-ray imaging became more and more desired and looked as a technical differentiating technology among CT manufacturers. In order to achieve this goal, several technologies and techniques are being investigated, including both hardware (integrated electronics and photon counting) and software (artificial intelligence and machine learning) based solutions. From a hardware point of view, electronic noise could indeed be a potential driver for low and ultra-low dose imaging. We demonstrated that the reduction or elimination of this term could lead to a reduction of dose without affecting image quality. Also, in this study, we will show that we can achieve this goal using conventional electronics (low cost and affordable technology), designed carefully and optimized for maximum detective quantum efficiency. We have conducted the tests using large imaging objects such as 30 cm water and 43 cm polyethylene phantoms. We compared the image quality with conventional imaging protocols with radiation as low as 10 mAs (<< 1 mGy). Clinical validation of such results has been performed as well.Keywords: computed tomography, electronic noise, scintillation detector, x-ray detector
Procedia PDF Downloads 13030402 Application of Artificial Neural Network and Background Subtraction for Determining Body Mass Index (BMI) in Android Devices Using Bluetooth
Authors: Neil Erick Q. Madariaga, Noel B. Linsangan
Abstract:
Body Mass Index (BMI) is one of the different ways to monitor the health of a person. It is based on the height and weight of the person. This study aims to compute for the BMI using an Android tablet by obtaining the height of the person by using a camera and measuring the weight of the person by using a weighing scale or load cell. The height of the person was estimated by applying background subtraction to the image captured and applying different processes such as getting the vanishing point and applying Artificial Neural Network. The weight was measured by using Wheatstone bridge load cell configuration and sending the value to the computer by using Gizduino microcontroller and Bluetooth technology after the amplification using AD620 instrumentation amplifier. The application will process the images and read the measured values and show the BMI of the person. The study met all the objectives needed and further studies will be needed to improve the design project.Keywords: body mass index, artificial neural network, vanishing point, bluetooth, wheatstone bridge load cell
Procedia PDF Downloads 32830401 Minimum Half Power Beam Width and Side Lobe Level Reduction of Linear Antenna Array Using Particle Swarm Optimization
Authors: Saeed Ur Rahman, Naveed Ullah, Muhammad Irshad Khan, Quensheng Cao, Niaz Muhammad Khan
Abstract:
In this paper the optimization performance of non-uniform linear antenna array is presented. The Particle Swarm Optimization (PSO) algorithm is presented to minimize Side Lobe Level (SLL) and Half Power Beamwidth (HPBW). The purpose of using the PSO algorithm is to get the optimum values for inter-element spacing and excitation amplitude of linear antenna array that provides a radiation pattern with minimum SLL and HPBW. Various design examples are considered and the obtain results using PSO are confirmed by comparing with results achieved using other nature inspired metaheuristic algorithms such as real coded genetic algorithm (RGA) and biogeography (BBO) algorithm. The comparative results show that optimization of linear antenna array using the PSO provides considerable enhancement in the SLL and HPBW.Keywords: linear antenna array, minimum side lobe level, narrow half power beamwidth, particle swarm optimization
Procedia PDF Downloads 55630400 Urban Planning Compilation Problems in China and the Corresponding Optimization Ideas under the Vision of the Hyper-Cycle Theory
Authors: Hong Dongchen, Chen Qiuxiao, Wu Shuang
Abstract:
Systematic science reveals the complex nonlinear mechanisms of behaviour in urban system. However, in China, when the current city planners face with the system, most of them are still taking simple linear thinking to consider the open complex giant system. This paper introduces the hyper-cycle theory, which is one of the basis theories of systematic science, based on the analysis of the reasons why the current urban planning failed, and proposals for optimization ideas that urban planning compilation should change, from controlling quantitative to the changes of relationship, from blueprint planning to progressive planning based on the nonlinear characteristics and from management control to dynamically monitor feedback.Keywords: systematic science, hyper-cycle theory, urban planning, urban management
Procedia PDF Downloads 41330399 Reliable Soup: Reliable-Driven Model Weight Fusion on Ultrasound Imaging Classification
Authors: Shuge Lei, Haonan Hu, Dasheng Sun, Huabin Zhang, Kehong Yuan, Jian Dai, Yan Tong
Abstract:
It remains challenging to measure reliability from classification results from different machine learning models. This paper proposes a reliable soup optimization algorithm based on the model weight fusion algorithm Model Soup, aiming to improve reliability by using dual-channel reliability as the objective function to fuse a series of weights in the breast ultrasound classification models. Experimental results on breast ultrasound clinical datasets demonstrate that reliable soup significantly enhances the reliability of breast ultrasound image classification tasks. The effectiveness of the proposed approach was verified via multicenter trials. The results from five centers indicate that the reliability optimization algorithm can enhance the reliability of the breast ultrasound image classification model and exhibit low multicenter correlation.Keywords: breast ultrasound image classification, feature attribution, reliability assessment, reliability optimization
Procedia PDF Downloads 8930398 Modelling Mode Choice Behaviour Using Cloud Theory
Authors: Leah Wright, Trevor Townsend
Abstract:
Mode choice models are crucial instruments in the analysis of travel behaviour. These models show the relationship between an individual’s choice of transportation mode for a given O-D pair and the individual’s socioeconomic characteristics such as household size and income level, age and/or gender, and the features of the transportation system. The most popular functional forms of these models are based on Utility-Based Choice Theory, which addresses the uncertainty in the decision-making process with the use of an error term. However, with the development of artificial intelligence, many researchers have started to take a different approach to travel demand modelling. In recent times, researchers have looked at using neural networks, fuzzy logic and rough set theory to develop improved mode choice formulas. The concept of cloud theory has recently been introduced to model decision-making under uncertainty. Unlike the previously mentioned theories, cloud theory recognises a relationship between randomness and fuzziness, two of the most common types of uncertainty. This research aims to investigate the use of cloud theory in mode choice models. This paper highlights the conceptual framework of the mode choice model using cloud theory. Merging decision-making under uncertainty and mode choice models is state of the art. The cloud theory model is expected to address the issues and concerns with the nested logit and improve the design of mode choice models and their use in travel demand.Keywords: Cloud theory, decision-making, mode choice models, travel behaviour, uncertainty
Procedia PDF Downloads 39030397 A New Approach to Predicting Physical Biometrics from Behavioural Biometrics
Authors: Raid R. O. Al-Nima, S. S. Dlay, W. L. Woo
Abstract:
A relationship between face and signature biometrics is established in this paper. A new approach is developed to predict faces from signatures by using artificial intelligence. A multilayer perceptron (MLP) neural network is used to generate face details from features extracted from signatures, here face is the physical biometric and signatures is the behavioural biometric. The new method establishes a relationship between the two biometrics and regenerates a visible face image from the signature features. Furthermore, the performance efficiencies of our new technique are demonstrated in terms of minimum error rates compared to published work.Keywords: behavioural biometric, face biometric, neural network, physical biometric, signature biometric
Procedia PDF Downloads 48130396 Optimization of Feeder Bus Routes at Urban Rail Transit Stations Based on Link Growth Probability
Authors: Yu Song, Yuefei Jin
Abstract:
Urban public transportation can be integrated when there is an efficient connection between urban rail lines, however, there are currently no effective or quick solutions being investigated for this connection. This paper analyzes the space-time distribution and travel demand of passenger connection travel based on taxi track data and data from the road network, excavates potential bus connection stations based on potential connection demand data, and introduces the link growth probability model in the complex network to solve the basic connection bus lines in order to ascertain the direction of the bus lines that are the most connected given the demand characteristics. Then, a tree view exhaustive approach based on constraints is suggested based on graph theory, which can hasten the convergence of findings while doing chain calculations. This study uses WEI QU NAN Station, the Xi'an Metro Line 2 terminal station in Shaanxi Province, as an illustration, to evaluate the model's and the solution method's efficacy. According to the findings, 153 prospective stations have been dug up in total, the feeder bus network for the entire line has been laid out, and the best route adjustment strategy has been found.Keywords: feeder bus, route optimization, link growth probability, the graph theory
Procedia PDF Downloads 8130395 Comparative Study of Bending Angle in Laser Forming Process Using Artificial Neural Network and Fuzzy Logic System
Authors: M. Hassani, Y. Hassani, N. Ajudanioskooei, N. N. Benvid
Abstract:
Laser Forming process as a non-contact thermal forming process is widely used to forming and bending of metallic and non-metallic sheets. In this process, according to laser irradiation along a specific path, sheet is bent. One of the most important output parameters in laser forming is bending angle that depends on process parameters such as physical and mechanical properties of materials, laser power, laser travel speed and the number of scan passes. In this paper, Artificial Neural Network and Fuzzy Logic System were used to predict of bending angle in laser forming process. Inputs to these models were laser travel speed and laser power. The comparison between artificial neural network and fuzzy logic models with experimental results has been shown both of these models have high ability to prediction of bending angles with minimum errors.Keywords: artificial neural network, bending angle, fuzzy logic, laser forming
Procedia PDF Downloads 60330394 Comparison of GIS-Based Soil Erosion Susceptibility Models Using Support Vector Machine, Binary Logistic Regression and Artificial Neural Network in the Southwest Amazon Region
Authors: Elaine Lima Da Fonseca, Eliomar Pereira Da Silva Filho
Abstract:
The modeling of areas susceptible to soil loss by hydro erosive processes consists of a simplified instrument of reality with the purpose of predicting future behaviors from the observation and interaction of a set of geoenvironmental factors. The models of potential areas for soil loss will be obtained through binary logistic regression, artificial neural networks, and support vector machines. The choice of the municipality of Colorado do Oeste in the south of the western Amazon is due to soil degradation due to anthropogenic activities, such as agriculture, road construction, overgrazing, deforestation, and environmental and socioeconomic configurations. Initially, a soil erosion inventory map constructed through various field investigations will be designed, including the use of remotely piloted aircraft, orbital imagery, and the PLANAFLORO/RO database. 100 sampling units with the presence of erosion will be selected based on the assumptions indicated in the literature, and, to complement the dichotomous analysis, 100 units with no erosion will be randomly designated. The next step will be the selection of the predictive parameters that exert, jointly, directly, or indirectly, some influence on the mechanism of occurrence of soil erosion events. The chosen predictors are altitude, declivity, aspect or orientation of the slope, curvature of the slope, composite topographic index, flow power index, lineament density, normalized difference vegetation index, drainage density, lithology, soil type, erosivity, and ground surface temperature. After evaluating the relative contribution of each predictor variable, the erosion susceptibility model will be applied to the municipality of Colorado do Oeste - Rondônia through the SPSS Statistic 26 software. Evaluation of the model will occur through the determination of the values of the R² of Cox & Snell and the R² of Nagelkerke, Hosmer and Lemeshow Test, Log Likelihood Value, and Wald Test, in addition to analysis of the Confounding Matrix, ROC Curve and Accumulated Gain according to the model specification. The validation of the synthesis map resulting from both models of the potential risk of soil erosion will occur by means of Kappa indices, accuracy, and sensitivity, as well as by field verification of the classes of susceptibility to erosion using drone photogrammetry. Thus, it is expected to obtain the mapping of the following classes of susceptibility to erosion very low, low, moderate, very high, and high, which may constitute a screening tool to identify areas where more detailed investigations need to be carried out, applying more efficient social resources.Keywords: modeling, susceptibility to erosion, artificial intelligence, Amazon
Procedia PDF Downloads 7330393 Intelligent Chatbot Generating Dynamic Responses Through Natural Language Processing
Authors: Aarnav Singh, Jatin Moolchandani
Abstract:
The proposed research work aims to build a query-based AI chatbot that can answer any question related to any topic. A chatbot is software that converses with users via text messages. In the proposed system, we aim to build a chatbot that generates a response based on the user’s query. For this, we use natural language processing to analyze the query and some set of texts to form a concise answer. The texts are obtained through web-scrapping and filtering all the credible sources from a web search. The objective of this project is to provide a chatbot that is able to provide simple and accurate answers without the user having to read through a large number of articles and websites. Creating an AI chatbot that can answer a variety of user questions on a variety of topics is the goal of the proposed research project. This chatbot uses natural language processing to comprehend user inquiries and provides succinct responses by examining a collection of writings that were scraped from the internet. The texts are carefully selected from reliable websites that are found via internet searches. This project aims to provide users with a chatbot that provides clear and precise responses, removing the need to go through several articles and web pages in great detail. In addition to exploring the reasons for their broad acceptance and their usefulness across many industries, this article offers an overview of the interest in chatbots throughout the world.Keywords: Chatbot, Artificial Intelligence, natural language processing, web scrapping
Procedia PDF Downloads 6930392 Design and Optimization of Sustainable Buildings by Combined Cooling, Heating and Power System (CCHP) Based on Exergy Analysis
Authors: Saeed Karimi, Ali Behbahaninia
Abstract:
In this study, the design and optimization of combined cooling, heating, and power system (CCHP) for a sustainable building are dealt with. Sustainable buildings are environmentally responsible and help us to save energy also reducing waste, pollution and environmental degradation. CCHP systems are widely used to save energy sources. In these systems, electricity, cooling, and heating are generating using just one primary energy source. The selection of the size of components based on the maximum demand of users will lead to an increase in the total cost of energy and equipment for the building complex. For this purpose, a system was designed in which the prime mover (gas turbine), heat recovery boiler, and absorption chiller are lower than the needed maximum. The difference in months with peak consumption is supplied with the help of electrical absorption chiller and auxiliary boiler (and the national electricity network). In this study, the optimum capacities of each of the equipment are determined based on Thermo economic method, in a way that the annual capital cost and energy consumption will be the lowest. The design was done for a gas turbine prime mover, and finally, the optimum designs were investigated using exergy analysis and were compared with a traditional energy supply system.Keywords: sustainable building, CCHP, energy optimization, gas turbine, exergy, thermo-economic
Procedia PDF Downloads 9630391 A Comparative Study of Sampling-Based Uncertainty Propagation with First Order Error Analysis and Percentile-Based Optimization
Authors: M. Gulam Kibria, Shourav Ahmed, Kais Zaman
Abstract:
In system analysis, the information on the uncertain input variables cause uncertainty in the system responses. Different probabilistic approaches for uncertainty representation and propagation in such cases exist in the literature. Different uncertainty representation approaches result in different outputs. Some of the approaches might result in a better estimation of system response than the other approaches. The NASA Langley Multidisciplinary Uncertainty Quantification Challenge (MUQC) has posed challenges about uncertainty quantification. Subproblem A, the uncertainty characterization subproblem, of the challenge posed is addressed in this study. In this subproblem, the challenge is to gather knowledge about unknown model inputs which have inherent aleatory and epistemic uncertainties in them with responses (output) of the given computational model. We use two different methodologies to approach the problem. In the first methodology we use sampling-based uncertainty propagation with first order error analysis. In the other approach we place emphasis on the use of Percentile-Based Optimization (PBO). The NASA Langley MUQC’s subproblem A is developed in such a way that both aleatory and epistemic uncertainties need to be managed. The challenge problem classifies each uncertain parameter as belonging to one the following three types: (i) An aleatory uncertainty modeled as a random variable. It has a fixed functional form and known coefficients. This uncertainty cannot be reduced. (ii) An epistemic uncertainty modeled as a fixed but poorly known physical quantity that lies within a given interval. This uncertainty is reducible. (iii) A parameter might be aleatory but sufficient data might not be available to adequately model it as a single random variable. For example, the parameters of a normal variable, e.g., the mean and standard deviation, might not be precisely known but could be assumed to lie within some intervals. It results in a distributional p-box having the physical parameter with an aleatory uncertainty, but the parameters prescribing its mathematical model are subjected to epistemic uncertainties. Each of the parameters of the random variable is an unknown element of a known interval. This uncertainty is reducible. From the study, it is observed that due to practical limitations or computational expense, the sampling is not exhaustive in sampling-based methodology. That is why the sampling-based methodology has high probability of underestimating the output bounds. Therefore, an optimization-based strategy to convert uncertainty described by interval data into a probabilistic framework is necessary. This is achieved in this study by using PBO.Keywords: aleatory uncertainty, epistemic uncertainty, first order error analysis, uncertainty quantification, percentile-based optimization
Procedia PDF Downloads 24430390 Evaluation of the Matching Optimization of Human-Machine Interface Matching in the Cab
Authors: Yanhua Ma, Lu Zhai, Xinchen Wang, Hongyu Liang
Abstract:
In this paper, by understanding the development status of the human-machine interface in today's automobile cab, a subjective and objective evaluation system for evaluating the optimization of human-machine interface matching in automobile cab was established. The man-machine interface of the car cab was divided into a software interface and a hard interface. Objective evaluation method of software human factor analysis is used to evaluate the hard interface matching; The analytic hierarchy process is used to establish the evaluation index system for the software interface matching optimization, and the multi-level fuzzy comprehensive evaluation method is used to evaluate hard interface machine. This article takes Dongfeng Sokon (DFSK) C37 model automobile as an example. The evaluation method given in the paper is used to carry out relevant analysis and evaluation, and corresponding optimization suggestions are given, which have certain reference value for designers.Keywords: analytic hierarchy process, fuzzy comprehension evaluation method, human-machine interface, matching optimization, software human factor analysis
Procedia PDF Downloads 16130389 Developing New Algorithm and Its Application on Optimal Control of Pumps in Water Distribution Network
Authors: R. Rajabpour, N. Talebbeydokhti, M. H. Ahmadi
Abstract:
In recent years, new techniques for solving complex problems in engineering are proposed. One of these techniques is JPSO algorithm. With innovative changes in the nature of the jump algorithm JPSO, it is possible to construct a graph-based solution with a new algorithm called G-JPSO. In this paper, a new algorithm to solve the optimal control problem Fletcher-Powell and optimal control of pumps in water distribution network was evaluated. Optimal control of pumps comprise of optimum timetable operation (status on and off) for each of the pumps at the desired time interval. Maximum number of status on and off for each pumps imposed to the objective function as another constraint. To determine the optimal operation of pumps, a model-based optimization-simulation algorithm was developed based on G-JPSO and JPSO algorithms. The proposed algorithm results were compared well with the ant colony algorithm, genetic and JPSO results. This shows the robustness of proposed algorithm in finding near optimum solutions with reasonable computational cost.Keywords: G-JPSO, operation, optimization, pumping station, water distribution networks
Procedia PDF Downloads 40630388 IACOP - Route Optimization in Wireless Networks Using Improved Ant Colony Optimization Protocol
Authors: S. Vasundra, D. Venkatesh
Abstract:
Wireless networks have gone through an extraordinary growth in the past few years, and will keep on playing a crucial role in future data communication. The present wireless networks aim to make communication possible anywhere and anytime. With the converging of mobile and wireless communications with Internet services, the boundary between mobile personal telecommunications and wireless computer networks is disappearing. Wireless networks of the next generation need the support of all the advances on new architectures, standards, and protocols. Since an ad hoc network may consist of a large number of mobile hosts, this imposes a significant challenge on the design of an effective and efficient routing protocol that can work well in an environment with frequent topological changes. This paper proposes improved ant colony optimization (IACO) technique. It also maintains load balancing in wireless networks. The simulation results show that the proposed IACO performs better than existing routing techniques.Keywords: wireless networks, ant colony optimization, load balancing, architecture
Procedia PDF Downloads 42930387 The Effectiveness of Teaching Emotional Intelligence on Reducing Marital Conflicts and Marital Adjustment in Married Students of Tehran University
Authors: Elham Jafari
Abstract:
The aim of this study was to evaluate the effectiveness of emotional intelligence training on reducing marital conflict and marital adjustment in married students of the University of Tehran. This research is an applied type in terms of purpose and a semi-experimental design of pre-test-post-test type with the control group and with follow-up test in terms of the data collection method. The statistical population of the present study consisted of all married students of the University of Tehran. In this study, 30 married students of the University of Tehran were selected by convenience sampling method as a sample that 15 people in the experimental group and 15 people in the control group were randomly selected. The method of data collection in this research was field and library. The data collection tool in the field section was two questionnaires of marital conflict and marital adjustment. To analyze the collected data, first at the descriptive level, using statistical indicators, the demographic characteristics of the sample were described by SPSS software. In inferential statistics, the statistical method used was the test of analysis of covariance. The results showed that the effect of the independent variable of emotional intelligence on the reduction of marital conflicts is statistically significant. And it can be inferred that emotional intelligence training has reduced the marital conflicts of married students of the University of Tehran in the experimental group compared to the control group. Also, the effect of the independent variable of emotional intelligence on marital adjustment was statistically significant. It can be inferred that emotional intelligence training has adjusted the marital adjustment of married students of the University of Tehran in the experimental group compared to the control group.Keywords: emotional intelligence, marital conflicts, marital compatibility, married students
Procedia PDF Downloads 25630386 Methods and Algorithms of Ensuring Data Privacy in AI-Based Healthcare Systems and Technologies
Authors: Omar Farshad Jeelani, Makaire Njie, Viktoriia M. Korzhuk
Abstract:
Recently, the application of AI-powered algorithms in healthcare continues to flourish. Particularly, access to healthcare information, including patient health history, diagnostic data, and PII (Personally Identifiable Information) is paramount in the delivery of efficient patient outcomes. However, as the exchange of healthcare information between patients and healthcare providers through AI-powered solutions increases, protecting a person’s information and their privacy has become even more important. Arguably, the increased adoption of healthcare AI has resulted in a significant concentration on the security risks and protection measures to the security and privacy of healthcare data, leading to escalated analyses and enforcement. Since these challenges are brought by the use of AI-based healthcare solutions to manage healthcare data, AI-based data protection measures are used to resolve the underlying problems. Consequently, this project proposes AI-powered safeguards and policies/laws to protect the privacy of healthcare data. The project presents the best-in-school techniques used to preserve the data privacy of AI-powered healthcare applications. Popular privacy-protecting methods like Federated learning, cryptographic techniques, differential privacy methods, and hybrid methods are discussed together with potential cyber threats, data security concerns, and prospects. Also, the project discusses some of the relevant data security acts/laws that govern the collection, storage, and processing of healthcare data to guarantee owners’ privacy is preserved. This inquiry discusses various gaps and uncertainties associated with healthcare AI data collection procedures and identifies potential correction/mitigation measures.Keywords: data privacy, artificial intelligence (AI), healthcare AI, data sharing, healthcare organizations (HCOs)
Procedia PDF Downloads 10030385 Improving the Performance of Gas Turbine Power Plant by Modified Axial Turbine
Authors: Hakim T. Kadhim, Faris A. Jabbar, Aldo Rona, Audrius Bagdanaviciu
Abstract:
Computer-based optimization techniques can be employed to improve the efficiency of energy conversions processes, including reducing the aerodynamic loss in a thermal power plant turbomachine. In this paper, towards mitigating secondary flow losses, a design optimization workflow is implemented for the casing geometry of a 1.5 stage axial flow turbine that improves the turbine isentropic efficiency. The improved turbine is used in an open thermodynamic gas cycle with regeneration and cogeneration. Performance estimates are obtained by the commercial software Cycle – Tempo. Design and off design conditions are considered as well as variations in inlet air temperature. Reductions in both the natural gas specific fuel consumption and in CO2 emissions are predicted by using the gas turbine cycle fitted with the new casing design. These gains are attractive towards enhancing the competitiveness and reducing the environmental impact of thermal power plant.Keywords: axial flow turbine, computational fluid dynamics, gas turbine power plant, optimization
Procedia PDF Downloads 16330384 Teachers’ Personal and Professional Characteristics: How They Relate to Teacher-Student Relationships and Students’ Behavior
Authors: Maria Poulou
Abstract:
The study investigated how teachers’ self-rated Emotional Intelligence (EI), competence in implementing Social and Emotional Learning (SEL) skills and teaching efficacy relate to teacher-student relationships and students’ emotional and behavioral difficulties. Participants were 98 elementary teachers from public schools in central Greece. They completed the Self-Rated Emotional Intelligence Scale (SREIS), the Teacher SEL Beliefs Scale, the Teachers’ Sense of Efficacy Scale (TSES), the Student-Teacher Relationships Scale-Short Form (STRS-SF) and the Strengths and Difficulties Questionnaire (SDQ) for 617 of their students, aged 6-11 years old. Structural equation modeling was used to examine an exploratory model of the variables. It was demonstrated that teachers’ emotional intelligence, SEL beliefs and teaching efficacy were significantly related to teacher-student relationships, but they were not related to students’ emotional and behavioral difficulties. Rather, teachers’ perceptions of teacher-students relationships were significantly related to these difficulties. These findings and their implications for research and practice are discussed.Keywords: emotional intelligence, social and emotional learning, teacher-student relationships, teaching efficacy
Procedia PDF Downloads 44330383 Exergetic Optimization on Solid Oxide Fuel Cell Systems
Authors: George N. Prodromidis, Frank A. Coutelieris
Abstract:
Biogas can be currently considered as an alternative option for electricity production, mainly due to its high energy content (hydrocarbon-rich source), its renewable status and its relatively low utilization cost. Solid Oxide Fuel Cell (SOFC) stacks convert fuel’s chemical energy to electricity with high efficiencies and reveal significant advantages on fuel flexibility combined with lower emissions rate, especially when utilize biogas. Electricity production by biogas constitutes a composite problem which incorporates an extensive parametric analysis on numerous dynamic variables. The main scope of the presented study is to propose a detailed thermodynamic model on the optimization of SOFC-based power plants’ operation based on fundamental thermodynamics, energy and exergy balances. This model named THERMAS (THERmodynamic MAthematical Simulation model) incorporates each individual process, during electricity production, mathematically simulated for different case studies that represent real life operational conditions. Also, THERMAS offers the opportunity to choose a great variety of different values for each operational parameter individually, thus allowing for studies within unexplored and experimentally impossible operational ranges. Finally, THERMAS innovatively incorporates a specific criterion concluded by the extensive energy analysis to identify the most optimal scenario per simulated system in exergy terms. Therefore, several dynamical parameters as well as several biogas mixture compositions have been taken into account, to cover all the possible incidents. Towards the optimization process in terms of an innovative OPF (OPtimization Factor), presented here, this research study reveals that systems supplied by low methane fuels can be comparable to these supplied by pure methane. To conclude, such an innovative simulation model indicates a perspective on the optimal design of a SOFC stack based system, in the direction of the commercialization of systems utilizing biogas.Keywords: biogas, exergy, efficiency, optimization
Procedia PDF Downloads 37330382 Sequential Covering Algorithm for Nondifferentiable Global Optimization Problem and Applications
Authors: Mohamed Rahal, Djaouida Guetta
Abstract:
In this paper, the one-dimensional unconstrained global optimization problem of continuous functions satifying a Hölder condition is considered. We extend the algorithm of sequential covering SCA for Lipschitz functions to a large class of Hölder functions. The convergence of the method is studied and the algorithm can be applied to systems of nonlinear equations. Finally, some numerical examples are presented and illustrate the efficiency of the present approach.Keywords: global optimization, Hölder functions, sequential covering method, systems of nonlinear equations
Procedia PDF Downloads 37630381 Modeling Visual Memorability Assessment with Autoencoders Reveals Characteristics of Memorable Images
Authors: Elham Bagheri, Yalda Mohsenzadeh
Abstract:
Image memorability refers to the phenomenon where certain images are more likely to be remembered by humans than others. It is a quantifiable and intrinsic attribute of an image. Understanding how visual perception and memory interact is important in both cognitive science and artificial intelligence. It reveals the complex processes that support human cognition and helps to improve machine learning algorithms by mimicking the brain's efficient data processing and storage mechanisms. To explore the computational underpinnings of image memorability, this study examines the relationship between an image's reconstruction error, distinctiveness in latent space, and its memorability score. A trained autoencoder is used to replicate human-like memorability assessment inspired by the visual memory game employed in memorability estimations. This study leverages a VGG-based autoencoder that is pre-trained on the vast ImageNet dataset, enabling it to recognize patterns and features that are common to a wide and diverse range of images. An empirical analysis is conducted using the MemCat dataset, which includes 10,000 images from five broad categories: animals, sports, food, landscapes, and vehicles, along with their corresponding memorability scores. The memorability score assigned to each image represents the probability of that image being remembered by participants after a single exposure. The autoencoder is finetuned for one epoch with a batch size of one, attempting to create a scenario similar to human memorability experiments where memorability is quantified by the likelihood of an image being remembered after being seen only once. The reconstruction error, which is quantified as the difference between the original and reconstructed images, serves as a measure of how well the autoencoder has learned to represent the data. The reconstruction error of each image, the error reduction, and its distinctiveness in latent space are calculated and correlated with the memorability score. Distinctiveness is measured as the Euclidean distance between each image's latent representation and its nearest neighbor within the autoencoder's latent space. Different structural and perceptual loss functions are considered to quantify the reconstruction error. The results indicate that there is a strong correlation between the reconstruction error and the distinctiveness of images and their memorability scores. This suggests that images with more unique distinct features that challenge the autoencoder's compressive capacities are inherently more memorable. There is also a negative correlation between the reduction in reconstruction error compared to the autoencoder pre-trained on ImageNet, which suggests that highly memorable images are harder to reconstruct, probably due to having features that are more difficult to learn by the autoencoder. These insights suggest a new pathway for evaluating image memorability, which could potentially impact industries reliant on visual content and mark a step forward in merging the fields of artificial intelligence and cognitive science. The current research opens avenues for utilizing neural representations as instruments for understanding and predicting visual memory.Keywords: autoencoder, computational vision, image memorability, image reconstruction, memory retention, reconstruction error, visual perception
Procedia PDF Downloads 9530380 Emotional Intelligence as a Predictor of Job Satisfaction in the Nigerian Construction Industry
Authors: Adedayo Johnson Ogungbile, Ayodeji Emmanuel Oke, Oluwaseyi Alabi Awodele
Abstract:
This study examines the role of emotional intelligence (EI) as a predictor of job satisfaction within the Nigerian construction industry. Utilizing a methodology that combines mean comparison and correlation analysis, the research explores how EI influences job satisfaction across diverse demographic and professional categories. The construction industry, known for its dynamic and often challenging work environment, provides a unique context to investigate how EI contributes to employee satisfaction. The findings reveal a significant positive correlation between EI and job satisfaction across the industry. Gender-based analysis shows that male employees typically report higher EI and job satisfaction levels compared to their female counterparts, although the impact of EI on job satisfaction is more substantial among women. The study further explores the relationship between trait EI and specific job satisfaction categories, identifying a general positive association with overall job satisfaction but not with supervisor-related satisfaction. Employees are categorized into four EI classes, consistently showing that higher EI levels correspond to greater job satisfaction. These findings align with existing literature, underscoring EI's pivotal role in enhancing job satisfaction in the construction sector. The study concludes that fostering EI among construction industry professionals can lead to improved job satisfaction and performance. Consequently, organizations are encouraged to integrate EI development into their professional growth programs to cultivate a more satisfied and effective workforce. In essence, this research highlights the importance of EI as a key predictor of job satisfaction in the Nigerian construction industry, providing valuable insights for both industry stakeholders and researchers into the benefits of prioritizing emotional intelligence in this high-stakes environment.Keywords: emotional intelligence, job satisfaction, construction industry, workforce productivity, demographics
Procedia PDF Downloads 2630379 A Novel Approach to Design and Implement Context Aware Mobile Phone
Authors: G. S. Thyagaraju, U. P. Kulkarni
Abstract:
Context-aware computing refers to a general class of computing systems that can sense their physical environment, and adapt their behaviour accordingly. Context aware computing makes systems aware of situations of interest, enhances services to users, automates systems and personalizes applications. Context-aware services have been introduced into mobile devices, such as PDA and mobile phones. In this paper we are presenting a novel approaches used to realize the context aware mobile. The context aware mobile phone (CAMP) proposed in this paper senses the users situation automatically and provides user context required services. The proposed system is developed by using artificial intelligence techniques like Bayesian Network, fuzzy logic and rough sets theory based decision table. Bayesian Network to classify the incoming call (high priority call, low priority call and unknown calls), fuzzy linguistic variables and membership degrees to define the context situations, the decision table based rules for service recommendation. To exemplify and demonstrate the effectiveness of the proposed methods, the context aware mobile phone is tested for college campus scenario including different locations like library, class room, meeting room, administrative building and college canteen.Keywords: context aware mobile, fuzzy logic, decision table, Bayesian probability
Procedia PDF Downloads 37130378 Speed Control of DC Motor Using Optimization Techniques Based PID Controller
Authors: Santosh Kumar Suman, Vinod Kumar Giri
Abstract:
The goal of this paper is to outline a speed controller of a DC motor by choice of a PID parameters utilizing genetic algorithms (GAs), the DC motor is extensively utilized as a part of numerous applications such as steel plants, electric trains, cranes and a great deal more. DC motor could be represented by a nonlinear model when nonlinearities such as attractive dissemination are considered. To provide effective control, nonlinearities and uncertainties in the model must be taken into account in the control design. The DC motor is considered as third order system. Objective of this paper three type of tuning techniques for PID parameter. In this paper, an independently energized DC motor utilizing MATLAB displaying, has been outlined whose velocity might be examined utilizing the Proportional, Integral, Derivative (KP, KI , KD) addition of the PID controller. Since, established controllers PID are neglecting to control the drive when weight parameters be likewise changed. The principle point of this paper is to dissect the execution of optimization techniques viz. The Genetic Algorithm (GA) for improve PID controllers parameters for velocity control of DC motor and list their points of interest over the traditional tuning strategies. The outcomes got from GA calculations were contrasted and that got from traditional technique. It was found that the optimization techniques beat customary tuning practices of ordinary PID controllers.Keywords: DC motor, PID controller, optimization techniques, genetic algorithm (GA), objective function, IAE
Procedia PDF Downloads 42630377 A Multi-Population DE with Adaptive Mutation and Local Search for Global Optimization
Authors: Zhoucheng Bao, Haiyan Zhu, Tingting Pang, Zuling Wang
Abstract:
This paper proposes a multi-population DE with adaptive mutation and local search for global optimization, named AMMADE. In order to better coordinate the cooperation between the populations and the rational use of resources. In AMMADE, the population is divided based on the Euclidean distance sorting method at each generation to appropriately coordinate the cooperation between subpopulations and the usage of resources, such that the best-performed subpopulation will get more computing resources in the next generation. Further, an adaptive local search strategy is employed on the best-performed subpopulation to achieve a balanced search. The proposed algorithm has been tested by solving optimization problems taken from CEC2014 benchmark problems. Experimental results show that our algorithm can achieve a competitive or better than related methods. The results also confirm the significance of devised strategies in the proposed algorithm.Keywords: differential evolution, multi-mutation strategies, memetic algorithm, adaptive local search
Procedia PDF Downloads 164