Search results for: Time domain analysis
12867 Multi-objective Optimisation of Composite Laminates under Heat and Moisture Effects using a Hybrid Neuro-GA Algorithm
Authors: M. R. Ghasemi, A. Ehsani
Abstract:
In this paper, the optimum weight and cost of a laminated composite plate is seeked, while it undergoes the heaviest load prior to a complete failure. Various failure criteria are defined for such structures in the literature. In this work, the Tsai-Hill theory is used as the failure criterion. The theory of analysis was based on the Classical Lamination Theory (CLT). A newly type of Genetic Algorithm (GA) as an optimization technique with a direct use of real variables was employed. Yet, since the optimization via GAs is a long process, and the major time is consumed through the analysis, Radial Basis Function Neural Networks (RBFNN) was employed in predicting the output from the analysis. Thus, the process of optimization will be carried out through a hybrid neuro-GA environment, and the procedure will be carried out until a predicted optimum solution is achieved.Keywords: Composite Laminates, GA, Multi-objectiveOptimisation, Neural Networks, RBFNN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 163512866 Joint Use of Factor Analysis (FA) and Data Envelopment Analysis (DEA) for Ranking of Data Envelopment Analysis
Authors: Reza Nadimi, Fariborz Jolai
Abstract:
This article combines two techniques: data envelopment analysis (DEA) and Factor analysis (FA) to data reduction in decision making units (DMU). Data envelopment analysis (DEA), a popular linear programming technique is useful to rate comparatively operational efficiency of decision making units (DMU) based on their deterministic (not necessarily stochastic) input–output data and factor analysis techniques, have been proposed as data reduction and classification technique, which can be applied in data envelopment analysis (DEA) technique for reduction input – output data. Numerical results reveal that the new approach shows a good consistency in ranking with DEA.Keywords: Effectiveness, Decision Making, Data EnvelopmentAnalysis, Factor Analysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 243212865 Prediction of Time to Crack Reinforced Concrete by Chloride Induced Corrosion
Authors: Anuruddha Jayasuriya, Thanakorn Pheeraphan
Abstract:
In this paper, a review of different mathematical models which can be used as prediction tools to assess the time to crack reinforced concrete (RC) due to corrosion is investigated. This investigation leads to an experimental study to validate a selected prediction model. Most of these mathematical models depend upon the mechanical behaviors, chemical behaviors, electrochemical behaviors or geometric aspects of the RC members during a corrosion process. The experimental program is designed to verify the accuracy of a well-selected mathematical model from a rigorous literature study. Fundamentally, the experimental program exemplifies both one-dimensional chloride diffusion using RC squared slab elements of 500 mm by 500 mm and two-dimensional chloride diffusion using RC squared column elements of 225 mm by 225 mm by 500 mm. Each set consists of three water-to-cement ratios (w/c); 0.4, 0.5, 0.6 and two cover depths; 25 mm and 50 mm. 12 mm bars are used for column elements and 16 mm bars are used for slab elements. All the samples are subjected to accelerated chloride corrosion in a chloride bath of 5% (w/w) sodium chloride (NaCl) solution. Based on a pre-screening of different models, it is clear that the well-selected mathematical model had included mechanical properties, chemical and electrochemical properties, nature of corrosion whether it is accelerated or natural, and the amount of porous area that rust products can accommodate before exerting expansive pressure on the surrounding concrete. The experimental results have shown that the selected model for both one-dimensional and two-dimensional chloride diffusion had ±20% and ±10% respective accuracies compared to the experimental output. The half-cell potential readings are also used to see the corrosion probability, and experimental results have shown that the mass loss is proportional to the negative half-cell potential readings that are obtained. Additionally, a statistical analysis is carried out in order to determine the most influential factor that affects the time to corrode the reinforcement in the concrete due to chloride diffusion. The factors considered for this analysis are w/c, bar diameter, and cover depth. The analysis is accomplished by using Minitab statistical software, and it showed that cover depth is the significant effect on the time to crack the concrete from chloride induced corrosion than other factors considered. Thus, the time predictions can be illustrated through the selected mathematical model as it covers a wide range of factors affecting the corrosion process, and it can be used to predetermine the durability concern of RC structures that are vulnerable to chloride exposure. And eventually, it is further concluded that cover thickness plays a vital role in durability in terms of chloride diffusion.
Keywords: Accelerated corrosion, chloride diffusion, corrosion cracks, passivation layer, reinforcement corrosion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 89812864 Numerical Modelling of Effective Diffusivity in Bone Tissue Engineering
Authors: Ayesha Sohail, Khadija Maqbool, Anila Asif, Haroon Ahmad
Abstract:
These days, the field of tissue engineering is getting serious attention due to its usefulness. Bone tissue engineering helps to address and sort-out the critical sized and non-healing orthopedic problems by the creation of manmade bone tissue. We will design and validate an efficient numerical model, which will simulate the effective diffusivity in bone tissue engineering. Our numerical model will be based on the finite element analysis of the diffusion-reaction equations. It will have the ability to optimize the diffusivity, even at multi-scale, with the variation of time. It will also have a special feature “parametric sweep”, with which we will be able to predict the oxygen, glucose and cell density dynamics, more accurately. We will fix these problems by modifying the governing equations, by selecting appropriate spatio-temporal finite element schemes and by transient analysis.
Keywords: Bone tissue engineering, Transient Analysis, Scaffolds, fabrication techniques.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 247112863 Computational Simulations on Stability of Model Predictive Control for Linear Discrete-time Stochastic Systems
Authors: Tomoaki Hashimoto
Abstract:
Model predictive control is a kind of optimal feedback control in which control performance over a finite future is optimized with a performance index that has a moving initial time and a moving terminal time. This paper examines the stability of model predictive control for linear discrete-time systems with additive stochastic disturbances. A sufficient condition for the stability of the closed-loop system with model predictive control is derived by means of a linear matrix inequality. The objective of this paper is to show the results of computational simulations in order to verify the effectiveness of the obtained stability condition.Keywords: Computational simulations, optimal control, predictive control, stochastic systems, discrete-time systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 185412862 Improving Patients Discharge Process in Hospitals by using Six Sigma Approach
Authors: Mahmoud A. El-Banna
Abstract:
The need to increase the efficiency of health care systems is becoming an obligation, and one of area of improvement is the discharge process. The objective of this work is to minimize the patients discharge time (for insured patients) to be less than 50 minutes by using six sigma approach, this improvement will also: lead to an increase in customer satisfaction, increase the number of admissions and turnover on the rooms, increase hospital profitability.Three different departments were considered in this study: Female, Male, and Paediatrics. Six Sigma approach coupled with simulation has been applied to reduce the patients discharge time for pediatrics, female, and male departments at hospital. Upon applying these recommendations at hospital: 60%, 80%, and 22% of insured female, male, and pediatrics patients respectively will have discharge time less than the upper specification time i.e. 50 min.Keywords: Discharge Time, Healthcare, Hospitals, Patients, Process Improvement, Six Sigma, Simulation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 475112861 Periodic Solutions for a Delayed Population Model on Time Scales
Authors: Kejun Zhuang, Zhaohui Wen
Abstract:
This paper deals with a delayed single population model on time scales. With the assistance of coincidence degree theory, sufficient conditions for existence of periodic solutions are obtained. Furthermore, the better estimations for bounds of periodic solutions are established.
Keywords: Coincidence degree, continuation theorem, periodic solutions, time scales
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 135412860 Lineup Optimization Model of Basketball Players Based on the Prediction of Recursive Neural Networks
Authors: Wang Yichen, Haruka Yamashita
Abstract:
In recent years, in the field of sports, decision making such as member in the game and strategy of the game based on then analysis of the accumulated sports data are widely attempted. In fact, in the NBA basketball league where the world's highest level players gather, to win the games, teams analyze the data using various statistical techniques. However, it is difficult to analyze the game data for each play such as the ball tracking or motion of the players in the game, because the situation of the game changes rapidly, and the structure of the data should be complicated. Therefore, it is considered that the analysis method for real time game play data is proposed. In this research, we propose an analytical model for "determining the optimal lineup composition" using the real time play data, which is considered to be difficult for all coaches. In this study, because replacing the entire lineup is too complicated, and the actual question for the replacement of players is "whether or not the lineup should be changed", and “whether or not Small Ball lineup is adopted”. Therefore, we propose an analytical model for the optimal player selection problem based on Small Ball lineups. In basketball, we can accumulate scoring data for each play, which indicates a player's contribution to the game, and the scoring data can be considered as a time series data. In order to compare the importance of players in different situations and lineups, we combine RNN (Recurrent Neural Network) model, which can analyze time series data, and NN (Neural Network) model, which can analyze the situation on the field, to build the prediction model of score. This model is capable to identify the current optimal lineup for different situations. In this research, we collected all the data of accumulated data of NBA from 2019-2020. Then we apply the method to the actual basketball play data to verify the reliability of the proposed model.Keywords: Recurrent Neural Network, players lineup, basketball data, decision making model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 84212859 Digital Control Algorithm Based on Delta-Operator for High-Frequency DC-DC Switching Converters
Authors: Renkai Wang, Tingcun Wei
Abstract:
In this paper, a digital control algorithm based on delta-operator is presented for high-frequency digitally-controlled DC-DC switching converters. The stability and the controlling accuracy of the DC-DC switching converters are improved by using the digital control algorithm based on delta-operator without increasing the hardware circuit scale. The design method of voltage compensator in delta-domain using PID (Proportion-Integration- Differentiation) control is given in this paper, and the simulation results based on Simulink platform are provided, which have verified the theoretical analysis results very well. It can be concluded that, the presented control algorithm based on delta-operator has better stability and controlling accuracy, and easier hardware implementation than the existed control algorithms based on z-operator, therefore it can be used for the voltage compensator design in high-frequency digitally- controlled DC-DC switching converters.
Keywords: Digitally-controlled DC-DC switching converter, finite word length, control algorithm based on delta-operator, high-frequency, stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 127212858 Effect of Time Delay on the Transmission of Dengue Fever
Authors: K. Patanarapelert, I.M. Tang
Abstract:
The effect of a time delay on the transmission on dengue fever is studied. The time delay is due to the presence of an incubation period for the dengue virus to develop in the mosquito before the mosquito becomes infectious. The conditions for the existence of a Hopf bifurcation to limit cycle behavior are established. The conditions are different from the usual one and they are based on whether a particular third degree polynomial has positive real roots. A theorem for determining whether for a given set of parameter values, a critical delay time exist is given. It is found that for a set of realistic values of the parameters in the model, a Hopf bifurcation can not occur. For a set of unrealistic values of some of the parameters, it is shown that a Hopf bifurcation can occur. Numerical solutions using this last set show the trajectory of two of the variables making a transition from a spiraling orbit to a limit cycle orbit.Keywords: Dengue fever transmission, time delay, Hopfbifurcation, limit cycle behavior
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 180412857 Time Dependent Biodistribution Modeling of 177Lu-DOTATOC Using Compartmental Analysis
Authors: M. Mousavi-Daramoroudi, H. Yousefnia, F. Abbasi-Davani, S. Zolghadri
Abstract:
In this study, 177Lu-DOTATOC was prepared under optimized conditions (radiochemical purity: > 99%, radionuclidic purity: > 99%). The percentage of injected dose per gram (%ID/g) was calculated for organs up to 168 h post injection. Compartmental model was applied to mathematical description of the drug behaviour in tissue at different times. The biodistribution data showed the significant excretion of the radioactivity from the kidneys. The adrenal and pancreas, as major expression sites for somatostatin receptor (SSTR), had significant uptake. A pharmacokinetic model of 177Lu-DOTATOC was presented by compartmental analysis which demonstrates the behavior of the complex.Keywords: Biodistribution, compartmental modeling, 177Lu, octreotide.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 82812856 Biospeckle Techniques in Quality Evaluation of Indian Fruits
Authors: MD Zaheer Ansari, A.K. Nirala
Abstract:
In this study spatial-temporal speckle correlation techniques have been applied for the quality evaluation of three different Indian fruits namely apple, pear and tomato for the first time. The method is based on the analysis of variations of laser light scattered from biological samples. The results showed that crosscorrelation coefficients of biospeckle patterns change subject to their freshness and the storage conditions. The biospeckle activity was determined by means of the cross-correlation functions of the intensity fluctuations. Significant changes in biospeckle activity were observed during their shelf lives. From the study, it is found that the biospeckle activity decreases with the shelf-life storage time. Further it has been shown that biospeckle activity changes according to their respiration rates.
Keywords: Biospeckle, cross-correlation, respiration, shelf-life.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 249712855 Personalized Learning: An Analysis Using Item Response Theory
Authors: A. Yacob, N. Hj. Ali, M. H. Yusoff, M. Y. MohdSaman, W. M. A. F. W. Hamzah
Abstract:
Personalized learning becomes increasingly popular which not be restricted by time, place or any other barriers. This study proposes an analysis of Personalized Learning using Item Response Theory which considers course material difficulty and learner ability.The study investigates twenty undergraduate students at TATI University College, who are taking programming subject. By using the IRT,it was found that, finding the most appropriate problem levels to each student include high and low level test items together is not a problem. Thus, the student abilities can be asses more accurately and fairly. Learners who experience more anxiety will affect a heavier cognitive load and receive lower test scores.Instructors are encouraged to provide a supportive learning environment to enhance learning effectiveness because Cognitive Load Theory concerns the limited capacity of the brain to absorb new information.
Keywords: Analysis, Cognitive Load Theory, Item Response Theory, Learning, Motivation, Performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 312412854 Three Steps of One-way Nested Grid for Energy Balance Equations by Wave Model
Authors: Worachat Wannawong, Usa W. Humphries, Prungchan Wongwises, Suphat Vongvisessomjai
Abstract:
The three steps of the standard one-way nested grid for a regional scale of the third generation WAve Model Cycle 4 (WAMC4) is scrutinized. The model application is enabled to solve the energy balance equation on a coarse resolution grid in order to produce boundary conditions for a smaller area by the nested grid technique. In the present study, the model takes a full advantage of the fine resolution of wind fields in space and time produced by the available U.S. Navy Global Atmospheric Prediction System (NOGAPS) model with 1 degree resolution. The nested grid application of the model is developed in order to gradually increase the resolution from the open ocean towards the South China Sea (SCS) and the Gulf of Thailand (GoT) respectively. The model results were compared with buoy observations at Ko Chang, Rayong and Huahin locations which were obtained from the Seawatch project. In addition, the results were also compared with Satun based weather station which was provided from Department of Meteorology, Thailand. The data collected from this station presented the significant wave height (Hs) reached 12.85 m. The results indicated that the tendency of the Hs from the model in the spherical coordinate propagation with deep water condition in the fine grid domain agreed well with the Hs from the observations.Keywords: energy balance equation, Gulf of Thailand, nested gridapplication, South China Sea, wave model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 160312853 Eliciting and Confirming Data, Information, Knowledge and Wisdom in a Specialist Health Care Setting: The WICKED Method
Authors: S. Impey, D. Berry, S. Furtado, M. Galvin, L. Grogan, O. Hardiman, L. Hederman, M. Heverin, V. Wade, L. Douris, D. O'Sullivan, G. Stephens
Abstract:
Healthcare is a knowledge-rich environment. This knowledge, while valuable, is not always accessible outside the borders of individual clinics. This research aims to address part of this problem (at a study site) by constructing a maximal data set (knowledge artefact) for motor neurone disease (MND). This data set is proposed as an initial knowledge base for a concurrent project to develop an MND patient data platform. It represents the domain knowledge at the study site for the duration of the research (12 months). A knowledge elicitation method was also developed from the lessons learned during this process - the WICKED method. WICKED is an anagram of the words: eliciting and confirming data, information, knowledge, wisdom. But it is also a reference to the concept of wicked problems, which are complex and challenging, as is eliciting expert knowledge. The method was evaluated at a second site, and benefits and limitations were noted. Benefits include that the method provided a systematic way to manage data, information, knowledge and wisdom (DIKW) from various sources, including healthcare specialists and existing data sets. Limitations surrounded the time required and how the data set produced only represents DIKW known during the research period. Future work is underway to address these limitations.
Keywords: Healthcare, knowledge acquisition, maximal data sets, action design science.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 57212852 Time-Delay Estimation Using Cross-ΨB-Energy Operator
Authors: Z. Saidi, A.O. Boudraa, J.C. Cexus, S. Bourennane
Abstract:
In this paper, a new time-delay estimation technique based on the cross IB-energy operator [5] is introduced. This quadratic energy detector measures how much a signal is present in another one. The location of the peak of the energy operator, corresponding to the maximum of interaction between the two signals, is the estimate of the delay. The method is a fully data-driven approach. The discrete version of the continuous-time form of the cross IBenergy operator, for its implementation, is presented. The effectiveness of the proposed method is demonstrated on real underwater acoustic signals arriving from targets and the results compared to the cross-correlation method.Keywords: Teager-Kaiser energy operator, Cross-energyoperator, Time-Delay, Underwater acoustic signals.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 565612851 Estimation of Buffer Size of Internet Gateway Server via G/M/1 Queuing Model
Authors: Dr. L.K. Singh, Dr. R. M. L, Riktesh Srivastava
Abstract:
How to efficiently assign system resource to route the Client demand by Gateway servers is a tricky predicament. In this paper, we tender an enhanced proposal for autonomous recital of Gateway servers under highly vibrant traffic loads. We devise a methodology to calculate Queue Length and Waiting Time utilizing Gateway Server information to reduce response time variance in presence of bursty traffic. The most widespread contemplation is performance, because Gateway Servers must offer cost-effective and high-availability services in the elongated period, thus they have to be scaled to meet the expected load. Performance measurements can be the base for performance modeling and prediction. With the help of performance models, the performance metrics (like buffer estimation, waiting time) can be determined at the development process. This paper describes the possible queue models those can be applied in the estimation of queue length to estimate the final value of the memory size. Both simulation and experimental studies using synthesized workloads and analysis of real-world Gateway Servers demonstrate the effectiveness of the proposed system.Keywords: Gateway Server, G/M/1 Queuing Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 160012850 Design Patterns for Emergency Management Processes
Authors: Tomáš Ludík, Jiří Barta, Josef Navrátil
Abstract:
Natural or human made disasters have a significant negative impact on the environment. At the same time there is an extensive effort to support management and decision making in emergency situations by information technologies. Therefore the purpose of the paper is to propose a design patterns applicable in emergency management, enabling better analysis and design of emergency management processes and therefore easier development and deployment of information systems in the field of emergency management. It will be achieved by detailed analysis of existing emergency management legislation, contingency plans and information systems. The result is a set of design patterns focused at emergency management processes that enable easier design of emergency plans or development of new information system. These results will have a major impact on the development of new information systems as well as to more effective and faster solving of emergencies.
Keywords: Analysis and Design, Business Process Modeling Notation, Contingency Plans, Design Patterns, Emergency Management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 310612849 Investigating the Effect of VR, Time Study and Ergonomics on the Design of Industrial Workstations
Authors: Aydin Azizi, Poorya Ghafoorpoor Yazdi
Abstract:
This paper presents the review of the studies on the ergonomics, virtual reality, and work measurement (time study) at the industrial workstations because each of these three individual techniques can be used to improve the design of workstations and task position. The objective of this paper is to give an overall literature review that if there is any relation between these three different techniques. Therefore, it is so important to review the scientific studies to find a better and effective way for improving design of workstations. On the other hand, manufacturers found that instead of using one of the approaches, utilizing the combination of these individual techniques are more effective to reduce the cost and production time.
Keywords: Ergonomics, time study, virtual reality, workplace.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 84212848 Numerical Simulation of Progressive Collapse for a Reinforced Concrete Building
Authors: Han-Soo Kim, Jae-Gyun Ahn, Hyo-Seung Ahn
Abstract:
Though nonlinear dynamic analysis using a specialized hydro-code such as AUTODYN is accurate and useful tool for progressive collapse assessment of a multi-story building subjected to blast load, it takes too much time to be applied to a practical simulation of progressive collapse of a tall building. In this paper, blast analysis of a RC frame structure using a simplified model with Reinforcement Contact technique provided in Ansys Workbench was introduced and investigated on its accuracy. Even though the simplified model has a fraction of elements of the detailed model, the simplified model with this modeling technique shows similar structural behavior under the blast load to the detailed model. The proposed modeling method can be effectively applied to blast loading progressive collapse analysis of a RC frame structure.Keywords: Autodyn, Blast Load, Progressive Collapse, Reinforcement Contact.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 427312847 Mathematical Analysis of Stock Prices Prediction in a Financial Market Using Geometric Brownian Motion Model
Authors: Edikan E. Akpanibah, Ogunmodimu Dupe Catherine
Abstract:
The relevance of geometric Brownian motion (GBM) in modelling the behaviour of stock market prices (SMP) cannot be over emphasized taking into consideration the volatility of the SMP. Consequently, there is need to investigate how GBM models are being estimated and used in financial market to predict SMP. To achieve this, the GBM estimation and its application to the SMP of some selected companies are studied. The normal and log-normal distributions were used to determine the expected value, variance and co-variance. Furthermore, the GBM model was used to predict the SMP of some selected companies over a period of time and the mean absolute percentage error (MAPE) were calculated and used to determine the accuracy of the GBM model in predicting the SMP of the four companies under consideration. It was observed that for all the four companies, their MAPE values were within the region of acceptance. Also, the MAPE values of our data were compared to an existing literature to test the accuracy of our prediction with respect to time of investment. Finally, some numerical simulations of the graphs of the SMP, expectations and variance of the four companies over a period of time were presented using MATLAB programming software.
Keywords: Stock Market, Geometric Brownian Motion, normal and log-normal distribution, mean absolute percentage error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27712846 A Deterministic Polynomial-time Algorithm for the Clique Problem and the Equality of P and NP Complexity Classes
Authors: Zohreh O. Akbari
Abstract:
In this paper a deterministic polynomial-time algorithm is presented for the Clique problem. The case is considered as the problem of omitting the minimum number of vertices from the input graph so that none of the zeroes on the graph-s adjacency matrix (except the main diagonal entries) would remain on the adjacency matrix of the resulting subgraph. The existence of a deterministic polynomial-time algorithm for the Clique problem, as an NP-complete problem will prove the equality of P and NP complexity classes.Keywords: Clique problem, Deterministic Polynomial-time Algorithm, Equality of P and NP Complexity Classes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 181912845 Retrieval Augmented Generation against the Machine: Merging Human Cyber Security Expertise with Generative AI
Authors: Brennan Lodge
Abstract:
Amidst a complex regulatory landscape, Retrieval Augmented Generation (RAG) emerges as a transformative tool for Governance Risk and Compliance (GRC) officers. This paper details the application of RAG in synthesizing Large Language Models (LLMs) with external knowledge bases, offering GRC professionals an advanced means to adapt to rapid changes in compliance requirements. While the development for standalone LLMs is exciting, such models do have their downsides. LLMs cannot easily expand or revise their memory, and they cannot straightforwardly provide insight into their predictions, and may produce “hallucinations.” Leveraging a pre-trained seq2seq transformer and a dense vector index of domain-specific data, this approach integrates real-time data retrieval into the generative process, enabling gap analysis and the dynamic generation of compliance and risk management content. We delve into the mechanics of RAG, focusing on its dual structure that pairs parametric knowledge contained within the transformer model with non-parametric data extracted from an updatable corpus. This hybrid model enhances decision-making through context-rich insights, drawing from the most current and relevant information, thereby enabling GRC officers to maintain a proactive compliance stance. Our methodology aligns with the latest advances in neural network fine-tuning, providing a granular, token-level application of retrieved information to inform and generate compliance narratives. By employing RAG, we exhibit a scalable solution that can adapt to novel regulatory challenges and cybersecurity threats, offering GRC officers a robust, predictive tool that augments their expertise. The granular application of RAG’s dual structure not only improves compliance and risk management protocols but also informs the development of compliance narratives with pinpoint accuracy. It underscores AI’s emerging role in strategic risk mitigation and proactive policy formation, positioning GRC officers to anticipate and navigate the complexities of regulatory evolution confidently.
Keywords: Retrieval Augmented Generation, Governance Risk and Compliance, Cybersecurity, AI-driven Compliance, Risk Management, Generative AI.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15912844 Stability Analysis of Linear Fractional Order Neutral System with Multiple Delays by Algebraic Approach
Authors: Lianglin Xiong, Yun Zhao, Tao Jiang
Abstract:
In this paper, we study the stability of n-dimensional linear fractional neutral differential equation with time delays. By using the Laplace transform, we introduce a characteristic equation for the above system with multiple time delays. We discover that if all roots of the characteristic equation have negative parts, then the equilibrium of the above linear system with fractional order is Lyapunov globally asymptotical stable if the equilibrium exist that is almost the same as that of classical differential equations. An example is provided to show the effectiveness of the approach presented in this paper.
Keywords: Fractional neutral differential equation, Laplace transform, characteristic equation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 231512843 Analysis of Delay and Throughput in MANET for DSR Protocol
Authors: Kumar Manoj, Ramesh Kumar, Kumari Arti
Abstract:
A wireless Ad-hoc network consists of wireless nodes communicating without the need for a centralized administration, in which all nodes potentially contribute to the routing process.In this paper, we report the simulation results of four different scenarios for wireless ad hoc networks having thirty nodes. The performances of proposed networks are evaluated in terms of number of hops per route, delay and throughput with the help of OPNET simulator. Channel speed 1 Mbps and simulation time 600 sim-seconds were taken for all scenarios. For the above analysis DSR routing protocols has been used. The throughput obtained from the above analysis (four scenario) are compared as shown in Figure 3. The average media access delay at node_20 for two routes and at node_20 for four different scenario are compared as shown in Figures 4 and 5. It is observed that the throughput will degrade when it will follow different hops for same source to destination (i.e. it has dropped from 1.55 Mbps to 1.43 Mbps which is around 9.7%, and then dropped to 0.48Mbps which is around 35%).Keywords: Throughput, Delay, DSR, OPNET, MANET, DSSS
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 267712842 Exploiting Two Intelligent Models to Predict Water Level: A Field Study of Urmia Lake, Iran
Authors: Shahab Kavehkar, Mohammad Ali Ghorbani, Valeriy Khokhlov, Afshin Ashrafzadeh, Sabereh Darbandi
Abstract:
Water level forecasting using records of past time series is of importance in water resources engineering and management. For example, water level affects groundwater tables in low-lying coastal areas, as well as hydrological regimes of some coastal rivers. Then, a reliable prediction of sea-level variations is required in coastal engineering and hydrologic studies. During the past two decades, the approaches based on the Genetic Programming (GP) and Artificial Neural Networks (ANN) were developed. In the present study, the GP is used to forecast daily water level variations for a set of time intervals using observed water levels. The measurements from a single tide gauge at Urmia Lake, Northwest Iran, were used to train and validate the GP approach for the period from January 1997 to July 2008. Statistics, the root mean square error and correlation coefficient, are used to verify model by comparing with a corresponding outputs from Artificial Neural Network model. The results show that both these artificial intelligence methodologies are satisfactory and can be considered as alternatives to the conventional harmonic analysis.
Keywords: Water-Level variation, forecasting, artificial neural networks, genetic programming, comparative analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 234312841 Influence of Moringa Leaves Extract on the Response of Hb Molecule to Dose Rates’ Changes: II. Relaxation Time and Its Thermodynamic Driven State Functions
Authors: Mohamed M. M. Elnasharty, Azhar M. Elwan
Abstract:
Irradiation deposits energy through ionisation changing the bio-system’s net dipole, allowing the use of dielectric parameters and thermodynamic state functions related to these parameters as biophysical detectors to electrical inhomogeneity within the biosystem. This part is concerned with the effect of Moringa leaves extract, natural supplement, on the response of the biosystem to two different dose rates of irradiation. Having Hb molecule as a representative to the biosystem to be least invasive to the biosystem, dielectric measurements were used to extract the relaxation time of certain process found in the Hb spectrum within the indicated frequency window and the interrelated thermodynamic state functions were calculated from the deduced relaxation time. The results showed that relaxation time was decreased for both dose rates indicating a strong influence of Moringa on the response of biosystem and consequently Hb molecule. This influence was presented in the relaxation time and other parameters as well.
Keywords: Activation energy, DC conductivity, dielectric relaxation, enthalpy change, moringa leaves extract, relaxation time.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 67212840 Exploring the Behavior of Recreational Bicyclists: an Empirical Study in Danshuei and Bali, Northern Taiwan
Authors: Chih-Cheng Chen, Meng-Lung Lin, Yung-Tan Lee Tsun-I Hsiao, Yi-Cheng Chen
Abstract:
The Taiwan government has invested approximately 21 billion NT dollars in the construction of bicycle paths since bicycling has gained huge popularity as a healthy leisure and recreational activity. This study focuses on the behavior of recreational bicyclists in Danshuei and Bali, northern Taiwan. Data were collected from a field investigation carried out along the Danshuei bicycle path and Bali left-bank bicycle path. A total of 578 questionnaires were gathered for data analysis. Descriptive statistics and Chi-Square tests were used to assess bicyclists- behaviors. The frequency shows that, in these areas, Danshuei and Bali, most bicyclists rented bicycles, rode the bicycle path in the afternoon for about 2 hours. The used the bicycle path one time per week. For most, it was the first time to ride these bicycle paths. There were significant differences in distribution of bicycle ownership, time of day, duration of ride, ride frequency, and whether riding occurred on weekdays or weekends. Results indicated that most bicyclists in Danshuei and Bali were infrequent users.Keywords: recreational bicyclist, bicyclist behaviors, NorthernTaiwan
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 153412839 Dynamic-Stochastic Influence Diagrams: Integrating Time-Slices IDs and Discrete Event Systems Modeling
Authors: Xin Zhao, Yin-fan Zhu, Wei-ping Wang, Qun Li
Abstract:
The Influence Diagrams (IDs) is a kind of Probabilistic Belief Networks for graphic modeling. The usage of IDs can improve the communication among field experts, modelers, and decision makers, by showing the issue frame discussed from a high-level point of view. This paper enhances the Time-Sliced Influence Diagrams (TSIDs, or called Dynamic IDs) based formalism from a Discrete Event Systems Modeling and Simulation (DES M&S) perspective, for Exploring Analysis (EA) modeling. The enhancements enable a modeler to specify times occurred of endogenous events dynamically with stochastic sampling as model running and to describe the inter- influences among them with variable nodes in a dynamic situation that the existing TSIDs fails to capture. The new class of model is named Dynamic-Stochastic Influence Diagrams (DSIDs). The paper includes a description of the modeling formalism and the hiberarchy simulators implementing its simulation algorithm, and shows a case study to illustrate its enhancements.
Keywords: Time-sliced influence diagrams, discrete event systems, dynamic-stochastic influence diagrams, modeling formalism, simulation algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 143812838 An Enhanced Distributed System to improve theTime Complexity of Binary Indexed Trees
Authors: Ahmed M. Elhabashy, A. Baes Mohamed, Abou El Nasr Mohamad
Abstract:
Distributed Computing Systems are usually considered the most suitable model for practical solutions of many parallel algorithms. In this paper an enhanced distributed system is presented to improve the time complexity of Binary Indexed Trees (BIT). The proposed system uses multi-uniform processors with identical architectures and a specially designed distributed memory system. The analysis of this system has shown that it has reduced the time complexity of the read query to O(Log(Log(N))), and the update query to constant complexity, while the naive solution has a time complexity of O(Log(N)) for both queries. The system was implemented and simulated using VHDL and Verilog Hardware Description Languages, with xilinx ISE 10.1, as the development environment and ModelSim 6.1c, similarly as the simulation tool. The simulation has shown that the overhead resulting by the wiring and communication between the system fragments could be fairly neglected, which makes it applicable to practically reach the maximum speed up offered by the proposed model.
Keywords: Binary Index Tree (BIT), Least Significant Bit (LSB), Parallel Adder (PA), Very High Speed Integrated Circuits HardwareDescription Language (VHDL), Distributed Parallel Computing System(DPCS).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1779