Search results for: analytical network design model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 30388

Search results for: analytical network design model

27088 Construction of the Large Scale Biological Networks from Microarrays

Authors: Fadhl Alakwaa

Abstract:

One of the sustainable goals of the system biology is understanding gene-gene interactions. Hence, gene regulatory networks (GRN) need to be constructed for understanding the disease ontology and to reduce the cost of drug development. To construct gene regulatory from gene expression we need to overcome many challenges such as data denoising and dimensionality. In this paper, we develop an integrated system to reduce data dimension and remove the noise. The generated network from our system was validated via available interaction databases and was compared to previous methods. The result revealed the performance of our proposed method.

Keywords: gene regulatory network, biclustering, denoising, system biology

Procedia PDF Downloads 239
27087 Culvert Blockage Evaluation Using Australian Rainfall And Runoff 2019

Authors: Rob Leslie, Taher Karimian

Abstract:

The blockage of cross drainage structures is a risk that needs to be understood and managed or lessened through the design. A blockage is a random event, influenced by site-specific factors, which needs to be quantified for design. Under and overestimation of blockage can have major impacts on flood risk and cost associated with drainage structures. The importance of this matter is heightened for those projects located within sensitive lands. It is a particularly complex problem for large linear infrastructure projects (e.g., rail corridors) located within floodplains where blockage factors can influence flooding upstream and downstream of the infrastructure. The selection of the appropriate blockage factors for hydraulic modeling has been subject to extensive research by hydraulic engineers. This paper has been prepared to review the current Australian Rainfall and Runoff 2019 (ARR 2019) methodology for blockage assessment by applying this method to a transport corridor brownfield upgrade case study in New South Wales. The results of applying the method are also validated against asset data and maintenance records. ARR 2019 – Book 6, Chapter 6 includes advice and an approach for estimating the blockage of bridges and culverts. This paper concentrates specifically on the blockage of cross drainage structures. The method has been developed to estimate the blockage level for culverts affected by sediment or debris due to flooding. The objective of the approach is to evaluate a numerical blockage factor that can be utilized in a hydraulic assessment of cross drainage structures. The project included an assessment of over 200 cross drainage structures. In order to estimate a blockage factor for use in the hydraulic model, a process has been advanced that considers the qualitative factors (e.g., Debris type, debris availability) and site-specific hydraulic factors that influence blockage. A site rating associated with the debris potential (i.e., availability, transportability, mobility) at each crossing was completed using the method outlined in ARR 2019 guidelines. The hydraulic results inputs (i.e., flow velocity, flow depth) and qualitative factors at each crossing were developed into an advanced spreadsheet where the design blockage level for cross drainage structures were determined based on the condition relating Inlet Clear Width and L10 (average length of the longest 10% of the debris reaching the site) and the Adjusted Debris Potential. Asset data, including site photos and maintenance records, were then reviewed and compared with the blockage assessment to check the validity of the results. The results of this assessment demonstrate that the estimated blockage factors at each crossing location using ARR 2019 guidelines are well-validated with the asset data. The primary finding of the study is that the ARR 2019 methodology is a suitable approach for culvert blockage assessment that has been validated against a case study spanning a large geographical area and multiple sub-catchments. The study also found that the methodology can be effectively coded within a spreadsheet or similar analytical tool to automate its application.

Keywords: ARR 2019, blockage, culverts, methodology

Procedia PDF Downloads 364
27086 Combining a Continuum of Hidden Regimes and a Heteroskedastic Three-Factor Model in Option Pricing

Authors: Rachid Belhachemi, Pierre Rostan, Alexandra Rostan

Abstract:

This paper develops a discrete-time option pricing model for index options. The model consists of two key ingredients. First, daily stock return innovations are driven by a continuous hidden threshold mixed skew-normal (HTSN) distribution which generates conditional non-normality that is needed to fit daily index return. The most important feature of the HTSN is the inclusion of a latent state variable with a continuum of states, unlike the traditional mixture distributions where the state variable is discrete with little number of states. The HTSN distribution belongs to the class of univariate probability distributions where parameters of the distribution capture the dependence between the variable of interest and the continuous latent state variable (the regime). The distribution has an interpretation in terms of a mixture distribution with time-varying mixing probabilities. It has been shown empirically that this distribution outperforms its main competitor, the mixed normal (MN) distribution, in terms of capturing the stylized facts known for stock returns, namely, volatility clustering, leverage effect, skewness, kurtosis and regime dependence. Second, heteroscedasticity in the model is captured by a threeexogenous-factor GARCH model (GARCHX), where the factors are taken from the principal components analysis of various world indices and presents an application to option pricing. The factors of the GARCHX model are extracted from a matrix of world indices applying principal component analysis (PCA). The empirically determined factors are uncorrelated and represent truly different common components driving the returns. Both factors and the eight parameters inherent to the HTSN distribution aim at capturing the impact of the state of the economy on price levels since distribution parameters have economic interpretations in terms of conditional volatilities and correlations of the returns with the hidden continuous state. The PCA identifies statistically independent factors affecting the random evolution of a given pool of assets -in our paper a pool of international stock indices- and sorting them by order of relative importance. The PCA computes a historical cross asset covariance matrix and identifies principal components representing independent factors. In our paper, factors are used to calibrate the HTSN-GARCHX model and are ultimately responsible for the nature of the distribution of random variables being generated. We benchmark our model to the MN-GARCHX model following the same PCA methodology and the standard Black-Scholes model. We show that our model outperforms the benchmark in terms of RMSE in dollar losses for put and call options, which in turn outperforms the analytical Black-Scholes by capturing the stylized facts known for index returns, namely, volatility clustering, leverage effect, skewness, kurtosis and regime dependence.

Keywords: continuous hidden threshold, factor models, GARCHX models, option pricing, risk-premium

Procedia PDF Downloads 297
27085 Create a Brand Value Assessment Model to Choosing a Cosmetic Brand in Tehran Combining DEMATEL Techniques and Multi-Stage ANFIS

Authors: Hamed Saremi, Suzan Taghavy, Seyed Mohammad Hanif Sanjari, Mostafa Kahali

Abstract:

One of the challenges in manufacturing and service companies to provide a product or service is recognized Brand to consumers in target markets. They provide most of their processes under the same capacity. But the constant threat of devastating internal and external resources to prevent a rise Brands and more companies are recognizing the stages are bankrupt. This paper has tried to identify and analyze effective indicators of brand equity and focuses on indicators and presents a model of intelligent create a model to prevent possible damage. In this study, the identified indicators of brand equity are based on literature study and according to expert opinions, set of indicators By techniques DEMATEL Then to used Multi-Step Adaptive Neural-Fuzzy Inference system (ANFIS) to design a multi-stage intelligent system for assessment of brand equity.

Keywords: brand, cosmetic product, ANFIS, DEMATEL

Procedia PDF Downloads 417
27084 Comparative Analysis of Sigmoidal Feedforward Artificial Neural Networks and Radial Basis Function Networks Approach for Localization in Wireless Sensor Networks

Authors: Ashish Payal, C. S. Rai, B. V. R. Reddy

Abstract:

With the increasing use and application of Wireless Sensor Networks (WSN), need has arisen to explore them in more effective and efficient manner. An important area which can bring efficiency to WSNs is the localization process, which refers to the estimation of the position of wireless sensor nodes in an ad hoc network setting, in reference to a coordinate system that may be internal or external to the network. In this paper, we have done comparison and analysed Sigmoidal Feedforward Artificial Neural Networks (SFFANNs) and Radial Basis Function (RBF) networks for developing localization framework in WSNs. The presented work utilizes the Received Signal Strength Indicator (RSSI), measured by static node on 100 x 100 m2 grid from three anchor nodes. The comprehensive evaluation of these approaches is done using MATLAB software. The simulation results effectively demonstrate that FFANNs based sensor motes will show better localization accuracy as compared to RBF.

Keywords: localization, wireless sensor networks, artificial neural network, radial basis function, multi-layer perceptron, backpropagation, RSSI, GPS

Procedia PDF Downloads 339
27083 A Mathematical Equation to Calculate Stock Price of Different Growth Model

Authors: Weiping Liu

Abstract:

This paper presents an equation to calculate stock prices of different growth model. This equation is mathematically derived by using discounted cash flow method. It has the advantages of being very easy to use and very accurate. It can still be used even when the first stage is lengthy. This equation is more generalized because it can be used for all the three popular stock price models. It can be programmed into financial calculator or electronic spreadsheets. In addition, it can be extended to a multistage model. It is more versatile and efficient than the traditional methods.

Keywords: stock price, multistage model, different growth model, discounted cash flow method

Procedia PDF Downloads 406
27082 The Role of Artificial Intelligence on Interior Space in College of Architecture and Design

Authors: Saif M. M. Obeidat

Abstract:

This research investigates the impact of artificial intelligence (AI) on interior spaces within a college of Architecture and Design. Employing a qualitative approach, the study conducts in-depth interviews and reviews AI-integrated design projects within the academic setting. The key objectives include assessing AI integration in design processes, examining the influence of AI on user experience, exploring its role in architectural innovation, identifying challenges, and assessing educational implications. The study aims to provide a comprehensive understanding of AI's role in shaping interior spaces within academia. It anticipates improved efficiency in design processes, positive user feedback on functionality and experiences, the emergence of innovative design solutions, and the identification of challenges like ethical considerations and technical limitations. Additionally, the research expects insights into how educational programs may need to adapt to incorporate AI knowledge and skills, ensuring students are well-prepared for the evolving landscape of architecture and design practice. By addressing these objectives, the research contributes valuable insights into the evolving relationship between technology and the field of architecture, particularly within educational contexts.

Keywords: interior design, artificial intelligence, academic settings, technology, education

Procedia PDF Downloads 95
27081 Smart Water Main Inspection and Condition Assessment Using a Systematic Approach for Pipes Selection

Authors: Reza Moslemi, Sebastien Perrier

Abstract:

Water infrastructure deterioration can result in increased operational costs owing to increased repair needs and non-revenue water and consequently cause a reduced level of service and customer service satisfaction. Various water main condition assessment technologies have been introduced to the market in order to evaluate the level of pipe deterioration and to develop appropriate asset management and pipe renewal plans. One of the challenges for any condition assessment and inspection program is to determine the percentage of the water network and the combination of pipe segments to be inspected in order to obtain a meaningful representation of the status of the entire water network with a desirable level of accuracy. Traditionally, condition assessment has been conducted by selecting pipes based on age or location. However, this may not necessarily offer the best approach, and it is believed that by using a smart sampling methodology, a better and more reliable estimate of the condition of a water network can be achieved. This research investigates three different sampling methodologies, including random, stratified, and systematic. It is demonstrated that selecting pipes based on the proposed clustering and sampling scheme can considerably improve the ability of the inspected subset to represent the condition of a wider network. With a smart sampling methodology, a smaller data sample can provide the same insight as a larger sample. This methodology offers increased efficiency and cost savings for condition assessment processes and projects.

Keywords: condition assessment, pipe degradation, sampling, water main

Procedia PDF Downloads 150
27080 Optimal Design of Step-Stress Partially Life Test Using Multiply Censored Exponential Data with Random Removals

Authors: Showkat Ahmad Lone, Ahmadur Rahman, Ariful Islam

Abstract:

The major assumption in accelerated life tests (ALT) is that the mathematical model relating the lifetime of a test unit and the stress are known or can be assumed. In some cases, such life–stress relationships are not known and cannot be assumed, i.e. ALT data cannot be extrapolated to use condition. So, in such cases, partially accelerated life test (PALT) is a more suitable test to be performed for which tested units are subjected to both normal and accelerated conditions. This study deals with estimating information about failure times of items under step-stress partially accelerated life tests using progressive failure-censored hybrid data with random removals. The life data of the units under test is considered to follow exponential life distribution. The removals from the test are assumed to have binomial distributions. The point and interval maximum likelihood estimations are obtained for unknown distribution parameters and tampering coefficient. An optimum test plan is developed using the D-optimality criterion. The performances of the resulting estimators of the developed model parameters are evaluated and investigated by using a simulation algorithm.

Keywords: binomial distribution, d-optimality, multiple censoring, optimal design, partially accelerated life testing, simulation study

Procedia PDF Downloads 320
27079 Generative Adversarial Network for Bidirectional Mappings between Retinal Fundus Images and Vessel Segmented Images

Authors: Haoqi Gao, Koichi Ogawara

Abstract:

Retinal vascular segmentation of color fundus is the basis of ophthalmic computer-aided diagnosis and large-scale disease screening systems. Early screening of fundus diseases has great value for clinical medical diagnosis. The traditional methods depend on the experience of the doctor, which is time-consuming, labor-intensive, and inefficient. Furthermore, medical images are scarce and fraught with legal concerns regarding patient privacy. In this paper, we propose a new Generative Adversarial Network based on CycleGAN for retinal fundus images. This method can generate not only synthetic fundus images but also generate corresponding segmentation masks, which has certain application value and challenge in computer vision and computer graphics. In the results, we evaluate our proposed method from both quantitative and qualitative. For generated segmented images, our method achieves dice coefficient of 0.81 and PR of 0.89 on DRIVE dataset. For generated synthetic fundus images, we use ”Toy Experiment” to verify the state-of-the-art performance of our method.

Keywords: retinal vascular segmentations, generative ad-versarial network, cyclegan, fundus images

Procedia PDF Downloads 144
27078 Resource Orchestration Based on Two-Sides Scheduling in Computing Network Control Sytems

Authors: Li Guo, Jianhong Wang, Dian Huang, Shengzhong Feng

Abstract:

Computing networks as a new network architecture has shown great promise in boosting the utilization of different resources, such as computing, caching, and communications. To maximise the efficiency of resource orchestration in computing network control systems (CNCSs), this work proposes a dynamic orchestration strategy of a different resource based on task requirements from computing power requestors (CPRs). Specifically, computing power providers (CPPs) in CNCSs could share information with each other through communication channels on the basis of blockchain technology, especially their current idle resources. This dynamic process is modeled as a cooperative game in which CPPs have the same target of maximising long-term rewards by improving the resource utilization ratio. Meanwhile, the task requirements from CPRs, including size, deadline, and calculation, are simultaneously considered in this paper. According to task requirements, the proposed orchestration strategy could schedule the best-fitting resource in CNCSs, achieving the maximum long-term rewards of CPPs and the best quality of experience (QoE) of CRRs at the same time. Based on the EdgeCloudSim simulation platform, the efficiency of the proposed strategy is achieved from both sides of CPRs and CPPs. Besides, experimental results show that the proposed strategy outperforms the other comparisons in all cases.

Keywords: computing network control systems, resource orchestration, dynamic scheduling, blockchain, cooperative game

Procedia PDF Downloads 115
27077 Statistical Characteristics of Code Formula for Design of Concrete Structures

Authors: Inyeol Paik, Ah-Ryang Kim

Abstract:

In this research, a statistical analysis is carried out to examine the statistical properties of the formula given in the design code for concrete structures. The design formulas of the Korea highway bridge design code - the limit state design method (KHBDC) which is the current national bridge design code and the design code for concrete structures by Korea Concrete Institute (KCI) are applied for the analysis. The safety levels provided by the strength formulas of the design codes are defined based on the probabilistic and statistical theory.KHBDC is a reliability-based design code. The load and resistance factors of this code were calibrated to attain the target reliability index. It is essential to define the statistical properties for the design formulas in this calibration process. In general, the statistical characteristics of a member strength are due to the following three factors. The first is due to the difference between the material strength of the actual construction and that used in the design calculation. The second is the difference between the actual dimensions of the constructed sections and those used in design calculation. The third is the difference between the strength of the actual member and the formula simplified for the design calculation. In this paper, the statistical study is focused on the third difference. The formulas for calculating the shear strength of concrete members are presented in different ways in KHBDC and KCI. In this study, the statistical properties of design formulas were obtained through comparison with the database which comprises the experimental results from the reference publications. The test specimen was either reinforced with the shear stirrup or not. For an applied database, the bias factor was about 1.12 and the coefficient of variation was about 0.18. By applying the statistical properties of the design formula to the reliability analysis, it is shown that the resistance factors of the current design codes satisfy the target reliability indexes of both codes. Also, the minimum resistance factors of the KHBDC which is written in the material resistance factor format and KCE which is in the member resistance format are obtained and the results are presented. A further research is underway to calibrate the resistance factors of the high strength and high-performance concrete design guide.

Keywords: concrete design code, reliability analysis, resistance factor, shear strength, statistical property

Procedia PDF Downloads 319
27076 Design and Implementation of a Lab Bench for Synthetic Aperture Radar Imaging System

Authors: Karthiyayini Nagarajan, P. V. RamaKrishna

Abstract:

Radar Imaging techniques provides extensive applications in the field of remote sensing, majorly Synthetic Aperture Radar(SAR) that provide high resolution target images. This paper work puts forward the effective and realizable signal generation and processing for SAR images. The major units in the system include camera, signal generation unit, signal processing unit and display screen. The real radio channel is replaced by its mathematical model based on optical image to calculate a reflected signal model in real time. Signal generation realizes the algorithm and forms the radar reflection model. Signal processing unit provides range and azimuth resolution through matched filtering and spectrum analysis procedure to form radar image on the display screen. The restored image has the same quality as that of the optical image. This SAR imaging system has been designed and implemented using MATLAB and Quartus II tools on Stratix III device as a System(lab bench) that works in real time to study/investigate on radar imaging rudiments and signal processing scheme for educational and research purposes.

Keywords: synthetic aperture radar, radio reflection model, lab bench

Procedia PDF Downloads 468
27075 Lead Removal From Ex- Mining Pond Water by Electrocoagulation: Kinetics, Isotherm, and Dynamic Studies

Authors: Kalu Uka Orji, Nasiman Sapari, Khamaruzaman W. Yusof

Abstract:

Exposure of galena (PbS), tealite (PbSnS2), and other associated minerals during mining activities release lead (Pb) and other heavy metals into the mining water through oxidation and dissolution. Heavy metal pollution has become an environmental challenge. Lead, for instance, can cause toxic effects to human health, including brain damage. Ex-mining pond water was reported to contain lead as high as 69.46 mg/L. Conventional treatment does not easily remove lead from water. A promising and emerging treatment technology for lead removal is the application of the electrocoagulation (EC) process. However, some of the problems associated with EC are systematic reactor design, selection of maximum EC operating parameters, scale-up, among others. This study investigated an EC process for the removal of lead from synthetic ex-mining pond water using a batch reactor and Fe electrodes. The effects of various operating parameters on lead removal efficiency were examined. The results obtained indicated that the maximum removal efficiency of 98.6% was achieved at an initial PH of 9, the current density of 15mA/cm2, electrode spacing of 0.3cm, treatment time of 60 minutes, Liquid Motion of Magnetic Stirring (LM-MS), and electrode arrangement = BP-S. The above experimental data were further modeled and optimized using a 2-Level 4-Factor Full Factorial design, a Response Surface Methodology (RSM). The four factors optimized were the current density, electrode spacing, electrode arrangements, and Liquid Motion Driving Mode (LM). Based on the regression model and the analysis of variance (ANOVA) at 0.01%, the results showed that an increase in current density and LM-MS increased the removal efficiency while the reverse was the case for electrode spacing. The model predicted the optimal lead removal efficiency of 99.962% with an electrode spacing of 0.38 cm alongside others. Applying the predicted parameters, the lead removal efficiency of 100% was actualized. The electrode and energy consumptions were 0.192kg/m3 and 2.56 kWh/m3 respectively. Meanwhile, the adsorption kinetic studies indicated that the overall lead adsorption system belongs to the pseudo-second-order kinetic model. The adsorption dynamics were also random, spontaneous, and endothermic. The higher temperature of the process enhances adsorption capacity. Furthermore, the adsorption isotherm fitted the Freundlish model more than the Langmuir model; describing the adsorption on a heterogeneous surface and showed good adsorption efficiency by the Fe electrodes. Adsorption of Pb2+ onto the Fe electrodes was a complex reaction, involving more than one mechanism. The overall results proved that EC is an efficient technique for lead removal from synthetic mining pond water. The findings of this study would have application in the scale-up of EC reactor and in the design of water treatment plants for feed-water sources that contain lead using the electrocoagulation method.

Keywords: ex-mining water, electrocoagulation, lead, adsorption kinetics

Procedia PDF Downloads 149
27074 Magnet Position Variation of the Electromagnetic Actuation System in a Torsional Scanner

Authors: Loke Kean Koay, Mani Maran Ratnam

Abstract:

A mechanically-resonant torsional spring scanner was developed in a recent study. Various methods were developed to improve the angular displacement of the scanner while maintaining the scanner frequency. However, the effects of rotor magnet radial position on scanner characteristics were not well investigated. In this study, the relationships between the magnet position and the scanner characteristics such as natural frequency, angular displacement and stress level were studied. A finite element model was created and an average deviation of 3.18% was found between the simulation and experimental results, qualifying the simulation results as a guide for further investigations. Three magnet positions on the transverse oscillating suspended plate were investigated by finite element analysis (FEA) and one of the positions were selected as the design position. The magnet position with the longest distance from the twist axis of the mirror was selected since it attains minimum stress level while exceeding the minimum critical flicker frequency and delivering the targeted angular displacement to the scanner.

Keywords: torsional scanner, design optimization, computer-aided design, magnet position variation

Procedia PDF Downloads 366
27073 Heat Transfer Characteristics of Film Condensation

Authors: M. Mosaad, J. H. Almutairi, A. S. Almutairi

Abstract:

In this paper, saturated-vapour film condensation on a vertical wall with the backside cooled by forced convection is analyzed as a conjugate problem. In the analysis, the temperature and heat flux at the wall sides are assumed unknown and determined from the solution. The model is presented in a dimensionless form to take a broad view of the solution. The dimensionless variables controlling this coupled heat transfer process are discovered from the analysis. These variables explain the relative impact of the interactive heat transfer mechanisms of forced convection and film condensation. The study shows that the conjugate treatment of film condensation process yields results different from that predicted by a non-conjugate Nusselt-type solution, wherein the effect of the cooling fluid is neglected.

Keywords: film condensation, forced convection, coupled heat transfer, analytical modelling

Procedia PDF Downloads 321
27072 Numerical Solutions of Fractional Order Epidemic Model

Authors: Sadia Arshad, Ayesha Sohail, Sana Javed, Khadija Maqbool, Salma Kanwal

Abstract:

The dynamical study of the carriers play an essential role in the evolution and global transmission of infectious diseases and will be discussed in this study. To make this approach novel, we will consider the fractional order model which is generalization of integer order derivative to an arbitrary number. Since the integration involved is non local therefore this property of fractional operator is very useful to study epidemic model for infectious diseases. An extended numerical method (ODE solver) is implemented on the model equations and we will present the simulations of the model for different values of fractional order to study the effect of carriers on transmission dynamics. Global dynamics of fractional model are established by using the reproduction number.

Keywords: Fractional differential equation, Numerical simulations, epidemic model, transmission dynamics

Procedia PDF Downloads 602
27071 Experimental and Analytical Dose Assessment of Patient's Family Members Treated with I-131

Authors: Marzieh Ebrahimi, Vahid Changizi, Mohammad Reza Kardan, Seyed Mahdi Hosseini Pooya, Parham Geramifar

Abstract:

Radiation exposure to the patient's family members is one of the major concerns during thyroid cancer radionuclide therapy. The aim of this study was to measure the total effective dose of the family members by means of thermoluminescence personal dosimeter, and compare with those calculated by analytical methods. Eighty-five adult family members of fifty-one patients volunteered to participate in this research study. Considering the minimum and maximum range of dose rate from 15 µsv/h to 120 µsv/h at patients' release time, the calculated mean and median dose values of family members were 0.45 mSv and 0.28 mSv, respectively. Moreover, almost all family members’ doses were measured to be less than the dose constraint of 5 mSv recommended by Basic Safety Standards. Considering the influence parameters such as patient dose rate and administrated activity, the total effective doses of family members were calculated by TEDE and NRC formulas and compared with those of experimental results. The results indicated that, it is fruitful to use the quantitative calculations for releasing patients treated with I-131 and correct estimation of patients' family doses.

Keywords: effective dose, thermoluminescence, I-131, thyroid cancer

Procedia PDF Downloads 400
27070 A Mathematical Model for Reliability Redundancy Optimization Problem of K-Out-Of-N: G System

Authors: Gak-Gyu Kim, Won Il Jung

Abstract:

According to a remarkable development of science and technology, function and role of the system of engineering fields has recently been diversified. The system has become increasingly more complex and precise, and thus, system designers intended to maximize reliability concentrate more effort at the design stage. This study deals with the reliability redundancy optimization problem (RROP) for k-out-of-n: G system configuration with cold standby and warm standby components. This paper further intends to present the optimal mathematical model through which the following three elements of (i) multiple components choices, (ii) redundant components quantity and (iii) the choice of redundancy strategies may be combined in order to maximize the reliability of the system. Therefore, we focus on the following three issues. First, we consider RROP that there exists warm standby state as well as cold standby state of the component. Second, as eliminating an approximation approach of the previous RROP studies, we construct a precise model for system reliability. Third, given transition time when the state of components changes, we present not simply a workable solution but the advanced method. For the wide applicability of RROPs, moreover, we use absorbing continuous time Markov chain and matrix analytic methods in the suggested mathematical model.

Keywords: RROP, matrix analytic methods, k-out-of-n: G system, MTTF, absorbing continuous time Markov Chain

Procedia PDF Downloads 254
27069 Machine Learning Based Gender Identification of Authors of Entry Programs

Authors: Go Woon Kwak, Siyoung Jun, Soyun Maeng, Haeyoung Lee

Abstract:

Entry is an education platform used in South Korea, created to help students learn to program, in which they can learn to code while playing. Using the online version of the entry, teachers can easily assign programming homework to the student and the students can make programs simply by linking programming blocks. However, the programs may be made by others, so that the authors of the programs should be identified. In this paper, as the first step toward author identification of entry programs, we present an artificial neural network based classification approach to identify genders of authors of a program written in an entry. A neural network has been trained from labeled training data that we have collected. Our result in progress, although preliminary, shows that the proposed approach could be feasible to be applied to the online version of entry for gender identification of authors. As future work, we will first use a machine learning technique for age identification of entry programs, which would be the second step toward the author identification.

Keywords: artificial intelligence, author identification, deep neural network, gender identification, machine learning

Procedia PDF Downloads 324
27068 Effects of Cash Transfers Mitigation Impacts in the Face of Socioeconomic External Shocks: Evidence from Egypt

Authors: Basma Yassa

Abstract:

Evidence on cash transfers’ effectiveness in mitigating macro and idiosyncratic shocks’ impacts has been mixed and is mostly concentrated in Latin America, Sub-Saharan Africa, and South Asia with very limited evidence from the MENA region. Yet conditional cash transfers schemes have been continually used, especially in Egypt, as the main social protection tool in response to the recent socioeconomic crises and macro shocks. We use 2 panel datasets and 1 cross-sectional dataset to estimate the effectiveness of cash transfers as a shock-mitigative mechanism in the Egyptian context. In this paper, the results from the different models (Panel Fixed Effects model and the Regression Discontinuity Design (RDD) model) confirm that micro and macro shocks lead to significant decline in several household-level welfare outcomes and that Takaful cash transfers have a significant positive impact in mitigating the negative shock impacts, especially on households’ debt incidence, debt levels, and asset ownership, but not necessarily on food, and non-food expenditure levels. The results indicate large positive significant effects on decreasing household incidence of debt by up to 12.4 percent and lowered the debt size by approximately 18 percent among Takaful beneficiaries compared to non-beneficiaries’. Similar evidence is found on asset ownership levels, as the RDD model shows significant positive effects on total asset ownership and productive asset ownership, but the model failed to detect positive impacts on per capita food and non-food expenditures. Further extensions are still in progress to compare the models’ results with the DID model results when using a nationally representative ELMPS panel data (2018/2024) rounds. Finally, our initial analysis suggests that conditional cash transfers are effective in buffering the negative shock impacts on certain welfare indicators even after successive macro-economic shocks in 2022 and 2023 in the Egyptian Context.

Keywords: cash transfers, fixed effects, household welfare, household debt, micro shocks, regression discontinuity design

Procedia PDF Downloads 47
27067 Ranking of Performance Measures of GSCM towards Sustainability: Using Analytic Hierarchy Process

Authors: Dixit Garg, S. Luthra, A. Haleem

Abstract:

During recent years, the natural environment has become a challenging topic that business organizations must consider due to the economic and ecological impacts and increasing awareness of environment protection among society. Organizations are trying to achieve the goals of improvement in environment, low cost, high quality, flexibility and more customer satisfaction. Performance measurement frameworks are very useful to monitor the performance of any organization. The basic goal of this paper is to identify performance measures and ranking of these performance measures of GSCM performance measurement towards sustainability framework. Five perspectives (Environment, Economic, Social, Operational and Cost performances) and nineteen performance measures of GSCM performance towards sustainability have been have been identified from extensive literature review. Analytical Hierarchy Process (AHP) technique has been utilized for ranking of these performance perspectives and measures. All pair comparisons in AHP have been made on the basis on the experts’ opinions (selected from academia and industry). Ranking of these performance perspectives and measures will help to understand the importance of environmental, economic, social, operational performances, and cost performances in the supply chain.

Keywords: analytical hierarchy process, green supply chain management, performance measures, sustainability

Procedia PDF Downloads 519
27066 Fair Federated Learning in Wireless Communications

Authors: Shayan Mohajer Hamidi

Abstract:

Federated Learning (FL) has emerged as a promising paradigm for training machine learning models on distributed data without the need for centralized data aggregation. In the realm of wireless communications, FL has the potential to leverage the vast amounts of data generated by wireless devices to improve model performance and enable intelligent applications. However, the fairness aspect of FL in wireless communications remains largely unexplored. This abstract presents an idea for fair federated learning in wireless communications, addressing the challenges of imbalanced data distribution, privacy preservation, and resource allocation. Firstly, the proposed approach aims to tackle the issue of imbalanced data distribution in wireless networks. In typical FL scenarios, the distribution of data across wireless devices can be highly skewed, resulting in unfair model updates. To address this, we propose a weighted aggregation strategy that assigns higher importance to devices with fewer samples during the aggregation process. By incorporating fairness-aware weighting mechanisms, the proposed approach ensures that each participating device's contribution is proportional to its data distribution, thereby mitigating the impact of data imbalance on model performance. Secondly, privacy preservation is a critical concern in federated learning, especially in wireless communications where sensitive user data is involved. The proposed approach incorporates privacy-enhancing techniques, such as differential privacy, to protect user privacy during the model training process. By adding carefully calibrated noise to the gradient updates, the proposed approach ensures that the privacy of individual devices is preserved without compromising the overall model accuracy. Moreover, the approach considers the heterogeneity of devices in terms of computational capabilities and energy constraints, allowing devices to adaptively adjust the level of privacy preservation to strike a balance between privacy and utility. Thirdly, efficient resource allocation is crucial for federated learning in wireless communications, as devices operate under limited bandwidth, energy, and computational resources. The proposed approach leverages optimization techniques to allocate resources effectively among the participating devices, considering factors such as data quality, network conditions, and device capabilities. By intelligently distributing the computational load, communication bandwidth, and energy consumption, the proposed approach minimizes resource wastage and ensures a fair and efficient FL process in wireless networks. To evaluate the performance of the proposed fair federated learning approach, extensive simulations and experiments will be conducted. The experiments will involve a diverse set of wireless devices, ranging from smartphones to Internet of Things (IoT) devices, operating in various scenarios with different data distributions and network conditions. The evaluation metrics will include model accuracy, fairness measures, privacy preservation, and resource utilization. The expected outcomes of this research include improved model performance, fair allocation of resources, enhanced privacy preservation, and a better understanding of the challenges and solutions for fair federated learning in wireless communications. The proposed approach has the potential to revolutionize wireless communication systems by enabling intelligent applications while addressing fairness concerns and preserving user privacy.

Keywords: federated learning, wireless communications, fairness, imbalanced data, privacy preservation, resource allocation, differential privacy, optimization

Procedia PDF Downloads 75
27065 An Analysis and Design of Mobile Payment System Based on NFC Technology

Authors: Shafiq ur Rehman, Zubair Ahmed Shaikh

Abstract:

This research provides the comparative study of different mobile payment system and proposes an efficient solution of mobile payment system. The research involves discovering how the mobile payment methods can be used and implemented keeping user and system interaction under consideration. The implementation of Nielsen’s heuristic and universal design principles enhanced the user’s interaction design and made the system more appropriate, understandable and visible to the end user. The design of application is greatly affected by the user driven factors. These factors help in the efficiency of the application usage.

Keywords: mobile payment system, m-commerce, usability, near field communication

Procedia PDF Downloads 453
27064 Seismic Fragility Assessment of Continuous Integral Bridge Frames with Variable Expansion Joint Clearances

Authors: P. Mounnarath, U. Schmitz, Ch. Zhang

Abstract:

Fragility analysis is an effective tool for the seismic vulnerability assessment of civil structures in the last several years. The design of the expansion joints according to various bridge design codes is almost inconsistent, and only a few studies have focused on this problem so far. In this study, the influence of the expansion joint clearances between the girder ends and the abutment backwalls on the seismic fragility assessment of continuous integral bridge frames is investigated. The gaps (ranging from 60 mm, 150 mm, 250 mm and 350 mm) are designed by following two different bridge design code specifications, namely, Caltrans and Eurocode 8-2. Five bridge models are analyzed and compared. The first bridge model serves as a reference. This model uses three-dimensional reinforced concrete fiber beam-column elements with simplified supports at both ends of the girder. The other four models also employ reinforced concrete fiber beam-column elements but include the abutment backfill stiffness and four different gap values. The nonlinear time history analysis is performed. The artificial ground motion sets, which have the peak ground accelerations (PGAs) ranging from 0.1 g to 1.0 g with an increment of 0.05 g, are taken as input. The soil-structure interaction and the P-Δ effects are also included in the analysis. The component fragility curves in terms of the curvature ductility demand to the capacity ratio of the piers and the displacement demand to the capacity ratio of the abutment sliding bearings are established and compared. The system fragility curves are then obtained by combining the component fragility curves. Our results show that in the component fragility analysis, the reference bridge model exhibits a severe vulnerability compared to that of other sophisticated bridge models for all damage states. In the system fragility analysis, the reference curves illustrate a smaller damage probability in the earlier PGA ranges for the first three damage states, they then show a higher fragility compared to other curves in the larger PGA levels. In the fourth damage state, the reference curve has the smallest vulnerability. In both the component and the system fragility analysis, the same trend is found that the bridge models with smaller clearances exhibit a smaller fragility compared to that with larger openings. However, the bridge model with a maximum clearance still induces a minimum pounding force effect.

Keywords: expansion joint clearance, fiber beam-column element, fragility assessment, time history analysis

Procedia PDF Downloads 435
27063 Performance Improvement of Information System of a Banking System Based on Integrated Resilience Engineering Design

Authors: S. H. Iranmanesh, L. Aliabadi, A. Mollajan

Abstract:

Integrated resilience engineering (IRE) is capable of returning banking systems to the normal state in extensive economic circumstances. In this study, information system of a large bank (with several branches) is assessed and optimized under severe economic conditions. Data envelopment analysis (DEA) models are employed to achieve the objective of this study. Nine IRE factors are considered to be the outputs, and a dummy variable is defined as the input of the DEA models. A standard questionnaire is designed and distributed among executive managers to be considered as the decision-making units (DMUs). Reliability and validity of the questionnaire is examined based on Cronbach's alpha and t-test. The most appropriate DEA model is determined based on average efficiency and normality test. It is shown that the proposed integrated design provides higher efficiency than the conventional RE design. Results of sensitivity and perturbation analysis indicate that self-organization, fault tolerance, and reporting culture respectively compose about 50 percent of total weight.

Keywords: banking system, Data Envelopment Analysis (DEA), Integrated Resilience Engineering (IRE), performance evaluation, perturbation analysis

Procedia PDF Downloads 188
27062 An Analytical Method for Maintenance Cost Estimating Relationships of Helicopters Using Linear Programming

Authors: Meesun Sun, Yongmin Kim

Abstract:

Estimating maintenance cost is crucial in defense management because it affects military budgets and availability of equipment. When it comes to estimating maintenance cost of the deployed equipment, time series forecasting can be applied with the actual historical cost data. It is more difficult issue to estimate maintenance cost of new equipment for which the actual costs are not provided. In this underlying context, this study proposes an analytical method for maintenance cost estimating relationships (CERs) development of helicopters using linear programming. The CERs can be applied to a new helicopter because they use non-cost independent variables such as the number of engines, the empty weight and so on. In the Republic of Korea, the maintenance cost of new equipment has been usually estimated by reflecting maintenance cost to unit price ratio of the legacy equipment. This study confirms that the CERs perform well for the 10 types of airmobile helicopters in terms of mean absolute percentage error by applying leave-one-out cross-validation. The suggested method is very useful to estimate the maintenance cost of new equipment and can help in the affordability assessment of acquisition program portfolios for total life cycle systems management.

Keywords: affordability analysis, cost estimating relationship, helicopter, linear programming, maintenance cost

Procedia PDF Downloads 139
27061 Learning Traffic Anomalies from Generative Models on Real-Time Observations

Authors: Fotis I. Giasemis, Alexandros Sopasakis

Abstract:

This study focuses on detecting traffic anomalies using generative models applied to real-time observations. By integrating a Graph Neural Network with an attention-based mechanism within the Spatiotemporal Generative Adversarial Network framework, we enhance the capture of both spatial and temporal dependencies in traffic data. Leveraging minute-by-minute observations from cameras distributed across Gothenburg, our approach provides a more detailed and precise anomaly detection system, effectively capturing the complex topology and dynamics of urban traffic networks.

Keywords: traffic, anomaly detection, GNN, GAN

Procedia PDF Downloads 8
27060 Thermal Buckling Response of Cylindrical Panels with Higher Order Shear Deformation Theory—a Case Study with Angle-Ply Laminations

Authors: Humayun R. H. Kabir

Abstract:

An analytical solution before used for static and free-vibration response has been extended for thermal buckling response on cylindrical panel with anti-symmetric laminations. The partial differential equations that govern kinematic behavior of shells produce five coupled differential equations. The basic displacement and rotational unknowns are similar to first order shear deformation theory---three displacement in spatial space, and two rotations about in-plane axes. No drilling degree of freedom is considered. Boundary conditions are considered as complete hinge in all edges so that the panel respond on thermal inductions. Two sets of double Fourier series are considered in the analytical solution process. The sets are selected that satisfy mixed type of natural boundary conditions. Numerical results are presented for the first 10 eigenvalues, and first 10 mode shapes for Ux, Uy, and Uz components. The numerical results are compared with a finite element based solution.

Keywords: higher order shear deformation, composite, thermal buckling, angle-ply laminations

Procedia PDF Downloads 373
27059 Hydrology and Hydraulics Analysis of Beko Abo Dam and Appurtenant Structre Design, Ethiopia

Authors: Azazhu Wassie

Abstract:

This study tried to evaluate the maximum design flood for appurtenance structure design using the given climatological and hydrological data analysis on the referenced study area. The maximum design flood is determined by using flood frequency analysis. Using this method, the peak discharge is 32,583.67 m3/s, but the data is transferred because the dam site is not on the gauged station. Then the peak discharge becomes 38,115 m3/s. The study was conducted in June 2023. This dam is built across a river to create a reservoir on its upstream side for impounding water. The water stored in the reservoir is used for various purposes, such as irrigation, hydropower, navigation, fishing, etc. The total average volume of annual runoff is estimated to be 115.1 billion m3. The total potential of the land for irrigation development can go beyond 3 million ha.

Keywords: dam design, flow duration curve, peak flood, rainfall, reservoir capacity, risk and reliability

Procedia PDF Downloads 29