Search results for: statistical estimation problem
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 12404

Search results for: statistical estimation problem

11564 Nurse Schedule Problem in Mubarak Al Kabeer Hospital

Authors: Khaled Al-Mansour, Nawaf Esmael, Abdulaziz Al-Zaid, Mohammed Al Ateeqi, Ali Al-Yousfi, Sayed Al-Zalzalah

Abstract:

In this project we will create the new schedule of nurse according to the preference of them. We did our project in Mubarak Al Kabeer Hospital (in Kuwait). The project aims to optimize the schedule of nurses in Mubarak Al Kabeer Hospital. The schedule of the nurses was studied and understood well to do any modification for their schedule to make the nurses feel as much comfort as they are. First constraints were found to know what things we can change and what things we can’t, the hard constraints are the hospital and ministry policies where we can’t change anything about, and the soft constraints are things that make nurses more comfortable. Data were collected and nurses were interviewed to know what is more better for them. All these constraints and date have been formulated to mathematical equations. This report will first contain an introduction to the topic which includes details of the problem definition. It will also contain information regarding the optimization of a nurse schedule and its contents and importance; furthermore, the report will contain information about the data needed to solve the problem and how it was collected. The problem requires formulation and that is also to be shown. The methodology will be explained which will state what has already been done. We used the lingo software to find the best schedule for the nurse. The schedule has been made according to what the nurses prefer, and also took consideration of the hospital policy when we make the schedule.

Keywords: nurse schedule problem, Kuwait, hospital policy, optimization of schedules

Procedia PDF Downloads 271
11563 Application of Multilinear Regression Analysis for Prediction of Synthetic Shear Wave Velocity Logs in Upper Assam Basin

Authors: Triveni Gogoi, Rima Chatterjee

Abstract:

Shear wave velocity (Vs) estimation is an important approach in the seismic exploration and characterization of a hydrocarbon reservoir. There are varying methods for prediction of S-wave velocity, if recorded S-wave log is not available. But all the available methods for Vs prediction are empirical mathematical models. Shear wave velocity can be estimated using P-wave velocity by applying Castagna’s equation, which is the most common approach. The constants used in Castagna’s equation vary for different lithologies and geological set-ups. In this study, multiple regression analysis has been used for estimation of S-wave velocity. The EMERGE module from Hampson-Russel software has been used here for generation of S-wave log. Both single attribute and multi attributes analysis have been carried out for generation of synthetic S-wave log in Upper Assam basin. Upper Assam basin situated in North Eastern India is one of the most important petroleum provinces of India. The present study was carried out using four wells of the study area. Out of these wells, S-wave velocity was available for three wells. The main objective of the present study is a prediction of shear wave velocities for wells where S-wave velocity information is not available. The three wells having S-wave velocity were first used to test the reliability of the method and the generated S-wave log was compared with actual S-wave log. Single attribute analysis has been carried out for these three wells within the depth range 1700-2100m, which corresponds to Barail group of Oligocene age. The Barail Group is the main target zone in this study, which is the primary producing reservoir of the basin. A system generated list of attributes with varying degrees of correlation appeared and the attribute with the highest correlation was concerned for the single attribute analysis. Crossplot between the attributes shows the variation of points from line of best fit. The final result of the analysis was compared with the available S-wave log, which shows a good visual fit with a correlation of 72%. Next multi-attribute analysis has been carried out for the same data using all the wells within the same analysis window. A high correlation of 85% has been observed between the output log from the analysis and the recorded S-wave. The almost perfect fit between the synthetic S-wave and the recorded S-wave log validates the reliability of the method. For further authentication, the generated S-wave data from the wells have been tied to the seismic and correlated them. Synthetic share wave log has been generated for the well M2 where S-wave is not available and it shows a good correlation with the seismic. Neutron porosity, density, AI and P-wave velocity are proved to be the most significant variables in this statistical method for S-wave generation. Multilinear regression method thus can be considered as a reliable technique for generation of shear wave velocity log in this study.

Keywords: Castagna's equation, multi linear regression, multi attribute analysis, shear wave logs

Procedia PDF Downloads 232
11562 An Earth Mover’s Distance Algorithm Based DDoS Detection Mechanism in SDN

Authors: Yang Zhou, Kangfeng Zheng, Wei Ni, Ren Ping Liu

Abstract:

Software-defined networking (SDN) provides a solution for scalable network framework with decoupled control and data plane. However, this architecture also induces a particular distributed denial-of-service (DDoS) attack that can affect or even overwhelm the SDN network. DDoS attack detection problem has to date been mostly researched as entropy comparison problem. However, this problem lacks the utilization of SDN, and the results are not accurate. In this paper, we propose a DDoS attack detection method, which interprets DDoS detection as a signature matching problem and is formulated as Earth Mover’s Distance (EMD) model. Considering the feasibility and accuracy, we further propose to define the cost function of EMD to be a generalized Kullback-Leibler divergence. Simulation results show that our proposed method can detect DDoS attacks by comparing EMD values with the ones computed in the case without attacks. Moreover, our method can significantly increase the true positive rate of detection.

Keywords: DDoS detection, EMD, relative entropy, SDN

Procedia PDF Downloads 341
11561 Statistical Comparison of Ensemble Based Storm Surge Forecasting Models

Authors: Amin Salighehdar, Ziwen Ye, Mingzhe Liu, Ionut Florescu, Alan F. Blumberg

Abstract:

Storm surge is an abnormal water level caused by a storm. Accurate prediction of a storm surge is a challenging problem. Researchers developed various ensemble modeling techniques to combine several individual forecasts to produce an overall presumably better forecast. There exist some simple ensemble modeling techniques in literature. For instance, Model Output Statistics (MOS), and running mean-bias removal are widely used techniques in storm surge prediction domain. However, these methods have some drawbacks. For instance, MOS is based on multiple linear regression and it needs a long period of training data. To overcome the shortcomings of these simple methods, researchers propose some advanced methods. For instance, ENSURF (Ensemble SURge Forecast) is a multi-model application for sea level forecast. This application creates a better forecast of sea level using a combination of several instances of the Bayesian Model Averaging (BMA). An ensemble dressing method is based on identifying best member forecast and using it for prediction. Our contribution in this paper can be summarized as follows. First, we investigate whether the ensemble models perform better than any single forecast. Therefore, we need to identify the single best forecast. We present a methodology based on a simple Bayesian selection method to select the best single forecast. Second, we present several new and simple ways to construct ensemble models. We use correlation and standard deviation as weights in combining different forecast models. Third, we use these ensembles and compare with several existing models in literature to forecast storm surge level. We then investigate whether developing a complex ensemble model is indeed needed. To achieve this goal, we use a simple average (one of the simplest and widely used ensemble model) as benchmark. Predicting the peak level of Surge during a storm as well as the precise time at which this peak level takes place is crucial, thus we develop a statistical platform to compare the performance of various ensemble methods. This statistical analysis is based on root mean square error of the ensemble forecast during the testing period and on the magnitude and timing of the forecasted peak surge compared to the actual time and peak. In this work, we analyze four hurricanes: hurricanes Irene and Lee in 2011, hurricane Sandy in 2012, and hurricane Joaquin in 2015. Since hurricane Irene developed at the end of August 2011 and hurricane Lee started just after Irene at the beginning of September 2011, in this study we consider them as a single contiguous hurricane event. The data set used for this study is generated by the New York Harbor Observing and Prediction System (NYHOPS). We find that even the simplest possible way of creating an ensemble produces results superior to any single forecast. We also show that the ensemble models we propose generally have better performance compared to the simple average ensemble technique.

Keywords: Bayesian learning, ensemble model, statistical analysis, storm surge prediction

Procedia PDF Downloads 310
11560 The Jurisprudential Evolution of Corruption Offenses in Spain: Before and after the Economic Crisis

Authors: Marta Fernandez Cabrera

Abstract:

The period of economic boom generated by the housing bubble created a climate of social indifference to the problem of corruption. This resulted in the persecution and conviction for these criminal offenses being low. After the economic recession, social awareness about the problem of corruption has increased. This has led to the Spanish citizenship requiring the public authorities to try to end the problem in the most effective way possible. In order to respond to the continuous social demands that require an exemplary punishment, the legislator has made changes in crimes against the public administration in the Spanish Criminal Code. However, from the point of view of criminal law, the social change has not served to modify only the law, but also the jurisprudence. After the recession, judges are punishing more severely these conducts than in the past. Before the crisis, it was usual for criminal judges to divert relevant behavior to other areas of the legal system such as administrative law and acquit in the criminal field. Criminal judges have considered that administrative law already has mechanisms that can effectively deal with this type of behavior in order to respect the principle of subsidiarity or ultima ratio. It has also been usual for criminal judges to acquit civil servants due to the absence of requirements unrelated to the applicable offense. For example, they have required an economic damage to the public administration when the offense in the criminal code does not require it. Nevertheless, for some years, these arguments have either partially disappeared or considerably transformed. Since 2010, a jurisprudential stream has been consolidated that aims to provide a more severe response to corruption than it had received until now. This change of opinion, together with greater prosecution of these behaviors by judges and prosecutors, has led to a significant increase in the number of individuals convicted of corruption crimes. This paper has two objectives. The first one is to show that even though judges apply the law impartially, they are flexible to social changes. The second one is to identify the erroneous arguments the courts have used up until now. To carry out the present paper, it has been done a detailed analysis of the judgments of the supreme court before and after the year 2010. Therefore, the jurisprudential analysis is complemented with the statistical data on corruption available.

Keywords: corruption, public administration, social perception, ultima ratio principle

Procedia PDF Downloads 147
11559 Pricing European Continuous-Installment Options under Regime-Switching Models

Authors: Saghar Heidari

Abstract:

In this paper, we study the valuation problem of European continuous-installment options under Markov-modulated models with a partial differential equation approach. Due to the opportunity for continuing or stopping to pay installments, the valuation problem under regime-switching models can be formulated as coupled partial differential equations (CPDE) with free boundary features. To value the installment options, we express the truncated CPDE as a linear complementarity problem (LCP), then a finite element method is proposed to solve the resulted variational inequality. Under some appropriate assumptions, we establish the stability of the method and illustrate some numerical results to examine the rate of convergence and accuracy of the proposed method for the pricing problem under the regime-switching model.

Keywords: continuous-installment option, European option, regime-switching model, finite element method

Procedia PDF Downloads 141
11558 Performance Evaluation of a Minimum Mean Square Error-Based Physical Sidelink Share Channel Receiver under Fading Channel

Authors: Yang Fu, Jaime Rodrigo Navarro, Jose F. Monserrat, Faiza Bouchmal, Oscar Carrasco Quilis

Abstract:

Cellular Vehicle to Everything (C-V2X) is considered a promising solution for future autonomous driving. From Release 16 to Release 17, the Third Generation Partnership Project (3GPP) has introduced the definitions and services for 5G New Radio (NR) V2X. Experience from previous generations has shown that establishing a simulator for C-V2X communications is an essential preliminary step to achieve reliable and stable communication links. This paper proposes a complete framework of a link-level simulator based on the 3GPP specifications for the Physical Sidelink Share Channel (PSSCH) of the 5G NR Physical Layer (PHY). In this framework, several algorithms in the receiver part, i.e., sliding window in channel estimation and Minimum Mean Square Error (MMSE)-based equalization, are developed. Finally, the performance of the developed PSSCH receiver is validated through extensive simulations under different assumptions.

Keywords: C-V2X, channel estimation, link-level simulator, sidelink, 3GPP

Procedia PDF Downloads 136
11557 Multiparametric Optimization of Water Treatment Process for Thermal Power Plants

Authors: Balgaisha Mukanova, Natalya Glazyrina, Sergey Glazyrin

Abstract:

The formulated problem of optimization of the technological process of water treatment for thermal power plants is considered in this article. The problem is of multiparametric nature. To optimize the process, namely, reduce the amount of waste water, a new technology was developed to reuse such water. A mathematical model of the technology of wastewater reuse was developed. Optimization parameters were determined. The model consists of a material balance equation, an equation describing the kinetics of ion exchange for the non-equilibrium case and an equation for the ion exchange isotherm. The material balance equation includes a nonlinear term that depends on the kinetics of ion exchange. A direct problem of calculating the impurity concentration at the outlet of the water treatment plant was numerically solved. The direct problem was approximated by an implicit point-to-point computation difference scheme. The inverse problem was formulated as relates to determination of the parameters of the mathematical model of the water treatment plant operating in non-equilibrium conditions. The formulated inverse problem was solved. Following the results of calculation the time of start of the filter regeneration process was determined, as well as the period of regeneration process and the amount of regeneration and wash water. Multi-parameter optimization of water treatment process for thermal power plants allowed decreasing the amount of wastewater by 15%.

Keywords: direct problem, multiparametric optimization, optimization parameters, water treatment

Procedia PDF Downloads 389
11556 Efficient Filtering of Graph Based Data Using Graph Partitioning

Authors: Nileshkumar Vaishnav, Aditya Tatu

Abstract:

An algebraic framework for processing graph signals axiomatically designates the graph adjacency matrix as the shift operator. In this setup, we often encounter a problem wherein we know the filtered output and the filter coefficients, and need to find out the input graph signal. Solution to this problem using direct approach requires O(N3) operations, where N is the number of vertices in graph. In this paper, we adapt the spectral graph partitioning method for partitioning of graphs and use it to reduce the computational cost of the filtering problem. We use the example of denoising of the temperature data to illustrate the efficacy of the approach.

Keywords: graph signal processing, graph partitioning, inverse filtering on graphs, algebraic signal processing

Procedia PDF Downloads 315
11555 A Heuristic for the Integrated Production and Distribution Scheduling Problem

Authors: Christian Meinecke, Bernd Scholz-Reiter

Abstract:

The integrated problem of production and distribution scheduling is relevant in many industrial applications. Thus, many heuristics to solve this integrated problem have been developed in the last decade. Most of these heuristics use a sequential working principal or a single decomposition and integration approach to separate and solve sub-problems. A heuristic using a multi-step decomposition and integration approach is presented in this paper and evaluated in a case study. The result show significant improved results compared with sequential scheduling heuristics.

Keywords: production and outbound distribution, integrated planning, heuristic, decomposition, integration

Procedia PDF Downloads 432
11554 R Software for Parameter Estimation of Spatio-Temporal Model

Authors: Budi Nurani Ruchjana, Atje Setiawan Abdullah, I. Gede Nyoman Mindra Jaya, Eddy Hermawan

Abstract:

In this paper, we propose the application package to estimate parameters of spatiotemporal model based on the multivariate time series analysis using the R open-source software. We build packages mainly to estimate the parameters of the Generalized Space Time Autoregressive (GSTAR) model. GSTAR is a combination of time series and spatial models that have parameters vary per location. We use the method of Ordinary Least Squares (OLS) and use the Mean Average Percentage Error (MAPE) to fit the model to spatiotemporal real phenomenon. For case study, we use oil production data from volcanic layer at Jatibarang Indonesia or climate data such as rainfall in Indonesia. Software R is very user-friendly and it is making calculation easier, processing the data is accurate and faster. Limitations R script for the estimation of model parameters spatiotemporal GSTAR built is still limited to a stationary time series model. Therefore, the R program under windows can be developed either for theoretical studies and application.

Keywords: GSTAR Model, MAPE, OLS method, oil production, R software

Procedia PDF Downloads 246
11553 Knowledge Representation Based on Interval Type-2 CFCM Clustering

Authors: Lee Myung-Won, Kwak Keun-Chang

Abstract:

This paper is concerned with knowledge representation and extraction of fuzzy if-then rules using Interval Type-2 Context-based Fuzzy C-Means clustering (IT2-CFCM) with the aid of fuzzy granulation. This proposed clustering algorithm is based on information granulation in the form of IT2 based Fuzzy C-Means (IT2-FCM) clustering and estimates the cluster centers by preserving the homogeneity between the clustered patterns from the IT2 contexts produced in the output space. Furthermore, we can obtain the automatic knowledge representation in the design of Radial Basis Function Networks (RBFN), Linguistic Model (LM), and Adaptive Neuro-Fuzzy Networks (ANFN) from the numerical input-output data pairs. We shall focus on a design of ANFN in this paper. The experimental results on an estimation problem of energy performance reveal that the proposed method showed a good knowledge representation and performance in comparison with the previous works.

Keywords: IT2-FCM, IT2-CFCM, context-based fuzzy clustering, adaptive neuro-fuzzy network, knowledge representation

Procedia PDF Downloads 326
11552 Hybrid Deep Learning and FAST-BRISK 3D Object Detection Technique for Bin-Picking Application

Authors: Thanakrit Taweesoontorn, Sarucha Yanyong, Poom Konghuayrob

Abstract:

Robotic arms have gained popularity in various industries due to their accuracy and efficiency. This research proposes a method for bin-picking tasks using the Cobot, combining the YOLOv5 CNNs model for object detection and pose estimation with traditional feature detection (FAST), feature description (BRISK), and matching algorithms. By integrating these algorithms and utilizing a small-scale depth sensor camera for capturing depth and color images, the system achieves real-time object detection and accurate pose estimation, enabling the robotic arm to pick objects correctly in both position and orientation. Furthermore, the proposed method is implemented within the ROS framework to provide a seamless platform for robotic control and integration. This integration of robotics, cameras, and AI technology contributes to the development of industrial robotics, opening up new possibilities for automating challenging tasks and improving overall operational efficiency.

Keywords: robotic vision, image processing, applications of robotics, artificial intelligent

Procedia PDF Downloads 99
11551 Optimization of Roster Construction In Sports

Authors: Elijah Cavan

Abstract:

In Major League Sports (MLB, NBA, NHL, NFL), it is the Front Office Staff (FOS) who make decisions about who plays for their respective team. The FOS bear the brunt of the responsibility for acquiring players through drafting, trading and signing players in free agency while typically contesting with maximum roster salary constraints. The players themselves are volatile assets of these teams- their value fluctuates with age and performance. A simple comparison can be made when viewing players as assets. The problem here is similar to that of optimizing your investment portfolio. The The goal is ultimately to maximize your periodic returns while tolerating a fixed risk (degree of uncertainty/ potential loss). Each franchise may value assets differently, and some may only tolerate lower risk levels- these are examples of factors that introduce additional constraints into the model. In this talk, we will detail the mathematical formulation of this problem as a constrained optimization problem- which can be solved with classical machine learning methods but is also well posed as a problem to be solved on quantum computers

Keywords: optimization, financial mathematics, sports analytics, simulated annealing

Procedia PDF Downloads 124
11550 Method to Find a ε-Optimal Control of Stochastic Differential Equation Driven by a Brownian Motion

Authors: Francys Souza, Alberto Ohashi, Dorival Leao

Abstract:

We present a general solution for finding the ε-optimal controls for non-Markovian stochastic systems as stochastic differential equations driven by Brownian motion, which is a problem recognized as a difficult solution. The contribution appears in the development of mathematical tools to deal with modeling and control of non-Markovian systems, whose applicability in different areas is well known. The methodology used consists to discretize the problem through a random discretization. In this way, we transform an infinite dimensional problem in a finite dimensional, thereafter we use measurable selection arguments, to find a control on an explicit form for the discretized problem. Then, we prove the control found for the discretized problem is a ε-optimal control for the original problem. Our theory provides a concrete description of a rather general class, among the principals, we can highlight financial problems such as portfolio control, hedging, super-hedging, pairs-trading and others. Therefore, our main contribution is the development of a tool to explicitly the ε-optimal control for non-Markovian stochastic systems. The pathwise analysis was made through a random discretization jointly with measurable selection arguments, has provided us with a structure to transform an infinite dimensional problem into a finite dimensional. The theory is applied to stochastic control problems based on path-dependent stochastic differential equations, where both drift and diffusion components are controlled. We are able to explicitly show optimal control with our method.

Keywords: dynamic programming equation, optimal control, stochastic control, stochastic differential equation

Procedia PDF Downloads 191
11549 Estimation of the Temperatures in an Asynchronous Machine Using Extended Kalman Filter

Authors: Yi Huang, Clemens Guehmann

Abstract:

In order to monitor the thermal behavior of an asynchronous machine with squirrel cage rotor, a 9th-order extended Kalman filter (EKF) algorithm is implemented to estimate the temperatures of the stator windings, the rotor cage and the stator core. The state-space equations of EKF are established based on the electrical, mechanical and the simplified thermal models of an asynchronous machine. The asynchronous machine with simplified thermal model in Dymola is compiled as DymolaBlock, a physical model in MATLAB/Simulink. The coolant air temperature, three-phase voltages and currents are exported from the physical model and are processed by EKF estimator as inputs. Compared to the temperatures exported from the physical model of the machine, three parts of temperatures can be estimated quite accurately by the EKF estimator. The online EKF estimator is independent from the machine control algorithm and can work under any speed and load condition if the stator current is nonzero current system.

Keywords: asynchronous machine, extended Kalman filter, resistance, simulation, temperature estimation, thermal model

Procedia PDF Downloads 286
11548 The Menu Planning Problem: A Systematic Literature Review

Authors: Dorra Kallel, Ines Kanoun, Diala Dhouib

Abstract:

This paper elaborates a Systematic Literature Review SLR) to select the most outstanding studies that address the Menu Planning Problem (MPP) and to classify them according to the to the three following criteria: the used methods, types of patients and the required constraints. At first, a set of 4165 studies was selected. After applying the SLR’s guidelines, this collection was filtered to 13 studies using specific inclusion and exclusion criteria as well as an accurate analysis of each study. Second, the selected papers were invested to answer the proposed research questions. Finally, data synthesis and new perspectives for future works are incorporated in the closing section.

Keywords: Menu Planning Problem (MPP), Systematic Literature Review (SLR), classification, exact and approaches methods

Procedia PDF Downloads 284
11547 Estimation of Uncertainty of Thermal Conductivity Measurement with Single Laboratory Validation Approach

Authors: Saowaluck Ukrisdawithid

Abstract:

The thermal conductivity of thermal insulation materials are measured by Heat Flow Meter (HFM) apparatus. The components of uncertainty are complex and difficult on routine measurement by modelling approach. In this study, uncertainty of thermal conductivity measurement was estimated by single laboratory validation approach. The within-laboratory reproducibility was 1.1%. The standard uncertainty of method and laboratory bias by using SRM1453 expanded polystyrene board was dominant at 1.4%. However, it was assessed that there was no significant bias. For sample measurement, the sources of uncertainty were repeatability, density of sample and thermal conductivity resolution of HFM. From this approach to sample measurements, the combined uncertainty was calculated. In summary, the thermal conductivity of sample, polystyrene foam, was reported as 0.03367 W/m·K ± 3.5% (k = 2) at mean temperature 23.5 °C. The single laboratory validation approach is simple key of routine testing laboratory for estimation uncertainty of thermal conductivity measurement by using HFM, according to ISO/IEC 17025-2017 requirements. These are meaningful for laboratory competent improvement, quality control on products, and conformity assessment.

Keywords: single laboratory validation approach, within-laboratory reproducibility, method and laboratory bias, certified reference material

Procedia PDF Downloads 156
11546 A Variable Neighborhood Search with Tabu Conditions for the Roaming Salesman Problem

Authors: Masoud Shahmanzari

Abstract:

The aim of this paper is to present a Variable Neighborhood Search (VNS) with Tabu Search (TS) conditions for the Roaming Salesman Problem (RSP). The RSP is a special case of the well-known traveling salesman problem (TSP) where a set of cities with time-dependent rewards and a set of campaign days are given. Each city can be visited on any day and a subset of cities can be visited multiple times. The goal is to determine an optimal campaign schedule consist of daily open/closed tours that visit some cities and maximizes the total net benefit while respecting daily maximum tour duration constraints and the necessity to return campaign base frequently. This problem arises in several real-life applications and particularly in election logistics where depots are not fixed. We formulate the problem as a mixed integer linear programming (MILP), in which we capture as many real-world aspects of the RSP as possible. We also present a hybrid metaheuristic algorithm based on a VNS with TS conditions. The initial feasible solution is constructed via a new matheuristc approach based on the decomposition of the original problem. Next, this solution is improved in terms of the collected rewards using the proposed local search procedure. We consider a set of 81 cities in Turkey and a campaign of 30 days as our largest instance. Computational results on real-world instances show that the developed algorithm could find near-optimal solutions effectively.

Keywords: optimization, routing, election logistics, heuristics

Procedia PDF Downloads 95
11545 Definition and Core Components of the Role-Partner Allocation Problem in Collaborative Networks

Authors: J. Andrade-Garda, A. Anguera, J. Ares-Casal, M. Hidalgo-Lorenzo, J.-A. Lara, D. Lizcano, S. Suárez-Garaboa

Abstract:

In the current constantly changing economic context, collaborative networks allow partners to undertake projects that would not be possible if attempted by them individually. These projects usually involve the performance of a group of tasks (named roles) that have to be distributed among the partners. Thus, an allocation/matching problem arises that will be referred to as Role-Partner Allocation problem. In real life this situation is addressed by negotiation between partners in order to reach ad hoc agreements. Besides taking a long time and being hard work, both historical evidence and economic analysis show that such approach is not recommended. Instead, the allocation process should be automated by means of a centralized matching scheme. However, as a preliminary step to start the search for such a matching mechanism (or even the development of a new one), the problem and its core components must be specified. To this end, this paper establishes (i) the definition of the problem and its constraints, (ii) the key features of the involved elements (i.e., roles and partners); and (iii) how to create preference lists both for roles and partners. Only this way it will be possible to conduct subsequent methodological research on the solution method.     

Keywords: collaborative network, matching, partner, preference list, role

Procedia PDF Downloads 240
11544 Block Based Imperial Competitive Algorithm with Greedy Search for Traveling Salesman Problem

Authors: Meng-Hui Chen, Chiao-Wei Yu, Pei-Chann Chang

Abstract:

Imperial competitive algorithm (ICA) simulates a multi-agent algorithm. Each agent is like a kingdom has its country, and the strongest country in each agent is called imperialist, others are colony. Countries are competitive with imperialist which in the same kingdom by evolving. So this country will move in the search space to find better solutions with higher fitness to be a new imperialist. The main idea in this paper is using the peculiarity of ICA to explore the search space to solve the kinds of combinational problems. Otherwise, we also study to use the greed search to increase the local search ability. To verify the proposed algorithm in this paper, the experimental results of traveling salesman problem (TSP) is according to the traveling salesman problem library (TSPLIB). The results show that the proposed algorithm has higher performance than the other known methods.

Keywords: traveling salesman problem, artificial chromosomes, greedy search, imperial competitive algorithm

Procedia PDF Downloads 462
11543 Mixture statistical modeling for predecting mortality human immunodeficiency virus (HIV) and tuberculosis(TB) infection patients

Authors: Mohd Asrul Affendi Bi Abdullah, Nyi Nyi Naing

Abstract:

The purpose of this study was to identify comparable manner between negative binomial death rate (NBDR) and zero inflated negative binomial death rate (ZINBDR) with died patients with (HIV + T B+) and (HIV + T B−). HIV and TB is a serious world wide problem in the developing country. Data were analyzed with applying NBDR and ZINBDR to make comparison which a favorable model is better to used. The ZINBDR model is able to account for the disproportionately large number of zero within the data and is shown to be a consistently better fit than the NBDR model. Hence, as a results ZINBDR model is a superior fit to the data than the NBDR model and provides additional information regarding the died mechanisms HIV+TB. The ZINBDR model is shown to be a use tool for analysis death rate according age categorical.

Keywords: zero inflated negative binomial death rate, HIV and TB, AIC and BIC, death rate

Procedia PDF Downloads 435
11542 Estimation of Functional Response Model by Supervised Functional Principal Component Analysis

Authors: Hyon I. Paek, Sang Rim Kim, Hyon A. Ryu

Abstract:

In functional linear regression, one typical problem is to reduce dimension. Compared with multivariate linear regression, functional linear regression is regarded as an infinite-dimensional case, and the main task is to reduce dimensions of functional response and functional predictors. One common approach is to adapt functional principal component analysis (FPCA) on functional predictors and then use a few leading functional principal components (FPC) to predict the functional model. The leading FPCs estimated by the typical FPCA explain a major variation of the functional predictor, but these leading FPCs may not be mostly correlated with the functional response, so they may not be significant in the prediction for response. In this paper, we propose a supervised functional principal component analysis method for a functional response model with FPCs obtained by considering the correlation of the functional response. Our method would have a better prediction accuracy than the typical FPCA method.

Keywords: supervised, functional principal component analysis, functional response, functional linear regression

Procedia PDF Downloads 81
11541 The Impact of Bim Technology on the Whole Process Cost Management of Civil Engineering Projects in Kenya

Authors: Nsimbe Allan

Abstract:

The study examines the impact of Building Information Modeling (BIM) on the cost management of engineering projects, focusing specifically on the Mombasa Port Area Development Project. The objective of this research venture is to determine the mechanisms through which Building Information Modeling (BIM) facilitates stakeholder collaboration, reduces construction-related expenses, and enhances the precision of cost estimation. Furthermore, the study investigates barriers to execution, assesses the impact on the project's transparency, and suggests approaches to maximize resource utilization. The study, selected for its practical significance and intricate nature, conducted a Systematic Literature Review (SLR) using credible databases, including ScienceDirect and IEEE Xplore. To constitute the diverse sample, 69 individuals, including project managers, cost estimators, and BIM administrators, were selected via stratified random sampling. The data were obtained using a mixed-methods approach, which prioritized ethical considerations. SPSS and Microsoft Excel were applied to the analysis. The research emphasizes the crucial role that project managers, architects, and engineers play in the decision-making process (47% of respondents). Furthermore, a significant improvement in cost estimation accuracy was reported by 70% of the participants. It was found that the implementation of BIM resulted in enhanced project visibility, which in turn optimized resource allocation and facilitated the process of budgeting. In brief, the study highlights the positive impacts of Building Information Modeling (BIM) on collaborative decision-making and cost estimation, addresses challenges related to implementation, and provides solutions for the efficient assimilation and understanding of BIM principles.

Keywords: cost management, resource utilization, stakeholder collaboration, project transparency

Procedia PDF Downloads 72
11540 Statistical Shape Analysis of the Human Upper Airway

Authors: Ramkumar Gunasekaran, John Cater, Vinod Suresh, Haribalan Kumar

Abstract:

The main objective of this project is to develop a statistical shape model using principal component analysis that could be used for analyzing the shape of the human airway. The ultimate goal of this project is to identify geometric risk factors for diagnosis and management of Obstructive Sleep Apnoea (OSA). Anonymous CBCT scans of 25 individuals were obtained from the Otago Radiology Group. The airways were segmented between the hard-palate and the aryepiglottic fold using snake active contour segmentation. The point data cloud of the segmented images was then fitted with a bi-cubic mesh, and pseudo landmarks were placed to perform PCA on the segmented airway to analyze the shape of the airway and to find the relationship between the shape and OSA risk factors. From the PCA results, the first four modes of variation were found to be significant. Mode 1 was interpreted to be the overall length of the airway, Mode 2 was related to the anterior-posterior width of the retroglossal region, Mode 3 was related to the lateral dimension of the oropharyngeal region and Mode 4 was related to the anterior-posterior width of the oropharyngeal region. All these regions are subjected to the risk factors of OSA.

Keywords: medical imaging, image processing, FEM/BEM, statistical modelling

Procedia PDF Downloads 516
11539 Promoting Authenticity in Employer Brands to Address the Global-Local Problem in Complex Organisations: The Case of a Developing Country

Authors: Saud Al Taj

Abstract:

Employer branding is considered as a useful tool for addressing the global-local problem facing complex organisations that have operations scattered across the globe and face challenges of dealing with the local environment alongside. Despite being an established field of study within the Western developed world, there is little empirical evidence concerning the relevance of employer branding to global companies that operate in the under-developed economies. This paper fills this gap by gaining rich insight into the implementation of employer branding programs in a foreign multinational operating in Pakistan dealing with the global-local problem. The study is qualitative in nature and employs semi-structured and focus group interviews with senior/middle managers and local frontline employees to deeply examine the phenomenon in case organisation. Findings suggest that authenticity is required in employer brands to enable them to respond to the local needs thereby leading to the resolution of the global-local problem. However, the role of signaling theory is key to the development of authentic employer brands as it stresses on the need to establish an efficient and effective signaling environment wherein signals travel in both directions (from signal designers to receivers and backwards) and facilitate firms with the global-local problem. The paper also identifies future avenues of research for the employer branding field.

Keywords: authenticity, counter-signals, employer branding, global-local problem, signaling theory

Procedia PDF Downloads 370
11538 Applying of an Adaptive Neuro-Fuzzy Inference System (ANFIS) for Estimation of Flood Hydrographs

Authors: Amir Ahmad Dehghani, Morteza Nabizadeh

Abstract:

This paper presents the application of an Adaptive Neuro-Fuzzy Inference System (ANFIS) to flood hydrograph modeling of Shahid Rajaee reservoir dam located in Iran. This was carried out using 11 flood hydrographs recorded in Tajan river gauging station. From this dataset, 9 flood hydrographs were chosen to train the model and 2 flood hydrographs to test the model. The different architectures of neuro-fuzzy model according to the membership function and learning algorithm were designed and trained with different epochs. The results were evaluated in comparison with the observed hydrographs and the best structure of model was chosen according the least RMSE in each performance. To evaluate the efficiency of neuro-fuzzy model, various statistical indices such as Nash-Sutcliff and flood peak discharge error criteria were calculated. In this simulation, the coordinates of a flood hydrograph including peak discharge were estimated using the discharge values occurred in the earlier time steps as input values to the neuro-fuzzy model. These results indicate the satisfactory efficiency of neuro-fuzzy model for flood simulating. This performance of the model demonstrates the suitability of the implemented approach to flood management projects.

Keywords: adaptive neuro-fuzzy inference system, flood hydrograph, hybrid learning algorithm, Shahid Rajaee reservoir dam

Procedia PDF Downloads 480
11537 Near-Infrared Hyperspectral Imaging Spectroscopy to Detect Microplastics and Pieces of Plastic in Almond Flour

Authors: H. Apaza, L. Chévez, H. Loro

Abstract:

Plastic and microplastic pollution in human food chain is a big problem for human health that requires more elaborated techniques that can identify their presences in different kinds of food. Hyperspectral imaging technique is an optical technique than can detect the presence of different elements in an image and can be used to detect plastics and microplastics in a scene. To do this statistical techniques are required that need to be evaluated and compared in order to find the more efficient ones. In this work, two problems related to the presence of plastics are addressed, the first is to detect and identify pieces of plastic immersed in almond seeds, and the second problem is to detect and quantify microplastic in almond flour. To do this we make use of the analysis hyperspectral images taken in the range of 900 to 1700 nm using 4 unmixing techniques of hyperspectral imaging which are: least squares unmixing (LSU), non-negatively constrained least squares unmixing (NCLSU), fully constrained least squares unmixing (FCLSU), and scaled constrained least squares unmixing (SCLSU). NCLSU, FCLSU, SCLSU techniques manage to find the region where the plastic is found and also manage to quantify the amount of microplastic contained in the almond flour. The SCLSU technique estimated a 13.03% abundance of microplastics and 86.97% of almond flour compared to 16.66% of microplastics and 83.33% abundance of almond flour prepared for the experiment. Results show the feasibility of applying near-infrared hyperspectral image analysis for the detection of plastic contaminants in food.

Keywords: food, plastic, microplastic, NIR hyperspectral imaging, unmixing

Procedia PDF Downloads 133
11536 Prevalence of Behavioral and Emotional Problems in School Going Adolescents in India

Authors: Anshu Gupta, Charu Gupta

Abstract:

Background: Adolescence is the transitional period between puberty and adulthood. It is marked by immense turmoil in emotional and behavioral spheres. Adolescents are at risk of an array of behavioral and emotional problems, resulting in social, academic and vocational function impairments. Conflicts in the family and inability of the parents to cope with the changing demands of an adolescent have a negative impact on the overall development of the child. This augers ill for the individual’s future, resulting in depression, delinquency and suicides among other problems. Aim: The aim of the study was to compare the prevalence of behavioral and emotional problems in school going adolescents aged 13 to 15 years residing in Ludhiana city. Method: A total of 1380 school children in the age group of 13 to 15 years were assessed by the adolescent health screening questionnaire (FAPS) and Youth Self-Report (2001) questionnaire. Statistical significance was ascertained by t-test, chi-square test (x²) and ANOVA, as appropriate. Results: A considerably high prevalence of behavioral and emotional problems was found in school going adolescents (26.5%), more in girls (31.7%) than in boys (24.4%). In case of boys, the maximum problem was in the 13 year age group, i.e., 28.2%, followed by a significant decline by the age of 14 years, i.e., 24.2% and 15 years, i.e., 19.6%. In case of girls also, the maximum problem was in the 13 year age group, i.e., 32.4% followed by a marginal decline in the 14 years i.e., 31.8% and 15 year age group, i.e., 30.2%. Demographic factors were non contributory. Internalizing syndrome (22.4%) was the most common problem followed by the neither internalizing nor externalizing (17.6%) group. In internalizing group, most (26.5%) of the students were observed to be anxious/ depressed. Social problem was observed to be the most frequent (10.6%) among neither internalizing nor externalizing group. Aggressive behavior was the commonest (8.4%) among externalizing group. Internalizing problems, mainly anxiety and depression, were commoner in females (30.6%) than males (24.6%). More boys (16%) than girls (13.4%) were reported to suffer from externalizing disorders. A critical review of the data showed that most of the adolescents had poor knowledge about reproductive health. Almost 36% reported that the source of their information on sexual and reproductive health being friends and the electronic media. There was a high percentage of adolescents who reported being worried about sexual abuse (20.2%) with majority of them being girls (93.6%) reflecting poorly on the social setup in the country. About 41% of adolescents reported being concerned about body weight and most of them being girls (92.4%). Up to 14.5% reported having thoughts of using alcohol or drugs perhaps due to the easy availability of substances of abuse in this part of the country. 12.8% (mostly girls) reported suicidal thoughts. Summary/conclusion: There is a high prevalence of emotional and behavioral problems among school-going adolescents. Resolution of these problems during adolescence is essential for attaining a healthy adulthood. The need of the hour is to spread awareness among caregivers and formulation of effective management strategies including school mental health programme.

Keywords: adolescence, behavioral, emotional, internalizing problem

Procedia PDF Downloads 290
11535 The Relationship between Confidence, Accuracy, and Decision Making in a Mobile Review Program

Authors: Carla Van De Sande, Jana Vandenberg

Abstract:

Just like physical skills, cognitive skills grow rusty over time unless they are regularly used and practiced, so academic breaks can have negative consequences on student learning and success. The Keeping in School Shape (KiSS) program is an engaging, accessible, and cost-effective intervention that harnesses the benefits of retrieval practice by using technology to help students maintain proficiency over breaks from school by delivering a daily review problem via text message or email. A growth mindset is promoted through feedback messages encouraging students to try again if they get a problem wrong and to take on a challenging problem if they get a problem correct. This paper reports on the relationship between confidence, accuracy, and decision-making during the implementation of the KiSS Program at a large university during winter break for students enrolled in an engineering introductory Calculus course sequence.

Keywords: growth mindset, learning loss, on-the-go learning, retrieval practice

Procedia PDF Downloads 209