Search results for: weighted overlay method
19295 Trends in Use of Millings in Pavement Maintenance
Authors: Rafiqul Tarefder, Mohiuddin Ahmad, Mohammad Hossain
Abstract:
While milling materials from old pavement surface can be an important component of cost effective maintenance operation, their use in maintenance projects are not uniform and well documented. This study documents the different maintenance practices followed by four transportation districts of New Mexico Department of Transportation (NMDOT) in an attempt to find whether millings are being used in maintenance projects by those districts. Based on existing literature, a questionnaire was developed related to six common maintenance practices. NMDOT district personal were interviewed face to face to discuss and get answers to that questionnaire. It revealed that NMDOT districts mainly use chip seal and patching. Other maintenance procedures such as sand seal, scrub seal, slurry seal, and thin overlay have limited use. Two out of four participating districts do not have any documents on chip sealing; rather they employ the experiences of the chip seal crew. All districts use polymer modified high float emulsion (HFE100P) for chip seal with an application rate ranging from 0.4 to 0.56 gallons per square yard. Chip application rate varies from 15 to 40 lb/ square yard. State wide, the thickness of chip seal varies from 3/8" to 1" and life varies from 3 to 10 years. NMDOT districts mainly use three type of patching: pothole, dig-out and blade patch. Pothole patches are used for small potholes and during emergency, dig-out patches are used for all type of potholes sometimes after pothole patching, and blade patch is used when a significant portion of the pavement is damaged. Pothole patches last as low as three days whereas, blade patch lasts as long as 3 years. It was observed that all participating districts use millings in maintenance projects.Keywords: chip seal, sand seal, scrub seal, slurry seal, overlay, patching, millings
Procedia PDF Downloads 34319294 A Method of Effective Planning and Control of Industrial Facility Energy Consumption
Authors: Aleksandra Aleksandrovna Filimonova, Lev Sergeevich Kazarinov, Tatyana Aleksandrovna Barbasova
Abstract:
A method of effective planning and control of industrial facility energy consumption is offered. The method allows to optimally arrange the management and full control of complex production facilities in accordance with the criteria of minimal technical and economic losses at the forecasting control. The method is based on the optimal construction of the power efficiency characteristics with the prescribed accuracy. The problem of optimal designing of the forecasting model is solved on the basis of three criteria: maximizing the weighted sum of the points of forecasting with the prescribed accuracy; the solving of the problem by the standard principles at the incomplete statistic data on the basis of minimization of the regularized function; minimizing the technical and economic losses due to the forecasting errors.Keywords: energy consumption, energy efficiency, energy management system, forecasting model, power efficiency characteristics
Procedia PDF Downloads 39319293 Statistical Convergence of the Szasz-Mirakjan-Kantorovich-Type Operators
Authors: Rishikesh Yadav, Ramakanta Meher, Vishnu Narayan Mishra
Abstract:
The main aim of this article is to investigate the statistical convergence of the summation of integral type operators and to obtain the weighted statistical convergence. The rate of statistical convergence by means of modulus of continuity and function belonging to the Lipschitz class are also studied. We discuss the convergence of the defined operators by graphical representation and put a better rate of convergence than the Szasz-Mirakjan-Kantorovich operators. In the last section, we extend said operators into bivariate operators to study about the rate of convergence in sense of modulus of continuity and by means of Lipschitz class by using function of two variables.Keywords: The Szasz-Mirakjan-Kantorovich operators, statistical convergence, modulus of continuity, Peeters K-functional, weighted modulus of continuity
Procedia PDF Downloads 21219292 A Lagrangian Hamiltonian Computational Method for Hyper-Elastic Structural Dynamics
Authors: Hosein Falahaty, Hitoshi Gotoh, Abbas Khayyer
Abstract:
Performance of a Hamiltonian based particle method in simulation of nonlinear structural dynamics is subjected to investigation in terms of stability and accuracy. The governing equation of motion is derived based on Hamilton's principle of least action, while the deformation gradient is obtained according to Weighted Least Square method. The hyper-elasticity models of Saint Venant-Kirchhoff and a compressible version similar to Mooney- Rivlin are engaged for the calculation of second Piola-Kirchhoff stress tensor, respectively. Stability along with accuracy of numerical model is verified by reproducing critical stress fields in static and dynamic responses. As the results, although performance of Hamiltonian based model is evaluated as being acceptable in dealing with intense extensional stress fields, however kinds of instabilities reveal in the case of violent collision which can be most likely attributed to zero energy singular modes.Keywords: Hamilton's principle of least action, particle-based method, hyper-elasticity, analysis of stability
Procedia PDF Downloads 34219291 Legal Thresholds in Germany Take Over: The Role of Minimum Offer Price Regulation
Authors: Osei Kwabena Brefo
Abstract:
The Germany takeover regulations mandate the minimum offer price to be set in accordance with section 31 (1) of the WpÜG. To safeguard this right of the shareholders, the WpÜG contains minimum price regulations that are based on the fundamental principle of equal treatment of shareholders. The minimum offer price is determined based on the maximum of either the highest price the bidder has offered other shareholders within the last six months (§ 4 WpÜG) or the three-month weighted average stock price (§ 5 Abs. 1 WpÜG) before the announcement of the offer. This study provides the first empirical evidence on differences between setting the minimum offer price based on the highest price paid to another shareholder within the last six months (“6MH offers”) as against the three-month weighted average stock before the announcement of the offer (“VWAP offers”) in terms of their impact on acquisition outcome. Using a sample of 358 acquisitions announced between 2004 and 2023, the study finds that the market and legal minimum premium in 6MH offers are significantly lower than three-month weighted average stock price offers. The study also employed Propensity Score Matching and endogenous Linear regression model to address for endogeneity concerns. Consistent with the Ordinary Least Square estimation, the study finds that after controlling for self-selection bias, market and legal premiums in 6MH offers are significantly lower than VWAP offers.Keywords: mergers and acquisition, offer premium, minimum price regulation, takeover regulation
Procedia PDF Downloads 1319290 Weighted-Distance Sliding Windows and Cooccurrence Graphs for Supporting Entity-Relationship Discovery in Unstructured Text
Authors: Paolo Fantozzi, Luigi Laura, Umberto Nanni
Abstract:
The problem of Entity relation discovery in structured data, a well covered topic in literature, consists in searching within unstructured sources (typically, text) in order to find connections among entities. These can be a whole dictionary, or a specific collection of named items. In many cases machine learning and/or text mining techniques are used for this goal. These approaches might be unfeasible in computationally challenging problems, such as processing massive data streams. A faster approach consists in collecting the cooccurrences of any two words (entities) in order to create a graph of relations - a cooccurrence graph. Indeed each cooccurrence highlights some grade of semantic correlation between the words because it is more common to have related words close each other than having them in the opposite sides of the text. Some authors have used sliding windows for such problem: they count all the occurrences within a sliding windows running over the whole text. In this paper we generalise such technique, coming up to a Weighted-Distance Sliding Window, where each occurrence of two named items within the window is accounted with a weight depending on the distance between items: a closer distance implies a stronger evidence of a relationship. We develop an experiment in order to support this intuition, by applying this technique to a data set consisting in the text of the Bible, split into verses.Keywords: cooccurrence graph, entity relation graph, unstructured text, weighted distance
Procedia PDF Downloads 15419289 Breast Cancer Survivability Prediction via Classifier Ensemble
Authors: Mohamed Al-Badrashiny, Abdelghani Bellaachia
Abstract:
This paper presents a classifier ensemble approach for predicting the survivability of the breast cancer patients using the latest database version of the Surveillance, Epidemiology, and End Results (SEER) Program of the National Cancer Institute. The system consists of two main components; features selection and classifier ensemble components. The features selection component divides the features in SEER database into four groups. After that it tries to find the most important features among the four groups that maximizes the weighted average F-score of a certain classification algorithm. The ensemble component uses three different classifiers, each of which models different set of features from SEER through the features selection module. On top of them, another classifier is used to give the final decision based on the output decisions and confidence scores from each of the underlying classifiers. Different classification algorithms have been examined; the best setup found is by using the decision tree, Bayesian network, and Na¨ıve Bayes algorithms for the underlying classifiers and Na¨ıve Bayes for the classifier ensemble step. The system outperforms all published systems to date when evaluated against the exact same data of SEER (period of 1973-2002). It gives 87.39% weighted average F-score compared to 85.82% and 81.34% of the other published systems. By increasing the data size to cover the whole database (period of 1973-2014), the overall weighted average F-score jumps to 92.4% on the held out unseen test set.Keywords: classifier ensemble, breast cancer survivability, data mining, SEER
Procedia PDF Downloads 32919288 Using Scale Invariant Feature Transform Features to Recognize Characters in Natural Scene Images
Authors: Belaynesh Chekol, Numan Çelebi
Abstract:
The main purpose of this work is to recognize individual characters extracted from natural scene images using scale invariant feature transform (SIFT) features as an input to K-nearest neighbor (KNN); a classification learner algorithm. For this task, 1,068 and 78 images of English alphabet characters taken from Chars74k data set is used to train and test the classifier respectively. For each character image, We have generated describing features by using SIFT algorithm. This set of features is fed to the learner so that it can recognize and label new images of English characters. Two types of KNN (fine KNN and weighted KNN) were trained and the resulted classification accuracy is 56.9% and 56.5% respectively. The training time taken was the same for both fine and weighted KNN.Keywords: character recognition, KNN, natural scene image, SIFT
Procedia PDF Downloads 28119287 Cognitive Weighted Polymorphism Factor: A New Cognitive Complexity Metric
Authors: T. Francis Thamburaj, A. Aloysius
Abstract:
Polymorphism is one of the main pillars of the object-oriented paradigm. It induces hidden forms of class dependencies which may impact software quality, resulting in higher cost factor for comprehending, debugging, testing, and maintaining the software. In this paper, a new cognitive complexity metric called Cognitive Weighted Polymorphism Factor (CWPF) is proposed. Apart from the software structural complexity, it includes the cognitive complexity on the basis of type. The cognitive weights are calibrated based on 27 empirical studies with 120 persons. A case study and experimentation of the new software metric shows positive results. Further, a comparative study is made and the correlation test has proved that CWPF complexity metric is a better, more comprehensive, and more realistic indicator of the software complexity than Abreu’s Polymorphism Factor (PF) complexity metric.Keywords: cognitive complexity metric, object-oriented metrics, polymorphism factor, software metrics
Procedia PDF Downloads 46119286 A Weighted Sum Particle Swarm Approach (WPSO) Combined with a Novel Feasibility-Based Ranking Strategy for Constrained Multi-Objective Optimization of Compact Heat Exchangers
Authors: Milad Yousefi, Moslem Yousefi, Ricarpo Poley, Amer Nordin Darus
Abstract:
Design optimization of heat exchangers is a very complicated task that has been traditionally carried out based on a trial-and-error procedure. To overcome the difficulties of the conventional design approaches especially when a large number of variables, constraints and objectives are involved, a new method based on a well-stablished evolutionary algorithm, particle swarm optimization (PSO), weighted sum approach and a novel constraint handling strategy is presented in this study. Since, the conventional constraint handling strategies are not effective and easy-to-implement in multi-objective algorithms, a novel feasibility-based ranking strategy is introduced which is both extremely user-friendly and effective. A case study from industry has been investigated to illustrate the performance of the presented approach. The results show that the proposed algorithm can find the near pareto-optimal with higher accuracy when it is compared to conventional non-dominated sorting genetic algorithm II (NSGA-II). Moreover, the difficulties of a trial-and-error process for setting the penalty parameters is solved in this algorithm.Keywords: Heat exchanger, Multi-objective optimization, Particle swarm optimization, NSGA-II Constraints handling.
Procedia PDF Downloads 55619285 Geospatial Analysis for Predicting Sinkhole Susceptibility in Greene County, Missouri
Authors: Shishay Kidanu, Abdullah Alhaj
Abstract:
Sinkholes in the karst terrain of Greene County, Missouri, pose significant geohazards, imposing challenges on construction and infrastructure development, with potential threats to lives and property. To address these issues, understanding the influencing factors and modeling sinkhole susceptibility is crucial for effective mitigation through strategic changes in land use planning and practices. This study utilizes geographic information system (GIS) software to collect and process diverse data, including topographic, geologic, hydrogeologic, and anthropogenic information. Nine key sinkhole influencing factors, ranging from slope characteristics to proximity to geological structures, were carefully analyzed. The Frequency Ratio method establishes relationships between attribute classes of these factors and sinkhole events, deriving class weights to indicate their relative importance. Weighted integration of these factors is accomplished using the Analytic Hierarchy Process (AHP) and the Weighted Linear Combination (WLC) method in a GIS environment, resulting in a comprehensive sinkhole susceptibility index (SSI) model for the study area. Employing Jenk's natural break classifier method, the SSI values are categorized into five distinct sinkhole susceptibility zones: very low, low, moderate, high, and very high. Validation of the model, conducted through the Area Under Curve (AUC) and Sinkhole Density Index (SDI) methods, demonstrates a robust correlation with sinkhole inventory data. The prediction rate curve yields an AUC value of 74%, indicating a 74% validation accuracy. The SDI result further supports the success of the sinkhole susceptibility model. This model offers reliable predictions for the future distribution of sinkholes, providing valuable insights for planners and engineers in the formulation of development plans and land-use strategies. Its application extends to enhancing preparedness and minimizing the impact of sinkhole-related geohazards on both infrastructure and the community.Keywords: sinkhole, GIS, analytical hierarchy process, frequency ratio, susceptibility, Missouri
Procedia PDF Downloads 7419284 Philippine Site Suitability Analysis for Biomass, Hydro, Solar, and Wind Renewable Energy Development Using Geographic Information System Tools
Authors: Jara Kaye S. Villanueva, M. Rosario Concepcion O. Ang
Abstract:
For the past few years, Philippines has depended most of its energy source on oil, coal, and fossil fuel. According to the Department of Energy (DOE), the dominance of coal in the energy mix will continue until the year 2020. The expanding energy needs in the country have led to increasing efforts to promote and develop renewable energy. This research is a part of the government initiative in preparation for renewable energy development and expansion in the country. The Philippine Renewable Energy Resource Mapping from Light Detection and Ranging (LiDAR) Surveys is a three-year government project which aims to assess and quantify the renewable energy potential of the country and to put them into usable maps. This study focuses on the site suitability analysis of the four renewable energy sources – biomass (coconut, corn, rice, and sugarcane), hydro, solar, and wind energy. The site assessment is a key component in determining and assessing the most suitable locations for the construction of renewable energy power plants. This method maximizes the use of both the technical methods in resource assessment, as well as taking into account the environmental, social, and accessibility aspect in identifying potential sites by utilizing and integrating two different methods: the Multi-Criteria Decision Analysis (MCDA) method and Geographic Information System (GIS) tools. For the MCDA, Analytical Hierarchy Processing (AHP) is employed to determine the parameters needed for the suitability analysis. To structure these site suitability parameters, various experts from different fields were consulted – scientists, policy makers, environmentalists, and industrialists. The need to have a well-represented group of people to consult with is relevant to avoid bias in the output parameter of hierarchy levels and weight matrices. AHP pairwise matrix computation is utilized to derive weights per level out of the expert’s gathered feedback. Whereas from the threshold values derived from related literature, international studies, and government laws, the output values were then consulted with energy specialists from the DOE. Geospatial analysis using GIS tools translate this decision support outputs into visual maps. Particularly, this study uses Euclidean distance to compute for the distance values of each parameter, Fuzzy Membership algorithm which normalizes the output from the Euclidean Distance, and the Weighted Overlay tool for the aggregation of the layers. Using the Natural Breaks algorithm, the suitability ratings of each of the map are classified into 5 discrete categories of suitability index: (1) not suitable (2) least suitable, (3) suitable, (4) moderately suitable, and (5) highly suitable. In this method, the classes are grouped based on the best groups similar values wherein each subdivision are set from the rest based on the big difference in boundary values. Results show that in the entire Philippine area of responsibility, biomass has the highest suitability rating with rice as the most suitable at 75.76% suitability percentage, whereas wind has the least suitability percentage with score 10.28%. Solar and Hydro fall in the middle of the two, with suitability values 28.77% and 21.27%.Keywords: site suitability, biomass energy, hydro energy, solar energy, wind energy, GIS
Procedia PDF Downloads 15119283 Seismic Perimeter Surveillance System (Virtual Fence) for Threat Detection and Characterization Using Multiple ML Based Trained Models in Weighted Ensemble Voting
Authors: Vivek Mahadev, Manoj Kumar, Neelu Mathur, Brahm Dutt Pandey
Abstract:
Perimeter guarding and protection of critical installations require prompt intrusion detection and assessment to take effective countermeasures. Currently, visual and electronic surveillance are the primary methods used for perimeter guarding. These methods can be costly and complicated, requiring careful planning according to the location and terrain. Moreover, these methods often struggle to detect stealthy and camouflaged insurgents. The object of the present work is to devise a surveillance technique using seismic sensors that overcomes the limitations of existing systems. The aim is to improve intrusion detection, assessment, and characterization by utilizing seismic sensors. Most of the similar systems have only two types of intrusion detection capability viz., human or vehicle. In our work we could even categorize further to identify types of intrusion activity such as walking, running, group walking, fence jumping, tunnel digging and vehicular movements. A virtual fence of 60 meters at GCNEP, Bahadurgarh, Haryana, India, was created by installing four underground geophones at a distance of 15 meters each. The signals received from these geophones are then processed to find unique seismic signatures called features. Various feature optimization and selection methodologies, such as LightGBM, Boruta, Random Forest, Logistics, Recursive Feature Elimination, Chi-2 and Pearson Ratio were used to identify the best features for training the machine learning models. The trained models were developed using algorithms such as supervised support vector machine (SVM) classifier, kNN, Decision Tree, Logistic Regression, Naïve Bayes, and Artificial Neural Networks. These models were then used to predict the category of events, employing weighted ensemble voting to analyze and combine their results. The models were trained with 1940 training events and results were evaluated with 831 test events. It was observed that using the weighted ensemble voting increased the efficiency of predictions. In this study we successfully developed and deployed the virtual fence using geophones. Since these sensors are passive, do not radiate any energy and are installed underground, it is impossible for intruders to locate and nullify them. Their flexibility, quick and easy installation, low costs, hidden deployment and unattended surveillance make such systems especially suitable for critical installations and remote facilities with difficult terrain. This work demonstrates the potential of utilizing seismic sensors for creating better perimeter guarding and protection systems using multiple machine learning models in weighted ensemble voting. In this study the virtual fence achieved an intruder detection efficiency of over 97%.Keywords: geophone, seismic perimeter surveillance, machine learning, weighted ensemble method
Procedia PDF Downloads 8119282 Detection Efficient Enterprises via Data Envelopment Analysis
Authors: S. Turkan
Abstract:
In this paper, the Turkey’s Top 500 Industrial Enterprises data in 2014 were analyzed by data envelopment analysis. Data envelopment analysis is used to detect efficient decision-making units such as universities, hospitals, schools etc. by using inputs and outputs. The decision-making units in this study are enterprises. To detect efficient enterprises, some financial ratios are determined as inputs and outputs. For this reason, financial indicators related to productivity of enterprises are considered. The efficient foreign weighted owned capital enterprises are detected via super efficiency model. According to the results, it is said that Mercedes-Benz is the most efficient foreign weighted owned capital enterprise in Turkey.Keywords: data envelopment analysis, super efficiency, logistic regression, financial ratios
Procedia PDF Downloads 32619281 Modelling and Maping Malnutrition Toddlers in Bojonegoro Regency with Mixed Geographically Weighted Regression Approach
Authors: Elvira Mustikawati P.H., Iis Dewi Ratih, Dita Amelia
Abstract:
Bojonegoro has proclaimed a policy of zero malnutrition. Therefore, as an effort to solve the cases of malnutrition children in Bojonegoro, this study used the approach geographically Mixed Weighted Regression (MGWR) to determine the factors that influence the percentage of malnourished children under five in which factors can be divided into locally influential factor in each district and global factors that influence throughout the district. Based on the test of goodness of fit models, R2 and AIC values in GWR models are better than MGWR models. R2 and AIC values in MGWR models are 84.37% and 14.28, while the GWR models respectively are 91.04% and -62.04. Based on the analysis with GWR models, District Sekar, Bubulan, Gondang, and Dander is a district with three predictor variables (percentage of vitamin A, the percentage of births assisted health personnel, and the percentage of clean water) that significantly influence the percentage of malnourished children under five. Procedia PDF Downloads 29719280 Groundwater Vulnerability of Halabja-Khurmal Sub-Basin
Authors: Lanja F. Rauf, Salahalddin S. Ali, Nadhir Al-Ansari
Abstract:
Evolving groundwater vulnerability from DRASTIC to modified DRASTIC methods helps choose the most accurate areas that are most delicate toward pollution. This study aims to modify DRASTIC with land use and water quality index for groundwater vulnerability assessment in the Halabja-Khurmal sub-basin, NE/Iraq. The Halabja- Khurmal sub-basin groundwater vulnerability index is calculated from nine hydrogeological parameters by the overlay weighting method. As a result, 1.3 % of the total area has a very high vulnerability value and 46.1 % with high vulnerability. The regions with high groundwater vulnerability have a high water table and groundwater recharge. Nitrate concentration was used to validate the result, and the Pearson correlation and recession analysis between the modified DRASTIC index and nitrate concentration depicted a strong relation with 0.76 and 0.7, respectively.Keywords: groundwater vulnerability, modified DRASTIC, land-use, nitrate pollution, water quality index
Procedia PDF Downloads 9719279 Integrating GIS and Analytical Hierarchy Process-Multicriteria Decision Analysis for Identification of Suitable Areas for Artificial Recharge with Reclaimed Water
Authors: Mahmoudi Marwa, Bahim Nadhem, Aydi Abdelwaheb, Issaoui Wissal, S. Najet
Abstract:
This work represents a coupling between the geographic information system (GIS) and the multicriteria analysis aiming at the selection of an artificial recharge site by the treated wastewater for the Ariana governorate. On regional characteristics, bibliography and available data on artificial recharge, 13 constraints and 5 factors were hierarchically structured for the adequacy of an artificial recharge. The factors are subdivided into two main groups: environmental factors and economic factors. The adopted methodology allows a preliminary assessment of a recharge site, the weighted linear combination (WLC) and the analytical hierarchy process (AHP) in a GIS. The standardization of the criteria is carried out by the application of the different membership functions. The form and control points of the latter are defined by the consultation of the experts. The weighting of the selected criteria is allocated according to relative importance using the AHP methodology. The weighted linear combination (WLC) integrates the different criteria and factors to delineate the most suitable areas for artificial recharge site selection by treated wastewater. The results of this study showed three potential candidate sites that appear when environmental factors are more important than economic factors. These sites are ranked in descending order using the ELECTRE III method. Nevertheless, decision making for the selection of an artificial recharge site will depend on the decision makers in force.Keywords: artificial recharge site, treated wastewater, analytical hierarchy process, ELECTRE III
Procedia PDF Downloads 16719278 A Modified Decoupled Semi-Analytical Approach Based On SBFEM for Solving 2D Elastodynamic Problems
Authors: M. Fakharian, M. I. Khodakarami
Abstract:
In this paper, a new trend for improvement in semi-analytical method based on scale boundaries in order to solve the 2D elastodynamic problems is provided. In this regard, only the boundaries of the problem domain discretization are by specific sub-parametric elements. Mapping functions are uses as a class of higher-order Lagrange polynomials, special shape functions, Gauss-Lobatto -Legendre numerical integration, and the integral form of the weighted residual method, the matrix is diagonal coefficients in the equations of elastodynamic issues. Differences between study conducted and prior research in this paper is in geometry production procedure of the interpolation function and integration of the different is selected. Validity and accuracy of the present method are fully demonstrated through two benchmark problems which are successfully modeled using a few numbers of DOFs. The numerical results agree very well with the analytical solutions and the results from other numerical methods.Keywords: 2D elastodynamic problems, lagrange polynomials, G-L-Lquadrature, decoupled SBFEM
Procedia PDF Downloads 44419277 Biologically Inspired Small Infrared Target Detection Using Local Contrast Mechanisms
Authors: Tian Xia, Yuan Yan Tang
Abstract:
In order to obtain higher small target detection accuracy, this paper presents an effective algorithm inspired by the local contrast mechanism. The proposed method can enhance target signal and suppress background clutter simultaneously. In the first stage, a enhanced image is obtained using the proposed Weighted Laplacian of Gaussian. In the second stage, an adaptive threshold is adopted to segment the target. Experimental results on two changeling image sequences show that the proposed method can detect the bright and dark targets simultaneously, and is not sensitive to sea-sky line of the infrared image. So it is fit for IR small infrared target detection.Keywords: small target detection, local contrast, human vision system, Laplacian of Gaussian
Procedia PDF Downloads 46919276 Identification of Suitable Rainwater Harvesting Sites Using Geospatial Techniques with AHP in Chacha Watershed, Jemma Sub-Basin Upper Blue Nile, Ethiopia
Authors: Abrha Ybeyn Gebremedhn, Yitea Seneshaw Getahun, Alebachew Shumye Moges, Fikrey Tesfay
Abstract:
Rainfed agriculture in Ethiopia has failed to produce enough food, to achieve the increasing demand for food. Pinpointing the appropriate site for rainwater harvesting (RWH) have a substantial contribution to increasing the available water and enhancing agricultural productivity. The current study related to the identification of the potential RWH sites was conducted at the Chacha watershed central highlands of Ethiopia which is endowed with rugged topography. The Geographic Information System with Analytical Hierarchy Process was used to generate the different maps for identifying appropriate sites for RWH. In this study, 11 factors that determine the RWH locations including slope, soil texture, runoff depth, land cover type, annual average rainfall, drainage density, lineament intensity, hydrologic soil group, antecedent moisture content, and distance to the roads were considered. The overall analyzed result shows that 10.50%, 71.10%, 17.90%, and 0.50% of the areas were found under highly, moderately, marginally suitable, and unsuitable areas for RWH, respectively. The RWH site selection was found highly dependent on a slope, soil texture, and runoff depth; moderately dependent on drainage density, annual average rainfall, and land use land cover; but less dependent on the other factors. The highly suitable areas for rainwater harvesting expansion are lands having a flat topography with a soil textural class of high-water holding capacity that can produce high runoff depth. The application of this study could be a baseline for planners and decision-makers and support any strategy adoption for appropriate RWH site selection.Keywords: runoff depth, antecedent moisture condition, AHP, weighted overlay, water resource
Procedia PDF Downloads 5419275 A Data-Driven Monitoring Technique Using Combined Anomaly Detectors
Authors: Fouzi Harrou, Ying Sun, Sofiane Khadraoui
Abstract:
Anomaly detection based on Principal Component Analysis (PCA) was studied intensively and largely applied to multivariate processes with highly cross-correlated process variables. Monitoring metrics such as the Hotelling's T2 and the Q statistics are usually used in PCA-based monitoring to elucidate the pattern variations in the principal and residual subspaces, respectively. However, these metrics are ill suited to detect small faults. In this paper, the Exponentially Weighted Moving Average (EWMA) based on the Q and T statistics, T2-EWMA and Q-EWMA, were developed for detecting faults in the process mean. The performance of the proposed methods was compared with that of the conventional PCA-based fault detection method using synthetic data. The results clearly show the benefit and the effectiveness of the proposed methods over the conventional PCA method, especially for detecting small faults in highly correlated multivariate data.Keywords: data-driven method, process control, anomaly detection, dimensionality reduction
Procedia PDF Downloads 29919274 Hardy Type Inequalities of Two-Dimensional on Time Scales via Steklov Operator
Authors: Wedad Albalawi
Abstract:
The mathematical inequalities have been the core of mathematical study and used in almost all branches of mathematics as well in various areas of science and engineering. The inequalities by Hardy, Littlewood and Polya were the first significant composition of several science. This work presents fundamental ideas, results and techniques and it has had much influence on research in various branches of analysis. Since 1934, various inequalities have been produced and studied in the literature. Furthermore, some inequalities have been formulated by some operators; in 1989, weighted Hardy inequalities have been obtained for integration operators. Then, they obtained weighted estimates for Steklov operators that were used in the solution of the Cauchy problem for the wave equation. They were improved upon in 2011 to include the boundedness of integral operators from the weighted Sobolev space to the weighted Lebesgue space. Some inequalities have been demonstrated and improved using the Hardy–Steklov operator. Recently, a lot of integral inequalities have been improved by differential operators. Hardy inequality has been one of the tools that is used to consider integrity solutions of differential equations. Then dynamic inequalities of Hardy and Coposon have been extended and improved by various integral operators. These inequalities would be interesting to apply in different fields of mathematics (functional spaces, partial differential equations, mathematical modeling). Some inequalities have been appeared involving Copson and Hardy inequalities on time scales to obtain new special version of them. A time scale is defined as a closed subset contains real numbers. Then the inequalities of time scales version have received a lot of attention and has had a major field in both pure and applied mathematics. There are many applications of dynamic equations on time scales to quantum mechanics, electrical engineering, neural networks, heat transfer, combinatorics, and population dynamics. This study focuses on double integrals to obtain new time-scale inequalities of Copson driven by Steklov operator. They will be applied in the solution of the Cauchy problem for the wave equation. The proof can be done by introducing restriction on the operator in several cases. In addition, the obtained inequalities done by using some concepts in time scale version such as time scales calculus, theorem of Fubini and the inequality of H¨older.Keywords: time scales, inequality of Hardy, inequality of Coposon, Steklov operator
Procedia PDF Downloads 7719273 Group Decision Making through Interval-Valued Intuitionistic Fuzzy Soft Set TOPSIS Method Using New Hybrid Score Function
Authors: Syed Talib Abbas Raza, Tahseen Ahmed Jilani, Saleem Abdullah
Abstract:
This paper presents interval-valued intuitionistic fuzzy soft sets based TOPSIS method for group decision making. The interval-valued intuitionistic fuzzy soft set is a mutation of an interval-valued intuitionistic fuzzy set and soft set. In group decision making problems IVIFSS makes the process much more algebraically elegant. We have used weighted arithmetic averaging operator for aggregating the information and define a new Hybrid Score Function as metric tool for comparison between interval-valued intuitionistic fuzzy values. In an illustrative example we have applied the developed method to a criminological problem. We have developed a group decision making model for integrating the imprecise and hesitant evaluations of multiple law enforcement agencies working on target killing cases in the country.Keywords: group decision making, interval-valued intuitionistic fuzzy soft set, TOPSIS, score function, criminology
Procedia PDF Downloads 60519272 Optimization of Smart Beta Allocation by Momentum Exposure
Authors: J. B. Frisch, D. Evandiloff, P. Martin, N. Ouizille, F. Pires
Abstract:
Smart Beta strategies intend to be an asset management revolution with reference to classical cap-weighted indices. Indeed, these strategies allow a better control on portfolios risk factors and an optimized asset allocation by taking into account specific risks or wishes to generate alpha by outperforming indices called 'Beta'. Among many strategies independently used, this paper focuses on four of them: Minimum Variance Portfolio, Equal Risk Contribution Portfolio, Maximum Diversification Portfolio, and Equal-Weighted Portfolio. Their efficiency has been proven under constraints like momentum or market phenomenon, suggesting a reconsideration of cap-weighting. To further increase strategy return efficiency, it is proposed here to compare their strengths and weaknesses inside time intervals corresponding to specific identifiable market phases, in order to define adapted strategies depending on pre-specified situations. Results are presented as performance curves from different combinations compared to a benchmark. If a combination outperforms the applicable benchmark in well-defined actual market conditions, it will be preferred. It is mainly shown that such investment 'rules', based on both historical data and evolution of Smart Beta strategies, and implemented according to available specific market data, are providing very interesting optimal results with higher return performance and lower risk. Such combinations have not been fully exploited yet and justify present approach aimed at identifying relevant elements characterizing them.Keywords: smart beta, minimum variance portfolio, equal risk contribution portfolio, maximum diversification portfolio, equal weighted portfolio, combinations
Procedia PDF Downloads 34019271 Kernel-Based Double Nearest Proportion Feature Extraction for Hyperspectral Image Classification
Authors: Hung-Sheng Lin, Cheng-Hsuan Li
Abstract:
Over the past few years, kernel-based algorithms have been widely used to extend some linear feature extraction methods such as principal component analysis (PCA), linear discriminate analysis (LDA), and nonparametric weighted feature extraction (NWFE) to their nonlinear versions, kernel principal component analysis (KPCA), generalized discriminate analysis (GDA), and kernel nonparametric weighted feature extraction (KNWFE), respectively. These nonlinear feature extraction methods can detect nonlinear directions with the largest nonlinear variance or the largest class separability based on the given kernel function. Moreover, they have been applied to improve the target detection or the image classification of hyperspectral images. The double nearest proportion feature extraction (DNP) can effectively reduce the overlap effect and have good performance in hyperspectral image classification. The DNP structure is an extension of the k-nearest neighbor technique. For each sample, there are two corresponding nearest proportions of samples, the self-class nearest proportion and the other-class nearest proportion. The term “nearest proportion” used here consider both the local information and other more global information. With these settings, the effect of the overlap between the sample distributions can be reduced. Usually, the maximum likelihood estimator and the related unbiased estimator are not ideal estimators in high dimensional inference problems, particularly in small data-size situation. Hence, an improved estimator by shrinkage estimation (regularization) is proposed. Based on the DNP structure, LDA is included as a special case. In this paper, the kernel method is applied to extend DNP to kernel-based DNP (KDNP). In addition to the advantages of DNP, KDNP surpasses DNP in the experimental results. According to the experiments on the real hyperspectral image data sets, the classification performance of KDNP is better than that of PCA, LDA, NWFE, and their kernel versions, KPCA, GDA, and KNWFE.Keywords: feature extraction, kernel method, double nearest proportion feature extraction, kernel double nearest feature extraction
Procedia PDF Downloads 34619270 Using the Smith-Waterman Algorithm to Extract Features in the Classification of Obesity Status
Authors: Rosa Figueroa, Christopher Flores
Abstract:
Text categorization is the problem of assigning a new document to a set of predetermined categories, on the basis of a training set of free-text data that contains documents whose category membership is known. To train a classification model, it is necessary to extract characteristics in the form of tokens that facilitate the learning and classification process. In text categorization, the feature extraction process involves the use of word sequences also known as N-grams. In general, it is expected that documents belonging to the same category share similar features. The Smith-Waterman (SW) algorithm is a dynamic programming algorithm that performs a local sequence alignment in order to determine similar regions between two strings or protein sequences. This work explores the use of SW algorithm as an alternative to feature extraction in text categorization. The dataset used for this purpose, contains 2,610 annotated documents with the classes Obese/Non-Obese. This dataset was represented in a matrix form using the Bag of Word approach. The score selected to represent the occurrence of the tokens in each document was the term frequency-inverse document frequency (TF-IDF). In order to extract features for classification, four experiments were conducted: the first experiment used SW to extract features, the second one used unigrams (single word), the third one used bigrams (two word sequence) and the last experiment used a combination of unigrams and bigrams to extract features for classification. To test the effectiveness of the extracted feature set for the four experiments, a Support Vector Machine (SVM) classifier was tuned using 20% of the dataset. The remaining 80% of the dataset together with 5-Fold Cross Validation were used to evaluate and compare the performance of the four experiments of feature extraction. Results from the tuning process suggest that SW performs better than the N-gram based feature extraction. These results were confirmed by using the remaining 80% of the dataset, where SW performed the best (accuracy = 97.10%, weighted average F-measure = 97.07%). The second best was obtained by the combination of unigrams-bigrams (accuracy = 96.04, weighted average F-measure = 95.97) closely followed by the bigrams (accuracy = 94.56%, weighted average F-measure = 94.46%) and finally unigrams (accuracy = 92.96%, weighted average F-measure = 92.90%).Keywords: comorbidities, machine learning, obesity, Smith-Waterman algorithm
Procedia PDF Downloads 29819269 Multi-Criteria Goal Programming Model for Sustainable Development of India
Authors: Irfan Ali, Srikant Gupta, Aquil Ahmed
Abstract:
Every country needs a sustainable development (SD) for its economic growth by forming suitable policies and initiative programs for the development of different sectors of the country. This paper is comprised of modeling and optimization of different sectors of India that form a multi-criterion model. In this paper, we developed a fractional goal programming (FGP) model that helps in providing the efficient allocation of resources simultaneously by achieving the sustainable goals in gross domestic product (GDP), electricity consumption (EC) and greenhouse gasses (GHG) emission by the year 2030. Also, a weighted model of FGP is presented to obtain varying solution according to the priorities set by the policy maker for achieving future goals of GDP growth, EC, and GHG emission. The presented models provide a useful insight to the decision makers for implementing strategies in a different sector.Keywords: sustainable and economic development, multi-objective fractional programming, fuzzy goal programming, weighted fuzzy goal programming
Procedia PDF Downloads 22319268 Optimal Design for SARMA(P,Q)L Process of EWMA Control Chart
Authors: Yupaporn Areepong
Abstract:
The main goal of this paper is to study Statistical Process Control (SPC) with Exponentially Weighted Moving Average (EWMA) control chart when observations are serially-correlated. The characteristic of control chart is Average Run Length (ARL) which is the average number of samples taken before an action signal is given. Ideally, an acceptable ARL of in-control process should be enough large, so-called (ARL0). Otherwise it should be small when the process is out-of-control, so-called Average of Delay Time (ARL1) or a mean of true alarm. We find explicit formulas of ARL for EWMA control chart for Seasonal Autoregressive and Moving Average processes (SARMA) with Exponential white noise. The results of ARL obtained from explicit formula and Integral equation are in good agreement. In particular, this formulas for evaluating (ARL0) and (ARL1) be able to get a set of optimal parameters which depend on smoothing parameter (λ) and width of control limit (H) for designing EWMA chart with minimum of (ARL1).Keywords: average run length, optimal parameters, exponentially weighted moving average (EWMA), control chart
Procedia PDF Downloads 56019267 Digital Library Evaluation by SWARA-WASPAS Method
Authors: Mehmet Yörükoğlu, Serhat Aydın
Abstract:
Since the discovery of the manuscript, mechanical methods for storing, transferring and using the information have evolved into digital methods over the time. In this process, libraries that are the center of the information have also become digitized and become accessible from anywhere and at any time in the world by taking on a structure that has no physical boundaries. In this context, some criteria for information obtained from digital libraries have become more important for users. This paper evaluates the user criteria from different perspectives that make a digital library more useful. The Step-Wise Weight Assessment Ratio Analysis-Weighted Aggregated Sum Product Assessment (SWARA-WASPAS) method is used with flexibility and easy calculation steps for the evaluation of digital library criteria. Three different digital libraries are evaluated by information technology experts according to five conflicting main criteria, ‘interface design’, ‘effects on users’, ‘services’, ‘user engagement’ and ‘context’. Finally, alternatives are ranked in descending order.Keywords: digital library, multi criteria decision making, SWARA-WASPAS method
Procedia PDF Downloads 15219266 An Application of Geographic Information System to Select Areas for Sanitary Landfill in Bang Nok- Khwaek Municipality
Authors: Musthaya Patchanee
Abstract:
The study of Sanitary landfill in Bang Nok-khwaek municipality consists of two procedures. First, to survey and create the spatial database by using physical factor, environmental factor, economical factor and social factor to follow the method of Geographic information system: GIS, second, to analyze the proper spatial for allocating the sanitary landfill in Bang Nok-khwaek municipality by using Overlay techniques to calculate the weighting linear total in Arc GIS program. The study found that there are 2.49 sq.km. proper spatial for the sanitary landfill in Bang Nok-khwaek municipals city which is 66.76% of the whole area. The highest proper spatial is 0.02 sq.km. which is 0.54%, The high proper spatial is 0.3 sq.km. which is 8.04%, the moderate spatial is 1.62 sq.km. which is 43.43% and the low proper spatial is 0.55 sq.km. which is 14.75%. These results will be used as the guideline to select the sanitary landfill area in accordance with sanitation standard for Subdistrict Administrative Organization and Subbdistrict Municipality in Samut Songkhram provice.Keywords: Geographic Information System (GIS), sanitary landfill, Bang Nok-Khwaek municipality, Subdistrict Administrative Organization
Procedia PDF Downloads 394