Search results for: equal weighted portfolio
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1983

Search results for: equal weighted portfolio

1803 E-Hailing Taxi Industry Management Mode Innovation Based on the Credit Evaluation

Authors: Yuan-lin Liu, Ye Li, Tian Xia

Abstract:

There are some shortcomings in Chinese existing taxi management modes. This paper suggests to establish the third-party comprehensive information management platform and put forward an evaluation model based on credit. Four indicators are used to evaluate the drivers’ credit, they are passengers’ evaluation score, driving behavior evaluation, drivers’ average bad record number, and personal credit score. A weighted clustering method is used to achieve credit level evaluation for taxi drivers. The management of taxi industry is based on the credit level, while the grade of the drivers is accorded to their credit rating. Credit rating determines the cost, income levels, the market access, useful period of license and the level of wage and bonus, as well as violation fine. These methods can make the credit evaluation effective. In conclusion, more credit data will help to set up a more accurate and detailed classification standard library.

Keywords: credit, mobile internet, e-hailing taxi, management mode, weighted cluster

Procedia PDF Downloads 288
1802 The Moment of the Optimal Average Length of the Multivariate Exponentially Weighted Moving Average Control Chart for Equally Correlated Variables

Authors: Edokpa Idemudia Waziri, Salisu S. Umar

Abstract:

The Hotellng’s T^2 is a well-known statistic for detecting a shift in the mean vector of a multivariate normal distribution. Control charts based on T have been widely used in statistical process control for monitoring a multivariate process. Although it is a powerful tool, the T statistic is deficient when the shift to be detected in the mean vector of a multivariate process is small and consistent. The Multivariate Exponentially Weighted Moving Average (MEWMA) control chart is one of the control statistics used to overcome the drawback of the Hotellng’s T statistic. In this paper, the probability distribution of the Average Run Length (ARL) of the MEWMA control chart when the quality characteristics exhibit substantial cross correlation and when the process is in-control and out-of-control was derived using the Markov Chain algorithm. The derivation of the probability functions and the moments of the run length distribution were also obtained and they were consistent with some existing results for the in-control and out-of-control situation. By simulation process, the procedure identified a class of ARL for the MEWMA control when the process is in-control and out-of-control. From our study, it was observed that the MEWMA scheme is quite adequate for detecting a small shift and a good way to improve the quality of goods and services in a multivariate situation. It was also observed that as the in-control average run length ARL0¬ or the number of variables (p) increases, the optimum value of the ARL0pt increases asymptotically and as the magnitude of the shift σ increases, the optimal ARLopt decreases. Finally, we use the example from the literature to illustrate our method and demonstrate its efficiency.

Keywords: average run length, markov chain, multivariate exponentially weighted moving average, optimal smoothing parameter

Procedia PDF Downloads 391
1801 Development of Disability Studies in Post-Transformational Central and East European Countries from the 80s until Present

Authors: Klaudia Muca

Abstract:

Disability studies as an international movement are still developing, especially in the Central and East European young democratic countries. It is crucial to recognize in what manner this development might lead to create a sustainable social environment. Thanks to disability studies the process of introducing disability studies and its main ideas might become as effective as in the 90s in the USA or other Western countries. In the Central and East Europe lack of activism in favor of the disabled in the early stages of democratic transition (i.e. the 80s and 90s) caused misrepresentation of the disabled and their experience in present political and social sphere of life. People with disabilities were made to hold a minor position in society due to political changes that introduced in fact non-equal democracy. The results of this study indicate that activism in favor of people with disabilities and works of art created by the disabled are tools that influence present disability politics. That suggests that young European democracies need to modify their current political path in order to establish more equal social policies.

Keywords: democratic transformation, disability as minority, misrepresentation of experience, non-equal democracy, sustainability

Procedia PDF Downloads 159
1800 Analysis of Spatial Heterogeneity of Residential Prices in Guangzhou: An Actual Study Based on Poi Geographically Weighted Regression Model

Authors: Zichun Guo

Abstract:

Guangzhou's housing prices have declined for a long time compared with the other three major cities. As Guangzhou's housing price ladder increases, the influencing factors of housing prices have gradually attracted attention. This article attempts to use housing price data and POI data and uses the Kriging spatial interpolation method and the geographically weighted regression model to explore the distribution of housing prices and the impact of factors. Caused, especially located in Huadu District and the city center. The response is mainly obvious in surrounding areas, which may be related to housing positioning. Economic POIs close to the city center have stronger responses. The factors affecting housing prices provide this method, which is conducive to the management and macro-control of relevant departments, better meets the demand for home purchases, and realizes financing-side reforms.

Keywords: housing prices, spatial heterogeneity, Guangzhou, POI

Procedia PDF Downloads 8
1799 Hardy Type Inequalities of Two-Dimensional on Time Scales via Steklov Operator

Authors: Wedad Albalawi

Abstract:

The mathematical inequalities have been the core of mathematical study and used in almost all branches of mathematics as well in various areas of science and engineering. The inequalities by Hardy, Littlewood and Polya were the first significant composition of several science. This work presents fundamental ideas, results and techniques and it has had much influence on research in various branches of analysis. Since 1934, various inequalities have been produced and studied in the literature. Furthermore, some inequalities have been formulated by some operators; in 1989, weighted Hardy inequalities have been obtained for integration operators. Then, they obtained weighted estimates for Steklov operators that were used in the solution of the Cauchy problem for the wave equation. They were improved upon in 2011 to include the boundedness of integral operators from the weighted Sobolev space to the weighted Lebesgue space. Some inequalities have been demonstrated and improved using the Hardy–Steklov operator. Recently, a lot of integral inequalities have been improved by differential operators. Hardy inequality has been one of the tools that is used to consider integrity solutions of differential equations. Then dynamic inequalities of Hardy and Coposon have been extended and improved by various integral operators. These inequalities would be interesting to apply in different fields of mathematics (functional spaces, partial differential equations, mathematical modeling). Some inequalities have been appeared involving Copson and Hardy inequalities on time scales to obtain new special version of them. A time scale is defined as a closed subset contains real numbers. Then the inequalities of time scales version have received a lot of attention and has had a major field in both pure and applied mathematics. There are many applications of dynamic equations on time scales to quantum mechanics, electrical engineering, neural networks, heat transfer, combinatorics, and population dynamics. This study focuses on double integrals to obtain new time-scale inequalities of Copson driven by Steklov operator. They will be applied in the solution of the Cauchy problem for the wave equation. The proof can be done by introducing restriction on the operator in several cases. In addition, the obtained inequalities done by using some concepts in time scale version such as time scales calculus, theorem of Fubini and the inequality of H¨older.

Keywords: time scales, inequality of Hardy, inequality of Coposon, Steklov operator

Procedia PDF Downloads 44
1798 Spatial Analysis of Flood Vulnerability in Highly Urbanized Area: A Case Study in Taipei City

Authors: Liang Weichien

Abstract:

Without adequate information and mitigation plan for natural disaster, the risk to urban populated areas will increase in the future as populations grow, especially in Taiwan. Taiwan is recognized as the world's high-risk areas, where an average of 5.7 times of floods occur per year should seek to strengthen coherence and consensus in how cities can plan for flood and climate change. Therefore, this study aims at understanding the vulnerability to flooding in Taipei city, Taiwan, by creating indicators and calculating the vulnerability of each study units. The indicators were grouped into sensitivity and adaptive capacity based on the definition of vulnerability of Intergovernmental Panel on Climate Change. The indicators were weighted by using Principal Component Analysis. However, current researches were based on the assumption that the composition and influence of the indicators were the same in different areas. This disregarded spatial correlation that might result in inaccurate explanation on local vulnerability. The study used Geographically Weighted Principal Component Analysis by adding geographic weighting matrix as weighting to get the different main flood impact characteristic in different areas. Cross Validation Method and Akaike Information Criterion were used to decide bandwidth and Gaussian Pattern as the bandwidth weight scheme. The ultimate outcome can be used for the reduction of damage potential by integrating the outputs into local mitigation plan and urban planning.

Keywords: flood vulnerability, geographically weighted principal components analysis, GWPCA, highly urbanized area, spatial correlation

Procedia PDF Downloads 266
1797 Using the Smith-Waterman Algorithm to Extract Features in the Classification of Obesity Status

Authors: Rosa Figueroa, Christopher Flores

Abstract:

Text categorization is the problem of assigning a new document to a set of predetermined categories, on the basis of a training set of free-text data that contains documents whose category membership is known. To train a classification model, it is necessary to extract characteristics in the form of tokens that facilitate the learning and classification process. In text categorization, the feature extraction process involves the use of word sequences also known as N-grams. In general, it is expected that documents belonging to the same category share similar features. The Smith-Waterman (SW) algorithm is a dynamic programming algorithm that performs a local sequence alignment in order to determine similar regions between two strings or protein sequences. This work explores the use of SW algorithm as an alternative to feature extraction in text categorization. The dataset used for this purpose, contains 2,610 annotated documents with the classes Obese/Non-Obese. This dataset was represented in a matrix form using the Bag of Word approach. The score selected to represent the occurrence of the tokens in each document was the term frequency-inverse document frequency (TF-IDF). In order to extract features for classification, four experiments were conducted: the first experiment used SW to extract features, the second one used unigrams (single word), the third one used bigrams (two word sequence) and the last experiment used a combination of unigrams and bigrams to extract features for classification. To test the effectiveness of the extracted feature set for the four experiments, a Support Vector Machine (SVM) classifier was tuned using 20% of the dataset. The remaining 80% of the dataset together with 5-Fold Cross Validation were used to evaluate and compare the performance of the four experiments of feature extraction. Results from the tuning process suggest that SW performs better than the N-gram based feature extraction. These results were confirmed by using the remaining 80% of the dataset, where SW performed the best (accuracy = 97.10%, weighted average F-measure = 97.07%). The second best was obtained by the combination of unigrams-bigrams (accuracy = 96.04, weighted average F-measure = 95.97) closely followed by the bigrams (accuracy = 94.56%, weighted average F-measure = 94.46%) and finally unigrams (accuracy = 92.96%, weighted average F-measure = 92.90%).

Keywords: comorbidities, machine learning, obesity, Smith-Waterman algorithm

Procedia PDF Downloads 270
1796 Numerical Modeling of Geogrid Reinforced Soil Bed under Strip Footings Using Finite Element Analysis

Authors: Ahmed M. Gamal, Adel M. Belal, S. A. Elsoud

Abstract:

This article aims to study the effect of reinforcement inclusions (geogrids) on the sand dunes bearing capacity under strip footings. In this research experimental physical model was carried out to study the effect of the first geogrid reinforcement depth (u/B), the spacing between the reinforcement (h/B) and its extension relative to the footing length (L/B) on the mobilized bearing capacity. This paper presents the numerical modeling using the commercial finite element package (PLAXIS version 8.2) to simulate the laboratory physical model, studying the same parameters previously handled in the experimental work (u/B, L/B & h/B) for the purpose of validation. In this study the soil, the geogrid, the interface element and the boundary condition are discussed with a set of finite element results and the validation. Then the validated FEM used for studying real material and dimensions of strip foundation. Based on the experimental and numerical investigation results, a significant increase in the bearing capacity of footings has occurred due to an appropriate location of the inclusions in sand. The optimum embedment depth of the first reinforcement layer (u/B) is equal to 0.25. The optimum spacing between each successive reinforcement layer (h/B) is equal to 0.75 B. The optimum Length of the reinforcement layer (L/B) is equal to 7.5 B. The optimum number of reinforcement is equal to 4 layers. The study showed a directly proportional relation between the number of reinforcement layer and the Bearing Capacity Ratio BCR, and an inversely proportional relation between the footing width and the BCR.

Keywords: reinforced soil, geogrid, sand dunes, bearing capacity

Procedia PDF Downloads 390
1795 Multi-Criteria Goal Programming Model for Sustainable Development of India

Authors: Irfan Ali, Srikant Gupta, Aquil Ahmed

Abstract:

Every country needs a sustainable development (SD) for its economic growth by forming suitable policies and initiative programs for the development of different sectors of the country. This paper is comprised of modeling and optimization of different sectors of India that form a multi-criterion model. In this paper, we developed a fractional goal programming (FGP) model that helps in providing the efficient allocation of resources simultaneously by achieving the sustainable goals in gross domestic product (GDP), electricity consumption (EC) and greenhouse gasses (GHG) emission by the year 2030. Also, a weighted model of FGP is presented to obtain varying solution according to the priorities set by the policy maker for achieving future goals of GDP growth, EC, and GHG emission. The presented models provide a useful insight to the decision makers for implementing strategies in a different sector.

Keywords: sustainable and economic development, multi-objective fractional programming, fuzzy goal programming, weighted fuzzy goal programming

Procedia PDF Downloads 195
1794 Filtering Momentum Life Cycles, Price Acceleration Signals and Trend Reversals for Stocks, Credit Derivatives and Bonds

Authors: Periklis Brakatsoulas

Abstract:

Recent empirical research shows a growing interest in investment decision-making under market anomalies that contradict the rational paradigm. Momentum is undoubtedly one of the most robust anomalies in the empirical asset pricing research and remains surprisingly lucrative ever since first documented. Although predominantly phenomena identified across equities, momentum premia are now evident across various asset classes. Yet few many attempts are made so far to provide traders a diversified portfolio of strategies across different assets and markets. Moreover, literature focuses on patterns from past returns rather than mechanisms to signal future price directions prior to momentum runs. The aim of this paper is to develop a diversified portfolio approach to price distortion signals using daily position data on stocks, credit derivatives, and bonds. An algorithm allocates assets periodically, and new investment tactics take over upon price momentum signals and across different ranking groups. We focus on momentum life cycles, trend reversals, and price acceleration signals. The main effort here concentrates on the density, time span and maturity of momentum phenomena to identify consistent patterns over time and measure the predictive power of buy-sell signals generated by these anomalies. To tackle this, we propose a two-stage modelling process. First, we generate forecasts on core macroeconomic drivers. Secondly, satellite models generate market risk forecasts using the core driver projections generated at the first stage as input. Moreover, using a combination of the ARFIMA and FIGARCH models, we examine the dependence of consecutive observations across time and portfolio assets since long memory behavior in volatilities of one market appears to trigger persistent volatility patterns across other markets. We believe that this is the first work that employs evidence of volatility transmissions among derivatives, equities, and bonds to identify momentum life cycle patterns.

Keywords: forecasting, long memory, momentum, returns

Procedia PDF Downloads 78
1793 Optimal Design for SARMA(P,Q)L Process of EWMA Control Chart

Authors: Yupaporn Areepong

Abstract:

The main goal of this paper is to study Statistical Process Control (SPC) with Exponentially Weighted Moving Average (EWMA) control chart when observations are serially-correlated. The characteristic of control chart is Average Run Length (ARL) which is the average number of samples taken before an action signal is given. Ideally, an acceptable ARL of in-control process should be enough large, so-called (ARL0). Otherwise it should be small when the process is out-of-control, so-called Average of Delay Time (ARL1) or a mean of true alarm. We find explicit formulas of ARL for EWMA control chart for Seasonal Autoregressive and Moving Average processes (SARMA) with Exponential white noise. The results of ARL obtained from explicit formula and Integral equation are in good agreement. In particular, this formulas for evaluating (ARL0) and (ARL1) be able to get a set of optimal parameters which depend on smoothing parameter (λ) and width of control limit (H) for designing EWMA chart with minimum of (ARL1).

Keywords: average run length, optimal parameters, exponentially weighted moving average (EWMA), control chart

Procedia PDF Downloads 531
1792 Detection Method of Federated Learning Backdoor Based on Weighted K-Medoids

Authors: Xun Li, Haojie Wang

Abstract:

Federated learning is a kind of distributed training and centralized training mode, which is of great value in the protection of user privacy. In order to solve the problem that the model is vulnerable to backdoor attacks in federated learning, a backdoor attack detection method based on a weighted k-medoids algorithm is proposed. First of all, this paper collates the update parameters of the client to construct a vector group, then uses the principal components analysis (PCA) algorithm to extract the corresponding feature information from the vector group, and finally uses the improved k-medoids clustering algorithm to identify the normal and backdoor update parameters. In this paper, the backdoor is implanted in the federation learning model through the model replacement attack method in the simulation experiment, and the update parameters from the attacker are effectively detected and removed by the defense method proposed in this paper.

Keywords: federated learning, backdoor attack, PCA, k-medoids, backdoor defense

Procedia PDF Downloads 78
1791 Implementation of the Recursive Formula for Evaluation of the Strength of Daniels' Bundle

Authors: Vaclav Sadilek, Miroslav Vorechovsky

Abstract:

The paper deals with the classical fiber bundle model of equal load sharing, sometimes referred to as the Daniels' bundle or the democratic bundle. Daniels formulated a multidimensional integral and also a recursive formula for evaluation of the strength cumulative distribution function. This paper describes three algorithms for evaluation of the recursive formula and also their implementations with source codes in high-level programming language Python. A comparison of the algorithms are provided with respect to execution time. Analysis of orders of magnitudes of addends in the recursion is also provided.

Keywords: equal load sharing, mpmath, python, strength of Daniels' bundle

Procedia PDF Downloads 376
1790 An Empirical Analysis of the Effects of Corporate Derivatives Use on the Underlying Stock Price Exposure: South African Evidence

Authors: Edson Vengesai

Abstract:

Derivative products have become essential instruments in portfolio diversification, price discovery, and, most importantly, risk hedging. Derivatives are complex instruments; their valuation, volatility implications, and real impact on the underlying assets' behaviour are not well understood. Little is documented empirically, with conflicting conclusions on how these instruments affect firm risk exposures. Given the growing interest in using derivatives in risk management and portfolio engineering, this study examines the practical impact of derivative usage on the underlying stock price exposure and systematic risk. The paper uses data from South African listed firms. The study employs GARCH models to understand the effect of derivative uses on conditional stock volatility. The GMM models are used to estimate the effect of derivatives use on stocks' systematic risk as measured by Beta and on the total risk of stocks as measured by the standard deviation of returns. The results provide evidence on whether derivatives use is instrumental in reducing stock returns' systematic and total risk. The results are subjected to numerous controls for robustness, including financial leverage, firm size, growth opportunities, and macroeconomic effects.

Keywords: derivatives use, hedging, volatility, stock price exposure

Procedia PDF Downloads 77
1789 Management as a Proxy for Firm Quality

Authors: Petar Dobrev

Abstract:

There is no agreed-upon definition of firm quality. While profitability and stock performance often qualify as popular proxies of quality, in this project, we aim to identify quality without relying on a firm’s financial statements or stock returns as selection criteria. Instead, we use firm-level data on management practices across small to medium-sized U.S. manufacturing firms from the World Management Survey (WMS) to measure firm quality. Each firm in the WMS dataset is assigned a mean management score from 0 to 5, with higher scores identifying better-managed firms. This management score serves as our proxy for firm quality and is the sole criteria we use to separate firms into portfolios comprised of high-quality and low-quality firms. We define high-quality (low-quality) firms as those firms with a management score of one standard deviation above (below) the mean. To study whether this proxy for firm quality can identify better-performing firms, we link this data to Compustat and The Center for Research in Security Prices (CRSP) to obtain firm-level data on financial performance and monthly stock returns, respectively. We find that from 1999 to 2019 (our sample data period), firms in the high-quality portfolio are consistently more profitable — higher operating profitability and return on equity compared to low-quality firms. In addition, high-quality firms also exhibit a lower risk of bankruptcy — a higher Altman Z-score. Next, we test whether the stocks of the firms in the high-quality portfolio earn superior risk-adjusted excess returns. We regress the monthly excess returns on each portfolio on the Fama-French 3-factor, 4-factor, and 5-factor models, the betting-against-beta factor, and the quality-minus-junk factor. We find no statistically significant differences in excess returns between both portfolios, suggesting that stocks of high-quality (well managed) firms do not earn superior risk-adjusted returns compared to low-quality (poorly managed) firms. In short, our proxy for firm quality, the WMS management score, can identify firms with superior financial performance (higher profitability and reduced risk of bankruptcy). However, our management proxy cannot identify stocks that earn superior risk-adjusted returns, suggesting no statistically significant relationship between managerial quality and stock performance.

Keywords: excess stock returns, management, profitability, quality

Procedia PDF Downloads 70
1788 Evaluation of Fetal brain using Magnetic Resonance Imaging

Authors: Mahdi Farajzadeh Ajirlou

Abstract:

Ordinary fetal brain development can be considered by in vivo attractive reverberation imaging (MRI) from the 18th gestational week (GW) to term and depends fundamentally on T2-weighted and diffusion-weighted (DW) arrangements. The foremost commonly suspected brain pathologies alluded to fetal MRI for assist assessment are ventriculomegaly, lost corpus callosum, and anomalies of the posterior fossa. Brain division could be a crucial to begin with step in neuroimage examination. Within the case of fetal MRI it is especially challenging and critical due to the subjective introduction of the hatchling, organs that encompass the fetal head, and irregular fetal movement. A few promising strategies have been proposed but are constrained in their execution in challenging cases and in realtime division. Fetal MRI is routinely performed on a 1.5-Tesla scanner without maternal or fetal sedation. The mother lies recumbent amid the course of the examination, the length of which is ordinarily 45 to 60 minutes. The accessibility and continuous approval of standardizing fetal brain development directions will give critical devices for early discovery of impeded fetal brain development upon which to oversee high-risk pregnancies.

Keywords: brain, fetal, MRI, imaging

Procedia PDF Downloads 45
1787 Multi-Objective Variable Neighborhood Search Algorithm to Solving Scheduling Problem with Transportation Times

Authors: Majid Khalili

Abstract:

This paper deals with a bi-objective hybrid no-wait flowshop scheduling problem minimizing the makespan and total weighted tardiness, in which we consider transportation times between stages. Obtaining an optimal solution for this type of complex, large-sized problem in reasonable computational time by using traditional approaches and optimization tools is extremely difficult. This paper presents a new multi-objective variable neighborhood algorithm (MOVNS). A set of experimental instances are carried out to evaluate the algorithm by advanced multi-objective performance measures. The algorithm is carefully evaluated for its performance against available algorithm by means of multi-objective performance measures and statistical tools. The related results show that a variant of our proposed MOVNS provides sound performance comparing with other algorithms.

Keywords: no-wait hybrid flowshop scheduling; multi-objective variable neighborhood algorithm; makespan; total weighted tardiness

Procedia PDF Downloads 397
1786 Comparison of Chest Weight of Pure and Mixed Races Kabood 30-Day Squab

Authors: Sepehr Moradi, Mehdi Asadi Rad

Abstract:

The aim of this study is to evaluate and compare chest weight of pure and mixed races Kabood 30-day Pigeons to investigate about their sex, race, and some auxiliary variables. In this paper, 62 pieces of pigeons as 31 male and female pairs with equal age are studied randomly. A natural incubation was done from each pair. All produced chickens were slaughtered at 30 days age after 12 hours hunger. Then their chests were weighted by a scale with one gram precision. A covariance analysis was used since there were many auxiliary variables and unequal observations. SAS software was used for statistical analysis. Mean weight of chests in pure race (Kabood-Kabood) with 8 records, 123.8±32.3g and mixed races of Kabood-Namebar, Kabood-Parvazy, Kabood-Tizpar, Namebar-Kabood, Tizpar-Kabood, and Parvazi-Kabood with 8, 8, 6, 12, 10, and 10 records were 139.4±23.5, 7/122±23.8, 124.7±30.1, 50.3±29.3, 51.4±26.4, and 137±28.6 gr, respectively. Mean weight of 30-day chests in male and female sex were 87.3±2.5 and 82.7±2.6g, respectively. Difference chest weight of 30-day chests of Kabood-Kabood race with Kabood-Namebar, Kabood-Parvazi, Tizpar-Kabood, Kabood-Tizpar, Namebar-Kabood and Parvazi-Kabood mixed races was not significant. Effect of sex was also significant in 5% level (P<0.05), but mutual effect of sex and race was not significant. Auxiliary variable of father weight was significant in 1% level (p < 0.01), but auxiliary variable of mother weight was not significant. The results showed that most and least weights belonged to Kabood-Namebar and Namebar-Kabood.

Keywords: squab, Kabood race, 30-day chest weight, pigeons

Procedia PDF Downloads 129
1785 The Design of a Die for the Processing of Aluminum through Equal Channel Angular Pressing

Authors: P. G. F. Siqueira, N. G. S. Almeida, P. M. A. Stemler, P. R. Cetlin, M. T. P. Aguilar

Abstract:

The processing of metals through Equal Channel Angular Pressing (ECAP) leads to their remarkable strengthening. The ECAP dies control the amount of strain imposed on the material through its geometry, especially through the angle between the die channels, and thus the microstructural and mechanical properties evolution of the material. The present study describes the design of an ECAP die whose utilization and maintenance are facilitated, and that also controls the eventual undesired flow of the material during processing. The proposed design was validated through numerical simulations procedures using commercial software. The die was manufactured according to the present design and tested. Tests using aluminum alloys also indicated to be suitable for the processing of higher strength alloys.

Keywords: ECAP, mechanical design, numerical methods, SPD

Procedia PDF Downloads 113
1784 Electricity Sector's Status in Lebanon and Portfolio Optimization for the Future Electricity Generation Scenarios

Authors: Nour Wehbe

Abstract:

The Lebanese electricity sector is at the heart of a deep crisis. Electricity in Lebanon is supplied by Électricité du Liban (EdL) which has to suffer from technical and financial deficiencies for decades and proved to be insufficient and deficient as the demand still exceeds the supply. As a result, backup generation is widespread throughout Lebanon. The sector costs massive government resources and, on top of it, consumers pay massive additional amounts for satisfying their electrical needs. While the developed countries have been investing in renewable energy for the past two decades, the Lebanese government realizes the importance of adopting such energy sourcing strategies for the upgrade of the electricity sector in the country. The diversification of the national electricity generation mix has increased considerably in Lebanon's energy planning agenda, especially that a detailed review of the energy potential in Lebanon has revealed a great potential of solar and wind energy resources, a considerable potential of biomass resource, and an important hydraulic potential in Lebanon. This paper presents a review of the energy status of Lebanon, and illustrates a detailed review of the EDL structure with the existing problems and recommended solutions. In addition, scenarios reflecting implementation of policy projects are presented, and conclusions are drawn on the usefulness of a proposed evaluation methodology and the effectiveness of the adopted new energy policy for the electrical sector in Lebanon.

Keywords: EdL Electricite du Liban, portfolio optimization, electricity generation mix, mean-variance approach

Procedia PDF Downloads 224
1783 Assessment of Ground Water Potential Zone: A Case Study of Paramakudi Taluk, Ramanathapuram, Tamilnadu, India

Authors: Shri Devi

Abstract:

This paper was conducted to see the ground water potential zones in Paramakudi taluk, Ramanathapuram,Tamilnadu India with a total areal extent of 745 sq. km. The various thematic map have been prepared for the study such as soil, geology, geomorphology, drainage, land use of the particular study area using the Toposheet of 1: 50000. The digital elevation model (DEM) has been generated from contour interval of 10m and also the slope was prepared. The ground water potential zone of the region was obtained using the weighted overlay analysis for which all the thematic maps were overlayed in arc gis 10.2. For the particular output the ranking has been given for all the parameters of each thematic layer with different weightage such as 25% was given to soil, 25% to geomorphology and land use land cover also 25%, slope 15%, lineament with 5% and drainage streams with 5 percentage. Using these entire potential zone maps was prepared which was overlayed with the village map to check the region which has good, moderate and low groundwater potential zone.

Keywords: GIS, ground water, Paramakudi, weighted overlay analysis

Procedia PDF Downloads 315
1782 Understanding the Complexities of Consumer Financial Spinning

Authors: Olivier Mesly

Abstract:

This research presents a conceptual framework termed “Consumer Financial Spinning” (CFS) to analyze consumer behavior in the financial/economic markets. This phenomenon occurs when consumers of high-stakes financial products accumulate unsustainable debt, leading them to detach from their initial financial hierarchy of needs, wealth-related goals, and preferences regarding their household portfolio of assets. The daring actions of these consumers, forming a dark financial triangle, are characterized by three behaviors: overconfidence, the use of rationed rationality, and deceitfulness. We show that we can incorporate CFS into the traditional CAPM and Markovitz’ portfolio optimization models to create a framework that explains such market phenomena as the global financial crisis, highlighting the antecedents and consequences of ill-conceived speculation. Because this is a conceptual paper, there is no methodology with respect to ground studies. However, we apply modeling principles derived from the data percolation methodology, which contains tenets explicating how to structure concepts. A simulation test of the proposed framework is conducted; it demonstrates the conditions under which the relationship between expected returns and risk may deviate from linearity. The analysis and conceptual findings are particularly relevant both theoretically and pragmatically as they shed light on the psychological conditions that drive intense speculation, which can lead to market turmoil. Armed with such understanding, regulators are better equipped to propose solutions before the economic problems become out of control.

Keywords: consumer financial spinning, rationality, deceitfulness, overconfidence, CAPM

Procedia PDF Downloads 17
1781 A Pedagogical Case Study on Consumer Decision Making Models: A Selection of Smart Phone Apps

Authors: Yong Bum Shin

Abstract:

This case focuses on Weighted additive difference, Conjunctive, Disjunctive, and Elimination by aspects methodologies in consumer decision-making models and the Simple additive weighting (SAW) approach in the multi-criteria decision-making (MCDM) area. Most decision-making models illustrate that the rank reversal phenomenon is unpreventable. This paper presents that rank reversal occurs in popular managerial methods such as Weighted Additive Difference (WAD), Conjunctive Method, Disjunctive Method, Elimination by Aspects (EBA) and MCDM methods as well as such as the Simple Additive Weighting (SAW) and finally Unified Commensurate Multiple (UCM) models which successfully addresses these rank reversal problems in most popular MCDM methods in decision-making area.

Keywords: multiple criteria decision making, rank inconsistency, unified commensurate multiple, analytic hierarchy process

Procedia PDF Downloads 58
1780 Building Blocks for the Next eGovernment Era: Exploratory Study Based on Dubai and UAE’s Ministry of Happiness Communication in 2020

Authors: Diamantino Ribeiro, António Pedro Costa, Jorge Remondes

Abstract:

Dubai and the UAE governments have been investing in technology and digital communication for a long time. These governments are pioneers in introducing innovative strategies, policies and projects. They are also recognized worldwide for defining and implementing long term public programs. In terms of eGovernment Dubai and the UAE rank among the world’s most advanced. Both governments have surprised the world a few years ago by creating a Happiness Ministry. This paper focuses on UAE’s government digital strategies and its approach to the next era. The main goal of this exploratory study is to understand the new era of eGovernment and transfer of the happiness and wellness programs. Data were collected from the corpus latente and analysis was anchored in qualitative methodology using content analysis and observation as analysis techniques. The study allowed to highlight that the 2020 government reshuffle has a strong focus on digital reorganisation and digital sustainability, one of the newest trends in sustainability. Regarding happiness and wellbeing portfolio, we were able to observe that there has been a major change within the government organisation: The Ministry of Happiness was extinct and the Ministry of Community Development will manage the so-called ‘Happiness Portfolio’. Additionally, our observation allowed to note the government dual approach to governance: one through digital transformation, thus enhancing the digital sustainability process and, the second one trough government development.

Keywords: ministry of happiness, eGovernment, communication, digital sustainability

Procedia PDF Downloads 119
1779 A Weighted K-Medoids Clustering Algorithm for Effective Stability in Vehicular Ad Hoc Networks

Authors: Rejab Hajlaoui, Tarek Moulahi, Hervé Guyennet

Abstract:

In a highway scenario, the vehicle speed can exceed 120 kmph. Therefore, any vehicle can enter or leave the network within a very short time. This mobility adversely affects the network connectivity and decreases the life time of all established links. To ensure an effective stability in vehicular ad hoc networks with minimum broadcasting storm, we have developed a weighted algorithm based on the k-medoids clustering algorithm (WKCA). Indeed, the number of clusters and the initial cluster heads will not be selected randomly as usual, but considering the available transmission range and the environment size. Then, to ensure optimal assignment of nodes to clusters in both k-medoids phases, the combined weight of any node will be computed according to additional metrics including direction, relative speed and proximity. Empirical results prove that in addition to the convergence speed that characterizes the k-medoids algorithm, our proposed model performs well both AODV-Clustering and OLSR-Clustering protocols under different densities and velocities in term of end-to-end delay, packet delivery ratio, and throughput.

Keywords: communication, clustering algorithm, k-medoids, sensor, vehicular ad hoc network

Procedia PDF Downloads 204
1778 An Investigation of the Effects of Emotional Experience Induction on Mirror Neurons System Activity with Regard to Spectrum of Depressive Symptoms

Authors: Elyas Akbari, Jafar Hasani, Newsha Dehestani, Mohammad Khaleghi, Alireza Moradi

Abstract:

The aim of the present study was to assess the effect of emotional experience induction in the mirror neurons systems (MNS) activity with regard to the spectrum of depressive symptoms. For this purpose, at first stage, 449 students of Kharazmi University of Tehran were selected randomly and completed the second version of the Beck Depression Inventory (BDI-II). Then, 36 students with standard Z-score equal or above +1.5 and equal or equal or below -1.5 were selected to construct two groups of high and low spectrum of depressive symptoms. In the next stage, the basic activity of MNS was recorded (mu wave) before presenting the positive and negative emotional video clips by Electroencephalography (EEG) technique. The findings related to emotion induction (neutral, negative and positive emotion) demonstrated that the activity of recorded mirror neuron areas had a significant difference between the depressive and non-depressive groups. These findings suggest that probably processing of negative emotions in depressive individuals is due to the idea that the mirror neurons in motor cortex matched up the activity of cognitive regions with the person’s schema. Considering the results of the present study, it could be said that the MNS provides a substrate where emotional disorders can be studied and evaluated.

Keywords: emotional experiences, mirror neurons, depressive symptoms, negative and positive emotion

Procedia PDF Downloads 330
1777 A Data Science Pipeline for Algorithmic Trading: A Comparative Study in Applications to Finance and Cryptoeconomics

Authors: Luyao Zhang, Tianyu Wu, Jiayi Li, Carlos-Gustavo Salas-Flores, Saad Lahrichi

Abstract:

Recent advances in AI have made algorithmic trading a central role in finance. However, current research and applications are disconnected information islands. We propose a generally applicable pipeline for designing, programming, and evaluating algorithmic trading of stock and crypto tokens. Moreover, we provide comparative case studies for four conventional algorithms, including moving average crossover, volume-weighted average price, sentiment analysis, and statistical arbitrage. Our study offers a systematic way to program and compare different trading strategies. Moreover, we implement our algorithms by object-oriented programming in Python3, which serves as open-source software for future academic research and applications.

Keywords: algorithmic trading, AI for finance, fintech, machine learning, moving average crossover, volume weighted average price, sentiment analysis, statistical arbitrage, pair trading, object-oriented programming, python3

Procedia PDF Downloads 113
1776 Site Selection of CNG Station by Using FUZZY-AHP Model (Case Study: Gas Zone 4, Tehran City Iran)

Authors: Hamidrza Joodaki

Abstract:

The most complex issue in urban land use planning is site selection that needs to assess the verity of elements and factors. Multi Criteria Decision Making (MCDM) methods are the best approach to deal with complex problems. In this paper, combination of the analytical hierarchy process (AHP) model and FUZZY logic was used as MCDM methods to select the best site for gas station in the 4th gas zone of Tehran. The first and the most important step in FUZZY-AHP model is selection of criteria and sub-criteria. Population, accessibility, proximity and natural disasters were considered as the main criteria in this study. After choosing the criteria, they were weighted based on AHP by EXPERT CHOICE software, and FUZZY logic was used to enhance accuracy and to approach the reality. After these steps, criteria layers were produced and weighted based on FUZZY-AHP model in GIS. Finally, through ARC GIS software, the layers were integrated and the 4th gas zone in TEHRAN was selected as the best site to locate gas station.

Keywords: multiple criteria decision making (MCDM), analytic hierarchy process (AHP), FUZZY logic, geographic information system (GIS)

Procedia PDF Downloads 328
1775 General Purpose Graphic Processing Units Based Real Time Video Tracking System

Authors: Mallikarjuna Rao Gundavarapu, Ch. Mallikarjuna Rao, K. Anuradha Bai

Abstract:

Real Time Video Tracking is a challenging task for computing professionals. The performance of video tracking techniques is greatly affected by background detection and elimination process. Local regions of the image frame contain vital information of background and foreground. However, pixel-level processing of local regions consumes a good amount of computational time and memory space by traditional approaches. In our approach we have explored the concurrent computational ability of General Purpose Graphic Processing Units (GPGPU) to address this problem. The Gaussian Mixture Model (GMM) with adaptive weighted kernels is used for detecting the background. The weights of the kernel are influenced by local regions and are updated by inter-frame variations of these corresponding regions. The proposed system has been tested with GPU devices such as GeForce GTX 280, GeForce GTX 280 and Quadro K2000. The results are encouraging with maximum speed up 10X compared to sequential approach.

Keywords: connected components, embrace threads, local weighted kernel, structuring elements

Procedia PDF Downloads 409
1774 Changing New York Financial Clusters in the 2000s: Modeling the Impact and Policy Implication of the Global Financial Crisis

Authors: Silvia Lorenzo, Hongmian Gong

Abstract:

With the influx of research assessing the economic impact of the global financial crisis of 2007-8, a spatial analysis based on empirical data is needed to better understand the spatial significance of the financial crisis in New York, a key international financial center also considered the origin of the crisis. Using spatial statistics, the existence of financial clusters specializing in credit and securities throughout the New York metropolitan area are identified for 2000 and 2010, the time period before and after the height of the global financial crisis. Geographically Weighted Regressions are then used to examine processes underlying the formation and movement of financial geographies across state, county and ZIP codes of the New York metropolitan area throughout the 2000s with specific attention to tax regimes, employment, household income, technology, and transportation hubs. This analysis provides useful inputs for financial risk management and public policy initiatives aimed at addressing regional economic sustainability across state boundaries, while also developing the groundwork for further research on a spatial analysis of the global financial crisis.

Keywords: financial clusters, New York, global financial crisis, geographically weighted regression

Procedia PDF Downloads 276