Search results for: optimization intelligence strategy
6261 Topological Sensitivity Analysis for Reconstruction of the Inverse Source Problem from Boundary Measurement
Authors: Maatoug Hassine, Mourad Hrizi
Abstract:
In this paper, we consider a geometric inverse source problem for the heat equation with Dirichlet and Neumann boundary data. We will reconstruct the exact form of the unknown source term from additional boundary conditions. Our motivation is to detect the location, the size and the shape of source support. We present a one-shot algorithm based on the Kohn-Vogelius formulation and the topological gradient method. The geometric inverse source problem is formulated as a topology optimization one. A topological sensitivity analysis is derived from a source function. Then, we present a non-iterative numerical method for the geometric reconstruction of the source term with unknown support using a level curve of the topological gradient. Finally, we give several examples to show the viability of our presented method.Keywords: geometric inverse source problem, heat equation, topological optimization, topological sensitivity, Kohn-Vogelius formulation
Procedia PDF Downloads 3006260 Hydrodynamic Behavior Study of Fast Mono Hull and Catamaran Vessels in Calm Waters Using Free Surface Flow Analysis
Authors: Mohammad Ali Badri, Pouya Molana, Amin Rezvanpour
Abstract:
In this paper, planning catamaran and mono-hull vessels resistance and trim in calm waters were considered. Hydrodynamic analysis of fast mono-hull planning vessel was also investigated. In order to hull form geometry optimization, numerical methods of different parameters were used for this type of vessels. Hull material was selected in carbon fiber composite. Exact architectural aspects were specified and stability calculations were performed as well. Hydrodynamic calculations to extract the resistance force using semi-analytical methods and numerical modeling were carried out. Free surface numerical analysis of vessel in designed draft using finite volume method and double phase were evaluated and verified by experimental tests.Keywords: fast vessel, hydrostatic and hydrodynamic optimization, free surface flow, computational fluid dynamics
Procedia PDF Downloads 5166259 The Application of Bayesian Heuristic for Scheduling in Real-Time Private Clouds
Authors: Sahar Sohrabi
Abstract:
The emergence of Cloud data centers has revolutionized the IT industry. Private Clouds in specific provide Cloud services for certain group of customers/businesses. In a real-time private Cloud each task that is given to the system has a deadline that desirably should not be violated. Scheduling tasks in a real-time private CLoud determine the way available resources in the system are shared among incoming tasks. The aim of the scheduling policy is to optimize the system outcome which for a real-time private Cloud can include: energy consumption, deadline violation, execution time and the number of host switches. Different scheduling policies can be used for scheduling. Each lead to a sub-optimal outcome in a certain settings of the system. A Bayesian Scheduling strategy is proposed for scheduling to further improve the system outcome. The Bayesian strategy showed to outperform all selected policies. It also has the flexibility in dealing with complex pattern of incoming task and has the ability to adapt.Keywords: cloud computing, scheduling, real-time private cloud, bayesian
Procedia PDF Downloads 3596258 Tree Species Classification Using Effective Features of Polarimetric SAR and Hyperspectral Images
Authors: Milad Vahidi, Mahmod R. Sahebi, Mehrnoosh Omati, Reza Mohammadi
Abstract:
Forest management organizations need information to perform their work effectively. Remote sensing is an effective method to acquire information from the Earth. Two datasets of remote sensing images were used to classify forested regions. Firstly, all of extractable features from hyperspectral and PolSAR images were extracted. The optical features were spectral indexes related to the chemical, water contents, structural indexes, effective bands and absorption features. Also, PolSAR features were the original data, target decomposition components, and SAR discriminators features. Secondly, the particle swarm optimization (PSO) and the genetic algorithms (GA) were applied to select optimization features. Furthermore, the support vector machine (SVM) classifier was used to classify the image. The results showed that the combination of PSO and SVM had higher overall accuracy than the other cases. This combination provided overall accuracy about 90.56%. The effective features were the spectral index, the bands in shortwave infrared (SWIR) and the visible ranges and certain PolSAR features.Keywords: hyperspectral, PolSAR, feature selection, SVM
Procedia PDF Downloads 4166257 On the Analysis of Strategies of Buechi Games
Authors: Ahmad Termimi Ab Ghani, Kojiro Higuchi
Abstract:
In this paper, we present some results of simultaneous infinite games. We mainly work with generalized reachability games and Buechi games. These games are two-player concurrent games where each player chooses simultaneously their moves at each step. Our goal is to give simple expressions of values for each game. Moreover, we are interested in the question of what type of optimal (ε-optimal) strategy exists for both players depending on the type of games. We first show the determinacy (optimal value) and optimal (ε-optimal) strategies in generalized reachability games. We provide a simple expressions of value of this game and prove the existence of memoryless randomized ε-optimal strategy for Player I in any generalized reachability games. We then observe games with more complex objectives, games with Buechi objectives. We present how to compute an ε-optimal strategies and approximate a value of game in some way. Specifically, the results of generalized reachability games are used to show the value of Buechi games can be approximated as values of some generalized reachability games.Keywords: optimal Strategies, generalized reachability games, Buechi games
Procedia PDF Downloads 5936256 The Development of Community Leadership Strategies for Career Development of the Benjarong Pottery Products in Eight Upper Central Provinces
Authors: Thanaporn Chaimongkol
Abstract:
The objective of this research was aimed to examine the factors that influence the development of community leadership strategies to further develop the career regarding the Benjarong pottery products in eight upper central provinces, Thailand. The sample included (1) 1200 Benjarong pottery operators, (2) 30 involved representatives at both the policy level and support, and (3) OTOP network of 24 people. In this quantitative study, investigating data was conducted on individual session basis. The research instruments used included questionnaires and interview. The results showed that the components of the development of the community leadership strategies for career development of the Benjarong pottery products in eight upper central provinces were high overall; the Five Competitive Forces were of the highest average, followed by bargaining power of suppliers, and McKinsey 7's framework, respectively; where the highest average was strategy.Keywords: community leadership, strategy development, Benjarong Pottery, 8 upper central provinces
Procedia PDF Downloads 3256255 An Approach to Building a Recommendation Engine for Travel Applications Using Genetic Algorithms and Neural Networks
Authors: Adrian Ionita, Ana-Maria Ghimes
Abstract:
The lack of features, design and the lack of promoting an integrated booking application are some of the reasons why most online travel platforms only offer automation of old booking processes, being limited to the integration of a smaller number of services without addressing the user experience. This paper represents a practical study on how to improve travel applications creating user-profiles through data-mining based on neural networks and genetic algorithms. Choices made by users and their ‘friends’ in the ‘social’ network context can be considered input data for a recommendation engine. The purpose of using these algorithms and this design is to improve user experience and to deliver more features to the users. The paper aims to highlight a broader range of improvements that could be applied to travel applications in terms of design and service integration, while the main scientific approach remains the technical implementation of the neural network solution. The motivation of the technologies used is also related to the initiative of some online booking providers that have made the fact that they use some ‘neural network’ related designs public. These companies use similar Big-Data technologies to provide recommendations for hotels, restaurants, and cinemas with a neural network based recommendation engine for building a user ‘DNA profile’. This implementation of the ‘profile’ a collection of neural networks trained from previous user choices, can improve the usability and design of any type of application.Keywords: artificial intelligence, big data, cloud computing, DNA profile, genetic algorithms, machine learning, neural networks, optimization, recommendation system, user profiling
Procedia PDF Downloads 1636254 Forest Products Pricing System in Community Forestry Program: An Analysis of Its Impacts on Forest Resources Management and Livelihood Improvement of Local People
Authors: Mohan Bikram Thapa
Abstract:
Despite the successful implementation of community forestry program, a number of pros and cons have been raised on Terai community forestry in the case of lowland locally called Terai region of Nepal, which climatically belongs to tropical humid and possessed high-quality forests in terms of ecology and economy. The study aims to investigate the local pricing strategy of forest products and its impacts on equitable forest benefits sharing, the collection of community fund and carrying out livelihood improvement activities. The study was carried out on six community forests revealed that local people have substantially benefited from the community forests. However, being the region is heterogeneous by socio-economic conditions and forest resources have higher economic potential, the decision of low pricing strategy made by the local people have created inequality problems while sharing the forest benefits, and poorly contributed to community fund collection and consequently carrying out limited activities of livelihood improvement. The paper argued that the decision of low pricing strategy of forest products is counterproductive to promote the equitable benefit-sharing in the areas of heterogeneous socio-economic conditions with high-value forests. The low pricing strategy has been increasing accessibility of better off households at a higher rate than poor, as such households always have the higher affording capacity. It is also defective to increase the community fund and carry out activities of livelihood improvement effectively. The study concluded that unilateral decentralized forest policy and decision-making autonomy to the local people seems questionable unless their decision-making capacities are enriched sufficiently. Therefore, it is recommended that empowerments of decision-making capacity of local people and their respective institutions together with policy and program formulation are prerequisite for efficient and equitable community forest management and its long-term sustainability.Keywords: benefit sharing, community forest, livelihood, pricing mechanism, Nepal
Procedia PDF Downloads 3676253 Finite Element and Split Bregman Methods for Solving a Family of Optimal Control Problem with Partial Differential Equation Constraint
Authors: Mahmoud Lot
Abstract:
In this article, we will discuss the solution of elliptic optimal control problem. First, by using the nite element method, we obtain the discrete form of the problem. The obtained discrete problem is actually a large scale constrained optimization problem. Solving this optimization problem with traditional methods is difficult and requires a lot of CPU time and memory. But split Bergman method converts the constrained problem to an unconstrained, and hence it saves time and memory requirement. Then we use the split Bregman method for solving this problem, and examples show the speed and accuracy of split Bregman methods for solving these types of problems. We also use the SQP method for solving the examples and compare with the split Bregman method.Keywords: Split Bregman Method, optimal control with elliptic partial differential equation constraint, finite element method
Procedia PDF Downloads 1526252 The Role of Artificial Intelligence in Patent Claim Interpretation: Legal Challenges and Opportunities
Authors: Mandeep Saini
Abstract:
The rapid advancement of Artificial Intelligence (AI) is transforming various fields, including intellectual property law. This paper explores the emerging role of AI in interpreting patent claims, a critical and highly specialized area within intellectual property rights. Patent claims define the scope of legal protection granted to an invention, and their precise interpretation is crucial in determining the boundaries of the patent holder's rights. Traditionally, this interpretation has relied heavily on the expertise of patent examiners, legal professionals, and judges. However, the increasing complexity of modern inventions, especially in fields like biotechnology, software, and electronics, poses significant challenges to human interpretation. Introducing AI into patent claim interpretation raises several legal and ethical concerns. This paper addresses critical issues such as the reliability of AI-driven interpretations, the potential for algorithmic bias, and the lack of transparency in AI decision-making processes. It considers the legal implications of relying on AI, particularly regarding accountability for errors and the potential challenges to AI interpretations in court. The paper includes a comparative study of AI-driven patent claim interpretations versus human interpretations across different jurisdictions to provide a comprehensive analysis. This comparison highlights the variations in legal standards and practices, offering insights into how AI could impact the harmonization of international patent laws. The paper proposes policy recommendations for the responsible use of AI in patent law. It suggests legal frameworks that ensure AI tools complement, rather than replace, human expertise in patent claim interpretation. These recommendations aim to balance the benefits of AI with the need for maintaining trust, transparency, and fairness in the legal process. By addressing these critical issues, this research contributes to the ongoing discourse on integrating AI into the legal field, specifically within intellectual property rights. It provides a forward-looking perspective on how AI could reshape patent law, offering both opportunities for innovation and challenges that must be carefully managed to protect the integrity of the legal system.Keywords: artificial intelligence (ai), patent claim interpretation, intellectual property rights, algorithmic bias, natural language processing, patent law harmonization, legal ethics
Procedia PDF Downloads 216251 Quantitative Risk Analysis for Major Subsystems and Project Success of a Highthrouput Satellite
Authors: Ibrahim Isa Ali (Pantami), Abdu Jaafaru Bambale, Abimbola Alale, Danjuma Ibrahim Ndihgihdah, Muhammad Alkali, Adamu Idris Umar, Babadoko Dantala Mohammed, Moshood Kareem Olawole
Abstract:
This paper dwells on the risk management required for High throughput Satellite (HTS) project, and major subsystems that pertains to the improved performance and reliability of the spacecraft. The paper gives a clear picture of high‐throughput satellites (HTS) and the associated technologies with performances as they align and differ with the traditional geostationary orbit or Geosynchronous Equatorial Orbit (GEO) Communication Satellites. The paper also highlights critical subsystems and processes in project conceptualization and execution. The paper discusses the configuration of the payload. The need for optimization of resources for the HTS project and successful integration of critical subsystems for spacecraft requires implementation of risk analysis and mitigation from the preliminary design stage; Assembly, Integration and Test (AIT); Launch and in-orbit- Management stage.Keywords: AIT, HTS, in-orbit management, optimization
Procedia PDF Downloads 1036250 Experimental Design and Optimization of Diesel Oil Desulfurization Process by Adsorption Processes
Authors: M. Firoz Kalam, Wilfried Schuetz, Jan Hendrik Bredehoeft
Abstract:
Thiophene sulfur compounds' removal from diesel oil by batch adsorption process using commercial powdered activated carbon was designed and optimized in two-level factorial design method. This design analysis was used to find out the effects of operating parameters directing the adsorption process, such as amount of adsorbent, temperature and stirring time. The desulfurization efficiency was considered the response or output variable. Results showed that the stirring time had the largest effects on sulfur removal efficiency as compared with other operating parameters and their interactions under the experimental ranges studied. A regression model was generated to observe the closeness between predicted and experimental values. The three-dimensional plots and contour plots of main factors were generated according to the regression results to observe the optimal points.Keywords: activated carbon, adsorptive desulfurization, factorial design, process optimization
Procedia PDF Downloads 1626249 Voltage and Frequency Regulation Using the Third-Party Mid-Size Battery
Authors: Roghieh A. Biroon, Zoleikha Abdollahi
Abstract:
The recent growth of renewables, e.g., solar panels, batteries, and electric vehicles (EVs) in residential and small commercial sectors, has potential impacts on the stability and operation of power grids. Considering approximately 50 percent share of the residential and the commercial sectors in the electricity demand market, the significance of these impacts, and the necessity of addressing them are more highlighted. Utilities and power system operators should manage the renewable electricity sources integration with power systems in such a way to extract the most possible advantages for the power systems. The most common effect of high penetration level of the renewables is the reverse power flow in the distribution feeders when the customers generate more power than their needs. The reverse power flow causes voltage rise and thermal issues in the power grids. To overcome the voltage rise issues in the distribution system, several techniques have been proposed including reducing transformers short circuit resistance and feeder impedance, installing autotransformers/voltage regulators along the line, absorbing the reactive power by distributed generators (DGs), and limiting the PV and battery sizes. In this study, we consider a medium-scale battery energy storage to manage the power energy and address the aforementioned issues on voltage deviation and power loss increase. We propose an optimization algorithm to find the optimum size and location for the battery. The optimization for the battery location and size is so that the battery maintains the feeder voltage deviation and power loss at a certain desired level. Moreover, the proposed optimization algorithm controls the charging/discharging profile of the battery to absorb the negative power flow from residential and commercial customers in the feeder during the peak time and sell the power back to the system during the off-peak time. The proposed battery regulates the voltage problem in the distribution system while it also can play frequency regulation role in islanded microgrids. This battery can be regulated and controlled by the utilities or a third-party ancillary service provider for the utilities to reduce the power system loss and regulate the distribution feeder voltage and frequency in standard level.Keywords: ancillary services, battery, distribution system and optimization
Procedia PDF Downloads 1316248 Understanding Profit Shifting by Multinationals in the Context of Cross-Border M&A: A Methodological Exploration
Authors: Michal Friedrich
Abstract:
Cross-border investment has never been easier than in today’s global economy. Despite recent initiatives tightening the international tax landscape, profit shifting and tax optimization by multinational entities (MNEs) in the context of cross-border M&A remain persistent and complex phenomena that warrant in-depth exploration. By synthesizing the outcomes of existing research, this study aims to first provide a methodological framework for identifying MNEs’ profit-shifting behavior and quantifying its fiscal impacts via various macroeconomic and microeconomic approaches. The study also proposes additional methods and qualitative/quantitative measures for extracting insight into the profit shifting behavior of MNEs in the context of their M&A activities at industry and entity levels. To develop the proposed methods, this study applies the knowledge of international tax laws and known profit shifting conduits (incl. dividends, interest, and royalties) on several model cases/types of cross-border acquisitions and post-acquisition integration activities by MNEs and highlights important factors that encourage or discourage tax optimization. Follow-up research is envisaged to apply the methods outlined in this study on published data on real-world M&A transactions to gain practical country-by-country, industry and entity-level insights. In conclusion, this study seeks to contribute to the ongoing discourse on profit shifting by providing a methodological toolkit for exploring profit shifting tendencies MNEs in connection with their M&A activities and to serve as a backbone for further research. The study is expected to provide valuable insight to policymakers, tax authorities, and tax professionals alike.Keywords: BEPS, cross-border M&A, international taxation, profit shifting, tax optimization
Procedia PDF Downloads 696247 Value-Based Argumentation Frameworks and Judicial Moral Reasoning
Authors: Sonia Anand Knowlton
Abstract:
As the use of Artificial Intelligence is becoming increasingly integrated in virtually every area of life, the need and interest to logically formalize the law and judicial reasoning is growing tremendously. The study of argumentation frameworks (AFs) provides promise in this respect. AF’s provide a way of structuring human reasoning using a formal system of non-monotonic logic. P.M. Dung first introduced this framework and demonstrated that certain arguments must prevail and certain arguments must perish based on whether they are logically “attacked” by other arguments. Dung labelled the set of prevailing arguments as the “preferred extension” of the given argumentation framework. Trevor Bench-Capon’s Value-based Argumentation Frameworks extended Dung’s AF system by allowing arguments to derive their force from the promotion of “preferred” values. In VAF systems, the success of an attack from argument A to argument B (i.e., the triumph of argument A) requires that argument B does not promote a value that is preferred to argument A. There has been thorough discussion of the application of VAFs to the law within the computer science literature, mainly demonstrating that legal cases can be effectively mapped out using VAFs. This article analyses VAFs from a jurisprudential standpoint to provide a philosophical and theoretical analysis of what VAFs tell the legal community about the judicial reasoning, specifically distinguishing between legal and moral reasoning. It highlights the limitations of using VAFs to account for judicial moral reasoning in theory and in practice.Keywords: nonmonotonic logic, legal formalization, computer science, artificial intelligence, morality
Procedia PDF Downloads 746246 Evolution under Length Constraints for Convolutional Neural Networks Architecture Design
Authors: Ousmane Youme, Jean Marie Dembele, Eugene Ezin, Christophe Cambier
Abstract:
In recent years, the convolutional neural networks (CNN) architectures designed by evolution algorithms have proven to be competitive with handcrafted architectures designed by experts. However, these algorithms need a lot of computational power, which is beyond the capabilities of most researchers and engineers. To overcome this problem, we propose an evolution architecture under length constraints. It consists of two algorithms: a search length strategy to find an optimal space and a search architecture strategy based on a genetic algorithm to find the best individual in the optimal space. Our algorithms drastically reduce resource costs and also keep good performance. On the Cifar-10 dataset, our framework presents outstanding performance with an error rate of 5.12% and only 4.6 GPU a day to converge to the optimal individual -22 GPU a day less than the lowest cost automatic evolutionary algorithm in the peer competition.Keywords: CNN architecture, genetic algorithm, evolution algorithm, length constraints
Procedia PDF Downloads 1286245 Cluster Analysis and Benchmarking for Performance Optimization of a Pyrochlore Processing Unit
Authors: Ana C. R. P. Ferreira, Adriano H. P. Pereira
Abstract:
Given the frequent variation of mineral properties throughout the Araxá pyrochlore deposit, even if a good homogenization work has been carried out before feeding the processing plants, an operation with quality and performance’s high variety standard is expected. These results could be improved and standardized if the blend composition parameters that most influence the processing route are determined, and then the types of raw materials are grouped by them, finally presenting a great reference with operational settings for each group. Associating the physical and chemical parameters of a unit operation through benchmarking or even an optimal reference of metallurgical recovery and product quality reflects in the reduction of the production costs, optimization of the mineral resource, and guarantee of greater stability in the subsequent processes of the production chain that uses the mineral of interest. Conducting a comprehensive exploratory data analysis to identify which characteristics of the ore are most relevant to the process route, associated with the use of Machine Learning algorithms for grouping the raw material (ore) and associating these with reference variables in the process’ benchmark is a reasonable alternative for the standardization and improvement of mineral processing units. Clustering methods through Decision Tree and K-Means were employed, associated with algorithms based on the theory of benchmarking, with criteria defined by the process team in order to reference the best adjustments for processing the ore piles of each cluster. A clean user interface was created to obtain the outputs of the created algorithm. The results were measured through the average time of adjustment and stabilization of the process after a new pile of homogenized ore enters the plant, as well as the average time needed to achieve the best processing result. Direct gains from the metallurgical recovery of the process were also measured. The results were promising, with a reduction in the adjustment time and stabilization when starting the processing of a new ore pile, as well as reaching the benchmark. Also noteworthy are the gains in metallurgical recovery, which reflect a significant saving in ore consumption and a consequent reduction in production costs, hence a more rational use of the tailings dams and life optimization of the mineral deposit.Keywords: mineral clustering, machine learning, process optimization, pyrochlore processing
Procedia PDF Downloads 1436244 Case Analysis of Bamboo Based Social Enterprises in India-Improving Profitability and Sustainability
Authors: Priyal Motwani
Abstract:
The current market for bamboo products in India is about Rs. 21000 crores and is highly unorganised and fragmented. In this study, we have closely analysed the structure and functions of a major bamboo craft based organisation in Kerela, India and elaborated about its value chain, product mix, pricing strategy and supply chain, collaborations and competitive landscape. We have identified six major bottlenecks that are prevalent in such organisations, based on the Indian context, in relation to their product mix, asset management, and supply chain- corresponding waste management and retail network. The study has identified that the target customers for the bamboo based products and alternative revenue streams (eco-tourism, microenterprises, training), by carrying out secondary and primary research (5000 sample space), that can boost the existing revenue by 150%. We have then recommended an optimum product mix-covering premium, medium and low valued processing, for medium sized bamboo based organisations, in accordance with their capacity to maximize their revenue potential. After studying such organisations and their counter parts, the study has established an optimum retail network, considering B2B, B2C physical and online retail, to maximize their sales to their target groups. On the basis of the results obtained from the analysis of the future and present trends, our study gives recommendations to improve the revenue potential of bamboo based organisation in India and promote sustainability.Keywords: bamboo, bottlenecks, optimization, product mix, retail network, value chain
Procedia PDF Downloads 2176243 Approximation of Convex Set by Compactly Semidefinite Representable Set
Authors: Anusuya Ghosh, Vishnu Narayanan
Abstract:
The approximation of convex set by semidefinite representable set plays an important role in semidefinite programming, especially in modern convex optimization. To optimize a linear function over a convex set is a hard problem. But optimizing the linear function over the semidefinite representable set which approximates the convex set is easy to solve as there exists numerous efficient algorithms to solve semidefinite programming problems. So, our approximation technique is significant in optimization. We develop a technique to approximate any closed convex set, say K by compactly semidefinite representable set. Further we prove that there exists a sequence of compactly semidefinite representable sets which give tighter approximation of the closed convex set, K gradually. We discuss about the convergence of the sequence of compactly semidefinite representable sets to closed convex set K. The recession cone of K and the recession cone of the compactly semidefinite representable set are equal. So, we say that the sequence of compactly semidefinite representable sets converge strongly to the closed convex set. Thus, this approximation technique is very useful development in semidefinite programming.Keywords: semidefinite programming, semidefinite representable set, compactly semidefinite representable set, approximation
Procedia PDF Downloads 3876242 A Semi-supervised Classification Approach for Trend Following Investment Strategy
Authors: Rodrigo Arnaldo Scarpel
Abstract:
Trend following is a widely accepted investment strategy that adopts a rule-based trading mechanism that rather than striving to predict market direction or on information gathering to decide when to buy and when to sell a stock. Thus, in trend following one must respond to market’s movements that has recently happen and what is currently happening, rather than on what will happen. Optimally, in trend following strategy, is to catch a bull market at its early stage, ride the trend, and liquidate the position at the first evidence of the subsequent bear market. For applying the trend following strategy one needs to find the trend and identify trade signals. In order to avoid false signals, i.e., identify fluctuations of short, mid and long terms and to separate noise from real changes in the trend, most academic works rely on moving averages and other technical analysis indicators, such as the moving average convergence divergence (MACD) and the relative strength index (RSI) to uncover intelligible stock trading rules following trend following strategy philosophy. Recently, some works has applied machine learning techniques for trade rules discovery. In those works, the process of rule construction is based on evolutionary learning which aims to adapt the rules to the current environment and searches for the global optimum rules in the search space. In this work, instead of focusing on the usage of machine learning techniques for creating trading rules, a time series trend classification employing a semi-supervised approach was used to early identify both the beginning and the end of upward and downward trends. Such classification model can be employed to identify trade signals and the decision-making procedure is that if an up-trend (down-trend) is identified, a buy (sell) signal is generated. Semi-supervised learning is used for model training when only part of the data is labeled and Semi-supervised classification aims to train a classifier from both the labeled and unlabeled data, such that it is better than the supervised classifier trained only on the labeled data. For illustrating the proposed approach, it was employed daily trade information, including the open, high, low and closing values and volume from January 1, 2000 to December 31, 2022, of the São Paulo Exchange Composite index (IBOVESPA). Through this time period it was visually identified consistent changes in price, upwards or downwards, for assigning labels and leaving the rest of the days (when there is not a consistent change in price) unlabeled. For training the classification model, a pseudo-label semi-supervised learning strategy was used employing different technical analysis indicators. In this learning strategy, the core is to use unlabeled data to generate a pseudo-label for supervised training. For evaluating the achieved results, it was considered the annualized return and excess return, the Sortino and the Sharpe indicators. Through the evaluated time period, the obtained results were very consistent and can be considered promising for generating the intended trading signals.Keywords: evolutionary learning, semi-supervised classification, time series data, trading signals generation
Procedia PDF Downloads 896241 Integrating AI in Education: Enhancing Learning Processes and Personalization
Authors: Waleed Afandi
Abstract:
Artificial intelligence (AI) has rapidly transformed various sectors, including education. This paper explores the integration of AI in education, emphasizing its potential to revolutionize learning processes, enhance teaching methodologies, and personalize education. We examine the historical context of AI in education, current applications, and the potential challenges and ethical considerations associated with its implementation. By reviewing a wide range of literature, this study aims to provide a comprehensive understanding of how AI can be leveraged to improve educational outcomes and the future directions of AI-driven educational innovations. Additionally, the paper discusses the impact of AI on student engagement, teacher support, and administrative efficiency. Case studies highlighting successful AI applications in diverse educational settings are presented, showcasing the practical benefits and real-world implications. The analysis also addresses potential disparities in access to AI technologies and suggests strategies to ensure equitable implementation. Through a balanced examination of the promises and pitfalls of AI in education, this study seeks to inform educators, policymakers, and technologists about the optimal pathways for integrating AI to foster an inclusive, effective, and innovative educational environment.Keywords: artificial intelligence, education, personalized learning, teaching methodologies, educational outcomes, AI applications, student engagement, teacher support, administrative efficiency, equity in education
Procedia PDF Downloads 326240 A Strategy of Direct Power Control for PWM Rectifier Reducing Ripple in Instantaneous Power
Authors: T. Mohammed Chikouche, K. Hartani
Abstract:
Based on the analysis of basic direct torque control, a parallel master slave for four in-wheel permanent magnet synchronous motors (PMSM) fed by two three phase inverters used in electric vehicle is proposed in this paper. A conventional system with multi-inverter and multi-machine comprises a three phase inverter for each machine to be controlled. Another approach consists in using only one three-phase inverter to supply several permanent magnet synchronous machines. A modified direct torque control (DTC) algorithm is used for the control of the bi-machine traction system. Simulation results show that the proposed control strategy is well adapted for the synchronism of this system and provide good speed tracking performance.Keywords: electric vehicle, multi-machine single-inverter system, multi-machine multi-inverter control, in-wheel motor, master-slave control
Procedia PDF Downloads 2216239 Credit Card Fraud Detection with Ensemble Model: A Meta-Heuristic Approach
Authors: Gong Zhilin, Jing Yang, Jian Yin
Abstract:
The purpose of this paper is to develop a novel system for credit card fraud detection based on sequential modeling of data using hybrid deep learning models. The projected model encapsulates five major phases are pre-processing, imbalance-data handling, feature extraction, optimal feature selection, and fraud detection with an ensemble classifier. The collected raw data (input) is pre-processed to enhance the quality of the data through alleviation of the missing data, noisy data as well as null values. The pre-processed data are class imbalanced in nature, and therefore they are handled effectively with the K-means clustering-based SMOTE model. From the balanced class data, the most relevant features like improved Principal Component Analysis (PCA), statistical features (mean, median, standard deviation) and higher-order statistical features (skewness and kurtosis). Among the extracted features, the most optimal features are selected with the Self-improved Arithmetic Optimization Algorithm (SI-AOA). This SI-AOA model is the conceptual improvement of the standard Arithmetic Optimization Algorithm. The deep learning models like Long Short-Term Memory (LSTM), Convolutional Neural Network (CNN), and optimized Quantum Deep Neural Network (QDNN). The LSTM and CNN are trained with the extracted optimal features. The outcomes from LSTM and CNN will enter as input to optimized QDNN that provides the final detection outcome. Since the QDNN is the ultimate detector, its weight function is fine-tuned with the Self-improved Arithmetic Optimization Algorithm (SI-AOA).Keywords: credit card, data mining, fraud detection, money transactions
Procedia PDF Downloads 1316238 Short Answer Grading Using Multi-Context Features
Authors: S. Sharan Sundar, Nithish B. Moudhgalya, Nidhi Bhandari, Vineeth Vijayaraghavan
Abstract:
Automatic Short Answer Grading is one of the prime applications of artificial intelligence in education. Several approaches involving the utilization of selective handcrafted features, graphical matching techniques, concept identification and mapping, complex deep frameworks, sentence embeddings, etc. have been explored over the years. However, keeping in mind the real-world application of the task, these solutions present a slight overhead in terms of computations and resources in achieving high performances. In this work, a simple and effective solution making use of elemental features based on statistical, linguistic properties, and word-based similarity measures in conjunction with tree-based classifiers and regressors is proposed. The results for classification tasks show improvements ranging from 1%-30%, while the regression task shows a stark improvement of 35%. The authors attribute these improvements to the addition of multiple similarity scores to provide ensemble of scoring criteria to the models. The authors also believe the work could reinstate that classical natural language processing techniques and simple machine learning models can be used to achieve high results for short answer grading.Keywords: artificial intelligence, intelligent systems, natural language processing, text mining
Procedia PDF Downloads 1336237 Proposal of a Model Supporting Decision-Making Based on Multi-Objective Optimization Analysis on Information Security Risk Treatment
Authors: Ritsuko Kawasaki (Aiba), Takeshi Hiromatsu
Abstract:
Management is required to understand all information security risks within an organization, and to make decisions on which information security risks should be treated in what level by allocating how much amount of cost. However, such decision-making is not usually easy, because various measures for risk treatment must be selected with the suitable application levels. In addition, some measures may have objectives conflicting with each other. It also makes the selection difficult. Moreover, risks generally have trends and it also should be considered in risk treatment. Therefore, this paper provides the extension of the model proposed in the previous study. The original model supports the selection of measures by applying a combination of weighted average method and goal programming method for multi-objective analysis to find an optimal solution. The extended model includes the notion of weights to the risks, and the larger weight means the priority of the risk.Keywords: information security risk treatment, selection of risk measures, risk acceptance, multi-objective optimization
Procedia PDF Downloads 4616236 Investigation of the Effects of Visually Disabled and Typical Development Students on Their Multiple Intelligence by Applying Abacus and Right Brain Training
Authors: Sidika Di̇lşad Kaya, Ahmet Seli̇m Kaya, Ibrahi̇m Eri̇k, Havva Yaldiz, Yalçin Kaya
Abstract:
The aim of this study was to reveal the effects of right brain development on reading, comprehension, learning and concentration levels and rapid processing skills in students with low vision and students with standard development, and to explore the effects of right and left brain integration on students' academic success and the permanence of the learned knowledge. A total of 68 students with a mean age of 10.01±0.12 were included in the study, 58 of them with standard development, 9 partially visually impaired and 1 totally visually disabled student. The student with a total visual impairment could not participate in the reading speed test due to her total visual impairment. The following data were measured in the participant students before the project; Reading speed measurement in 1 minute, Reading comprehension questions, Burdon attention test, 50 questions of math quiz timed with a stopwatch. Participants were trained for 3 weeks, 5 days a week, for a total of two hours a day. In this study, right-brain developing exercises were carried out with the use of an abacus, and it was aimed to develop both mathematical and attention of students with questions prepared with numerical data taken from fairy tale activities. Among these problems, the study was supported with multiple-choice, 5W (what, where, who, why, when?), 1H (how?) questions along with true-false and fill-in-the-blank activities. By using memory cards, students' short-term memories were strengthened, photographic memory studies were conducted and their visual intelligence was supported. Auditory intelligence was supported by aiming to make calculations by using the abacus in the minds of the students with the numbers given aurally. When calculating the numbers by touching the real abacus, the development of students' tactile intelligence is enhanced. Research findings were analyzed in SPSS program, Kolmogorov Smirnov test was used for normality analysis. Since the variables did not show normal distribution, Wilcoxon test, one of the non-parametric tests, was used to compare the dependent groups. Statistical significance level was accepted as 0.05. The reading speed of the participants was 83.54±33.03 in the pre-test and 116.25±38.49 in the post-test. Narration pre-test 69.71±25.04 post-test 97.06±6.70; BURDON pretest 84.46±14.35 posttest 95.75±5.67; rapid math processing skills pretest 90.65±10.93, posttest 98.18±2.63 (P<0.05). It was determined that the pre-test and post-test averages of students with typical development and students with low vision were also significant for all four values (p<0.05). As a result of the data obtained from the participants, it is seen that the study was effective in terms of measurement parameters, and the findings were statistically significant. Therefore, it is recommended to use the method widely.Keywords: Abacus, reading speed, multiple intelligences, right brain training, visually impaired
Procedia PDF Downloads 1836235 Optimization of High Flux Density Design for Permanent Magnet Motor
Authors: Dong-Woo Kang
Abstract:
This paper presents an optimal magnet shape of a spoke-shaped interior permanent magnet synchronous motor by using ferrite magnets. Generally, the permanent magnet motor used the ferrite magnets has lower output power and efficiency than a rare-earth magnet motor, because the ferrite magnet has lower magnetic energy than the rare-earth magnet. Nevertheless, the ferrite magnet motor is used to many industrial products owing to cost effectiveness. In this paper, the authors propose a high power density design of the ferrite permanent magnet synchronous motor. Furthermore, because the motor design has to be taken a manufacturing process into account, the design is simulated by using the finite element method for analyzing the demagnetization, the magnetizing, and the structure stiffness. Especially, the magnet shape and dimensions are decided for satisfying these properties. Finally, the authors design an optimal motor for applying our system. That final design is manufactured and evaluated from experimentations.Keywords: demagnetization, design optimization, magnetic analysis, permanent magnet motors
Procedia PDF Downloads 3776234 Sparse Signal Restoration Algorithm Based on Piecewise Adaptive Backtracking Orthogonal Least Squares
Authors: Linyu Wang, Jiahui Ma, Jianhong Xiang, Hanyu Jiang
Abstract:
the traditional greedy compressed sensing algorithm needs to know the signal sparsity when recovering the signal, but the signal sparsity in the practical application can not be obtained as a priori information, and the recovery accuracy is low, which does not meet the needs of practical application. To solve this problem, this paper puts forward Piecewise adaptive backtracking orthogonal least squares algorithm. The algorithm is divided into two stages. In the first stage, the sparsity pre-estimation strategy is adopted, which can quickly approach the real sparsity and reduce time consumption. In the second stage iteration, the correction strategy and adaptive step size are used to accurately estimate the sparsity, and the backtracking idea is introduced to improve the accuracy of signal recovery. Through experimental simulation, the algorithm can accurately recover the estimated signal with fewer iterations when the sparsity is unknown.Keywords: compressed sensing, greedy algorithm, least square method, adaptive reconstruction
Procedia PDF Downloads 1486233 Event-Led Strategy for Cultural Tourism Development: The Case of Liverpool as the 2008 European Capital of Culture
Authors: Yi-De Liu
Abstract:
Cultural tourism is one of the largest and fastest growing global tourism markets and the cultures are increasingly being used to promote cities and to increase their competitiveness and attractiveness. One of the major forms of cultural tourism development undertaken throughout Europe has been the staging of a growing number of cultural events. The event of European Capitals of Culture (ECOC) is probably the best example of the new trends of cultural tourism in Europe, which is therefore used in this article to demonstrate some of the key issues surrounding the event-led strategy for cultural tourism development. Based on the experience of the 2008 ECOC Liverpool, UK, the study’s findings point to a number of ways in which the ECOC constitutes a boost for the development of cultural tourism in terms of realising experience economy, enhancing city image, facilitating urban regeneration, promoting cultural production and consumption, as well as establishing partnerships. This study is concluded by drawing some critical factors that event and tourism organisers should consider.Keywords: cultural tourism, event tourism, cultural event, European capital of culture, Liverpool
Procedia PDF Downloads 7056232 Central Energy Management for Optimizing Utility Grid Power Exchange with a Network of Smart Homes
Authors: Sima Aznavi, Poria Fajri, Hanif Livani
Abstract:
Smart homes are small energy systems which may be equipped with renewable energy sources, storage devices, and loads. Energy management strategy plays a main role in the efficient operation of smart homes. Effective energy scheduling of the renewable energy sources and storage devices guarantees efficient energy management in households while reducing the energy imports from the grid. Nevertheless, despite such strategies, independently day ahead energy schedules for multiple households can cause undesired effects such as high power exchange with the grid at certain times of the day. Therefore, the interactions between multiple smart home day ahead energy projections is a challenging issue in a smart grid system and if not managed appropriately, the imported energy from the power network can impose additional burden on the distribution grid. In this paper, a central energy management strategy for a network consisting of multiple households each equipped with renewable energy sources, storage devices, and Plug-in Electric Vehicles (PEV) is proposed. The decision-making strategy alongside the smart home energy management system, minimizes the energy purchase cost of the end users, while at the same time reducing the stress on the utility grid. In this approach, the smart home energy management system determines different operating scenarios based on the forecasted household daily load and the components connected to the household with the objective of minimizing the end user overall cost. Then, selected projections for each household that are within the same cost range are sent to the central decision-making system. The central controller then organizes the schedules to reduce the overall peak to average ratio of the total imported energy from the grid. To validate this approach simulations are carried out for a network of five smart homes with different load requirements and the results confirm that by applying the proposed central energy management strategy, the overall power demand from the grid can be significantly flattened. This is an effective approach to alleviate the stress on the network by distributing its energy to a network of multiple households over a 24- hour period.Keywords: energy management, renewable energy sources, smart grid, smart home
Procedia PDF Downloads 248