Search results for: precise time domain expanding algorithm
22035 A New Intelligent, Dynamic and Real Time Management System of Sewerage
Authors: R. Tlili Yaakoubi, H.Nakouri, O. Blanpain, S. Lallahem
Abstract:
The current tools for real time management of sewer systems are based on two software tools: the software of weather forecast and the software of hydraulic simulation. The use of the first ones is an important cause of imprecision and uncertainty, the use of the second requires temporal important steps of decision because of their need in times of calculation. This way of proceeding fact that the obtained results are generally different from those waited. The major idea of this project is to change the basic paradigm by approaching the problem by the "automatic" face rather than by that "hydrology". The objective is to make possible the realization of a large number of simulations at very short times (a few seconds) allowing to take place weather forecasts by using directly the real time meditative pluviometric data. The aim is to reach a system where the decision-making is realized from reliable data and where the correction of the error is permanent. A first model of control laws was realized and tested with different return-period rainfalls. The gains obtained in rejecting volume vary from 19 to 100 %. The development of a new algorithm was then used to optimize calculation time and thus to overcome the subsequent combinatorial problem in our first approach. Finally, this new algorithm was tested with 16- year-rainfall series. The obtained gains are 40 % of total volume rejected to the natural environment and of 65 % in the number of discharges.Keywords: automation, optimization, paradigm, RTC
Procedia PDF Downloads 30122034 A Genetic Algorithm Based Permutation and Non-Permutation Scheduling Heuristics for Finite Capacity Material Requirement Planning Problem
Authors: Watchara Songserm, Teeradej Wuttipornpun
Abstract:
This paper presents a genetic algorithm based permutation and non-permutation scheduling heuristics (GAPNP) to solve a multi-stage finite capacity material requirement planning (FCMRP) problem in automotive assembly flow shop with unrelated parallel machines. In the algorithm, the sequences of orders are iteratively improved by the GA characteristics, whereas the required operations are scheduled based on the presented permutation and non-permutation heuristics. Finally, a linear programming is applied to minimize the total cost. The presented GAPNP algorithm is evaluated by using real datasets from automotive companies. The required parameters for GAPNP are intently tuned to obtain a common parameter setting for all case studies. The results show that GAPNP significantly outperforms the benchmark algorithm about 30% on average.Keywords: capacitated MRP, genetic algorithm, linear programming, automotive industries, flow shop, application in industry
Procedia PDF Downloads 49022033 A BERT-Based Model for Financial Social Media Sentiment Analysis
Authors: Josiel Delgadillo, Johnson Kinyua, Charles Mutigwe
Abstract:
The purpose of sentiment analysis is to determine the sentiment strength (e.g., positive, negative, neutral) from a textual source for good decision-making. Natural language processing in domains such as financial markets requires knowledge of domain ontology, and pre-trained language models, such as BERT, have made significant breakthroughs in various NLP tasks by training on large-scale un-labeled generic corpora such as Wikipedia. However, sentiment analysis is a strong domain-dependent task. The rapid growth of social media has given users a platform to share their experiences and views about products, services, and processes, including financial markets. StockTwits and Twitter are social networks that allow the public to express their sentiments in real time. Hence, leveraging the success of unsupervised pre-training and a large amount of financial text available on social media platforms could potentially benefit a wide range of financial applications. This work is focused on sentiment analysis using social media text on platforms such as StockTwits and Twitter. To meet this need, SkyBERT, a domain-specific language model pre-trained and fine-tuned on financial corpora, has been developed. The results show that SkyBERT outperforms current state-of-the-art models in financial sentiment analysis. Extensive experimental results demonstrate the effectiveness and robustness of SkyBERT.Keywords: BERT, financial markets, Twitter, sentiment analysis
Procedia PDF Downloads 15422032 Hyperspectral Image Classification Using Tree Search Algorithm
Authors: Shreya Pare, Parvin Akhter
Abstract:
Remotely sensing image classification becomes a very challenging task owing to the high dimensionality of hyperspectral images. The pixel-wise classification methods fail to take the spatial structure information of an image. Therefore, to improve the performance of classification, spatial information can be integrated into the classification process. In this paper, the multilevel thresholding algorithm based on a modified fuzzy entropy function is used to perform the segmentation of hyperspectral images. The fuzzy parameters of the MFE function have been optimized by using a new meta-heuristic algorithm based on the Tree-Search algorithm. The segmented image is classified by a large distribution machine (LDM) classifier. Experimental results are shown on a hyperspectral image dataset. The experimental outputs indicate that the proposed technique (MFE-TSA-LDM) achieves much higher classification accuracy for hyperspectral images when compared to state-of-art classification techniques. The proposed algorithm provides accurate segmentation and classification maps, thus becoming more suitable for image classification with large spatial structures.Keywords: classification, hyperspectral images, large distribution margin, modified fuzzy entropy function, multilevel thresholding, tree search algorithm, hyperspectral image classification using tree search algorithm
Procedia PDF Downloads 18022031 An Enhanced Hybrid Backoff Technique for Minimizing the Occurrence of Collision in Mobile Ad Hoc Networks
Authors: N. Sabiyath Fatima, R. K. Shanmugasundaram
Abstract:
In Mobile Ad-hoc Networks (MANETS), every node performs both as transmitter and receiver. The existing backoff models do not exactly forecast the performance of the wireless network. Also, the existing models experience elevated packet collisions. Every time a collision happens, the station’s contention window (CW) is doubled till it arrives at the utmost value. The main objective of this paper is to diminish collision by means of contention window Multiplicative Increase Decrease Backoff (CWMIDB) scheme. The intention of rising CW is to shrink the collision possibility by distributing the traffic into an outsized point in time. Within wireless Ad hoc networks, the CWMIDB algorithm dynamically controls the contention window of the nodes experiencing collisions. During packet communication, the backoff counter is evenly selected from the given choice of [0, CW-1]. At this point, CW is recognized as contention window and its significance lies on the amount of unsuccessful transmission that had happened for the packet. On the initial transmission endeavour, CW is put to least amount value (C min), if transmission effort fails, subsequently the value gets doubled, and once more the value is set to least amount on victorious broadcast. CWMIDB is simulated inside NS2 environment and its performance is compared with Binary Exponential Backoff Algorithm. The simulation results show improvement in transmission probability compared to that of the existing backoff algorithm.Keywords: backoff, contention window, CWMIDB, MANET
Procedia PDF Downloads 27922030 Recent Developments in the Application of Deep Learning to Stock Market Prediction
Authors: Shraddha Jain Sharma, Ratnalata Gupta
Abstract:
Predicting stock movements in the financial market is both difficult and rewarding. Analysts and academics are increasingly using advanced approaches such as machine learning techniques to anticipate stock price patterns, thanks to the expanding capacity of computing and the recent advent of graphics processing units and tensor processing units. Stock market prediction is a type of time series prediction that is incredibly difficult to do since stock prices are influenced by a variety of financial, socioeconomic, and political factors. Furthermore, even minor mistakes in stock market price forecasts can result in significant losses for companies that employ the findings of stock market price prediction for financial analysis and investment. Soft computing techniques are increasingly being employed for stock market prediction due to their better accuracy than traditional statistical methodologies. The proposed research looks at the need for soft computing techniques in stock market prediction, the numerous soft computing approaches that are important to the field, past work in the area with their prominent features, and the significant problems or issue domain that the area involves. For constructing a predictive model, the major focus is on neural networks and fuzzy logic. The stock market is extremely unpredictable, and it is unquestionably tough to correctly predict based on certain characteristics. This study provides a complete overview of the numerous strategies investigated for high accuracy prediction, with a focus on the most important characteristics.Keywords: stock market prediction, artificial intelligence, artificial neural networks, fuzzy logic, accuracy, deep learning, machine learning, stock price, trading volume
Procedia PDF Downloads 9222029 Inverse Cauchy Problem of Doubly Connected Domains via Spectral Meshless Radial Point Interpolation
Authors: Elyas Shivanian
Abstract:
In this paper, the spectral meshless radial point interpolation (SMRPI) technique is applied to the Cauchy problems of two-dimensional elliptic PDEs in doubly connected domains. It is obtained the unknown data on the inner boundary of the domain while overspecified boundary data are imposed on the outer boundary of the domain by using the SMRPI. Shape functions, which are constructed through point interpolation method using the radial basis functions, help us to treat problem locally with the aim of high order convergence rate. In this way, localization in SMRPI can reduce the ill-conditioning for Cauchy problem. Furthermore, we improve previous results and it is revealed the SMRPI is more accurate and stable by adding strong perturbations.Keywords: cauchy problem, doubly connected domain, radial basis function, shape function
Procedia PDF Downloads 27822028 Consumer Load Profile Determination with Entropy-Based K-Means Algorithm
Authors: Ioannis P. Panapakidis, Marios N. Moschakis
Abstract:
With the continuous increment of smart meter installations across the globe, the need for processing of the load data is evident. Clustering-based load profiling is built upon the utilization of unsupervised machine learning tools for the purpose of formulating the typical load curves or load profiles. The most commonly used algorithm in the load profiling literature is the K-means. While the algorithm has been successfully tested in a variety of applications, its drawback is the strong dependence in the initialization phase. This paper proposes a novel modified form of the K-means that addresses the aforementioned problem. Simulation results indicate the superiority of the proposed algorithm compared to the K-means.Keywords: clustering, load profiling, load modeling, machine learning, energy efficiency and quality
Procedia PDF Downloads 16522027 A Supervised Goal Directed Algorithm in Economical Choice Behaviour: An Actor-Critic Approach
Authors: Keyvanl Yahya
Abstract:
This paper aims to find a algorithmic structure that affords to predict and explain economic choice behaviour particularly under uncertainty (random policies) by manipulating the prevalent Actor-Critic learning method that complies with the requirements we have been entrusted ever since the field of neuroeconomics dawned on us. Whilst skimming some basics of neuroeconomics that might be relevant to our discussion, we will try to outline some of the important works which have so far been done to simulate choice making processes. Concerning neurological findings that suggest the existence of two specific functions that are executed through Basal Ganglia all the way down to sub-cortical areas, namely 'rewards' and 'beliefs', we will offer a modified version of actor/critic algorithm to shed a light on the relation between these functions and most importantly resolve what is referred to as a challenge for actor-critic algorithms, that is lack of inheritance or hierarchy which avoids the system being evolved in continuous time tasks whence the convergence might not emerge.Keywords: neuroeconomics, choice behaviour, decision making, reinforcement learning, actor-critic algorithm
Procedia PDF Downloads 39722026 Improving Load Frequency Control of Multi-Area Power System by Considering Uncertainty by Using Optimized Type 2 Fuzzy Pid Controller with the Harmony Search Algorithm
Authors: Mehrdad Mahmudizad, Roya Ahmadi Ahangar
Abstract:
This paper presents the method of designing the type 2 fuzzy PID controllers in order to solve the problem of Load Frequency Control (LFC). The Harmony Search (HS) algorithm is used to regulate the measurement factors and the effect of uncertainty of membership functions of Interval Type 2 Fuzzy Proportional Integral Differential (IT2FPID) controllers in order to reduce the frequency deviation resulted from the load oscillations. The simulation results implicitly show that the performance of the proposed IT2FPID LFC in terms of error, settling time and resistance against different load oscillations is more appropriate and preferred than PID and Type 1 Fuzzy Proportional Integral Differential (T1FPID) controllers.Keywords: load frequency control, fuzzy-pid controller, type 2 fuzzy system, harmony search algorithm
Procedia PDF Downloads 27922025 Motor Control Recovery Minigame
Authors: Taha Enes Kon, Vanshika Reddy
Abstract:
This project focuses on developing a gamified mobile application to aid in stroke rehabilitation by enhancing motor skills through interactive activities. The primary goal was to design a companion app for a passive haptic rehab glove, incorporating Google MediaPipe for gesture tracking and vibrotactile feedback. The app simulates farming activities, offering a fun and engaging experience while addressing the monotony of traditional rehabilitation methods. The prototype focuses on a single minigame, Flower Picking, which uses gesture recognition to interact with virtual elements, encouraging users to perform exercises that improve hand dexterity. The development process involved creating accessible and user-centered designs using Figma, integrating gesture recognition algorithms, and implementing unity-based game mechanics. Real-time feedback and progressive difficulty levels ensured a personalized experience, motivating users to adhere to rehabilitation routines. The prototype achieved a gesture detection precision of 90%, effectively recognizing predefined gestures such as the Fist and OK symbols. Quantitative analysis highlighted a 40% increase in average session duration compared to traditional exercises, while qualitative feedback praised the app’s immersive design and ease of use. Despite its success, challenges included rigidity in gesture recognition, requiring precise hand orientations, and limited gesture support. Future improvements include expanding gesture adaptability and incorporating additional minigames to target a broader range of exercises. The project demonstrates the potential of gamification in stroke rehabilitation, offering a scalable and accessible solution that complements clinical treatments, making recovery engaging and effective for users.Keywords: stroke rehabilitation, haptic feedback, gamification, MediaPipe, motor control
Procedia PDF Downloads 722024 Forward Stable Computation of Roots of Real Polynomials with Only Real Distinct Roots
Authors: Nevena Jakovčević Stor, Ivan Slapničar
Abstract:
Any polynomial can be expressed as a characteristic polynomial of a complex symmetric arrowhead matrix. This expression is not unique. If the polynomial is real with only real distinct roots, the matrix can be chosen as real. By using accurate forward stable algorithm for computing eigen values of real symmetric arrowhead matrices we derive a forward stable algorithm for computation of roots of such polynomials in O(n^2 ) operations. The algorithm computes each root to almost full accuracy. In some cases, the algorithm invokes extended precision routines, but only in the non-iterative part. Our examples include numerically difficult problems, like the well-known Wilkinson’s polynomials. Our algorithm compares favorably to other method for polynomial root-finding, like MPSolve or Newton’s method.Keywords: roots of polynomials, eigenvalue decomposition, arrowhead matrix, high relative accuracy
Procedia PDF Downloads 41822023 Vector Quantization Based on Vector Difference Scheme for Image Enhancement
Authors: Biji Jacob
Abstract:
Vector quantization algorithm which uses minimum distance calculation for codebook generation, a time consuming calculation performed on each pixel values leads to computation complexity. The codebook is updated by comparing the distance of each vector to their centroid vector and measure for their closeness. In this paper vector quantization is modified based on vector difference algorithm for image enhancement purpose. In the proposed scheme, vector differences between the vectors are considered as the new generation vectors or new codebook vectors. The codebook is updated by comparing the new generation vector with a threshold value having minimum error with the parent vector. The minimum error decides the fitness of each newly generated vector. Thus the codebook is generated in an adaptive manner and the fitness value is determined for the suppression of the degraded portion of the image and thereby leads to the enhancement of the image through the adaptive searching capability of the vector quantization through vector difference algorithm. Experimental results shows that the vector difference scheme efficiently modifies the vector quantization algorithm for enhancing the image with peak signal to noise ratio (PSNR), mean square error (MSE), Euclidean distance (E_dist) as the performance parameters.Keywords: codebook, image enhancement, vector difference, vector quantization
Procedia PDF Downloads 26822022 An Approach for Association Rules Ranking
Authors: Rihab Idoudi, Karim Saheb Ettabaa, Basel Solaiman, Kamel Hamrouni
Abstract:
Medical association rules induction is used to discover useful correlations between pertinent concepts from large medical databases. Nevertheless, ARs algorithms produce huge amount of delivered rules and do not guarantee the usefulness and interestingness of the generated knowledge. To overcome this drawback, we propose an ontology based interestingness measure for ARs ranking. According to domain expert, the goal of the use of ARs is to discover implicit relationships between items of different categories such as ‘clinical features and disorders’, ‘clinical features and radiological observations’, etc. That’s to say, the itemsets which are composed of ‘similar’ items are uninteresting. Therefore, the dissimilarity between the rule’s items can be used to judge the interestingness of association rules; the more different are the items, the more interesting the rule is. In this paper, we design a distinct approach for ranking semantically interesting association rules involving the use of an ontology knowledge mining approach. The basic idea is to organize the ontology’s concepts into a hierarchical structure of conceptual clusters of targeted subjects, where each cluster encapsulates ‘similar’ concepts suggesting a specific category of the domain knowledge. The interestingness of association rules is, then, defined as the dissimilarity between corresponding clusters. That is to say, the further are the clusters of the items in the AR, the more interesting the rule is. We apply the method in our domain of interest – mammographic domain- using an existing mammographic ontology called Mammo with the goal of deriving interesting rules from past experiences, to discover implicit relationships between concepts modeling the domain.Keywords: association rule, conceptual clusters, interestingness measures, ontology knowledge mining, ranking
Procedia PDF Downloads 32222021 Ontologies for Social Media Digital Evidence
Authors: Edlira Kalemi, Sule Yildirim-Yayilgan
Abstract:
Online Social Networks (OSNs) are nowadays being used widely and intensively for crime investigation and prevention activities. As they provide a lot of information they are used by the law enforcement and intelligence. An extensive review on existing solutions and models for collecting intelligence from this source of information and making use of it for solving crimes has been presented in this article. The main focus is on smart solutions and models where ontologies have been used as the main approach for representing criminal domain knowledge. A framework for a prototype ontology named SC-Ont will be described. This defines terms of the criminal domain ontology and the relations between them. The terms and the relations are extracted during both this review and the discussions carried out with domain experts. The development of SC-Ont is still ongoing work, where in this paper, we report mainly on the motivation for using smart ontology models and the possible benefits of using them for solving crimes.Keywords: criminal digital evidence, social media, ontologies, reasoning
Procedia PDF Downloads 39022020 Characterization of the Catalytic and Structural Roles of the Human Hexokinase 2 in Cancer Progression
Authors: Mir Hussain Nawaz, Lyudmila Nedyalkova, Haizhong Zhu, Wael M. Rabeh
Abstract:
In this study, we aim to biochemically and structurally characterize the interactions of human HK2 with the mitochondria in addition to the role of its N-terminal domain in catalysis and stability of the full-length enzyme. Here, we solved the crystal structure of human HK2 in complex with glucose and glucose-6-phosphate (PDB code: 2NZT), where it is a homodimer with catalytically active N- and C-terminal domains linked by a seven-turn α-helix. Different from the inactive N-terminal domains of isozymes 1 and 3, the N- domain of HK2 not only capable to catalyze a reaction but it is responsible for the thermodynamic stabilizes of the full-length enzyme. Deletion of first α-helix of the N-domain that binds to the mitochondria altered the stability and catalytic activity of the full-length HK2. In addition, we found the linker helix between the N- and C-terminal domains to play an important role in controlling the catalytic activity of the N-terminal domain. HK2 is a major step in the regulation of glucose metabolism in cancer making it an ideal target for the development of new anticancer therapeutics. Characterizing the structural and molecular mechanisms of human HK2 and its role in cancer metabolism will accelerate the design and development of new cancer therapeutics that are safe and cancer specific.Keywords: cancer metabolism, enzymology, drug discovery, protein stability
Procedia PDF Downloads 26522019 Structure Domains Tuning Magnetic Anisotropy and Motivating Novel Electric Behaviors in LaCoO₃ Films
Authors: Dechao Meng, Yongqi Dong, Qiyuan Feng, Zhangzhang Cui, Xiang Hu, Haoliang Huang, Genhao Liang, Huanhua Wang, Hua Zhou, Hawoong Hong, Jinghua Guo, Qingyou Lu, Xiaofang Zhai, Yalin Lu
Abstract:
Great efforts have been taken to reveal the intrinsic origins of emerging ferromagnetism (FM) in strained LaCoO₃ (LCO) films. However, some macro magnetic performances of LCO are still not well understood and even controversial, such as magnetic anisotropy. Determining and understanding magnetic anisotropy might help to find the true causes of FM in turn. Perpendicular magnetic anisotropy (PMA) was the first time to be directly observed in high-quality LCO films with different thickness. The in-plane (IP) and out of plane (OOP) remnant magnetic moment ratio of 30 unit cell (u.c.) films is as large as 20. The easy axis lays in the OOP direction with an IP/OOP coercive field ratio of 10. What's more, the PMA could be simply tuned by changing the thickness. With the thickness increases, the IP/OOP magnetic moment ratio remarkably decrease with magnetic easy axis changing from OOP to IP. Such a huge and tunable PMA performance exhibit strong potentials in fundamental researches or applications. What causes PMA is the first concern. More OOP orbitals occupation may be one of the micro reasons of PMA. A cluster-like magnetic domain pattern was found in 30 u.c. with no obvious color contrasts, similar to that of LaAlO₃/SrTiO₃ films. And the nanosize domains could not be totally switched even at a large OOP magnetic field of 23 T. It indicates strong IP characters or none OOP magnetism of some clusters. The IP magnetic domains might influence the magnetic performance and help to form PMA. Meanwhile some possible nonmagnetic clusters might be the reason why the measured moments of LCO films are smaller than the calculated values 2 μB/Co, one of the biggest confusions in LCO films.What tunes PMA seems much more interesting. Totally different magnetic domain patterns were found in 180 u.c. films with cluster magnetic domains surrounded by < 110 > cross-hatch lines. These lines were regarded as structure domain walls (DWs) determined by 3D reciprocal space mapping (RSM). Two groups of in-plane features with fourfold symmetry were observed near the film diffraction peaks in (002) 3D-RSM. One is along < 110 > directions with a larger intensity, which is well match the lines on the surfaces. The other is much weaker and along < 100 > directions, which is from the normal lattice titling of films deposited on cubic substrates. The < 110 > domain features obtained from (103) and (113) 3D-RSMs exhibit similar evolution of the DWs percentages and magnetic behavior. Structure domains and domain walls are believed to tune PMA performances by transform more IP magnetic moments to OOP. Last but not the least, thick films with lots of structure domains exhibit different electrical transport behaviors. A metal-to-insulator transition (MIT) and an angular dependent negative magnetic resistivity were observed near 150 K, higher than FM transition temperature but similar to that of spin-orbital coupling related 1/4 order diffraction peaks.Keywords: structure domain, magnetic anisotropy, magnetic domain, domain wall, 3D-RSM, strain
Procedia PDF Downloads 15322018 M-Machine Assembly Scheduling Problem to Minimize Total Tardiness with Non-Zero Setup Times
Authors: Harun Aydilek, Asiye Aydilek, Ali Allahverdi
Abstract:
Our objective is to minimize the total tardiness in an m-machine two-stage assembly flowshop scheduling problem. The objective is an important performance measure because of the fact that the fulfillment of due dates of customers has to be taken into account while making scheduling decisions. In the literature, the problem is considered with zero setup times which may not be realistic and appropriate for some scheduling environments. Considering separate setup times from processing times increases machine utilization by decreasing the idle time and reduces total tardiness. We propose two new algorithms and adapt four existing algorithms in the literature which are different versions of simulated annealing and genetic algorithms. Moreover, a dominance relation is developed based on the mathematical formulation of the problem. The developed dominance relation is incorporated in our proposed algorithms. Computational experiments are conducted to investigate the performance of the newly proposed algorithms. We find that one of the proposed algorithms performs significantly better than the others, i.e., the error of the best algorithm is less than those of the other algorithms by minimum 50%. The newly proposed algorithm is also efficient for the case of zero setup times and performs better than the best existing algorithm in the literature.Keywords: algorithm, assembly flowshop, scheduling, simulation, total tardiness
Procedia PDF Downloads 33222017 A New Method Presentation for Locating Fault in Power Distribution Feeders Considering DG
Authors: Rahman Dashti, Ehsan Gord
Abstract:
In this paper, an improved impedance based fault location method is proposed. In this method, online fault locating is performed using voltage and current information at the beginning of the feeder. Determining precise fault location in a short time increases reliability and efficiency of the system. The proposed method utilizes information about main component of voltage and current at the beginning of the feeder and distributed generation unit (DGU) in order to precisely locate different faults in acceptable time. To evaluate precision and accuracy of the proposed method, a 13-node is simulated and tested using MATLAB.Keywords: distribution network, fault section determination, distributed generation units, distribution protection equipment
Procedia PDF Downloads 40422016 Artificial Intelligence in Bioscience: The Next Frontier
Authors: Parthiban Srinivasan
Abstract:
With recent advances in computational power and access to enough data in biosciences, artificial intelligence methods are increasingly being used in drug discovery research. These methods are essentially a series of advanced statistics based exercises that review the past to indicate the likely future. Our goal is to develop a model that accurately predicts biological activity and toxicity parameters for novel compounds. We have compiled a robust library of over 150,000 chemical compounds with different pharmacological properties from literature and public domain databases. The compounds are stored in simplified molecular-input line-entry system (SMILES), a commonly used text encoding for organic molecules. We utilize an automated process to generate an array of numerical descriptors (features) for each molecule. Redundant and irrelevant descriptors are eliminated iteratively. Our prediction engine is based on a portfolio of machine learning algorithms. We found Random Forest algorithm to be a better choice for this analysis. We captured non-linear relationship in the data and formed a prediction model with reasonable accuracy by averaging across a large number of randomized decision trees. Our next step is to apply deep neural network (DNN) algorithm to predict the biological activity and toxicity properties. We expect the DNN algorithm to give better results and improve the accuracy of the prediction. This presentation will review all these prominent machine learning and deep learning methods, our implementation protocols and discuss these techniques for their usefulness in biomedical and health informatics.Keywords: deep learning, drug discovery, health informatics, machine learning, toxicity prediction
Procedia PDF Downloads 36022015 Capacity Optimization in Cooperative Cognitive Radio Networks
Authors: Mahdi Pirmoradian, Olayinka Adigun, Christos Politis
Abstract:
Cooperative spectrum sensing is a crucial challenge in cognitive radio networks. Cooperative sensing can increase the reliability of spectrum hole detection, optimize sensing time and reduce delay in cooperative networks. In this paper, an efficient central capacity optimization algorithm is proposed to minimize cooperative sensing time in a homogenous sensor network using OR decision rule subject to the detection and false alarm probabilities constraints. The evaluation results reveal significant improvement in the sensing time and normalized capacity of the cognitive sensors.Keywords: cooperative networks, normalized capacity, sensing time
Procedia PDF Downloads 63622014 Mean Shift-Based Preprocessing Methodology for Improved 3D Buildings Reconstruction
Authors: Nikolaos Vassilas, Theocharis Tsenoglou, Djamchid Ghazanfarpour
Abstract:
In this work we explore the capability of the mean shift algorithm as a powerful preprocessing tool for improving the quality of spatial data, acquired from airborne scanners, from densely built urban areas. On one hand, high resolution image data corrupted by noise caused by lossy compression techniques are appropriately smoothed while at the same time preserving the optical edges and, on the other, low resolution LiDAR data in the form of normalized Digital Surface Map (nDSM) is upsampled through the joint mean shift algorithm. Experiments on both the edge-preserving smoothing and upsampling capabilities using synthetic RGB-z data show that the mean shift algorithm is superior to bilateral filtering as well as to other classical smoothing and upsampling algorithms. Application of the proposed methodology for 3D reconstruction of buildings of a pilot region of Athens, Greece results in a significant visual improvement of the 3D building block model.Keywords: 3D buildings reconstruction, data fusion, data upsampling, mean shift
Procedia PDF Downloads 31622013 Cognitive eTransformation Framework for Education Sector
Authors: A. Hol
Abstract:
21st century brought waves of business and industry eTransformations. The impact of change is also being seen in education. To identify the extent of this, scenario analysis methodology was utilised with the aim to assess business transformations across industry sectors ranging from craftsmanship, medicine, finance and manufacture to innovations and adoptions of new technologies and business models. Firstly, scenarios were drafted based on the current eTransformation models and its dimensions. Following this, eTransformation framework was utilised with the aim to derive the key eTransformation parameters, the essential characteristics that have enabled eTransformations across the sectors. Following this, identified key parameters were mapped to the transforming domain-education. The mapping assisted in deriving a cognitive eTransformation framework for education sector. The framework highlights the importance of context and the notion that education today needs not only to deliver content to students but it also needs to be able to meet the dynamically changing demands of specific student and industry groups. Furthermore, it pinpoints that for such processes to be supported, specific technology is required, so that instant, on demand and periodic feedback as well as flexible, dynamically expanding study content can be sought and received via multiple education mediums.Keywords: education sector, business transformation, eTransformation model, cognitive model, cognitive systems, eTransformation
Procedia PDF Downloads 13722012 Genetic Algorithm Optimization of Microcantilever Based Resonator
Authors: Manjula Sutagundar, B. G. Sheeparamatti, D. S. Jangamshetti
Abstract:
Micro Electro Mechanical Systems (MEMS) resonators have shown the potential of replacing quartz crystal technology for sensing and high frequency signal processing applications because of inherent advantages like small size, high quality factor, low cost, compatibility with integrated circuit chips. This paper presents the optimization and modelling and simulation of the optimized micro cantilever resonator. The objective of the work is to optimize the dimensions of a micro cantilever resonator for a specified range of resonant frequency and specific quality factor. Optimization is carried out using genetic algorithm. The genetic algorithm is implemented using MATLAB. The micro cantilever resonator is modelled in CoventorWare using the optimized dimensions obtained from genetic algorithm. The modeled cantilever is analysed for resonance frequency.Keywords: MEMS resonator, genetic algorithm, modelling and simulation, optimization
Procedia PDF Downloads 55022011 A Variable Incremental Conductance MPPT Algorithm Applied to Photovoltaic Water Pumping System
Authors: Sarah Abdourraziq, Rachid Elbachtiri
Abstract:
The use of solar energy as a source for pumping water is one of the promising areas in the photovoltaic (PV) application. The energy of photovoltaic pumping systems (PVPS) can be widely improved by employing an MPPT algorithm. This will lead consequently to maximize the electrical motor speed of the system. This paper presents a modified incremental conductance (IncCond) MPPT algorithm with direct control method applied to a standalone PV pumping system. The influence of the algorithm parameters on system behavior is investigated and compared with the traditional (INC) method. The studied system consists of a PV panel, a DC-DC boost converter, and a PMDC motor-pump. The simulation of the system by MATLAB-SIMULINK is carried out. Simulation results found are satisfactory.Keywords: photovoltaic pumping system (PVPS), incremental conductance (INC), MPPT algorithm, boost converter
Procedia PDF Downloads 40222010 An Evolutionary Approach for QAOA for Max-Cut
Authors: Francesca Schiavello
Abstract:
This work aims to create a hybrid algorithm, combining Quantum Approximate Optimization Algorithm (QAOA) with an Evolutionary Algorithm (EA) in the place of traditional gradient based optimization processes. QAOA’s were first introduced in 2014, where, at the time, their algorithm performed better than the traditional best known classical algorithm for Max-cut graphs. Whilst classical algorithms have improved since then and have returned to being faster and more efficient, this was a huge milestone for quantum computing, and their work is often used as a benchmarking tool and a foundational tool to explore variants of QAOA’s. This, alongside with other famous algorithms like Grover’s or Shor’s, highlights to the world the potential that quantum computing holds. It also presents the reality of a real quantum advantage where, if the hardware continues to improve, this could constitute a revolutionary era. Given that the hardware is not there yet, many scientists are working on the software side of things in the hopes of future progress. Some of the major limitations holding back quantum computing are the quality of qubits and the noisy interference they generate in creating solutions, the barren plateaus that effectively hinder the optimization search in the latent space, and the availability of number of qubits limiting the scale of the problem that can be solved. These three issues are intertwined and are part of the motivation for using EAs in this work. Firstly, EAs are not based on gradient or linear optimization methods for the search in the latent space, and because of their freedom from gradients, they should suffer less from barren plateaus. Secondly, given that this algorithm performs a search in the solution space through a population of solutions, it can also be parallelized to speed up the search and optimization problem. The evaluation of the cost function, like in many other algorithms, is notoriously slow, and the ability to parallelize it can drastically improve the competitiveness of QAOA’s with respect to purely classical algorithms. Thirdly, because of the nature and structure of EA’s, solutions can be carried forward in time, making them more robust to noise and uncertainty. Preliminary results show that the EA algorithm attached to QAOA can perform on par with the traditional QAOA with a Cobyla optimizer, which is a linear based method, and in some instances, it can even create a better Max-Cut. Whilst the final objective of the work is to create an algorithm that can consistently beat the original QAOA, or its variants, due to either speedups or quality of the solution, this initial result is promising and show the potential of EAs in this field. Further tests need to be performed on an array of different graphs with the parallelization aspect of the work commencing in October 2023 and tests on real hardware scheduled for early 2024.Keywords: evolutionary algorithm, max cut, parallel simulation, quantum optimization
Procedia PDF Downloads 6022009 Job Satisfaction among Brigadista in Nicaragua: A Lesson to Be Considered for Task-Shifting
Authors: Rashed Shah, Jeanne Koepsell, Dixmer Rivera, Eric Swedberg, David Marsh
Abstract:
Success of primary health care goals of health promotion and disease prevention may well be determined by community based health workers’ overall job satisfaction. It is also important to understand the ways community health workers perceive their jobs and the importance they give to the various factors influencing their job satisfaction, which is critical before making a decision for task-shifting and for expanding their scope of work. Although brigadistas are unpaid volunteers, they are formally recognized and receive support and supervision from the Ministry of Health in Nicaragua. Brigadistas are responsible for classifying and diagnosing illnesses, administering treatment, counseling mothers and care givers within the community, encouraging referral in case of serious illness and making follow-up visits at home. Some brigadistas provide more technically advanced services, including treatment for pneumonia, diarrhea, malaria and tuberculosis and/or distribution of contraceptives. Expanding brigadistas’ duties could threaten their heretofore ‘job satisfaction’. This study primarily aims to report on job satisfaction of brigadistas in Nicaragua before expanding the scope of their work by adding more responsibilities. The study was guided by the following research questions: 1) What aspects of their job made the brigadistas satisfied or dissatisfied? 2) What is the job satisfaction level of brigadistas in Nicaragua? This cross-sectional study was conducted during March – July 2014, to assess brigadistas’ job satisfaction, prior to deciding on inclusion of care for sick newborns and young infants (<2 months of age) to brigadistas’ existing service package of community case management for children of 2-59 months of age. Following stratified random sampling strategy, 15 brigadistas were randomly selected from each of the following four strata: [(1) females under 25 years of age, (2) females over 30 years of age, (3) males under 25 years of age, and (4) males over 30 years of age. Out of 45 completed in-person interview with eligible and available brigadistas, 20 (44.4%) were with female and 25 (55.6%) were with male respondents; the mean age (±sd) was found as 32.0 (±3.2) years. About 53% (24/45) brigadista mentioned “Training” as the most helpful for performing their job. Another 31% (14/45) mentioned that “feeling of doing good, supporting community, women and children” was helpful to perform their job well. When asked about difficulty, about 35.5% (16/45) brigadistas mentioned about “Lack of time” due to their responsibilities in family, farm, other work places, study and such time constraint made their job performance difficult. Measured on a 0-5 scale, estimated average job satisfaction was 4.2. Current trends in task-shifting and integrated program delivery require community health workers (like the brigadistas) to deliver several essential services, including maternal, newborn and child health, and family planning, and thereby increasing their responsibilities. Given the reported level of job satisfaction among brigadistas (4.2 out of 5), and the mentioned difficulty in performing their current job (as ‘Lack of Time’) in this study results, the policy makers and program managers in MOH should be cautious enough before making a decision to expand current scope of work for brigadistas in Nicaragua.Keywords: Brigadisata, job satisfaction, Nicaragua, task-shifting
Procedia PDF Downloads 23122008 Development of Fuzzy Logic Control Ontology for E-Learning
Authors: Muhammad Sollehhuddin A. Jalil, Mohd Ibrahim Shapiai, Rubiyah Yusof
Abstract:
Nowadays, ontology is common in many areas like artificial intelligence, bioinformatics, e-commerce, education and many more. Ontology is one of the focus areas in the field of Information Retrieval. The purpose of an ontology is to describe a conceptual representation of concepts and their relationships within a particular domain. In other words, ontology provides a common vocabulary for anyone who needs to share information in the domain. There are several ontology domains in various fields including engineering and non-engineering knowledge. However, there are only a few available ontology for engineering knowledge. Fuzzy logic as engineering knowledge is still not available as ontology domain. In general, fuzzy logic requires step-by-step guidelines and instructions of lab experiments. In this study, we presented domain ontology for Fuzzy Logic Control (FLC) knowledge. We give Table of Content (ToC) with middle strategy based on the Uschold and King method to develop FLC ontology. The proposed framework is developed using Protégé as the ontology tool. The Protégé’s ontology reasoner, known as the Pellet reasoner is then used to validate the presented framework. The presented framework offers better performance based on consistency and classification parameter index. In general, this ontology can provide a platform to anyone who needs to understand FLC knowledge.Keywords: engineering knowledge, fuzzy logic control ontology, ontology development, table of content
Procedia PDF Downloads 30022007 Inference for Compound Truncated Poisson Lognormal Model with Application to Maximum Precipitation Data
Authors: M. Z. Raqab, Debasis Kundu, M. A. Meraou
Abstract:
In this paper, we have analyzed maximum precipitation data during a particular period of time obtained from different stations in the Global Historical Climatological Network of the USA. One important point to mention is that some stations are shut down on certain days for some reason or the other. Hence, the maximum values are recorded by excluding those readings. It is assumed that the number of stations that operate follows zero-truncated Poisson random variables, and the daily precipitation follows a lognormal random variable. We call this model a compound truncated Poisson lognormal model. The proposed model has three unknown parameters, and it can take a variety of shapes. The maximum likelihood estimators can be obtained quite conveniently using Expectation-Maximization (EM) algorithm. Approximate maximum likelihood estimators are also derived. The associated confidence intervals also can be obtained from the observed Fisher information matrix. Simulation results have been performed to check the performance of the EM algorithm, and it is observed that the EM algorithm works quite well in this case. When we analyze the precipitation data set using the proposed model, it is observed that the proposed model provides a better fit than some of the existing models.Keywords: compound Poisson lognormal distribution, EM algorithm, maximum likelihood estimation, approximate maximum likelihood estimation, Fisher information, skew distribution
Procedia PDF Downloads 10922006 Composite Approach to Extremism and Terrorism Web Content Classification
Authors: Kolade Olawande Owoeye, George Weir
Abstract:
Terrorism and extremism activities on the internet are becoming the most significant threats to national security because of their potential dangers. In response to this challenge, law enforcement and security authorities are actively implementing comprehensive measures by countering the use of the internet for terrorism. To achieve the measures, there is need for intelligence gathering via the internet. This includes real-time monitoring of potential websites that are used for recruitment and information dissemination among other operations by extremist groups. However, with billions of active webpages, real-time monitoring of all webpages become almost impossible. To narrow down the search domain, there is a need for efficient webpage classification techniques. This research proposed a new approach tagged: SentiPosit-based method. SentiPosit-based method combines features of the Posit-based method and the Sentistrenght-based method for classification of terrorism and extremism webpages. The experiment was carried out on 7500 webpages obtained through TENE-webcrawler by International Cyber Crime Research Centre (ICCRC). The webpages were manually grouped into three classes which include the ‘pro-extremist’, ‘anti-extremist’ and ‘neutral’ with 2500 webpages in each category. A supervised learning algorithm is then applied on the classified dataset in order to build the model. Results obtained was compared with existing classification method using the prediction accuracy and runtime. It was observed that our proposed hybrid approach produced a better classification accuracy compared to existing approaches within a reasonable runtime.Keywords: sentiposit, classification, extremism, terrorism
Procedia PDF Downloads 280