Search results for: linear and polynomial model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18504

Search results for: linear and polynomial model

6204 Topic Sentiments toward the COVID-19 Vaccine on Twitter

Authors: Melissa Vang, Raheyma Khan, Haihua Chen

Abstract:

The coronavirus disease 2019 (COVID‐19) pandemic has changed people's lives from all over the world. More people have turned to Twitter to engage online and discuss the COVID-19 vaccine. This study aims to present a text mining approach to identify people's attitudes towards the COVID-19 vaccine on Twitter. To achieve this purpose, we collected 54,268 COVID-19 vaccine tweets from September 01, 2020, to November 01, 2020, then the BERT model is used for the sentiment and topic analysis. The results show that people had more negative than positive attitudes about the vaccine, and countries with an increasing number of confirmed cases had a higher percentage of negative attitudes. Additionally, the topics discussed in positive and negative tweets are different. The tweet datasets can be helpful to information professionals to inform the public about vaccine-related informational resources. Our findings may have implications for understanding people's cognitions and feelings about the vaccine.

Keywords: BERT, COVID-19 vaccine, sentiment analysis, topic modeling

Procedia PDF Downloads 128
6203 Standard Languages for Creating a Database to Display Financial Statements on a Web Application

Authors: Vladimir Simovic, Matija Varga, Predrag Oreski

Abstract:

XHTML and XBRL are the standard languages for creating a database for the purpose of displaying financial statements on web applications. Today, XBRL is one of the most popular languages for business reporting. A large number of countries in the world recognize the role of XBRL language for financial reporting and the benefits that the reporting format provides in the collection, analysis, preparation, publication and the exchange of data (information) which is the positive side of this language. Here we present all advantages and opportunities that a company may have by using the XBRL format for business reporting. Also, this paper presents XBRL and other languages that are used for creating the database, such XML, XHTML, etc. The role of the AJAX complex model and technology will be explained in detail, and during the exchange of financial data between the web client and web server. Here will be mentioned basic layers of the network for data exchange via the web.

Keywords: XHTML, XBRL, XML, JavaScript, AJAX technology, data exchange

Procedia PDF Downloads 382
6202 Single-Camera Basketball Tracker through Pose and Semantic Feature Fusion

Authors: Adrià Arbués-Sangüesa, Coloma Ballester, Gloria Haro

Abstract:

Tracking sports players is a widely challenging scenario, specially in single-feed videos recorded in tight courts, where cluttering and occlusions cannot be avoided. This paper presents an analysis of several geometric and semantic visual features to detect and track basketball players. An ablation study is carried out and then used to remark that a robust tracker can be built with Deep Learning features, without the need of extracting contextual ones, such as proximity or color similarity, nor applying camera stabilization techniques. The presented tracker consists of: (1) a detection step, which uses a pretrained deep learning model to estimate the players pose, followed by (2) a tracking step, which leverages pose and semantic information from the output of a convolutional layer in a VGG network. Its performance is analyzed in terms of MOTA over a basketball dataset with more than 10k instances.

Keywords: basketball, deep learning, feature extraction, single-camera, tracking

Procedia PDF Downloads 120
6201 Analysis of the Aquifer Vulnerability of a Miopliocene Arid Area Using Drastic and SI Models

Authors: H. Majour, L. Djabri

Abstract:

Many methods in the groundwater vulnerability have been developed in the world (methods like PRAST, DRIST, APRON/ARAA, PRASTCHIM, GOD). In this study, our choice dealt with two recent complementary methods using category mapping of index with weighting criteria (Point County Systems Model MSCP) namely the standard DRASTIC method and SI (Susceptibility Index). At present, these two methods are the most used for the mapping of the intrinsic vulnerability of groundwater. Two classes of groundwater vulnerability in the Biskra sandy aquifer were identified by the DRASTIC method (average and high) and the SI method (very high and high). Integrated analysis has revealed that the high class is predominant for the DRASTIC method whereas for that of SI the preponderance is for the very high class. Furthermore, we notice that the method SI estimates better the vulnerability for the pollution in nitrates, with a rate of 85 % between the concentrations in nitrates of groundwater and the various established classes of vulnerability, against 75 % for the DRASTIC method. By including the land use parameter, the SI method produced more realistic results.

Keywords: DRASTIC, SI, GIS, Biskra sandy aquifer, Algeria

Procedia PDF Downloads 469
6200 Integrating ICT- Based Applications for Sustainable Tourism Development in Algeria

Authors: Boutkhil Guemide, Chellali Benachaiba

Abstract:

Information and Communication Technology (ICT) has an inevitable impact on different industries and their performances. The tourism industry as the largest and fastest growing industry in the world cannot be excluded from this technology and its huge impacts. ICT provides information about tourist attractions in the different destinations before travelling and may improve tourists’ satisfaction. Although Algeria has great tourism potentials, it still needs to be performed well in promoting its attractions to international tourists via ICT tools yet. This research explores the impact of ICT on foreign tourists’ satisfaction of the tourism industry and uses Algerian tourist agencies as a case study, and proposes a model for the impact of ICT on sustainable tourism. Finally, it is concluded that e-ticketing, e-reservation, online payment, multilingual and updated information websites are essential needs for planning strategies in the field of e-tourism. Also, it is recommended that the tourism authorities should develop e-tourism infrastructures in order to keep up with the competitiveness of this field to enable the country to benefit from the global benefits of the tourism industry.

Keywords: Information and Communications Technology (ICT), tourism, tourists’ satisfaction, sustainable tourism

Procedia PDF Downloads 517
6199 Environmental Impact of Pallets in the Supply Chain: Including Logistics and Material Durability in a Life Cycle Assessment Approach

Authors: Joana Almeida, Kendall Reid, Jonas Bengtsson

Abstract:

Pallets are devices that are used for moving and storing freight and are nearly omnipresent in supply chains. The market is dominated by timber pallets, with plastic being a common alternative. Either option underpins the use of important resources (oil, land, timber), the emission of greenhouse gases and additional waste generation in most supply chains. This study uses a dynamic approach to the life cycle assessment (LCA) of pallets. It demonstrates that what ultimately defines the environmental burden of pallets in the supply chain is how often the length of its lifespan, which depends on the durability of the material and on how pallets are utilized. This study proposes a life cycle assessment (LCA) of pallets in supply chains supported by an algorithm that estimates pallet durability in function of material resilience and of logistics. The LCA runs from cradle-to-grave, including raw material provision, manufacture, transport and end of life. The scope is representative of timber and plastic pallets in the Australian and South-East Asia markets. The materials included in this analysis are: -tropical mixed hardwood, unsustainably harvested in SE Asia; -certified softwood, sustainably harvested; -conventional plastic, a mix of virgin and scrap plastic; -recycled plastic pallets, 100% mixed plastic scrap, which are being pioneered by Re > Pal. The logistical model purports that more complex supply chains and rougher handling subject pallets to higher stress loads. More stress shortens the lifespan of pallets in function of their composition. Timber pallets can be repaired, extending their lifespan, while plastic pallets cannot. At the factory gate, softwood pallets have the lowest carbon footprint. Re > pal follows closely due to its burden-free feedstock. Tropical mixed hardwood and plastic pallets have the highest footprints. Harvesting tropical mixed hardwood in SE Asia often leads to deforestation, leading to emissions from land use change. The higher footprint of plastic pallets is due to the production of virgin plastic. Our findings show that manufacture alone does not determine the sustainability of pallets. Even though certified softwood pallets have lower carbon footprint and their lifespan can be extended by repair, the need for re-supply of materials and disposal of waste timber offsets this advantage. It also leads to most waste being generated among all pallets. In a supply chain context, Re > Pal pallets have the lowest footprint due to lower replacement and disposal needs. In addition, Re > Pal are nearly ‘waste neutral’, because the waste that is generated throughout their life cycle is almost totally offset by the scrap uptake for production. The absolute results of this study can be confirmed by progressing the logistics model, improving data quality, expanding the range of materials and utilization practices. Still, this LCA demonstrates that considering logistics, raw materials and material durability is central for sustainable decision-making on pallet purchasing, management and disposal.

Keywords: carbon footprint, life cycle assessment, recycled plastic, waste

Procedia PDF Downloads 205
6198 Using Probabilistic Neural Network (PNN) for Extracting Acoustic Microwaves (Bulk Acoustic Waves) in Piezoelectric Material

Authors: Hafdaoui Hichem, Mehadjebia Cherifa, Benatia Djamel

Abstract:

In this paper, we propose a new method for Bulk detection of an acoustic microwave signal during the propagation of acoustic microwaves in a piezoelectric substrate (Lithium Niobate LiNbO3). We have used the classification by probabilistic neural network (PNN) as a means of numerical analysis in which we classify all the values of the real part and the imaginary part of the coefficient attenuation with the acoustic velocity in order to build a model from which we note the Bulk waves easily. These singularities inform us of presence of Bulk waves in piezoelectric materials. By which we obtain accurate values for each of the coefficient attenuation and acoustic velocity for Bulk waves. This study will be very interesting in modeling and realization of acoustic microwaves devices (ultrasound) based on the propagation of acoustic microwaves.

Keywords: piezoelectric material, probabilistic neural network (PNN), classification, acoustic microwaves, bulk waves, the attenuation coefficient

Procedia PDF Downloads 417
6197 University-Industry Technology Transfer and Technology Transfer Offices in Emerging Economies

Authors: José Carlos Rodríguez, Mario Gómez

Abstract:

The aim of this paper is to get insight on the nature of university-industry technology transfer (UITT) and technology transfer offices (TTOs) activity at universities in the case of emerging economies. In relation to the process of transferring knowledge/technology in the case of emerging economies, knowledge/technology transfer in these economies are more reactive than in developed economies due to differences in maturity of technologies. It is assumed in this paper that knowledge/technology transfer is a complex phenomenon, and thus the paper contributes to get insight on the nature of UITT and TTOs creation in the case of emerging economies by using a system dynamics model of knowledge/technology transfer in these countries. The paper recognizes the differences between industrialized countries and emerging economies on these phenomena.

Keywords: university-industry technology transfer, technology transfer offices, technology transfer models, emerging economies

Procedia PDF Downloads 230
6196 Modal Analysis for Optimal Location of Doubly Fed Induction-Generator-Based Wind Farms for Reduction of Small Signal Oscillation

Authors: Meet Patel, Darshan Patel, Nilay Shah

Abstract:

Excess growth of wind-based renewable energy sources is required to identify the optimal location and damping capacity of doubly fed induction-generator-based (DFIG) wind farms while it penetrates into the transmission network. In this analysis, various ratings of DFIG wind farms are penetrated into the Single Machine Infinite Bus (SMIB ) at a different distance of the transmission line. On the basis of detailed examinations, a prime position is evaluated to maximize the stability of overall systems. A damping controller is designed at an optimum location to mitigate the small oscillations. The proposed model was validated using eigenvalue analysis, calculation of the participation factor, and time-domain simulation.

Keywords: DFIG, small signal stability, eigenvalues, time domain simulation

Procedia PDF Downloads 91
6195 Attack Redirection and Detection using Honeypots

Authors: Chowduru Ramachandra Sharma, Shatunjay Rawat

Abstract:

A false positive state is when the IDS/IPS identifies an activity as an attack, but the activity is acceptable behavior in the system. False positives in a Network Intrusion Detection System ( NIDS ) is an issue because they desensitize the administrator. It wastes computational power and valuable resources when rules are not tuned properly, which is the main issue with anomaly NIDS. Furthermore, most false positives reduction techniques are not performed during the real-time of attempted intrusions; instead, they have applied afterward on collected traffic data and generate alerts. Of course, false positives detection in ‘offline mode’ is tremendously valuable. Nevertheless, there is room for improvement here; automated techniques still need to reduce False Positives in real-time. This paper uses the Snort signature detection model to redirect the alerted attacks to Honeypots and verify attacks.

Keywords: honeypot, TPOT, snort, NIDS, honeybird, iptables, netfilter, redirection, attack detection, docker, snare, tanner

Procedia PDF Downloads 139
6194 Co-Evolution of Urban Lake System and Rapid Urbanization: Case of Raipur, Chhattisgarh

Authors: Kamal Agrawal, Ved Prakash Nayak, Akshay Patil

Abstract:

Raipur is known as a city of water bodies. The city had around 200 man-made and natural lakes of varying sizes. These structures were constructed to collect rainwater and control flooding in the city. Due to the transition from community participation to state government, as well as rapid urbanisation, Raipur now has only about 80 lakes left. Rapid and unplanned growth has resulted in pollution, encroachment, and eutrophication of the city's lakes. The state government keeps these lakes in good condition by cleaning them and proposing lakefront developments. However, maintaining individual lakes is insufficient because urban lakes are not distinct entities. It is a system comprised of the lake, shore, catchment, and other components. While Urban lake system (ULS) is a combination of multiple such lake systems interacting in a complex urban setting. Thus, the project aims to propose a co-evolution model for urban lake systems (ULS) and rapid urbanization in Raipur. The goals are to comprehend the ULS and to identify elements and dimensions of urbanization that influence the ULS. Evaluate the impact of rapid urbanization on the ULS & vice versa in the study area. Determine how to maximize the positive impact while minimizing the negative impact identified in the study area. Propose short-, medium-, and long-term planning interventions to support the ULS's co-evolution with rapid urbanization. A complexity approach is used to investigate the ULS. It is a technique for understanding large, complex systems. A complex system is one with many interconnected and interdependent elements and dimensions. Thus, elements of ULS and rapid urbanization are identified through a literature study to evaluate statements of their impacts (Beneficial/ Adverse) on one another. Rapid urbanization has been identified as having elements such as demography, urban legislation, informal settlement, urban infrastructure, and tourism. Similarly, the catchment area of the lake, the lake's water quality, the water spread area, and lakefront developments are all being impacted by rapid urbanisation. These nine elements serve as parameters for the subsequent analysis. Elements are limited to physical parameters only. The city has designated a study area based on the definition provided by the National Plan for the Conservation of Aquatic Ecosystems. Three lakes are discovered within a one-kilometer radius, establishing a tiny urban lake system. Because the condition of a lake is directly related to the condition of its catchment area, the catchment area of these three lakes is delineated as the study area. Data is collected to identify impact statements, and the interdependence diagram generated between the parameters yields results in terms of interlinking between each parameter and their impact on the system as a whole. The planning interventions proposed for the ULS and rapid urbanisation co-evolution model include spatial proposals as well as policy recommendations for the short, medium, and long term. This study's next step will be to determine how to implement the proposed interventions based on the availability of resources, funds, and governance patterns.

Keywords: urban lake system, co-evolution, rapid urbanization, complex system

Procedia PDF Downloads 57
6193 Decentralized Peak-Shaving Strategies for Integrated Domestic Batteries

Authors: Corentin Jankowiak, Aggelos Zacharopoulos, Caterina Brandoni

Abstract:

In a context of increasing stress put on the electricity network by the decarbonization of many sectors, energy storage is likely to be the key mitigating element, by acting as a buffer between production and demand. In particular, the highest potential for storage is when connected closer to the loads. Yet, low voltage storage struggles to penetrate the market at a large scale due to the novelty and complexity of the solution, and the competitive advantage of fossil fuel-based technologies regarding regulations. Strong and reliable numerical simulations are required to show the benefits of storage located near loads and promote its development. The present study was restrained from excluding aggregated control of storage: it is assumed that the storage units operate independently to one another without exchanging information – as is currently mostly the case. A computationally light battery model is presented in detail and validated by direct comparison with a domestic battery operating in real conditions. This model is then used to develop Peak-Shaving (PS) control strategies as it is the decentralized service from which beneficial impacts are most likely to emerge. The aggregation of flatter, peak- shaved consumption profiles is likely to lead to flatter and arbitraged profile at higher voltage layers. Furthermore, voltage fluctuations can be expected to decrease if spikes of individual consumption are reduced. The crucial part to achieve PS lies in the charging pattern: peaks depend on the switching on and off of appliances in the dwelling by the occupants and are therefore impossible to predict accurately. A performant PS strategy must, therefore, include a smart charge recovery algorithm that can ensure enough energy is present in the battery in case it is needed without generating new peaks by charging the unit. Three categories of PS algorithms are introduced in detail. First, using a constant threshold or power rate for charge recovery, followed by algorithms using the State Of Charge (SOC) as a decision variable. Finally, using a load forecast – of which the impact of the accuracy is discussed – to generate PS. A performance metrics was defined in order to quantitatively evaluate their operating regarding peak reduction, total energy consumption, and self-consumption of domestic photovoltaic generation. The algorithms were tested on load profiles with a 1-minute granularity over a 1-year period, and their performance was assessed regarding these metrics. The results show that constant charging threshold or power are far from optimal: a certain value is not likely to fit the variability of a residential profile. As could be expected, forecast-based algorithms show the highest performance. However, these depend on the accuracy of the forecast. On the other hand, SOC based algorithms also present satisfying performance, making them a strong alternative when the reliable forecast is not available.

Keywords: decentralised control, domestic integrated batteries, electricity network performance, peak-shaving algorithm

Procedia PDF Downloads 102
6192 A Numerical Study on the Connection of an SC Wall to an RC Foundation

Authors: Siamak Epackachi, Andrew S. Whittaker, Amit H. Varma

Abstract:

There are a large number of methods to connect SC walls to RC foundations. An experimental study of the cyclic nonlinear behavior of SC walls in the NEES laboratory at the University at Buffalo used a connection detail involving the post-tensioning of a steel baseplate to the SC wall to a RC foundation. This type of connection introduces flexibility that influenced substantially the global response of the SC walls. The assumption of a rigid base, which would be commonly made by practitioners, would lead to a substantial overestimation of initial stiffness. This paper presents an analytical approach to characterize the rotational flexibility and to predict the initial stiffness of flexure-critical SC wall piers with baseplate connection. The good agreement between the analytical and test results confirmed the utility of the proposed method for calculating the initial stiffness of an SC wall with baseplate connection.

Keywords: steel-plate composite shear wall, flexure-critical wall, cyclic loading, analytical model

Procedia PDF Downloads 325
6191 Development of a System for Fitting Clothes and Accessories Using Augmented Reality

Authors: Dinmukhamed T., Vassiliy S.

Abstract:

This article suggests the idea of fitting clothes and accessories based on augmented reality. A logical data model has been developed, taking into account the decision-making module (colors, style, type, material, popularity, etc.) based on personal data (age, gender, weight, height, leg size, hoist length, geolocation, photogrammetry, number of purchases of certain types of clothing, etc.) and statistical data of the purchase history (number of items, price, size, color, style, etc.). Also, in order to provide information to the user, it is planned to develop an augmented reality system using a QR code. This system of selection and fitting of clothing and accessories based on augmented reality will be used in stores to reduce the time for the buyer to make a decision on the choice of clothes.

Keywords: augmented reality, online store, decision-making module, like QR code, clothing store, queue

Procedia PDF Downloads 136
6190 A Preliminary Study for Building an Arabic Corpus of Pair Questions-Texts from the Web: Aqa-Webcorp

Authors: Wided Bakari, Patrce Bellot, Mahmoud Neji

Abstract:

With the development of electronic media and the heterogeneity of Arabic data on the Web, the idea of building a clean corpus for certain applications of natural language processing, including machine translation, information retrieval, question answer, become more and more pressing. In this manuscript, we seek to create and develop our own corpus of pair’s questions-texts. This constitution then will provide a better base for our experimentation step. Thus, we try to model this constitution by a method for Arabic insofar as it recovers texts from the web that could prove to be answers to our factual questions. To do this, we had to develop a java script that can extract from a given query a list of html pages. Then clean these pages to the extent of having a database of texts and a corpus of pair’s question-texts. In addition, we give preliminary results of our proposal method. Some investigations for the construction of Arabic corpus are also presented in this document.

Keywords: Arabic, web, corpus, search engine, URL, question, corpus building, script, Google, html, txt

Procedia PDF Downloads 308
6189 Impedimetric Phage-Based Sensor for the Rapid Detection of Staphylococcus aureus from Nasal Swab

Authors: Z. Yousefniayejahr, S. Bolognini, A. Bonini, C. Campobasso, N. Poma, F. Vivaldi, M. Di Luca, A. Tavanti, F. Di Francesco

Abstract:

Pathogenic bacteria represent a threat to healthcare systems and the food industry because their rapid detection remains challenging. Electrochemical biosensors are gaining prominence as a novel technology for the detection of pathogens due to intrinsic features such as low cost, rapid response time, and portability, which make them a valuable alternative to traditional methodologies. These sensors use biorecognition elements that are crucial for the identification of specific bacteria. In this context, bacteriophages are promising tools for their inherent high selectivity towards bacterial hosts, which is of fundamental importance when detecting bacterial pathogens in complex biological samples. In this study, we present the development of a low-cost and portable sensor based on the Zeno phage for the rapid detection of Staphylococcus aureus. Screen-printed gold electrodes functionalized with the Zeno phage were used, and electrochemical impedance spectroscopy was applied to evaluate the change of the charge transfer resistance (Rct) as a result of the interaction with S. aureus MRSA ATCC 43300. The phage-based biosensor showed a linear range from 101 to 104 CFU/mL with a 20-minute response time and a limit of detection (LOD) of 1.2 CFU/mL under physiological conditions. The biosensor’s ability to recognize various strains of staphylococci was also successfully demonstrated in the presence of clinical isolates collected from different geographic areas. Assays using S. epidermidis were also carried out to verify the species-specificity of the phage sensor. We only observed a remarkable change of the Rct in the presence of the target S. aureus bacteria, while no substantial binding to S. epidermidis occurred. This confirmed that the Zeno phage sensor only targets S. aureus species within the genus Staphylococcus. In addition, the biosensor's specificity with respect to other bacterial species, including gram-positive bacteria like Enterococcus faecium and the gram-negative bacterium Pseudomonas aeruginosa, was evaluated, and a non-significant impedimetric signal was observed. Notably, the biosensor successfully identified S. aureus bacterial cells in a complex matrix such as a nasal swab, opening the possibility of its use in a real-case scenario. We diluted different concentrations of S. aureus from 108 to 100 CFU/mL with a ratio of 1:10 in the nasal swap matrices collected from healthy donors. Three different sensors were applied to measure various concentrations of bacteria. Our sensor indicated high selectivity to detect S. aureus in biological matrices compared to time-consuming traditional methods, such as enzyme-linked immunosorbent assay (ELISA), polymerase chain reaction (PCR), and radioimmunoassay (RIA), etc. With the aim to study the possibility to use this biosensor to address the challenge associated to pathogen detection, ongoing research is focused on the assessment of the biosensor’s analytical performances in different biological samples and the discovery of new phage bioreceptors.

Keywords: electrochemical impedance spectroscopy, bacteriophage, biosensor, Staphylococcus aureus

Procedia PDF Downloads 50
6188 Privacy Preserving Data Publishing Based on Sensitivity in Context of Big Data Using Hive

Authors: P. Srinivasa Rao, K. Venkatesh Sharma, G. Sadhya Devi, V. Nagesh

Abstract:

Privacy Preserving Data Publication is the main concern in present days because the data being published through the internet has been increasing day by day. This huge amount of data was named as Big Data by its size. This project deals the privacy preservation in the context of Big Data using a data warehousing solution called hive. We implemented Nearest Similarity Based Clustering (NSB) with Bottom-up generalization to achieve (v,l)-anonymity. (v,l)-Anonymity deals with the sensitivity vulnerabilities and ensures the individual privacy. We also calculate the sensitivity levels by simple comparison method using the index values, by classifying the different levels of sensitivity. The experiments were carried out on the hive environment to verify the efficiency of algorithms with Big Data. This framework also supports the execution of existing algorithms without any changes. The model in the paper outperforms than existing models.

Keywords: sensitivity, sensitive level, clustering, Privacy Preserving Data Publication (PPDP), bottom-up generalization, Big Data

Procedia PDF Downloads 276
6187 Forecasting Exchange Rate between Thai Baht and the US Dollar Using Time Series Analysis

Authors: Kunya Bowornchockchai

Abstract:

The objective of this research is to forecast the monthly exchange rate between Thai baht and the US dollar and to compare two forecasting methods. The methods are Box-Jenkins’ method and Holt’s method. Results show that the Box-Jenkins’ method is the most suitable method for the monthly Exchange Rate between Thai Baht and the US Dollar. The suitable forecasting model is ARIMA (1,1,0)  without constant and the forecasting equation is Yt = Yt-1 + 0.3691 (Yt-1 - Yt-2) When Yt  is the time series data at time t, respectively.

Keywords: Box–Jenkins method, Holt’s method, mean absolute percentage error (MAPE), exchange rate

Procedia PDF Downloads 238
6186 Fast and Robust Long-term Tracking with Effective Searching Model

Authors: Thang V. Kieu, Long P. Nguyen

Abstract:

Kernelized Correlation Filter (KCF) based trackers have gained a lot of attention recently because of their accuracy and fast calculation speed. However, this algorithm is not robust in cases where the object is lost by a sudden change of direction, being obscured or going out of view. In order to improve KCF performance in long-term tracking, this paper proposes an anomaly detection method for target loss warning by analyzing the response map of each frame, and a classification algorithm for reliable target re-locating mechanism by using Random fern. Being tested with Visual Tracker Benchmark and Visual Object Tracking datasets, the experimental results indicated that the precision and success rate of the proposed algorithm were 2.92 and 2.61 times higher than that of the original KCF algorithm, respectively. Moreover, the proposed tracker handles occlusion better than many state-of-the-art long-term tracking methods while running at 60 frames per second.

Keywords: correlation filter, long-term tracking, random fern, real-time tracking

Procedia PDF Downloads 124
6185 Robot Movement Using the Trust Region Policy Optimization

Authors: Romisaa Ali

Abstract:

The Policy Gradient approach is one of the deep reinforcement learning families that combines deep neural networks (DNN) with reinforcement learning RL to discover the optimum of the control problem through experience gained from the interaction between the robot and its surroundings. In contrast to earlier policy gradient algorithms, which were unable to handle these two types of error because of over-or under-estimation introduced by the deep neural network model, this article will discuss the state-of-the-art SOTA policy gradient technique, trust region policy optimization (TRPO), by applying this method in various environments compared to another policy gradient method, the Proximal Policy Optimization (PPO), to explain their robust optimization, using this SOTA to gather experience data during various training phases after observing the impact of hyper-parameters on neural network performance.

Keywords: deep neural networks, deep reinforcement learning, proximal policy optimization, state-of-the-art, trust region policy optimization

Procedia PDF Downloads 151
6184 The End Is Just the Beginning: The Importance of Project Post-Implementation Reviews

Authors: Catalin-Teodor Dogaru, Ana-Maria Dogaru

Abstract:

Success means different things to different people. For us, project managers, it becomes even harder to find a definition. Many factors have to be included in the evaluation. Moreover, literature is not very helpful, lacking consensus and neutrality. Post-implementation reviews (PIR) can be an efficient tool in evaluating how things worked on a certain project. Despite the visible progress, PIR is not a very detailed subject yet and there is not a common understanding in this matter. This may be the reason that some organizations include it in the projects’ lifecycle and some do not. Through this paper, we point out the reasons why all project managers should pay proper attention to this important step and to the elements, which can be assessed, beside the already famous triple constraints: cost, budget, and time. It is essential to take notice that PIR is not a checklist. It brings the edge in eliminating subjectivity and judging projects based on actual proof. Based on our experience, our success indicator model, presented in this paper, contributes to the success of the project! In the same time, it increases trust among customers who will perceive success more objectively.

Keywords: project, post implementation, review, success, indicators

Procedia PDF Downloads 353
6183 An Evaluation of a First Year Introductory Statistics Course at a University in Jamaica

Authors: Ayesha M. Facey

Abstract:

The evaluation sought to determine the factors associated with the high failure rate among students taking a first-year introductory statistics course. By utilizing Tyler’s Objective Based Model, the main objectives were: to assess the effectiveness of the lecturer’s teaching strategies; to determine the proportion of students who attends lectures and tutorials frequently and to determine the impact of infrequent attendance on performance; to determine how the assigned activities assisted in students understanding of the course content; to ascertain the possible issues being faced by students in understanding the course material and obtain possible solutions to the challenges and to determine whether the learning outcomes have been achieved based on an assessment of the second in-course examination. A quantitative survey research strategy was employed and the study population was students enrolled in semester one of the academic year 2015/2016. A convenience sampling approach was employed resulting in a sample of 98 students. Primary data was collected using self-administered questionnaires over a one-week period. Secondary data was obtained from the results of the second in-course examination. Data were entered and analyzed in SPSS version 22 and both univariate and bivariate analyses were conducted on the information obtained from the questionnaires. Univariate analyses provided description of the sample through means, standard deviations and percentages while bivariate analyses were done using Spearman’s Rho correlation coefficient and Chi-square analyses. For secondary data, an item analysis was performed to obtain the reliability of the examination questions, difficulty index and discriminant index. The examination results also provided information on the weak areas of the students and highlighted the learning outcomes that were not achieved. Findings revealed that students were more likely to participate in lectures than tutorials and that attendance was high for both lectures and tutorials. There was a significant relationship between participation in lectures and performance on examination. However, a high proportion of students has been absent from three or more tutorials as well as lectures. A higher proportion of students indicated that they completed the assignments obtained from the lectures sometimes while they rarely completed tutorial worksheets. Students who were more likely to complete their assignments were significantly more likely to perform well on their examination. Additionally, students faced a number of challenges in understanding the course content and the topics of probability, binomial distribution and normal distribution were the most challenging. The item analysis also highlighted these topics as problem areas. Problems doing mathematics and application and analyses were their major challenges faced by students and most students indicated that some of the challenges could be alleviated if additional examples were worked in lectures and they were given more time to solve questions. Analysis of the examination results showed that a number of learning outcomes were not achieved for a number of topics. Based on the findings recommendations were made that suggested adjustments to grade allocations, delivery of lectures and methods of assessment.

Keywords: evaluation, item analysis, Tyler’s objective based model, university statistics

Procedia PDF Downloads 175
6182 Finite Element Analysis of RC Frames with Retrofitted Infill Walls

Authors: M. Ömer Timurağaoğlu, Adem Doğangün, Ramazan Livaoğlu

Abstract:

The evaluation of performance of infilled reinforced concrete (RC) frames has been a significant challenge for engineers. The strengthening of infill walls has been an important concern to enhance the behavior of RC infilled frames. The aim of this study is to investigate the behaviour of retrofitted infill walls of RC frames using finite element analysis. For this purpose, a one storey, one bay infilled and strengthened infilled RC frame which have the same geometry and material properties with the frames tested in laboratory are modelled using different analytical approaches. A fibrous material is used to strengthen infill walls and frame. As a consequence, the results of the finite element analysis were evaluated of whether these analytical approaches estimate the behavior or not. To model the infilled and strengthened infilled RC frames, a finite element program ABAQUS is used. Finally, data obtained from the nonlinear finite element analysis is compared with the experimental results.

Keywords: finite element analysis, infilled RC frames, infill wall, strengthening

Procedia PDF Downloads 513
6181 Text Similarity in Vector Space Models: A Comparative Study

Authors: Omid Shahmirzadi, Adam Lugowski, Kenneth Younge

Abstract:

Automatic measurement of semantic text similarity is an important task in natural language processing. In this paper, we evaluate the performance of different vector space models to perform this task. We address the real-world problem of modeling patent-to-patent similarity and compare TFIDF (and related extensions), topic models (e.g., latent semantic indexing), and neural models (e.g., paragraph vectors). Contrary to expectations, the added computational cost of text embedding methods is justified only when: 1) the target text is condensed; and 2) the similarity comparison is trivial. Otherwise, TFIDF performs surprisingly well in other cases: in particular for longer and more technical texts or for making finer-grained distinctions between nearest neighbors. Unexpectedly, extensions to the TFIDF method, such as adding noun phrases or calculating term weights incrementally, were not helpful in our context.

Keywords: big data, patent, text embedding, text similarity, vector space model

Procedia PDF Downloads 154
6180 Solution of Insurance Pricing Model Giving Optimum Premium Level for Both Insured and Insurer by Game Theory

Authors: Betul Zehra Karagul

Abstract:

A game consists of strategies that each actor has in his/her own choice strategies, and a game regulates the certain rules in the strategies that the actors choose, express how they evaluate their knowledge and the utility of output results. Game theory examines the human behaviors (preferences) of strategic situations in which each actor of a game regards the action that others will make in spite of his own moves. There is a balance between each player playing a game with the final number of players and the player with a certain probability of choosing the players, and this is called Nash equilibrium. The insurance is a two-person game where the insurer and insured are the actors. Both sides have the right to act in favor of utility functions. The insured has to pay a premium to buy the insurance cover. The insured will want to pay a low premium while the insurer is willing to get a high premium. In this study, the state of equilibrium for insurance pricing was examined in terms of the insurer and insured with game theory.

Keywords: game theory, insurance pricing, Nash equilibrium, utility function

Procedia PDF Downloads 337
6179 Effects of Cattaneo-Christov Heat Flux on 3D Magnetohydrodynamic Viscoelastic Fluid Flow with Variable Thermal Conductivity

Authors: Muhammad Ramzan

Abstract:

A mathematical model has been envisaged to discuss three-dimensional Viscoelastic fluid flow with an effect of Cattaneo-Christov heat flux in attendance of magnetohydrodynamic (MHD). Variable thermal conductivity with the impact of homogeneous-heterogeneous reactions and convective boundary condition is also taken into account. Homotopy analysis method is engaged to obtain series solutions. Graphical illustrations depicting behaviour of sundry parameters on skin friction coefficient and all involved distributions are also given. It is observed that velocity components are decreasing functions of Viscoelastic fluid parameter. Furthermore, strength of homogeneous and heterogeneous reactions have opposite effects on concentration distribution. A comparison with a published paper has also been established and an excellent agreement is obtained; hence reliable results are being presented.

Keywords: Cattaneo Christov heat flux, homogenous-heterogeneous reactions, magnetic field, variable thermal conductivity

Procedia PDF Downloads 184
6178 Risk Management of Water Derivatives: A New Commodity in The Market

Authors: Daniel Mokatsanyane, Johnny Jansen Van Rensburg

Abstract:

This paper is a concise introduction of the risk management on the water derivatives market. Water, a new commodity in the market, is one of the most important commodity on earth. As important to life and planet as crops, metals, and energy, none of them matters without water. This paper presents a brief overview of water as a tradable commodity via a new first of its kind futures contract on the Nasdaq Veles California Water Index (NQH2O) derivative instrument, TheGeneralised Autoregressive Conditional Heteroscedasticity (GARCH) statistical model will be the used to measure the water price volatility of the instrument and its performance since it’s been traded. describe the main products and illustrate their usage in risk management and also discuss key challenges with modeling and valuation of water as a traded commodity and finally discuss how water derivatives may be taken as an alternative asset investment class.

Keywords: water derivatives, commodity market, nasdaq veles california water Index (NQH2O, water price, risk management

Procedia PDF Downloads 115
6177 Precision Assessment of the Orthometric Heights Determination in the Northern Part of Libya

Authors: Jamal A. Gledan, Akrm H. Algnin

Abstract:

The Global Positioning System (GPS) satellite-based technology has been utilized extensively in the last few years in a wide range of Geomatics and Geographic Information Systems (GIS) applications. One of the main challenges dealing with GPS-based heights consists of converting them into Mean Sea Level (MSL) heights which is used in surveys and mapping. In this research work, differences in heights of 50 points, in northern part of Libya were carried out using both ordinary levelling (in which Geoid is the reference datum) and GPS techniques (in which Ellipsoid is the reference datum). In addition, this study has utilized the EGM2008 model to obtain the undulation values between the ellipsoidal and orthometric heights. From these values with ellipsoidal heights which can be obtained from GPS observations to compute the orthomteric heights. This research presented a suitable alternative, from an economical point of view, to substitute the expensive traditional levelling technique particularly for topographic mapping.

Keywords: geoid undulation, GPS, ordinary and geodetic levelling, orthometric height

Procedia PDF Downloads 420
6176 Optimal Tracking Control of a Hydroelectric Power Plant Incorporating Neural Forecasting for Uncertain Input Disturbances

Authors: Marlene Perez Villalpando, Kelly Joel Gurubel Tun

Abstract:

In this paper, we propose an optimal control strategy for a hydroelectric power plant subject to input disturbances like meteorological phenomena. The engineering characteristics of the system are described by a nonlinear model. The random availability of renewable sources is predicted by a high-order neural network trained with an extended Kalman filter, whereas the power generation is regulated by the optimal control law. The main advantage of the system is the stabilization of the amount of power generated in the plant. A control supervisor maintains stability and availability in hydropower reservoirs water levels for power generation. The proposed approach demonstrated a good performance to stabilize the reservoir level and the power generation along their desired trajectories in the presence of disturbances.

Keywords: hydropower, high order neural network, Kalman filter, optimal control

Procedia PDF Downloads 281
6175 The Effect of Spatial Variability on Axial Pile Design of Closed Ended Piles in Sand

Authors: Cormac Reale, Luke J. Prendergast, Kenneth Gavin

Abstract:

While significant improvements have been made in axial pile design methods over recent years, the influence of soils natural variability has not been adequately accounted for within them. Soil variability is a crucial parameter to consider as it can account for large variations in pile capacity across the same site. This paper seeks to address this knowledge deficit, by demonstrating how soil spatial variability can be accommodated into existing cone penetration test (CPT) based pile design methods, in the form of layered non-homogeneous random fields. These random fields model the scope of a given property’s variance and define how it varies spatially. A Monte Carlo analysis of the pile will be performed taking into account parameter uncertainty and spatial variability, described using the measured scales of fluctuation. The results will be discussed in light of Eurocode 7 and the effect of spatial averaging on design capacities will be analysed.

Keywords: pile axial design, reliability, spatial variability, CPT

Procedia PDF Downloads 230