Search results for: distributed model
18160 An Efficient FPGA Realization of Fir Filter Using Distributed Arithmetic
Authors: M. Iruleswari, A. Jeyapaul Murugan
Abstract:
Most fundamental part used in many Digital Signal Processing (DSP) application is a Finite Impulse Response (FIR) filter because of its linear phase, stability and regular structure. Designing a high-speed and hardware efficient FIR filter is a very challenging task as the complexity increases with the filter order. In most applications the higher order filters are required but the memory usage of the filter increases exponentially with the order of the filter. Using multipliers occupy a large chip area and need high computation time. Multiplier-less memory-based techniques have gained popularity over past two decades due to their high throughput processing capability and reduced dynamic power consumption. This paper describes the design and implementation of highly efficient Look-Up Table (LUT) based circuit for the implementation of FIR filter using Distributed arithmetic algorithm. It is a multiplier less FIR filter. The LUT can be subdivided into a number of LUT to reduce the memory usage of the LUT for higher order filter. Analysis on the performance of various filter orders with different address length is done using Xilinx 14.5 synthesis tool. The proposed design provides less latency, less memory usage and high throughput.Keywords: finite impulse response, distributed arithmetic, field programmable gate array, look-up table
Procedia PDF Downloads 45718159 Using the Weakest Precondition to Achieve Self-Stabilization in Critical Networks
Authors: Antonio Pizzarello, Oris Friesen
Abstract:
Networks, such as the electric power grid, must demonstrate exemplary performance and integrity. Integrity depends on the quality of both the system design model and the deployed software. Integrity of the deployed software is key, for both the original versions and the many that occur throughout numerous maintenance activity. Current software engineering technology and practice do not produce adequate integrity. Distributed systems utilize networks where each node is an independent computer system. The connections between them is realized via a network that is normally redundantly connected to guarantee the presence of a path between two nodes in the case of failure of some branch. Furthermore, at each node, there is software which may fail. Self-stabilizing protocols are usually present that recognize failure in the network and perform a repair action that will bring the node back to a correct state. These protocols first introduced by E. W. Dijkstra are currently present in almost all Ethernets. Super stabilization protocols capable of reacting to a change in the network topology due to the removal or addition of a branch in the network are less common but are theoretically defined and available. This paper describes how to use the Software Integrity Assessment (SIA) methodology to analyze self-stabilizing software. SIA is based on the UNITY formalism for parallel and distributed programming, which allows the analysis of code for verifying the progress property p leads-to q that describes the progress of all computations starting in a state satisfying p to a state satisfying q via the execution of one or more system modules. As opposed to demonstrably inadequate test and evaluation methods SIA allows the analysis and verification of any network self-stabilizing software as well as any other software that is designed to recover from failure without external intervention of maintenance personnel. The model to be analyzed is obtained by automatic translation of the system code to a transition system that is based on the use of the weakest precondition.Keywords: network, power grid, self-stabilization, software integrity assessment, UNITY, weakest precondition
Procedia PDF Downloads 22318158 Building Education Leader Capacity through an Integrated Information and Communication Technology Leadership Model and Tool
Authors: Sousan Arafeh
Abstract:
Educational systems and schools worldwide are increasingly reliant on information and communication technology (ICT). Unfortunately, most educational leadership development programs do not offer formal curricular and/or field experiences that prepare students for managing ICT resources, personnel, and processes. The result is a steep learning curve for the leader and his/her staff and dissipated organizational energy that compromises desired outcomes. To address this gap in education leaders’ development, Arafeh’s Integrated Technology Leadership Model (AITLM) was created. It is a conceptual model and tool that educational leadership students can use to better understand the ICT ecology that exists within their schools. The AITL Model consists of six 'infrastructure types' where ICT activity takes place: technical infrastructure, communications infrastructure, core business infrastructure, context infrastructure, resources infrastructure, and human infrastructure. These six infrastructures are further divided into 16 key areas that need management attention. The AITL Model was created by critically analyzing existing technology/ICT leadership models and working to make something more authentic and comprehensive regarding school leaders’ purview and experience. The AITL Model then served as a tool when it was distributed to over 150 educational leadership students who were asked to review it and qualitatively share their reactions. Students said the model presented crucial areas of consideration that they had not been exposed to before and that the exercise of reviewing and discussing the AITL Model as a group was useful for identifying areas of growth that they could pursue in the leadership development program and in their professional settings. While development in all infrastructures and key areas was important for students’ understanding of ICT, they noted that they were least aware of the importance of the intangible area of the resources infrastructure. The AITL Model will be presented and session participants will have an opportunity to review and reflect on its impact and utility. Ultimately, the AITL Model is one that could have significant policy and practice implications. At the very least, it might help shape ICT content in educational leadership development programs through curricular and pedagogical updates.Keywords: education leadership, information and communications technology, ICT, leadership capacity building, leadership development
Procedia PDF Downloads 11618157 Distributed Processing for Content Based Lecture Video Retrieval on Hadoop Framework
Authors: U. S. N. Raju, Kothuri Sai Kiran, Meena G. Kamal, Vinay Nikhil Pabba, Suresh Kanaparthi
Abstract:
There is huge amount of lecture video data available for public use, and many more lecture videos are being created and uploaded every day. Searching for videos on required topics from this huge database is a challenging task. Therefore, an efficient method for video retrieval is needed. An approach for automated video indexing and video search in large lecture video archives is presented. As the amount of video lecture data is huge, it is very inefficient to do the processing in a centralized computation framework. Hence, Hadoop Framework for distributed computing for Big Video Data is used. First, step in the process is automatic video segmentation and key-frame detection to offer a visual guideline for the video content navigation. In the next step, we extract textual metadata by applying video Optical Character Recognition (OCR) technology on key-frames. The OCR and detected slide text line types are adopted for keyword extraction, by which both video- and segment-level keywords are extracted for content-based video browsing and search. The performance of the indexing process can be improved for a large database by using distributed computing on Hadoop framework.Keywords: video lectures, big video data, video retrieval, hadoop
Procedia PDF Downloads 53318156 Safe and Scalable Framework for Participation of Nodes in Smart Grid Networks in a P2P Exchange of Short-Term Products
Authors: Maciej Jedrzejczyk, Karolina Marzantowicz
Abstract:
Traditional utility value chain is being transformed during last few years into unbundled markets. Increased distributed generation of energy is one of considerable challenges faced by Smart Grid networks. New sources of energy introduce volatile demand response which has a considerable impact on traditional middlemen in E&U market. The purpose of this research is to search for ways to allow near-real-time electricity markets to transact with surplus energy based on accurate time synchronous measurements. A proposed framework evaluates the use of secure peer-2-peer (P2P) communication and distributed transaction ledgers to provide flat hierarchy, and allow real-time insights into present and forecasted grid operations, as well as state and health of the network. An objective is to achieve dynamic grid operations with more efficient resource usage, higher security of supply and longer grid infrastructure life cycle. Methods used for this study are based on comparative analysis of different distributed ledger technologies in terms of scalability, transaction performance, pluggability with external data sources, data transparency, privacy, end-to-end security and adaptability to various market topologies. An intended output of this research is a design of a framework for safer, more efficient and scalable Smart Grid network which is bridging a gap between traditional components of the energy network and individual energy producers. Results of this study are ready for detailed measurement testing, a likely follow-up in separate studies. New platforms for Smart Grid achieving measurable efficiencies will allow for development of new types of Grid KPI, multi-smart grid branches, markets, and businesses.Keywords: autonomous agents, Distributed computing, distributed ledger technologies, large scale systems, micro grids, peer-to-peer networks, Self-organization, self-stabilization, smart grids
Procedia PDF Downloads 30018155 Pareto System of Optimal Placement and Sizing of Distributed Generation in Radial Distribution Networks Using Particle Swarm Optimization
Authors: Sani M. Lawal, Idris Musa, Aliyu D. Usman
Abstract:
The Pareto approach of optimal solutions in a search space that evolved in multi-objective optimization problems is adopted in this paper, which stands for a set of solutions in the search space. This paper aims at presenting an optimal placement of Distributed Generation (DG) in radial distribution networks with an optimal size for minimization of power loss and voltage deviation as well as maximizing voltage profile of the networks. And these problems are formulated using particle swarm optimization (PSO) as a constraint nonlinear optimization problem with both locations and sizes of DG being continuous. The objective functions adopted are the total active power loss function and voltage deviation function. The multiple nature of the problem, made it necessary to form a multi-objective function in search of the solution that consists of both the DG location and size. The proposed PSO algorithm is used to determine optimal placement and size of DG in a distribution network. The output indicates that PSO algorithm technique shows an edge over other types of search methods due to its effectiveness and computational efficiency. The proposed method is tested on the standard IEEE 34-bus and validated with 33-bus test systems distribution networks. Results indicate that the sizing and location of DG are system dependent and should be optimally selected before installing the distributed generators in the system and also an improvement in the voltage profile and power loss reduction have been achieved.Keywords: distributed generation, pareto, particle swarm optimization, power loss, voltage deviation
Procedia PDF Downloads 36418154 A Research Using Remote Monitoring Technology for Pump Output Monitoring in Distributed Fuel Stations in Nigeria
Authors: Ofoegbu Ositadinma Edward
Abstract:
This research paper discusses a web based monitoring system that enables effective monitoring of fuel pump output and sales volume from distributed fuel stations under the domain of a single company/organization. The traditional method of operation by these organizations in Nigeria is non-automated and accounting for dispensed product is usually approximated and manual as there is little or no technology implemented to presently provide information relating to the state of affairs in the station both to on-ground staff and to supervisory staff that are not physically present in the station. This results in unaccountable losses in product and revenue as well as slow decision making. Remote monitoring technology as a vast research field with numerous application areas incorporating various data collation techniques and sensor networks can be applied to provide information relating to fuel pump status in distributed fuel stations reliably. Thus, the proposed system relies upon a microcontroller, keypad and pump to demonstrate the traditional fuel dispenser. A web-enabled PC with an accompanying graphic user interface (GUI) was designed using virtual basic which is connected to the microcontroller via the serial port which is to provide the web implementation.Keywords: fuel pump, microcontroller, GUI, web
Procedia PDF Downloads 43318153 Investigation of Chord Protocol in Peer to Peer Wireless Mesh Network with Mobility
Authors: P. Prasanna Murali Krishna, M. V. Subramanyam, K. Satya Prasad
Abstract:
File sharing in networks are generally achieved using Peer-to-Peer (P2P) applications. Structured P2P approaches are widely used in adhoc networks due to its distributed and scalability features. Efficient mechanisms are required to handle the huge amount of data distributed to all peers. The intrinsic characteristics of P2P system makes for easier content distribution when compared to client-server architecture. All the nodes in a P2P network act as both client and server, thus, distributing data takes lesser time when compared to the client-server method. CHORD protocol is a resource routing based where nodes and data items are structured into a 1- dimensional ring. The structured lookup algorithm of Chord is advantageous for distributed P2P networking applications. Though, structured approach improves lookup performance in a high bandwidth wired network it could contribute to unnecessary overhead in overlay networks leading to degradation of network performance. In this paper, the performance of existing CHORD protocol on Wireless Mesh Network (WMN) when nodes are static and dynamic is investigated.Keywords: wireless mesh network (WMN), structured P2P networks, peer to peer resource sharing, CHORD Protocol, DHT
Procedia PDF Downloads 48018152 Service Interactions Coordination Using a Declarative Approach: Focuses on Deontic Rule from Semantics of Business Vocabulary and Rules Models
Authors: Nurulhuda A. Manaf, Nor Najihah Zainal Abidin, Nur Amalina Jamaludin
Abstract:
Coordinating service interactions are a vital part of developing distributed applications that are built up as networks of autonomous participants, e.g., software components, web services, online resources, involve a collaboration between a diverse number of participant services on different providers. The complexity in coordinating service interactions reflects how important the techniques and approaches require for designing and coordinating the interaction between participant services to ensure the overall goal of a collaboration between participant services is achieved. The objective of this research is to develop capability of steering a complex service interaction towards a desired outcome. Therefore, an efficient technique for modelling, generating, and verifying the coordination of service interactions is developed. The developed model describes service interactions using service choreographies approach and focusing on a declarative approach, advocating an Object Management Group (OMG) standard, Semantics of Business Vocabulary and Rules (SBVR). This model, namely, SBVR model for service choreographies focuses on a declarative deontic rule expressing both obligation and prohibition, which can be more useful in working with coordinating service interactions. The generated SBVR model is then be formulated and be transformed into Alloy model using Alloy Analyzer for verifying the generated SBVR model. The transformation of SBVR into Alloy allows to automatically generate the corresponding coordination of service interactions (service choreography), hence producing an immediate instance of execution that satisfies the constraints of the specification and verifies whether a specific request can be realised in the given choreography in the generated choreography.Keywords: service choreography, service coordination, behavioural modelling, complex interactions, declarative specification, verification, model transformation, semantics of business vocabulary and rules, SBVR
Procedia PDF Downloads 15418151 Seasonal Short-Term Effect of Air Pollution on Cardiovascular Mortality in Belgium
Authors: Natalia Bustos Sierra, Katrien Tersago
Abstract:
It is currently proven that both extremes of temperature are associated with increased mortality and that air pollution is associated with temperature. This relationship is complex, and in countries with important seasonal variations in weather such as Belgium, some effects can appear as non-significant when the analysis is done over the entire year. We, therefore, analyzed the effect of short-term outdoor air pollution exposure on cardiovascular mortality during the warmer and colder months separately. We used daily cardiovascular deaths from acute cardiovascular diagnostics according to the International Classification of Diseases, 10th Revision (ICD-10: I20-I24, I44-I49, I50, I60-I66) during the period 2008-2013. The environmental data were population-weighted concentrations of particulates with an aerodynamic diameter less than 10 µm (PM₁₀) and less than 2.5 µm (PM₂.₅) (daily average), nitrogen dioxide (NO₂) (daily maximum of the hourly average) and ozone (O₃) (daily maximum of the 8-hour running mean). A Generalized linear model was applied adjusting for the confounding effect of season, temperature, dew point temperature, the day of the week, public holidays and the incidence of influenza-like illness (ILI) per 100,000 inhabitants. The relative risks (RR) were calculated for an increase of one interquartile range (IQR) of the air pollutant (μg/m³). These were presented for the four hottest months (June, July, August, September) and coldest months (November, December, January, February) in Belgium. We applied both individual lag model and unconstrained distributed lag model methods. The cumulative effect of a four-day exposure (day of exposure and three consecutive days) was calculated from the unconstrained distributed lag model. The IQR for PM₁₀, PM₂.₅, NO₂, and O₃ were respectively 8.2, 6.9, 12.9 and 25.5 µg/m³ during warm months and 18.8, 17.6, 18.4 and 27.8 µg/m³ during cold months. The association with CV mortality was statistically significant for the four pollutants during warm months and only for NO₂ during cold months. During the warm months, the cumulative effect of an IQR increase of ozone for the age groups 25-64, 65-84 and 85+ was 1.066 (95%CI: 1.002-1.135), 1.041 (1.008-1.075) and 1.036 (1.013-1.058) respectively. The cumulative effect of an IQR increase of NO₂ for the age group 65-84 was 1.066 (1.020-1.114) during warm months and 1.096 (1.030-1.166) during cold months. The cumulative effect of an IQR increase of PM₁₀ during warm months reached 1.046 (1.011-1.082) and 1.038 (1.015-1.063) for the age groups 65-84 and 85+ respectively. Similar results were observed for PM₂.₅. The short-term effect of air pollution on cardiovascular mortality is greater during warm months for lower pollutant concentrations compared to cold months. Spending more time outside during warm months increases population exposure to air pollution and can, therefore, be a confounding factor for this association. Age can also affect the length of time spent outdoors and the type of physical activity exercised. This study supports the deleterious effect of air pollution on cardiovascular mortality (CV) which varies according to season and age groups in Belgium. Public health measures should, therefore, be adapted to seasonality.Keywords: air pollution, cardiovascular, mortality, season
Procedia PDF Downloads 16518150 A Reduced Distributed Sate Space for Modular Petri Nets
Authors: Sawsen Khlifa, Chiheb AMeur Abid, Belhassan Zouari
Abstract:
Modular verification approaches have been widely attempted to cope with the well known state explosion problem. This paper deals with the modular verification of modular Petri nets. We propose a reduced version for the modular state space of a given modular Petri net. The new structure allows the creation of smaller modular graphs. Each one draws the behavior of the corresponding module and outlines some global information. Hence, this version helps to overcome the explosion problem and to use less memory space. In this condensed structure, the verification of some generic properties concerning one module is limited to the exploration of its associated graph.Keywords: distributed systems, modular verification, petri nets, state space explosition
Procedia PDF Downloads 11518149 Reviewing Privacy Preserving Distributed Data Mining
Authors: Sajjad Baghernezhad, Saeideh Baghernezhad
Abstract:
Nowadays considering human involved in increasing data development some methods such as data mining to extract science are unavoidable. One of the discussions of data mining is inherent distribution of the data usually the bases creating or receiving such data belong to corporate or non-corporate persons and do not give their information freely to others. Yet there is no guarantee to enable someone to mine special data without entering in the owner’s privacy. Sending data and then gathering them by each vertical or horizontal software depends on the type of their preserving type and also executed to improve data privacy. In this study it was attempted to compare comprehensively preserving data methods; also general methods such as random data, coding and strong and weak points of each one are examined.Keywords: data mining, distributed data mining, privacy protection, privacy preserving
Procedia PDF Downloads 52518148 Cross-Layer Design of Event-Triggered Adaptive OFDMA Resource Allocation Protocols with Application to Vehicle Clusters
Authors: Shaban Guma, Naim Bajcinca
Abstract:
We propose an event-triggered algorithm for the solution of a distributed optimization problem by means of the projected subgradient method. Thereby, we invoke an OFDMA resource allocation scheme by applying an event-triggered sensitivity analysis at the access point. The optimal resource assignment of the subcarriers to the involved wireless nodes is carried out by considering the sensitivity analysis of the overall objective function as defined by the control of vehicle clusters with respect to the information exchange between the nodes.Keywords: consensus, cross-layer, distributed, event-triggered, multi-vehicle, protocol, resource, OFDMA, wireless
Procedia PDF Downloads 33118147 A Petri Net Model to Obtain the Throughput of Unreliable Production Lines in the Buffer Allocation Problem
Authors: Joselito Medina-Marin, Alexandr Karelin, Ana Tarasenko, Juan Carlos Seck-Tuoh-Mora, Norberto Hernandez-Romero, Eva Selene Hernandez-Gress
Abstract:
A production line designer faces with several challenges in manufacturing system design. One of them is the assignment of buffer slots in between every machine of the production line in order to maximize the throughput of the whole line, which is known as the Buffer Allocation Problem (BAP). The BAP is a combinatorial problem that depends on the number of machines and the total number of slots to be distributed on the production line. In this paper, we are proposing a Petri Net (PN) Model to obtain the throughput in unreliable production lines, based on PN mathematical tools and the decomposition method. The results obtained by this methodology are similar to those presented in previous works, and the number of machines is not a hard restriction.Keywords: buffer allocation problem, Petri Nets, throughput, production lines
Procedia PDF Downloads 30718146 Behavioural Intention to Use Learning Management System (LMS) among Postgraduate Students: An Application of Utaut Model
Authors: Kamaludeen Samaila, Khashyaullah Abdulfattah, Fahimi Ahmad Bin Amir
Abstract:
The study was conducted to examine the relationship between selected factors (performance expectancy, effort expectancy, social influence and facilitating condition) and students’ intention to use the learning management system (LMS), as well as investigating the factors predicting students’ intention to use the LMS. The study was specifically conducted at the Faculty of Educational Study of University Putra Malaysia. Questionnaires were distributed to 277 respondents using a random sampling technique. SPSS Version 22 was employed in analyzing the data; the findings of this study indicated that performance expectancy (r = .69, p < .01), effort expectancy (r=.60, p < .01), social influence (r = .61, p < .01), and facilitating condition (r=.42, p < .01), were significantly related to students’ intention to use the LMS. In addition, the result also revealed that performance expectancy (β = .436, p < .05), social influence (β=.232, p < .05), and effort expectancy (β = .193, p < .05) were strong predictors of students’ intention to use the LMS. The analysis further indicated that (R2) is 0.054 which means that 54% of variation in the dependent variable is explained by the entire predictor variables entered into the regression model. Understanding the factors that affect students’ intention to use the LMS could help the lecturers, LMS managers and university management to develop the policies that may attract students to use the LMS.Keywords: LMS, postgraduate students, PutraBlas, students’ intention, UPM, UTAUT model
Procedia PDF Downloads 51018145 A Survey on the Requirements of University Course Timetabling
Authors: Nurul Liyana Abdul Aziz, Nur Aidya Hanum Aizam
Abstract:
Course timetabling problems occur every semester in a university which includes the allocation of resources (subjects, lecturers and students) to a number of fixed rooms and timeslots. The assignment is carried out in a way such that there are no conflicts within rooms, students and lecturers, as well as fulfilling a range of constraints. The constraints consist of rules and policies set up by the universities as well as lecturers’ and students’ preferences of courses to be allocated in specific timeslots. This paper specifically focuses on the preferences of the course timetabling problem in one of the public universities in Malaysia. The demands will be considered into our existing mathematical model to make it more generalized and can be used widely. We have distributed questionnaires to a number of lecturers and students of the university to investigate their demands and preferences for their desired course timetable. We classify the preferences thus converting them to construct one mathematical model that can produce such timetable.Keywords: university course timetabling problem, integer programming, preferences, constraints
Procedia PDF Downloads 36618144 Design of Geochemical Maps of Industrial City Using Gradient Boosting and Geographic Information System
Authors: Ruslan Safarov, Zhanat Shomanova, Yuri Nossenko, Zhandos Mussayev, Ayana Baltabek
Abstract:
Geochemical maps of distribution of polluting elements V, Cr, Mn, Co, Ni, Cu, Zn, Mo, Cd, Pb on the territory of the Pavlodar city (Kazakhstan), which is an industrial hub were designed. The samples of soil were taken from 100 locations. Elemental analysis has been performed using XRF. The obtained data was used for training of the computational model with gradient boosting algorithm. The optimal parameters of model as well as the loss function were selected. The computational model was used for prediction of polluting elements concentration for 1000 evenly distributed points. Based on predicted data geochemical maps were created. Additionally, the total pollution index Zc was calculated for every from 1000 point. The spatial distribution of the Zc index was visualized using GIS (QGIS). It was calculated that the maximum coverage area of the territory of the Pavlodar city belongs to the moderately hazardous category (89.7%). The visualization of the obtained data allowed us to conclude that the main source of contamination goes from the industrial zones where the strategic metallurgical and refining plants are placed.Keywords: Pavlodar, geochemical map, gradient boosting, CatBoost, QGIS, spatial distribution, heavy metals
Procedia PDF Downloads 8218143 Potential Impact of Climate Change on Suspended Sediment Changes in Mekong River Basin
Authors: Zuliziana Suif, Nordila Ahmad, Sengheng Hul
Abstract:
This paper evaluates the impact of climate change on suspended sediment changes in the Mekong River Basin. In this study, the distributed process-based sediment transport model is used to examine the potential impact of future climate on suspended sediment dynamic changes in the Mekong River Basin. To this end, climate scenarios from two General Circulation Model (GCMs) were considered in the scenario analysis. The simulation results show that the sediment load and concentration shows 0.64% to 69% increase in the near future (2041-2050) and 2.5% to 95% in the far future (2090- 2099). As the projected climate change impact on sediment varies remarkably between the different climate models, the uncertainty should be taken into account in sediment management. Overall, the changes in sediment load and concentration can have a great implication for related sediment management.Keywords: climate change, suspended sediment, Mekong River Basin, GCMs
Procedia PDF Downloads 44318142 Advanced Hybrid Particle Swarm Optimization for Congestion and Power Loss Reduction in Distribution Networks with High Distributed Generation Penetration through Network Reconfiguration
Authors: C. Iraklis, G. Evmiridis, A. Iraklis
Abstract:
Renewable energy sources and distributed power generation units already have an important role in electrical power generation. A mixture of different technologies penetrating the electrical grid, adds complexity in the management of distribution networks. High penetration of distributed power generation units creates node over-voltages, huge power losses, unreliable power management, reverse power flow and congestion. This paper presents an optimization algorithm capable of reducing congestion and power losses, both described as a function of weighted sum. Two factors that describe congestion are being proposed. An upgraded selective particle swarm optimization algorithm (SPSO) is used as a solution tool focusing on the technique of network reconfiguration. The upgraded SPSO algorithm is achieved with the addition of a heuristic algorithm specializing in reduction of power losses, with several scenarios being tested. Results show significant improvement in minimization of losses and congestion while achieving very small calculation times.Keywords: congestion, distribution networks, loss reduction, particle swarm optimization, smart grid
Procedia PDF Downloads 44518141 An Econometric Analysis of the Impacts of Inflation on the Economic Growth of South Africa
Authors: Gisele Mah, Paul Saah
Abstract:
The rising rates of inflation are hindering economic growth in developing nations. Hence, this study investigated the effects of inflation rates on the economic growth of South Africa using the secondary time series data from 1987 to 2022. The main objectives of this study were to investigate the long run relationship between inflation and economic growth, and also to determine the causality direction between these two variables. The study utilized the Autoregressive Distributed Lag (ARDL) bounds test of co-integration to investigate whether there is a long-run relationship between inflation and economic growth. The Pairwise Granger causality approach was employed to determine the second objective, which is the direction of causality. The study discovered only one co-integration relationship between our variables and it was between inflation and economic growth. The results showed that there is a negative and significant relationship between inflation and economic growth. There appeared to be a positive and significant relationship between economic growth and exchange rate. The interest rates have shown to be negative and insignificant in explaining economic growth. The study also established that inflation does Granger cause economic growth which is given as GDP. Similarly, the study discovered that inflation Granger causes exchange rates. Therefore, the study recommends that inflation should be decreased in South Africa, in order for economic growth to increase. Contrary, this study recommends that South Africa should increase its exchange rates, in order for economic growth to also increase.Keywords: inflation rate, economic growth, South Africa, autoregressive distributed lag model
Procedia PDF Downloads 4818140 Unveiling the Mystery of Innovation in Higher Education Institutions
Authors: Ana Martins, Isabel Martins
Abstract:
The purpose of this research is to ascertain whether students at HEIs cultivate distributed leadership and higher-level skills to inspire knowledge creation. Critical reflection of extant literature illustrates the need for a culture of innovation in organizational sustainability. New age leadership behaviors harmonize innovation. The leadership self-efficacy construct supports organizational learning. This exploratory study applies the pragmatic paradigm methodology using the survey research method for primary data collection. A questionnaire was distributed to a sample of university students based in the Southern Anatolian region of Turkey, from both under and postgraduate Business degree programs. An analysis of the findings reveals a greater connection in influencing behavior relying more on the task-centered perspective rather than with the people perspective. These results reveal the need for HEIs to instill a humanistic perspective in curricula enabling graduates to be capable leaders with the awareness soft skills to energize creativity and innovation. A limitation of this research is that one university makes it difficult to generalize to a broader population. This study is of added value for scholars and organizations in the current knowledge and innovation economy.Keywords: distributed leadership, exploration, higher education institutions, innovation, knowledge creation, learning, self-efficacy
Procedia PDF Downloads 19518139 Calibration and Validation of ArcSWAT Model for Estimation of Surface Runoff and Sediment Yield from Dhangaon Watershed
Authors: M. P. Tripathi, Priti Tiwari
Abstract:
Soil and Water Assessment Tool (SWAT) is a distributed parameter continuous time model and was tested on daily and fortnightly basis for a small agricultural watershed (Dhangaon) of Chhattisgarh state in India. The SWAT model recently interfaced with ArcGIS and called as ArcSWAT. The watershed and sub-watershed boundaries, drainage networks, slope and texture maps were generated in the environment of ArcGIS of ArcSWAT. Supervised classification method was used for land use/cover classification from satellite imageries of the years 2009 and 2012. Manning's roughness coefficient 'n' for overland flow and channel flow and Fraction of Field Capacity (FFC) were calibrated for monsoon season of the years 2009 and 2010. The model was validated on a daily basis for the years 2011 and 2012 by using the observed daily rainfall and temperature data. Calibration and validation results revealed that the model was predicting the daily surface runoff and sediment yield satisfactorily. Sensitivity analysis showed that the annual sediment yield was inversely proportional to the overland and channel 'n' values whereas; annual runoff and sediment yields were directly proportional to the FFC. The model was also tested (calibrated and validated) for the fortnightly runoff and sediment yield for the year 2009-10 and 2011-12, respectively. Simulated values of fortnightly runoff and sediment yield for the calibration and validation years compared well with their observed counterparts. The calibration and validation results revealed that the ArcSWAT model could be used for identification of critical sub-watershed and for developing management scenarios for the Dhangaon watershed. Further, the model should be tested for simulating the surface runoff and sediment yield using generated rainfall and temperature before applying it for developing the management scenario for the critical or priority sub-watersheds.Keywords: watershed, hydrologic and water quality, ArcSWAT model, remote sensing, GIS, runoff and sediment yield
Procedia PDF Downloads 37918138 Stochastic Control of Decentralized Singularly Perturbed Systems
Authors: Walid S. Alfuhaid, Saud A. Alghamdi, John M. Watkins, M. Edwin Sawan
Abstract:
Designing a controller for stochastic decentralized interconnected large scale systems usually involves a high degree of complexity and computation ability. Noise, observability, and controllability of all system states, connectivity, and channel bandwidth are other constraints to design procedures for distributed large scale systems. The quasi-steady state model investigated in this paper is a reduced order model of the original system using singular perturbation techniques. This paper results in an optimal control synthesis to design an observer based feedback controller by standard stochastic control theory techniques using Linear Quadratic Gaussian (LQG) approach and Kalman filter design with less complexity and computation requirements. Numerical example is given at the end to demonstrate the efficiency of the proposed method.Keywords: decentralized, optimal control, output, singular perturb
Procedia PDF Downloads 37018137 Analyzing the Market Growth in Application Programming Interface Economy Using Time-Evolving Model
Authors: Hiroki Yoshikai, Shin’ichi Arakawa, Tetsuya Takine, Masayuki Murata
Abstract:
API (Application Programming Interface) economy is expected to create new value by converting corporate services such as information processing and data provision into APIs and using these APIs to connect services. Understanding the dynamics of a market of API economy under the strategies of participants is crucial to fully maximize the values of the API economy. To capture the behavior of a market in which the number of participants changes over time, we present a time-evolving market model for a platform in which API providers who provide APIs to service providers participate in addition to service providers and consumers. Then, we use the market model to clarify the role API providers play in expanding market participants and forming ecosystems. The results show that the platform with API providers increased the number of market participants by 67% and decreased the cost to develop services by 25% compared to the platform without API providers. Furthermore, during the expansion phase of the market, it is found that the profits of participants are mostly the same when 70% of the revenue from consumers is distributed to service providers and API providers. It is also found that when the market is mature, the profits of the service provider and API provider will decrease significantly due to their competition, and the profit of the platform increases.Keywords: API economy, ecosystem, platform, API providers
Procedia PDF Downloads 9118136 Signature Verification System for a Banking Business Process Management
Authors: A. Rahaf, S. Liyakathunsia
Abstract:
In today’s world, unprecedented operational pressure is faced by banks that test the efficiency, effectiveness, and agility of their business processes. In a typical banking process, a person’s authorization is usually based on his signature on most all of the transactions. Signature verification is considered as one of the highly significant information needed for any bank document processing. Banks usually use Signature Verification to authenticate the identity of individuals. In this paper, a business process model has been proposed in order to increase the quality of the verification process and to reduce time and needed resources. In order to understand the current process, a survey has been conducted and distributed among bank employees. After analyzing the survey, a process model has been created using Bizagi modeler which helps in simulating the process after assigning time and cost of it. The outcomes show that the automation of signature verification process is highly recommended for a banking business process.Keywords: business process management, process modeling, quality, Signature Verification
Procedia PDF Downloads 42518135 Species Distribution Model for Zanthoxylum Rhetsa Genus in Thailand
Authors: Yosiya Chanta, Jantrararuk Tovaranont
Abstract:
Species distribution model (SDMs) is one of the powerful tools used to create a suitability map used to predict and address ecology and conservation approaches. MaxEnt is a tool used among SDMs that is highly popular because it only uses presence data. Zanthoxylum rhetsa has more than 200 species distributed in the tropics. Most commonly found in cooler forest environments, there are 8-9 species found in Thailand. In northern Thailand, 3 varieties are commonly grown: Zanthoxylum myriacanthum, Zanthoxylum rhetsa and Zanthoxylum armatum. In the northern regions, these varieties are mainly used as a spice and as a cooking ingredient. MaxEnt has been used in this study to predict potential habitats for these Zanthoxylums in current and future times (2041and 2060). Suitable habitats are predicted using data from the EC-Earth3-Veg general circulation model with 19 climatic variables. The results indicate that the suitability of future habitats of Zanthoxylum rhetsa may expand into the lower northern part of Thailand. The habitat suitability map obtained from the MaxEnt tool shows that the Precipitation of Wettest Quarter (Bio16) is the most important climatic variable influencing the current and future spread of Zanthoxylum rhetsa.Keywords: MaxEnt, Zanthoxylum rhets, species distribution modelling, climate change
Procedia PDF Downloads 9818134 Mathematical Model to Quantify the Phenomenon of Democracy
Authors: Mechlouch Ridha Fethi
Abstract:
This paper presents a recent mathematical model in political sciences concerning democracy. The model is represented by a logarithmic equation linking the Relative Index of Democracy (RID) to Participation Ratio (PR). Firstly the meanings of the different parameters of the model were presented; and the variation curve of the RID according to PR with different critical areas was discussed. Secondly, the model was applied to a virtual group where we show that the model can be applied depending on the gender. Thirdly, it was observed that the model can be extended to different language models of democracy and that little use to assess the state of democracy for some International organizations like UNO.Keywords: democracy, mathematic, modelization, quantification
Procedia PDF Downloads 36818133 Application of Simulated Annealing to Threshold Optimization in Distributed OS-CFAR System
Authors: L. Abdou, O. Taibaoui, A. Moumen, A. Talib Ahmed
Abstract:
This paper proposes an application of the simulated annealing to optimize the detection threshold in an ordered statistics constant false alarm rate (OS-CFAR) system. Using conventional optimization methods, such as the conjugate gradient, can lead to a local optimum and lose the global optimum. Also for a system with a number of sensors that is greater than or equal to three, it is difficult or impossible to find this optimum; Hence, the need to use other methods, such as meta-heuristics. From a variety of meta-heuristic techniques, we can find the simulated annealing (SA) method, inspired from a process used in metallurgy. This technique is based on the selection of an initial solution and the generation of a near solution randomly, in order to improve the criterion to optimize. In this work, two parameters will be subject to such optimisation and which are the statistical order (k) and the scaling factor (T). Two fusion rules; “AND” and “OR” were considered in the case where the signals are independent from sensor to sensor. The results showed that the application of the proposed method to the problem of optimisation in a distributed system is efficiency to resolve such problems. The advantage of this method is that it allows to browse the entire solutions space and to avoid theoretically the stagnation of the optimization process in an area of local minimum.Keywords: distributed system, OS-CFAR system, independent sensors, simulating annealing
Procedia PDF Downloads 49718132 Prosperous Digital Image Watermarking Approach by Using DCT-DWT
Authors: Prabhakar C. Dhavale, Meenakshi M. Pawar
Abstract:
In this paper, everyday tons of data is embedded on digital media or distributed over the internet. The data is so distributed that it can easily be replicated without error, putting the rights of their owners at risk. Even when encrypted for distribution, data can easily be decrypted and copied. One way to discourage illegal duplication is to insert information known as watermark, into potentially valuable data in such a way that it is impossible to separate the watermark from the data. These challenges motivated researchers to carry out intense research in the field of watermarking. A watermark is a form, image or text that is impressed onto paper, which provides evidence of its authenticity. Digital watermarking is an extension of the same concept. There are two types of watermarks visible watermark and invisible watermark. In this project, we have concentrated on implementing watermark in image. The main consideration for any watermarking scheme is its robustness to various attacksKeywords: watermarking, digital, DCT-DWT, security
Procedia PDF Downloads 42218131 A New Method Presentation for Locating Fault in Power Distribution Feeders Considering DG
Authors: Rahman Dashti, Ehsan Gord
Abstract:
In this paper, an improved impedance based fault location method is proposed. In this method, online fault locating is performed using voltage and current information at the beginning of the feeder. Determining precise fault location in a short time increases reliability and efficiency of the system. The proposed method utilizes information about main component of voltage and current at the beginning of the feeder and distributed generation unit (DGU) in order to precisely locate different faults in acceptable time. To evaluate precision and accuracy of the proposed method, a 13-node is simulated and tested using MATLAB.Keywords: distribution network, fault section determination, distributed generation units, distribution protection equipment
Procedia PDF Downloads 401