Search results for: e-content producing algorithm
1751 An Experimental Testbed Using Virtual Containers for Distributed Systems
Authors: Parth Patel, Ying Zhu
Abstract:
Distributed systems have become ubiquitous, and they continue their growth through a range of services. With advances in resource virtualization technology such as Virtual Machines (VM) and software containers, developers no longer require high-end servers to test and develop distributed software. Even in commercial production, virtualization has streamlined the process of rapid deployment and service management. This paper introduces a distributed systems testbed that utilizes virtualization to enable distributed systems development on commodity computers. The testbed can be used to develop new services, implement theoretical distributed systems concepts for understanding, and experiment with virtual network topologies. We show its versatility through two case studies that utilize the testbed for implementing a theoretical algorithm and developing our own methodology to find high-risk edges. The results of using the testbed for these use cases have proven the effectiveness and versatility of this testbed across a range of scenarios.Keywords: distributed systems, experimental testbed, peer-to-peer networks, virtual container technology
Procedia PDF Downloads 1471750 Simulation of Wave Propagation in Multiphase Medium
Authors: Edip Kemal, Sheshov Vlatko, Bojadjieva Julijana, Bogdanovic ALeksandra, Gjorgjeska Irena
Abstract:
The wave propagation phenomenon in porous domains is of great importance in the field of geotechnical earthquake engineering. In these kinds of problems, the elastic waves propagate from the interior to the exterior domain and require special treatment at the computational level since apart from displacement in the solid-state there is a p-wave that takes place in the pore water phase. In this paper, a study on the implementation of multiphase finite elements is presented. The proposed algorithm is implemented in the ANSYS finite element software and tested on one-dimensional wave propagation considering both pore pressure wave propagation and displacement fields. In the simulation of porous media such as soils, the behavior is governed largely by the interaction of the solid skeleton with water and/or air in the pores. Therefore, coupled problems of fluid flow and deformation of the solid skeleton are considered in a detailed way.Keywords: wave propagation, multiphase model, numerical methods, finite element method
Procedia PDF Downloads 1661749 Applying Spanning Tree Graph Theory for Automatic Database Normalization
Authors: Chetneti Srisa-an
Abstract:
In Knowledge and Data Engineering field, relational database is the best repository to store data in a real world. It has been using around the world more than eight decades. Normalization is the most important process for the analysis and design of relational databases. It aims at creating a set of relational tables with minimum data redundancy that preserve consistency and facilitate correct insertion, deletion, and modification. Normalization is a major task in the design of relational databases. Despite its importance, very few algorithms have been developed to be used in the design of commercial automatic normalization tools. It is also rare technique to do it automatically rather manually. Moreover, for a large and complex database as of now, it make even harder to do it manually. This paper presents a new complete automated relational database normalization method. It produces the directed graph and spanning tree, first. It then proceeds with generating the 2NF, 3NF and also BCNF normal forms. The benefit of this new algorithm is that it can cope with a large set of complex function dependencies.Keywords: relational database, functional dependency, automatic normalization, primary key, spanning tree
Procedia PDF Downloads 3531748 The Development of GPS Buoy for Ocean Surface Monitoring: Initial Results
Authors: Anuar Mohd Salleh, Mohd Effendi Daud
Abstract:
This study presents a kinematic positioning approach which is use the GPS buoy for precise ocean surface monitoring. A GPS buoy data from two experiments have been processed using a precise, medium-range differential kinematic technique. In each case the data were collected for more than 24 hours at nearby coastal site at a high rate (1 Hz), along with measurements from neighboring tidal stations, to verify the estimated sea surface heights. Kinematic coordinates of GPS buoy were estimated using the epoch-wise pre-elimination and the backward substitution algorithm. Test results show the centimeter level accuracy in sea surface height determination can be successfully achieved using proposed technique. The centimeter level agreement between two methods also suggests the possibility of using this inexpensive and more flexible GPS buoy equipment to enhance (or even replace) the current use of tidal gauge stations.Keywords: global positioning system, kinematic GPS, sea surface height, GPS buoy, tide gauge
Procedia PDF Downloads 5461747 Determination of Frequency Relay Setting during Distributed Generators Islanding
Authors: Tarek Kandil, Ameen Ali
Abstract:
Distributed generation (DG) has recently gained a lot of momentum in power industry due to market deregulation and environmental concerns. One of the most technical challenges facing DGs is islanding of distributed generators. The current industry practice is to disconnect all distributed generators immediately after the occurrence of islands within 200 to 350 ms after loss of main supply. To achieve such goal, each DG must be equipped with an islanding detection device. Frequency relays are one of the most commonly used loss of mains detection method. However, distribution utilities may be faced with concerns related to false operation of these frequency relays due to improper settings. The commercially available frequency relays are considering standard tight setting. This paper investigates some factors related to relays internal algorithm that contribute to their different operating responses. Further, the relay operation in the presence of multiple distributed at the same network is analyzed. Finally, the relay setting can be accurately determined based on these investigation and analysis.Keywords: frequency relay, distributed generation, islanding detection, relay setting
Procedia PDF Downloads 5341746 Catalytic Hydrothermal Decarboxylation of Lipid from Activated Sludge for Renewable Diesel Production
Authors: Ifeanyichukwu Edeh, Tim Overton, Steve Bowra
Abstract:
Currently biodiesel is produced from plant oils or animal’s fats by a liquid-phase catalysed transesterification process at low temperature. Although biodiesel is renewable and to a large extent sustainable, inherent properties such as poor cold flow, low oxidation stability, low cetane value restrict application to blends with fossil fuels. An alternative to biodiesel is renewable diesel produced by catalytic hydrotreating of oils and fats and is considered a drop in fuel because its properties are similar to petroleum diesel. In addition to developing alternative productions routes there is continued interest in reducing the cost of the feed stock, waste cooking oils and fats are increasingly used as the feedstocks due to low cost. However, use of oils and fat are highly adulterated resulting in high free fatty acid content which turn impacts on the efficiency of FAME production. Therefore, in light of the need to develop, alternative lipid feed stocks and related efficient catalysis the present study investigates the potential of producing renewable diesel from the lipids-extracted from activated sludge, a waste water treatment by-product, through catalytic hydrothermal decarboxylation. The microbial lipids were first extracted from the activated sludge using the Folch et al method before hydrothermal decarboxylation reactions were carried out using palladium (Pd/C) and platinum (Pt/C) on activated carbon as the catalysts in a batch reactor. The impact of three temperatures 290, 300, 330 °C and residence time between 30 min and 4hrs was assessed. At the end of the reaction, the products were recovered using organic solvents and characterized using gas chromatography (GC). The principle products of the reaction were pentadecane and heptadecane. The highest yields of pentadecane and heptadecane from lipid-extract were 23.23% and 15.21%, respectively. These yields were obtained at 290 °C and residence time 1h using Pt/C. To the best of our knowledge, the current work is the first investigation on the hydrothermal decarboxylation of lipid-extract from activated sludge.Keywords: activated sludge, lipid, hydrothermal decarboxylation, renewable diesel
Procedia PDF Downloads 3191745 Optimal Design of Reference Node Placement for Wireless Indoor Positioning Systems in Multi-Floor Building
Authors: Kittipob Kondee, Chutima Prommak
Abstract:
In this paper, we propose an optimization technique that can be used to optimize the placements of reference nodes and improve the location determination performance for the multi-floor building. The proposed technique is based on Simulated Annealing algorithm (SA) and is called MSMR-M. The performance study in this work is based on simulation. We compare other node-placement techniques found in the literature with the optimal node-placement solutions obtained from our optimization. The results show that using the optimal node-placement obtained by our proposed technique can improve the positioning error distances up to 20% better than those of the other techniques. The proposed technique can provide an average error distance within 1.42 meters.Keywords: indoor positioning system, optimization system design, multi-floor building, wireless sensor networks
Procedia PDF Downloads 2471744 Interactions and Integration: Implications of Victim-Agent Portrayals for Refugees and Asylum Seekers in Germany
Authors: Denise Muro
Abstract:
Conflict in Syria, producing over 11 million displaced persons, has incited global attention to displacement. Although neighboring countries have borne the largest part of the displacement burden, due to the influx of refugees into Europe, the so-called ‘refugee crisis’ is taking place on two fronts: Syria’s neighboring countries, with millions of refugees, and Europe, a destination goal for so many that European states face unprecedented challenges. With increasing attention to displacement, forcibly displaced persons are consistently portrayed as either un-agentic victims, or as dangerous free agents. Recognizing that these dominant portrayals involve discourses of power and inequality, this research investigates the extent to which this victim-agent dichotomy affects refugees and organizations that work closely with them during initial integration processes in Berlin, Germany. The research measures initial integration based on German policy measures regarding integration juxtaposed with the way refugees and those who work with them understand integration. Additionally, the study examines day-to-day interactions of refugees in Germany as a way to gauge social integration in a bottom-up approach. This study involved a discourse analysis of portrayals of refugees and participant observation and interviews with refugees and those who work closely with them, which took place during fieldwork in Berlin in the summer of 2016. Germany is unique regarding their migration history and lack of successful integration, in part due to the persistent refrain, ‘Wir sind kein einwanderungsland’ (‘We are not an immigration country’). Still, their accepted asylum seeker population has grown exponentially in the past few years. Findings suggest that the victim-agent dichotomy is present and impactful in the process of refugees entering and integrating into Germany. Integration is hindered due to refugees either being patronized or criminalized to such an extent that, despite being constantly told that they must integrate, they cannot become part of German society.Keywords: discourse analysis, Germany, integration, refugee crisis
Procedia PDF Downloads 2741743 Efficient Ground Targets Detection Using Compressive Sensing in Ground-Based Synthetic-Aperture Radar (SAR) Images
Authors: Gherbi Nabil
Abstract:
Detection of ground targets in SAR radar images is an important area for radar information processing. In the literature, various algorithms have been discussed in this context. However, most of them are of low robustness and accuracy. To this end, we discuss target detection in SAR images based on compressive sensing. Firstly, traditional SAR image target detection algorithms are discussed, and their limitations are highlighted. Secondly, a compressive sensing method is proposed based on the sparsity of SAR images. Next, the detection problem is solved using Multiple Measurements Vector configuration. Furthermore, a robust Alternating Direction Method of Multipliers (ADMM) is developed to solve the optimization problem. Finally, the detection results obtained using raw complex data are presented. Experimental results on real SAR images have verified the effectiveness of the proposed algorithm.Keywords: compressive sensing, raw complex data, synthetic aperture radar, ADMM
Procedia PDF Downloads 221742 Solving Single Machine Total Weighted Tardiness Problem Using Gaussian Process Regression
Authors: Wanatchapong Kongkaew
Abstract:
This paper proposes an application of probabilistic technique, namely Gaussian process regression, for estimating an optimal sequence of the single machine with total weighted tardiness (SMTWT) scheduling problem. In this work, the Gaussian process regression (GPR) model is utilized to predict an optimal sequence of the SMTWT problem, and its solution is improved by using an iterated local search based on simulated annealing scheme, called GPRISA algorithm. The results show that the proposed GPRISA method achieves a very good performance and a reasonable trade-off between solution quality and time consumption. Moreover, in the comparison of deviation from the best-known solution, the proposed mechanism noticeably outperforms the recently existing approaches.Keywords: Gaussian process regression, iterated local search, simulated annealing, single machine total weighted tardiness
Procedia PDF Downloads 3091741 A Security Cloud Storage Scheme Based Accountable Key-Policy Attribute-Based Encryption without Key Escrow
Authors: Ming Lun Wang, Yan Wang, Ning Ruo Sun
Abstract:
With the development of cloud computing, more and more users start to utilize the cloud storage service. However, there exist some issues: 1) cloud server steals the shared data, 2) sharers collude with the cloud server to steal the shared data, 3) cloud server tampers the shared data, 4) sharers and key generation center (KGC) conspire to steal the shared data. In this paper, we use advanced encryption standard (AES), hash algorithms, and accountable key-policy attribute-based encryption without key escrow (WOKE-AKP-ABE) to build a security cloud storage scheme. Moreover, the data are encrypted to protect the privacy. We use hash algorithms to prevent the cloud server from tampering the data uploaded to the cloud. Analysis results show that this scheme can resist conspired attacks.Keywords: cloud storage security, sharing storage, attributes, Hash algorithm
Procedia PDF Downloads 3901740 Estimation of Structural Parameters in Time Domain Using One Dimensional Piezo Zirconium Titanium Patch Model
Authors: N. Jinesh, K. Shankar
Abstract:
This article presents a method of using the one dimensional piezo-electric patch on beam model for structural identification. A hybrid element constituted of one dimensional beam element and a PZT sensor is used with reduced material properties. This model is convenient and simple for identification of beams. Accuracy of this element is first verified against a corresponding 3D finite element model (FEM). The structural identification is carried out as an inverse problem whereby parameters are identified by minimizing the deviation between the predicted and measured voltage response of the patch, when subjected to excitation. A non-classical optimization algorithm Particle Swarm Optimization is used to minimize this objective function. The signals are polluted with 5% Gaussian noise to simulate experimental noise. The proposed method is applied on beam structure and identified parameters are stiffness and damping. The model is also validated experimentally.Keywords: inverse problem, particle swarm optimization, PZT patches, structural identification
Procedia PDF Downloads 3101739 Switching to the Latin Alphabet in Kazakhstan: A Brief Overview of Character Recognition Methods
Authors: Ainagul Yermekova, Liudmila Goncharenko, Ali Baghirzade, Sergey Sybachin
Abstract:
In this article, we address the problem of Kazakhstan's transition to the Latin alphabet. The transition process started in 2017 and is scheduled to be completed in 2025. In connection with these events, the problem of recognizing the characters of the new alphabet is raised. Well-known character recognition programs such as ABBYY FineReader, FormReader, MyScript Stylus did not recognize specific Kazakh letters that were used in Cyrillic. The author tries to give an assessment of the well-known method of character recognition that could be in demand as part of the country's transition to the Latin alphabet. Three methods of character recognition: template, structured, and feature-based, are considered through the algorithms of operation. At the end of the article, a general conclusion is made about the possibility of applying a certain method to a particular recognition process: for example, in the process of population census, recognition of typographic text in Latin, or recognition of photos of car numbers, store signs, etc.Keywords: text detection, template method, recognition algorithm, structured method, feature method
Procedia PDF Downloads 1891738 Sludge Densification: Emerging and Efficient Way to Look at Biological Nutrient Removal Treatment
Authors: Raj Chavan
Abstract:
Currently, there are over 14,500 Water Resource Recovery Facilities (WRRFs) in the United States, with ~35% of them having some type of nutrient limits in place. These WRRFs account for about 1% of overall power demand and 2% of total greenhouse gas emissions (GHG) in the United States and contribute for 10 to 15% of the overall nutrient load to surface rivers in the United States. The evolution of densification technologies toward more compact and energy-efficient nutrient removal processes has been impacted by a number of factors. Existing facilities that require capacity expansion or biomass densification for higher treatability within the same footprint are being subjected to more stringent requirements relating to nutrient removal prior to surface water discharge. Densification of activated sludge has received recent widespread interest as a means for achieving process intensification and nutrient removal at WRRFs. At the core of the technology are the aerobic sludge granules where the biological processes occur. There is considerable interest in the prospect of producing granular sludge in continuous (or traditional) activated sludge processes (CAS) or densification of biomass by moving activated sludge flocs to a denser aggregate of biomass as a highly effective technique of intensification. This presentation will provide a fundamental understanding of densification by presenting insights and practical issues. The topics that will be discussed include methods used to generate and retain densified granules; the mechanisms that allow biological flocs to densify; the role that physical selectors play in the densification of biological flocs; some viable ways for managing biological flocs that have become densified; effects of physical selection design parameters on the retention of densified biological flocs and finally some operational solutions for customizing the flocs and granules required to meet performance and capacity targets. In addition, it will present some case studies where biological and physical parameters were used to generate aerobic granular sludge in the continuous flow system.Keywords: densification, aerobic granular sludge, nutrient removal, intensification
Procedia PDF Downloads 1891737 Producing Carbon Nanoparticles from Agricultural and Municipal Wastes
Authors: Kanik Sharma
Abstract:
In the year of 2011, the global production of carbon nano-materials (CNMs) was around 3,500 tons, and it is projected to expand at a compound annual growth rate of 30.6%. Expanding markets for applications of CNMs, such as carbon nano-tubes (CNTs) and carbon nano-fibers (CNFs), place ever-increasing demands on lowering their production costs. Current technologies for CNM generation require intensive premium feedstock consumption and employ costly catalysts; they also require input of external energy. Industrial-scale CNM production is conventionally achieved through chemical vapor deposition (CVD) methods which consume a variety of expensive premium chemical feedstocks such as ethylene, carbon monoxide (CO) and hydrogen (H2); or by flame synthesis techniques, which also consume premium feedstock fuels. Additionally, CVD methods are energy-intensive. Renewable and replenishable feedstocks, such as those found in municipal, industrial, agricultural recycling streams have a more judicious reason for usage, in the light of current emerging needs for sustainability. Agricultural sugarcane bagasse and corn residues, scrap tire chips as well as post-consumer polyethylene (PE) and polyethylene terephthalate (PET) bottle shreddings when either thermally treated by sole pyrolysis or by sequential pyrolysis and partial oxidation result in the formation of gaseous carbon-bearing effluents which when channeled into a heated reactor, produce CNMs, including carbon nano-tubes, catalytically synthesized therein on stainless steel meshes. The structure of the nano-material synthesized depends on the type of feedstock available for pyrolysis, and can be determined by analysing the feedstock. These feedstocks could supersede the use of costly and often toxic or highly-flammable chemicals such as hydrocarbon gases, carbon monoxide and hydrogen, which are commonly used as feedstocks in current nano-manufacturing process for CNMs.Keywords: nanomaterials, waste plastics, sugarcane bagasse, pyrolysis
Procedia PDF Downloads 2301736 Analysing the Behaviour of Local Hurst Exponent and Lyapunov Exponent for Prediction of Market Crashes
Authors: Shreemoyee Sarkar, Vikhyat Chadha
Abstract:
In this paper, the local fractal properties and chaotic properties of financial time series are investigated by calculating two exponents, the Local Hurst Exponent: LHE and Lyapunov Exponent in a moving time window of a financial series.y. For the purpose of this paper, the Dow Jones Industrial Average (DIJA) and S&P 500, two of the major indices of United States have been considered. The behaviour of the above-mentioned exponents prior to some major crashes (1998 and 2008 crashes in S&P 500 and 2002 and 2008 crashes in DIJA) is discussed. Also, the optimal length of the window for obtaining the best possible results is decided. Based on the outcomes of the above, an attempt is made to predict the crashes and accuracy of such an algorithm is decided.Keywords: local hurst exponent, lyapunov exponent, market crash prediction, time series chaos, time series local fractal properties
Procedia PDF Downloads 1531735 Solving Linear Systems Involved in Convex Programming Problems
Authors: Yixun Shi
Abstract:
Many interior point methods for convex programming solve an (n+m)x(n+m)linear system in each iteration. Many implementations solve this system in each iteration by considering an equivalent mXm system (4) as listed in the paper, and thus the job is reduced into solving the system (4). However, the system(4) has to be solved exactly since otherwise the error would be entirely passed onto the last m equations of the original system. Often the Cholesky factorization is computed to obtain the exact solution of (4). One Cholesky factorization is to be done in every iteration, resulting in higher computational costs. In this paper, two iterative methods for solving linear systems using vector division are combined together and embedded into interior point methods. Instead of computing one Cholesky factorization in each iteration, it requires only one Cholesky factorization in the entire procedure, thus significantly reduces the amount of computation needed for solving the problem. Based on that, a hybrid algorithm for solving convex programming problems is proposed.Keywords: convex programming, interior point method, linear systems, vector division
Procedia PDF Downloads 4021734 A Survey on Countermeasures of Cache-Timing Attack on AES Systems
Authors: Settana M. Abdulh, Naila A. Sadalla, Yaseen H. Taha, Howaida Elshoush
Abstract:
Side channel attacks are based on side channel information, which is information that is leaked from encryption systems. This includes timing information, power consumption as well as electromagnetic or even sound leaking which can exploited by an attacker. Implementing side channel attacks are possible if and only if an attacker has access to a cryptosystem. In this case, the attacker can exploit bad implementation in software or hardware which is not controlled by encryption implementer. Thus, he/she will represent a real threat to the security system. Several countermeasures have been proposed to eliminate side channel information vulnerability.Cache timing attack is a special type of side channel attack. Here, timing information is collected and analyzed by an attacker to guess sensitive information such as encryption key or plaintext. This paper reviews the technique applied in this attack and surveys the countermeasures against it, evaluating the feasibility and usability of each. Based on this evaluation, finally we pose several recommendations about using these countermeasures.Keywords: AES algorithm, side channel attack, cache timing attack, cache timing countermeasure
Procedia PDF Downloads 3011733 A Weighted Approach to Unconstrained Iris Recognition
Authors: Yao-Hong Tsai
Abstract:
This paper presents a weighted approach to unconstrained iris recognition. Nowadays, commercial systems are usually characterized by strong acquisition constraints based on the subject’s cooperation. However, it is not always achievable for real scenarios in our daily life. Researchers have been focused on reducing these constraints and maintaining the performance of the system by new techniques at the same time. With large variation in the environment, there are two main improvements to develop the proposed iris recognition system. For solving extremely uneven lighting condition, statistic based illumination normalization is first used on eye region to increase the accuracy of iris feature. The detection of the iris image is based on Adaboost algorithm. Secondly, the weighted approach is designed by Gaussian functions according to the distance to the center of the iris. Furthermore, local binary pattern (LBP) histogram is then applied to texture classification with the weight. Experiment showed that the proposed system provided users a more flexible and feasible way to interact with the verification system through iris recognition.Keywords: authentication, iris recognition, adaboost, local binary pattern
Procedia PDF Downloads 2251732 Printed Thai Character Recognition Using Particle Swarm Optimization Algorithm
Authors: Phawin Sangsuvan, Chutimet Srinilta
Abstract:
This Paper presents the applications of Particle Swarm Optimization (PSO) Method for Thai optical character recognition (OCR). OCR consists of the pre-processing, character recognition and post-processing. Before enter into recognition process. The Character must be “Prepped” by pre-processing process. The PSO is an optimization method that belongs to the swarm intelligence family based on the imitation of social behavior patterns of animals. Route of each particle is determined by an individual data among neighborhood particles. The interaction of the particles with neighbors is the advantage of Particle Swarm to determine the best solution. So PSO is interested by a lot of researchers in many difficult problems including character recognition. As the previous this research used a Projection Histogram to extract printed digits features and defined the simple Fitness Function for PSO. The results reveal that PSO gives 67.73% for testing dataset. So in the future there can be explored enhancement the better performance of PSO with improve the Fitness Function.Keywords: character recognition, histogram projection, particle swarm optimization, pattern recognition techniques
Procedia PDF Downloads 4781731 Analysis and Rule Extraction of Coronary Artery Disease Data Using Data Mining
Authors: Rezaei Hachesu Peyman, Oliyaee Azadeh, Salahzadeh Zahra, Alizadeh Somayyeh, Safaei Naser
Abstract:
Coronary Artery Disease (CAD) is one major cause of disability in adults and one main cause of death in developed. In this study, data mining techniques including Decision Trees, Artificial neural networks (ANNs), and Support Vector Machine (SVM) analyze CAD data. Data of 4948 patients who had suffered from heart diseases were included in the analysis. CAD is the target variable, and 24 inputs or predictor variables are used for the classification. The performance of these techniques is compared in terms of sensitivity, specificity, and accuracy. The most significant factor influencing CAD is chest pain. Elderly males (age > 53) have a high probability to be diagnosed with CAD. SVM algorithm is the most useful way for evaluation and prediction of CAD patients as compared to non-CAD ones. Application of data mining techniques in analyzing coronary artery diseases is a good method for investigating the existing relationships between variables.Keywords: classification, coronary artery disease, data-mining, knowledge discovery, extract
Procedia PDF Downloads 6591730 Autonomous Position Control of an Unmanned Aerial Vehicle Based on Accelerometer Response for Indoor Navigation Using Kalman Filtering
Authors: Syed Misbahuddin, Sagufta Kapadia
Abstract:
Autonomous indoor drone navigation has been posed with various challenges, including the inability to use a Global Positioning System (GPS). As of now, Unmanned Aerial Vehicles (UAVs) either rely on 3D mapping systems or utilize external camera arrays to track the UAV in an enclosed environment. The objective of this paper is to develop an algorithm that utilizes Kalman Filtering to reduce noise, allowing the UAV to be navigated indoors using only the flight controller and an onboard companion computer. In this paper, open-source libraries are used to control the UAV, which will only use the onboard accelerometer on the flight controller to estimate the position through double integration. One of the advantages of such a system is that it allows for low-cost and lightweight UAVs to autonomously navigate indoors without advanced mapping of the environment or the use of expensive high-precision-localization sensors.Keywords: accelerometer, indoor-navigation, Kalman-filtering, position-control
Procedia PDF Downloads 3501729 Statecraft: Building a Hindu Nationalist Intellectual Ecosystem in India
Authors: Anuradha Sajjanhar
Abstract:
The rise of authoritarian populist regimes has been accompanied by hardened nationalism and heightened divisions between 'us' and 'them'. Political actors reinforce these sentiments through coercion, but also through inciting fear about imagined threats and by transforming public discourse about policy concerns. Extremist ideas can penetrate national policy, as newly appointed intellectuals and 'experts' in knowledge-producing institutions, such as government committees, universities, and think tanks, succeed in transforming public discourse. While attacking left and liberal academics, universities, and the press, the current Indian government is building new institutions to provide authority to its particularly rigid, nationalist discourse. This paper examines the building of a Hindu-nationalist intellectual ecosystem in India, interrogating the key role of hyper-nationalist think tanks. While some are explicit about their political and ideological leanings, others claim neutrality and pursue their agenda through coded technocratic language and resonant historical narratives. Their key is to change thinking by normalizing it. Six years before winning the election in 2014, India’s Hindu-nationalist party, the BJP, put together its own network of elite policy experts. In a national newspaper, the vice-president of the BJP described this as an intentional shift: from 'being action-oriented to solidifying its ideological underpinnings in a policy framework'. When the BJP came to power in 2014, 'experts' from these think tanks filled key positions in the central government. The BJP has since been circulating dominant ideas of Hindu supremacy through regional parties, grassroots political organisations, and civil society organisations. These think tanks have the authority to articulate and legitimate Hindu nationalism within a credible technocratic policy framework. This paper is based on ethnography and over 50 interviews in New Delhi, before and after the BJP’s staggering election victory in 2019. It outlines the party’s attempt to take over existing institutions while developing its own cadre of nationalist policy-making professionals.Keywords: ideology, politics, South Asia, technocracy
Procedia PDF Downloads 1221728 Application of the MOOD Technique to the Steady-State Euler Equations
Authors: Gaspar J. Machado, Stéphane Clain, Raphael Loubère
Abstract:
The goal of the present work is to numerically study steady-state nonlinear hyperbolic equations in the context of the finite volume framework. We will consider the unidimensional Burgers' equation as the reference case for the scalar situation and the unidimensional Euler equations for the vectorial situation. We consider two approaches to solve the nonlinear equations: a time marching algorithm and a direct steady-state approach. We first develop the necessary and sufficient conditions to obtain the existence and unicity of the solution. We treat regular examples and solutions with a steady shock and to provide very-high-order finite volume approximations we implement a method based on the MOOD technology (Multi-dimensional Optimal Order Detection). The main ingredient consists in using an 'a posteriori' limiting strategy to eliminate non physical oscillations deriving from the Gibbs phenomenon while keeping a high accuracy for the smooth part.Keywords: Euler equations, finite volume, MOOD, steady-state
Procedia PDF Downloads 2781727 Electron Beam Melting Process Parameter Optimization Using Multi Objective Reinforcement Learning
Authors: Michael A. Sprayberry, Vincent C. Paquit
Abstract:
Process parameter optimization in metal powder bed electron beam melting (MPBEBM) is crucial to ensure the technology's repeatability, control, and industry-continued adoption. Despite continued efforts to address the challenges via the traditional design of experiments and process mapping techniques, there needs to be more successful in an on-the-fly optimization framework that can be adapted to MPBEBM systems. Additionally, data-intensive physics-based modeling and simulation methods are difficult to support by a metal AM alloy or system due to cost restrictions. To mitigate the challenge of resource-intensive experiments and models, this paper introduces a Multi-Objective Reinforcement Learning (MORL) methodology defined as an optimization problem for MPBEBM. An off-policy MORL framework based on policy gradient is proposed to discover optimal sets of beam power (P) – beam velocity (v) combinations to maintain a steady-state melt pool depth and phase transformation. For this, an experimentally validated Eagar-Tsai melt pool model is used to simulate the MPBEBM environment, where the beam acts as the agent across the P – v space to maximize returns for the uncertain powder bed environment producing a melt pool and phase transformation closer to the optimum. The culmination of the training process yields a set of process parameters {power, speed, hatch spacing, layer depth, and preheat} where the state (P,v) with the highest returns corresponds to a refined process parameter mapping. The resultant objects and mapping of returns to the P-v space show convergence with experimental observations. The framework, therefore, provides a model-free multi-objective approach to discovery without the need for trial-and-error experiments.Keywords: additive manufacturing, metal powder bed fusion, reinforcement learning, process parameter optimization
Procedia PDF Downloads 941726 Error Correction Method for 2D Ultra-Wideband Indoor Wireless Positioning System Using Logarithmic Error Model
Authors: Phornpat Chewasoonthorn, Surat Kwanmuang
Abstract:
Indoor positioning technologies have been evolved rapidly. They augment the Global Positioning System (GPS) which requires line-of-sight to the sky to track the location of people or objects. This study developed an error correction method for an indoor real-time location system (RTLS) based on an ultra-wideband (UWB) sensor from Decawave. Multiple stationary nodes (anchor) were installed throughout the workspace. The distance between stationary and moving nodes (tag) can be measured using a two-way-ranging (TWR) scheme. The result has shown that the uncorrected ranging error from the sensor system can be as large as 1 m. To reduce ranging error and thus increase positioning accuracy, This study purposes an online correction algorithm using the Kalman filter. The results from experiments have shown that the system can reduce ranging error down to 5 cm.Keywords: indoor positioning, ultra-wideband, error correction, Kalman filter
Procedia PDF Downloads 1601725 Development of an Automatic Control System for ex vivo Heart Perfusion
Authors: Pengzhou Lu, Liming Xin, Payam Tavakoli, Zhonghua Lin, Roberto V. P. Ribeiro, Mitesh V. Badiwala
Abstract:
Ex vivo Heart Perfusion (EVHP) has been developed as an alternative strategy to expand cardiac donation by enabling resuscitation and functional assessment of hearts donated from marginal donors, which were previously not accepted. EVHP parameters, such as perfusion flow (PF) and perfusion pressure (PP) are crucial for optimal organ preservation. However, with the heart’s constant physiological changes during EVHP, such as coronary vascular resistance, manual control of these parameters is rendered imprecise and cumbersome for the operator. Additionally, low control precision and the long adjusting time may lead to irreversible damage to the myocardial tissue. To solve this problem, an automatic heart perfusion system was developed by applying a Human-Machine Interface (HMI) and a Programmable-Logic-Controller (PLC)-based circuit to control PF and PP. The PLC-based control system collects the data of PF and PP through flow probes and pressure transducers. It has two control modes: the RPM-flow mode and the pressure mode. The RPM-flow control mode is an open-loop system. It influences PF through providing and maintaining the desired speed inputted through the HMI to the centrifugal pump with a maximum error of 20 rpm. The pressure control mode is a closed-loop system where the operator selects a target Mean Arterial Pressure (MAP) to control PP. The inputs of the pressure control mode are the target MAP, received through the HMI, and the real MAP, received from the pressure transducer. A PID algorithm is applied to maintain the real MAP at the target value with a maximum error of 1mmHg. The precision and control speed of the RPM-flow control mode were examined by comparing the PLC-based system to an experienced operator (EO) across seven RPM adjustment ranges (500, 1000, 2000 and random RPM changes; 8 trials per range) tested in a random order. System’s PID algorithm performance in pressure control was assessed during 10 EVHP experiments using porcine hearts. Precision was examined through monitoring the steady-state pressure error throughout perfusion period, and stabilizing speed was tested by performing two MAP adjustment changes (4 trials per change) of 15 and 20mmHg. A total of 56 trials were performed to validate the RPM-flow control mode. Overall, the PLC-based system demonstrated the significantly faster speed than the EO in all trials (PLC 1.21±0.03, EO 3.69±0.23 seconds; p < 0.001) and greater precision to reach the desired RPM (PLC 10±0.7, EO 33±2.7 mean RPM error; p < 0.001). Regarding pressure control, the PLC-based system has the median precision of ±1mmHg error and the median stabilizing times in changing 15 and 20mmHg of MAP are 15 and 19.5 seconds respectively. The novel PLC-based control system was 3 times faster with 60% less error than the EO for RPM-flow control. In pressure control mode, it demonstrates a high precision and fast stabilizing speed. In summary, this novel system successfully controlled perfusion flow and pressure with high precision, stability and a fast response time through a user-friendly interface. This design may provide a viable technique for future development of novel heart preservation and assessment strategies during EVHP.Keywords: automatic control system, biomedical engineering, ex-vivo heart perfusion, human-machine interface, programmable logic controller
Procedia PDF Downloads 1751724 Main Chaos-Based Image Encryption Algorithm
Authors: Ibtissem Talbi
Abstract:
During the last decade, a variety of chaos-based cryptosystems have been investigated. Most of them are based on the structure of Fridrich, which is based on the traditional confusion-diffusion architecture proposed by Shannon. Compared with traditional cryptosystems (DES, 3DES, AES, etc.), the chaos-based cryptosystems are more flexible, more modular and easier to be implemented, which make them suitable for large scale-data encyption, such as images and videos. The heart of any chaos-based cryptosystem is the chaotic generator and so, a part of the efficiency (robustness, speed) of the system depends greatly on it. In this talk, we give an overview of the state of the art of chaos-based block ciphers and we describe some of our schemes already proposed. Also we will focus on the essential characteristics of the digital chaotic generator, The needed performance of a chaos-based block cipher in terms of security level and speed of calculus depends on the considered application. There is a compromise between the security and the speed of the calculation. The security of these block block ciphers will be analyzed.Keywords: chaos-based cryptosystems, chaotic generator, security analysis, structure of Fridrich
Procedia PDF Downloads 6851723 A Strategy of Direct Power Control for PWM Rectifier Reducing Ripple in Instantaneous Power
Authors: T. Mohammed Chikouche, K. Hartani
Abstract:
Based on the analysis of basic direct torque control, a parallel master slave for four in-wheel permanent magnet synchronous motors (PMSM) fed by two three phase inverters used in electric vehicle is proposed in this paper. A conventional system with multi-inverter and multi-machine comprises a three phase inverter for each machine to be controlled. Another approach consists in using only one three-phase inverter to supply several permanent magnet synchronous machines. A modified direct torque control (DTC) algorithm is used for the control of the bi-machine traction system. Simulation results show that the proposed control strategy is well adapted for the synchronism of this system and provide good speed tracking performance.Keywords: electric vehicle, multi-machine single-inverter system, multi-machine multi-inverter control, in-wheel motor, master-slave control
Procedia PDF Downloads 2211722 Bit Error Rate Monitoring for Automatic Bias Control of Quadrature Amplitude Modulators
Authors: Naji Ali Albakay, Abdulrahman Alothaim, Isa Barshushi
Abstract:
The most common quadrature amplitude modulator (QAM) applies two Mach-Zehnder Modulators (MZM) and one phase shifter to generate high order modulation format. The bias of MZM changes over time due to temperature, vibration, and aging factors. The change in the biasing causes distortion to the generated QAM signal which leads to deterioration of bit error rate (BER) performance. Therefore, it is critical to be able to lock MZM’s Q point to the required operating point for good performance. We propose a technique for automatic bias control (ABC) of QAM transmitter using BER measurements and gradient descent optimization algorithm. The proposed technique is attractive because it uses the pertinent metric, BER, which compensates for bias drifting independently from other system variations such as laser source output power. The proposed scheme performance and its operating principles are simulated using OptiSystem simulation software for 4-QAM and 16-QAM transmitters.Keywords: automatic bias control, optical fiber communication, optical modulation, optical devices
Procedia PDF Downloads 190