Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2
Search results for: ReinforcementLearning
2 Acquiring Contour Following Behaviour in Robotics through Q-Learning and Image-based States
Authors: Carlos V. Regueiro, Jose E. Domenech, Roberto Iglesias, Jose L. Correa
Abstract:
In this work a visual and reactive contour following behaviour is learned by reinforcement. With artificial vision the environment is perceived in 3D, and it is possible to avoid obstacles that are invisible to other sensors that are more common in mobile robotics. Reinforcement learning reduces the need for intervention in behaviour design, and simplifies its adjustment to the environment, the robot and the task. In order to facilitate its generalisation to other behaviours and to reduce the role of the designer, we propose a regular image-based codification of states. Even though this is much more difficult, our implementation converges and is robust. Results are presented with a Pioneer 2 AT on a Gazebo 3D simulator.Keywords: Image-based State Codification, Mobile Robotics, ReinforcementLearning, Visual Behaviour.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16071 Adaptive PID Controller based on Reinforcement Learning for Wind Turbine Control
Authors: M. Sedighizadeh, A. Rezazadeh
Abstract:
A self tuning PID control strategy using reinforcement learning is proposed in this paper to deal with the control of wind energy conversion systems (WECS). Actor-Critic learning is used to tune PID parameters in an adaptive way by taking advantage of the model-free and on-line learning properties of reinforcement learning effectively. In order to reduce the demand of storage space and to improve the learning efficiency, a single RBF neural network is used to approximate the policy function of Actor and the value function of Critic simultaneously. The inputs of RBF network are the system error, as well as the first and the second-order differences of error. The Actor can realize the mapping from the system state to PID parameters, while the Critic evaluates the outputs of the Actor and produces TD error. Based on TD error performance index and gradient descent method, the updating rules of RBF kernel function and network weights were given. Simulation results show that the proposed controller is efficient for WECS and it is perfectly adaptable and strongly robust, which is better than that of a conventional PID controller.Keywords: Wind energy conversion systems, reinforcementlearning; Actor-Critic learning; adaptive PID control; RBF network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4942