1 / 39

Sérgio Ronaldo Barros dos Santos Cairo Lúcio Nascimento Júnior

Design of Attitude and Path Tracking Controllers for Quad-Rotor Robots using Reinforcement Learning. Sérgio Ronaldo Barros dos Santos Cairo Lúcio Nascimento Júnior Instituto Tecnológico de Aeronáutica (ITA) Brazil Sidney Nascimento Givigi Júnior Royal Military College of Canada (RMCC)

lisbet
Download Presentation

Sérgio Ronaldo Barros dos Santos Cairo Lúcio Nascimento Júnior

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Design of Attitude and Path Tracking Controllers for Quad-Rotor Robots using Reinforcement Learning Sérgio Ronaldo Barros dos Santos Cairo Lúcio Nascimento Júnior Instituto Tecnológico de Aeronáutica (ITA) Brazil Sidney Nascimento Givigi Júnior Royal Military College of Canada (RMCC) Canada

  2. Introduction • Quad-rotor robots have attracted the attention of many researchers in the past few years. • Examples of applications: • Military applications: surveillance, border patrolling, crowd control. • Civilian applications: rescue missions during floods and earthquakes, monitoring pipelines and electric transmission liones.

  3. Introduction • A quad-rotor consists of four independent propellers attached to the corners of a cross-shaped frame, turning in opposite directions.

  4. Quad-Rotor Dynamics All rotational and translational movements of a quad-rotor can be achieved by adjusting its rotor speeds.

  5. Introduction • Quad-rotor robots are affected by a number of physical effects such as: • Aerodynamic effects, • Gravity effect, • Ground effect, • Gyroscopic effect, • Friction. • Due to these nonlinear effects, it is difficult to design good controllers for a quad-rotor.

  6. Introduction • Typically quad-rotor applications use controllers derived using linearized models. • These controllers exhibit poor performance for fast maneuvers or in the presence of disturbances such as wind and the ground effect. • In order to perform path tracking in the presence of nonlinear disturbances, a machine learning technique (RL-LA) will be applied.

  7. Objectives • To present a solution for testing and evaluation of attitude stabilization and path tracking controllers for quad-rotors. • To use a Reinforcement Learning algorithm (Learning Automata) to adjust the controllers parameters using a simulation environment that includes wind and ground effects.

  8. Quad-Rotor Dynamics • An inertial frame and a body fixed frame whose origin is in the center of mass of the quad-rotor are used.

  9. Quad-Rotor Dynamics • The dynamic model is derived under the following assumptions. • the vehicle frame is rigid and symmetrical, • the body fixed frame is located at the vehicle center of mass, • the propellers are also rigid.

  10. Quad-Rotor Dynamics • The dynamic model of the quad-rotor can de derived using Newton-Euler formalism.

  11. Robot Controllers • The control architecture for the robot involves two loops: inner and outer. The roll, pitch, and yaw angles are represented by Φ, θ and ψ, respectively.

  12. Robot Controllers • Three nonlinear control strategies are used: - Nonlinear PID Control, - Backstepping technique - Sliding Model Control.

  13. Robot Controllers • The parameters of the 6 controllers are tuned using the RL algorithm.

  14. Simulation Environment • A simulation setup is proposed to train and evaluate the quad-rotor controller under more realistic conditions.

  15. Simulation Environment

  16. Simulation Environment

  17. Simulation Environment • Using the Plane-Marker, a X-Plane model of the X3D-BL quad-rotor (manufactured by Ascending Technologies) was created.

  18. Simulation Environment • The responses of the X-Plane and SIMULINK models are compared for a hovering maneuver.

  19. Reinforcement Learning • Learning Automata (LA) is an alternative approach that can be used to adjust the parameters of the controllers.

  20. Reinforcement Learning • Steps of the learning process: • Initialize the probability and parameters vectors of each controller; • Select the parameters for each controller using its associated probability vector; • Execute the desired task, obtain its response and use a cost function to measure its performance. • Compute the reinforcement signal; • Adjust the probability vectors; • Check the probability vectors for convergence, otherwise return to step 2.

  21. Reinforcement Learning • Supervisory level: LA adjusts the parameters of the attitude and path tracking controllers.

  22. Reinforcement Learning • Learning the parameters of the controllers was executed using the X-Plane model in 3 stages with increasing levels of difficulty : • without the presence of any external disturbances, • considering only the presence of wind, • considering the wind and ground effects.

  23. Reinforcement Learning

  24. Reinforcement Learning

  25. Reinforcement Learning • A cost function evaluates the response of each controller (i) for the selected task at the end of each trial (k) :

  26. Reinforcement Learning • The reinforcement signal is computed for each controller (i) at the end of each trial (k):

  27. Reinforcement Learning The element of the probability vector associated with the selected controller parameter is adjusted: The probability vector is then normalized.

  28. Reinforcement Learning • Learning the desired trajectory using the PID controller during the first stage.

  29. Results • The nonlinear PID controllers results obtained during simulation. The trajectory is formed by the points (0,0) - (0,10) - (10,10) - (10,0) meters.

  30. Results • The quad-rotor robot during the execution of a pre-defined trajectory visualized in the X-Plane.

  31. Results • The backstepping controller results in the presence of wind and ground effects

  32. Results • The path tracking of quad-rotor obtained by the backstepping controllers in the presence of wind and ground effects, visualized in the X-Plane.

  33. Results • The sliding mode controller response using the in presence of wind and ground effects.

  34. Results • The quad-rotor trajectory obtained by the sliding controllers in presence of wind and ground effects, visualized in the X-Plane.

  35. Results • Evaluation of the controllers tracking of desired path after the learning process.

  36. Conclusions • The proposed method (Learning Automata) allows one to tune the parameters of different controllers for a quad-rotor aircraft, considering external disturbances such as wind and ground effects. • It was shown that the proposed simulation framework can be useful to investigate the application of learning algorithms to adjust the control laws of quad-rotors for different flight maneuvers.

  37. Future Research • Evaluate the controllers (obtained using LA, the simulated model, the simulation environment) using real quad-rotors. • On-line learning: useful to correct inaccuracies of the simulated (model + environment).

  38. Future Research • Comparison to other RL methods (e.g., Q-Learning) and other search procedures (e.g., genetic algorithms). • Limitation of learning: generalization to other tasks • Problem: selection of tasks to be executed during training (adaptive control: choice of excitation signal).

  39. Thank You !

More Related