Autonomous landing of quadrotor on a moving UGV with the optimal control policies
View/ Open
Date
2020-12-01Author
Manda, Suhas Priyatham
0000-0003-2846-9743
Metadata
Show full item recordAbstract
This thesis proposes an offline method that uses an integral reinforcement learning (IRL) technique along with the system identification to determine the optimal control of a system with completely unknown dynamics. Unmanned aerial vehicles (UAV) that are particularly deployed to track and land on an arbitrarily moving unmanned ground vehicles (UGV), demand a high performance controller to perform precise tracking. One way of designing an optimal tracking controller is developing linear quadratic integrators (LQI) with a quadratic type of cost function that solves Riccati equation. However, this approach requires prior knowledge of the linearized UAV system dynamics. We overcome this problem by employing an IRL technique that solves LQI through system identification. Usually, IRL techniques adopt a conventional way of solving the Hamilton–Jacobi–Bellman (HJB) equation with value function approximation. The proposed approach evaluates the optimal control using IRL that solves the HJB equation using system identification instead of value function approximation. Assuming that the UAV system dynamics are linear time-invariant over a particular flight condition, we identify the linear model by analyzing the input and output data samples from a linear regression perspective, where we use the conjugate gradient descent optimization algorithm. This approach addresses the challenge to compute optimal control without the need to know UAV dynamics. We have rigorously tested and simulated the proposed method on various flight trajectories. The test results have shown significant improvement in the control policy over each iteration of IRL. After validating the proposed method in simulation, we have implemented this approach on a real UAV to track and land on a UGV.