AB - © 2018 Australasian Robotics and Automation Association. All rights reserved. Small Autonomous Underwater Vehicles (AUV) in shallow water might not be stabilized well by feedback or model predictive control. This is because wave and current disturbances may frequently exceed AUV thrust capabilities and disturbance estimation and prediction models available are not sufficiently accurate. In contrast to classical model-free Reinforcement Learning (RL), this paper presents an improved RL for Excessive disturbance rejection Control (REC) that is able to learn and utilize disturbance behaviour, through formulating the disturbed AUV dynamics as a multi-order Markov chain. The unobserved disturbance behaviour is then encoded in the AUV state-action history of fixed length, its embeddings are learned within the policy optimization. The proposed REC is further enhanced by a base controller that is pre-trained on iterative Linear Quadratic Regulator (iLQR) solutions for a reduced AUV dynamic model, resulting in hybrid-REC. Numerical simulations on pose regulation tasks have demonstrated that REC significantly outperforms a canonical controller and classical RL, and that the hybrid-REC leads to more efficient and safer sampling and motion than REC. AU - Wang, T AU - Lu, W AU - Liu, D DA - 2018/01/01 JO - Australasian Conference on Robotics and Automation, ACRA PY - 2018/01/01 TI - Excessive disturbance rejection control of autonomous underwater vehicle using reinforcement learning VL - 2018-December Y1 - 2018/01/01 Y2 - 2024/03/29 ER -