Sample Efficiency in Model-Based Reinforcement Learning for Ship Seakeeping & Manoeuvring

Keywords

Loading...
Thumbnail Image

Issue Date

2023-04-01

Language

en

Document type

Journal Title

Journal ISSN

Volume Title

Publisher

Title

ISSN

Volume

Issue

Startpage

Endpage

DOI

Abstract

Ships travel in stochastic environments. The stochasticity does not only come from the waves but also from the wind, current, bathymetry changes and presence of other objects. Maintaining the course of the ship in such an environment is challenging as it involves dynamic interaction of the ship and its appendages with this stochastic environment. For course keeping of ships, the potential of reinforcement learning (RL) is mainly unexplored as classical feedback controllers are typically used. The training of RL models, however, requires a large number of interactions with the environment, which makes the real world application of these models infeasible. This research project investigates a model-based approach to increase the sample-efficiency in RL. Model-based RL learns a dynamics model alongside the policy to estimate the state transitions in the environment. In this project, a model-based RL agent is trained to maintain the course of a ship in waves, and is compared to a linear-quadratic regulator (LQR) and a model-free RL agent. By employing the model-based agent, the required number of agent-environment interactions to learn a policy was significantly reduced compared to the model-free agent. Furthermore, the performance of the agent was improved in the sense that the rudder usage was reduced considerably while maintaining the course as well as the model-free RL agent and the LQR.

Description

Citation

Faculty

Faculteit der Sociale Wetenschappen