An algorithm has been developed for the resistive control of a nonlinear model of a wave energy converter using least-squares policy iteration, which incorporates function approximation, with tabular and radial… Click to show full abstract
An algorithm has been developed for the resistive control of a nonlinear model of a wave energy converter using least-squares policy iteration, which incorporates function approximation, with tabular and radial basis functions being used as features. With this method, the controller learns the optimal power take-off damping coefficient in each sea state for the maximization of the mean generated power. The performance of the algorithm is assessed against two online reinforcement learning schemes: Q-learning and SARSA. In both regular and irregular waves, least-squares policy iteration outperforms the other strategies, especially when starting from unfavorable conditions for learning. Similar performance is observed for both basis functions, with a smaller number of radial basis functions underfitting the Q-function. The shorter learning time is fundamental for a practical application on a real wave energy converter. Furthermore, this paper shows that least-squares policy iteration is able to maximize the energy absorption of a wave energy converter despite strongly nonlinear effects due to its model-free nature, which removes the influence of modeling errors. Additionally, the floater geometry has been changed during a simulation to show that reinforcement learning control is able to adapt to variations in the system dynamics.
               
Click one of the above tabs to view related content.