As a powerful tool for solving nonlinear complex system control problems, the model‐free reinforcement learning hardly guarantees system stability in the early stage of learning, especially with high complicity learning… Click to show full abstract
As a powerful tool for solving nonlinear complex system control problems, the model‐free reinforcement learning hardly guarantees system stability in the early stage of learning, especially with high complicity learning components applied. In this paper, a reinforcement learning framework imitating many cognitive mechanisms of brain such as attention, competition, and integration is proposed to realize sample‐efficient self‐stabilized online learning control. Inspired by the generation of consciousness in human brain, multiple actors that work either competitively for best interaction results or cooperatively for more accurate modeling and predictions were applied. A deep reinforcement learning implementation for challenging control tasks and a real‐time control implementation of the proposed framework are respectively given to demonstrate the high sample efficiency and the capability of maintaining system stability in the online learning process without requiring an initial admissible control.
               
Click one of the above tabs to view related content.