The energy-efficient train operation methodology is the focus of this paper, and a Q-Learning-based eco-driving approach is proposed. Firstly, the core idea of energy-distribution-based method (EDBM) that converts the eco-driving… Click to show full abstract
The energy-efficient train operation methodology is the focus of this paper, and a Q-Learning-based eco-driving approach is proposed. Firstly, the core idea of energy-distribution-based method (EDBM) that converts the eco-driving problem to the finite Markov decision process is presented. Secondly, Q-Learning approach is proposed to determine the optimal energy distribution policy. Specifically, two different state definitions, i.e., trip-time-relevant (TT) and energy-distribution-relevant (ED) state definitions, are introduced. Finally, the effectiveness of the proposed approach is verified in a deterministic and a stochastic environment. It is also illustrated that TT-state approach takes about 20 times more computation time compared with ED-state approach while the space complexity of TT-state approach is nearly constant. The hyperparameter sensitivity analysis demonstrates the robustness of the proposed approach.
               
Click one of the above tabs to view related content.