Formula-E race strategy development using distributed policy gradient reinforcement learning

Date

2021-01-20

Free to read from

Supervisor/s

Journal Title

Journal ISSN

Volume Title

Publisher

Elsevier

Department

Type

Article

ISSN

0950-7051

Format

Citation

Liu X, Fotouhi A, Auger DJ. (2021) Formula-E race strategy development using distributed policy gradient reinforcement learning. Knowledge-Based Systems, Volume 216, March 2021, Article number 106781

Abstract

Energy and thermal management is a crucial element in Formula-E race strategy development. In this study, the race-level strategy development is formulated into a Markov decision process (MDP) problem featuring a hybrid-type action space. Deep Deterministic Policy Gradient (DDPG) reinforcement learning is implemented under distributed architecture Ape-X and integrated with the prioritized experience replay and reward shaping techniques to optimize a hybrid-type set of actions of both continuous and discrete components. Soft boundary violation penalties in reward shaping, significantly improves the performance of DDPG and makes it capable of generating faster race finishing solutions. The new proposed method has shown superior performance in comparison to the Monte Carlo Tree Search (MCTS) with policy gradient reinforcement learning, which solves this problem in a fully discrete action space as presented in the literature. The advantages are faster race finishing time and better handling of ambient temperature rise.

Description

Software Description

Software Language

Github

Keywords

Energy management, Formula-E race strategy, Deep deterministic policy gradient, Reinforcement leaning

DOI

Rights

Attribution-NonCommercial-NoDerivatives 4.0 International

Relationships

Relationships

Supplements

Funder/s