Manoeuvrability Pursuit-Evasion Games for Robots

An overall overview of the framework, including the architecture of MERL-GP and the application of multi-robot pursuit-evasion games.

Recently, multi-agent reinforcement learning (MARL) has made great progress in pursuit-evasion scenarios. A number of MARL methods are proposed, but they have their own disadvantages. Although these methods can improve the cooperative ability of robots, these methods do not apply to environments with a large number of robots as they require the use of all the robots’ states or observations in constructing their critic networks. As the number of robots increases, these methods will become increasingly difficult to be trained.

A whole novel method, called manoeuvrability enhanced reinforcement learning via gaussian process (MERL-GP), is proposed to deal with problems including escape strategy, local optima, and uncertainty for multi-robot high-dimensional data. MERL-GP contains manoeuvrability action, composite reward mechanism, and gaussian process. Specifically, manoeuvrability action provides more escape strategies. Composite reward mechanism overcomes the sparse reward and local optima problems. Gaussian process approximation solves the Q-function and allows an accurate online update of the parameters of the posterior mean and covariance. (Check our video)

Related