RUS  ENG
Full version
JOURNALS // Doklady Rossijskoj Akademii Nauk. Mathematika, Informatika, Processy Upravlenia // Archive

Dokl. RAN. Math. Inf. Proc. Upr., 2023 Volume 514, Number 2, Pages 250–261 (Mi danma470)

SPECIAL ISSUE: ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING TECHNOLOGIES

Hierarchical method for cooperative multi-agent reinforcement learning in Markov decision processes

V. È. Bol'shakov, A. N. Alfimtsev

Bauman Moscow State Technical University, Moscow, Russia

Abstract: In the rapidly evolving field of reinforcement learning, combination of hierarchical and multi-agent learning methods presents unique challenges and opens up new opportunities. This paper discusses a combination of multi-level hierarchical learning with subgoal discovery and multi-agent reinforcement learning with hindsight experience replay. Combining these approaches leads to the creation of Multi-Agent Subgoal Hierarchy Algorithm (MASHA) that allows multiple agents to learn efficiently in complex environments, including environments with sparse rewards. We demonstrate the results of the proposed approach in one of these environments inside the StarCraft II strategy game, in addition to making comparisons with other existing approaches. The proposed algorithm is developed in the paradigm of centralized learning with decentralized execution, which makes it possible to achieve a balance between coordination and autonomy of agents.

Keywords: multi-agent reinforcement learning, hierarchical learning, subgoal discovery, hindsight experience replay, centralized learning with decentralized execution, sparse rewards.

UDC: 004.8

Presented: A. A. Shananin
Received: 01.09.2023
Revised: 29.09.2023
Accepted: 18.10.2023

DOI: 10.31857/S2686954323601501


 English version:
Doklady Mathematics, 2023, 108:suppl. 2, S382–S392

Bibliographic databases:


© Steklov Math. Inst. of RAS, 2024