Welcome to Journal of University of Chinese Academy of Sciences,Today is

Journal of University of Chinese Academy of Sciences ›› 2024, Vol. 41 ›› Issue (1): 117-126.DOI: 10.7523/j.ucas.2022.028

• Research Articles • Previous Articles     Next Articles

Local observation reconstruction for Ad-Hoc cooperation

CHEN Hao1,2, YANG Likun1,2, YIN Qiyue1,2, HUANG Kaiqi1,2,3   

  1. 1. CRISE, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China;
    2. School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China;
    3 CAS Center for Excellence in Brain Science and Intelligence Technology, Shanghai 200031, China
  • Received:2022-03-02 Revised:2022-04-01 Online:2024-01-15

Abstract: In recent years, multi-agent reinforcement learning has received a lot of attention from researchers. In the study of multi-agent reinforcement learning, the question of how to perform ad-hoc cooperation, i.e., how to adapt to a changing variety and number of teammates, is a key problem. Existing methods either have strong prior knowledge assumptions or use hard-coded protocols for cooperation, which lack generality and can not be generalized to more general ad-hoc cooperation scenarios. To address this problem, this paper proposes a local observation reconstruction algorithm for ad-hoc cooperation, which uses attention mechanisms and sampling networks to reconstruct local observations, enabling the algorithm to recognize and make full use of high-dimensional state representations in different situations and achieve zero-shot generalization in ad-hoc cooperation scenarios. In this paper, the performance of the algorithm is compared and analyzed with representative algorithms on the StarCraft micromanagement environment and ad-hoc cooperation scenarios to verify the effectiveness of the algorithm.

Key words: multi-agent, deep reinforcement learning, credit assignment, Ad-Hoc cooperation

CLC Number: