期刊论文详细信息
IEEE Access
Off-Policy Meta-Reinforcement Learning With Belief-Based Task Inference
Yoshimasa Tsuruoka1  Takahisa Imagawa1  Takuya Hiraoka1 
[1] National Institute of Advanced Industrial Science and Technology, Tokyo, Japan;
关键词: Artificial intelligence;    inference;    meta learning;    reinforcement learning;    uncertainty;   
DOI  :  10.1109/ACCESS.2022.3170582
来源: DOAJ
【 摘 要 】

Meta-reinforcement learning (RL) addresses the problem of sample inefficiency in deep RL by using experience obtained in past tasks for solving a new task. However, most existing meta-RL methods require partially or fully on-policy data, which hinders the improvement of sample efficiency. To alleviate this problem, we propose a novel off-policy meta-RL method, embedding learning and uncertainty evaluation (ELUE). An ELUE agent is characterized by the learning of what we call a task embedding space, an embedding space for representing the features of tasks. The agent learns a belief model over the task embedding space and trains a belief-conditional policy and Q-function. The belief model is designed to be agnostic to the order in which task information is obtained, thereby reducing the difficulty of task embedding learning. For a new task, the ELUE agent collects data by the pretrained policy, and updates its belief on the basis of the belief model. Thanks to the belief update, the performance of the agent improves with a small amount of data. In addition, the agent updates the parameters of its policy and Q-function so that it can adjust the pretrained relationships when there are enough data. We demonstrate that ELUE outperforms state-of-the-art meta RL methods through experiments on meta-RL benchmarks.

【 授权许可】

Unknown   

  文献评价指标  
  下载次数:0次 浏览次数:0次