File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: Model-based reinforcement learning via imagination with derived memory
Title | Model-based reinforcement learning via imagination with derived memory |
---|---|
Authors | |
Keywords | Model-based Reinforcement Learning Policy Robustness Visual Control Task Derived Memory World Models |
Issue Date | 2021 |
Publisher | Neural Information Processing Systems Foundation. |
Citation | 35th Conference on Neural Information Processing Systems (NeurIPS 2021) (Virtual), December 6-14, 2021. In Advances In Neural Information Processing Systems: 35th conference on neural information processing systems (NeurIPS 2021), p. 9493-9505 How to Cite? |
Abstract | Model-based reinforcement learning aims to improve the sample efficiency of policy learning by modeling the dynamics of the environment. Recently, the latent dynamics model is further developed to enable fast planning in a compact space. It summarizes the high-dimensional experiences of an agent, which mimics the memory function of humans. Learning policies via imagination with the latent model shows great potential for solving complex tasks. However, only considering memories from the true experiences in the process of imagination could limit its advantages. Inspired by the memory prosthesis proposed by neuroscientists, we present a novel model-based reinforcement learning framework called Imagining with Derived Memory (IDM). It enables the agent to learn policy from enriched diverse imagination with prediction-reliability weight, thus improving sample efficiency and policy robustness. Experiments on various high-dimensional visual control tasks in the DMControl benchmark demonstrate that IDM outperforms previous state-of-the-art methods in terms of policy robustness and further improves the sample efficiency of the model-based method. |
Persistent Identifier | http://hdl.handle.net/10722/315861 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Mu, Y | - |
dc.contributor.author | Zhuang, Y | - |
dc.contributor.author | Wang, B | - |
dc.contributor.author | Zhu, G | - |
dc.contributor.author | Liu, W | - |
dc.contributor.author | Chen, J | - |
dc.contributor.author | Luo, P | - |
dc.contributor.author | Li, SE | - |
dc.contributor.author | Zhang, C | - |
dc.contributor.author | Hao, J | - |
dc.date.accessioned | 2022-08-19T09:05:47Z | - |
dc.date.available | 2022-08-19T09:05:47Z | - |
dc.date.issued | 2021 | - |
dc.identifier.citation | 35th Conference on Neural Information Processing Systems (NeurIPS 2021) (Virtual), December 6-14, 2021. In Advances In Neural Information Processing Systems: 35th conference on neural information processing systems (NeurIPS 2021), p. 9493-9505 | - |
dc.identifier.uri | http://hdl.handle.net/10722/315861 | - |
dc.description.abstract | Model-based reinforcement learning aims to improve the sample efficiency of policy learning by modeling the dynamics of the environment. Recently, the latent dynamics model is further developed to enable fast planning in a compact space. It summarizes the high-dimensional experiences of an agent, which mimics the memory function of humans. Learning policies via imagination with the latent model shows great potential for solving complex tasks. However, only considering memories from the true experiences in the process of imagination could limit its advantages. Inspired by the memory prosthesis proposed by neuroscientists, we present a novel model-based reinforcement learning framework called Imagining with Derived Memory (IDM). It enables the agent to learn policy from enriched diverse imagination with prediction-reliability weight, thus improving sample efficiency and policy robustness. Experiments on various high-dimensional visual control tasks in the DMControl benchmark demonstrate that IDM outperforms previous state-of-the-art methods in terms of policy robustness and further improves the sample efficiency of the model-based method. | - |
dc.language | eng | - |
dc.publisher | Neural Information Processing Systems Foundation. | - |
dc.relation.ispartof | Advances In Neural Information Processing Systems: 35th conference on neural information processing systems (NeurIPS 2021) | - |
dc.subject | Model-based Reinforcement Learning | - |
dc.subject | Policy Robustness | - |
dc.subject | Visual Control Task | - |
dc.subject | Derived Memory | - |
dc.subject | World Models | - |
dc.title | Model-based reinforcement learning via imagination with derived memory | - |
dc.type | Conference_Paper | - |
dc.identifier.email | Luo, P: pluo@hku.hk | - |
dc.identifier.authority | Luo, P=rp02575 | - |
dc.identifier.hkuros | 335597 | - |
dc.identifier.spage | 9493 | - |
dc.identifier.epage | 9505 | - |
dc.publisher.place | United States | - |