File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: Do RNN and LSTM have Long Memory?
Title | Do RNN and LSTM have Long Memory? |
---|---|
Authors | |
Issue Date | 2021 |
Citation | Proceedings of the 37th International Conference on Machine Learning, PMLR, v. 119, p. 11365-11375 How to Cite? |
Abstract | The LSTM network was proposed to overcome the difficulty in learning long-term dependence, and has made significant advancements in applications. With its success and drawbacks in mind, this paper raises the question - do RNN and LSTM have long memory? We answer it partially by proving that RNN and LSTM do not have long memory from a statistical perspective. A new definition for long memory networks is further introduced, and it requires the model weights to decay at a polynomial rate. To verify our theory, we convert RNN and LSTM into long memory networks by making a minimal modification, and their superiority is illustrated in modeling long-term dependence of various datasets. |
Persistent Identifier | http://hdl.handle.net/10722/320755 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | ZHAO, J | - |
dc.contributor.author | HUANG, F | - |
dc.contributor.author | Lv, J | - |
dc.contributor.author | Duan, Y | - |
dc.contributor.author | Qin, Z | - |
dc.contributor.author | Li, G | - |
dc.contributor.author | Tian, G | - |
dc.date.accessioned | 2022-10-21T07:59:16Z | - |
dc.date.available | 2022-10-21T07:59:16Z | - |
dc.date.issued | 2021 | - |
dc.identifier.citation | Proceedings of the 37th International Conference on Machine Learning, PMLR, v. 119, p. 11365-11375 | - |
dc.identifier.uri | http://hdl.handle.net/10722/320755 | - |
dc.description.abstract | The LSTM network was proposed to overcome the difficulty in learning long-term dependence, and has made significant advancements in applications. With its success and drawbacks in mind, this paper raises the question - do RNN and LSTM have long memory? We answer it partially by proving that RNN and LSTM do not have long memory from a statistical perspective. A new definition for long memory networks is further introduced, and it requires the model weights to decay at a polynomial rate. To verify our theory, we convert RNN and LSTM into long memory networks by making a minimal modification, and their superiority is illustrated in modeling long-term dependence of various datasets. | - |
dc.language | eng | - |
dc.relation.ispartof | Proceedings of the 37th International Conference on Machine Learning, PMLR | - |
dc.title | Do RNN and LSTM have Long Memory? | - |
dc.type | Conference_Paper | - |
dc.identifier.email | Li, G: gdli@hku.hk | - |
dc.identifier.authority | Li, G=rp00738 | - |
dc.identifier.hkuros | 339989 | - |
dc.identifier.volume | 119 | - |
dc.identifier.spage | 11365 | - |
dc.identifier.epage | 11375 | - |