File Download

There are no files associated with this item.

Supplementary

Conference Paper: Do RNN and LSTM have Long Memory?

TitleDo RNN and LSTM have Long Memory?
Authors
Issue Date2021
Citation
Proceedings of the 37th International Conference on Machine Learning, PMLR, v. 119, p. 11365-11375 How to Cite?
AbstractThe LSTM network was proposed to overcome the difficulty in learning long-term dependence, and has made significant advancements in applications. With its success and drawbacks in mind, this paper raises the question - do RNN and LSTM have long memory? We answer it partially by proving that RNN and LSTM do not have long memory from a statistical perspective. A new definition for long memory networks is further introduced, and it requires the model weights to decay at a polynomial rate. To verify our theory, we convert RNN and LSTM into long memory networks by making a minimal modification, and their superiority is illustrated in modeling long-term dependence of various datasets.
Persistent Identifierhttp://hdl.handle.net/10722/320755

 

DC FieldValueLanguage
dc.contributor.authorZHAO, J-
dc.contributor.authorHUANG, F-
dc.contributor.authorLv, J-
dc.contributor.authorDuan, Y-
dc.contributor.authorQin, Z-
dc.contributor.authorLi, G-
dc.contributor.authorTian, G-
dc.date.accessioned2022-10-21T07:59:16Z-
dc.date.available2022-10-21T07:59:16Z-
dc.date.issued2021-
dc.identifier.citationProceedings of the 37th International Conference on Machine Learning, PMLR, v. 119, p. 11365-11375-
dc.identifier.urihttp://hdl.handle.net/10722/320755-
dc.description.abstractThe LSTM network was proposed to overcome the difficulty in learning long-term dependence, and has made significant advancements in applications. With its success and drawbacks in mind, this paper raises the question - do RNN and LSTM have long memory? We answer it partially by proving that RNN and LSTM do not have long memory from a statistical perspective. A new definition for long memory networks is further introduced, and it requires the model weights to decay at a polynomial rate. To verify our theory, we convert RNN and LSTM into long memory networks by making a minimal modification, and their superiority is illustrated in modeling long-term dependence of various datasets.-
dc.languageeng-
dc.relation.ispartofProceedings of the 37th International Conference on Machine Learning, PMLR-
dc.titleDo RNN and LSTM have Long Memory?-
dc.typeConference_Paper-
dc.identifier.emailLi, G: gdli@hku.hk-
dc.identifier.authorityLi, G=rp00738-
dc.identifier.hkuros339989-
dc.identifier.volume119-
dc.identifier.spage11365-
dc.identifier.epage11375-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats