File Download
Supplementary

Conference Paper: Adaptive vector quantization for reinforcement learning

TitleAdaptive vector quantization for reinforcement learning
Authors
KeywordsLearning algorithm
intelligent control
vector quantization
robot navigation
automated guided vehicles
Issue Date2002
PublisherInternational Federation of Automatic Control.
Citation
The 15th World Congress of International Federation of Automatic Control (2002 IFAC), Barcelona, Spain, 21-26 July 2002 How to Cite?
AbstractDynamic programming methods are capable of solving reinforcement learning problems, in which an agent must improve its behavior through trial-and-error interactions with a dynamic environment. However, these computational algorithms suffer from the curse of dimensionality (Bellman, 1957) that the number of computational operations increases exponentially with the cardinality of the state space. In practice, this usually results in a very long training time and applications in continuous domain are far from trivial. In order to ease this problem, we propose the use of vector quantization to adaptively partition the state space based on the recent estimate of the action-value function. In particular, this state-space partitioning operation is performed incrementally to reflect the experience accumulated by the agent as it explores the underlying environment.
Persistent Identifierhttp://hdl.handle.net/10722/100010

 

DC FieldValueLanguage
dc.contributor.authorLau, HYKen_HK
dc.contributor.authorMak, KLen_HK
dc.contributor.authorLee, SKen_HK
dc.date.accessioned2010-09-25T18:53:19Z-
dc.date.available2010-09-25T18:53:19Z-
dc.date.issued2002en_HK
dc.identifier.citationThe 15th World Congress of International Federation of Automatic Control (2002 IFAC), Barcelona, Spain, 21-26 July 2002-
dc.identifier.urihttp://hdl.handle.net/10722/100010-
dc.description.abstractDynamic programming methods are capable of solving reinforcement learning problems, in which an agent must improve its behavior through trial-and-error interactions with a dynamic environment. However, these computational algorithms suffer from the curse of dimensionality (Bellman, 1957) that the number of computational operations increases exponentially with the cardinality of the state space. In practice, this usually results in a very long training time and applications in continuous domain are far from trivial. In order to ease this problem, we propose the use of vector quantization to adaptively partition the state space based on the recent estimate of the action-value function. In particular, this state-space partitioning operation is performed incrementally to reflect the experience accumulated by the agent as it explores the underlying environment.-
dc.languageengen_HK
dc.publisherInternational Federation of Automatic Control.-
dc.relation.ispartofProceedings of the 15th World Congress of International Federation of Automatic Controlen_HK
dc.subjectLearning algorithm-
dc.subjectintelligent control-
dc.subjectvector quantization-
dc.subjectrobot navigation-
dc.subjectautomated guided vehicles-
dc.titleAdaptive vector quantization for reinforcement learningen_HK
dc.typeConference_Paperen_HK
dc.identifier.emailLau, HYK: hyklau@hkucc.hku.hken_HK
dc.identifier.emailMak, KL: makkl@hkucc.hku.hken_HK
dc.identifier.authorityLau, HYK=rp00137en_HK
dc.identifier.authorityMak, KL=rp00154en_HK
dc.description.naturelink_to_OA_fulltext-
dc.identifier.hkuros81153en_HK

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats