File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1016/j.engappai.2004.08.005
- Scopus: eid_2-s2.0-5444230072
- WOS: WOS:000224909500002
- Find via
Supplementary
-
Bookmarks:
- CiteULike: 1
- Citations:
- Appears in Collections:
Article: Adaptive state space partitioning for reinforcement learning
Title | Adaptive state space partitioning for reinforcement learning |
---|---|
Authors | |
Keywords | Navigation Nearest neighbor quantizer Peg-in-hole Reinforcement learning State space partitioning |
Issue Date | 2004 |
Publisher | Elsevier Ltd. The Journal's web site is located at http://www.elsevier.com/locate/engappai |
Citation | Engineering Applications Of Artificial Intelligence, 2004, v. 17 n. 6, p. 577-588 How to Cite? |
Abstract | The convergence property of reinforcement learning has been extensively investigated in the field of machine learning, however, its applications to real-world problems are still constrained due to its computational complexity. A novel algorithm to improve the applicability and efficacy of reinforcement learning algorithms via adaptive state space partitioning is presented. The proposed temporal difference learning with adaptive vector quantization (TD-AVQ) is an online algorithm and does not assume any a priori knowledge with respect to the learning task and environment. It utilizes the information generated from the reinforcement learning algorithms. Therefore, no additional computations on the decisions of how to partition a particular state space are required. A series of simulations are provided to demonstrate the practical values and performance of the proposed algorithms in solving robot motion planning problems. © 2004 Elsevier Ltd. All rights reserved. |
Persistent Identifier | http://hdl.handle.net/10722/74293 |
ISSN | 2023 Impact Factor: 7.5 2023 SCImago Journal Rankings: 1.749 |
ISI Accession Number ID | |
References |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Lee, ISK | en_HK |
dc.contributor.author | Lau, HYK | en_HK |
dc.date.accessioned | 2010-09-06T06:59:51Z | - |
dc.date.available | 2010-09-06T06:59:51Z | - |
dc.date.issued | 2004 | en_HK |
dc.identifier.citation | Engineering Applications Of Artificial Intelligence, 2004, v. 17 n. 6, p. 577-588 | en_HK |
dc.identifier.issn | 0952-1976 | en_HK |
dc.identifier.uri | http://hdl.handle.net/10722/74293 | - |
dc.description.abstract | The convergence property of reinforcement learning has been extensively investigated in the field of machine learning, however, its applications to real-world problems are still constrained due to its computational complexity. A novel algorithm to improve the applicability and efficacy of reinforcement learning algorithms via adaptive state space partitioning is presented. The proposed temporal difference learning with adaptive vector quantization (TD-AVQ) is an online algorithm and does not assume any a priori knowledge with respect to the learning task and environment. It utilizes the information generated from the reinforcement learning algorithms. Therefore, no additional computations on the decisions of how to partition a particular state space are required. A series of simulations are provided to demonstrate the practical values and performance of the proposed algorithms in solving robot motion planning problems. © 2004 Elsevier Ltd. All rights reserved. | en_HK |
dc.language | eng | en_HK |
dc.publisher | Elsevier Ltd. The Journal's web site is located at http://www.elsevier.com/locate/engappai | en_HK |
dc.relation.ispartof | Engineering Applications of Artificial Intelligence | en_HK |
dc.subject | Navigation | en_HK |
dc.subject | Nearest neighbor quantizer | en_HK |
dc.subject | Peg-in-hole | en_HK |
dc.subject | Reinforcement learning | en_HK |
dc.subject | State space partitioning | en_HK |
dc.title | Adaptive state space partitioning for reinforcement learning | en_HK |
dc.type | Article | en_HK |
dc.identifier.openurl | http://library.hku.hk:4550/resserv?sid=HKU:IR&issn=0952-1976&volume=17&issue=6&spage=289&epage=312&date=2005&atitle=Adaptive+state+space+partitioning+for+reinforcement+learning | en_HK |
dc.identifier.email | Lau, HYK:hyklau@hkucc.hku.hk | en_HK |
dc.identifier.authority | Lau, HYK=rp00137 | en_HK |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1016/j.engappai.2004.08.005 | en_HK |
dc.identifier.scopus | eid_2-s2.0-5444230072 | en_HK |
dc.identifier.hkuros | 103233 | en_HK |
dc.relation.references | http://www.scopus.com/mlt/select.url?eid=2-s2.0-5444230072&selection=ref&src=s&origin=recordpage | en_HK |
dc.identifier.volume | 17 | en_HK |
dc.identifier.issue | 6 | en_HK |
dc.identifier.spage | 577 | en_HK |
dc.identifier.epage | 588 | en_HK |
dc.identifier.isi | WOS:000224909500002 | - |
dc.publisher.place | United Kingdom | en_HK |
dc.identifier.scopusauthorid | Lee, ISK=26663339300 | en_HK |
dc.identifier.scopusauthorid | Lau, HYK=7201497761 | en_HK |
dc.identifier.citeulike | 1723595 | - |
dc.identifier.issnl | 0952-1976 | - |