메뉴 건너뛰기




Volumn 4131 LNCS - I, Issue , 2006, Pages 790-800

Optimal tuning of continual online exploration in reinforcement learning

Author keywords

[No Author keywords available]

Indexed keywords

ALGORITHMS; ENTROPY; ONLINE SYSTEMS; OPTIMAL SYSTEMS; OPTIMIZATION; PROBABILITY DISTRIBUTIONS;

EID: 33749864692     PISSN: 03029743     EISSN: 16113349     Source Type: Book Series    
DOI: 10.1007/11840817_82     Document Type: Conference Paper
Times cited : (13)

References (24)
  • 1
    • 34047216334 scopus 로고    scopus 로고
    • Timing continual exploration in reinforcement learning
    • Y. Achbany, F. Fouss, L. Yen, A. Pirotte, and M. Saerens, Timing continual exploration in reinforcement learning. Technical report, http://www.isys.ucl.ac.be/staff/francois/Articles/Achbany2005a.pdf, 2005.
    • (2005) Technical Report
    • Achbany, Y.1    Fouss, F.2    Yen, L.3    Pirotte, A.4    Saerens, M.5
  • 16
    • 0029753630 scopus 로고    scopus 로고
    • Reinforcement learning with replacing eligibility traces
    • S. Singh and R. Sutton. Reinforcement learning with replacing eligibility traces. Machine Learning, 22:123-158, 1996.
    • (1996) Machine Learning , vol.22 , pp. 123-158
    • Singh, S.1    Sutton, R.2
  • 20
    • 0002210775 scopus 로고
    • The role of exploration in learning control
    • D. White and D. Sofge, editors. Van Nostrand Reinhold, Florence, Kentucky 41022
    • S. Thrun. The role of exploration in learning control. In D. White and D. Sofge, editors, Handbook for Intelligent Control: Neural, Fuzzy and Adaptive Approaches. Van Nostrand Reinhold, Florence, Kentucky 41022, 1992.
    • (1992) Handbook for Intelligent Control: Neural, Fuzzy and Adaptive Approaches
    • Thrun, S.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.