메뉴 건너뛰기




Volumn 2, Issue , 1999, Pages 1324-1331

A sparse sampling algorithm for near-optimal planning in large Markov decision processes

Author keywords

[No Author keywords available]

Indexed keywords

DEGREE OF APPROXIMATION; DISCOUNT FACTORS; INFINITE STATE SPACE; MARKOV DECISION PROCESSES; NEAR-OPTIMAL POLICIES; OPTIMAL POLICIES; STOCHASTIC ENVIRONMENT; TRADITIONAL PLANNING;

EID: 84880649215     PISSN: 10450823     EISSN: None     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (138)

References (11)
  • 3
    • 0031639838 scopus 로고    scopus 로고
    • Applying online-search to reinforcement learning
    • AAAI Press
    • Scott Davies, Andrew Y. Ng, and Andrew Moore. Applying online-search to reinforcement learning. In Proceedings of AAAI-98, pages 753-760. AAAI Press, 1998.
    • (1998) Proceedings of AAAI-98 , pp. 753-760
    • Davies, S.1    Ng, A.Y.2    Moore, A.3
  • 5
    • 84899026236 scopus 로고    scopus 로고
    • Finite-sample convergence rates for Q-learning and indirect algorithms
    • MIT Press
    • Michael Kearns and Satinder Singh. Finite-sample convergence rates for Q-learning and indirect algorithms. In Neural Information Processing Systems 12. MIT Press, 1999.
    • (1999) Neural Information Processing Systems , vol.12
    • Kearns, M.1    Singh, S.2
  • 7
    • 84880658797 scopus 로고    scopus 로고
    • Personal Communication
    • D. McAllester and S. Singh. 1999. Personal Communication.
    • (1999)
    • McAllester, D.1    Singh, S.2
  • 11
    • 0028497385 scopus 로고
    • An upper bound on the loss from approximate optimal-value functions
    • Satinder Singh and Richard Yee. An upper bound on the loss from approximate optimal-value functions. Machine Learning, 16:227-233, 1994.
    • (1994) Machine Learning , vol.16 , pp. 227-233
    • Singh, S.1    Yee, R.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.