메뉴 건너뛰기




Volumn , Issue , 2011, Pages 335-338

Sample-based planning for continuous action Markov decision processes

Author keywords

[No Author keywords available]

Indexed keywords

ACTION SPACES; BANDIT PROBLEMS; DISCRETIZATIONS; EMPIRICAL RESULTS; MARKOV DECISION PROCESSES;

EID: 80054835987     PISSN: None     EISSN: None     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (86)

References (10)
  • 1
    • 0036568025 scopus 로고    scopus 로고
    • Finite-time analysis of the multi-armed bandit problem
    • Auer, P.; Fischer, P.; and Cesa-Bianchi, N. 2002. Finite-time analysis of the multi-armed bandit problem. Machine Learning 47.
    • (2002) Machine Learning , vol.47
    • Auer, P.1    Fischer, P.2    Cesa-Bianchi, N.3
  • 4
    • 0028442413 scopus 로고
    • Associative reinforcement learning: Functions ink-dnf
    • Kaelbling, L. 1994. Associative reinforcement learning: Functions ink-dnf. Machine Learning 15(3).
    • (1994) Machine Learning , vol.15 , Issue.3
    • Kaelbling, L.1
  • 5
    • 84880649215 scopus 로고    scopus 로고
    • A sparse sampling algorithm for near-optimal planning in large Markov decision processes
    • Kearns, M.; Mansour, S.; and Ng, A. 1999. A sparse sampling algorithm for near-optimal planning in large Markov decision processes. In IJCAI.
    • (1999) IJCAI
    • Kearns, M.1    Mansour, S.2    Ng, A.3
  • 10
    • 0031231885 scopus 로고    scopus 로고
    • Experiments with reinforcement learning in problems with continuous state and action spaces
    • Santamaría, J. C.; Sutton, R.; and Ram, A. 1998. Experiments with reinforcement learning in problems with continuous state and action spaces. In Adaptive Behavior 6.
    • (1998) Adaptive Behavior , vol.6
    • Santamaría, J.C.1    Sutton, R.2    Ram, A.3


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.