메뉴 건너뛰기




Volumn , Issue , 2002, Pages 260-266

Nearly deterministic abstractions of Markov decision processes

Author keywords

[No Author keywords available]

Indexed keywords

ALGORITHMS; COMPUTER SIMULATION; MARKOV PROCESSES; NAVIGATION; OPTIMIZATION; PROBLEM SOLVING;

EID: 0036931070     PISSN: None     EISSN: None     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (14)

References (23)
  • 3
    • 0034248853 scopus 로고    scopus 로고
    • Stochastic dynamic programming with factored representations
    • Boutilier, C.; Dearden, R.; and Goldszmidt, M. 2000. Stochastic dynamic programming with factored representations. Artificial Intelligence 121(1-2):49-107.
    • (2000) Artificial Intelligence , vol.121 , Issue.1-2 , pp. 49-107
    • Boutilier, C.1    Dearden, R.2    Goldszmidt, M.3
  • 9
    • 0013061165 scopus 로고    scopus 로고
    • Kluwer Academic Publishers, chapter Experimental Analysis of Heuristics for the STSP. To appear
    • Johnson, D. S., and McGeoch, L. A. 2001. The Traveling Salesman Problem and its Variations. Kluwer Academic Publishers, chapter Experimental Analysis of Heuristics for the STSP. To appear.
    • (2001) The Traveling Salesman Problem and its Variations
    • Johnson, D.S.1    McGeoch, L.A.2
  • 18
    • 0001205548 scopus 로고    scopus 로고
    • Complexity of finite-horizon markov decision process problems
    • Mundhenk, M.; Goldsmith, J.; Lusena, C.; and Allender, E. 2000. Complexity of finite-horizon markov decision process problems. Journal of the ACM 47(4):681-720.
    • (2000) Journal of the ACM , vol.47 , Issue.4 , pp. 681-720
    • Mundhenk, M.1    Goldsmith, J.2    Lusena, C.3    Allender, E.4
  • 21
    • 0003392384 scopus 로고    scopus 로고
    • Ph.D. Dissertation, University of Massachusetts, Amherst, Department of Computer Science
    • Precup, D. 2000. Temporal Abstraction in Reinforcement Learning. Ph.D. Dissertation, University of Massachusetts, Amherst, Department of Computer Science.
    • (2000) Temporal Abstraction in Reinforcement Learning
    • Precup, D.1
  • 23
    • 0033170372 scopus 로고    scopus 로고
    • Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning
    • Sutton, R. S.; Precup, D.; and Singh, S. 1999. Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning. Artificial Intelligence 112:181-211.
    • (1999) Artificial Intelligence , vol.112 , pp. 181-211
    • Sutton, R.S.1    Precup, D.2    Singh, S.3


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.