메뉴 건너뛰기




Volumn 27, Issue 1, 2013, Pages 1-51

A survey of point-based POMDP solvers

Author keywords

Decision theoretic planning; Partially observable Markov decision processes; Reinforcement learning

Indexed keywords

ITERATIVE METHODS; REINFORCEMENT LEARNING; SURVEYS;

EID: 84873747053     PISSN: 13872532     EISSN: 15737454     Source Type: Journal    
DOI: 10.1007/s10458-012-9200-2     Document Type: Article
Times cited : (427)

References (70)
  • 4
    • 0029210635 scopus 로고
    • Learning to act using real-time dynamic programming
    • doi:10.1016/0004-3702(94)00011-O
    • Barto A. G., Bradtke S. J., Singh S. P. (1995) Learning to act using real-time dynamic programming. Artificial Intelligence 72: 81-138. doi: 10. 1016/0004-3702(94)00011-O.
    • (1995) Artificial Intelligence , vol.72 , pp. 81-138
    • Barto, A.G.1    Bradtke, S.J.2    Singh, S.P.3
  • 5
    • 85012688561 scopus 로고
    • Princeton: Princeton University Press
    • Bellman R. (1957) Dynamic programming. Princeton University Press, Princeton.
    • (1957) Dynamic Programming
    • Bellman, R.1
  • 11
    • 0001909869 scopus 로고    scopus 로고
    • Incremental Pruning: A simple, fast, exact method for partially observable Markov decision processes
    • Cassandra, A., Littman, M. L., & Zhang, N. L. (1997). Incremental Pruning: A simple, fast, exact method for partially observable Markov decision processes. In Conference on uncertainty in artificial intelligence (UAI) (pp. 54-61). http://www. cs. duke. edu/~mlittman/docs/uai97-pomdp. ps.
    • (1997) Conference on uncertainty in artificial intelligence (UAI) , pp. 54-61
    • Cassandra, A.1    Littman, M.L.2    Zhang, N.L.3
  • 18
    • 0031385618 scopus 로고    scopus 로고
    • Incremental methods for computing bounds in partially observable Markov decision processes
    • Hauskrecht, M. (1997). Incremental methods for computing bounds in partially observable Markov decision processes. In: National conference on artificial intelligence (pp. 734-739).
    • (1997) National conference on artificial intelligence , pp. 734-739
    • Hauskrecht, M.1
  • 19
    • 0001770240 scopus 로고    scopus 로고
    • Value-function approximations for partially observable Markov decision processes
    • Hauskrecht, M. (2000). Value-function approximations for partially observable Markov decision processes. Journal of Artificial Intelligence Research (JAIR), 13, 33-94. http://www. cs. washington. edu/research/jair/abstracts/hauskrecht00a. html.
    • (2000) Journal of Artificial Intelligence Research (JAIR) , vol.13 , pp. 33-94
    • Hauskrecht, M.1
  • 20
    • 0034160101 scopus 로고    scopus 로고
    • Planning treatment of ischemic heart disease with partially observable Markov decision processes
    • Hauskrecht M., Fraser H. S. F. (2000) Planning treatment of ischemic heart disease with partially observable Markov decision processes. Artificial Intelligence in Medicine 18(3): 221-244.
    • (2000) Artificial Intelligence in Medicine , vol.18 , Issue.3 , pp. 221-244
    • Hauskrecht, M.1    Fraser, H.S.F.2
  • 21
    • 77950932056 scopus 로고    scopus 로고
    • Automated handwashing assistance for persons with dementia using video and a partially observable Markov decision process
    • Hoey J., Poupart P., von Bertoldi A., Craig T., Boutilier C., Mihailidis A. (2010) Automated handwashing assistance for persons with dementia using video and a partially observable Markov decision process. Computer Vision and Image Understanding 114(5): 503-519.
    • (2010) Computer Vision and Image Understanding , vol.114 , Issue.5 , pp. 503-519
    • Hoey, J.1    Poupart, P.2    von Bertoldi, A.3    Craig, T.4    Boutilier, C.5    Mihailidis, A.6
  • 28
    • 0032073263 scopus 로고    scopus 로고
    • Planning and acting in partially observable stochastic domains
    • Kaelbling, L., Littman, M., & Cassandra, A. (1998). Planning and acting in partially observable stochastic domains. In Artificial intelligence (pp. 99-134).
    • (1998) Artificial intelligence , pp. 99-134
    • Kaelbling, L.1    Littman, M.2    Cassandra, A.3
  • 30
    • 70349645087 scopus 로고    scopus 로고
    • SARSOP: Efficient point-based POMDP planning by approximating optimally reachable belief spaces
    • Kurniawati, H., Hsu, D., & Lee, W. (2008). SARSOP: Efficient point-based POMDP planning by approximating optimally reachable belief spaces. In Robotics: Science and systems (RSS).
    • (2008) Robotics: Science and systems (RSS)
    • Kurniawati, H.1    Hsu, D.2    Lee, W.3
  • 31
    • 0003861655 scopus 로고    scopus 로고
    • PhD thesis, Department of Computer Science, Brown University, Providence, RI, Also Technical Report CS-96-09
    • Littman, M. L. (1996). Algorithms for sequential decision making. PhD thesis, Department of Computer Science, Brown University, Providence, RI. ftp://ftp. cs. brown. edu/pub/techreports/96/cs96-09. ps. Z. Also Technical Report CS-96-09.
    • (1996) Algorithms for sequential decision making
    • Littman, M.L.1
  • 35
    • 0000494894 scopus 로고
    • Computationally feasible bounds for partially observed Markov decision processes
    • Lovejoy W. S. (1991) Computationally feasible bounds for partially observed Markov decision processes. Operations Research 39(1): 162-175.
    • (1991) Operations Research , vol.39 , Issue.1 , pp. 162-175
    • Lovejoy, W.S.1
  • 56
    • 0029753630 scopus 로고    scopus 로고
    • Reinforcement learning with replacing eligibility traces
    • Singh S. P., Sutton R. S. (1996) Reinforcement learning with replacing eligibility traces. Machine Learning 22: 123-158.
    • (1996) Machine Learning , vol.22 , pp. 123-158
    • Singh, S.P.1    Sutton, R.S.2
  • 60
    • 0017943242 scopus 로고
    • The optimal control of partially observable Markov processes over the infinite horizon: Discounted costs
    • Sondik E. J. (1978) The optimal control of partially observable Markov processes over the infinite horizon: Discounted costs. Operations Research 26: 282-304.
    • (1978) Operations Research , vol.26 , pp. 282-304
    • Sondik, E.J.1
  • 67
    • 33750703175 scopus 로고    scopus 로고
    • Partially observable Markov decision processes for spoken dialog systems
    • Williams J. D., Young S. (2007) Partially observable Markov decision processes for spoken dialog systems. Computer Speech & Language 21(2): 393-422.
    • (2007) Computer Speech & Language , vol.21 , Issue.2 , pp. 393-422
    • Williams, J.D.1    Young, S.2
  • 69
    • 0036374229 scopus 로고    scopus 로고
    • Speeding up the convergence of value iteration in partially observable Markov decision processes
    • Zhang N. L., Zhang S. (2001) Speeding up the convergence of value iteration in partially observable Markov decision processes. Journal of Artificial Intelligence Research (JAIR) 14: 29-51.
    • (2001) Journal of Artificial Intelligence Research (JAIR) , vol.14 , pp. 29-51
    • Zhang, N.L.1    Zhang, S.2
  • 70
    • 0030230721 scopus 로고    scopus 로고
    • Using anytime algorithms in intelligent systems
    • Zilberstein S. (1996) Using anytime algorithms in intelligent systems. AI Magazine 17: 73-83.
    • (1996) AI Magazine , vol.17 , pp. 73-83
    • Zilberstein, S.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.