메뉴 건너뛰기




Volumn 7, Issue , 2006, Pages 2329-2367

Point-based value iteration for continuous POMDPs

Author keywords

Continuous action space; Continuous observation space; Continuous state space; Partially observable Markov decision processes; Planning under uncertainty; Point based value iteration

Indexed keywords

COMPUTATIONAL COMPLEXITY; ITERATIVE METHODS; LEARNING ALGORITHMS; MARKOV PROCESSES; OPTIMIZATION;

EID: 33750724397     PISSN: 15337928     EISSN: 15337928     Source Type: Journal    
DOI: None     Document Type: Article
Times cited : (210)

References (59)
  • 1
    • 50549213583 scopus 로고
    • Optimal control of Markov decision processes with incomplete state estimation
    • K. J. Åström. Optimal control of Markov decision processes with incomplete state estimation. Journal of Mathematical Analysis and Applications, 10:174-205, 1965.
    • (1965) Journal of Mathematical Analysis and Applications , vol.10 , pp. 174-205
    • Åström, K.J.1
  • 10
    • 0030349220 scopus 로고    scopus 로고
    • Computing optimal policies for partially observable decision processes using compact representations
    • Portland, OR
    • C. Boutilier and D. Poole. Computing optimal policies for partially observable decision processes using compact representations. In Proceedings of the National Conference on Artificial Intelligence, pages 1168-1175, Portland, OR, 1996.
    • (1996) Proceedings of the National Conference on Artificial Intelligence , pp. 1168-1175
    • Boutilier, C.1    Poole, D.2
  • 15
    • 84898804274 scopus 로고    scopus 로고
    • Active gesture recognition using partially observable Markov decision processes
    • Vienna, Austria
    • T. Darrell and A. P. Pentland. Active gesture recognition using partially observable Markov decision processes. In IEEE International Conference on Pattern Recognition, pages 984-988, Vienna, Austria, 1996.
    • (1996) IEEE International Conference on Pattern Recognition , pp. 984-988
    • Darrell, T.1    Pentland, A.P.2
  • 21
    • 0344445520 scopus 로고    scopus 로고
    • Adapting the sample size in particle niters through kld-sampling
    • D. Fox. Adapting the sample size in particle niters through kld-sampling. International Journal of Robotics Research, 22(10-11):985-1004, 2003.
    • (2003) International Journal of Robotics Research , vol.22 , Issue.10-11 , pp. 985-1004
    • Fox, D.1
  • 24
    • 0032136153 scopus 로고    scopus 로고
    • Condensation - Conditional density propagation for visual tracking
    • M. Isard and A. Blake. Condensation - conditional density propagation for visual tracking. International Journal of Computer Vision, 29(1):5-28, 1998.
    • (1998) International Journal of Computer Vision , vol.29 , Issue.1 , pp. 5-28
    • Isard, M.1    Blake, A.2
  • 25
    • 0032073263 scopus 로고    scopus 로고
    • Planning and acting in partially observable stochastic domains
    • L. P. Kaelbling, M. L. Littman, and A. R. Cassandra. Planning and acting in partially observable stochastic domains. Artificial Intelligence, 101(1-2):99-134, 1998.
    • (1998) Artificial Intelligence , vol.101 , Issue.1-2 , pp. 99-134
    • Kaelbling, L.P.1    Littman, M.L.2    Cassandra, A.R.3
  • 27
    • 0032596468 scopus 로고    scopus 로고
    • On the undecidability of probabilistic planning and infinitehorizon partially observable Markov decision problems
    • Orlando, FL
    • O. Madani, S. Hanks, and A. Condon. On the undecidability of probabilistic planning and infinitehorizon partially observable Markov decision problems. In Proceedings of the National Conference on Artificial Intelligence, pages 541-548, Orlando, FL, 1999.
    • (1999) Proceedings of the National Conference on Artificial Intelligence , pp. 541-548
    • Madani, O.1    Hanks, S.2    Condon, A.3
  • 29
    • 0019909899 scopus 로고
    • A survey of partially observable Markov decision processes: Theory, models, and algorithms
    • G. E. Monahan. A survey of partially observable Markov decision processes: Theory, models, and algorithms. Management Science, 28(1): 1-16, 1982.
    • (1982) Management Science , vol.28 , Issue.1 , pp. 1-16
    • Monahan, G.E.1
  • 37
    • 84959528732 scopus 로고    scopus 로고
    • Robot planning in partially observable continuous domains
    • MIT, Cambridge, MA
    • J. M. Porta, M. T. J. Spaan, and N. Vlassis. Robot planning in partially observable continuous domains. In Robotics: Science and Systems I, pages 217-224, MIT, Cambridge, MA, 2005.
    • (2005) Robotics: Science and Systems I , pp. 217-224
    • Porta, J.M.1    Spaan, M.T.J.2    Vlassis, N.3
  • 56
    • 0029250080 scopus 로고
    • Reinforcement learning of non-Markov decision processes
    • Steven D. Whitehead and Long-Ji Lin. Reinforcement learning of non-Markov decision processes. Artificial Intelligence, 73(1-2):271-306, 1995.
    • (1995) Artificial Intelligence , vol.73 , Issue.1-2 , pp. 271-306
    • Whitehead, S.D.1    Lin, L.-J.2
  • 58
    • 33645958652 scopus 로고    scopus 로고
    • Planning and acting under uncertainty: A new model for spoken dialogue systems
    • Seattle, WA
    • B. Zhang, Q. Cai, J. Mao, and B. Guo. Planning and acting under uncertainty: a new model for spoken dialogue systems. In Proceedings of Uncertainty in Artificial Intelligence, pages 572-579, Seattle, WA, 2001.
    • (2001) Proceedings of Uncertainty in Artificial Intelligence , pp. 572-579
    • Zhang, B.1    Cai, Q.2    Mao, J.3    Guo, B.4
  • 59
    • 0036374229 scopus 로고    scopus 로고
    • Speeding up the convergence of value iteration in partially observable Markov decision processes
    • N. L. Zhang and W. Zhang. Speeding up the convergence of value iteration in partially observable Markov decision processes. Journal of Artificial Intelligence Research, 14:29-51, 2001.
    • (2001) Journal of Artificial Intelligence Research , vol.14 , pp. 29-51
    • Zhang, N.L.1    Zhang, W.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.