메뉴 건너뛰기




Volumn 14, Issue , 2001, Pages 29-51

Speeding up the convergence of value iteration in partially observable Markov decision processes

Author keywords

[No Author keywords available]

Indexed keywords

ALGORITHMS; CONVERGENCE OF NUMERICAL METHODS; ITERATIVE METHODS; MARKOV PROCESSES; MATHEMATICAL MODELS;

EID: 0036374229     PISSN: 10769757     EISSN: None     Source Type: Journal    
DOI: 10.1613/jair.761     Document Type: Article
Times cited : (105)

References (30)
  • 1
    • 50549213583 scopus 로고
    • Optimal control of Markov decision processes with the incomplete state estimation
    • Aström, K. J. (1965). Optimal control of Markov decision processes with the incomplete state estimation. Journal of Computer and System Sciences, 10, 174-205.
    • (1965) Journal of Computer and System Sciences , vol.10 , pp. 174-205
    • Aström, K.J.1
  • 7
    • 0021486586 scopus 로고
    • The optimal search for a moving target when the search path is constrained
    • Eagle, J. N. (1984). The optimal search for a moving target when the search path is constrained. Operations Research, 32(5), 1107-1115.
    • (1984) Operations Research , vol.32 , Issue.5 , pp. 1107-1115
    • Eagle, J.N.1
  • 12
    • 0001770240 scopus 로고    scopus 로고
    • Value function approximations for partially observable Markov decision processes
    • Hauskrecht, M. (2000). Value function approximations for partially observable Markov decision processes, Journal of Artificial Intelligence Research, 13, 33-95.
    • (2000) Journal of Artificial Intelligence Research , vol.13 , pp. 33-95
    • Hauskrecht, M.1
  • 13
    • 0003596835 scopus 로고
    • Efficient dynamic-programming updates in partially observable Markov decision processes
    • Brown University
    • Littman, M. L., Cassandra, A. R. and Kaelbling, L. P. (1995a). Efficient dynamic-programming updates in partially observable Markov decision processes. Technical Report CS-95-19, Brown University.
    • (1995) Technical Report CS-95-19
    • Littman, M.L.1    Cassandra, A.R.2    Kaelbling, L.P.3
  • 17
    • 0000494894 scopus 로고
    • Computationally feasible bounds for partially observed Markov decision processes
    • Lovejoy, W. S. (1991). Computationally feasible bounds for partially observed Markov decision processes. Operations Research, 39, 192-175.
    • (1991) Operations Research , vol.39 , pp. 192-1175
    • Lovejoy, W.S.1
  • 18
    • 0001095688 scopus 로고
    • Suboptimal policies with bounds for parameter adaptive decision processes
    • Lovejoy, W. S. (1993). Suboptimal policies with bounds for parameter adaptive decision processes. Operations Research, 41, 583-599.
    • (1993) Operations Research , vol.41 , pp. 583-599
    • Lovejoy, W.S.1
  • 19
    • 0019909899 scopus 로고
    • A survey of partially observable Markov decision processes: Theory, models, and algorithms
    • Monahan, G. E. (1982). A survey of partially observable Markov decision processes: theory, models, and algorithms. Management Science, 28 (1), 1-16.
    • (1982) Management Science , vol.28 , Issue.1 , pp. 1-16
    • Monahan, G.E.1
  • 22
    • 0019037868 scopus 로고
    • Optimal infinite-horizon undiscounted control of finite probabilistic systems
    • Platzman, L. K.(1980). Optimal infinite-horizon undiscounted control of finite probabilistic systems. SIAM Journal of Control and Optimization, 18, 362-380.
    • (1980) SIAM Journal of Control and Optimization , vol.18 , pp. 362-380
    • Platzman, L.K.1
  • 23
    • 77957101448 scopus 로고
    • Markov decision processes
    • D. P. Heyman and M. J. Sobel (eds.), Elsevier Science Publishers
    • Puterman, M. L. (1990), Markov decision processes, in D. P. Heyman and M. J. Sobel (eds.), Handbooks in OR & MS., Vol. 2, 331-434, Elsevier Science Publishers.
    • (1990) Handbooks in or & MS , vol.2 , pp. 331-434
    • Puterman, M.L.1
  • 24
    • 0015658957 scopus 로고
    • The optimal control of partially observable processes over a finite horizon
    • Smallwood, R. D. and Sondik, E. J. (1973). The optimal control of partially observable processes over a finite horizon. Operations Research, 21, 1071-1088.
    • (1973) Operations Research , vol.21 , pp. 1071-1088
    • Smallwood, R.D.1    Sondik, E.J.2
  • 26
    • 0015658957 scopus 로고
    • The optimal control of partially observable Markov processes over the infinite horizon
    • Sondik, E. J. (1978). The optimal control of partially observable Markov processes over the infinite horizon, Operations Research, 21, 1071-1088.
    • (1978) Operations Research , vol.21 , pp. 1071-1088
    • Sondik, E.J.1
  • 27
    • 0024739631 scopus 로고
    • Solution procedures for partially observed Markov decision processes
    • White, C. C. III and Scherer, W. T. (1989). Solution procedures for partially observed Markov decision processes, Operations Research, 37(5), 791-797.
    • (1989) Operations Research , vol.37 , Issue.5 , pp. 791-797
    • White III, C.C.1    Scherer, W.T.2
  • 29
    • 85016628903 scopus 로고    scopus 로고
    • A model approximation scheme for planning in stochastic domains
    • Zhang, N. L. and W. Liu (1997). A model approximation scheme for planning in stochastic domains, Journal of Artificial Intelligence Research, 7, 199-230.
    • (1997) Journal of Artificial Intelligence Research , vol.7 , pp. 199-230
    • Zhang, N.L.1    Liu, W.2
  • 30
    • 84867833986 scopus 로고    scopus 로고
    • A POMDP approximation algorithm that anticipates the need to observe
    • To appear in Proceedings of the Pacific Rim Conference on Artificial Intelligence (PRICAI-2000), New York: Springer-Verlag
    • Zubek, V. B. and Dietterich, T. G.(2000). A POMDP approximation algorithm that anticipates the need to observe. To appear in Proceedings of the Pacific Rim Conference on Artificial Intelligence (PRICAI-2000), Lecture Notes in Computer Science, New York: Springer-Verlag.
    • (2000) Lecture Notes in Computer Science
    • Zubek, V.B.1    Dietterich, T.G.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.