-
1
-
-
50549213583
-
Optimal control of Markov decision processes with incomplete state estimation
-
K. J. Åström. Optimal control of Markov decision processes with incomplete state estimation. Journal of Mathematical Analysis and Applications, 10:174-205, 1965.
-
(1965)
Journal of Mathematical Analysis and Applications
, vol.10
, pp. 174-205
-
-
Åström, K.J.1
-
2
-
-
1942514241
-
Scaling internal-state policy-gradient methods for POMDPs
-
Sydney, Australia
-
D. Aberdeen and J. Baxter. Scaling internal-state policy-gradient methods for POMDPs. In ICML, pages 3-10, Sydney, Australia, 2002.
-
(2002)
ICML
, pp. 3-10
-
-
Aberdeen, D.1
Baxter, J.2
-
4
-
-
84880705310
-
A decision-theoretic approach to task assistance for persons with dementia
-
J. Boger, P. Poupart, J. Hoey, C. Boutilier, G. Fernie, and Alex Mihailidis. A decision-theoretic approach to task assistance for persons with dementia. In Proc. IJCAI, Edinburgh, 2005.
-
Proc. IJCAI, Edinburgh, 2005
-
-
Boger, J.1
Poupart, P.2
Hoey, J.3
Boutilier, C.4
Fernie, G.5
Mihailidis, A.6
-
5
-
-
84880747107
-
Acting optimally in partially observable stochastic domains
-
A. R. Cassandra, L. P. Kaelbling, and M. L. Littman. Acting optimally in partially observable stochastic domains. In AAAI, Seattle, WA, 1994.
-
AAAI, Seattle, WA, 1994
-
-
Cassandra, A.R.1
Kaelbling, L.P.2
Littman, M.L.3
-
9
-
-
84947403595
-
Probability inequalities for sums of bounded random variables
-
W. Hoeffding. Probability inequalities for sums of bounded random variables. Journal of the American Statistical Association, 58(301):13-30, 1963.
-
(1963)
Journal of the American Statistical Association
, vol.58
, Issue.301
, pp. 13-30
-
-
Hoeffding, W.1
-
12
-
-
0032073263
-
Planning and acting in partially observable stochastic domains
-
Leslie Pack Kaelbling, Michael Littman, and Anthony R. Cassandra. Planning and acting in partially observable stochastic domains. Artificial Intelligence, 101:99-134, 1998.
-
(1998)
Artificial Intelligence
, vol.101
, pp. 99-134
-
-
Kaelbling, L.P.1
Littman, M.2
Cassandra, A.R.3
-
13
-
-
33646430192
-
Learning finite-state controllers for partially observable environments
-
Stockholm
-
N. Meuleau, L. Peshkin, K.-E. Kim, and L. P. Kaelbling. Learning finite-state controllers for partially observable environments. In UAI, pages 427-436, Stockholm, 1999.
-
(1999)
UAI
, pp. 427-436
-
-
Meuleau, N.1
Peshkin, L.2
Kim, K.-E.3
Kaelbling, L.P.4
-
14
-
-
4644247530
-
The use of computer vision in an intelligent environment to support aging-inplace, safety, and independence in the home
-
A. Mihailidis, B. Carmichael, and J. Boger. The use of computer vision in an intelligent environment to support aging-inplace, safety, and independence in the home. IEEE Trans. on Information Technology in Biomedicine (Spec. Issue on Pervasive Healthcare), 8(3):1-11, 2004.
-
(2004)
IEEE Trans. on Information Technology in Biomedicine (Spec. Issue on Pervasive Healthcare)
, vol.8
, Issue.3
, pp. 1-11
-
-
Mihailidis, A.1
Carmichael, B.2
Boger, J.3
-
15
-
-
0036931186
-
Experiences with a mobile robotic guide for the elderly
-
Edmonton, AB
-
M. Montemerlo, J. Pineau, N. Roy, S. Thrun, and V. Verma. Experiences with a mobile robotic guide for the elderly. In AAAI, pages 587-592, Edmonton, AB, 2002.
-
(2002)
AAAI
, pp. 587-592
-
-
Montemerlo, M.1
Pineau, J.2
Roy, N.3
Thrun, S.4
Verma, V.5
-
16
-
-
0141819580
-
PEGASUS: A policy search method for large MDPs and POMDPs
-
Stanford, CA
-
A. Y. Ng and M. Jordan. PEGASUS: A policy search method for large MDPs and POMDPs. In UAI, pages 406-415, Stanford, CA, 2000.
-
(2000)
UAI
, pp. 406-415
-
-
Ng, A.Y.1
Jordan, M.2
-
17
-
-
84880772945
-
Point-based value iteration: An anytime algorithm for POMDPs
-
J. Pineau, G. Gordon, and S. Thrun. Point-based value iteration: an anytime algorithm for POMDPs. In IJCAI, Acapulco, Mexico, 2003.
-
IJCAI, Acapulco, Mexico, 2003
-
-
Pineau, J.1
Gordon, G.2
Thrun, S.3
-
20
-
-
85156196231
-
Exponential family PCA for belief compression in POMDPs
-
Vancouver, BC
-
N. Roy and G. Gordon. Exponential family PCA for belief compression in POMDPs. In NIPS, pages 1635-1642, Vancouver, BC, 2002.
-
(2002)
NIPS
, pp. 1635-1642
-
-
Roy, N.1
Gordon, G.2
-
22
-
-
0015658957
-
The optimal control of partially observable Markov processes over a finite horizon
-
R. Smallwood and E. Sondik. The optimal control of partially observable Markov processes over a finite horizon. Operations Research, 21:1071-1088, 1973.
-
(1973)
Operations Research
, vol.21
, pp. 1071-1088
-
-
Smallwood, R.1
Sondik, E.2
-
23
-
-
84898978676
-
Monte Carlo POMDPs
-
Denver
-
S. Thrun. Monte Carlo POMDPs. In NIPS, pages 1064-1070, Denver, 1999.
-
(1999)
NIPS
, pp. 1064-1070
-
-
Thrun, S.1
-
25
-
-
33750709342
-
-
Technical Report CUED/F-INFEG/TR.520, Cambridge University, Engineering Department
-
J. Williams, P. Poupart, and S. Young. Using factored Markov decision processes with continuous observations for dialogue management. Technical Report CUED/F-INFEG/TR.520, Cambridge University, Engineering Department, 2005.
-
(2005)
Using Factored Markov Decision Processes with Continuous Observations for Dialogue Management
-
-
Williams, J.1
Poupart, P.2
Young, S.3
-
27
-
-
0036374229
-
Speeding up the convergence of value iteration in partially observable Markov decision processes
-
N. Zhang and W. Zhang. Speeding up the convergence of value-iteration in partially observable Markov decision processes. Journal of Artificial Intelligence Research, 14:29-51, 2001. (Pubitemid 33738058)
-
(2001)
Journal of Artificial Intelligence Research
, vol.14
, pp. 29-51
-
-
Zhang, N.L.1
Zhang, W.2
|