-
1
-
-
39649090194
-
Learning in non-stationary partially observable Markov decision processes
-
R. Jaulmes, J. Pineau, and D. Precup. Learning in non-stationary partially observable Markov decision processes. ECML Workshop, 2005.
-
(2005)
ECML Workshop
-
-
Jaulmes, R.1
Pineau, J.2
Precup, D.3
-
3
-
-
51649091499
-
Bayesian reinforcement learning in continuous POMDPs with application to robot navigation
-
Stephane Ross, Brahim Chaib-draa, and Joelle Pineau. Bayesian reinforcement learning in continuous POMDPs with application to robot navigation. In ICRA, 2008.
-
(2008)
ICRA
-
-
Ross, S.1
Chaib-Draa, B.2
Pineau, J.3
-
4
-
-
56449086386
-
Reinforcement learning with limited reinforcement: Using Bayes risk for active learning in POMDPs
-
Finale Doshi, Joelle Pineau, and Nicholas Roy. Reinforcement learning with limited reinforcement: Using Bayes risk for active learning in POMDPs. In International Conference on Machine Learning, volume 25, 2008.
-
(2008)
International Conference on Machine Learning
, vol.25
-
-
Doshi, F.1
Pineau, J.2
Roy, N.3
-
6
-
-
84862279024
-
Inverse optimal heuristic control for imitation learning
-
Nathan Ratliff, Brian Ziebart, Kevin Peterson, J. Andrew Bagnell, Martial Hebert, Anind K. Dey, and Siddhartha Srinivasa. Inverse optimal heuristic control for imitation learning. In Proc. AISTATS, pages 424-431, 2009.
-
(2009)
Proc. AISTATS
, pp. 424-431
-
-
Ratliff, N.1
Ziebart, B.2
Peterson, K.3
Bagnell, J.A.4
Hebert, M.5
Dey, A.K.6
Srinivasa, S.7
-
7
-
-
77950356463
-
Model-based Bayesian reinforcement learning in partially observable domains
-
P. Poupart and N. Vlassis. Model-based Bayesian reinforcement learning in partially observable domains. In ISAIM, 2008.
-
(2008)
ISAIM
-
-
Poupart, P.1
Vlassis, N.2
-
8
-
-
14344258433
-
A Bayesian framework for reinforcement learning
-
M. Strens. A Bayesian framework for reinforcement learning. In ICML, 2000.
-
(2000)
ICML
-
-
Strens, M.1
-
9
-
-
78649507911
-
A Bayesian sampling approach to exploration in reinforcement learning
-
John Asmuth, Lihong Li, Michael Littman, Ali Nouri, and David Wingate. A Bayesian sampling approach to exploration in reinforcement learning. In Uncertainty in Artificial Intelligence (UAI), 2009.
-
(2009)
Uncertainty in Artificial Intelligence (UAI)
-
-
Asmuth, J.1
Li, L.2
Littman, M.3
Nouri, A.4
Wingate, D.5
-
12
-
-
77958539351
-
The infinite partially observable Markov decision process
-
Y. Bengio, D. Schuurmans, J. Lafferty, C. K. I. Williams, and A. Culotta, editors
-
Finale Doshi-Velez. The infinite partially observable Markov decision process. In Y. Bengio, D. Schuurmans, J. Lafferty, C. K. I. Williams, and A. Culotta, editors, Advances in Neural Information Processing Systems 22, pages 477-485. 2009.
-
(2009)
Advances in Neural Information Processing Systems
, vol.22
, pp. 477-485
-
-
Doshi-Velez, F.1
-
14
-
-
33749249312
-
Hierarchical Dirichlet processes
-
Yee Whye Teh, Michael I. Jordan, Matthew J. Beal, and David M. Blei. Hierarchical Dirichlet processes. Journal of the American Statistical Association, 101:1566-1581, 2006.
-
(2006)
Journal of the American Statistical Association
, vol.101
, pp. 1566-1581
-
-
Teh, Y.W.1
Jordan, M.I.2
Beal, M.J.3
Blei, D.M.4
-
17
-
-
56449130659
-
Beam sampling for the infinite hidden Markov model
-
J. van Gael, Y. Saatci, Y. W. Teh, and Z. Ghahramani. Beam sampling for the infinite hidden Markov model. In ICML, volume 25, 2008.
-
(2008)
ICML
, vol.25
-
-
Van Gael, J.1
Saatci, Y.2
Teh, Y.W.3
Ghahramani, Z.4
-
18
-
-
84880772945
-
Point-based value iteration: An anytime algorithm for POMDPs
-
J. Pineau, G. Gordon, and S. Thrun. Point-based value iteration: An anytime algorithm for POMDPs. IJCAI, 2003.
-
(2003)
IJCAI
-
-
Pineau, J.1
Gordon, G.2
Thrun, S.3
-
20
-
-
85138579181
-
Learning policies for partially observable environments: Scaling up
-
M. L. Littman, A. R. Cassandra, and L. P. Kaelbling. Learning policies for partially observable environments: scaling up. ICML, 1995.
-
(1995)
ICML
-
-
Littman, M.L.1
Cassandra, A.R.2
Kaelbling, L.P.3
-
21
-
-
0026998041
-
Reinforcement learning with perceptual aliasing: The perceptual distinctions approach
-
AAAI Press
-
Lonnie Chrisman. Reinforcement learning with perceptual aliasing: The perceptual distinctions approach. In In Proceedings of the Tenth National Conference on Artificial Intelligence, pages 183-188. AAAI Press, 1992.
-
(1992)
Proceedings of the Tenth National Conference on Artificial Intelligence
, pp. 183-188
-
-
Chrisman, L.1
-
22
-
-
31144465830
-
Heuristic search value iteration for POMDPs
-
Banff, Alberta
-
T. Smith and R. Simmons. Heuristic search value iteration for POMDPs. In Proc. of UAI 2004, Banff, Alberta, 2004.
-
(2004)
Proc. of UAI 2004
-
-
Smith, T.1
Simmons, R.2
|