-
2
-
-
0002278788
-
Hierarchical reinforcement learning with the MAXQ value function decomposition
-
Dietterich, T. G. (2000). Hierarchical reinforcement learning with the MAXQ value function decomposition. Artificial Intelligence Research, 13, 227-303.
-
(2000)
Artificial Intelligence Research
, vol.13
, pp. 227-303
-
-
Dietterich, T.G.1
-
3
-
-
0036618011
-
Multiple model-based reinforcement learning
-
To appear in
-
Doya, K., Samejima, K., Katagiri, K., & Kawato, M. (2002). Multiple model-based reinforcement learning. To appear in Neural Computation.
-
(2002)
Neural Computation
-
-
Doya, K.1
Samejima, K.2
Katagiri, K.3
Kawato, M.4
-
4
-
-
0038517214
-
Equivalence notions and model minimization in Markov decision processes
-
To appear in
-
Givan, R., Dean, T., & Greig, M. (2003). Equivalence notions and model minimization in Markov decision processes. To appear in Artificial Intelligence.
-
(2003)
Artificial Intelligence
-
-
Givan, R.1
Dean, T.2
Greig, M.3
-
5
-
-
0034272032
-
Bounded-parameter Markov decision processes
-
Givan, R., Leach, S., & Dean, T. (2000). Bounded-parameter Markov decision processes. Artificial Intelligence, 122, 71-109.
-
(2000)
Artificial Intelligence
, vol.122
, pp. 71-109
-
-
Givan, R.1
Leach, S.2
Dean, T.3
-
7
-
-
0035487297
-
MOSAIC model for sensorimotor learning and control
-
Haruno, M., Wolpert, D. M., & Kawato, M. (2001). MOSAIC model for sensorimotor learning and control. Neural Computation, 13, 2201-2220.
-
(2001)
Neural Computation
, vol.13
, pp. 2201-2220
-
-
Haruno, M.1
Wolpert, D.M.2
Kawato, M.3
-
9
-
-
0000148778
-
A heuristic approach to the discovery of macro-operators
-
Iba, G. A. (1989). A heuristic approach to the discovery of macro-operators. Machine Learning, 3, 285-317.
-
(1989)
Machine Learning
, vol.3
, pp. 285-317
-
-
Iba, G.A.1
-
10
-
-
84898927961
-
Automated state abstraction for options using the u-tree algorithm
-
Cambridge, MA: MIT Press
-
Jonsson, A., & Barto, A. G. (2001). Automated state abstraction for options using the u-tree algorithm. Proceedings of Advances in Neural Information Processing Systems 13 (pp. 1054-1060). Cambridge, MA: MIT Press.
-
(2001)
Proceedings of Advances in Neural Information Processing Systems
, vol.13
, pp. 1054-1060
-
-
Jonsson, A.1
Barto, A.G.2
-
14
-
-
84956854078
-
Model minimization in hierarchical reinforcement learning
-
New York, NY: Springer-Verlag
-
Ravindran, B., & Barto, A. G. (2002). Model minimization in hierarchical reinforcement learning. Proceedings of the Fifth Symposium on Abstraction, Reformulation and Approximation (SARA 2002) (pp. 196-211). New York, NY: Springer-Verlag.
-
(2002)
Proceedings of the Fifth Symposium on Abstraction, Reformulation and Approximation (SARA 2002)
, pp. 196-211
-
-
Ravindran, B.1
Barto, A.G.2
-
16
-
-
0033170372
-
Between MDPs and Semi-MDPs: A framework for temporal abstraction in reinforcement learning
-
Sutton, R. S., Precup, D., & Singh, S. (1999). Between MDPs and Semi-MDPs: A framework for temporal abstraction in reinforcement learning. Artificial Intelligence, 112, 181-211.
-
(1999)
Artificial Intelligence
, vol.112
, pp. 181-211
-
-
Sutton, R.S.1
Precup, D.2
Singh, S.3
|