-
1
-
-
84899003140
-
Multi-time models for temporally abstract planning
-
The MIT Press
-
D. Precup and R. S. Sutton, "Multi-time models for temporally abstract planning," in NIPS 10, The MIT Press, 1998.
-
(1998)
NIPS 10
-
-
Precup, D.1
Sutton, R.S.2
-
2
-
-
0003899594
-
Between mdps and semi-mdps: Learning, planning, and representing knowledge at multiple temporal scales
-
Amherst, MA
-
R. S. Sutton, D. Precup, and S. Singh, "Between MDPs and Semi-MDPs: Learning, planning, and representing knowledge at multiple temporal scales," tech. rep., Univ. Mass., Dept. Comp. Inf. Sci., Amherst, MA, 1998.
-
(1998)
Tech. Rep., Univ. Mass., Dept. Comp. Inf. Sci
-
-
Sutton, R.S.1
Precup, D.2
Singh, S.3
-
3
-
-
84898956770
-
Reinforcement learning with hierarchies of machines
-
The MIT Press
-
R. Parr and S. Russell, "Reinforcement learning with hierarchies of machines," in NIPS-10, The MIT Press, 1998.
-
(1998)
NIPS-10
-
-
Parr, R.1
Russell, S.2
-
4
-
-
0001027894
-
Transfer of learning by composing solutions of elemental sequential tasks
-
S. P. Singh, "Transfer of learning by composing solutions of elemental sequential tasks," Machine Learning, vol. 8, p. 323, 1992.
-
(1992)
Machine Learning
, vol.8
, pp. 323
-
-
Singh, S.P.1
-
5
-
-
0000908087
-
Hierarchical reinforcement learning: Preliminary results
-
Morgan Kaufmann
-
L. P. Kaelbling, "Hierarchical reinforcement learning: Preliminary results," in Proceedings ICML-10, pp. 167-173, Morgan Kaufmann, 1993.
-
(1993)
Proceedings ICML-10
, pp. 167-173
-
-
Kaelbling, L.P.1
-
6
-
-
84867797982
-
Hierarchical solution of Markov decision processes using macro-actions
-
Providence, RI
-
M. Hauskrecht, N. Meuleau, C. Boutilier, L. Kaelbling, and T. Dean, "Hierarchical solution of Markov decision processes using macro-actions," tech. rep., Brown Univ., Dept. Comp. Sci., Providence, RI, 1998.
-
(1998)
Tech. Rep., Brown Univ., Dept. Comp. Sci
-
-
Hauskrecht, M.1
Meuleau, N.2
Boutilier, C.3
Kaelbling, L.4
Dean, T.5
-
7
-
-
0001234682
-
Feudal reinforcement learning
-
San Francisco, CA: Morgan Kaufmann
-
P. Dayan and G. Hinton, "Feudal reinforcement learning," in NIPS-5, pp. 271-278, San Francisco, CA: Morgan Kaufmann, 1993.
-
(1993)
NIPS-5
, pp. 271-278
-
-
Dayan, P.1
Hinton, G.2
-
8
-
-
0001806701
-
The MAXQ method for hierarchical reinforcement learning
-
Morgan Kaufmann
-
T. G. Dietterich, "The MAXQ method for hierarchical reinforcement learning," in ICML-15, Morgan Kaufmann, 1998.
-
(1998)
ICML-15
-
-
Dietterich, T.G.1
-
9
-
-
0346087506
-
Convergence results for single-step on-policy reinforcement-learning algorithms
-
Boulder, CO
-
S. Singh, T. Jaakkola, M. L. Littman, and C. Szpesvari, "Convergence results for single-step on-policy reinforcement-learning algorithms," tech. rep., Univ. Col, Dept. Comp. Sci., Boulder, CO, 1998.
-
(1998)
Tech. Rep., Univ. Col, Dept. Comp. Sci
-
-
Singh, S.1
Jaakkola, T.2
Littman, M.L.3
Szpesvari, C.4
-
11
-
-
0000439891
-
On the convergence of stochastic iterative dynamic programming algorithms
-
T. Jaakkola, M. I. Jordan, and S. P. Singh, "On the convergence of stochastic iterative dynamic programming algorithms," Neur. Comp., vol. 6, no. 6, pp. 1185-1201, 1994.
-
(1994)
Neur. Comp.
, vol.6
, Issue.6
, pp. 1185-1201
-
-
Jaakkola, T.1
Jordan, M.I.2
Singh, S.P.3
-
12
-
-
85166207010
-
Exploiting structure in policy construction
-
C. Boutilier, R. Dearden, and M. Goldszmidt, "Exploiting structure in policy construction," in Proceedings IJCAI-95, pp. 1104-1111, 1995.
-
(1995)
Proceedings IJCAI-95
, pp. 1104-1111
-
-
Boutilier, C.1
Dearden, R.2
Goldszmidt, M.3
|