-
1
-
-
33847202724
-
Learning to predict by the methods of temporal differences
-
R. S. Sutton, "Learning to predict by the methods of temporal differences," Machine Learning, no. 3, pp. 9-44, 1988.
-
(1988)
Machine Learning
, vol.3
, pp. 9-44
-
-
Sutton, R.S.1
-
3
-
-
25844441312
-
Design of self-learning controllers using FYNESSE
-
T. Furuhashi, S. Tano, and H.A. Jacobsen, Eds. Physica, to appear
-
R. Schoknecht, M. Spott, and M. Riedmiller, "Design of self-learning controllers using FYNESSE," in Deep Fusion of Computational and Symbolic Processing, T. Furuhashi, S. Tano, and H.A. Jacobsen, Eds. Physica, 1999, to appear.
-
(1999)
Deep Fusion of Computational and Symbolic Processing
-
-
Schoknecht, R.1
Spott, M.2
Riedmiller, M.3
-
4
-
-
0000123778
-
Self-improving reactive agents based on reinforcement learning, planning and teaching
-
L.-J. Lin, "Self-improving reactive agents based on reinforcement learning, planning and teaching," Machine Learning, no. 8, pp. 293-321, 1992.
-
(1992)
Machine Learning
, vol.8
, pp. 293-321
-
-
Lin, L.-J.1
-
5
-
-
25844486799
-
Approaches for the integration of a priori knowledge into an autonomously learning control architecture
-
Aachen
-
M. Spott, R. Schoknecht, and M. Riedmiller, "Approaches for the integration of a priori knowledge into an autonomously learning control architecture," in Proceedings of EUFIT '99, Aachen, 1999.
-
(1999)
Proceedings of EUFIT '99
-
-
Spott, M.1
Schoknecht, R.2
Riedmiller, M.3
-
6
-
-
0003787146
-
-
Princeton University Press, Princeton, NJ
-
R. E. Bellman, Dynamic Programming, Princeton University Press, Princeton, NJ, 1957.
-
(1957)
Dynamic Programming
-
-
Bellman, R.E.1
-
8
-
-
0029210635
-
Learning to act using real-time dynamic programming
-
A. G. Barto, S. J. Bradtke, and S. P. Singh, "Learning to act using real-time dynamic programming," Artificial Intelligence, no. 72, pp. 81-138, 1995.
-
(1995)
Artificial Intelligence
, vol.72
, pp. 81-138
-
-
Barto, A.G.1
Bradtke, S.J.2
Singh, S.P.3
-
11
-
-
0003270924
-
Issues in using function ap proximation for reinforcement learning
-
Hillsdale, NJ, Dec., Lawrence Erlbaum Publisher
-
S. Thrun and A. Schwartz, "Issues in using function ap proximation for reinforcement learning," in Proceedings of the Fourth Connectionist Models Summer School, Hillsdale, NJ, Dec. 1993, Lawrence Erlbaum Publisher.
-
(1993)
Proceedings of the Fourth Connectionist Models Summer School
-
-
Thrun, S.1
Schwartz, A.2
-
12
-
-
0020970738
-
Neuron-like adaptive elements that can solve difficult learning control problems
-
A. G. Barto, R. S. Sutton, and C. W. Anderson, "Neuron-like adaptive elements that can solve difficult learning control problems," IEEE Transactions on Systems, Man, and Cybernetics, vol. 13, pp. 834-846, 1983.
-
(1983)
IEEE Transactions on Systems, Man, and Cybernetics
, vol.13
, pp. 834-846
-
-
Barto, A.G.1
Sutton, R.S.2
Anderson, C.W.3
-
13
-
-
0000676676
-
Learning to control an unstable system with forward modeling
-
D. S. Touretzky, Ed., Morgan Kaufmann, San Mateo, California
-
M.I. Jordan and R.A. Jacobs, "Learning to control an unstable system with forward modeling," in Advances in Neural Information Processing Systems, D. S. Touretzky, Ed., vol. 2, pp. 84-97. Morgan Kaufmann, San Mateo, California, 1989.
-
(1989)
Advances in Neural Information Processing Systems
, vol.2
, pp. 84-97
-
-
Jordan, M.I.1
Jacobs, R.A.2
-
14
-
-
0242580448
-
Variable resolution discretization in optimal control
-
submitted
-
R. Munos and A. Moore, "Variable Resolution Discretization in Optimal Control," Machine Learning, 1999, submitted.
-
(1999)
Machine Learning
-
-
Munos, R.1
Moore, A.2
-
15
-
-
0346872401
-
Fynesse: A hybrid architecture for selflearning control
-
I. Cloete and J. Zurada, Eds. MIT Press, (to appear)
-
M. Riedmiller, M. Spott, and J. Weisbrod, "Fynesse: A hybrid architecture for selflearning control," in Knowledge-Based Neurocomputing, I. Cloete and J. Zurada, Eds. MIT Press, 1999, (to appear).
-
(1999)
Knowledge-Based Neurocomputing
-
-
Riedmiller, M.1
Spott, M.2
Weisbrod, J.3
-
16
-
-
0001133021
-
Generalization in reinforcement learning: Safely approximating the value function
-
Morgan Kaufmann
-
Boyan and Moore, "Generalization in reinforcement learning: Safely approximating the value function," in Advances in Neural Information Processing Systems 7. 1995, Morgan Kaufmann.
-
(1995)
Advances in Neural Information Processing Systems
, pp. 7
-
-
Boyan1
Moore2
|