-
1
-
-
0003942195
-
-
Byte Books, Peterborough, NH
-
J. Albus, Brains, Behavior, and Robotics, Byte Books, Peterborough, NH, 1981.
-
(1981)
Brains, Behavior, and Robotics
-
-
Albus, J.1
-
2
-
-
33746599972
-
Reinforcement learning in motor control
-
M. A. Arbib, (ed.), MIT Press, Cambridge, MA
-
A. G. Barto, Reinforcement learning in motor control, in M. A. Arbib, (ed.), The Handbook of Brain Theory and Neural Networks, Second Edition, pp. 968-972, MIT Press, Cambridge, MA, 2003.
-
(2003)
The Handbook of Brain Theory and Neural Networks, Second Edition
, pp. 968-972
-
-
Barto, A.G.1
-
3
-
-
0020970738
-
Neuronlike elements that can solve difficult learning control problems
-
A. G. Barto, R. S. Sutton, and C. W. Anderson, Neuronlike elements that can solve difficult learning control problems, IEEE Trans. Systems, Man, and Cybernetics, vol. 13, pp. 835-846,1983.
-
(1983)
IEEE Trans. Systems, Man, and Cybernetics
, vol.13
, pp. 835-846
-
-
Barto, A.G.1
Sutton, R.S.2
Anderson, C.W.3
-
8
-
-
85152521744
-
A teaching method for reinforcement learning
-
Morgan Kaufmann, San Francisco, CA
-
J. A. Clouse and P. E. Utgoff, A teaching method for reinforcement learning, Proc. Ninth International Conference on Machine Learning, pp. 92-101, Morgan Kaufmann, San Francisco, CA, 1992.
-
(1992)
Proc. Ninth International Conference on Machine Learning
, pp. 92-101
-
-
Clouse, J.A.1
Utgoff, P.E.2
-
10
-
-
0028739953
-
Robot shaping: Developing autonomous agents through
-
M. Dorigo and M. Colombetti, Robot shaping: developing autonomous agents through learning, Artificial Intelligence, vol. 71, no. 2, pp. 321-370,1994.
-
(1994)
Learning, Artificial Intelligence
, vol.71
, Issue.2
, pp. 321-370
-
-
Dorigo, M.1
Colombetti, M.2
-
11
-
-
0025600638
-
A stochastic reinforcement learning algorithm for learning realvalued functions
-
V. Gullapalli, A stochastic reinforcement learning algorithm for learning realvalued functions, Neural Networks, vol. 3, no. 6, pp. 671-692, 1990.
-
(1990)
Neural Networks
, vol.3
, Issue.6
, pp. 671-692
-
-
Gullapalli, V.1
-
12
-
-
0031343489
-
A feedback control structure for on-line learning tasks
-
M. Huber and R. A. Grupen, A feedback control structure for on-line learning tasks, Robotics and Autonomous Systems, vol. 22, no. 3-4, pp. 303-315, 1997.
-
(1997)
Robotics and Autonomous Systems
, vol.22
, Issue.3-4
, pp. 303-315
-
-
Huber, M.1
Grupen, R.A.2
-
13
-
-
44049116478
-
Forward models: Supervised learning with a distal teacher
-
M. I. Jordan and D. E. Rumelhart, Forward models: Supervised learning with a distal teacher, Cognitive Science, vol. 16, no. 3, pp. 307-354,1992.
-
(1992)
Cognitive Science
, vol.16
, Issue.3
, pp. 307-354
-
-
Jordan, M.I.1
Rumelhart, D.E.2
-
14
-
-
0029697750
-
Building elementary robot skills from human demonstration
-
IEEE, Piscataway, NJ
-
M. Kaiser and R. Dillmann, Building elementary robot skills from human demonstration, Proc. IEEE International Conference on Robotics and Automation, pp. 2700-2705, IEEE, Piscataway, NJ, 1996.
-
(1996)
Proc. IEEE International Conference on Robotics and Automation
, pp. 2700-2705
-
-
Kaiser, M.1
Dillmann, R.2
-
15
-
-
0035977495
-
Robust reinforcement learning control with static and dynamic stability
-
R. M. Kretchmar, P. M. Young, C. W. Anderson, D. C. Hittle, M. L. Anderson, C. C. Delnero, and J. Tu, Robust reinforcement learning control with static and dynamic stability, International Journal of Robust and Nonlinear Control, vol. 11, pp. 1469-1500,2001.
-
(2001)
International Journal of Robust and Nonlinear Control
, vol.11
, pp. 1469-1500
-
-
Kretchmar, R.M.1
Young, P.M.2
Anderson, C.W.3
Hittle, D.C.4
Anderson, M.L.5
Delnero, C.C.6
Tu, J.7
-
16
-
-
0000123778
-
Self-improving reactive agents based on reinforcement learning, planning and teaching
-
L.-J. Lin, Self-improving reactive agents based on reinforcement learning, planning and teaching, Machine Learning, vol. 8, no. 3-4, pp. 293-321,1992.
-
(1992)
Machine Learning
, vol.8
, Issue.3-4
, pp. 293-321
-
-
Lin, L.-J.1
-
17
-
-
0029732210
-
Creating advice-taking reinforcement learners
-
R. Maclin and J. W. Shavlik, Creating advice-taking reinforcement learners, Machine Learning, vol. 22, no. 1-3, pp. 251-281,1996.
-
(1996)
Machine Learning
, vol.22
, Issue.1-3
, pp. 251-281
-
-
Maclin, R.1
Shavlik, J.W.2
-
19
-
-
0003200022
-
Sensory-motor primitives as a basis for imitation: Linking perception to action and biology to robotics
-
C. Nehaniv and K. Dautenhahn, (eds.), MIT Press, Cambridge, MA
-
M. J. Mataric, Sensory-motor primitives as a basis for imitation: linking perception to action and biology to robotics, in C. Nehaniv and K. Dautenhahn, (eds.), Imitation in Animals and Artifacts, MIT Press, Cambridge, MA, 2000.
-
(2000)
Imitation in Animals and Artifacts
-
-
Mataric, M.J.1
-
20
-
-
0141596576
-
Policy invariance under reward transformations: Theory and applications to reward shaping
-
Morgan Kaufmann, San Francisco, CA
-
A. Y. Ng, D. Harada, and S. Russell, Policy invariance under reward transformations: Theory and applications to reward shaping, Proc. Sixteenth International Conference on Machine Learning, pp. 278-287, Morgan Kaufmann, San Francisco, CA, 1999.
-
(1999)
Proc. Sixteenth International Conference on Machine Learning
, pp. 278-287
-
-
Ng, A.Y.1
Harada, D.2
Russell, S.3
-
21
-
-
0013530450
-
Lyapunov-constrained action sets for reinforcement learning
-
C. Brodley and A. Danyluk, (eds.), Morgan Kaufmann, San Francisco, CA
-
T. J. Perkins and A. G. Barto, Lyapunov-constrained action sets for reinforcement learning, in C. Brodley and A. Danyluk, (eds.), Proc. 18th International Conference on Machine Learning, pp. 409-416, Morgan Kaufmann, San Francisco, CA, 2001.
-
(2001)
Proc. 18Th International Conference on Machine Learning
, pp. 409-416
-
-
Perkins, T.J.1
Barto, A.G.2
-
23
-
-
0010276944
-
Implicit imitation in multiagent reinforcement learning
-
I. Bratko and S. Dzeroski, (eds.), Morgan Kaufmann, San Francisco, CA
-
B. Price and C. Boutilier, Implicit imitation in multiagent reinforcement learning, in I. Bratko and S. Dzeroski, (eds.), Proc. 16th International Conference on Machine Learning, pp. 325-334, Morgan Kaufmann, San Francisco, CA, 1999.
-
(1999)
Proc. 16Th International Conference on Machine Learning
, pp. 325-334
-
-
Price, B.1
Boutilier, C.2
-
24
-
-
0031231885
-
Experiments with reinforcement learning in problems with continuous state and action spaces
-
J. C. Santamaria, R. S. Sutton, and A. Ram, Experiments with reinforcement learning in problems with continuous state and action spaces, Adaptive Behavior, vol. 6, pp. 163-217, 1997.
-
(1997)
Adaptive Behavior
, vol.6
, pp. 163-217
-
-
Santamaria, J.C.1
Sutton, R.S.2
Ram, A.3
-
25
-
-
84898995067
-
Learning from demonstration
-
M. C. Mozer, M. I. Jordan, and T. Petsche, (eds.), MIT Press, Cambridge, MA
-
S. Schaal, Learning from demonstration, in M. C. Mozer, M. I. Jordan, and T. Petsche, (eds.), Advances In Neural Information Processing Systems 9, pp. 1040-1046, MIT Press, Cambridge, MA, 1997.
-
(1997)
Advances in Neural Information Processing Systems 9
, pp. 1040-1046
-
-
Schaal, S.1
-
26
-
-
0033151712
-
Is imitation learning the route to humanoid robots?
-
S. Schaal, Is imitation learning the route to humanoid robots? Trends in Cognitive Science, vol. 3, pp. 233-242,1999.
-
(1999)
Trends in Cognitive Science
, vol.3
, pp. 233-242
-
-
Schaal, S.1
-
27
-
-
0002933526
-
Linearization and gain-scheduling
-
W. S. Levine, (ed.), CRC Press, Boca Raton, FL
-
J. S. Shamma, Linearization and gain-scheduling, in W. S. Levine, (ed.), The Control Handbook, pp. 388-396, CRC Press, Boca Raton, FL, 1996.
-
(1996)
The Control Handbook
, pp. 388-396
-
-
Shamma, J.S.1
-
28
-
-
0029753630
-
-
S. P. Singh and R. S. Sutton, Reinforcement learning with replacing eligibility traces^ Machine Learning, vol. 22, no. 1-3, pp. 123-158, 1996.
-
(1996)
Reinforcement Learning with Replacing Eligibility Traces^ Machine Learning
, vol.22
, Issue.1-3
, pp. 123-158
-
-
Singh, S.P.1
Sutton, R.S.2
-
29
-
-
0036058423
-
Effective reinforcement learning for mobile robots
-
IEEE, Piscataway, NJ
-
W. D. Smart and L. P. Kaelbling, Effective reinforcement learning for mobile robots, Proc. IEEE International Conference on Robotics and Automation, pp. 3404-3410, IEEE, Piscataway, NJ, 2002.
-
(2002)
Proc. IEEE International Conference on Robotics and Automation
, pp. 3404-3410
-
-
Smart, W.D.1
Kaelbling, L.P.2
-
30
-
-
33847202724
-
Learning to predict by the method of temporal differences
-
R. S. Sutton, Learning to predict by the method of temporal differences, Machine Learning, vol. 3, pp. 9-44,1988.
-
(1988)
Machine Learning
, vol.3
, pp. 9-44
-
-
Sutton, R.S.1
-
32
-
-
34249833101
-
Q-leaming
-
C. J. C. H. Watkins and P. Dayan, Q-leaming, Machine Learning, vol. 8, no. 3-4, pp 279-292,1992.
-
(1992)
Machine Learning
, vol.8
, Issue.3-4
, pp. 279-292
-
-
Watkins, C.J.C.H.1
Dayan, P.2
-
33
-
-
0000337576
-
Simple statistical gradient-following algorithms for connectionist reinforcement learning
-
R. J. Williams, Simple statistical gradient-following algorithms for connectionist reinforcement learning, Machine Learning, vol. 8, pp. 229-256,1992.
-
(1992)
Machine Learning
, vol.8
, pp. 229-256
-
-
Williams, R.J.1
|