-
3
-
-
0029539342
-
Dynamics of behaviour: Theory and application for autonomous robot architecture
-
Schoner, G., Dose, M., and Engels, C., 1995. Dynamics of behaviour:theory and application for autonomous robot architecture. Robotics Autonomous Syst., 16:213–245.
-
(1995)
Robotics Autonomous Syst.
, vol.16
, pp. 213-245
-
-
Schoner, G.1
Dose, M.2
Engels, C.3
-
5
-
-
3042617895
-
Autonomous reaching and obstacle avoidance with anthropomorphic arm of a robotics assistant using the attractor dynamics approach
-
New Orleans, LA
-
Iossifidis, I., and Schoner, G., 2004. “ Autonomous reaching and obstacle avoidance with anthropomorphic arm of a robotics assistant using the attractor dynamics approach ”. In Proc. IEEE Int. Conf. on Robotics and Automation 4295–4300. New Orleans, LA
-
(2004)
Proc. IEEE Int. Conf. on Robotics and Automation
, pp. 4295-4300
-
-
Iossifidis, I.1
Schoner, G.2
-
6
-
-
33845640093
-
Programmable central pattern generators: A application to biped locomotion control
-
Orlando, FL
-
Righetti, L., and Ijspeert, A., 2006. “ Programmable central pattern generators:a application to biped locomotion control ”. In Proc. IEEE Int. Conf. on Robotics and Automation 1585–1590. Orlando, FL
-
(2006)
Proc. IEEE Int. Conf. on Robotics and Automation
, pp. 1585-1590
-
-
Righetti, L.1
Ijspeert, A.2
-
8
-
-
34250620650
-
Learning movement primitives
-
Siena
-
Schaal, S., Peters, J., Nakanishi, J., and Ijspeert, A., 2003. “ Learning movement primitives ”. In Proc. Int. Symp. on Robotics Research 1805–1815. Siena
-
(2003)
Proc. Int. Symp. on Robotics Research
, pp. 1805-1815
-
-
Schaal, S.1
Peters, J.2
Nakanishi, J.3
Ijspeert, A.4
-
9
-
-
0035979437
-
Acquisition of stand-up behavior by a real robot using hierarchical reinforcement learning
-
Morimoto, J., and Doya, K., 2001. Acquisition of stand-up behavior by a real robot using hierarchical reinforcement learning. Robotics Autonomous Syst., 36:37–51.
-
(2001)
Robotics Autonomous Syst.
, vol.36
, pp. 37-51
-
-
Morimoto, J.1
Doya, K.2
-
10
-
-
14044272927
-
Reinforcement learning for motion control of humanoid robots
-
Sendai
-
Iida, S., Kanoh, M., Kato, S., and Itoh, H., 2004. “ Reinforcement learning for motion control of humanoid robots ”. In Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and Systems 353–358. Sendai
-
(2004)
Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and Systems
, pp. 353-358
-
-
Iida, S.1
Kanoh, M.2
Kato, S.3
Itoh, H.4
-
11
-
-
85150714688
-
Reinforcement learning methods for continuous-time Markov decision problems
-
Denver
-
Bratke, S. J., and Duff, M. O., 1994. “ Reinforcement learning methods for continuous-time Markov decision problems ”. In Proc. Neural Information Processing Systems Conf 393–400. Denver
-
(1994)
Proc. Neural Information Processing Systems Conf
, pp. 393-400
-
-
Bratke, S.J.1
Duff, M.O.2
-
12
-
-
0033629916
-
Reinforcement learning in continuous time and space
-
Doya, K., 2000. Reinforcement learning in continuous time and space. Neural Comput., 12:219–245.
-
(2000)
Neural Comput.
, vol.12
, pp. 219-245
-
-
Doya, K.1
-
15
-
-
34447553096
-
Reinforcement learning for humanoid robotics
-
München
-
Peters, J., Vijayakumar, S., and Schaal, S., 2003. “ Reinforcement learning for humanoid robotics ”. In Proc. IEEE-RAS Int. Conf. on Humanoid Robots (Humanoids2003) 225–230. München
-
(2003)
Proc. IEEE-RAS Int. Conf. on Humanoid Robots (Humanoids2003)
, pp. 225-230
-
-
Peters, J.1
Vijayakumar, S.2
Schaal, S.3
-
16
-
-
33646413135
-
Natural actor—critic
-
Porto
-
Peters, J., Vijayakumar, S., and Schaal, S., 2005. “ Natural actor—critic ”. In Proc. 16th Eur. Conf. on Machine Learning 280–291. Porto
-
(2005)
Proc. 16th Eur. Conf. on Machine Learning
, pp. 280-291
-
-
Peters, J.1
Vijayakumar, S.2
Schaal, S.3
-
17
-
-
84873015924
-
Least-squares policy evaluation algorithms with linear function approximation
-
Cambridge
-
Nedic, A., and Bertsekas, D., 2001. “ Least-squares policy evaluation algorithms with linear function approximation ”. In LIDS Report LIDS-P-2537, Dec. 2001 Cambridge
-
(2001)
LIDS Report LIDS-P-2537, Dec. 2001
-
-
Nedic, A.1
Bertsekas, D.2
-
18
-
-
0000337576
-
Simple statistical gradient-following algorithms for connectionist reinforcement learning
-
Williams, R., 1992. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learn., 8:229–256.
-
(1992)
Machine Learn.
, vol.8
, pp. 229-256
-
-
Williams, R.1
-
19
-
-
84898939480
-
Policy gradient methods for reinforcement learning with function approximation
-
Sutton, R. S., McAllester, D., Singh, S., and Mansour, Y., 2000. Policy gradient methods for reinforcement learning with function approximation. Adv. Neural Information Process. Syst., 12:1057–1064.
-
(2000)
Adv. Neural Information Process. Syst.
, vol.12
, pp. 1057-1064
-
-
Sutton, R.S.1
McAllester, D.2
Singh, S.3
Mansour, Y.4
-
21
-
-
0000396062
-
Natural gradient works efficiently in learning
-
Amari, S., 2000. Natural gradient works efficiently in learning. Neural Comput., 10:251–276.
-
(2000)
Neural Comput.
, vol.10
, pp. 251-276
-
-
Amari, S.1
-
22
-
-
33749243349
-
Autonomous shaping: Knowledge transfer in reinforcement learning
-
Pittsburgh, PA
-
Konidaris, G., and Barto, A., 2006. “ Autonomous shaping:knowledge transfer in reinforcement learning ”. In Proc. Int. Conf. on Machine Learning 497–504. Pittsburgh, PA
-
(2006)
Proc. Int. Conf. on Machine Learning
, pp. 497-504
-
-
Konidaris, G.1
Barto, A.2
-
23
-
-
33749242451
-
Using inaccurate models in reinforcement learning
-
Pittsburgh, PA
-
Abbeel, P., and Quigley, M., 2006. “ Using inaccurate models in reinforcement learning ”. In Proc. Int. Conf. on Machine Learning 9–16. Pittsburgh, PA
-
(2006)
Proc. Int. Conf. on Machine Learning
, pp. 9-16
-
-
Abbeel, P.1
Quigley, M.2
-
24
-
-
33749261645
-
An intrinsic reward mechanism for efficient exploration
-
Pittsburgh, PA
-
Simsek, O., and Barto, A., 2006. “ An intrinsic reward mechanism for efficient exploration ”. In Proc. Int. Conf. on Machine Learning 841–848. Pittsburgh, PA
-
(2006)
Proc. Int. Conf. on Machine Learning
, pp. 841-848
-
-
Simsek, O.1
Barto, A.2
-
25
-
-
34047104556
-
Learning dynamical system modulation for constraint reaching tasks
-
Genova
-
Hersch, M., Guenter, F., Calinon, S., and Billard, A., 2006. “ Learning dynamical system modulation for constraint reaching tasks ”. In Proc. IEEE-RAS International Conference on Humanoid Robots 444–449. Genova
-
(2006)
Proc. IEEE-RAS International Conference on Humanoid Robots
, pp. 444-449
-
-
Hersch, M.1
Guenter, F.2
Calinon, S.3
Billard, A.4
-
26
-
-
34047173490
-
On learning, representing and generalizing a task in a humanoid robot
-
Calinon S., Guenter F., and Billard A., On learning, representing and generalizing a task in a humanoid robot, IEEE Trans. Syst. Man Cybernet. B (Special Issue on Robot Learning by Observation, Demonstration and Imitation) 37, in press (2007).
-
(2007)
IEEE Trans. Syst. Man Cybernet. B (Special Issue on Robot Learning by Observation, Demonstration and Imitation
, vol.37
-
-
Calinon, S.1
Guenter, F.2
Billard, A.3
-
28
-
-
33646162402
-
Discriminative and adaptative imitation in uni-manual and bi-manual tasks
-
Billard, A., Calinon, S., and Guenter, F., 2006. Discriminative and adaptative imitation in uni-manual and bi-manual tasks. Robotics and Autonomous Syst., 54:370–384.
-
(2006)
Robotics and Autonomous Syst.
, vol.54
, pp. 370-384
-
-
Billard, A.1
Calinon, S.2
Guenter, F.3
-
29
-
-
0036832950
-
Technical update: Least-squares temporal difference learning
-
Boyan, J. A., 2002. Technical update:least-squares temporal difference learning. Machine Learn., 49:233–246.
-
(2002)
Machine Learn.
, vol.49
, pp. 233-246
-
-
Boyan, J.A.1
|