메뉴 건너뛰기




Volumn 17, Issue 2, 2010, Pages 20-29

Learning control in robotics

Author keywords

Learning control; Optimal control; Reinforcement learning; Robot learning

Indexed keywords

BAYESIAN; CONTROL POLICY; CONTROLLED SYSTEM; CONTROLLER PARAMETER; INTERNAL MODELS; INVERSE MODELS; LEARNING CONTROL; LEARNING CONTROL TECHNIQUES; LOCALLY WEIGHTED REGRESSION; OPTIMAL CONTROL; OPTIMAL CONTROLS; ROBOT ARMS; ROBUST CONTROLLERS; STATE VECTOR; TRAJECTORY-BASED;

EID: 77953330028     PISSN: 10709932     EISSN: None     Source Type: Journal    
DOI: 10.1109/MRA.2010.936957     Document Type: Article
Times cited : (159)

References (83)
  • 1
    • 73549098121 scopus 로고    scopus 로고
    • The new robotics-Towards human-centeredmachines
    • S. Schaal, "The new robotics-Towards human-centeredmachines," HFSP J. Frontiers Interdisciplinary Res. Life Sci., vol.1, no.2, pp. 115-126, 2007.
    • (2007) HFSP J. Frontiers Interdisciplinary Res. Life Sci. , vol.1 , Issue.2 , pp. 115-126
    • Schaal, S.1
  • 6
    • 27344443125 scopus 로고    scopus 로고
    • Finding approximate POMDP solutions through belief compression
    • M. Roy, G. Gordon, and S. Thrun, "Finding approximate POMDP solutions through belief compression," J. Artif. Intell. Res., vol.23, pp. 1-40, 2005.
    • (2005) J. Artif. Intell. Res. , vol.23 , pp. 1-40
    • Roy, M.1    Gordon, G.2    Thrun, S.3
  • 9
    • 85012688561 scopus 로고
    • Princeton NJ Princeton Univ. Press
    • R. Bellman, Dynamic Programming. Princeton, NJ: Princeton Univ. Press, 1957.
    • (1957) Dynamic Programming
    • Bellman, R.1
  • 12
    • 44049116478 scopus 로고
    • Supervised learning with a distal teacher
    • I. M. Jordan, D. E. Rumelhart, "Supervised learning with a distal teacher," Cogn. Sci., vol.16, pp. 307-354, 1992.
    • (1992) Cogn. Sci. , vol.16 , pp. 307-354
    • Jordan, I.M.1    Rumelhart, D.E.2
  • 14
    • 0027382368 scopus 로고
    • A self-organizing neural model of motor equivalent reaching and tool use by a multijoint arm
    • D. Bullock, S. Grossberg, and F. H. Guenther, "A self-organizing neural model of motor equivalent reaching and tool use by a multijoint arm," J. Cogn. Neurosci., vol.5, no.4, pp. 408-435, 1993.
    • (1993) J. Cogn. Neurosci. , vol.5 , Issue.4 , pp. 408-435
    • Bullock, D.1    Grossberg, S.2    Guenther, F.H.3
  • 15
    • 38649095925 scopus 로고    scopus 로고
    • Learning to control in operational space
    • J. Peters and S. Schaal, "Learning to control in operational space," Int. J. Robot. Res., vol.27, pp. 197-212, 2008.
    • (2008) Int. J. Robot. Res. , vol.27 , pp. 197-212
    • Peters, J.1    Schaal, S.2
  • 16
    • 0001551844 scopus 로고
    • Supervised learning from incomplete data via an em approach
    • J. D. Cowan, G. Tesauro, and J. Alspector, Eds. San Mateo, CA: Morgan Kaufmann
    • Z. Ghahramani and M. I. Jordan, "Supervised learning from incomplete data via an EM approach," in Advances in Neural Information Processing Systems 6, J. D. Cowan, G. Tesauro, and J. Alspector, Eds. San Mateo, CA: Morgan Kaufmann, 1994, pp. 120-127.
    • (1994) Advances in Neural Information Processing Systems , vol.6 , pp. 120-127
    • Ghahramani, Z.1    Jordan, M.I.2
  • 17
    • 0001108227 scopus 로고    scopus 로고
    • Constructive incremental learning from only local information
    • S. Schaal and C. G. Atkeson, "Constructive incremental learning from only local information," Neural Comput., vol.10, no.8, pp. 2047-2084, 1998.
    • (1998) Neural Comput. , vol.10 , Issue.8 , pp. 2047-2084
    • Schaal, S.1    Atkeson, C.G.2
  • 18
    • 84936916896 scopus 로고
    • Robust locally weighted regression and smoothing scatterplots
    • W. S. Cleveland, "Robust locally weighted regression and smoothing scatterplots," J. Amer. Statist. Assoc., vol.74, pp. 829-836, 1979.
    • (1979) J. Amer. Statist. Assoc. , vol.74 , pp. 829-836
    • Cleveland, W.S.1
  • 19
    • 2342560362 scopus 로고
    • Using local models to control movement
    • D. Touretzky, Ed. San Mateo, CA: Morgan Kaufmann
    • C. G. Atkeson, "Using local models to control movement," in Advances in Neural Information Processing Systems 1, D. Touretzky, Ed. San Mateo, CA: Morgan Kaufmann, 1989, pp. 157-183.
    • (1989) Advances in Neural Information Processing Systems , vol.1 , pp. 157-183
    • Atkeson, C.G.1
  • 21
    • 0031073475 scopus 로고    scopus 로고
    • Locally weighted learning for control
    • C. G. Atkeson, A. W. Moore, and S. Schaal, "Locally weighted learning for control," Artif. Intell. Rev., vol.11, no.1-5, pp. 75-113, 1997.
    • (1997) Artif. Intell. Rev. , vol.11 , Issue.1-5 , pp. 75-113
    • Atkeson, C.G.1    Moore, A.W.2    Schaal, S.3
  • 22
    • 27144556425 scopus 로고    scopus 로고
    • Incremental online learning in high dimensions
    • S. Vijayakumar, A. D'Souza, and S. Schaal, "Incremental online learning in high dimensions," Neural Comput., vol.17, no.12, pp. 2602-2634, 2005.
    • (2005) Neural Comput. , vol.17 , Issue.12 , pp. 2602-2634
    • Vijayakumar, S.1    D'Souza, A.2    Schaal, S.3
  • 25
    • 77953336006 scopus 로고    scopus 로고
    • Local gaussian process regression for real time online model learning and control
    • D. Schuurmans, J. Benigio, and D. Koller, Eds. Vancouver, BC, Dec. 8-11
    • D. Nguyen-Tuong, M. Seeger, and J. Peters, "Local gaussian process regression for real time online model learning and control," in Proc. Advances in Neural Information Processing Systems 21 (NIPS 2008), D. Schuurmans, J. Benigio, and D. Koller, Eds. Vancouver, BC, Dec. 8-11, 2009, pp. 1193-1200.
    • (2009) Proc. Advances in Neural Information Processing Systems 21 (NIPS 2008) , pp. 1193-1200
    • Nguyen-Tuong, D.1    Seeger, M.2    Peters, J.3
  • 26
    • 61849173491 scopus 로고    scopus 로고
    • Gaussian process dynamic programming
    • M. P. Deisenroth, C. E. Rasmussen, and J. Peters, "Gaussian process dynamic programming," Neurocomputing, vol.72, no.7-9, pp. 1508- 1524, 2009.
    • (2009) Neurocomputing , vol.72 , Issue.7-9 , pp. 1508-1524
    • Deisenroth, M.P.1    Rasmussen, C.E.2    Peters, J.3
  • 28
    • 0032192424 scopus 로고    scopus 로고
    • Multiple paired forward and inverse models for motor control
    • D. M. Wolpert and M. Kawato, "Multiple paired forward and inverse models for motor control," Neural Netw., vol.11, no.7-8, pp. 1317- 1329, 1998.
    • (1998) Neural Netw. , vol.11 , Issue.7-8 , pp. 1317-1329
    • Wolpert, D.M.1    Kawato, M.2
  • 30
    • 0033629916 scopus 로고    scopus 로고
    • Reinforcement learning in continuous time and space
    • Jan.
    • K. Doya, "Reinforcement learning in continuous time and space," Neural Comput., vol.12, no.1, pp. 219-245, Jan. 2000.
    • (2000) Neural Comput. , vol.12 , Issue.1 , pp. 219-245
    • Doya, K.1
  • 32
    • 0036832953 scopus 로고    scopus 로고
    • Variable resolution discretization in optimal control
    • R. Munos and A. Moore, "Variable resolution discretization in optimal control," Mach. Learn., vol. 49, no. 2/3, p. 33, 2002.
    • (2002) Mach. Learn. , vol.49 , Issue.2-3 , pp. 33
    • Munos, R.1    Moore, A.2
  • 33
    • 49049094416 scopus 로고    scopus 로고
    • Random sampling of states in dynamic programming
    • C. G. Atkeson and B. J. Stephens, "Random sampling of states in dynamic programming," IEEE Trans. Syst., Man, Cybern. B, vol.38, no.4, pp. 924-929, 2008.
    • (2008) IEEE Trans. Syst., Man, Cybern. B , vol.38 , Issue.4 , pp. 924-929
    • Atkeson, C.G.1    Stephens, B.J.2
  • 35
    • 77950579842 scopus 로고    scopus 로고
    • Control of a walking biped using a combination of simple policies
    • Paris, France, Dec. 7-10
    • E. Whitman and C. G. Atkeson, "Control of a walking biped using a combination of simple policies," in Proc. IEEE/RAS Int. Conf. Humanoid Robotics, Paris, France, Dec. 7-10, 2009, pp. 520-527.
    • (2009) Proc. IEEE/RAS Int. Conf. Humanoid Robotics , pp. 520-527
    • Whitman, E.1    Atkeson, C.G.2
  • 36
    • 64849106540 scopus 로고    scopus 로고
    • Tomlab Optimization Inc. [Online]. Available
    • Tomlab Optimization Inc. (2010). PROPT-Matlab optimal control software [Online]. Available: http://tomdyn.com/
    • (2010) PROPT-Matlab Optimal Control Software
  • 38
    • 77953331240 scopus 로고    scopus 로고
    • [Online]. Available
    • Stanford Business Software Corporation. (2010). SNOPT; Software for large-scale nonlinear programming [Online]. Available: http://www.sbsisol- optimize.com/asp/sol-product-snopt.htm
    • (2010) SNOPT; Software for Large-scale Nonlinear Programming
  • 39
    • 12844272111 scopus 로고    scopus 로고
    • Synthesizing physically realistic human motion in low-dimensional, behavior-specific spaces
    • A. Safonova, J. K. Hodgins, and N. S. Pollard, "Synthesizing physically realistic human motion in low-dimensional, behavior-specific spaces," ACM Trans. Graph. J.(SIGGRAPH 2004 Proc.), vol.23, no.3, pp. 514- 521, 2004.
    • (2004) ACM Trans. Graph. J.(SIGGRAPH 2004 Proc.) , vol.23 , Issue.3 , pp. 514-521
    • Safonova, A.1    Hodgins, J.K.2    Pollard, N.S.3
  • 41
    • 0141819580 scopus 로고    scopus 로고
    • Pegasus: A policy search method for large MDPs and POMDPs
    • presented at the
    • A. Ng, "Pegasus: A policy search method for large MDPs and POMDPs," presented at the Uncertainty in Artificial Intelligence (UAI), 2000.
    • (2000) Uncertainty in Artificial Intelligence (UAI)
    • Ng, A.1
  • 42
    • 0015615562 scopus 로고
    • Wide-sense adaptive dual control for nonlinear stochastic systems
    • E. Tse, Y. Bar-Shalom, and L. Meier, III, "Wide-sense adaptive dual control for nonlinear stochastic systems," IEEE Trans. Automat. Contr., vol.18, no.2, pp. 98-108, 1973.
    • (1973) IEEE Trans. Automat. Contr. , vol.18 , Issue.2 , pp. 98-108
    • Tse, E.1    Bar-Shalom, Y.2    Meier III, L.3
  • 43
    • 0041443966 scopus 로고
    • Caution, probing and the value of information in the control of uncertain systems
    • Y. Bar-Shalom and E. Tse, "Caution, probing and the value of information in the control of uncertain systems," Ann. Econ. Social Meas., vol.4, no.3, pp. 323-338, 1976.
    • (1976) Ann. Econ. Social Meas. , vol.4 , Issue.3 , pp. 323-338
    • Bar-Shalom, Y.1    Tse, E.2
  • 44
    • 0002130986 scopus 로고    scopus 로고
    • Robot learning from demonstration
    • D. H. Fisher, Jr., Ed. Nashville, TN, July 8-12
    • C. G. Atkeson and S. Schaal, "Robot learning from demonstration," in Proc. 14th Int. Conf. Machine Learning (ICML'97), D. H. Fisher, Jr., Ed. Nashville, TN, July 8-12, 1997, pp. 12-20.
    • (1997) Proc. 14th Int. Conf. Machine Learning (ICML'97) , pp. 12-20
    • Atkeson, C.G.1    Schaal, S.2
  • 45
    • 33847202724 scopus 로고
    • Learning to predict by the methods of temporal differences
    • R. S. Sutton, "Learning to predict by the methods of temporal differences," Mach. Learn., vol.3, no.1, pp. 9-44, 1988.
    • (1988) Mach. Learn. , vol.3 , Issue.1 , pp. 9-44
    • Sutton, R.S.1
  • 47
    • 0035979437 scopus 로고    scopus 로고
    • Acquisition of stand-up behavior by a real robot using hierarchical reinforcement learning
    • J. Morimoto and K. Doya, "Acquisition of stand-up behavior by a real robot using hierarchical reinforcement learning," Robot. Auton. Syst., vol.36, no.1, pp. 37-51, 2001.
    • (2001) Robot. Auton. Syst. , vol.36 , Issue.1 , pp. 37-51
    • Morimoto, J.1    Doya, K.2
  • 48
    • 0033151712 scopus 로고    scopus 로고
    • Is imitation learning the route to humanoid robots?
    • S. Schaal, "Is imitation learning the route to humanoid robots?" Trends Cogn. Sci., vol.3, no.6, pp. 233-242, 1999.
    • (1999) Trends Cogn. Sci. , vol.3 , Issue.6 , pp. 233-242
    • Schaal, S.1
  • 51
    • 84898995067 scopus 로고    scopus 로고
    • Learning from demonstration
    • M. C. Mozer, M. Jordan, and T. Petsche, Eds. Cambridge, MA
    • S. Schaal, "Learning from demonstration," in Proc. Advances in Neural Information Processing Systems 9, M. C. Mozer, M. Jordan, and T. Petsche, Eds. Cambridge, MA, 1997, pp. 1040-1046.
    • (1997) Proc. Advances in Neural Information Processing Systems , vol.9 , pp. 1040-1046
    • Schaal, S.1
  • 52
    • 21844465127 scopus 로고    scopus 로고
    • Tree-based batch mode reinforcement learning
    • D. Ernst, P. Geurts, and L. Wehenkel, "Tree-based batch mode reinforcement learning," J. Mach. Learn. Res., vol.6, pp. 503-556, 2005.
    • (2005) J. Mach. Learn. Res. , vol.6 , pp. 503-556
    • Ernst, D.1    Geurts, P.2    Wehenkel, L.3
  • 53
    • 70049104729 scopus 로고    scopus 로고
    • Fitted Q-iteration by advantage weighted regression
    • D. Schuurmans, J. Benigio, and D. Koller, Eds. Vancouver, BC, Dec. 8-1
    • G. Neumann and J. Peters, "Fitted Q-iteration by advantage weighted regression," in Proc. Advances in Neural Information Processing Systems 21 (NIPS 2008), D. Schuurmans, J. Benigio, and D. Koller, Eds. Vancouver, BC, Dec. 8-11, 2009, pp. 1177-1184.
    • (2009) Proc. Advances in Neural Information Processing Systems 21 (NIPS 2008) , pp. 1177-1184
    • Neumann, G.1    Peters, J.2
  • 54
    • 84898939480 scopus 로고    scopus 로고
    • Policy gradient methods for reinforcement learning with function approximation
    • S. A. Solla, T. K. Leen, and K.-R.Muller, Eds. Denver, CO
    • R. S. Sutton, D. McAllester, S. Singh, and Y. Mansour, "Policy gradient methods for reinforcement learning with function approximation," in Proc. Advances in Neural Processing Systems 12, S. A. Solla, T. K. Leen, and K.-R.Muller, Eds. Denver, CO, 2000.
    • (2000) Proc. Advances in Neural Processing Systems , vol.12
    • Sutton, R.S.1    McAllester, D.2    Singh, S.3    Mansour, Y.4
  • 55
    • 44949241322 scopus 로고    scopus 로고
    • Reinforcement learning of motor skills with policy gradients
    • May
    • J. Peters and S. Schaal, "Reinforcement learning of motor skills with policy gradients," Neural Netw., vol.21, no.4, pp. 682-697, May 2008.
    • (2008) Neural Netw. , vol.21 , Issue.4 , pp. 682-697
    • Peters, J.1    Schaal, S.2
  • 56
    • 0030703463 scopus 로고    scopus 로고
    • Optimal random perturbations for stochastic approximation using a simultaneous perturbation gradient approximation
    • presented at the
    • P. Sadegh and J. Spall, "Optimal random perturbations for stochastic approximation using a simultaneous perturbation gradient approximation," presented at the Proc. American Control Conf., 1997.
    • (1997) Proc. American Control Conf.
    • Sadegh, P.1    Spall, J.2
  • 57
    • 0000337576 scopus 로고
    • Simple statistical gradient-following algorithms for connectionist reinforcement learning
    • R. J. Williams, "Simple statistical gradient-following algorithms for connectionist reinforcement learning," Mach. Learn., vol.8, no.3-4, pp. 229- 256, 1992.
    • (1992) Mach. Learn. , vol.8 , Issue.3-4 , pp. 229-256
    • Williams, R.J.1
  • 58
    • 0025600638 scopus 로고
    • A stochastic reinforcement learning algorithm for learning real-valued functions
    • V. Gullapalli, "A stochastic reinforcement learning algorithm for learning real-valued functions," Neural Netw., vol.3, no.6, pp. 671-692, 1990.
    • (1990) Neural Netw. , vol.3 , Issue.6 , pp. 671-692
    • Gullapalli, V.1
  • 60
    • 40649106649 scopus 로고    scopus 로고
    • Natural actor critic
    • J. Peters and S. Schaal, "Natural actor critic," Neurocomputing, vol.71, no.7-9, pp. 1180-1190, 2008.
    • (2008) Neurocomputing , vol.71 , Issue.7-9 , pp. 1180-1190
    • Peters, J.1    Schaal, S.2
  • 61
    • 0033570817 scopus 로고    scopus 로고
    • Natural gradient learning for over- and under-complete bases in ICA
    • Nov.
    • S. Amari, "Natural gradient learning for over- and under-complete bases In ICA," Neural Comput., vol.11, no.8, pp. 1875-1883, Nov. 1999.
    • (1999) Neural Comput. , vol.11 , Issue.8 , pp. 1875-1883
    • Amari, S.1
  • 64
    • 38649142135 scopus 로고    scopus 로고
    • Learning CPG-based biped locomotion with a policy gradient method: Application to a humanoid robot
    • G. Endo, J. Morimoto, T. Matsubara, J. Nakanish, and G. Cheng, "Learning CPG-based biped locomotion with a policy gradient method: Application to a humanoid robot," Int. J. Robot. Res., vol.27, no.2, pp. 213-228, 2008.
    • (2008) Int. J. Robot. Res. , vol.27 , Issue.2 , pp. 213-228
    • Endo, G.1    Morimoto, J.2    Matsubara, T.3    Nakanish, J.4    Cheng, G.5
  • 67
    • 0346982426 scopus 로고    scopus 로고
    • Using em for reinforcement learning
    • P. Dayan and G. Hinton, "Using EM for reinforcement learning," Neural Comput., vol.9, no.2, pp. 271-278, 1997.
    • (1997) Neural Comput. , vol.9 , Issue.2 , pp. 271-278
    • Dayan, P.1    Hinton, G.2
  • 68
    • 0002629270 scopus 로고
    • Maximum likelihood from incomplete data via the em algorithm
    • A. P. Dempster, N. M. Laird, and D. B. Rubin, "Maximum likelihood from incomplete data via the EM algorithm," J. R. Statist. Soc. B, vol.39, no.1, pp. 1-38, 1977.
    • (1977) J. R. Statist. Soc. B , vol.39 , Issue.1 , pp. 1-38
    • Dempster, A.P.1    Laird, N.M.2    Rubin, D.B.3
  • 69
    • 78049446733 scopus 로고    scopus 로고
    • Learning motor primitives in robotics
    • D. Schuurmans, J. Benigio, and D. Koller, Eds. Vancouver, BC, Dec. 8-11
    • J. Kober and J. Peters, "Learning motor primitives in robotics," in Proc. Advances in Neural Information Processing Systems 21 (NIPS 2008), D. Schuurmans, J. Benigio, and D. Koller, Eds. Vancouver, BC, Dec. 8-11, 2009, pp. 297-304.
    • (2009) Proc. Advances in Neural Information Processing Systems 21 (NIPS 2008) , pp. 297-304
    • Kober, J.1    Peters, J.2
  • 70
    • 34250728061 scopus 로고    scopus 로고
    • Probabilistic inference for solving discrete and continuous state Markov decision processes
    • presented at the
    • M. Toussaint and A. Storkey, "Probabilistic inference for solving discrete and continuous state Markov decision processes," presented at the 23nd Int. Conf. Machine Learning (ICML 2006), 2006.
    • (2006) 23nd Int. Conf. Machine Learning (ICML 2006)
    • Toussaint, M.1    Storkey, A.2
  • 71
    • 70349327392 scopus 로고    scopus 로고
    • Learning modelfree control by a Monte-Carlo em algorithm
    • N. Vlassis, M. Toussaint, G. Kontes, and S. Piperidis, "Learning modelfree control by a Monte-Carlo EM algorithm," Auton. Robots, vol.27, no.2, pp. 123-130, 2009.
    • (2009) Auton. Robots , vol.27 , Issue.2 , pp. 123-130
    • Vlassis, N.1    Toussaint, M.2    Kontes, G.3    Piperidis, S.4
  • 72
    • 28844435646 scopus 로고    scopus 로고
    • Linear theory for control of nonlinear stochastic systems
    • Nov.
    • H. J. Kappen, "Linear theory for control of nonlinear stochastic systems," Phys. Rev. Lett., vol.95, no.20, pp. 200201-200204, Nov. 2005.
    • (2005) Phys. Rev. Lett. , vol.95 , Issue.20 , pp. 200201-200204
    • Kappen, H.J.1
  • 73
    • 33947410345 scopus 로고    scopus 로고
    • An introduction to stochastic control theory, path integrals and reinforcement learning
    • J. Marro, P. L. Garrido, and J. J. Torres, Eds
    • H. J. Kappen, "An introduction to stochastic control theory, path integrals and reinforcement learning," in Cooperative Behavior in Neural Systems, vol.887, J. Marro, P. L. Garrido, and J. J. Torres, Eds. 2007, pp. 149-181.
    • (2007) Cooperative Behavior in Neural Systems , vol.887 , pp. 149-181
    • Kappen, H.J.1
  • 75
    • 67650915125 scopus 로고    scopus 로고
    • Efficient computation of optimal actions
    • July
    • E. Todorov, "Efficient computation of optimal actions," Proc. Nat. Acad. Sci. USA, vol.106, no.28, pp. 11478-11483, July 2009.
    • (2009) Proc. Nat. Acad. Sci. USA , vol.106 , Issue.28 , pp. 11478-11483
    • Todorov, E.1
  • 76
    • 84899019754 scopus 로고    scopus 로고
    • Learning attractor landscapes for learning motor primitives
    • S. Becker, S. Thrun, and K.Obermayer, Eds.
    • A. Ijspeert, J. Nakanishi, and S. Schaal, "Learning attractor landscapes for learning motor primitives," in Advances in Neural Information Processing Systems 15, S. Becker, S. Thrun, and K.Obermayer, Eds. 2003, pp. 1547-1554.
    • (2003) Advances in Neural Information Processing Systems , vol.15 , pp. 1547-1554
    • Ijspeert, A.1    Nakanishi, J.2    Schaal, S.3
  • 78
    • 58249141653 scopus 로고    scopus 로고
    • Robot programming by demonstration
    • B. Siciliano andO.Khatib, Eds. Cambridge,MA:MIT Press ch. 59
    • A. Billard, S. Calinon, R. Dillmann, and S. Schaal, "Robot programming by demonstration," in Handbook of Robotics, vol.1, B. Siciliano andO.Khatib, Eds. Cambridge,MA:MIT Press, 2008, ch. 59.
    • (2008) Handbook of Robotics , vol.1
    • Billard, A.1    Calinon, S.2    Dillmann, R.3    Schaal, S.4
  • 79
    • 0027832075 scopus 로고
    • Trajectory formation of arm movement by a neural network with forward and inverse dynamics models
    • Y. Wada and M. Kawato, "Trajectory formation of arm movement by a neural network with forward and inverse dynamics models," Syst. Comput. Jpn., vol.24, pp. 37-50, 1994.
    • (1994) Syst. Comput. Jpn. , vol.24 , pp. 37-50
    • Wada, Y.1    Kawato, M.2
  • 80
    • 2442636320 scopus 로고    scopus 로고
    • Embodied symbol emergence based on mimesis theory
    • Apr.-May
    • T. Inamura, I. Toshima, H. Tanie, and Y. Nakamura, "Embodied symbol emergence based on mimesis theory," Int. J. Robot. Res., vol.23, no.4-5, p. 363, Apr.-May 2004.
    • (2004) Int. J. Robot. Res. , vol.23 , Issue.4-5 , pp. 363
    • Inamura, T.1    Toshima, I.2    Tanie, H.3    Nakamura, Y.4
  • 83
    • 67650957592 scopus 로고    scopus 로고
    • Learning to search: Functional gradient techniques for imitation learning
    • N. Ratliff, D. Silver, and J. A. Bagnell, "Learning to search: Functional gradient techniques for imitation learning," Auton. Robots, vol.27, no.1, pp. 25-53, 2009.
    • (2009) Auton. Robots , vol.27 , Issue.1 , pp. 25-53
    • Ratliff, N.1    Silver, D.2    Bagnell, J.A.3


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.