메뉴 건너뛰기




Volumn 6, Issue , 2016, Pages 4135-4148

Continuous deep q-learning with model-based acceleration

Author keywords

[No Author keywords available]

Indexed keywords

ARTIFICIAL INTELLIGENCE; COMPLEX NETWORKS; DEEP LEARNING; EFFICIENCY; ITERATIVE METHODS; LEARNING ALGORITHMS;

EID: 84998579328     PISSN: None     EISSN: None     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (352)

References (37)
  • 1
    • 0031073475 scopus 로고    scopus 로고
    • Locally weighted learning for control
    • Springer
    • Atkeson, Christopher G, Moore, Andrew W, and Schaal, Stefan. Locally weighted learning for control. In Lazy learning, pp. 75-113. Springer, 1997.
    • (1997) Lazy Learning , pp. 75-113
    • Atkeson, C.G.1    Moore, A.W.2    Schaal, S.3
  • 2
    • 0004370245 scopus 로고
    • Technical report, DTIC Document
    • Baird III, Leemon C. Advantage updating. Technical report, DTIC Document, 1993.
    • (1993) Advantage Updating
    • Baird, L.C.1
  • 3
    • 84998965670 scopus 로고    scopus 로고
    • The importance of experience replay database composition in deep reinforcement learning
    • NIPS
    • de Bruin, Tim, Kober, Jens, Tuyls, Karl, and Babuska, Robert. The importance of experience replay database composition in deep reinforcement learning. Deep Reinforcement Learning Workshop, NIPS, 2015.
    • (2015) Deep Reinforcement Learning Workshop
    • De Bruin, T.1    Kober, J.2    Tuyls, K.3    Babuska, R.4
  • 7
    • 79958779459 scopus 로고    scopus 로고
    • Reinforcement learning in feedback control
    • Hafner, Roland and Riedmiller, Martin. Reinforcement learning in feedback control. Machine learning, 84(1-2): 137-169, 2011.
    • (2011) Machine Learning , vol.84 , Issue.1-2 , pp. 137-169
    • Hafner, R.1    Riedmiller, M.2
  • 12
    • 84892593209 scopus 로고    scopus 로고
    • Reinforcement learning in robotics: A survey
    • Springer
    • Kober, Jens and Peters, Jan. Reinforcement learning in robotics: A survey. In Reinforcement Learning, pp. 579-610. Springer, 2012.
    • (2012) Reinforcement Learning , pp. 579-610
    • Kober, J.1    Peters, J.2
  • 16
    • 84937822296 scopus 로고    scopus 로고
    • Learning neural network policies with guided policy search under unknown dynamics
    • Levine, Sergey and Abbeel, Pieter. Learning neural network policies with guided policy search under unknown dynamics. In Advances in Neural Information Processing Systems (NIPS), pp. 1071-1079, 2014.
    • (2014) Advances in Neural Information Processing Systems (NIPS) , pp. 1071-1079
    • Levine, S.1    Abbeel, P.2
  • 18
    • 84979924150 scopus 로고    scopus 로고
    • End-to-end training of deep visuomotor policies
    • Levine, Sergey, Finn, Chelsea, Darrell, Trevor, and Abbeel, Pieter. End-to-end training of deep visuomotor policies. JMLR 17, 2016.
    • (2016) JMLR , vol.17
    • Levine, S.1    Finn, C.2    Darrell, T.3    Abbeel, P.4
  • 19
    • 17444424051 scopus 로고    scopus 로고
    • Iterative linear quadratic regulator design for nonlinear biological movement systems
    • Li, Weiwei and Todorov, Emanuel. Iterative linear quadratic regulator design for nonlinear biological movement systems. In ICINCO (1), pp. 222-229, 2004.
    • (2004) ICINCO , Issue.1 , pp. 222-229
    • Li, W.1    Todorov, E.2
  • 24
    • 85167411371 scopus 로고    scopus 로고
    • Relative entropy policy search
    • Atlanta
    • Peters, Jan, Mulling, Katharina, and Altun, Yasemin. Relative entropy policy search. In AAAI. Atlanta, 2010.
    • (2010) AAAI
    • Peters, J.1    Mulling, K.2    Altun, Y.3
  • 31
    • 85132026293 scopus 로고
    • Integrated architectures for learning, planning, and reacting based on approximating dynamic programming
    • Sutton, Richard S. Integrated architectures for learning, planning, and reacting based on approximating dynamic programming. In International Conference on Machine Learning (ICML), pp. 216-224, 1990.
    • (1990) International Conference on Machine Learning (ICML) , pp. 216-224
    • Sutton, R.S.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.