메뉴 건너뛰기




Volumn , Issue , 2017, Pages

Loss is its own reward: Self-supervision for reinforcement learning

Author keywords

[No Author keywords available]

Indexed keywords

END TO END; EXPECTED RETURN; OPTIMISATIONS; OPTIMIZING POLICIES; PRE-TRAINING; REINFORCEMENT LEARNINGS;

EID: 85144242642     PISSN: None     EISSN: None     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (70)

References (11)
  • 3
    • 85088229507 scopus 로고    scopus 로고
    • Learning to act by predicting the future
    • A. Dosovitskiy and V. Koltun. Learning to act by predicting the future. ICLR, 2017.
    • (2017) ICLR
    • Dosovitskiy, A.1    Koltun, V.2
  • 6
    • 84941943714 scopus 로고    scopus 로고
    • Learning state representations with robotic priors
    • R. Jonschkowski and O. Brock. Learning state representations with robotic priors. Autonomous Robots, 39(3):407-428, 2015.
    • (2015) Autonomous Robots , vol.39 , Issue.3 , pp. 407-428
    • Jonschkowski, R.1    Brock, O.2
  • 7
    • 85083952350 scopus 로고    scopus 로고
    • Data-dependent initializations of convolutional neural networks
    • Philipp Krähenbühl, Carl Doersch, Jeff Donahue, and Trevor Darrell. Data-dependent initializations of convolutional neural networks. In ICLR, 2016.
    • (2016) ICLR
    • Krähenbühl, P.1    Doersch, C.2    Donahue, J.3    Darrell, T.4
  • 11
    • 84965129327 scopus 로고    scopus 로고
    • Embed to control: A locally linear latent dynamics model for control from raw images
    • M. Watter, J. Springenberg, J. Boedecker, and M. Riedmiller. Embed to control: A locally linear latent dynamics model for control from raw images. In NIPS, pp. 2746-2754, 2015.
    • (2015) NIPS , pp. 2746-2754
    • Watter, M.1    Springenberg, J.2    Boedecker, J.3    Riedmiller, M.4


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.