메뉴 건너뛰기




Volumn , Issue , 2009, Pages 242-264

Transfer learning

Author keywords

[No Author keywords available]

Indexed keywords


EID: 84866381873     PISSN: None     EISSN: None     Source Type: Book    
DOI: 10.4018/978-1-60566-766-9.ch011     Document Type: Chapter
Times cited : (1782)

References (63)
  • 5
    • 0031189914 scopus 로고    scopus 로고
    • Multitask learning
    • doi:10.1023/A:1007379606734
    • Caruana, R. (1997). Multitask learning. Machine Learning, 28, 41-75. doi:10.1023/A:1007379606734
    • (1997) Machine Learning , vol.28 , pp. 41-75
    • Caruana, R.1
  • 10
    • 0002278788 scopus 로고    scopus 로고
    • Hierarchical reinforcement learning with the maxq value function decomposition
    • Dietterich, T. (2000). Hierarchical reinforcement learning with the MAXQ value function decomposition. Journal of Artificial Intelligence Research, 13, 227-303.
    • (2000) Journal of Artificial Intelligence Research , vol.13 , pp. 227-303
    • Dietterich, T.1
  • 14
    • 0024767488 scopus 로고
    • The structure-mapping engine: Algorithm and examples
    • doi:10.1016/0004-3702(89)90077-5
    • Falkenhainer, B., Forbus, K., & Gentner, D. (1989). The structure-mapping engine: Algorithm and examples. Artificial Intelligence, 41, 1-63. doi:10.1016/0004-3702(89)90077-5
    • (1989) Artificial Intelligence , vol.41 , pp. 1-63
    • Falkenhainer, B.1    Forbus, K.2    Gentner, D.3
  • 16
    • 0031211090 scopus 로고    scopus 로고
    • A decision-theoretic generalization of on-line learning and an application to boosting
    • doi:10.1006/jcss.1997.1504
    • Freund, Y., & Schapire, R. (1997). A decision-theoretic generalization of on-line learning and an application to boosting. Journal of Computer and System Sciences, 55(1), 119-139. doi:10.1006/jcss.1997.1504
    • (1997) Journal of Computer and System Sciences , vol.55 , Issue.1 , pp. 119-139
    • Freund, Y.1    Schapire, R.2
  • 23
    • 3843062299 scopus 로고    scopus 로고
    • Transfer of experience between reinforcement learning environments with progressive difficulty
    • doi:10.1023/B:AIRE.0000036264.95672.64
    • Madden, M., & Howley, T. (2004). Transfer of experience between reinforcement learning environments with progressive difficulty. Artificial Intelligence Review, 21, 375-398. doi:10.1023/B:AIRE.0000036264.95672.64
    • (2004) Artificial Intelligence Review , vol.21 , pp. 375-398
    • Madden, M.1    Howley, T.2
  • 34
    • 32044466073 scopus 로고    scopus 로고
    • Markov logic networks
    • doi:10.1007/s10994-006-5833-1
    • Richardson, M., & Domingos, P. (2006). Markov logic networks. Machine Learning, 62(1-2), 107-136. doi:10.1007/s10994-006-5833-1
    • (2006) Machine Learning , vol.62 , Issue.1-2 , pp. 107-136
    • Richardson, M.1    Domingos, P.2
  • 38
    • 74049137639 scopus 로고    scopus 로고
    • Action-space knowledge transfer in mdps: Formalism, suboptimality bounds, and algorithms
    • In
    • Sherstov, A., & Stone, P. (2005). Action-space knowledge transfer in MDPs: Formalism, suboptimality bounds, and algorithms. In Proceedings of the Conference on Learning Theory.
    • (2005) Proceedings of the Conference on Learning Theory
    • Sherstov, A.1    Stone, P.2
  • 40
    • 0001027894 scopus 로고
    • Transfer of learning by composing solutions of elemental sequential tasks
    • doi:10.1007/BF00992700
    • Singh, S. (1992). Transfer of learning by composing solutions of elemental sequential tasks. Machine Learning, 8(3-4), 323-339. doi:10.1007/BF00992700
    • (1992) Machine Learning , vol.8 , Issue.3-4 , pp. 323-339
    • Singh, S.1
  • 44
    • 33847202724 scopus 로고
    • Learning to predict by the methods of temporal differences
    • Sutton, R. (1988). Learning to predict by the methods of temporal differences. Machine Learning, 3, 9-44.
    • (1988) Machine Learning , vol.3 , pp. 9-44
    • Sutton, R.1
  • 62
    • 0004049893 scopus 로고
    • Unpublished doctoral dissertation, University of Cambridge
    • Watkins, C. (1989). Learning from delayed rewards. Unpublished doctoral dissertation, University of Cambridge.
    • (1989) Learning from Delayed Rewards
    • Watkins, C.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.