메뉴 건너뛰기




Volumn 4894 LNAI, Issue , 2008, Pages 254-268

Relational macros for transfer in reinforcement learning

Author keywords

[No Author keywords available]

Indexed keywords

COMPUTER SIMULATION; DECISION MAKING; FORMAL LOGIC; KNOWLEDGE ACQUISITION; LOGIC PROGRAMMING;

EID: 40249114836     PISSN: 03029743     EISSN: 16113349     Source Type: Book Series    
DOI: 10.1007/978-3-540-78469-2_25     Document Type: Conference Paper
Times cited : (27)

References (22)
  • 2
    • 0002278788 scopus 로고    scopus 로고
    • Hierarchical reinforcement learning with the MAXQ value function decomposition
    • Dietterich, T.: Hierarchical reinforcement learning with the MAXQ value function decomposition. Journal of Artificial Intelligence Research 13, 227-303 (2000)
    • (2000) Journal of Artificial Intelligence Research , vol.13 , pp. 227-303
    • Dietterich, T.1
  • 3
    • 4444312102 scopus 로고    scopus 로고
    • Integrating guidance into relational reinforcement learning
    • Driessens, K., Dzeroski, S.: Integrating guidance into relational reinforcement learning. Machine Learning 57(3), 271-304 (2004)
    • (2004) Machine Learning , vol.57 , Issue.3 , pp. 271-304
    • Driessens, K.1    Dzeroski, S.2
  • 10
    • 0142121953 scopus 로고    scopus 로고
    • Using options for knowledge transfer in reinforcement learning
    • Technical Report UM-CS-1999-034
    • Perkins, T., Precup, D.: Using options for knowledge transfer in reinforcement learning. Technical Report UM-CS-1999-034 (1999)
    • (1999)
    • Perkins, T.1    Precup, D.2
  • 11
    • 33750690679 scopus 로고    scopus 로고
    • Using homomorphisms to transfer options across continuous reinforcement learning domains
    • Soni, V., Singh, S.: Using homomorphisms to transfer options across continuous reinforcement learning domains. In: AAAI Conference on Artificial Intelligence (2006)
    • (2006) AAAI Conference on Artificial Intelligence
    • Soni, V.1    Singh, S.2
  • 15
    • 33847202724 scopus 로고
    • Learning to predict by the methods of temporal differences
    • Sutton, R.: Learning to predict by the methods of temporal differences. Machine Learning 3, 9-44 (1988)
    • (1988) Machine Learning , vol.3 , pp. 9-44
    • Sutton, R.1
  • 20
    • 33750335095 scopus 로고    scopus 로고
    • Torrey, L., Shavlik, J., Walker, T., Maclin, R.: Skill acquisition via transfer learning and advice taking. In: Fürnkranz, J., Scheffer, T., Spiliopoulou, M. (eds.) ECML 2006. LNCS (LNAI), 4212, Springer, Heidelberg (2006)
    • Torrey, L., Shavlik, J., Walker, T., Maclin, R.: Skill acquisition via transfer learning and advice taking. In: Fürnkranz, J., Scheffer, T., Spiliopoulou, M. (eds.) ECML 2006. LNCS (LNAI), vol. 4212, Springer, Heidelberg (2006)
  • 21
    • 33646413134 scopus 로고    scopus 로고
    • Torrey, L., Walker, T., Shavlik, J., Maclin, R.: Using advice to transfer knowledge acquired in one reinforcement learning task to another. In: Gama, J., Camacho, R., Brazdil, P.B., Jorge, A.M., Torgo, L. (eds.) ECML 2005. LNCS (LNAI), 3720, Springer, Heidelberg (2005)
    • Torrey, L., Walker, T., Shavlik, J., Maclin, R.: Using advice to transfer knowledge acquired in one reinforcement learning task to another. In: Gama, J., Camacho, R., Brazdil, P.B., Jorge, A.M., Torgo, L. (eds.) ECML 2005. LNCS (LNAI), vol. 3720, Springer, Heidelberg (2005)


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.