메뉴 건너뛰기




Volumn 30, Issue 5, 1999, Pages 341-363

An analysis of experience replay in temporal difference learning

Author keywords

[No Author keywords available]

Indexed keywords

ADAPTIVE SYSTEMS; LEARNING ALGORITHMS; NUMERICAL METHODS; SET THEORY; THEOREM PROVING;

EID: 0032649518     PISSN: 01969722     EISSN: 10876553     Source Type: Journal    
DOI: 10.1080/019697299125127     Document Type: Article
Times cited : (21)

References (18)
  • 2
    • 0007512578 scopus 로고
    • Truncating temporal differences: On the efficient implementation of TD(A) for reinforcement learning
    • Cichosz, P. 1995. Truncating temporal differences: On the efficient implementation of TD(A) for reinforcement learning. Journal of Artificial Intelligence Research 2:287-318.
    • (1995) Journal of Artificial Intelligence Research , vol.2 , pp. 287-318
    • Cichosz, P.1
  • 3
    • 0008666497 scopus 로고    scopus 로고
    • Ph.D. thesis, Warsaw University of Technology, Department of Electronics and Information Technology
    • Cichosz, P. 1997. Reinforcement Learning by Truncating Temporal Differences. Ph.D. thesis, Warsaw University of Technology, Department of Electronics and Information Technology.
    • (1997) Reinforcement Learning by Truncating Temporal Differences
    • Cichosz, P.1
  • 8
    • 0026880130 scopus 로고
    • Automatic programming of behavior-based robots using reinforcement learning
    • Mahadevan, S. and J. Connell. 1992. Automatic programming of behavior-based robots using reinforcement learning. Artificial Intelligence 55:311-365.
    • (1992) Artificial Intelligence , vol.55 , pp. 311-365
    • Mahadevan, S.1    Connell, J.2
  • 11
    • 0029753630 scopus 로고    scopus 로고
    • Reinforcement learning with replacing eligibility traces
    • Singh, S. P. and R. S. Sutton. 1996. Reinforcement learning with replacing eligibility traces. Machine Learning 22:123-158.
    • (1996) Machine Learning , vol.22 , pp. 123-158
    • Singh, S.P.1    Sutton, R.S.2
  • 12
    • 0003617454 scopus 로고
    • Ph.D. thesis, University of Massachusetts, Department of Computer and Information Science, Boston, MA
    • Sutton, R. S. 1984. Temporal Credit Assignment in Reinforcement Learning. Ph.D. thesis, University of Massachusetts, Department of Computer and Information Science, Boston, MA.
    • (1984) Temporal Credit Assignment in Reinforcement Learning
    • Sutton, R.S.1
  • 13
    • 33847202724 scopus 로고
    • Learning to predict by the methods of temporal differences
    • Sutton, R. S. 1988. Learning to predict by the methods of temporal differences. Machine Learning 3:9-44.
    • (1988) Machine Learning , vol.3 , pp. 9-44
    • Sutton, R.S.1
  • 14
    • 85156221438 scopus 로고    scopus 로고
    • Generalization in reinforcement learning: Successful examples using sparse coarse coding
    • Cambridge, MA: MIT Press. Morgan Kaufmann
    • Sutton, R. S. 1996. Generalization in reinforcement learning: Successful examples using sparse coarse coding. In Advances in Neural Information Processing Systems 8, pp. 1038-1044. Cambridge, MA: MIT Press. Morgan Kaufmann.
    • (1996) Advances in Neural Information Processing Systems , vol.8 , pp. 1038-1044
    • Sutton, R.S.1
  • 16
    • 0001046225 scopus 로고
    • Practical issues in temporal difference learning
    • Tesauro, G. 1992. Practical issues in temporal difference learning. Machine Learning 8:257-277.
    • (1992) Machine Learning , vol.8 , pp. 257-277
    • Tesauro, G.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.