메뉴 건너뛰기




Volumn 1, Issue , 2003, Pages 424-431

Reinforcement Learning as Classification: Leveraging Modern Classifiers

Author keywords

[No Author keywords available]

Indexed keywords

ALGORITHMS; CLASSIFICATION (OF INFORMATION); CRYSTAL ORIENTATION; FEATURE EXTRACTION; MARKOV PROCESSES; MONTE CARLO METHODS; NEURAL NETWORKS; PARAMETER ESTIMATION;

EID: 1942420814     PISSN: None     EISSN: None     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (148)

References (18)
  • 2
    • 0000913324 scopus 로고    scopus 로고
    • SVMTorch: Support vector machines for large-scale regression problems
    • Collobert, R., & Bengio, S. (2001). SVMTorch: Support vector machines for large-scale regression problems. Journal of Machine Learning Research (JMLR), 1, 143-160.
    • (2001) Journal of Machine Learning Research (JMLR) , vol.1 , pp. 143-160
    • Collobert, R.1    Bengio, S.2
  • 6
    • 0000439891 scopus 로고
    • On the convergence of stochastic iterative dynamic programming algorithms
    • Jaakkola, T., Jordan, M., & Singh, S. (1994). On the convergence of stochastic iterative dynamic programming algorithms. Neural Computation, 6, 1185-1201.
    • (1994) Neural Computation , vol.6 , pp. 1185-1201
    • Jaakkola, T.1    Jordan, M.2    Singh, S.3
  • 7
    • 85153938292 scopus 로고
    • Reinforcement learning algorithm for partially observable Markov decision problems
    • Cambridge, Massachusetts: MIT Press
    • Jaakkola, T., Singh, S. P., & Jordan, M. I. (1995). Reinforcement learning algorithm for partially observable Markov decision problems. Advances in Neural Information Processing Systems 7 (pp. 345-352). Cambridge, Massachusetts: MIT Press.
    • (1995) Advances in Neural Information Processing Systems , vol.7 , pp. 345-352
    • Jaakkola, T.1    Singh, S.P.2    Jordan, M.I.3
  • 12
    • 0141596576 scopus 로고    scopus 로고
    • Policy invariance under reward transformations: Theory and application to reward shaping
    • Morgan Kaufmann, San Francisco, CA
    • Ng, A. Y., Harada, D., & Russell, S. (1999). Policy invariance under reward transformations: theory and application to reward shaping. Proc. 16th International Conf. on Machine Learning (pp. 278-287). Morgan Kaufmann, San Francisco, CA.
    • (1999) Proc. 16th International Conf. on Machine Learning , pp. 278-287
    • Ng, A.Y.1    Harada, D.2    Russell, S.3
  • 16
    • 0030082891 scopus 로고    scopus 로고
    • An approach to fuzzy control of nonlinear systems: Stability and design issues
    • Wang, H. Tanaka, K., & Griffin, M. (1996). An approach to fuzzy control of nonlinear systems: Stability and design issues. IEEE Transactions on Fuzzy Systems, 4, 14-23.
    • (1996) IEEE Transactions on Fuzzy Systems , vol.4 , pp. 14-23
    • Wang, H.1    Tanaka, K.2    Griffin, M.3
  • 17
    • 0000337576 scopus 로고
    • Simple statistical gradient-following algorithms for connectionist reinforcement learning
    • Williams, R. J. (1992). Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8, 229-256.
    • (1992) Machine Learning , vol.8 , pp. 229-256
    • Williams, R.J.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.