메뉴 건너뛰기




Volumn 8188 LNAI, Issue PART 1, 2013, Pages 241-256

Greedy confidence pursuit: A pragmatic approach to multi-bandit optimization

Author keywords

[No Author keywords available]

Indexed keywords

BASELINE METHODS; HIGH CONFIDENCE; MULTI ARMED BANDIT; PRACTICAL PROBLEMS; PURSUIT PROBLEMS;

EID: 84886556435     PISSN: 03029743     EISSN: 16113349     Source Type: Book Series    
DOI: 10.1007/978-3-642-40988-2_16     Document Type: Conference Paper
Times cited : (1)

References (16)
  • 1
    • 84886540275 scopus 로고    scopus 로고
    • Analysis of thompson sampling for the multi-armed bandit problem
    • Agrawal, S., Goyal, N.: Analysis of thompson sampling for the multi-armed bandit problem. In: COLT (2012)
    • (2012) COLT
    • Agrawal, S.1    Goyal, N.2
  • 2
    • 84864970677 scopus 로고    scopus 로고
    • Best arm identification in multi-armed bandits
    • Audibert, J.-Y., Bubeck, S., Munos, R.: Best arm identification in multi-armed bandits. In: COLT (2010)
    • (2010) COLT
    • Audibert, J.-Y.1    Bubeck, S.2    Munos, R.3
  • 4
    • 77952070805 scopus 로고    scopus 로고
    • Pure exploration in multi-armed bandits problems
    • Gavaldà, R., Lugosi, G., Zeugmann, T., Zilles, S. (eds.) ALT 2009. Springer, Heidelberg
    • Bubeck, S., Munos, R., Stoltz, G.: Pure exploration in multi-armed bandits problems. In: Gavaldà, R., Lugosi, G., Zeugmann, T., Zilles, S. (eds.) ALT 2009. LNCS, vol. 5809, pp. 23-37. Springer, Heidelberg (2009)
    • (2009) LNCS , vol.5809 , pp. 23-37
    • Bubeck, S.1    Munos, R.2    Stoltz, G.3
  • 7
    • 33745295134 scopus 로고    scopus 로고
    • Action elimination and stopping conditions for the multi-armed bandit and reinforcement learning problems
    • Even-Dar, E., Mannor, S., Mansour, Y.: Action elimination and stopping conditions for the multi-armed bandit and reinforcement learning problems. Journal of Machine Learning Research 7, 1079-1105 (2006)
    • (2006) Journal of Machine Learning Research , vol.7 , pp. 1079-1105
    • Even-Dar, E.1    Mannor, S.2    Mansour, Y.3
  • 11
    • 85029696856 scopus 로고    scopus 로고
    • Open problem: Regret bounds for thompson sampling
    • Li, L., Chappelle, O.: Open problem: Regret bounds for thompson sampling. In: COLT (2012)
    • (2012) COLT
    • Li, L.1    Chappelle, O.2
  • 12
    • 77953631184 scopus 로고    scopus 로고
    • The budgeted multi-armed bandit problem
    • Madani, O., Lizotte, D.J., Greiner, R.: The budgeted multi-armed bandit problem. In: COLT (2004)
    • (2004) COLT
    • Madani, O.1    Lizotte, D.J.2    Greiner, R.3
  • 13
    • 30044441333 scopus 로고    scopus 로고
    • The sample complexity of exploration in the multiarmed bandit problem
    • Mannor, S., Tsitsiklis, J.N.: The sample complexity of exploration in the multiarmed bandit problem. Journal of Machine Learning Research 5, 623-648 (2004)
    • (2004) Journal of Machine Learning Research , vol.5 , pp. 623-648
    • Mannor, S.1    Tsitsiklis, J.N.2
  • 16
    • 0001395850 scopus 로고
    • On the likelihood that one unknown probability exceeds another in view of the evidence of two samples
    • Thompson,W.R.: On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika 25(3-4), 285-294 (1933)
    • (1933) Biometrika , vol.25 , Issue.3-4 , pp. 285-294
    • Thompson, W.R.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.