-
2
-
-
84919787147
-
Taming the monster: A fast and simple algorithm for contextual bandits
-
Alekh Agarwal, Daniel Hsu, Satyen Kale, John Langford, Lihong Li, and Robert Schapire. Taming the monster: A fast and simple algorithm for contextual bandits. In 31st Intl. Conf. on Machine Learning (ICML), 2013.
-
(2013)
31st Intl. Conf. on Machine Learning (ICML)
-
-
Agarwal, A.1
Hsu, D.2
Kale, S.3
Langford, J.4
Li, L.5
Schapire, R.6
-
3
-
-
84939620711
-
Bandits with concave rewards and convex knapsacks
-
Shipra Agrawal and Nikhil R. Devanur. Bandits with concave rewards and convex knapsacks. In 15th, 2014.
-
(2014)
15th
-
-
Agrawal, S.1
Devanur, N.R.2
-
4
-
-
0041966002
-
Using confidence bounds for exploitation-exploration trade-offs
-
Preliminary version in 41st IEEE FOCS
-
Peter Auer. Using confidence bounds for exploitation-exploration trade-offs. J. of Machine Learning Research (JMLR), 3:397-422, 2002. Preliminary version in 41st IEEE FOCS, 2000.
-
(2000)
J. of Machine Learning Research (JMLR)
, vol.3
, pp. 397-422
-
-
Auer, P.1
-
5
-
-
0037709910
-
The nonstochastic multiarmed bandit problem
-
Preliminary version in 36th IEEE FOCS
-
Peter Auer, Nicolò Cesa-Bianchi, Yoav Freund, and Robert E. Schapire. The nonstochastic multiarmed bandit problem. SIAM J. Comput., 32(1):48-77, 2002. Preliminary version in 36th IEEE FOCS, 1995.
-
(1995)
SIAM J. Comput.
, vol.32
, Issue.1
, pp. 48-77
-
-
Auer, P.1
Cesa-Bianchi, N.2
Freund, Y.3
Schapire, R.E.4
-
10
-
-
70350251174
-
Dynamic pricing without knowing the demand function: Risk bounds and near-optimal algorithms
-
Omar Besbes and Assaf Zeevi. Dynamic pricing without knowing the demand function: Risk bounds and near-optimal algorithms. Operations Research, 57:1407-1420, 2009.
-
(2009)
Operations Research
, vol.57
, pp. 1407-1420
-
-
Besbes, O.1
Zeevi, A.2
-
11
-
-
84871887590
-
Blind network revenue management
-
Omar Besbes and Assaf J. Zeevi. Blind network revenue management. Operations Research, 60(6):1537-1550, 2012.
-
(2012)
Operations Research
, vol.60
, Issue.6
, pp. 1537-1550
-
-
Besbes, O.1
Zeevi, A.J.2
-
12
-
-
84939621750
-
Efficient optimal leanring for contextual bandits
-
Alina Beygelzimer, John Langford, Lihong Li, Lev Reyzin, and Robert E. Schapire. Efficient optimal leanring for contextual bandits. In 14th Intl. Conf. on Artificial Intelligence and Statistics (AISTATS), 2011.
-
(2011)
14th Intl. Conf. on Artificial Intelligence and Statistics (AISTATS)
-
-
Beygelzimer, A.1
Langford, J.2
Li, L.3
Reyzin, L.4
Schapire, R.E.5
-
13
-
-
84874045238
-
Regret analysis of stochastic and nonstochastic multiarmed bandit problems
-
Sébastien Bubeck and Nicolo Cesa-Bianchi. Regret Analysis of Stochastic and Nonstochastic Multiarmed Bandit Problems. Foundations and Trends in Machine Learning, 5(1):1-122, 2012.
-
(2012)
Foundations and Trends in Machine Learning
, vol.5
, Issue.1
, pp. 1-122
-
-
Bubeck, S.1
Cesa-Bianchi, N.2
-
15
-
-
4544235463
-
The spending constraint model for market equilibrium: Algorithmic, existence and uniqueness results
-
Nikhil Devanur and Vijay Vazirani. The spending constraint model for market equilibrium: Algorithmic, existence and uniqueness results. In 36th ACM Symp. on Theory of Computing (STOC), 2004.
-
(2004)
36th ACM Symp. on Theory of Computing (STOC)
-
-
Devanur, N.1
Vazirani, V.2
-
16
-
-
76749148680
-
The AdWords problem: Online keyword matching with budgeted bidders under random permutations
-
Nikhil R. Devanur and Thomas P. Hayes. The AdWords problem: Online keyword matching with budgeted bidders under random permutations. In 10th ACM Conf. on Electronic Commerce (EC), pages 71-78, 2009.
-
(2009)
10th ACM Conf. on Electronic Commerce (EC)
, pp. 71-78
-
-
Devanur, N.R.1
Hayes, T.P.2
-
17
-
-
79959593176
-
Near optimal online algorithms and fast approximation algorithms for resource allocation problems
-
Nikhil R. Devanur, Kamal Jain, Balasubramanian Sivan, and Christopher A. Wilkens. Near optimal online algorithms and fast approximation algorithms for resource allocation problems. In 12th ACM Conf. on Electronic Commerce (EC), pages 29-38, 2011.
-
(2011)
12th ACM Conf. on Electronic Commerce (EC)
, pp. 29-38
-
-
Devanur, N.R.1
Jain, K.2
Sivan, B.3
Wilkens, C.A.4
-
18
-
-
80053154335
-
Efficient optimal leanring for contextual bandits
-
Miroslav Dudik, Daniel Hsu, Satyen Kale, Nikos Karampatziakis, John Langford, Lev Reyzin, and Tong Zhang. Efficient optimal leanring for contextual bandits. In 27th Conf. on Uncertainty in Artificial Intelligence (UAI), 2011.
-
(2011)
27th Conf. on Uncertainty in Artificial Intelligence (UAI)
-
-
Dudik, M.1
Hsu, D.2
Kale, S.3
Karampatziakis, N.4
Langford, J.5
Reyzin, L.6
Zhang, T.7
-
19
-
-
0002384441
-
On tail probabilities for martingales
-
D. A. Freedman. On tail probabilities for martingales. The Annals of Probability, 3:100-118, 1975.
-
(1975)
The Annals of Probability
, vol.3
, pp. 100-118
-
-
Freedman, D.A.1
-
22
-
-
69449097218
-
Approximation algorithms for restless bandit problems.
-
Combined final version of papers in
-
Sudipta Guha, Kamesh Munagala, and Peng Shi. Approximation algorithms for restless bandit problems., 2010. Combined final version of papers in IEEE FOCS 2007 and ACM-SIAM SODA 2009.
-
(2009)
IEEE FOCS 2007 and ACM-SIAM SODA
-
-
Guha, S.1
Munagala, K.2
Shi, P.3
-
23
-
-
84863332792
-
Approximation algorithms for correlated knapsacks and non-martingale bandits
-
Anupam Gupta, Ravishankar Krishnaswamy, Marco Molinaro, and R. Ravi. Approximation algorithms for correlated knapsacks and non-martingale bandits. In 52nd IEEE Symp. on Foundations of Computer Science (FOCS), pages 827-836, 2011.
-
(2011)
52nd IEEE Symp. on Foundations of Computer Science (FOCS)
, pp. 827-836
-
-
Gupta, A.1
Krishnaswamy, R.2
Molinaro, M.3
Ravi, R.4
-
24
-
-
84880882858
-
Continuous time associative bandit problems
-
András György, Levente Kocsis, Ivett Szabó, and Csaba Szepesvári. Continuous time associative bandit problems. In 20th Intl. Joint Conf. on Artificial Intelligence (IJCAI), pages 830-835, 2007.
-
(2007)
20th Intl. Joint Conf. on Artificial Intelligence (IJCAI)
, pp. 830-835
-
-
György, A.1
Kocsis, L.2
Szabó, I.3
Szepesvári, C.4
-
29
-
-
84893043989
-
Truthful incentives in crowdsourcing tasks using regret minimization mechanisms
-
Adish Singla and Andreas Krause. Truthful incentives in crowdsourcing tasks using regret minimization mechanisms. In 22nd Intl. World Wide Web Conf. (WWW), pages 1167-1178, 2013.
-
(2013)
22nd Intl. World Wide Web Conf. (WWW)
, pp. 1167-1178
-
-
Singla, A.1
Krause, A.2
-
30
-
-
84972513554
-
On general minimax theorems
-
Maurice Sion. On general minimax theorems. Pac. J. Math., 8:171176, 1958.
-
(1958)
Pac. J. Math.
, vol.8
, pp. 171176
-
-
Sion, M.1
-
31
-
-
84874058621
-
Contextual bandits with similarity information
-
To appear in J. of Machine Learning Research JMLR
-
Aleksandrs Slivkins. Contextual Bandits with Similarity Information. In 24th Conf. on Learning Theory (COLT), 2011. To appear in J. of Machine Learning Research (JMLR), 2014.
-
(2011)
24th Conf. on Learning Theory (COLT)
-
-
Slivkins, A.1
-
33
-
-
84963496106
-
Online decision making in crowdsourcing markets: Theoretical challenges
-
December, Position Paper and survey
-
Aleksandrs Slivkins and Jennifer Wortman Vaughan. Online decision making in crowdsourcing markets: Theoretical challenges. SIGecom Exchanges, 12(2), December 2013. Position Paper and survey.
-
(2013)
SIGecom Exchanges
, vol.12
, Issue.2
-
-
Slivkins, A.1
Vaughan, J.W.2
-
34
-
-
0001395850
-
On the likelihood that one unknown probability exceeds another in view of the evidence of two samples
-
William R. Thompson. On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika, 25(3-4):285294, 1933.
-
(1933)
Biometrika
, vol.25
, Issue.3-4
, pp. 285294
-
-
Thompson, W.R.1
-
35
-
-
77958583895
-
E-first policies for budget-limited multi-armed bandits
-
Long Tran-Thanh, Archie Chapman, Enrique Munoz De Cote, Alex Rogers, and Nicholas R. Jennings. e-first policies for budget-limited multi-armed bandits. In 24th AAAI Conference on Artificial Intelligence (AAAI), pages 1211-1216, 2010.
-
(2010)
24th AAAI Conference on Artificial Intelligence (AAAI)
, pp. 1211-1216
-
-
Tran-Thanh, L.1
Chapman, A.2
De Cote, E.M.3
Rogers, A.4
Jennings, N.R.5
-
36
-
-
84868281643
-
Knapsack based optimal policies for budget-limited multi-armed bandits
-
Long Tran-Thanh, Archie Chapman, Alex Rogers, and Nicholas R. Jennings. Knapsack based optimal policies for budget-limited multi-armed bandits. In 26th AAAI Conference on Artificial Intelligence (AAAI), pages 1134-1140, 2012.
-
(2012)
26th AAAI Conference on Artificial Intelligence (AAAI)
, pp. 1134-1140
-
-
Tran-Thanh, L.1
Chapman, A.2
Rogers, A.3
Jennings, N.R.4
|