-
1
-
-
80052395875
-
Preference-based policy learning
-
Dimitrios Gunopulos, Thomas Hofmann, Donato Malerba, and Michalis Vazirgiannis, editors, volume 6911 of Lecture Notes in Computer Science, Springer
-
R. Akrour, M. Schoenauer, and M. Sebag. Preference-based policy learning. In Dimitrios Gunopulos, Thomas Hofmann, Donato Malerba, and Michalis Vazirgiannis, editors, Proc. ECML/PKDD'11, Part I, volume 6911 of Lecture Notes in Computer Science, pages 12-27. Springer, 2011.
-
(2011)
Proc. ECML/PKDD'11, Part i
, pp. 12-27
-
-
Akrour, R.1
Schoenauer, M.2
Sebag, M.3
-
2
-
-
0037262814
-
An introduction to mcmc for machine learning
-
Christophe Andrieu, Nando de Freitas, Arnaud Doucet, and Michael I. Jordan. An introduction to mcmc for machine learning. Machine Learning, 50(1-2):5-43, 2003.
-
(2003)
Machine Learning
, vol.50
, Issue.1-2
, pp. 5-43
-
-
Andrieu, C.1
De Freitas, N.2
Doucet, A.3
Jordan, M.I.4
-
3
-
-
63149159130
-
A survey of robot learning from demonstration
-
May
-
Brenna D. Argall, Sonia Chernova, Manuela Veloso, and Brett Browning. A survey of robot learning from demonstration. Robot. Auton. Syst., 57(5):469-483, May 2009.
-
(2009)
Robot. Auton. Syst.
, vol.57
, Issue.5
, pp. 469-483
-
-
Argall, B.D.1
Chernova, S.2
Veloso, M.3
Browning, B.4
-
4
-
-
0000626524
-
Expected information as expected utility
-
J M Bernardo. Expected information as expected utility. Annals of Statistics, 7(3):686-690, 1979.
-
(1979)
Annals of Statistics
, vol.7
, Issue.3
, pp. 686-690
-
-
Bernardo, J.M.1
-
5
-
-
80052414142
-
Preferencebased policy iteration: Leveraging preference learning for reinforcement learning
-
Springer
-
Weiwei Cheng, Johannes Fürnkranz, Eyke Hüllermeier, and Sang-Hyeun Park. Preferencebased policy iteration: Leveraging preference learning for reinforcement learning. In Proceedings of the 22nd European Conference on Machine Learning (ECML 2011), pages 312-327. Springer, 2011.
-
(2011)
Proceedings of the 22nd European Conference on Machine Learning (ECML 2011)
, pp. 312-327
-
-
Cheng, W.1
Fürnkranz, J.2
Hüllermeier, E.3
Park, S.-H.4
-
6
-
-
31844453751
-
Preference learning with gaussian processes
-
New York, NY, USA, ACM
-
Wei Chu and Zoubin Ghahramani. Preference learning with gaussian processes. In Proceedings of the 22nd international conference on Machine learning, ICML '05, pages 137-144, New York, NY, USA, 2005. ACM.
-
(2005)
Proceedings of the 22nd International Conference on Machine Learning, ICML '05
, pp. 137-144
-
-
Chu, W.1
Ghahramani, Z.2
-
8
-
-
4243137056
-
Hybrid monte carlo
-
Simon Duane, A. D. Kennedy, Brian J. Pendleton, and Duncan Roweth. Hybrid monte carlo. Physics Letters B, 195(2):216-222, 1987.
-
(1987)
Physics Letters B
, vol.195
, Issue.2
, pp. 216-222
-
-
Duane, S.1
Kennedy, A.D.2
Pendleton, B.J.3
Roweth, D.4
-
9
-
-
0031209604
-
Selective sampling using the query by committee algorithm
-
Yoav Freund, H. Sebastian Seung, Eli Shamir, and Naftali Tishby. Selective sampling using the query by committee algorithm. Machine Learning, 28(2-3):133-168, 1997.
-
(1997)
Machine Learning
, vol.28
, Issue.2-3
, pp. 133-168
-
-
Freund, Y.1
Seung, H.S.2
Shamir, E.3
Tishby, N.4
-
11
-
-
0001249987
-
On a measure of the information provided by an experiment
-
D. V. Lindley. On a Measure of the Information Provided by an Experiment. The Annals of Mathematical Statistics, 27(4):986-1005, 1956.
-
(1956)
The Annals of Mathematical Statistics
, vol.27
, Issue.4
, pp. 986-1005
-
-
Lindley, D.V.1
-
12
-
-
0042547347
-
Algorithms for inverse reinforcement learning
-
Andrew Y. Ng and Stuart J. Russell. Algorithms for inverse reinforcement learning. In ICML, pages 663-670, 2000.
-
(2000)
ICML
, pp. 663-670
-
-
Ng, A.Y.1
Russell, S.J.2
-
13
-
-
27344432348
-
Accelerating reinforcement learning through implicit imitation
-
Bob Price and Craig Boutilier. Accelerating reinforcement learning through implicit imitation. J. Artif. Intell. Res. (JAIR), 19:569-629, 2003.
-
(2003)
J. Artif. Intell. Res. (JAIR)
, vol.19
, pp. 569-629
-
-
Price, B.1
Boutilier, C.2
-
14
-
-
1642401055
-
Learning to drive a bicycle using reinforcement learning and shaping
-
Jette Randløv and Preben Alstrøm. Learning to drive a bicycle using reinforcement learning and shaping. In ICML, pages 463-471, 1998.
-
(1998)
ICML
, pp. 463-471
-
-
Randløv, J.1
Alstrøm, P.2
-
15
-
-
84898995067
-
Learning from demonstration
-
Stefan Schaal. Learning from demonstration. In NIPS, pages 1040-1046, 1996.
-
(1996)
NIPS
, pp. 1040-1046
-
-
Schaal, S.1
-
16
-
-
34548168342
-
Active learning for logistic regression: An evaluation
-
October
-
Andrew I. Schein and Lyle H. Ungar. Active learning for logistic regression: an evaluation. Mach. Learn., 68(3):235-265, October 2007.
-
(2007)
Mach. Learn.
, vol.68
, Issue.3
, pp. 235-265
-
-
Schein, A.I.1
Ungar, L.H.2
-
17
-
-
0026981853
-
Query by committee
-
New York, NY, USA. ACM
-
H. S. Seung, M. Opper, and H. Sompolinsky. Query by committee. In Proceedings of the fifth annual workshop on Computational learning theory, COLT '92, pages 287-294, New York, NY, USA, 1992. ACM.
-
(1992)
Proceedings of the Fifth Annual Workshop on Computational Learning Theory, COLT '92
, pp. 287-294
-
-
Seung, H.S.1
Opper, M.2
Sompolinsky, H.3
|