-
1
-
-
0034870014
-
Vision-based reinforcement learning for robot navigation
-
Washington, DC
-
W. Zhu and S. Levinson, "Vision-based reinforcement learning for robot navigation," in Proc. Int. Joint Conf. Neural Netw., Washington, DC, 2001, vol. 2, pp. 1025-1030.
-
(2001)
Proc. Int. Joint Conf. Neural Netw.
, vol.2
, pp. 1025-1030
-
-
Zhu, W.1
Levinson, S.2
-
2
-
-
8444249971
-
Learning from reinforcement and advice using composite reward functions
-
St. Augustine, FL
-
V. N. Papudesi and M. Huber, "Learning from reinforcement and advice using composite reward functions," in Proc. 16th Int. FLAIRS Conf., St. Augustine, FL, 2003, pp. 361-365.
-
(2003)
Proc. 16th Int. FLAIRS Conf.
, pp. 361-365
-
-
Papudesi, V.N.1
Huber, M.2
-
3
-
-
0346872398
-
Integrating user commands and autonomous task performance in a reinforcement learning framework
-
Stanford, CA, Stanford Univ.
-
V. N. Papudesi, Y. Wang, M. Huber, and D. J. Cook, "Integrating user commands and autonomous task performance in a reinforcement learning framework," in Proc. AAAI Spring Symp. Human Interaction Autonom. Syst.Complex Environ., Stanford, CA, 2003, pp. 160-165, Stanford Univ.
-
(2003)
Proc. AAAI Spring Symp. Human Interaction Autonom. Syst.Complex Environ.
, pp. 160-165
-
-
Papudesi, V.N.1
Wang, Y.2
Huber, M.3
Cook, D.J.4
-
4
-
-
72149126901
-
Stabilization of biped robot based on two mode Q-learning
-
P. Kui-Hong, J. Jun, and K. Jong-Hwan, "Stabilization of biped robot based on two mode Q-learning," in Proc. 2nd Int. Conf. Autonom. Robots Agents, 2004, pp. 446-451.
-
(2004)
Proc. 2nd Int. Conf. Autonom. Robots Agents
, pp. 446-451
-
-
Kui-Hong, P.1
Jun, J.2
Jong-Hwan, K.3
-
5
-
-
33846153561
-
Robot learning in partially observable, noisy, continuous worlds
-
Barcelona, Spain
-
R. Broadbent and T. Peterson, "Robot learning in partially observable, noisy, continuous worlds," in Proc. 2005 IEEE Int. Conf. Robot. Autom., Barcelona, Spain, 2005, pp. 4386-4393.
-
(2005)
Proc. 2005 IEEE Int. Conf. Robot. Autom.
, pp. 4386-4393
-
-
Broadbent, R.1
Peterson, T.2
-
6
-
-
33845607326
-
Quasi-online reinforcement learning for robots
-
B. Bakker, V. Zhumatiy, G. Gruener, and J. Schmidhuber, "Quasi-online reinforcement learning for robots," in Proc. 2006 IEEE Int. Conf. Robot. Autom., 2006, pp. 2997-3002.
-
(2006)
Proc. 2006 IEEE Int. Conf. Robot. Autom.
, pp. 2997-3002
-
-
Bakker, B.1
Zhumatiy, V.2
Gruener, G.3
Schmidhuber, J.4
-
7
-
-
34548119090
-
Human-robot collaborative learning system for inspection
-
Taipei, Taiwan, Oct.
-
U. Kartoun, H. Stern, and Y. Edan, "Human-robot collaborative learning system for inspection," in Proc. IEEE Int. Conf. Syst., Man, Cybern., Taipei, Taiwan, Oct. 2006, pp. 4249-4255.
-
(2006)
Proc. IEEE Int. Conf. Syst., Man, Cybern.
, pp. 4249-4255
-
-
Kartoun, U.1
Stern, H.2
Edan, Y.3
-
8
-
-
33746102906
-
Using active relocation to aid reinforcement
-
Melbourne Beach, FL
-
L. Mihalkova and R. Mooney, "Using active relocation to aid reinforcement," in Proc. 19th Int. FLAIRS Conf. (FLAIRS-2006), Melbourne Beach, FL, 2006, pp. 580-585.
-
(2006)
Proc. 19th Int. FLAIRS Conf. (FLAIRS-2006)
, pp. 580-585
-
-
Mihalkova, L.1
Mooney, R.2
-
9
-
-
34147169178
-
A multiresolution analysis-assisted reinforcement learning approach to run-by-run control
-
R. Ganesan, T. K. Das, and K. M. Ramachandran, "A multiresolution analysis-assisted reinforcement learning approach to run-by-run control," IEEE Trans. Autom. Sci. Eng., vol. 4, no. 2, pp. 182-193, 2007.
-
(2007)
IEEE Trans. Autom. Sci. Eng.
, vol.4
, Issue.2
, pp. 182-193
-
-
Ganesan, R.1
Das, T.K.2
Ramachandran, K.M.3
-
10
-
-
0004049893
-
-
Ph.D. dissertation, Psychol. Dept. Cambridge Univ. Cambridge, U.K.
-
C. J. C. H. Watkins, "Learning from delayed rewards," Ph.D. dissertation, Psychol. Dept., Cambridge Univ., Cambridge, U.K., 1989.
-
(1989)
Learning from Delayed Rewards
-
-
Watkins, C.J.C.H.1
-
11
-
-
0000955979
-
Incremental multi-step Q-learning
-
J. Peng and R. Williams, "Incremental multi-step Q-learning," Mach. Learn., vol. 22, no. 1-3, pp. 283-290, 1996.
-
(1996)
Mach. Learn.
, vol.22
, Issue.1-3
, pp. 283-290
-
-
Peng, J.1
Williams, R.2
-
12
-
-
34548115726
-
Seek of an optimal way by Q-learning
-
Y. Dahmani and A. Benyettou, "Seek of an optimal way by Q-learning," J. Comput. Sci., vol. 1, no. 1, pp. 28-30, 2005.
-
(2005)
J. Comput. Sci.
, vol.1
, Issue.1
, pp. 28-30
-
-
Dahmani, Y.1
Benyettou, A.2
-
13
-
-
32644453016
-
Representation for knot-tying tasks
-
J. Takamatsu, T. Morita, K. Ogawara, H. Kimura, and K. Ikeuchi, "Representation for knot-tying tasks," IEEE Trans. Robot., vol. 22, no. 1, pp. 65-78, 2006.
-
(2006)
IEEE Trans. Robot.
, vol.22
, Issue.1
, pp. 65-78
-
-
Takamatsu, J.1
Morita, T.2
Ogawara, K.3
Kimura, H.4
Ikeuchi, K.5
-
14
-
-
33645677666
-
Knotting/unknotting manipulation of deformable linear objects
-
H. Wakamatsu, A. Eiji, and H. Shinichi, "Knotting/unknotting manipulation of deformable linear objects," Int. J. Robot. Res., vol. 25, no. 4, pp. 371-395, 2006.
-
(2006)
Int. J. Robot. Res.
, vol.25
, Issue.4
, pp. 371-395
-
-
Wakamatsu, H.1
Eiji, A.2
Shinichi, H.3
-
15
-
-
34250630848
-
Manipulation of flexible rope using topological model based on sensor information
-
T. Matsuno and T. Fukuda, "Manipulation of flexible rope using topological model based on sensor information," in Proc. Int. Conf. Intell. Robots Syst., 2006, pp. 2638-2643.
-
(2006)
Proc. Int. Conf. Intell. Robots Syst.
, pp. 2638-2643
-
-
Matsuno, T.1
Fukuda, T.2
-
16
-
-
37549020027
-
Motion planning for robotic manipulation of deformable linear objects
-
M. Saha and P. Isto, "Motion planning for robotic manipulation of deformable linear objects," in Proc. Int. Conf. Intell. Robots Autom., 2007, vol. 23, no. 6, pp. 1141-1150.
-
(2007)
Proc. Int. Conf. Intell. Robots Autom.
, vol.23
, Issue.6
, pp. 1141-1150
-
-
Saha, M.1
Isto, P.2
-
17
-
-
0036570250
-
Reinforcement learning agents
-
C. Ribeiro, "Reinforcement learning agents," Artificial Intelligence Review, vol. 17, no. 3, pp. 223-250, 2002.
-
(2002)
Artificial Intelligence Review
, vol.17
, Issue.3
, pp. 223-250
-
-
Ribeiro, C.1
-
20
-
-
33845529505
-
Reinforcement learning: An overview
-
Aachen, Germany
-
P. Y. Glorennec, "Reinforcement learning: An overview," in Eur. Symp. Intell. Tech., Aachen, Germany, 2000, pp. 17-35.
-
(2000)
Eur. Symp. Intell. Tech.
, pp. 17-35
-
-
Glorennec, P.Y.1
-
21
-
-
0348041527
-
User-guided reinforcement learning of robot assistive tasks for an intelligent environment
-
Y. Wang, M. Huber, V. N. Papudesi, and D. J. Cook, "User-guided reinforcement learning of robot assistive tasks for an intelligent environment," in Proc. IEEE/RJS Int. Conf. Intell. Robots Syst., 2003, vol. 1, pp. 424-429.
-
(2003)
Proc. IEEE/RJS Int. Conf. Intell. Robots Syst.
, vol.1
, pp. 424-429
-
-
Wang, Y.1
Huber, M.2
Papudesi, V.N.3
Cook, D.J.4
-
22
-
-
33845430870
-
Bag classification using support vector machines
-
Heidelberg, Germany: Springer Berlin
-
U. Kartoun, H. Stern, and Y. Edan, "Bag classification using support vector machines," in Applied Soft Computing Technologies: The Challenge of Complexity Series: Advances in Soft Computing. Heidelberg, Germany: Springer Berlin, 2006, pp. 665-674.
-
(2006)
Applied Soft Computing Technologies: The Challenge of Complexity Series: Advances in Soft Computing
, pp. 665-674
-
-
Kartoun, U.1
Stern, H.2
Edan, Y.3
-
23
-
-
73849140859
-
-
Ph.D. dissertation, Dept. of Ind. Eng. Manage. Ben-Gurion Univ. of the Negev, Beer-Sheva, Israel
-
U. Kartoun, "Human-Robot Collaborative Learning Methods," Ph.D. dissertation, Dept. of Ind. Eng. Manage., Ben-Gurion Univ. of the Negev, Beer-Sheva, Israel, 2007.
-
(2007)
Human-robot Collaborative Learning Methods
-
-
Kartoun, U.1
|