Access the full text.
Sign up today, get DeepDyve free for 14 days.
G. Iba (1989)
A Heuristic Approach to the Discovery of Macro-OperatorsMachine Learning, 3
M. Stolle, C. Atkeson (2007)
Knowledge Transfer Using Local Features2007 IEEE International Symposium on Approximate Dynamic Programming and Reinforcement Learning
T. Howard, A. Kelly (2007)
Optimal Rough Terrain Trajectory Generation for Wheeled Mobile RobotsThe International Journal of Robotics Research, 26
L. Kavraki, P. Svestka, J. Latombe, M. Overmars (1996)
Probabilistic roadmaps for path planning in high-dimensional configuration spacesIEEE Trans. Robotics Autom., 12
S. Bailey, R. L. Grossman, L. Gu, D. Hanley (1996)
Hybrid systems III, Proceedings of the DIMACS workshop on verification and control of hybrid systems
Balaraman Ravindran, A. Barto (2003)
SMDP homomorphisms: an algebraic approach to abstraction in semi-Markov decision processes
Elly Winner, M. Veloso (2002)
Automatically Acquiring Planning Templates from Example Plans
B. Ravindran, A. G. Barto (2003)
Proceedings of the eighteenth international joint conference on artificial intelligence
M. Stolle, Hanns Tappeiner, Joel Chestnutt, C. Atkeson (2007)
Transfer of policies based on trajectory libraries2007 IEEE/RSJ International Conference on Intelligent Robots and Systems
M. Boddy, T. Dean (1994)
Deliberation Scheduling for Problem Solving in Time-Constrained EnvironmentsArtif. Intell., 67
S. LaValle (2006)
Planning algorithms
(2007)
Images of mazes used
B. Argall, Brett Browning, M. Veloso (2008)
Learning robot motion control with demonstration and advice-operators2008 IEEE/RSJ International Conference on Intelligent Robots and Systems
J. Friedman, J. Bentley, R. Finkel (1976)
An Algorithm for Finding Best Matches in Logarithmic Expected TimeACM Trans. Math. Softw., 3
David Conner, A. Rizzi, H. Choset (2003)
Composition of local potential functions for global robot control and navigationProceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453), 4
J. Laird, P. Rosenbloom, A. Newell (1986)
Chunking in Soar: The Anatomy of a General Learning MechanismMachine Learning, 1
(2001)
DIRCOL. http://www.sim.informatik.tudarmstadt.de/sw/dircol.html.en, URL http://www.sim.informatik. tu-darmstadt.de/sw/dircol.html.en
Shie Mannor, Ishai Menache, Amit Hoze, Uri Klein (2004)
Dynamic abstraction in reinforcement learning via clusteringProceedings of the twenty-first international conference on Machine learning
Emilio Frazzoli, M. Dahleh, E. Feron (2000)
Robust hybrid control for autonomous vehicle motion planningProceedings of the 39th IEEE Conference on Decision and Control (Cat. No.00CH37187), 1
J. Lee, J. Chai, P. S. A. Reitsma, J. K. Hodgins, N. S. Pollard (2002)
SIGGRAPH ’02: Proceedings of the 29th annual conference on computer graphics and interactive techniques
(2012)
Auton Robot
D. C. Conner, A. Rizzi, H. Choset (2003)
Proceedings of the international conference on intelligent robots and systems (IROS 2003)
M. Littman, R. Sutton, Satinder Singh (2001)
Predictive Representations of State
(1997)
Multidimensional interpolation and triangulation for reinforcement learning
M. Stolle, C. Atkeson (2006)
Policies based on trajectory librariesProceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006.
D. Grollman, O. Jenkins (2008)
Sparse incremental learning for interactive robot control policy estimation2008 IEEE International Conference on Robotics and Automation
S. Mahadevan (1992)
Enhancing Transfer in Reinforcement Learning by Building Stochastic Models of Robot Actions
Libo Yang, S. LaValle (2004)
The sampling-based neighborhood graph: an approach to computing and executing feedback motion strategiesIEEE Transactions on Robotics and Automation, 20
T. Röfer (2004)
Evolutionary Gait-Optimization Using a Fitness Function Based on Proprioception
E. Rimon, D. E. Koditschek (1992)
Exact robot navigation using artificial potential fieldsIEEE Transactions on Robotics and Automation, 8
Elizabeth Mcgovern, A. Barto (2002)
Autonomous discovery of temporal abstractions from interaction with an environment
M. Veloso (1992)
Learning by analogical reasoning in general problem-solving
Nate Kohl, P. Stone (2004)
Machine Learning for Fast Quadrupedal Locomotion
Paul Reitsma, N. Pollard (2004)
Evaluating motion graphs for character navigation
C. Connolly, R. Grupen (1993)
The applications of harmonic functions to roboticsJ. Field Robotics, 10
S. Bailey, R. Grossman, L. Gu, D. Hanley (1996)
A Data Intensive Computing Approach to Path Planning and Mode Management for Hybrid Systems
Joel Chestnutt, Manfred Lau, G. Cheung, J. Kuffner, J. Hodgins, T. Kanade (2005)
Footstep Planning for the Honda ASIMO HumanoidProceedings of the 2005 IEEE International Conference on Robotics and Automation
Carlos Guestrin, D. Koller, C. Gearhart, Neal Kanodia (2003)
Generalizing plans to new environments in relational MDPs
R. Fikes, P. Hart, N. Nilsson (1993)
Learning and Executing Generalized Robot PlansArtif. Intell., 3
M. L. Littman, R. S. Sutton, S. Singh (2002)
Advances in neural information processing systems
SermanetPierre, HadsellRaia, ScoffierMarco, GrimesMatt, BenJan, ErkanAyse, CrudeleChris, MillerUrs, LecunYann (2009)
A multirange architecture for collision-free off-road robot navigationJournal of Field Robotics
(2008)
Finding and transferring policies using stored behaviors
A. Fern, S. Yoon, R. Givan (2006)
Approximate policy iteration with a policy language bias: solving relational Markov decision processesJournal of Artificial Intelligence Research, 25
Özgür Simsek, A. Barto (2004)
Using relative novelty to identify useful temporal abstractions in reinforcement learningProceedings of the twenty-first international conference on Machine learning
R. Grossman, S. Mehta, X. Qin (1993)
Path planning by querying persistent stores of trajectory segments
H. Rosenbrock, D. Jacobson, D. Mayne (1972)
Differential dynamic programmingThe Mathematical Gazette, 56
C. Connolly, R. Grupen (1993)
The application of harmonic potential functions to roboticsJournal of Robotic Systems, 10
Alan Fern, S. Yoon, R. Givan (2003)
Approximate Policy Iteration with a Policy Language Bias
P. Montague (2005)
Reinforcement Learning: An IntroductionIEEE Transactions on Neural Networks
M. Stolle, D. Precup (2002)
Lecture notes in computer science
Alan Fern, S. Yoon, R. Givan (2004)
Learning Domain-Specific Control Knowledge from Random Walks
Jehee Lee, Jinxiang Chai, Paul Reitsma, J. Hodgins, N. Pollard (2002)
Interactive control of avatars animated with human motion dataProceedings of the 29th annual conference on Computer graphics and interactive techniques
R. Munos, A. Moore (2002)
Variable Resolution Discretization in Optimal ControlMachine Learning, 49
C. Atkeson (1993)
Using Local Trajectory Optimizers to Speed Up Global Optimization in Dynamic Programming
S. Chernova, M. Veloso (2004)
Learning and using models of kicking motions for legged robotsIEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004, 5
S. Chernova, M. Veloso (2008)
Learning equivalent action choices from demonstration2008 IEEE/RSJ International Conference on Intelligent Robots and Systems
S. Yoon, Alan Fern, R. Givan (2008)
Learning Control Knowledge for Forward Search PlanningJ. Mach. Learn. Res., 9
Darrin Bentivegna, C. Atkeson (2001)
Learning from observation using primitivesProceedings 2001 ICRA. IEEE International Conference on Robotics and Automation (Cat. No.01CH37164), 2
J. D. Weingarten, G. A. D. Lopes, M. Buehler, R. E. Groff, D. E. Koditschek (2004)
International conference in robotics and automation
S. LaValle, J. Kuffner (1999)
Randomized Kinodynamic PlanningThe International Journal of Robotics Research, 20
S. Chernova, M. Veloso (2004)
An evolutionary approach to gait learning for four-legged robots2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566), 3
L. Yang, S. M. LaValle (2004)
The sampling-based neighborhood graph: a framework for planning and executing feedback motion strategiesIEEE Transactions on Robotics and Automation, 20
J. Kolter, P. Abbeel, A. Ng (2007)
Hierarchical Apprenticeship Learning with Application to Quadruped Locomotion
M. Stolle, Doina Precup (2002)
Learning Options in Reinforcement Learning
J. Weingarten, G. Lopes, M. Buehler, R. Groff, D. Koditschek (2004)
Automated gait adaptation for legged robotsIEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004, 3
C. Atkeson, J. Morimoto (2002)
Nonparametric Representation of Policies and Value Functions: A Trajectory-Based Approach
Darrin Bentivegna, C. Atkeson, G. Cheng (2006)
Learning Similar Tasks From Observation and Practice2006 IEEE/RSJ International Conference on Intelligent Robots and Systems
E. Rimon, D. Koditschek (1992)
Exact robot navigation using artificial potential functionsIEEE Trans. Robotics Autom., 8
J. Pearl (1984)
Heuristics : intelligent search strategies for computer problem solving
Richard Sutton, A. Barto (1998)
Introduction to Reinforcement Learning
Manfred Lau, J. Kuffner (2005)
Behavior planning for character animation
R. Caruana (1993)
Multitask Learning: A Knowledge-Based Source of Inductive Bias
M. Lau, J. J. Kuffner (2005)
SCA ’05: proceedings of the 2005 ACM SIGGRAPH/Eurographics symposium on computer animation
We present several algorithms that aim to advance the state-of-the-art in reinforcement learning and planning algorithms. One key idea is to transfer knowledge across problems by representing it using local features. This idea is used to speed up a dynamic programming based generalized policy iteration.
Autonomous Robots – Springer Journals
Published: May 6, 2010
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.