Andrew G. Barto

Orcid: 0000-0002-3581-1896

Affiliations:
  • Department of Computer Science, University of Massachusetts Amherst


According to our database1, Andrew G. Barto authored at least 105 papers between 1978 and 2021.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2021
Looking Back on the Actor-Critic Architecture.
IEEE Trans. Syst. Man Cybern. Syst., 2021

2019
Editorial: Intrinsically Motivated Open-Ended Learning in Autonomous Robots.
Frontiers Neurorobotics, 2019

Reinforcement Learning: Connections, Surprises, and Challenge.
AI Mag., 2019

2017
Adaptive Real-Time Dynamic Programming.
Proceedings of the Encyclopedia of Machine Learning and Data Mining, 2017

On Ensuring that Intelligent Machines Are Well-Behaved.
CoRR, 2017

2015
Learning grounded finite-state representations from unstructured demonstrations.
Int. J. Robotics Res., 2015

Online Bayesian changepoint detection for articulated motion models.
Proceedings of the IEEE International Conference on Robotics and Automation, 2015

2014
Commentary on Utility and Bounds.
Top. Cogn. Sci., 2014

Optimal Behavioral Hierarchy.
PLoS Comput. Biol., 2014

Learning parameterized motor skills on a humanoid robot.
Proceedings of the 2014 IEEE International Conference on Robotics and Automation, 2014

Active Learning of Parameterized Skills.
Proceedings of the 31th International Conference on Machine Learning, 2014

2013
Variable risk control via stochastic optimization.
Int. J. Robotics Res., 2013

Incremental Semantically Grounded Learning from Demonstration.
Proceedings of the Robotics: Science and Systems IX, Technische Universität Berlin, Berlin, Germany, June 24, 2013

Intrinsic Motivation and Reinforcement Learning.
Proceedings of the Intrinsically Motivated Learning in Natural and Artificial Systems, 2013

Behavioral Hierarchy: Exploration and Representation.
Proceedings of the Computational and Robotic Models of the Hierarchical Organization of Behavior, 2013

2012
Transfer in Reinforcement Learning via Shared Features.
J. Mach. Learn. Res., 2012

Robot learning from demonstration by constructing skill trees.
Int. J. Robotics Res., 2012

Variational Bayesian Optimization for Runtime Risk-Sensitive Control.
Proceedings of the Robotics: Science and Systems VIII, 2012

Learning and generalization of complex tasks from unstructured demonstrations.
Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2012

Learning Parameterized Skills.
Proceedings of the 29th International Conference on Machine Learning, 2012

Motor primitive discovery.
Proceedings of the 2012 IEEE International Conference on Development and Learning and Epigenetic Robotics, 2012

TD-DeltaPi: A Model-Free Algorithm for Efficient Exploration.
Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence, 2012

Adaptive Step-Size for Online Temporal Difference Learning.
Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence, 2012

2011
Conjugate Markov Decision Processes.
Proceedings of the 28th International Conference on Machine Learning, 2011

Learning dynamic arm motions for postural recovery.
Proceedings of the 11th IEEE-RAS International Conference on Humanoid Robots (Humanoids 2011), 2011

Evolution of reward functions for reinforcement learning.
Proceedings of the 13th Annual Genetic and Evolutionary Computation Conference, 2011

Clustering via Dirichlet Process Mixture Models for Portable Skill Discovery.
Proceedings of the Lifelong Learning, 2011

Autonomous Skill Acquisition on a Mobile Manipulator.
Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence, 2011

2010
Adaptive Real-Time Dynamic Programming.
Proceedings of the Encyclopedia of Machine Learning, 2010

Intrinsically Motivated Hierarchical Skill Learning in Structured Environments.
IEEE Trans. Auton. Ment. Dev., 2010

Intrinsically Motivated Reinforcement Learning: An Evolutionary Perspective.
IEEE Trans. Auton. Ment. Dev., 2010

Genetic Programming for Reward Function Search.
IEEE Trans. Auton. Ment. Dev., 2010

Constructing Skill Trees for Reinforcement Learning Agents from Demonstration Trajectories.
Proceedings of the Advances in Neural Information Processing Systems 23: 24th Annual Conference on Neural Information Processing Systems 2010. Proceedings of a meeting held 6-9 December 2010, 2010

Competence progress intrinsic motivation.
Proceedings of the 2010 IEEE 9th International Conference on Development and Learning, 2010

2009
Skill Discovery in Continuous Reinforcement Learning Domains using Skill Chaining.
Proceedings of the Advances in Neural Information Processing Systems 22: 23rd Annual Conference on Neural Information Processing Systems 2009. Proceedings of a meeting held 7-10 December 2009, 2009

Efficient Skill Learning using Abstraction Selection.
Proceedings of the IJCAI 2009, 2009

2008
Skill Characterization Based on Betweenness.
Proceedings of the Advances in Neural Information Processing Systems 21, 2008

Hierarchical Representations of Behavior for Efficient Creative Search.
Proceedings of the Creative Intelligent Systems, 2008

2007
Temporal difference learning.
Scholarpedia, 2007

Adaptive Control of Duty Cycling in Energy-Harvesting Wireless Sensor Networks.
Proceedings of the Fourth Annual IEEE Communications Society Conference on Sensor, 2007

Active Learning of Dynamic Bayesian Networks in Markov Decision Processes.
Proceedings of the Abstraction, 2007

Deictic Option Schemas.
Proceedings of the IJCAI 2007, 2007

Building Portable Options: Skill Transfer in Reinforcement Learning.
Proceedings of the IJCAI 2007, 2007

Repairing Disengagement With Non-Invasive Interventions.
Proceedings of the Artificial Intelligence in Education, 2007

2006
Learning at the level of synergies for a robot weightlifter.
Robotics Auton. Syst., 2006

Causal Graph Based Decomposition of Factored MDPs.
J. Mach. Learn. Res., 2006

An Adaptive Robot Motivational System.
Proceedings of the From Animals to Animats 9, 2006

Improving Intelligent Tutoring Systems: Using Expectation Maximization to Learn Student Skill Levels.
Proceedings of the Intelligent Tutoring Systems, 8th International Conference, 2006

An intrinsic reward mechanism for efficient exploration.
Proceedings of the Machine Learning, 2006

Autonomous shaping: knowledge transfer in reinforcement learning.
Proceedings of the Machine Learning, 2006

Decision Tree Methods for Finding Reusable MDP Homomorphisms.
Proceedings of the Proceedings, 2006

2005
Learning Skills in Reinforcement Learning Using Relative Novelty.
Proceedings of the Abstraction, 2005

Identifying useful subgoals in reinforcement learning by local graph partitioning.
Proceedings of the Machine Learning, 2005

A causal approach to hierarchical decomposition of factored MDPs.
Proceedings of the Machine Learning, 2005

2004
Intrinsically Motivated Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 17 [Neural Information Processing Systems, 2004

Using relative novelty to identify useful temporal abstractions in reinforcement learning.
Proceedings of the Machine Learning, 2004

Reinforcement learning with supervision by a stable controller.
Proceedings of the 2004 American Control Conference, 2004

2003
Recent Advances in Hierarchical Reinforcement Learning.
Discret. Event Dyn. Syst., 2003

SMDP Homomorphisms: An Algebraic Approach to Abstraction in Semi-Markov Decision Processes.
Proceedings of the IJCAI-03, 2003

Relativized Options: Choosing the Right Transformation.
Proceedings of the Machine Learning, 2003

2002
Building a Basic Block Instruction Scheduler with Reinforcement Learning and Rollouts.
Mach. Learn., 2002

Lyapunov Design for Safe Reinforcement Learning.
J. Mach. Learn. Res., 2002

The emergence of movement units through learning with noisy efferent signals and delayed sensory feedback.
Neurocomputing, 2002

Model Minimization in Hierarchical Reinforcement Learning.
Proceedings of the Abstraction, 2002

PolicyBlocks: An Algorithm for Creating Useful Macro-Actions in Reinforcement Learning.
Proceedings of the Machine Learning, 2002

2001
The Emergence of Multiple Movement Units in the Presence of Noise and Feedback Delay.
Proceedings of the Advances in Neural Information Processing Systems 14 [Neural Information Processing Systems: Natural and Synthetic, 2001

Robot Weightlifting By Direct Policy Search.
Proceedings of the Seventeenth International Joint Conference on Artificial Intelligence, 2001

Heuristic Search in Infinite State Spaces Guided by Lyapunov Analysis.
Proceedings of the Seventeenth International Joint Conference on Artificial Intelligence, 2001

Lyapunov-Constrained Action Sets for Reinforcement Learning.
Proceedings of the Eighteenth International Conference on Machine Learning (ICML 2001), Williams College, Williamstown, MA, USA, June 28, 2001

Automatic Discovery of Subgoals in Reinforcement Learning using Diverse Density.
Proceedings of the Eighteenth International Conference on Machine Learning (ICML 2001), Williams College, Williamstown, MA, USA, June 28, 2001

2000
Automated State Abstraction for Options using the U-Tree Algorithm.
Proceedings of the Advances in Neural Information Processing Systems 13, 2000

Combining Reinforcement Learning with a Local Control Algorithm.
Proceedings of the Seventeenth International Conference on Machine Learning (ICML 2000), Stanford University, Stanford, CA, USA, June 29, 2000

Machine Learning for Subproblem Selection.
Proceedings of the Seventeenth International Conference on Machine Learning (ICML 2000), Stanford University, Stanford, CA, USA, June 29, 2000

1999
A Cerebellar Model of Timing and Prediction in the Control of Reaching.
Neural Comput., 1999

1998
Reinforcement Learning: An Introduction.
IEEE Trans. Neural Networks, 1998

Elevator Group Control Using Multiple Reinforcement Learning Agents.
Mach. Learn., 1998

Learning Instance-Independent Value Functions to Enhance Local Search.
Proceedings of the Advances in Neural Information Processing Systems 11, [NIPS Conference, Denver, Colorado, USA, November 30, 1998

Reinforcement learning - an introduction.
Adaptive computation and machine learning, MIT Press, ISBN: 978-0-262-19398-6, 1998

1997
Automated Aircraft Recovery via Reinforcement Learning: Initial Experiments.
Proceedings of the Advances in Neural Information Processing Systems 10, 1997

Cerebellar learning for control of a two-link arm in muscle space.
Proceedings of the 1997 IEEE International Conference on Robotics and Automation, 1997

A model of cerebellar learning for control of arm movements using muscle synergies.
Proceedings of the Proceedings 1997 IEEE International Symposium on Computational Intelligence in Robotics and Automation CIRA'97, 1997

1996
Linear Least-Squares Algorithms for Temporal Difference Learning.
Mach. Learn., 1996

Text-Based Information Retrieval Using Exponentiated Gradient Descent.
Proceedings of the Advances in Neural Information Processing Systems 9, 1996

Reinforcement Learning for Mixed Open-loop and Closed-loop Control.
Proceedings of the Advances in Neural Information Processing Systems 9, 1996

Local Bandit Approximation for Optimal Learning Problems.
Proceedings of the Advances in Neural Information Processing Systems 9, 1996

1995
Learning to Act Using Real-Time Dynamic Programming.
Artif. Intell., 1995

Improving Elevator Performance Using Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 8, 1995

A Predictive Switching Model of Cerebellar Movement Control.
Proceedings of the Advances in Neural Information Processing Systems 8, 1995

1994
An Actor/Critic Algorithm that is Equivalent to Q-Learning.
Proceedings of the Advances in Neural Information Processing Systems 7, 1994

Learning Admittance Mappings for Force-Guided Assembly.
Proceedings of the 1994 International Conference on Robotics and Automation, 1994

1993
Task Decompostiion Through Competition in a Modular Connectionist Architecture: The What and Where Vision Tasks.
Proceedings of the Machine Learning: From Theory to Applications, 1993

Robust Reinforcement Learning in Motion Planning.
Proceedings of the Advances in Neural Information Processing Systems 6, 1993

Convergence of Indirect Adaptive Asynchronous Value Iteration Algorithms.
Proceedings of the Advances in Neural Information Processing Systems 6, 1993

Monte Carlo Matrix Inversion and Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 6, 1993

1992
Learning reactive admittance control.
Proceedings of the 1992 IEEE International Conference on Robotics and Automation, 1992

1991
Task Decomposition Through Competition in a Modular Connectionist Architecture: The What and Where Vision Tasks.
Cogn. Sci., 1991

Linear systems analysis of the relationship between firing of deep cerebellar neurons and the classically conditioned nictitating membrane response in rabbits.
Biol. Cybern., 1991

A Cortico-Cerebellar Model that Learns to Generate Distributed Motor Commands to Control a Kinematic Arm.
Proceedings of the Advances in Neural Information Processing Systems 4, 1991

1990
Cerebellar control of endpoint position-a simulation model.
Proceedings of the IJCNN 1990, 1990

Explaining Temporal Differences to Create Useful Concepts for Evaluating States.
Proceedings of the 8th National Conference on Artificial Intelligence. Boston, Massachusetts, USA, July 29, 1990

1989
Sequential Decision Probelms and Neural Networks.
Proceedings of the Advances in Neural Information Processing Systems 2, 1989

1985
Pattern-recognizing stochastic learning automata.
IEEE Trans. Syst. Man Cybern., 1985

Training and Tracking in Robotics.
Proceedings of the 9th International Joint Conference on Artificial Intelligence. Los Angeles, 1985

1983
Neuronlike adaptive elements that can solve difficult learning control problems.
IEEE Trans. Syst. Man Cybern., 1983

1978
A Note on Pattern Reproduction in Tessellation Structures.
J. Comput. Syst. Sci., 1978


  Loading...