Hado van Hasselt

According to our database1, Hado van Hasselt authored at least 23 papers between 2008 and 2018.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Other 

Links

On csauthors.net:

Bibliography

2018
Meta-Gradient Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Learning to Coordinate with Coordination Graphs in Repeated Single-Stage Multi-Agent Decision Problems.
Proceedings of the 35th International Conference on Machine Learning, 2018

Rainbow: Combining Improvements in Deep Reinforcement Learning.
Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, 2018

2017
Natural Value Approximators: Learning when to Trust Past Estimates.
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017

Successor Features for Transfer in Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017

The Predictron: End-To-End Learning and Planning.
Proceedings of the 34th International Conference on Machine Learning, 2017

2016
Learning values across many orders of magnitude.
Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, 2016

Dueling Network Architectures for Deep Reinforcement Learning.
Proceedings of the 33nd International Conference on Machine Learning, 2016

Deep Reinforcement Learning with Double Q-Learning.
Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, 2016

2014
Off-policy TD( l) with a true online equivalence.
Proceedings of the Thirtieth Conference on Uncertainty in Artificial Intelligence, 2014

Weighted importance sampling for off-policy learning with linear function approximation.
Proceedings of the Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, 2014

A new Q(lambda) with interim forward view and Monte Carlo equivalence.
Proceedings of the 31th International Conference on Machine Learning, 2014

2013
Stacking under uncertainty: We know how to predict, but how should we act?
Proceedings of the IEEE Symposium on Computational Intelligence In Production And Logistics Systems, 2013

2011
Insights in reinforcement rearning : formal analysis and empirical evaluation of temporal-difference learning algorithms.
PhD thesis, 2011

Exploiting Best-Match Equations for Efficient Reinforcement Learning.
Journal of Machine Learning Research, 2011

Reinforcement learning algorithms for solving classification problems.
Proceedings of the 2011 IEEE Symposium on Adaptive Dynamic Programming And Reinforcement Learning, 2011

2010
Double Q-learning.
Proceedings of the Advances in Neural Information Processing Systems 23: 24th Annual Conference on Neural Information Processing Systems 2010. Proceedings of a meeting held 6-9 December 2010, 2010

2009
Using continuous action spaces to solve discrete problems.
Proceedings of the International Joint Conference on Neural Networks, 2009

Adaptive Serious Games Using Agent Organizations.
Proceedings of the Agents for Games and Simulations, 2009

The QV family compared to other reinforcement learning algorithms.
Proceedings of the IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, 2009

A theoretical and empirical analysis of Expected Sarsa.
Proceedings of the IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, 2009

2008
Ensemble Algorithms in Reinforcement Learning.
IEEE Trans. Systems, Man, and Cybernetics, Part B, 2008

On-line adapting games using agent organizations.
Proceedings of the 2008 IEEE Symposium on Computational Intelligence and Games, 2008


  Loading...