Ian Osband

According to our database1, Ian Osband authored at least 40 papers between 2013 and 2023.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2023
Ensembles for Uncertainty Estimation: Benefits of Prior Functions and Bootstrapping.
Trans. Mach. Learn. Res., 2023

Reinforcement Learning, Bit by Bit.
Found. Trends Mach. Learn., 2023

Approximate Thompson Sampling via Epistemic Neural Networks.
Proceedings of the Uncertainty in Artificial Intelligence, 2023

Epistemic Neural Networks.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

2022
Fine-Tuning Language Models via Epistemic Neural Networks.
CoRR, 2022

Robustness of Epinets against Distributional Shifts.
CoRR, 2022

Evaluating high-order predictive distributions in deep learning.
Proceedings of the Uncertainty in Artificial Intelligence, 2022

The Neural Testbed: Evaluating Joint Predictions.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

2021
Evaluating Predictive Distributions: Does Bayesian Deep Learning Work?
CoRR, 2021

Evaluating Probabilistic Inference in Deep Learning: Beyond Marginal Predictions.
CoRR, 2021

Epistemic Neural Networks.
CoRR, 2021

Matrix games with bandit feedback.
Proceedings of the Thirty-Seventh Conference on Uncertainty in Artificial Intelligence, 2021

2020
Stochastic matrix games with bandit feedback.
CoRR, 2020

Behaviour Suite for Reinforcement Learning.
Proceedings of the 8th International Conference on Learning Representations, 2020

Making Sense of Reinforcement Learning and Probabilistic Inference.
Proceedings of the 8th International Conference on Learning Representations, 2020

Hypermodels for Exploration.
Proceedings of the 8th International Conference on Learning Representations, 2020

2019
Deep Exploration via Randomized Value Functions.
J. Mach. Learn. Res., 2019

Meta-learning of Sequential Strategies.
CoRR, 2019

2018
A Tutorial on Thompson Sampling.
Found. Trends Mach. Learn., 2018

Randomized Prior Functions for Deep Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Scalable Coordinated Exploration in Concurrent Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

The Uncertainty Bellman Equation and Exploration.
Proceedings of the 35th International Conference on Machine Learning, 2018

Noisy Networks For Exploration.
Proceedings of the 6th International Conference on Learning Representations, 2018

Deep Q-learning From Demonstrations.
Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, 2018

2017
On Optimistic versus Randomized Exploration in Reinforcement Learning.
CoRR, 2017

Gaussian-Dirichlet Posterior Dominance in Sequential Learning.
CoRR, 2017

Learning from Demonstrations for Real World Reinforcement Learning.
CoRR, 2017

Noisy Networks for Exploration.
CoRR, 2017

A Tutorial on Thompson Sampling.
CoRR, 2017

Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Proceedings of the 34th International Conference on Machine Learning, 2017

Minimax Regret Bounds for Reinforcement Learning.
Proceedings of the 34th International Conference on Machine Learning, 2017

2016
On Lower Bounds for Regret in Reinforcement Learning.
CoRR, 2016

Posterior Sampling for Reinforcement Learning Without Episodes.
CoRR, 2016

Deep Exploration via Bootstrapped DQN.
Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, 2016

Generalization and Exploration via Randomized Value Functions.
Proceedings of the 33nd International Conference on Machine Learning, 2016

2015
Bootstrapped Thompson Sampling and Deep Exploration.
CoRR, 2015

2014
Near-optimal Regret Bounds for Reinforcement Learning in Factored MDPs.
CoRR, 2014

Model-based Reinforcement Learning and the Eluder Dimension.
Proceedings of the Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, 2014

Near-optimal Reinforcement Learning in Factored MDPs.
Proceedings of the Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, 2014

2013
(More) Efficient Reinforcement Learning via Posterior Sampling.
Proceedings of the Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meeting held December 5-8, 2013


  Loading...