Emilie Kaufmann

According to our database1, Emilie Kaufmann authored at least 57 papers between 2012 and 2023.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2023
Towards Instance-Optimality in Online PAC Reinforcement Learning.
CoRR, 2023

Bandit Pareto Set Identification: the Fixed Budget Setting.
CoRR, 2023

Adaptive Algorithms for Relaxed Pareto Set Identification.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

An ε-Best-Arm Identification Algorithm for Fixed-Confidence and Beyond.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Active Coverage for PAC Reinforcement Learning.
Proceedings of the Thirty Sixth Annual Conference on Learning Theory, 2023

Optimistic PAC Reinforcement Learning: the Instance-Dependent View.
Proceedings of the International Conference on Algorithmic Learning Theory, 2023

Dealing with Unknown Variances in Best-Arm Identification.
Proceedings of the International Conference on Algorithmic Learning Theory, 2023

2022
Efficient Change-Point Detection for Tackling Piecewise-Stationary Bandits.
J. Mach. Learn. Res., 2022

Near Instance-Optimal PAC Reinforcement Learning for Deterministic MDPs.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Near-Optimal Collaborative Learning in Bandits.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Top Two Algorithms Revisited.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Efficient Algorithms for Extreme Bandits.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2022

2021
Mixture Martingales Revisited with Applications to Sequential Tests and Confidence Intervals.
J. Mach. Learn. Res., 2021

On Multi-Armed Bandit Designs for Dose-Finding Trials.
J. Mach. Learn. Res., 2021

Fast active learning for pure exploration in reinforcement learning.
Proceedings of the 38th International Conference on Machine Learning, 2021

Kernel-Based Reinforcement Learning: A Finite-Time Analysis.
Proceedings of the 38th International Conference on Machine Learning, 2021

Optimal Thompson Sampling strategies for support-aware CVaR bandits.
Proceedings of the 38th International Conference on Machine Learning, 2021

Adaptive Reward-Free Exploration.
Proceedings of the Algorithmic Learning Theory, 2021

Episodic Reinforcement Learning in Finite MDPs: Minimax Lower Bounds Revisited.
Proceedings of the Algorithmic Learning Theory, 2021

Top-m identification for linear bandits.
Proceedings of the 24th International Conference on Artificial Intelligence and Statistics, 2021

A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces.
Proceedings of the 24th International Conference on Artificial Intelligence and Statistics, 2021

2020
Thompson Sampling for CVaR Bandits.
CoRR, 2020

Regret Bounds for Kernel-Based Reinforcement Learning.
CoRR, 2020

Planning in Markov Decision Processes with Gap-Dependent Sample Complexity.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Sub-sampling for Efficient Non-Parametric Bandit Exploration.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Solving Bernoulli Rank-One Bandits with Unimodal Thompson Sampling.
Proceedings of the Algorithmic Learning Theory, 2020

Fixed-confidence guarantees for Bayesian best-arm identification.
Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, 2020

A Practical Algorithm for Multiplayer Bandits when Arm Means Vary Among Players.
Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, 2020

2019
Asymptotically optimal algorithms for budgeted multiple play bandits.
Mach. Learn., 2019

On Multi-Armed Bandit Designs for Phase I Clinical Trials.
CoRR, 2019

The Generalized Likelihood Ratio Test meets klUCB: an Improved Algorithm for Piece-Wise Non-Stationary Bandits.
CoRR, 2019

New Algorithms for Multiplayer Bandits when Arm Means Vary Among Players.
CoRR, 2019

General parallel optimization a without metric.
Proceedings of the Algorithmic Learning Theory, 2019

2018
A spectral algorithm with additive clustering for the recovery of overlapping communities in networks.
Theor. Comput. Sci., 2018

What Doubling Tricks Can and Can't Do for Multi-Armed Bandits.
CoRR, 2018

Aggregation of multi-armed bandits learning algorithms for opportunistic spectrum access.
Proceedings of the 2018 IEEE Wireless Communications and Networking Conference, 2018

Sequential Test for the Lowest Mean: From Thompson to Murphy Sampling.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Corrupt Bandits for Preserving Local Privacy.
Proceedings of the Algorithmic Learning Theory, 2018

{Multi-Player Bandits Revisited}.
Proceedings of the Algorithmic Learning Theory, 2018

Pure Exploration in Infinitely-Armed Bandit Models with Fixed-Confidence.
Proceedings of the Algorithmic Learning Theory, 2018

2017
Multi-Player Bandits Models Revisited.
CoRR, 2017

Corrupt Bandits for Privacy Preserving Input.
CoRR, 2017

Learning the distribution with largest mean: two bandit frameworks.
CoRR, 2017

Monte-Carlo Tree Search by Best Arm Identification.
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017

Multi-Armed Bandit Learning in IoT Networks: Learning Helps Even in Non-stationary Settings.
Proceedings of the Cognitive Radio Oriented Wireless Networks, 2017

2016
On the Complexity of Best-Arm Identification in Multi-Armed Bandit Models.
J. Mach. Learn. Res., 2016

Asymptotically Optimal Algorithms for Multiple Play Bandits with Partial Feedback.
CoRR, 2016

On Explore-Then-Commit strategies.
Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, 2016

Maximin Action Identification: A New Bandit Framework for Games.
Proceedings of the 29th Conference on Learning Theory, 2016

Optimal Best Arm Identification with Fixed Confidence.
Proceedings of the 29th Conference on Learning Theory, 2016

2014
Analysis of bayesian and frequentist strategies for sequential resource allocation. (Analyse de stratégies bayésiennes et fréquentistes pour l'allocation séquentielle de ressources).
PhD thesis, 2014

On the Complexity of A/B Testing.
Proceedings of The 27th Conference on Learning Theory, 2014

2013
Thompson Sampling for 1-Dimensional Exponential Family Bandits.
Proceedings of the Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meeting held December 5-8, 2013

Information Complexity in Bandit Subset Selection.
Proceedings of the COLT 2013, 2013

2012
On Bayesian Upper Confidence Bounds for Bandit Problems.
Proceedings of the Fifteenth International Conference on Artificial Intelligence and Statistics, 2012

Thompson Sampling: An Optimal Finite Time Analysis
CoRR, 2012

Thompson Sampling: An Asymptotically Optimal Finite-Time Analysis.
Proceedings of the Algorithmic Learning Theory - 23rd International Conference, 2012


  Loading...