Yu Bai

Affiliations:
  • Salesforce Research, Palo Alto, CA, USA
  • Stanford University, CA, USA (PhD 2019)


According to our database1, Yu Bai authored at least 45 papers between 2019 and 2023.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2023
Is Inverse Reinforcement Learning Harder than Standard Reinforcement Learning?
CoRR, 2023

How Do Transformers Learn In-Context Beyond Simple Functions? A Case Study on Learning with Representations.
CoRR, 2023

Transformers as Decision Makers: Provable In-Context Reinforcement Learning via Supervised Pretraining.
CoRR, 2023

Sample-Efficient Learning of POMDPs with Multiple Observations In Hindsight.
CoRR, 2023

Transformers as Statisticians: Provable In-Context Learning with In-Context Algorithm Selection.
CoRR, 2023

What can a Single Attention Layer Learn? A Study Through the Random Features Lens.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Efficient RL with Impaired Observability: Learning to Act with Delayed and Missing State Observations.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Lower Bounds for Learning in Revealing POMDPs.
Proceedings of the International Conference on Machine Learning, 2023

Improved Online Conformal Prediction via Strongly Adaptive Online Learning.
Proceedings of the International Conference on Machine Learning, 2023

The Role of Coverage in Online Reinforcement Learning.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Partially Observable RL with B-Stability: Unified Structural Condition and Sharp Sample-Efficient Algorithms.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Learning Rationalizable Equilibria in Multiplayer Games.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Breaking the Curse of Multiagency: Provably Efficient Decentralized Multi-Agent RL with Function Approximation.
Proceedings of the Thirty Sixth Annual Conference on Learning Theory, 2023

2022
Unified Algorithms for RL with Decision-Estimation Coefficients: No-Regret, PAC, and Reward-Free Learning.
CoRR, 2022

Efficient Φ-Regret Minimization in Extensive-Form Games via Online Mirror Descent.
CoRR, 2022

Local calibration: metrics and recalibration.
Proceedings of the Uncertainty in Artificial Intelligence, 2022

Policy Optimization for Markov Games: Unified Framework and Faster Convergence.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Sample-Efficient Learning of Correlated Equilibria in Extensive-Form Games.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Identifying good directions to escape the NTK regime and efficiently learn low-degree plus sparse polynomials.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Efficient Phi-Regret Minimization in Extensive-Form Games via Online Mirror Descent.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Near-Optimal Learning of Extensive-Form Games with Imperfect Information.
Proceedings of the International Conference on Machine Learning, 2022

When Can We Learn General-Sum Markov Games with a Large Number of Players Sample-Efficiently?
Proceedings of the Tenth International Conference on Learning Representations, 2022

Efficient and Differentiable Conformal Prediction with General Function Classes.
Proceedings of the Tenth International Conference on Learning Representations, 2022

Conformal Predictor for Improving Zero-Shot Text Classification Efficiency.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

2021
Localized Calibration: Metrics and Recalibration.
CoRR, 2021

Near-Optimal Offline Reinforcement Learning via Double Variance Reduction.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Policy Finetuning: Bridging Sample-Efficient Offline and Online Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Understanding the Under-Coverage Bias in Uncertainty Estimation.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Sample-Efficient Learning of Stackelberg Equilibria in General-Sum Games.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Exact Gap between Generalization Error and Uniform Convergence in Random Feature Models.
Proceedings of the 38th International Conference on Machine Learning, 2021

A Sharp Analysis of Model-based Reinforcement Learning with Self-Play.
Proceedings of the 38th International Conference on Machine Learning, 2021

Don't Just Blame Over-parametrization for Over-confidence: Theoretical Analysis of Calibration in Binary Classification.
Proceedings of the 38th International Conference on Machine Learning, 2021

How Important is the Train-Validation Split in Meta-Learning?
Proceedings of the 38th International Conference on Machine Learning, 2021

Near-Optimal Provable Uniform Convergence in Offline Policy Evaluation for Reinforcement Learning.
Proceedings of the 24th International Conference on Artificial Intelligence and Statistics, 2021

2020
Near Optimal Provable Uniform Convergence in Off-Policy Evaluation for Reinforcement Learning.
CoRR, 2020

Taylorized Training: Towards Better Approximation of Neural Network Training at Finite Width.
CoRR, 2020

Towards Understanding Hierarchical Learning: Benefits of Neural Representations.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Near-Optimal Reinforcement Learning with Self-Play.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Provable Self-Play Algorithms for Competitive Reinforcement Learning.
Proceedings of the 37th International Conference on Machine Learning, 2020

Beyond Linearization: On Quadratic and Higher-Order Approximation of Wide Neural Networks.
Proceedings of the 8th International Conference on Learning Representations, 2020

2019
Proximal algorithms for constrained composite optimization, with applications to solving low-rank SDPs.
CoRR, 2019

Provably Efficient Q-Learning with Low Switching Cost.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

ProxQuant: Quantized Neural Networks via Proximal Operators.
Proceedings of the 7th International Conference on Learning Representations, 2019

Approximability of Discriminators Implies Diversity in GANs.
Proceedings of the 7th International Conference on Learning Representations, 2019

Subgradient Descent Learns Orthogonal Dictionaries.
Proceedings of the 7th International Conference on Learning Representations, 2019


  Loading...