Jiaqi Yang

Orcid: 0000-0002-6148-8738

Affiliations:
  • University of California, Berkeley, CA, USA
  • Tsinghua University, Beijing, China (former)


According to our database1, Jiaqi Yang authored at least 10 papers between 2020 and 2022.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2022
Nearly Minimax Algorithms for Linear Bandits with Shared Representation.
CoRR, 2022

2021
Variance-Aware Confidence Set: Variance-Dependent Bound for Linear Bandits and Horizon-Free Bound for Linear Mixture MDP.
CoRR, 2021

Linear bandits with limited adaptivity and learning distributional optimal design.
Proceedings of the STOC '21: 53rd Annual ACM SIGACT Symposium on Theory of Computing, 2021

Improved Variance-Aware Confidence Sets for Linear Bandits and Linear Mixture MDP.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Optimal Gradient-based Algorithms for Non-concave Bandit Optimization.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Going Beyond Linear RL: Sample Efficient Neural Function Approximation.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Provable Model-based Nonlinear Bandit and Reinforcement Learning: Shelve Optimism, Embrace Virtual Curvature.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Impact of Representation Learning in Linear Bandits.
Proceedings of the 9th International Conference on Learning Representations, 2021

Fully Gap-Dependent Bounds for Multinomial Logit Bandit.
Proceedings of the 24th International Conference on Artificial Intelligence and Statistics, 2021

2020
Provable Benefits of Representation Learning in Linear Bandits.
CoRR, 2020


  Loading...