Siwei Wang

Orcid: 0000-0003-0764-5592

Affiliations:
  • Tsinghua University, Beijing, China


According to our database1, Siwei Wang authored at least 15 papers between 2018 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Provable Risk-Sensitive Distributional Reinforcement Learning with General Function Approximation.
CoRR, 2024

2023
Contextual Combinatorial Bandits with Probabilistically Triggered Arms.
CoRR, 2023

Contextual Combinatorial Bandits with Probabilistically Triggered Arms.
Proceedings of the International Conference on Machine Learning, 2023

Provably Efficient Risk-Sensitive Reinforcement Learning: Iterated CVaR and Worst Path.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

2022
The pure exploration problem with general reward functions depending on full distributions.
Mach. Learn., 2022

Risk-Sensitive Reinforcement Learning: Iterated CVaR and the Worst Path.
CoRR, 2022

Batch-Size Independent Regret Bounds for Combinatorial Semi-Bandits with Probabilistically Triggered Arms or Independent Arms.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

2021
Pure Exploration Bandit Problem with General Reward Functions Depending on Full Distributions.
CoRR, 2021

Continuous Mean-Covariance Bandits.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Adaptive Algorithms for Multi-armed Bandit with Composite and Anonymous Feedback.
Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, 2021

A One-Size-Fits-All Solution to Conservative Bandit Problems.
Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, 2021

2020
Restless-UCB, an Efficient and Low-complexity Algorithm for Online Restless Bandits.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Dueling Bandits: From Two-dueling to Multi-dueling.
Proceedings of the 19th International Conference on Autonomous Agents and Multiagent Systems, 2020

2018
Multi-armed Bandits with Compensation.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Thompson Sampling for Combinatorial Semi-Bandits.
Proceedings of the 35th International Conference on Machine Learning, 2018


  Loading...