Cheems Wang

Orcid: 0009-0003-7758-725X

Affiliations:
  • Tsinghua University, Department of Automation, Beijing, China
  • National University of Defense Technology, College of Science, Changsha, China (former)
  • University of Amsterdam, Machine Learning Lab (AMLab), Amsterdam, The Netherlands (former, PhD 2022)


According to our database1, Cheems Wang authored at least 38 papers between 2017 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
Adaptive Neighborhood-Constrained Q Learning for Offline Reinforcement Learning.
CoRR, November, 2025

Group & reweight: a novel cost-sensitive approach to mitigating class imbalance in network traffic classification.
Int. J. Mach. Learn. Cybern., October, 2025

Utility-Diversity Aware Online Batch Selection for LLM Supervised Fine-tuning.
CoRR, October, 2025

Gains: Fine-grained Federated Domain Adaptation in Open Set.
CoRR, October, 2025

HOB: A Holistically Optimized Bidding Strategy under Heterogeneous Auction Mechanisms with Organic Traffic.
CoRR, October, 2025

A Unified Multi-Task Learning Framework for Generative Auto-Bidding with Validation-Aligned Optimization.
CoRR, October, 2025

Enhancing Generative Auto-bidding with Offline Reward Evaluation and Policy Search.
CoRR, September, 2025

Can Prompt Difficulty be Online Predicted for Accelerating RL Finetuning of Reasoning Models?
CoRR, July, 2025

Model Predictive Task Sampling for Efficient and Robust Adaptation.
CoRR, January, 2025

Robust Fast Adaptation from Adversarially Explicit Task Distribution Generation.
Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining, V.1, 2025

Fast and Robust: Task Sampling with Posterior and Diversity Synergies for Adaptive Decision-Makers in Randomized Environments.
Proceedings of the Forty-second International Conference on Machine Learning, 2025

DynaPrompt: Dynamic Test-Time Prompt Tuning.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

Latent Reward: LLM-Empowered Credit Assignment in Episodic Reinforcement Learning.
Proceedings of the AAAI-25, Sponsored by the Association for the Advancement of Artificial Intelligence, February 25, 2025

2024
Non-informative noise-enhanced stochastic neural networks for improving adversarial robustness.
Inf. Fusion, 2024

Theoretical Investigations and Practical Enhancements on Tail Task Risk Minimization in Meta Learning.
CoRR, 2024

Choices are More Important than Efforts: LLM Enables Efficient Multi-Agent Exploration.
CoRR, 2024

Group Distributionally Robust Optimization can Suppress Class Imbalance Effect in Network Traffic Classification.
CoRR, 2024

GO4Align: Group Optimization for Multi-Task Alignment.
CoRR, 2024

Offline Reinforcement Learning with OOD State Correction and OOD Action Suppression.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

Doubly Mild Generalization for Offline Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

Balanced Confidence Calibration for Graph Neural Networks.
Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2024

Reducing Fine-Tuning Memory Overhead by Approximate and Memory-Sharing Backpropagation.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

2023
Large-scale Generative Simulation Artificial Intelligence: the Next Hotspot in Generative AI.
CoRR, 2023

A Simple Yet Effective Strategy to Robustify the Meta Learning Paradigm.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Bridge the Inference Gaps of Neural Processes via Expectation Maximization.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

2022
Learning Expressive Meta-Representations with Mixture of Expert Neural Processes.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Model-based Meta Reinforcement Learning using Graph Structured Surrogate Models and Amortized Policy Search.
Proceedings of the International Conference on Machine Learning, 2022

2021
Model-based Meta Reinforcement Learning using Graph Structured Surrogate Models.
CoRR, 2021

2020
CSAN: A neural network benchmark model for crime forecasting in spatio-temporal scale.
Knowl. Based Syst., 2020

Urban Fire Situation Forecasting: Deep sequence learning with spatio-temporal dynamics.
Appl. Soft Comput., 2020

Doubly Stochastic Variational Inference for Neural Processes with Hierarchical Latent Variables.
Proceedings of the 37th International Conference on Machine Learning, 2020

2019
Crime-GAN: A Context-based Sequence Generative Network for Crime Forecasting with Adversarial Loss.
Proceedings of the 2019 IEEE International Conference on Big Data (IEEE BigData), 2019

2018
VMAV-C: A Deep Attention-based Reinforcement Learning Algorithm for Model-based Control.
CoRR, 2018

The Mathematical Modeling of the Two-Echelon Ground Vehicle and Its Mounted Unmanned Aerial Vehicle Cooperated Routing Problem.
Proceedings of the 2018 IEEE Intelligent Vehicles Symposium, 2018

Addressing the Task of Rocket Recycling with Deep Reinforcement Learning.
Proceedings of the 6th International Conference on Information Technology: IoT and Smart City, 2018

2017
A Novel Ensemble Method for Imbalanced Data Learning: Bagging of Extrapolation-SMOTE SVM.
Comput. Intell. Neurosci., 2017

PAIRS: Prediction of Activation/Inhibition Regulation Signaling Pathway.
Comput. Intell. Neurosci., 2017

Online Shopping Recommendation with Bayesian Probabilistic Matrix Factorization.
Proceedings of the Intelligence Science I - Second IFIP TC 12 International Conference, 2017


  Loading...