Zhuomin He

Orcid: 0000-0002-3477-0607

According to our database1, Zhuomin He authored at least 5 papers between 2024 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
Online Context Caching for Distributed Large Language Models Serving.
Proceedings of the IEEE INFOCOM 2025, 2025

AdaSkip: Adaptive Sublayer Skipping for Accelerating Long-Context LLM Inference.
Proceedings of the AAAI-25, Sponsored by the Association for the Advancement of Artificial Intelligence, February 25, 2025

2024
AttentionStore: Cost-effective Attention Reuse across Multi-turn Conversations in Large Language Model Serving.
CoRR, 2024

Cost-Efficient Large Language Model Serving for Multi-turn Conversations with CachedAttention.
Proceedings of the 2024 USENIX Annual Technical Conference, 2024

IMI: In-memory Multi-job Inference Acceleration for Large Language Models.
Proceedings of the 53rd International Conference on Parallel Processing, 2024


  Loading...