Qinghao Hu

Orcid: 0000-0003-1034-7502

Affiliations:
  • MIT, Cambridge, MA, USA


According to our database1, Qinghao Hu authored at least 18 papers between 2021 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
Sailor: Automating Distributed Training over Dynamic, Heterogeneous, and Geo-distributed Clusters.
CoRR, April, 2025

DeltaZip: Efficient Serving of Multiple Full-Model-Tuned LLMs.
Proceedings of the Twentieth European Conference on Computer Systems, 2025

2024
Deep Learning Workload Scheduling in GPU Datacenters: A Survey.
ACM Comput. Surv., June, 2024

Efficient Training of Large Language Models on Distributed Infrastructures: A Survey.
CoRR, 2024

LoongTrain: Efficient Training of Long-Sequence LLMs with Head-Context Parallelism.
CoRR, 2024

InternEvo: Efficient Long-sequence Large Language Model Training via Hybrid Parallelism and Redundant Sharding.
CoRR, 2024

FedDSE: Distribution-aware Sub-model Extraction for Federated Learning over Resource-constrained Devices.
Proceedings of the ACM on Web Conference 2024, 2024

TorchGT: A Holistic System for Large-Scale Graph Transformer Training.
Proceedings of the International Conference for High Performance Computing, 2024

Characterization of Large Language Model Development in the Datacenter.
Proceedings of the 21st USENIX Symposium on Networked Systems Design and Implementation, 2024

Lins: Reducing Communication Overhead of ZeRO for Efficient LLM Training.
Proceedings of the 32nd IEEE/ACM International Symposium on Quality of Service, 2024

Sylvie: 3D-Adaptive and Universal System for Large-Scale Graph Neural Network Training.
Proceedings of the 40th IEEE International Conference on Data Engineering, 2024

2023
AMSP: Super-Scaling LLM Training via Advanced Model States Partitioning.
CoRR, 2023

Boosting Distributed Full-graph GNN Training with Asynchronous One-bit Communication.
CoRR, 2023

Hydro: Surrogate-Based Hyperparameter Tuning Service in Datacenters.
Proceedings of the 17th USENIX Symposium on Operating Systems Design and Implementation, 2023

Lucid: A Non-intrusive, Scalable and Interpretable Scheduler for Deep Learning Training Jobs.
Proceedings of the 28th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, 2023

2022
Deep Learning Workload Scheduling in GPU Datacenters: Taxonomy, Challenges and Vision.
CoRR, 2022

Primo: Practical Learning-Augmented Systems with Interpretable Models.
Proceedings of the 2022 USENIX Annual Technical Conference, 2022

2021
Characterization and prediction of deep learning workloads in large-scale GPU datacenters.
Proceedings of the International Conference for High Performance Computing, 2021


  Loading...