Jian Sha

Orcid: 0009-0006-9004-5959

According to our database1, Jian Sha authored at least 10 papers between 2023 and 2025.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
Every FLOP Counts: Scaling a 300B Mixture-of-Experts LING LLM without Premium GPUs.
CoRR, March, 2025

mLoRA: Fine-Tuning LoRA Adapters via Highly-Efficient Pipeline Parallelism in Multiple GPUs.
Proc. VLDB Endow., February, 2025

XPUTimer: Anomaly Diagnostics for Divergent LLM Training in GPU Clusters of Thousand-Plus Scale.
CoRR, February, 2025

EDiT: A Local-SGD-Based Efficient Distributed Training Method for Large Language Models.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

MEPipe: Democratizing LLM Training with Memory-Efficient Slice-Level Pipeline Scheduling on Cost-Effective Accelerators.
Proceedings of the Twentieth European Conference on Computer Systems, 2025

2024
DLRover-RM: Resource Optimization for Deep Recommendation Models Training in the cloud.
Proc. VLDB Endow., August, 2024

Enhancing hydrological extremes prediction accuracy: Integrating diverse loss functions in Transformer models.
Environ. Model. Softw., 2024

Couler: Unified Machine Learning Workflow Optimization in Cloud.
CoRR, 2024

Couler: Unified Machine Learning Workflow Optimization in Cloud.
Proceedings of the 40th IEEE International Conference on Data Engineering, 2024

2023
ASPEN: High-Throughput LoRA Fine-Tuning of Large Language Models with a Single GPU.
CoRR, 2023


  Loading...