Xiaonan Nie
Orcid: 0000-0001-6766-757X
According to our database1,
Xiaonan Nie
authored at least 31 papers
between 2021 and 2025.
Collaborative distances:
Collaborative distances:
Timeline
Legend:
Book In proceedings Article PhD thesis Dataset OtherLinks
On csauthors.net:
Bibliography
2025
Proc. ACM Manag. Data, June, 2025
Malleus: Straggler-Resilient Hybrid Parallel Training of Large-scale Models via Malleable Data and Model Parallelization.
Proc. ACM Manag. Data, June, 2025
Efficient and scalable huge embedding model training via distributed cache management.
VLDB J., May, 2025
Proc. ACM Manag. Data, February, 2025
ByteScale: Efficient Scaling of LLM Training with a 2048K Context Length on More Than 12,000 GPUs.
CoRR, February, 2025
Proceedings of the Thirteenth International Conference on Learning Representations, 2025
Proceedings of the 41st IEEE International Conference on Data Engineering, 2025
2024
IEEE Trans. Knowl. Data Eng., 2024
FlashFlex: Accommodating Large Language Model Training over Heterogeneous Environment.
CoRR, 2024
DataSculpt: Crafting Data Landscapes for LLM Post-Training through Multi-objective Partitioning.
CoRR, 2024
BaichuanSEED: Sharing the Potential of ExtensivE Data Collection and Deduplication by Introducing a Competitive Large Language Model Baseline.
CoRR, 2024
CoRR, 2024
Proceedings of the ACM SIGOPS 30th Symposium on Operating Systems Principles, 2024
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024
2023
Sci. China Inf. Sci., January, 2023
Proc. VLDB Endow., 2023
FlexMoE: Scaling Large-scale Sparse Pre-trained Model Training via Dynamic Device Placement.
Proc. ACM Manag. Data, 2023
CoRR, 2023
Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, 2023
2022
Galvatron: Efficient Transformer Training over Multiple GPUs Using Automatic Parallelism.
Proc. VLDB Endow., 2022
CoRR, 2022
Proceedings of the SIGMOD '22: International Conference on Management of Data, Philadelphia, PA, USA, June 12, 2022
TSPLIT: Fine-grained GPU Memory Management for Efficient DNN Training via Tensor Splitting.
Proceedings of the 38th IEEE International Conference on Data Engineering, 2022
2021
HET: Scaling out Huge Embedding Model Training via Cache-enabled Distributed Framework.
Proc. VLDB Endow., 2021
Proceedings of the SIGMOD '21: International Conference on Management of Data, 2021