Fangcheng Fu

Orcid: 0000-0003-1658-0380

According to our database1, Fangcheng Fu authored at least 57 papers between 2018 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
Efficient Mixed-Precision Large Language Model Inference with TurboMind.
CoRR, August, 2025

PQCache: Product Quantization-based KVCache for Long Context LLM Inference.
Proc. ACM Manag. Data, June, 2025

Malleus: Straggler-Resilient Hybrid Parallel Training of Large-scale Models via Malleable Data and Model Parallelization.
Proc. ACM Manag. Data, June, 2025

How Significant Are the Real Performance Gains? An Unbiased Evaluation Framework for GraphRAG.
CoRR, June, 2025

Cascadia: A Cascade Serving System for Large Language Models.
CoRR, June, 2025

SALE : Low-bit Estimation for Efficient Sparse Attention in Long-context LLM Prefilling.
CoRR, May, 2025

Thinking Short and Right Over Thinking Long: Serving LLM Reasoning Efficiently and Accurately.
CoRR, May, 2025

LobRA: Multi-tenant Fine-tuning over Heterogeneous Data.
Proc. VLDB Endow., April, 2025

Galvatron: An Automatic Distributed System for Efficient Foundation Model Training.
CoRR, April, 2025

Hetu v2: A General and Scalable Deep Learning System with Hierarchical and Heterogeneous Single Program Multiple Data Annotations.
CoRR, April, 2025

Detecting and Analyzing Motifs in Large-Scale Online Transaction Networks.
IEEE Trans. Knowl. Data Eng., February, 2025

MEMO: Fine-grained Tensor Management For Ultra-long Context LLM Training.
Proc. ACM Manag. Data, February, 2025

ByteScale: Efficient Scaling of LLM Training with a 2048K Context Length on More Than 12,000 GPUs.
CoRR, February, 2025

Training-free and Adaptive Sparse Attention for Efficient Long Video Generation.
CoRR, February, 2025

ThunderServe: High-performance and Cost-efficient LLM Serving in Cloud Environments.
CoRR, February, 2025

Demystifying Cost-Efficiency in LLM Serving over Heterogeneous GPUs.
CoRR, February, 2025

NetMoE: Accelerating MoE Training through Dynamic Sample Placement.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

Hounding Data Diversity: Towards Participant Selection in Vertical Federated Learning.
Proceedings of the 41st IEEE International Conference on Data Engineering, 2025

Towards Scalable and Efficient Graph Structure Learning.
Proceedings of the 41st IEEE International Conference on Data Engineering, 2025

Spindle: Efficient Distributed Training of Multi-Task Large Models via Wavefront Scheduling.
Proceedings of the 30th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, 2025

FlexSP: Accelerating Large Language Model Training via Flexible Sequence Parallelism.
Proceedings of the 30th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, 2025

Enhancing Unsupervised Sentence Embeddings via Knowledge-Driven Data Augmentation and Gaussian-Decayed Contrastive Learning.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

HaCore: Efficient Coreset Construction with Locality Sensitive Hashing for Vertical Federated Learning.
Proceedings of the AAAI-25, Sponsored by the Association for the Advancement of Artificial Intelligence, February 25, 2025

2024
ProjPert: Projection-Based Perturbation for Label Protection in Split Learning Based Vertical Federated Learning.
IEEE Trans. Knowl. Data Eng., July, 2024

Improving Automatic Parallel Training via Balanced Memory Workload Optimization.
IEEE Trans. Knowl. Data Eng., 2024

Demystifying Workload Imbalances in Large Transformer Model Training over Variable-length Sequences.
CoRR, 2024

Data-Centric and Heterogeneity-Adaptive Sequence Parallelism for Efficient LLM Training.
CoRR, 2024

Gradual Learning: Optimizing Fine-Tuning with Partially Mastered Knowledge in Large Language Models.
CoRR, 2024

Retrofitting Temporal Graph Neural Networks with Transformer.
CoRR, 2024

Efficient Multi-Task Large Model Training via Data Heterogeneity-aware Model Management.
CoRR, 2024

Efficiently Training 7B LLM with 1 Million Sequence Length on 8 GPUs.
CoRR, 2024

Retrieval-Augmented Generation for AI-Generated Content: A Survey.
CoRR, 2024

Enabling Parallelism Hot Switching for Efficient Training of Large Language Models.
Proceedings of the ACM SIGOPS 30th Symposium on Operating Systems Principles, 2024

Efficient Multi-task LLM Quantization and Serving for Multiple LoRA Adapters.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

LSH-MoE: Communication-efficient MoE Training via Locality-Sensitive Hashing.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

X-former Elucidator: Reviving Efficient Attention for Long Context Language Modeling.
Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence, 2024

Generative and Contrastive Paradigms Are Complementary for Graph Self-Supervised Learning.
Proceedings of the 40th IEEE International Conference on Data Engineering, 2024

Accelerating Text-to-Image Editing via Cache-Enabled Sparse Diffusion Inference.
Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence, 2024

2023
P<sup>2</sup>CG: a privacy preserving collaborative graph neural network training framework.
VLDB J., July, 2023

Angel-PTM: A Scalable and Economical Large-scale Pre-training System in Tencent.
Proc. VLDB Endow., 2023

Generative and Contrastive Paradigms Are Complementary for Graph Self-Supervised Learning.
CoRR, 2023

Improving Automatic Parallel Training via Balanced Memory Workload Optimization.
CoRR, 2023

FISEdit: Accelerating Text-to-image Editing via Cache-enabled Sparse Diffusion Inference.
CoRR, 2023

OSDP: Optimal Sharded Data Parallel for Distributed Deep Learning.
Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, 2023

KVSAgg: Secure Aggregation of Distributed Key-Value Sets.
Proceedings of the 39th IEEE International Conference on Data Engineering, 2023

2022
Towards Communication-efficient Vertical Federated Learning Training via Cache-enabled Local Update.
Proc. VLDB Endow., 2022

Towards Communication-efficient Vertical Federated Learning Training via Cache-enabled Local Updates.
CoRR, 2022

BlindFL: Vertical Federated Machine Learning without Peeking into Your Data.
Proceedings of the SIGMOD '22: International Conference on Management of Data, Philadelphia, PA, USA, June 12, 2022

K-core decomposition on super large graphs with limited resources.
Proceedings of the SAC '22: The 37th ACM/SIGAPP Symposium on Applied Computing, Virtual Event, April 25, 2022

VF-PS: How to Select Important Participants in Vertical Federated Learning, Efficiently and Securely?
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Analyzing Online Transaction Networks with Network Motifs.
Proceedings of the KDD '22: The 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, August 14, 2022

2021
VF<sup>2</sup>Boost: Very Fast Vertical Federated Gradient Boosting for Cross-Enterprise Learning.
Proceedings of the SIGMOD '21: International Conference on Management of Data, 2021

2020
SKCompress: compressing sparse and nonuniform gradient in distributed machine learning.
VLDB J., 2020

Don't Waste Your Bits! Squeeze Activations and Gradients for Deep Neural Networks via TinyScript.
Proceedings of the 37th International Conference on Machine Learning, 2020

2019
An Experimental Evaluation of Large Scale GBDT Systems.
Proc. VLDB Endow., 2019

2018
SketchML: Accelerating Distributed Machine Learning with Data Sketches.
Proceedings of the 2018 International Conference on Management of Data, 2018

DimBoost: Boosting Gradient Boosting Decision Tree to Higher Dimensions.
Proceedings of the 2018 International Conference on Management of Data, 2018


  Loading...