Haocheng Xi

According to our database1, Haocheng Xi authored at least 20 papers between 2023 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
DC-VideoGen: Efficient Video Generation with Deep Compression Video Autoencoder.
CoRR, September, 2025

DC-Gen: Post-Training Diffusion Acceleration with Deeply Compressed Latent Space.
CoRR, September, 2025

SLA: Beyond Sparsity in Diffusion Transformers via Fine-Tunable Sparse-Linear Attention.
CoRR, September, 2025

SciML Agents: Write the Solver, Not the Solution.
CoRR, September, 2025

Jet-Nemotron: Efficient Language Model with Post Neural Architecture Search.
CoRR, August, 2025

XQuant: Breaking the Memory Wall for LLM Inference with KV Cache Rematerialization.
CoRR, August, 2025

Radial Attention: O(n log n) Sparse Attention with Energy Decay for Long Video Generation.
CoRR, June, 2025

Sparse VideoGen2: Accelerate Video Generation with Sparse Attention via Semantic-Aware Permutation.
CoRR, May, 2025

Oscillation-Reduced MXFP4 Training for Vision Transformers.
CoRR, February, 2025

SpargeAttn: Accurate Sparse Attention Accelerating Any Model Inference.
CoRR, February, 2025

QuantSpec: Self-Speculative Decoding with Hierarchical Quantized KV Cache.
CoRR, February, 2025

Sparse VideoGen: Accelerating Video Diffusion Transformers with Spatial-Temporal Sparsity.
CoRR, February, 2025

COAT: Compressing Optimizer states and Activations for Memory-Efficient FP8 Training.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

NVILA: Efficient Frontier Visual Language Models.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2025

2024
NVILA: Efficient Frontier Visual Language Models.
CoRR, 2024

COAT: Compressing Optimizer states and Activation for Memory-Efficient FP8 Training.
CoRR, 2024

Jetfire: Efficient and Accurate Transformer Pretraining with INT8 Data Flow and Per-Block Quantization.
CoRR, 2024

T-Rex: Text-assisted Retrosynthesis Prediction.
CoRR, 2024

Jetfire: Efficient and Accurate Transformer Pretraining with INT8 Data Flow and Per-Block Quantization.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

2023
Training Transformers with 4-bit Integers.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023


  Loading...