Jinqi Xiao

Orcid: 0009-0004-7311-9413

According to our database1, Jinqi Xiao authored at least 13 papers between 2023 and 2025.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
HeadInfer: Memory-Efficient LLM Inference by Head-wise Offloading.
CoRR, February, 2025

DisDet: Exploring Detectability of Backdoor Attack on Diffusion Models.
Trans. Mach. Learn. Res., 2025

TopV: Compatible Token Pruning with Inference Time Optimization for Fast and Low-Memory Multimodal Vision Language Model.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2025

COAP: Memory-Efficient Training with Correlation-Aware Gradient Projection.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2025

2024
COAP: Memory-Efficient Training with Correlation-Aware Gradient Projection.
CoRR, 2024

MoE-I<sup>2</sup>: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank Decomposition.
CoRR, 2024

ELRT: Efficient Low-Rank Training for Compact Convolutional Neural Networks.
CoRR, 2024

MoE-I²: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank Decomposition.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2024, 2024

Clean and Compact: Efficient Data-Free Backdoor Defense with Model Compactness.
Proceedings of the Computer Vision - ECCV 2024, 2024

2023
GraphMP: Graph Neural Network-based Motion Planning with Efficient Graph Search.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

ETTE: Efficient Tensor-Train-based Computing Engine for Deep Neural Networks.
Proceedings of the 50th Annual International Symposium on Computer Architecture, 2023

COMCAT: Towards Efficient Compression and Customization of Attention-Based Vision Models.
Proceedings of the International Conference on Machine Learning, 2023

HALOC: Hardware-Aware Automatic Low-Rank Compression for Compact Neural Networks.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023


  Loading...