Xing Hu

Orcid: 0009-0003-4510-898X

Affiliations:
  • Houmo AI, China
  • Hohai University, College of Computer and Information, Nanjing, China (former)


According to our database1, Xing Hu authored at least 13 papers between 2022 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
MoEQuant: Enhancing Quantization for Mixture-of-Experts Large Language Models via Expert-Balanced Sampling and Affinity Guidance.
CoRR, May, 2025

RWKVQuant: Quantizing the RWKV Family with Proxy Guided Hybrid of Scalar and Vector Quantization.
CoRR, May, 2025

Q-PETR: Quant-aware Position Embedding Transformation for Multi-View 3D Object Detection.
CoRR, February, 2025

MQuant: Unleashing the Inference Potential of Multimodal Large Language Models via Full Static Quantization.
CoRR, February, 2025

OstQuant: Refining Large Language Model Quantization with Orthogonal and Scaling Transformations for Better Distribution Fitting.
CoRR, January, 2025

AIM: Software and Hardware Co-design for Architecture-level IR-drop Mitigation in High-performance PIM.
Proceedings of the 52nd Annual International Symposium on Computer Architecture, 2025

MambaQuant: Quantizing the Mamba Family with Variance Aligned Rotation Methods.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

OSTQuant: Refining Large Language Model Quantization with Orthogonal and Scaling Transformations for Better Distribution Fitting.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

PillarHist: A Quantization-aware Pillar Feature Encoder based on Height-aware Histogram.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2025

2024
Post-training quantization for re-parameterization via coarse & fine weight splitting.
J. Syst. Archit., February, 2024

I-LLM: Efficient Integer-Only Inference for Fully-Quantized Low-Bit Large Language Models.
CoRR, 2024

34.3 A 22nm 64kb Lightning-Like Hybrid Computing-in-Memory Macro with a Compressed Adder Tree and Analog-Storage Quantizers for Transformer and CNNs.
Proceedings of the IEEE International Solid-State Circuits Conference, 2024

2022
3D Object Detection Based on Multi-scale Feature Fusion and Contrastive Learning.
Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, 2022


  Loading...