Wei Huang

Orcid: 0009-0007-9885-0028

Affiliations:
  • University of Hong Kong, Department of Electrical and Electronic Engineering, Hong Kong


According to our database1, Wei Huang authored at least 19 papers between 2023 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
OmniVinci: Enhancing Architecture and Data for Omni-Modal Understanding LLM.
CoRR, October, 2025

QeRL: Beyond Efficiency - Quantization-enhanced Reinforcement Learning for LLMs.
CoRR, October, 2025

MC#: Mixture Compressor for Mixture-of-Experts Large Models.
CoRR, October, 2025

LongLive: Real-time Interactive Long Video Generation.
CoRR, September, 2025

Scaling RL to Long Videos.
CoRR, July, 2025

DBellQuant: Breaking the Bell with Double-Bell Transformation for LLMs Post Training Binarization.
CoRR, July, 2025

MindOmni: Unleashing Reasoning Generation in Vision Language Models with RGPO.
CoRR, May, 2025

SliM-LLM: Salience-Driven Mixed-Precision Quantization for Large Language Models.
Proceedings of the Forty-second International Conference on Machine Learning, 2025

Data Pruning by Information Maximization.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

Mixture Compressor for Mixture-of-Experts LLMs Gains More.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

VideoEspresso: A Large-Scale Chain-of-Thought Dataset for Fine-Grained Video Reasoning via Core Frame Selection.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2025

2024
An empirical study of LLaMA3 quantization: from LLMs to MLLMs.
Vis. Intell., 2024

MC-MoE: Mixture Compressor for Mixture-of-Experts LLMs Gains More.
CoRR, 2024

SliM-LLM: Salience-Driven Mixed-Precision Quantization for Large Language Models.
CoRR, 2024

How Good Are Low-bit Quantized LLaMA3 Models? An Empirical Study.
CoRR, 2024

BiLLM: Pushing the Limit of Post-Training Quantization for LLMs.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

SNNGX: Securing Spiking Neural Networks with Genetic XOR Encryption on RRAM-based Neuromorphic Accelerator.
Proceedings of the 43rd IEEE/ACM International Conference on Computer-Aided Design, 2024

2023
OHQ: On-chip Hardware-aware Quantization.
CoRR, 2023

An innovative experimental teaching method of hardware-software co-design-Taking a hardware accelerator of neural network using FPGA.
Proceedings of the IEEE Frontiers in Education Conference, 2023


  Loading...