Lu Hou

Orcid: 0000-0002-4694-1821

Affiliations:
  • Huawei Technologies, Noah's Ark Lab, Shenzhen, China
  • Hong Kong University of Science and Technology, Hong Kong (PhD 2019)


According to our database1, Lu Hou authored at least 55 papers between 2016 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
Pre-Trained Models for Search and Recommendation: Introduction to the Special Issue - Part 2.
ACM Trans. Inf. Syst., September, 2025

Think Before You Talk: Enhancing Meaningful Dialogue Generation in Full-Duplex Speech Language Models with Planning-Inspired Text Guidance.
CoRR, August, 2025

The Synergy Dilemma of Long-CoT SFT and RL: Investigating Post-Training Techniques for Reasoning VLMs.
CoRR, July, 2025

A Simple Linear Patch Revives Layer-Pruned Large Language Models.
CoRR, May, 2025

Quantization Hurts Reasoning? An Empirical Study on Quantized Reasoning Models.
CoRR, April, 2025

ILLUME+: Illuminating Unified MLLM with Dual Visual Tokenization and Diffusion Refinement.
CoRR, April, 2025

Pre-Trained Models for Search and Recommendation: Introduction to the Special Issue - Part 1.
ACM Trans. Inf. Syst., March, 2025

HiRes-LLaVA: Restoring Fragmentation Input in High-Resolution Large Vision-Language Models.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2025

EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2025

2024
Power Law in Deep Neural Networks: Sparse Network Generation and Continual Learning With Preferential Attachment.
IEEE Trans. Neural Networks Learn. Syst., July, 2024

Embedding Compression in Recommender Systems: A Survey.
ACM Comput. Surv., May, 2024

ILLUME: Illuminating Your LLMs to See, Draw, and Self-Enhance.
CoRR, 2024

FlatQuant: Flatness Matters for LLM Quantization.
CoRR, 2024

EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions.
CoRR, 2024

DeCo: Decoupling Token Compression from Semantic Abstraction in Multimodal Large Language Models.
CoRR, 2024

Towards Multimodal Video Paragraph Captioning Models Robust to Missing Modality.
CoRR, 2024

Visually Guided Generative Text-Layout Pre-training for Document Intelligence.
CoRR, 2024

UNIT: Unifying Image and Text Recognition in One Vision Encoder.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

Visually Guided Generative Text-Layout Pre-training for Document Intelligence.
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024

Plug-and-Play: An Efficient Post-training Pruning Method for Large Language Models.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

VITATECS: A Diagnostic Dataset for Temporal Concept Understanding of Video-Language Models.
Proceedings of the Computer Vision - ECCV 2024, 2024

TimeChat: A Time-sensitive Multimodal Large Language Model for Long Video Understanding.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024

MoPE-CLIP: Structured Pruning for Efficient Vision-Language Models with Module-Wise Pruning Error Metric.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024

TempCompass: Do Video LLMs Really Understand Videos?
Proceedings of the Findings of the Association for Computational Linguistics, 2024

IntactKV: Improving Large Language Model Quantization by Keeping Pivot Tokens Intact.
Proceedings of the Findings of the Association for Computational Linguistics, 2024

2023
CTRL: Connect Tabular and Language Model for CTR Prediction.
CoRR, 2023

FETV: A Benchmark for Fine-Grained Evaluation of Open-Domain Text-to-Video Generation.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

TESTA: Temporal-Spatial Token Aggregation for Long-form Video-Language Understanding.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, 2023

Structured Pruning for Efficient Generative Pre-trained Language Models.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, 2023

mCLIP: Multilingual CLIP via Cross-lingual Transfer.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

Wukong-Reader: Multi-modal Pre-training for Fine-grained Visual Document Understanding.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

Adaptive Low-Precision Training for Embeddings in Click-Through Rate Prediction.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023

2022
Wukong-Reader: Multi-modal Pre-training for Fine-grained Visual Document Understanding.
CoRR, 2022

Wukong: 100 Million Large-scale Chinese Cross-modal Pre-training Dataset and A Foundation Framework.
CoRR, 2022

Wukong: A 100 Million Large-scale Chinese Cross-modal Pre-training Benchmark.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Towards Efficient Post-training Quantization of Pre-trained Language Models.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

FILIP: Fine-grained Interactive Language-Image Pre-Training.
Proceedings of the Tenth International Conference on Learning Representations, 2022

LiteVL: Efficient Video-Language Learning with Enhanced Spatial-Temporal Modeling.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

Compression of Generative Pre-trained Language Models via Quantization.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022

Enabling Multimodal Generation on CLIP via Vision-Language Knowledge Distillation.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2022, 2022

2021
Improved OOD Generalization via Adversarial Training and Pre-training.
CoRR, 2021

Improved OOD Generalization via Adversarial Training and Pretraing.
Proceedings of the 38th International Conference on Machine Learning, 2021

Reweighting Augmented Samples by Minimizing the Maximal Expected Loss.
Proceedings of the 9th International Conference on Learning Representations, 2021

GhostBERT: Generate More Features with Cheap Operations for BERT.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021

BinaryBERT: Pushing the Limit of BERT Quantization.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021

2020
BinaryBERT: Pushing the Limit of BERT Quantization.
CoRR, 2020

DynaBERT: Dynamic BERT with Adaptive Width and Depth.
CoRR, 2020

DynaBERT: Dynamic BERT with Adaptive Width and Depth.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

TernaryBERT: Distillation-aware Ultra-low Bit BERT.
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020

2019
Normalization Helps Training of Quantized LSTM.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Analysis of Quantized Models.
Proceedings of the 7th International Conference on Learning Representations, 2019

2018
Power Law in Sparsified Deep Neural Networks.
CoRR, 2018

Loss-aware Weight Quantization of Deep Networks.
Proceedings of the 6th International Conference on Learning Representations, 2018

2017
Loss-aware Binarization of Deep Networks.
Proceedings of the 5th International Conference on Learning Representations, 2017

2016
Efficient Learning of Timeseries Shapelets.
Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, 2016


  Loading...