Yehui Tang

Orcid: 0000-0002-0322-4283

According to our database1, Yehui Tang authored at least 87 papers between 2019 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
LLM Data Selection and Utilization via Dynamic Bi-level Optimization.
CoRR, July, 2025

DualTHOR: A Dual-Arm Humanoid Simulation Platform for Contingency-Aware Planning.
CoRR, June, 2025

EAQuant: Enhancing Post-Training Quantization for MoE Models via Expert-Aware Optimization.
CoRR, June, 2025

SlimLLM: Accurate Structured Pruning for Large Language Models.
CoRR, May, 2025

Pangu Embedded: An Efficient Dual-system LLM Reasoner with Metacognition.
CoRR, May, 2025

Pangu Pro MoE: Mixture of Grouped Experts for Efficient Sparsity.
CoRR, May, 2025

Pangu Light: Weight Re-Initialization for Pruning and Accelerating LLMs.
CoRR, May, 2025

Pangu Ultra MoE: How to Train Your Big MoE on Ascend NPUs.
CoRR, May, 2025

Pangu Ultra: Pushing the Limits of Dense Large Language Models on Ascend NPUs.
CoRR, April, 2025

Saliency-driven Dynamic Token Pruning for Large Language Models.
CoRR, April, 2025

SpeCache: Speculative Key-Value Caching for Efficient Generation of LLMs.
CoRR, March, 2025

Mixture of Lookup Experts.
CoRR, March, 2025

Post-Training Quantization for Diffusion Transformer via Hierarchical Timestep Grouping.
CoRR, March, 2025

Rethinking Video Tokenization: A Conditioned Diffusion-based Approach.
CoRR, March, 2025

GenVidBench: A Challenging Benchmark for Detecting AI-Generated Video.
CoRR, January, 2025

GPT4Image: Large Pre-trained Models Help Vision Models Learn Better on Perception Task.
Proceedings of the Companion Proceedings of the ACM on Web Conference 2025, 2025

EMS-SD: Efficient Multi-sample Speculative Decoding for Accelerating Large Language Models.
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies, 2025

DenseSSM: State Space Models with Dense Hidden Connection for Efficient Large Language Models.
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies, 2025

QuaDiM: A Conditional Diffusion Model For Quantum State Property Estimation.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

CBQ: Cross-Block Quantization for Large Language Models.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

TinySAM: Pushing the Envelope for Efficient Segment Anything Model.
Proceedings of the AAAI-25, Sponsored by the Association for the Advancement of Artificial Intelligence, February 25, 2025

Eve: Efficient Multimodal Vision Language Models with Elastic Visual Experts.
Proceedings of the AAAI-25, Sponsored by the Association for the Advancement of Artificial Intelligence, February 25, 2025

2024
Forest-of-Thought: Scaling Test-Time Compute for Enhancing LLM Reasoning.
CoRR, 2024

Free Video-LLM: Prompt-guided Visual Perception for Efficient Training-free Video LLMs.
CoRR, 2024

Multi-Granularity Semantic Revision for Large Language Model Distillation.
CoRR, 2024

No Time to Waste: Squeeze Time into Channel for Mobile Video Understanding.
CoRR, 2024

Kangaroo: Lossless Self-Speculative Decoding via Double Early Exiting.
CoRR, 2024

GhostNetV3: Exploring the Training Strategies for Compact Models.
CoRR, 2024

DenseMamba: State Space Models with Dense Hidden Connection for Efficient Large Language Models.
CoRR, 2024

SAM-DiffSR: Structure-Modulated Diffusion Model for Image Super-Resolution.
CoRR, 2024

A Survey on Transformer Compression.
CoRR, 2024

Star-Agents: Automatic Data Optimization with LLM Agents for Instruction Tuning.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

Kangaroo: Lossless Self-Speculative Decoding for Accelerating LLMs via Double Early Exiting.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

MemoryFormer : Minimize Transformer Computation by Removing Fully-Connected Layers.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

SSL4Q: Semi-Supervised Learning of Quantum Data with Application to Quantum State Classification.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Rethinking Optimization and Architecture for Tiny Language Models.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

ExCP: Extreme LLM Checkpoint Compression via Weight-Momentum Joint Shrinking.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Memory-Space Visual Prompting for Efficient Vision-Language Fine-Tuning.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Data-efficient Large Vision Models through Sequential Autoregression.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

SLAB: Efficient Transformers with Simplified Linear Attention and Progressive Re-parameterized Batch Normalization.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Towards LLM4QPE: Unsupervised Pretraining of Quantum Property Estimation and A Benchmark.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

Node2ket: Efficient High-Dimensional Network Embedding in Quantum Hilbert Space.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

Adapt Without Forgetting: Distill Proximity from Dual Teachers in Vision-Language Models.
Proceedings of the Computer Vision - ECCV 2024, 2024

Visual Prompting via Partial Optimal Transport.
Proceedings of the Computer Vision - ECCV 2024, 2024

Context-Guided Spatial Feature Reconstruction for Efficient Semantic Segmentation.
Proceedings of the Computer Vision - ECCV 2024, 2024

Token Compensator: Altering Inference Cost of Vision Transformer Without Re-tuning.
Proceedings of the Computer Vision - ECCV 2024, 2024

Circuit Design and Efficient Simulation of Quantum Inner Product and Empirical Studies of Its Effect on Near-Term Hybrid Quantum-Classic Machine Learning.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024

2023
A Survey on Vision Transformer.
IEEE Trans. Pattern Anal. Mach. Intell., 2023

PanGu-π: Enhancing Language Model Architectures via Nonlinearity Compensation.
CoRR, 2023

LightCLIP: Learning Multi-Level Interaction for Lightweight Vision-Language Models.
CoRR, 2023

GPT4Image: Can Large Pre-trained Models Help Vision Models on Perception Tasks?
CoRR, 2023

One-for-All: Bridge the Gap Between Heterogeneous Architectures in Knowledge Distillation.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Masked Image Modeling with Local Multi-Scale Reconstruction.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023

Network Expansion For Practical Training Acceleration.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023

2022
Recent progress and perspectives on quantum computing for finance.
Serv. Oriented Comput. Appl., 2022

FastMIM: Expediting Masked Image Modeling Pre-training for Vision.
CoRR, 2022

From Quantum Graph Computing to Quantum Graph Learning: A Survey.
CoRR, 2022

PyramidTNT: Improved Transformer-in-Transformer Baselines with Pyramid Architecture.
CoRR, 2022

GraphQNTK: Quantum Neural Tangent Kernel for Graph Data.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

GhostNetV2: Enhance Cheap Operation with Long-Range Attention.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Learning Efficient Vision Transformers via Fine-Grained Manifold Distillation.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Bridge the Gap Between Architecture Spaces via A Cross-Domain Predictor.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Vision GNN: An Image is Worth Graph of Nodes.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Towards a Native Quantum Paradigm for Graph Representation Learning: A Sampling-based Recurrent Embedding Approach.
Proceedings of the KDD '22: The 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, August 14, 2022

Spatial-Channel Token Distillation for Vision MLPs.
Proceedings of the International Conference on Machine Learning, 2022

An Image Patch is a Wave: Phase-Aware Vision MLP.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

Patch Slimming for Efficient Vision Transformers.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

Hire-MLP: Vision MLP via Hierarchical Rearrangement.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

CMT: Convolutional Neural Networks Meet Vision Transformers.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

Source-Free Domain Adaptation via Distribution Estimation.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

2021
CMT: Convolutional Neural Networks Meet Vision Transformers.
CoRR, 2021

Efficient Vision Transformers via Fine-Grained Manifold Distillation.
CoRR, 2021

Visual Transformer Pruning.
CoRR, 2021

Learning Frequency Domain Approximation for Binary Neural Networks.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Augmented Shortcuts for Vision Transformers.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Homogeneous Architecture Augmentation for Neural Predictor.
Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, 2021

ReNAS: Relativistic Evaluation of Neural Architecture Search.
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2021

Manifold Regularized Dynamic Network Pruning.
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2021

2020
A Survey on Visual Transformer.
CoRR, 2020

SCOP: Scientific Control for Reliable Neural Network Pruning.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

A Semi-Supervised Assessor of Neural Architectures.
Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020

Neuromorphic Camera Guided High Dynamic Range Imaging.
Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020

Frequency Domain Compact 3D Convolutional Neural Networks.
Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020

Reborn Filters: Pruning Convolutional Neural Networks with Limited Data.
Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, 2020

Beyond Dropout: Feature Map Distortion to Regularize Deep Neural Networks.
Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, 2020

2019
RNAS: Architecture Ranking for Powerful Networks.
CoRR, 2019

Bringing Giant Neural Networks Down to Earth with Unlabeled Data.
CoRR, 2019


  Loading...