Jinhao Li

Orcid: 0009-0009-4286-6359

Affiliations:
  • Shanghai Jiao Tong University, Qing Yuan Research Institute, China


According to our database1, Jinhao Li authored at least 13 papers between 2024 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
BalanceGS: Algorithm-System Co-design for Efficient 3D Gaussian Splatting Training on GPU.
CoRR, October, 2025

Enabling Efficient Sparse Multiplications on GPUs With Heuristic Adaptability.
IEEE Trans. Comput. Aided Des. Integr. Circuits Syst., June, 2025

SpecEE: Accelerating Large Language Model Inference with Speculative Early Exiting.
Proceedings of the 52nd Annual International Symposium on Computer Architecture, 2025

TB-STC: Transposable Block-wise N: M Structured Sparse Tensor Core.
Proceedings of the IEEE International Symposium on High Performance Computer Architecture, 2025

FlightVGM: Efficient Video Generation Model Inference with Online Sparsification and Hybrid Precision on FPGAs.
Proceedings of the 2025 ACM/SIGDA International Symposium on Field Programmable Gate Arrays, 2025

DyLGNN: Efficient LM-GNN Fine-Tuning with Dynamic Node Partitioning, Low-Degree Sparsity, and Asynchronous Sub-Batch.
Proceedings of the Design, Automation & Test in Europe Conference, 2025

SoftmAP: Software-Hardware Co-Design for Integer-Only Softmax on Associative Processors.
Proceedings of the Design, Automation & Test in Europe Conference, 2025

Harnessing Conventional Video Processing Insights for Emerging 3D Video Generation Models: A Comprehensive Attention-aware Way.
Proceedings of the 62nd ACM/IEEE Design Automation Conference, 2025

Accelerator for LLM-Enhanced GNN with Product Quantization and Unified Indexing.
Proceedings of the 30th Asia and South Pacific Design Automation Conference, 2025

LLSM: LLM-enhanced Logic Synthesis Model with EDA-guided CoT Prompting, Hybrid Embedding and AIG-tailored Acceleration.
Proceedings of the 30th Asia and South Pacific Design Automation Conference, 2025

2024
Large Language Model Inference Acceleration: A Comprehensive Hardware Perspective.
CoRR, 2024

Fast and Efficient 2-bit LLM Inference on GPU: 2/4/16-bit in a Weight Matrix with Asynchronous Dequantization.
Proceedings of the 43rd IEEE/ACM International Conference on Computer-Aided Design, 2024

MARCA: Mamba Accelerator with Reconfigurable Architecture.
Proceedings of the 43rd IEEE/ACM International Conference on Computer-Aided Design, 2024


  Loading...