Daehyun Ahn

Orcid: 0000-0001-8050-9963

According to our database1, Daehyun Ahn authored at least 17 papers between 2018 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
QUICK: Quantization-aware Interleaving and Conflict-free Kernel for efficient LLM inference.
CoRR, 2024

2023
V-LSTM: An Efficient LSTM Accelerator Using Fixed Nonzero-Ratio Viterbi-Based Pruning.
IEEE Trans. Comput. Aided Des. Integr. Circuits Syst., October, 2023

Squeezing Large-Scale Diffusion Models for Mobile.
CoRR, 2023

Searching for Robust Binary Neural Networks via Bimodal Parameter Perturbation.
Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2023

Temporal Dynamic Quantization for Diffusion Models.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Leveraging Early-Stage Robustness in Diffusion Models for Efficient and High-Quality Image Synthesis.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

2022
Workload-Balanced Graph Attention Network Accelerator with Top-K Aggregation Candidates.
Proceedings of the 41st IEEE/ACM International Conference on Computer-Aided Design, 2022

2021
Maximizing Parallel Activation of Word-Lines in MRAM-Based Binary Neural Network Accelerators.
IEEE Access, 2021

SPRITE: Sparsity-Aware Neural Processing Unit with Constant Probability of Index-Matching.
Proceedings of the Design, Automation & Test in Europe Conference & Exhibition, 2021

Energy-efficient charge sharing-based 8T2C SRAM in-memory accelerator for binary neural networks in 28nm CMOS.
Proceedings of the IEEE Asian Solid-State Circuits Conference, 2021

2020
Balancing Computation Loads and Optimizing Input Vector Loading in LSTM Accelerators.
IEEE Trans. Comput. Aided Des. Integr. Circuits Syst., 2020

OPTIMUS: OPTImized matrix MUltiplication Structure for Transformer neural network accelerator.
Proceedings of Machine Learning and Systems 2020, 2020

Time-step interleaved weight reuse for LSTM neural network computing.
Proceedings of the ISLPED '20: ACM/IEEE International Symposium on Low Power Electronics and Design, 2020

V-LSTM: An Efficient LSTM Accelerator Using Fixed Nonzero-Ratio Viterbi-Based Pruning.
Proceedings of the FPGA '20: The 2020 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays, 2020

2019
Double Viterbi: Weight Encoding for High Compression Ratio and Fast On-Chip Reconstruction for Deep Neural Network.
Proceedings of the 7th International Conference on Learning Representations, 2019

2018
Input-Splitting of Large Neural Networks for Power-Efficient Accelerator with Resistive Crossbar Memory Array.
Proceedings of the International Symposium on Low Power Electronics and Design, 2018

Viterbi-based Pruning for Sparse Matrix with Fixed and High Index Compression Ratio.
Proceedings of the 6th International Conference on Learning Representations, 2018


  Loading...