Denis Kuznedelev

Orcid: 0009-0005-2420-9620

According to our database1, Denis Kuznedelev authored at least 25 papers between 2022 and 2025.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
Hogwild! Inference: Parallel LLM Generation via Concurrent Attention.
CoRR, April, 2025

Scale-wise Distillation of Diffusion Models.
CoRR, March, 2025

Cache Me If You Must: Adaptive Key-Value Quantization for Large Language Models.
CoRR, January, 2025

TACO Vision Models Can Be Efficiently Specialized via Few-Shot Task-Aware Compression.
Trans. Mach. Learn. Res., 2025


2024
Accurate Neural Network Pruning Requires Rethinking Sparse Optimization.
Trans. Mach. Learn. Res., 2024

Label Privacy in Split Learning for Large Models with Parameter-Efficient Training.
CoRR, 2024

Switti: Designing Scale-Wise Transformers for Text-to-Image Synthesis.
CoRR, 2024

EvoPress: Towards Optimal Dynamic Model Compression via Evolutionary Search.
CoRR, 2024

Accurate Compression of Text-to-Image Diffusion Models via Vector Quantization.
CoRR, 2024

Does Diffusion Beat GAN in Image Super Resolution?
CoRR, 2024

YaART: Yet Another ART Rendering Technology.
CoRR, 2024

The Iterative Optimal Brain Surgeon: Faster Sparse Recovery by Leveraging Second-Order Information.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

PV-Tuning: Beyond Straight-Through Estimation for Extreme LLM Compression.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

Extreme Compression of Large Language Models via Additive Quantization.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

2023
Sparse Fine-tuning for Inference Acceleration of Large Language Models.
CoRR, 2023

Vision Models Can Be Efficiently Specialized via Few-Shot Task-Aware Compression.
CoRR, 2023

Characterizing Graph Datasets for Node Classification: Homophily-Heterophily Dichotomy and Beyond.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

CAP: Correlation-Aware Pruning for Highly-Accurate Sparse Vision Models.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Evaluating Robustness and Uncertainty of Graph Models Under Structural Distributional Shifts.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

A view of mini-batch SGD via generating functions: conditions of convergence, phase transitions, benefit from negative momenta.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

A critical look at the evaluation of GNNs under heterophily: Are we really making progress?
Proceedings of the Eleventh International Conference on Learning Representations, 2023

2022
oViT: An Accurate Second-Order Pruning Framework for Vision Transformers.
CoRR, 2022

Characterizing Graph Datasets for Node Classification: Beyond Homophily-Heterophily Dichotomy.
CoRR, 2022


  Loading...