Tim Dettmers

According to our database1, Tim Dettmers authored at least 22 papers between 2016 and 2023.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2023
MatFormer: Nested Transformer for Elastic Inference.
CoRR, 2023

SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression.
CoRR, 2023

Towards A Unified View of Sparse Feed-Forward Network in Pretraining Large Language Model.
CoRR, 2023

Stable and low-precision training for large-scale vision-language models.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

QLoRA: Efficient Finetuning of Quantized LLMs.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Distributed Inference and Fine-tuning of Large Language Models Over The Internet.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient.
Proceedings of the International Conference on Machine Learning, 2023

The case for 4-bit precision: k-bit Inference Scaling Laws.
Proceedings of the International Conference on Machine Learning, 2023

Towards A Unified View of Sparse Feed-Forward Network in Pretraining Large Language Model.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

Petals: Collaborative Inference and Fine-tuning of Large Models.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics: System Demonstrations, 2023

2022
Petals: Collaborative Inference and Fine-tuning of Large Models.
CoRR, 2022

LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale.
CoRR, 2022

Branch-Train-Merge: Embarrassingly Parallel Training of Expert Language Models.
CoRR, 2022

GPT3.int8(): 8-bit Matrix Multiplication for Transformers at Scale.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

8-bit Optimizers via Block-wise Quantization.
Proceedings of the Tenth International Conference on Learning Representations, 2022

2021
Training Transformers Together.
Proceedings of the NeurIPS 2021 Competitions and Demonstrations Track, 2021

BASE Layers: Simplifying Training of Large, Sparse Models.
Proceedings of the 38th International Conference on Machine Learning, 2021

2020
High Performance Natural Language Processing.
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts, 2020

2019
Sparse Networks from Scratch: Faster Training without Losing Performance.
CoRR, 2019

2018
Jack the Reader - A Machine Reading Framework.
Proceedings of ACL 2018, Melbourne, Australia, July 15-20, 2018, System Demonstrations, 2018

Convolutional 2D Knowledge Graph Embeddings.
Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, 2018

2016
8-Bit Approximations for Parallelism in Deep Learning.
Proceedings of the 4th International Conference on Learning Representations, 2016


  Loading...