Ofir Press

According to our database1, Ofir Press authored at least 13 papers between 2017 and 2023.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2023
SWE-bench: Can Language Models Resolve Real-World GitHub Issues?
CoRR, 2023

How Language Model Hallucinations Can Snowball.
CoRR, 2023

Measuring and Narrowing the Compositionality Gap in Language Models.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, 2023

2022
What Language Model to Train if You Have One Million GPU Hours?
CoRR, 2022

Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation.
Proceedings of the Tenth International Conference on Learning Representations, 2022

What Language Model to Train if You Have One Million GPU Hours?
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, 2022

Transformer Language Models without Positional Encodings Still Learn Positional Information.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, 2022

2021
Shortformer: Better Language Modeling using Shorter Inputs.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021

2020
Improving Transformer Models by Reordering their Sublayers.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020

2019
Partially Shuffling the Training Data to Improve Language Models.
CoRR, 2019

2018
You May Not Need Attention.
CoRR, 2018

2017
Language Generation with Recurrent Generative Adversarial Networks without Pre-training.
CoRR, 2017

Using the Output Embedding to Improve Language Models.
Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, 2017


  Loading...