Iz Beltagy

According to our database1, Iz Beltagy authored at least 52 papers between 2018 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
OLMo: Accelerating the Science of Language Models.
CoRR, 2024

Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research.
CoRR, 2024

TESS: Text-to-Text Self-Conditioned Simplex Diffusion.
Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics, 2024

2023
Paloma: A Benchmark for Evaluating Language Model Fit.
CoRR, 2023

Catwalk: A Unified Language Model Evaluation Framework for Many Datasets.
CoRR, 2023

Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2.
CoRR, 2023

Efficiency Pentathlon: A Standardized Arena for Efficiency Evaluation.
CoRR, 2023

How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources.
CoRR, 2023

Large Language Model Distillation Doesn't Need a Teacher.
CoRR, 2023

TESS: Text-to-Text Self-Conditioned Simplex Diffusion.
CoRR, 2023

The Semantic Scholar Open Data Platform.
CoRR, 2023

How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

FiD-ICL: A Fusion-in-Decoder Approach for Efficient In-Context Learning.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

Z-ICL: Zero-Shot In-Context Learning with Pseudo-Demonstrations.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

2022
BLOOM: A 176B-Parameter Open-Access Multilingual Language Model.
CoRR, 2022

What Language Model to Train if You Have One Million GPU Hours?
CoRR, 2022

Transparency Helps Reveal When Language Models Learn Meaning.
CoRR, 2022

What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization?
CoRR, 2022

Don't Say What You Don't Know: Improving the Consistency of Abstractive Summarization by Constraining Beam Search.
CoRR, 2022

Infrastructure for Rapid Open Knowledge Network Development.
AI Mag., 2022

MultiVerS: Improving scientific claim verification with weak supervision and full-document context.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2022, 2022

Few-Shot Self-Rationalization with Natural Language Prompts.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2022, 2022

What Language Model Architecture and Pretraining Objective Works Best for Zero-Shot Generalization?
Proceedings of the International Conference on Machine Learning, 2022

Staged Training for Transformer Language Models.
Proceedings of the International Conference on Machine Learning, 2022

Continued Pretraining for Better Zero- and Few-Shot Promptability.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

SciFact-Open: Towards open-domain scientific claim verification.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, 2022

What Language Model to Train if You Have One Million GPU Hours?
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, 2022

PRIMERA: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022

Zero- and Few-Shot NLP with Pretrained Language Models.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, 2022

2021
LongChecker: Improving scientific claim verification by modeling full-abstract context.
CoRR, 2021

PRIMER: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization.
CoRR, 2021

MS2: Multi-Document Summarization of Medical Studies.
CoRR, 2021

Cross-Document Language Modeling.
CoRR, 2021

FLEX: Unifying Evaluation for Few-Shot NLP.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

A Dataset of Information-Seeking Questions and Answers Anchored in Research Papers.
Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021

MS\^2: Multi-Document Summarization of Medical Studies.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021

CDLM: Cross-Document Language Modeling.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2021, 2021

Scientific Language Models for Biomedical Knowledge Base Completion: An Empirical Study.
Proceedings of the 3rd Conference on Automated Knowledge Base Construction, 2021

SciCo: Hierarchical Cross-Document Coreference for Scientific Concepts.
Proceedings of the 3rd Conference on Automated Knowledge Base Construction, 2021

2020
Longformer: The Long-Document Transformer.
CoRR, 2020

SciREX: A Challenge Dataset for Document-Level Information Extraction.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020

Don't Stop Pretraining: Adapt Language Models to Domains and Tasks.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020

SPECTER: Document-level Representation Learning using Citation-informed Transformers.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020

2019
SciBERT: Pretrained Contextualized Embeddings for Scientific Text.
CoRR, 2019

Combining Distant and Direct Supervision for Neural Relation Extraction.
Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019

Pretrained Language Models for Sequential Sentence Classification.
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019

SciBERT: A Pretrained Language Model for Scientific Text.
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019

ScispaCy: Fast and Robust Models for Biomedical Natural Language Processing.
Proceedings of the 18th BioNLP Workshop and Shared Task, 2019

2018
Improving Distant Supervision with Maxpooled Attention and Sentence-Level Supervision.
CoRR, 2018

Machine Learning for Reliable mmWave Systems: Blockage Prediction and Proactive Handoff.
CoRR, 2018

Construction of the Literature Graph in Semantic Scholar.
Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2018

Machine Learning for Reliable MMWAVE Systems: Blockage Prediction and Proactive Handoff.
Proceedings of the 2018 IEEE Global Conference on Signal and Information Processing, 2018


  Loading...