Armen Aghajanyan

According to our database1, Armen Aghajanyan authored at least 28 papers between 2015 and 2023.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2023
DOMINO: A Dual-System for Multi-step Visual Language Reasoning.
CoRR, 2023

Jointly Training Large Autoregressive Multimodal Models.
CoRR, 2023

Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning.
CoRR, 2023

MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

D4: Improving LLM Pretraining via Document De-Duplication and Diversification.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Retrieval-Augmented Multimodal Language Modeling.
Proceedings of the International Conference on Machine Learning, 2023

Scaling Laws for Generative Mixed-Modal Language Models.
Proceedings of the International Conference on Machine Learning, 2023

InCoder: A Generative Model for Code Infilling and Synthesis.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

2022
BARTSmiles: Generative Masked Language Models for Molecular Representations.
CoRR, 2022

CM3: A Causal Masked Multimodal Model of the Internet.
CoRR, 2022

Memorization Without Overfitting: Analyzing the Training Dynamics of Large Language Models.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

CCQA: A New Web-Scale Question Answering Dataset for Model Pre-Training.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2022, 2022

HTLM: Hyper-Text Pre-Training and Prompting of Language Models.
Proceedings of the Tenth International Conference on Learning Representations, 2022

Improving Passage Retrieval with Zero-Shot Question Generation.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

RetroNLU: Retrieval Augmented Task-Oriented Semantic Parsing.
Proceedings of the 4th Workshop on NLP for Conversational AI, 2022

2021
Non-Autoregressive Semantic Parsing for Compositional Task-Oriented Dialog.
Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021

Better Fine-Tuning by Reducing Representational Collapse.
Proceedings of the 9th International Conference on Learning Representations, 2021

VideoCLIP: Contrastive Pre-training for Zero-shot Video-Text Understanding.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021

Muppet: Massive Multi-task Representations with Pre-Finetuning.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021

Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tuning.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021

2020
Conversational Semantic Parsing.
CoRR, 2020

Pre-training via Paraphrasing.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Conversational Semantic Parsing.
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020

2019
Towards Language Agnostic Universal Representations.
Proceedings of the 57th Conference of the Association for Computational Linguistics, 2019

2017
Convolution Aware Initialization.
CoRR, 2017

Charged Point Normalization: An Efficient Solution to the Saddle Point Problem.
Proceedings of the 5th International Conference on Learning Representations, 2017

SoftTarget Regularization: An Effective Technique to Reduce Over-Fitting in Neural Networks.
Proceedings of the 3rd IEEE International Conference on Cybernetics, 2017

2015
Gravitational Clustering.
CoRR, 2015


  Loading...