Xiaodong Liu

Orcid: 0000-0003-2050-8576

Affiliations:
  • Microsoft Research, Redmond, WA, USA
  • Nara Institute of Science and Technology, Takayama, Japan (PhD 2015)
  • Beijing University of Posts and Telecommunications, Center for Intelligence Science and Technology, Beijing, China (former)
  • University of Tokushima, Department of Information Sciences and Intelligent Systems, Tokushima, Japan (former)


According to our database1, Xiaodong Liu authored at least 80 papers between 2010 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Language Models as Inductive Reasoners.
Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics, 2024

2023
Fine-tuning large neural language models for biomedical natural language processing.
Patterns, April, 2023

Augmenting Language Models with Long-Term Memory.
CoRR, 2023

Chain-of-Skills: A Configurable Model for Open-domain Question Answering.
CoRR, 2023

Pre-training Transformers for Knowledge Graph Completion.
CoRR, 2023

Visually-Augmented Language Modeling.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Pre-training Multi-task Contrastive Learning Models for Scientific Literature Understanding.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, 2023

Task-Aware Specialization for Efficient and Robust Dense Retrieval for Open-Domain Question Answering.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2023

2022
Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing.
ACM Trans. Comput. Heal., 2022

Efficient Long Sequence Modeling via State Space Augmented Transformer.
CoRR, 2022

AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers.
CoRR, 2022

AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models.
CoRR, 2022

METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals.
CoRR, 2022

Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer.
CoRR, 2022

A Survey of Knowledge-Intensive NLP with Pre-Trained Language Models.
CoRR, 2022

AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models.
CoRR, 2022

LiST: Lite Prompted Self-training Makes Parameter-efficient Few-shot Learners.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2022, 2022

Human Parity on CommonsenseQA: Augmenting Self-Attention with External Attention.
Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, 2022

Taming Sparsely Activated Transformer with Stochastic Experts.
Proceedings of the Tenth International Conference on Learning Representations, 2022

No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models.
Proceedings of the Tenth International Conference on Learning Representations, 2022

AdaMix: Mixture-of-Adaptations for Parameter-efficient Model Tuning.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

Open-domain Question Answering via Chain of Reasoning over Heterogeneous Knowledge.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, 2022

Knowledge-Rich Self-Supervision for Biomedical Entity Linking.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, 2022

Open Domain Question Answering with A Unified Knowledge Interface.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022

2021
Knowledge-Rich Self-Supervised Entity Linking.
CoRR, 2021

CLUES: Few-Shot Learning Evaluation in Natural Language Understanding.
CoRR, 2021

Open Domain Question Answering over Virtual Documents: A Unified Approach for Data and Text.
CoRR, 2021

LiST: Lite Self-training Makes Efficient Few-shot Learners.
CoRR, 2021

Adversarial Training as Stackelberg Game: An Unrolled Optimization Approach.
CoRR, 2021

Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Few-Shot Learning Evaluation in Natural Language Understanding.
Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks 1, 2021

Targeted Adversarial Training for Natural Language Understanding.
Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021

Posterior Differential Regularization with f-divergence for Improving Model Robustness.
Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021

Deberta: decoding-Enhanced Bert with Disentangled Attention.
Proceedings of the 9th International Conference on Learning Representations, 2021

Adversarial Regularization as Stackelberg Game: An Unrolled Optimization Approach.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021

ARCH: Efficient Adversarial Regularized Training with Caching.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2021, 2021

Token-wise Curriculum Learning for Neural Machine Translation.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2021, 2021

HittER: Hierarchical Transformers for Knowledge Graph Embeddings.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021

Reader-Guided Passage Reranking for Open-Domain Question Answering.
Proceedings of the Findings of the Association for Computational Linguistics: ACL/IJCNLP 2021, 2021

Generation-Augmented Retrieval for Open-Domain Question Answering.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021

Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021

UnitedQA: A Hybrid Approach for Open Domain Question Answering.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021

2020
Very Deep Transformers for Neural Machine Translation.
CoRR, 2020

Adversarial Training for Large Neural Language Models.
CoRR, 2020

Adversarial Training for Commonsense Inference.
Proceedings of the 5th Workshop on Representation Learning for NLP, 2020


UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training.
Proceedings of the 37th International Conference on Machine Learning, 2020

On the Variance of the Adaptive Learning Rate and Beyond.
Proceedings of the 8th International Conference on Learning Representations, 2020

Understanding the Difficulty of Training Transformers.
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020

A Tale of Two Linkings: Dynamically Gating between Schema Linking and Structural Linking for Text-to-SQL Parsing.
Proceedings of the 28th International Conference on Computational Linguistics, 2020

RAT-SQL: Relation-Aware Schema Encoding and Linking for Text-to-SQL Parsers.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020

The Microsoft Toolkit of Multi-Task Deep Neural Networks for Natural Language Understanding.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, 2020

SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020

2019
A Hybrid Neural Network Model for Commonsense Reasoning.
CoRR, 2019

Improving Multi-Task Deep Neural Networks via Knowledge Distillation for Natural Language Understanding.
CoRR, 2019

Unified Language Model Pre-training for Natural Language Understanding and Generation.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Multi-task Learning with Sample Re-weighting for Machine Reading Comprehension.
Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019

Unsupervised Deep Structured Semantic Models for Commonsense Reasoning.
Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019

Cyclical Annealing Schedule: A Simple Approach to Mitigating KL Vanishing.
Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019

Adversarial Domain Adaptation for Machine Reading Comprehension.
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019

A Hybrid Retrieval-Generation Neural Conversation Model.
Proceedings of the 28th ACM International Conference on Information and Knowledge Management, 2019

DoubleTransfer at MEDIQA 2019: Multi-Source Transfer Learning for Natural Language Understanding in the Medical Domain.
Proceedings of the 18th BioNLP Workshop and Shared Task, 2019

Conversing by Reading: Contentful Neural Conversation with On-demand Machine Reading.
Proceedings of the 57th Conference of the Association for Computational Linguistics, 2019

Multi-Task Deep Neural Networks for Natural Language Understanding.
Proceedings of the 57th Conference of the Association for Computational Linguistics, 2019

2018
ReCoRD: Bridging the Gap between Human and Machine Commonsense Reading Comprehension.
CoRR, 2018

Stochastic Answer Networks for SQuAD 2.0.
CoRR, 2018

Multi-Task Learning for Machine Reading Comprehension.
CoRR, 2018

Stochastic Answer Networks for Natural Language Inference.
CoRR, 2018

Navigating with Graph Representations for Fast and Scalable Decoding of Neural Language Models.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Language-Based Image Editing With Recurrent Attentive Models.
Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, 2018

Stochastic Answer Networks for Machine Reading Comprehension.
Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, 2018

2017
Towards Human-level Machine Reading Comprehension: Reasoning and Inference with Multiple Strategies.
CoRR, 2017

An Empirical Analysis of Multiple-Turn Reasoning Strategies in Reading Comprehension Tasks.
Proceedings of the Eighth International Joint Conference on Natural Language Processing, 2017

Lexical Simplification with the Deep Structured Similarity Model.
Proceedings of the Eighth International Joint Conference on Natural Language Processing, 2017

2015
Multilingual Topic Models for Bilingual Dictionary Extraction.
ACM Trans. Asian Low Resour. Lang. Inf. Process., 2015

A Hybrid Ranking Approach to Chinese Spelling Check.
ACM Trans. Asian Low Resour. Lang. Inf. Process., 2015

Representation Learning Using Multi-Task Deep Neural Networks for Semantic Classification and Information Retrieval.
Proceedings of the NAACL HLT 2015, The 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Denver, Colorado, USA, May 31, 2015

2013
Topic Models + Word Alignment = A Flexible Framework for Extracting Bilingual Dictionary from Comparable Corpus.
Proceedings of the Seventeenth Conference on Computational Natural Language Learning, 2013

A Hybrid Chinese Spelling Correction Using Language Model and Statistical Machine Translation with Reranking.
Proceedings of the Seventh SIGHAN Workshop on Chinese Language Processing, 2013

2010
Use relative weight to improve the kNN for unbalanced text category.
Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering, 2010


  Loading...