Canwen Xu

Orcid: 0000-0002-1552-999X

According to our database1, Canwen Xu authored at least 37 papers between 2019 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
StarCoder 2 and The Stack v2: The Next Generation.
CoRR, 2024

2023
Contrastive Post-training Large Language Models on Data Curriculum.
CoRR, 2023

RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems.
CoRR, 2023

Small Models are Valuable Plug-ins for Large Language Models.
CoRR, 2023

Mirror: A Natural Language Interface for Data Querying, Summarization, and Visualization.
Proceedings of the Companion Proceedings of the ACM Web Conference 2023, 2023

LongCoder: A Long-Range Pre-trained Language Model for Code Completion.
Proceedings of the International Conference on Machine Learning, 2023

Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

Spoiler Detection as Semantic Text Matching.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

A Survey on Dynamic Neural Networks for Natural Language Processing.
Proceedings of the Findings of the Association for Computational Linguistics: EACL 2023, 2023

A Survey on Model Compression and Acceleration for Pretrained Language Models.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023

2022
BLOOM: A 176B-Parameter Open-Access Multilingual Language Model.
CoRR, 2022

A Survey on Model Compression for Natural Language Processing.
CoRR, 2022

PromptSource: An Integrated Development Environment and Repository for Natural Language Prompts.
CoRR, 2022

Automatic Multi-Label Prompting: Simple and Interpretable Few-Shot Classification.
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2022


Efficiently Tuned Parameters Are Task Embeddings.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

InforMask: Unsupervised Informative Masking for Language Model Pretraining.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

BERT Learns to Teach: Knowledge Distillation with Meta Learning.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022

LaPraDoR: Unsupervised Pretrained Dense Retriever for Zero-Shot Text Retrieval.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2022, 2022


Leashing the Inner Demons: Self-Detoxification for Language Models.
Proceedings of the Thirty-Sixth AAAI Conference on Artificial Intelligence, 2022

2021
Multitask Prompted Training Enables Zero-Shot Task Generalization.
CoRR, 2021

Meta Learning for Knowledge Distillation.
CoRR, 2021

Blow the Dog Whistle: A Chinese Dataset for Cant Understanding with Common Sense and World Knowledge.
Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021

Improving Sequence-to-Sequence Pre-training via Sequence Span Rewriting.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021

Beyond Preserved Accuracy: Evaluating Loyalty and Robustness of BERT Compression.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021


2020
BERT Loses Patience: Fast and Robust Inference with Early Exit.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

UnihanLM: Coarse-to-Fine Chinese-Japanese Language Model Pretraining with the Unihan Database.
Proceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, 2020

BERT-of-Theseus: Compressing BERT by Progressive Module Replacing.
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020

Transformers: State-of-the-Art Natural Language Processing.
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, 2020

MATINF: A Jointly Labeled Large-Scale Dataset for Classification, Question Answering and Summarization.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020

Pre-train and Plug-in: Flexible Conditional Text Generation with Variational Auto-Encoders.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020

2019
Pre-train and Plug-in: Flexible Conditional Text Generation with Variational Auto-Encoders.
CoRR, 2019

Obj-GloVe: Scene-Based Contextual Object Embedding.
CoRR, 2019

DLocRL: A Deep Learning Pipeline for Fine-Grained Location Recognition and Linking in Tweets.
Proceedings of the World Wide Web Conference, 2019

Exploiting Multiple Embeddings for Chinese Named Entity Recognition.
Proceedings of the 28th ACM International Conference on Information and Knowledge Management, 2019


  Loading...