Zhilin Yang

Orcid: 0009-0008-0681-9603

Affiliations:
  • Tsinghua University, Institute for Interdisciplinary Information Sciences, Beijing, China
  • Shanghai Qi Zhi Institute, China
  • Carnegie Mellon University, School of Computer Science, Pittsburgh, PA, USA (PhD 2020)


According to our database1, Zhilin Yang authored at least 51 papers between 2013 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
GPT understands, too.
AI Open, 2024

2023
CodeGeeX: A Pre-Trained Model for Code Generation with Multilingual Evaluations on HumanEval-X.
CoRR, 2023

CodeGeeX: A Pre-Trained Model for Code Generation with Multilingual Benchmarking on HumanEval-X.
Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2023

Not All Tasks Are Born Equal: Understanding Zero-Shot Generalization.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Compositional Task Representations for Large Language Models.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

A Universal Discriminator for Zero-Shot Generalization.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

Prompt-Based Metric Learning for Few-Shot NER.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, 2023

2022
Zero-Label Prompt Selection.
CoRR, 2022

NLP From Scratch Without Large-Scale Pretraining: A Simple and Efficient Framework.
Proceedings of the International Conference on Machine Learning, 2022

Learning to Detect Noisy Labels Using Model-Based Features.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, 2022

FlipDA: Effective and Robust Data Augmentation for Few-Shot Learning.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022

FewNLU: Benchmarking State-of-the-Art Methods for Few-Shot Natural Language Understanding.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022

P-Tuning: Prompt Tuning Can Be Comparable to Fine-tuning Across Scales and Tasks.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2022

GLM: General Language Model Pretraining with Autoregressive Blank Infilling.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022

2021
P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks.
CoRR, 2021

FewNLU: Benchmarking State-of-the-Art Methods for Few-Shot Natural Language Understanding.
CoRR, 2021

FlipDA: Effective and Robust Data Augmentation for Few-Shot Learning.
CoRR, 2021

FastMoE: A Fast Mixture-of-Expert Training System.
CoRR, 2021

All NLP Tasks Are Generation Tasks: A General Pretraining Framework.
CoRR, 2021

WuDaoCorpora: A super large-scale Chinese corpora for pre-training language models.
AI Open, 2021

Controllable Generation from Pre-trained Language Models via Inverse Prompting.
Proceedings of the KDD '21: The 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2021

The International Workshop on Pretraining: Algorithms, Architectures, and Applications ([email protected] 2021).
Proceedings of the KDD '21: The 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2021

2019
Mixtape: Breaking the Softmax Bottleneck Efficiently.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

XLNet: Generalized Autoregressive Pretraining for Language Understanding.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Transformer-XL: Attentive Language Models beyond a Fixed-Length Context.
Proceedings of the 57th Conference of the Association for Computational Linguistics, 2019

2018
GLoMo: Unsupervisedly Learned Relational Graphs as Transferable Representations.
CoRR, 2018

GLoMo: Unsupervised Learning of Transferable Relational Graphs.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Neural Models for Reasoning over Multiple Mentions Using Coreference.
Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2018

Mastering the Dungeon: Grounded Language Learning by Mechanical Turker Descent.
Proceedings of the 6th International Conference on Learning Representations, 2018

Breaking the Softmax Bottleneck: A High-Rank RNN Language Model.
Proceedings of the 6th International Conference on Learning Representations, 2018

HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering.
Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31, 2018

Neural Cross-lingual Named Entity Recognition with Minimal Resources.
Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31, 2018

2017
Differentiable Learning of Logical Rules for Knowledge Base Completion.
CoRR, 2017

A Probabilistic Framework for Location Inference from Social Media.
CoRR, 2017

Linguistic Knowledge as Memory for Recurrent Neural Networks.
CoRR, 2017

Differentiable Learning of Logical Rules for Knowledge Base Reasoning.
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017

Good Semi-supervised Learning That Requires a Bad GAN.
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017

Transfer Learning for Sequence Tagging with Hierarchical Recurrent Networks.
Proceedings of the 5th International Conference on Learning Representations, 2017

Words or Characters? Fine-grained Gating for Reading Comprehension.
Proceedings of the 5th International Conference on Learning Representations, 2017

Semi-Supervised QA with Generative Domain-Adaptive Nets.
Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 2017

Gated-Attention Readers for Text Comprehension.
Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 2017

2016
Encode, Review, and Decode: Reviewer Module for Caption Generation.
CoRR, 2016

Multi-Task Cross-Lingual Sequence Tagging from Scratch.
CoRR, 2016

Review Networks for Caption Generation.
Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, 2016

Multi-Modal Bayesian Embeddings for Learning Social Knowledge Graphs.
Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, 2016

Revisiting Semi-Supervised Learning with Graph Embeddings.
Proceedings of the 33nd International Conference on Machine Learning, 2016

2015
Multi-Source Bayesian Embeddings for Learning Social Knowledge Graphs.
CoRR, 2015

COSNET: Connecting Heterogeneous Social Networks with Local and Global Consistency.
Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2015

2014
Active learning for networked data based on non-progressive diffusion model.
Proceedings of the Seventh ACM International Conference on Web Search and Data Mining, 2014

Active Learning for Streaming Networked Data.
Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, 2014

2013
SAE: social analytic engine for large networks.
Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2013


  Loading...