Shizhe Diao

Orcid: 0000-0002-3325-9209

According to our database1, Shizhe Diao authored at least 34 papers between 2017 and 2024.

Collaborative distances:
  • Dijkstra number2 of five.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
LISA: Layerwise Importance Sampling for Memory-Efficient Large Language Model Fine-Tuning.
CoRR, 2024

Arithmetic Control of LLMs for Diverse User Preferences: Directional Preference Alignment with Multi-Objective Rewards.
CoRR, 2024

Can We Verify Step by Step for Incorrect Answer Detection?
CoRR, 2024

The Instinctive Bias: Spurious Images lead to Hallucination in MLLMs.
CoRR, 2024

ConstraintChecker: A Plugin for Large Language Models to Reason on Commonsense Knowledge Bases.
Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics, 2024

2023
Black-Box Prompt Learning for Pre-trained Language Models.
Trans. Mach. Learn. Res., 2023

R-Tuning: Teaching Large Language Models to Refuse Unknown Questions.
CoRR, 2023

Plum: Prompt Learning using Metaheuristic.
CoRR, 2023

MarineGPT: Unlocking Secrets of Ocean to the Public.
CoRR, 2023

UniTime: A Language-Empowered Unified Model for Cross-Domain Time Series Forecasting.
CoRR, 2023

Speciality vs Generality: An Empirical Study on Catastrophic Forgetting in Fine-tuning Foundation Models.
CoRR, 2023

LMFlow: An Extensible Toolkit for Finetuning and Inference of Large Foundation Models.
CoRR, 2023

On the Difference of BERT-style and CLIP-style Text Encoders.
CoRR, 2023

RAFT: Reward rAnked FineTuning for Generative Foundation Model Alignment.
CoRR, 2023

Active Prompting with Chain-of-Thought for Large Language Models.
CoRR, 2023

Hashtag-Guided Low-Resource Tweet Classification.
Proceedings of the ACM Web Conference 2023, 2023

Write and Paint: Generative Vision-Language Models are Unified Modal Learners.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Towards Unifying Medical Vision-and-Language Pre-training via Soft Prompts.
Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023

Automatic Prompt Augmentation and Selection with Chain-of-Thought from Labeled Data.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, 2023

DetGPT: Detect What You Need via Reasoning.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

Doolittle: Benchmarks and Corpora for Academic Writing Formalization.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

Mixture-of-Domain-Adapters: Decoupling and Injecting Domain Knowledge to Pre-trained Language Models' Memories.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

On the Difference of BERT-style and CLIP-style Text Encoders.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, 2023

2022
ExtremeBERT: A Toolkit for Accelerating Pretraining of Customized BERT.
CoRR, 2022

Normalizing Flow with Variational Latent Representation.
CoRR, 2022

Prefix Language Models are Unified Modal Learners.
CoRR, 2022

VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models.
CoRR, 2022

Black-box Prompt Learning for Pre-trained Language Models.
CoRR, 2022

VLUE: A Multi-Task Multi-Dimension Benchmark for Evaluating Vision-Language Pre-training.
Proceedings of the International Conference on Machine Learning, 2022

2021
Efficient Neural Network Training via Forward and Backward Propagation Sparsification.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Taming Pre-trained Language Models with N-gram Representations for Low-Resource Domain Adaptation.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021

TILGAN: Transformer-based Implicit Latent GAN for Diverse and Coherent Text Generation.
Proceedings of the Findings of the Association for Computational Linguistics: ACL/IJCNLP 2021, 2021

2020
ZEN: Pre-training Chinese Text Encoder Enhanced by N-gram Representations.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, 2020

2017
GubaLex: Guba-Oriented Sentiment Lexicon for Big Texts in Finance.
Proceedings of the 13th International Conference on Semantics, Knowledge and Grids, 2017


  Loading...