Mingqi Gao

Orcid: 0009-0002-6227-3402

Affiliations:
  • Peking University, Beijing, China


According to our database1, Mingqi Gao authored at least 19 papers between 2022 and 2025.

Collaborative distances:
  • Dijkstra number2 of five.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
Minos: A Multimodal Evaluation Model for Bidirectional Generation Between Image and Text.
CoRR, June, 2025

Exploring the Multilingual NLG Evaluation Abilities of LLM-Based Evaluators.
CoRR, March, 2025

Aspect-Guided Multi-Level Perturbation Analysis of Large Language Models in Automated Peer Review.
CoRR, February, 2025

Re-evaluating Automatic LLM System Ranking for Alignment with Human Preference.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2025, Albuquerque, New Mexico, USA, April 29, 2025

Analyzing and Evaluating Correlation Measures in NLG Meta-Evaluation.
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies, 2025

Towards A "Novel" Benchmark: Evaluating Literary Fiction with Large Language Models.
Proceedings of the Findings of the Association for Computational Linguistics, 2025

A Dual-Perspective NLG Meta-Evaluation Framework with Automatic Benchmark and Better Interpretability.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

2024
Themis: Towards Flexible and Interpretable NLG Evaluation.
CoRR, 2024

LLM-based NLG Evaluation: Current Status and Challenges.
CoRR, 2024

Themis: A Reference-free NLG Evaluation Language Model with Flexibility and Interpretability.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Is Summary Useful or Not? An Extrinsic Human Evaluation of Text Summaries on Downstream Tasks.
Proceedings of the 2024 Joint International Conference on Computational Linguistics, 2024

Are LLM-based Evaluators Confusing NLG Quality Criteria?
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2024

Better than Random: Reliable NLG Human Evaluation with Constrained Active Sampling.
Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence, 2024

2023
Summarization is (Almost) Dead.
CoRR, 2023

Human-like Summarization Evaluation with ChatGPT.
CoRR, 2023

Evaluating Factuality in Cross-lingual Summarization.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, 2023

Reference Matters: Benchmarking Factual Error Correction for Dialogue Summarization with Fine-grained Evaluation Framework.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

2022
Social Biases in Automatic Evaluation Metrics for NLG.
CoRR, 2022

DialSummEval: Revisiting Summarization Evaluation for Dialogues.
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2022


  Loading...