Zhiyuan Zeng

Orcid: 0000-0002-3979-3322

Affiliations:
  • Fudan University, School of Computer Science, Shanghai, China
  • Tianjin University, School of New Media and Communication, China (former)


According to our database1, Zhiyuan Zeng authored at least 15 papers between 2021 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
Implicit Reward as the Bridge: A Unified View of SFT and DPO Connections.
CoRR, July, 2025

Dynamic and Generalizable Process Reward Modeling.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

Revisiting the Test-Time Scaling of o1-like Models: Do they Truly Possess Test-Time Scaling Capabilities?
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

2024
Scaling of Search and Learning: A Roadmap to Reproduce o1 from Reinforcement Learning Perspective.
CoRR, 2024

WanJuan-CC: A Safe and High-Quality Open-sourced English Webtext Dataset.
CoRR, 2024

Query of CC: Unearthing Large Scale Domain-Specific Knowledge from Public Corpora.
CoRR, 2024

Memorize Step by Step: Efficient Long-Context Prefilling with Incremental Memory and Decremental Chunk.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Turn Waste into Worth: Rectifying Top-k Router of MoE.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Explicit Memory Learning with Expectation Maximization.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Aggregation of Reasoning: A Hierarchical Framework for Enhancing Answer Selection in Large Language Models.
Proceedings of the 2024 Joint International Conference on Computational Linguistics, 2024

Reasoning in Flux: Enhancing Large Language Models Reasoning through Uncertainty-aware Adaptive Guidance.
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2024

2023
Unsupervised and Few-Shot Parsing from Pretrained Language Models (Extended Abstract).
Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, 2023

SCoMoE: Efficient Mixtures of Experts with Structured Communication.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

2022
Unsupervised and few-shot parsing from pretrained language models.
Artif. Intell., 2022

2021
An Empirical Study on Adversarial Attack on NMT: Languages and Positions Matter.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021


  Loading...