Shihan Dou

Orcid: 0009-0002-6013-3035

According to our database1, Shihan Dou authored at least 67 papers between 2020 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
Intern-S1: A Scientific Multimodal Foundation Model.
CoRR, August, 2025

LLMEval-3: A Large-Scale Longitudinal Study on Robust and Fair Evaluation of Large Language Models.
CoRR, August, 2025

VRPO: Rethinking Value Modeling for Robust RL Training under Noisy Supervision.
CoRR, August, 2025

Mitigating Attention Hacking in Preference-Based Reward Modeling via Interaction Distillation.
CoRR, August, 2025

SpeechRole: A Large-Scale Dataset and Benchmark for Evaluating Speech Role-Playing Agents.
CoRR, August, 2025

Pre-Trained Policy Discriminators are General Reward Models.
CoRR, July, 2025

LLMEval-Med: A Real-world Clinical Benchmark for Medical LLMs with Physician Validation.
CoRR, June, 2025

Progressive Mastery: Customized Curriculum Learning with Guided Prompting for Mathematical Reasoning.
CoRR, June, 2025

EvaLearn: Quantifying the Learning Capability and Efficiency of LLMs via Sequential Problem Solving.
CoRR, June, 2025

Compression Hacking: A Supplementary Perspective on Informatics Metric of Language Models from Geometric Distortion.
CoRR, May, 2025

Two Minds Better Than One: Collaborative Reward Modeling for LLM Alignment.
CoRR, May, 2025

Improving RL Exploration for LLM Reasoning through Retrospective Replay.
CoRR, April, 2025

Seed1.5-Thinking: Advancing Superb Reasoning Models with Reinforcement Learning.
CoRR, April, 2025

Detecting Essence Code Clones via Information Theoretic Analysis.
CoRR, February, 2025

Measuring Data Diversity for Instruction Tuning: A Systematic Analysis and A Reliable Metric.
CoRR, February, 2025

Predicting Large Language Model Capabilities on Closed-Book QA Tasks Using Only Information Available Prior to Training.
CoRR, February, 2025

The rise and potential of large language model based agents: a survey.
Sci. China Inf. Sci., 2025

RMB: Comprehensively benchmarking reward models in LLM alignment.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

ToolEyes: Fine-Grained Evaluation for Tool Learning Capabilities of Large Language Models in Real-world Scenarios.
Proceedings of the 31st International Conference on Computational Linguistics, 2025

Revisiting Jailbreaking for Large Language Models: A Representation Engineering Perspective.
Proceedings of the 31st International Conference on Computational Linguistics, 2025

PFDial: A Structured Dialogue Instruction Fine-tuning Method Based on UML Flowcharts.
Proceedings of the Findings of the Association for Computational Linguistics, 2025

Measuring Data Diversity for Instruction Tuning: A Systematic Analysis and A Reliable Metric.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

Lost in the Context: Insufficient and Distracted Attention to Contexts in Preference Modeling.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

DocFusion: A Unified Framework for Document Parsing Tasks.
Proceedings of the Findings of the Association for Computational Linguistics, 2025

Alleviating Shifted Distribution in Human Preference Alignment through Meta-Learning.
Proceedings of the AAAI-25, Sponsored by the Association for the Advancement of Artificial Intelligence, February 25, 2025

2024
COCL: An Intelligent Framework for Enhancing Deep Learning-Based Vulnerability Detection.
IEEE Trans. Ind. Informatics, March, 2024

CC2Vec: Combining Typed Tokens with Contrastive Learning for Effective Code Clone Detection.
Proc. ACM Softw. Eng., 2024

Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision.
CoRR, 2024

Multi-Programming Language Sandbox for LLMs.
CoRR, 2024

What's Wrong with Your Code Generated by Large Language Models? An Extensive Study.
CoRR, 2024

SafeAligner: Safety Alignment against Jailbreak Attacks via Response Disparity Guidance.
CoRR, 2024

Aligning Large Language Models from Self-Reference AI Feedback with one General Principle.
CoRR, 2024

MetaRM: Shifted Distributions Alignment via Meta-Learning.
CoRR, 2024

EasyJailbreak: A Unified Framework for Jailbreaking Large Language Models.
CoRR, 2024

CodeChameleon: Personalized Encryption Framework for Jailbreaking Large Language Models.
CoRR, 2024

Advancing Translation Preference Modeling with RLHF: A Step Towards Cost-Effective Solution.
CoRR, 2024

StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback.
CoRR, 2024

MouSi: Poly-Visual-Expert Vision-Language Models.
CoRR, 2024

Linear Alignment: A Closed-form Solution for Aligning Human Preferences without Tuning and Feedback.
CoRR, 2024

Secrets of RLHF in Large Language Models Part II: Reward Modeling.
CoRR, 2024

ToolEyes: Fine-Grained Evaluation for Tool Learning Capabilities of Large Language Models in Real-world Scenarios.
CoRR, 2024

CausalAPM: Generalizable Literal Disentanglement for NLU Debiasing.
Proceedings of the Natural Language Processing and Chinese Computing, 2024

Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Linear Alignment: A Closed-form Solution for Aligning Human Preferences without Tuning and Feedback.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Improving Generalization of Alignment with Human Preferences through Group Invariant Learning.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

TransferTOD: A Generalizable Chinese Multi-Domain Task-Oriented Dialogue System with Transfer Capabilities.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

LoRAMoE: Alleviating World Knowledge Forgetting in Large Language Models via MoE-Style Plugin.
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2024

StepCoder: Improving Code Generation with Reinforcement Learning from Compiler Feedback.
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2024

2023
LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment.
CoRR, 2023

The Rise and Potential of Large Language Model Based Agents: A Survey.
CoRR, 2023

Towards Understanding the Capability of Large Language Models on Code Clone Detection: A Survey.
CoRR, 2023

Secrets of RLHF in Large Language Models Part I: PPO.
CoRR, 2023

CausalAPM: Generalizable Literal Disentanglement for NLU Debiasing.
CoRR, 2023

Gitor: Scalable Code Clone Detection by Building Global Sample Graph.
Proceedings of the 31st ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering, 2023

Loose lips sink ships: Mitigating Length Bias in Reinforcement Learning from Human Feedback.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, 2023

Detecting Adversarial Samples through Sharpness of Loss Landscape.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, 2023

On the Universal Adversarial Perturbations for Efficient Data-free Adversarial Detection.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, 2023

DSRM: Boost Textual Adversarial Training with Distribution Shift Risk Minimization.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

2022
Decorrelate Irrelevant, Purify Relevant: Overcome Textual Spurious Correlations from a Feature Perspective.
CoRR, 2022

VulCNN: An Image-inspired Scalable Vulnerability Detection System.
Proceedings of the 44th IEEE/ACM 44th International Conference on Software Engineering, 2022

Kernel-Whitening: Overcome Dataset Bias with Isotropic Sentence Embedding.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

Decorrelate Irrelevant, Purify Relevant: Overcome Textual Spurious Correlations from a Feature Perspective.
Proceedings of the 29th International Conference on Computational Linguistics, 2022

MINER: Improving Out-of-Vocabulary Named Entity Recognition from an Information Theoretic Perspective.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022

2021
IntDroid: Android Malware Detection Based on API Intimacy Analysis.
ACM Trans. Softw. Eng. Methodol., 2021

Boosting the Capability of Intelligent Vulnerability Detection by Training in a Human-Learning Manner.
CoRR, 2021

Obfuscation-resilient Android Malware Analysis Based on Contrastive Learning.
CoRR, 2021

2020
SCDetector: Software Functional Clone Detection Based on Semantic Tokens Analysis.
Proceedings of the 35th IEEE/ACM International Conference on Automated Software Engineering, 2020


  Loading...