Zhenhong Zhou

Orcid: 0000-0003-4065-1740

According to our database1, Zhenhong Zhou authored at least 24 papers between 2021 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
Hidden in the Noise: Unveiling Backdoors in Audio LLMs Alignment through Latent Acoustic Pattern Triggers.
CoRR, August, 2025

Jailbreaking Large Language Diffusion Models: Revealing Hidden Safety Flaws in Diffusion-Based Text Generation.
CoRR, July, 2025

RECALLED: An Unbounded Resource Consumption Attack on Large Vision-Language Models.
CoRR, July, 2025

Goal-Aware Identification and Rectification of Misinformation in Multi-Agent Systems.
CoRR, June, 2025

PD<sup>3</sup>F: A Pluggable and Dynamic DoS-Defense Framework Against Resource Consumption Attacks Targeting Large Language Models.
CoRR, May, 2025

LIFEBench: Evaluating Length Instruction Following in Large Language Models.
CoRR, May, 2025

A Vision for Auto Research with LLM Agents.
CoRR, April, 2025

A Comprehensive Survey in LLM(-Agent) Full Stack Safety: Data, Training and Deployment.
CoRR, April, 2025

CORBA: Contagious Recursive Blocking Attacks on Multi-Agent Systems Based on Large Language Models.
CoRR, February, 2025

DemonAgent: Dynamically Encrypted Multi-Backdoor Implantation Attack on LLM-based Agent.
CoRR, February, 2025

Reinforced Lifelong Editing for Language Models.
CoRR, February, 2025

On the Role of Attention Heads in Large Language Model Safety.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

Crabs: Consuming Resource via Auto-generation for LLM-DoS Attack under Black-box Settings.
Proceedings of the Findings of the Association for Computational Linguistics, 2025

2024
Enforcing group fairness in privacy-preserving Federated Learning.
Future Gener. Comput. Syst., 2024

Crabs: Consuming Resrouce via Auto-generation for LLM-DoS Attack under Black-box Settings.
CoRR, 2024

On the Role of Attention Heads in Large Language Model Safety.
CoRR, 2024

Alignment-Enhanced Decoding:Defending via Token-Level Adaptive Refining of Probability Distributions.
CoRR, 2024

Course-Correction: Safety Alignment Using Synthetic Preferences.
CoRR, 2024

Speak Out of Turn: Safety Vulnerability of Large Language Models in Multi-turn Dialogue.
CoRR, 2024

How Alignment and Jailbreak Work: Explain LLM Safety through Intermediate Hidden States.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2024, 2024

Course-Correction: Safety Alignment Using Synthetic Preferences.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: EMNLP 2024, 2024

Alignment-Enhanced Decoding: Defending Jailbreaks via Token-Level Adaptive Refining of Probability Distributions.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Quantifying and Analyzing Entity-Level Memorization in Large Language Models.
Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence, 2024

2021
Three-Dimensional Reconstruction of Huizhou Landscape Combined with Multimedia Technology and Geographic Information System.
Mob. Inf. Syst., 2021


  Loading...