Xiaogeng Liu

Orcid: 0009-0008-7677-2310

According to our database1, Xiaogeng Liu authored at least 26 papers between 2021 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
MetaAgent: Automatically Constructing Multi-Agent Systems Based on Finite State Machines.
CoRR, July, 2025

DRIFT: Dynamic Rule-Based Defense with Injection Isolation for Securing LLM Agents.
CoRR, June, 2025

OET: Optimization-based prompt injection Evaluation Toolkit.
CoRR, May, 2025

Doxing via the Lens: Revealing Privacy Leakage in Image Geolocation for Agentic Multi-Modal Large Reasoning Model.
CoRR, April, 2025

CVE-Bench: Benchmarking LLM-based Software Engineering Agent's Ability to Repair Real-World CVE Vulnerabilities.
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies, 2025

RePD: Defending Jailbreak Attack through a Retrieval-based Prompt Decomposition Process.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2025, Albuquerque, New Mexico, USA, April 29, 2025

Can Watermarks be Used to Detect LLM IP Infringement For Free?
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

AutoDAN-Turbo: A Lifelong Agent for Strategy Self-Exploration to Jailbreak LLMs.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

AGrail: A Lifelong Agent Guardrail with Effective and Adaptive Safety Detection.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

PIGuard: Prompt Injection Guardrail via Mitigating Overdefense for Free.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

2024
InjecGuard: Benchmarking and Mitigating Over-defense in Prompt Injection Guardrail Models.
CoRR, 2024

Visual-RolePlay: Universal Jailbreak Attack on MultiModal Large Language Models via Role-playing Image Characte.
CoRR, 2024

JailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against Jailbreak Attacks.
CoRR, 2024

Automatic and Universal Prompt Injection Attacks against Large Language Models.
CoRR, 2024

Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models.
Proceedings of the 33rd USENIX Security Symposium, 2024

Why Does Little Robustness Help? A Further Step Towards Understanding Adversarial Transferability.
Proceedings of the IEEE Symposium on Security and Privacy, 2024

AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

AdaShield : Safeguarding Multimodal Large Language Models from Structure-Based Attack via Adaptive Shield Prompting.
Proceedings of the Computer Vision - ECCV 2024, 2024

2023
DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions.
CoRR, 2023

Why Does Little Robustness Help? Understanding Adversarial Transferability From Surrogate Training.
CoRR, 2023

PointCRT: Detecting Backdoor in 3D Point Cloud via Corruption Robustness.
Proceedings of the 31st ACM International Conference on Multimedia, 2023

Detecting Backdoors During the Inference Stage Based on Corruption Robustness Consistency.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023

2022
Towards Efficient Data-Centric Robust Machine Learning with Noise-based Augmentation.
CoRR, 2022

Protecting Facial Privacy: Generating Adversarial Identity Masks via Style-robust Makeup Transfer.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

2021
AdvHash: Set-to-set Targeted Attack on Deep Hashing with One Single Adversarial Patch.
Proceedings of the MM '21: ACM Multimedia Conference, Virtual Event, China, October 20, 2021


  Loading...