Xinyue Shen

Orcid: 0009-0006-9954-587X

Affiliations:
  • CISPA Helmholtz Center for Information Security, Saarbrucken, Saarland, Germany


According to our database1, Xinyue Shen authored at least 24 papers between 2021 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
HateBench: Benchmarking Hate Speech Detectors on LLM-Generated Content and Hate Campaigns.
CoRR, January, 2025

HateBench: Benchmarking Hate Speech Detectors on LLM-Generated Content and Hate Campaigns.
Dataset, January, 2025

HateBench: Benchmarking Hate Speech Detectors on LLM-Generated Content and Hate Campaigns.
Dataset, January, 2025

On the Effectiveness of Prompt Stealing Attacks on In-the-Wild Prompts.
Proceedings of the IEEE Symposium on Security and Privacy, 2025

GPTracker: A Large-Scale Measurement of Misused GPTs.
Proceedings of the IEEE Symposium on Security and Privacy, 2025

JailbreakRadar: Comprehensive Assessment of Jailbreak Attacks Against LLMs.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

Are We in the AI-Generated Text World Already? Quantifying and Monitoring AIGT on Social Media.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

When GPT Spills the Tea: Comprehensive Assessment of Knowledge File Leakage in GPTs.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

2024
<i>ModSCAN</i>: Measuring Stereotypical Bias in Large Vision-Language Models from Vision and Language Modalities.
CoRR, 2024

Voice Jailbreak Attacks Against GPT-4o.
CoRR, 2024

UnsafeBench: Benchmarking Image Safety Classifiers on Real-World and AI-Generated Images.
CoRR, 2024

Comprehensive Assessment of Jailbreak Attacks Against LLMs.
CoRR, 2024

Prompt Stealing Attacks Against Text-to-Image Generation Models.
Proceedings of the 33rd USENIX Security Symposium, 2024

Games and Beyond: Analyzing the Bullet Chats of Esports Livestreaming.
Proceedings of the Eighteenth International AAAI Conference on Web and Social Media, 2024

The Death and Life of Great Prompts: Analyzing the Evolution of LLM Prompts from the Structural Perspective.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

ModSCAN: Measuring Stereotypical Bias in Large Vision-Language Models from Vision and Language Modalities.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

"Do Anything Now": Characterizing and Evaluating In-The-Wild Jailbreak Prompts on Large Language Models.
Proceedings of the 2024 on ACM SIGSAC Conference on Computer and Communications Security, 2024

MGTBench: Benchmarking Machine-Generated Text Detection.
Proceedings of the 2024 on ACM SIGSAC Conference on Computer and Communications Security, 2024

2023
Comprehensive Assessment of Toxicity in ChatGPT.
CoRR, 2023

In ChatGPT We Trust? Measuring and Characterizing the Reliability of ChatGPT.
CoRR, 2023

Unsafe Diffusion: On the Generation of Unsafe Images and Hateful Memes From Text-To-Image Models.
Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, 2023

2022
Backdoor Attacks in the Supply Chain of Masked Image Modeling.
CoRR, 2022

On Xing Tian and the Perseverance of Anti-China Sentiment Online.
Proceedings of the Sixteenth International AAAI Conference on Web and Social Media, 2022

2021
Evil Under the Sun: Understanding and Discovering Attacks on Ethereum Decentralized Applications.
Proceedings of the 30th USENIX Security Symposium, 2021


  Loading...