Yang Zhang

Orcid: 0000-0003-3612-7348

Affiliations:
  • CISPA, Saarbrücken, Germany
  • Saarland University, Saarland Informatics Campus, Saarbrücken, Germany


According to our database1, Yang Zhang authored at least 121 papers between 2013 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Efficient Data-Free Model Stealing with Label Diversity.
CoRR, 2024

Comprehensive Assessment of Jailbreak Attacks Against LLMs.
CoRR, 2024

2023
FAKEPCD: Fake Point Cloud Detection via Source Attribution.
CoRR, 2023

Comprehensive Assessment of Toxicity in ChatGPT.
CoRR, 2023

Generated Distributions Are All You Need for Membership Inference Attacks Against Generative Models.
CoRR, 2023

On the Proactive Generation of Unsafe Images From Text-To-Image Models Using Benign Prompts.
CoRR, 2023

SecurityNet: Assessing Machine Learning Vulnerabilities on Public Models.
CoRR, 2023

Quantifying Privacy Risks of Prompts in Visual Prompt Learning.
CoRR, 2023

Last One Standing: A Comparative Analysis of Security and Privacy of Soft Prompt Tuning, LoRA, and In-Context Learning.
CoRR, 2023

Composite Backdoor Attacks Against Large Language Models.
CoRR, 2023

Prompt Backdoors in Visual Prompt Learning.
CoRR, 2023

Robustness Over Time: Understanding Adversarial Examples' Effectiveness on Longitudinal Versions of Large Language Models.
CoRR, 2023

White-box Membership Inference Attacks against Diffusion Models.
CoRR, 2023

You Only Prompt Once: On the Capabilities of Prompt Learning on Large Language Models to Tackle Toxic Content.
CoRR, 2023

You Are How You Walk: Quantifying Privacy Risks in Step Count Data.
CoRR, 2023

"Do Anything Now": Characterizing and Evaluating In-The-Wild Jailbreak Prompts on Large Language Models.
CoRR, 2023

Mondrian: Prompt Abstraction Attack Against Large Language Models for Cheaper API Pricing.
CoRR, 2023

Generative Watermarking Against Unauthorized Subject-Driven Image Synthesis.
CoRR, 2023

Vision-language models boost food composition compilation.
CoRR, 2023

Watermarking Diffusion Model.
CoRR, 2023

In ChatGPT We Trust? Measuring and Characterizing the Reliability of ChatGPT.
CoRR, 2023

MGTBench: Benchmarking Machine-Generated Text Detection.
CoRR, 2023

From Visual Prompt Learning to Zero-Shot Transfer: Mapping Is All You Need.
CoRR, 2023

Prompt Stealing Attacks Against Text-to-Image Generation Models.
CoRR, 2023

A Plot is Worth a Thousand Words: Model Information Stealing Attacks via Scientific Plots.
Proceedings of the 32nd USENIX Security Symposium, 2023

PrivTrace: Differentially Private Trajectory Synthesis by Adaptive Markov Models.
Proceedings of the 32nd USENIX Security Symposium, 2023

Two-in-One: A Model Hijacking Attack Against Text Generation Models.
Proceedings of the 32nd USENIX Security Symposium, 2023

UnGANable: Defending Against GAN-based Face Manipulation.
Proceedings of the 32nd USENIX Security Symposium, 2023

FACE-AUDITOR: Data Auditing in Facial Recognition Systems.
Proceedings of the 32nd USENIX Security Symposium, 2023

On the Evolution of (Hateful) Memes by Means of Multimodal Contrastive Learning.
Proceedings of the 44th IEEE Symposium on Security and Privacy, 2023

Backdoor Attacks Against Dataset Distillation.
Proceedings of the 30th Annual Network and Distributed System Security Symposium, 2023

Generated Graph Detection.
Proceedings of the International Conference on Machine Learning, 2023

Data Poisoning Attacks Against Multimodal Encoders.
Proceedings of the International Conference on Machine Learning, 2023

Is Adversarial Training Really a Silver Bullet for Mitigating Data Poisoning?
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Can't Steal? Cont-Steal! Contrastive Stealing Attacks Against Image Encoders.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023

DE-FAKE: Detection and Attribution of Fake Images Generated by Text-to-Image Generation Models.
Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, 2023

Unsafe Diffusion: On the Generation of Unsafe Images and Hateful Memes From Text-To-Image Models.
Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, 2023

Differentially Private Resource Allocation.
Proceedings of the Annual Computer Security Applications Conference, 2023

NOTABLE: Transferable Backdoor Attacks Against Prompt-based NLP Models.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

Pseudo Label-Guided Model Inversion Attack via Conditional Generative Adversarial Network.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023

2022
FairSR: Fairness-aware Sequential Recommendation through Multi-Task Learning with Preference Graph Embeddings.
ACM Trans. Intell. Syst. Technol., 2022

Fine-Tuning Is All You Need to Mitigate Backdoor Attacks.
CoRR, 2022

DE-FAKE: Detection and Attribution of Fake Images Generated by Text-to-Image Diffusion Models.
CoRR, 2022

Backdoor Attacks in the Supply Chain of Masked Image Modeling.
CoRR, 2022

Membership Inference Attacks Against Text-to-image Generation Models.
CoRR, 2022

PrivTrace: Differentially Private Trajectory Synthesis by Adaptive Markov Model.
CoRR, 2022

Membership-Doctor: Comprehensive Assessment of Membership Inference Against Machine Learning Models.
CoRR, 2022

ML-Doctor: Holistic Risk Assessment of Inference Attacks Against Machine Learning Models.
Proceedings of the 31st USENIX Security Symposium, 2022

Inference Attacks Against Graph Neural Networks.
Proceedings of the 31st USENIX Security Symposium, 2022

Teacher Model Fingerprinting Attacks Against Transfer Learning.
Proceedings of the 31st USENIX Security Symposium, 2022

Model Stealing Attacks Against Inductive Graph Neural Networks.
Proceedings of the 43rd IEEE Symposium on Security and Privacy, 2022

Amplifying Membership Exposure via Data Poisoning.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Property Inference Attacks Against GANs.
Proceedings of the 29th Annual Network and Distributed System Security Symposium, 2022

Get a Model! Model Hijacking Attack Against Machine Learning Models.
Proceedings of the 29th Annual Network and Distributed System Security Symposium, 2022

On Xing Tian and the Perseverance of Anti-China Sentiment Online.
Proceedings of the Sixteenth International AAAI Conference on Web and Social Media, 2022

Dynamic Backdoor Attacks Against Machine Learning Models.
Proceedings of the 7th IEEE European Symposium on Security and Privacy, 2022

Semi-Leak: Membership Inference Attacks Against Semi-supervised Learning.
Proceedings of the Computer Vision - ECCV 2022, 2022

Why So Toxic?: Measuring and Triggering Toxic Behavior in Open-Domain Chatbots.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

Finding MNEMON: Reviving Memories of Node Embeddings.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

Membership Inference Attacks by Exploiting Loss Trajectory.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

Auditing Membership Leakages of Multi-Exit Networks.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

On the Privacy Risks of Cell-Based NAS Architectures.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

SSLGuard: A Watermarking Scheme for Self-supervised Learning Pre-trained Encoders.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

Graph Unlearning.
Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022

2021
DatingSec: Detecting Malicious Accounts in Dating Apps Using a Content-Based Attention Network.
IEEE Trans. Dependable Secur. Comput., 2021

On exploring feature representation learning of items to forecast their rise and fall in social media.
J. Intell. Inf. Syst., 2021

Node-Level Membership Inference Attacks Against Graph Neural Networks.
CoRR, 2021

"Go eat a bat, Chang!": On the Emergence of Sinophobic Behavior on Web Communities in the Face of COVID-19.
Proceedings of the WWW '21: The Web Conference 2021, 2021

Stealing Links from Graph Neural Networks.
Proceedings of the 30th USENIX Security Symposium, 2021

PrivSyn: Differentially Private Data Synthesis.
Proceedings of the 30th USENIX Security Symposium, 2021

MLCapsule: Guarded Offline Deployment of Machine Learning as a Service.
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2021

Membership Inference Attacks Against Recommender Systems.
Proceedings of the CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, Republic of Korea, November 15, 2021

Membership Leakage in Label-Only Exposures.
Proceedings of the CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, Republic of Korea, November 15, 2021

Quantifying and Mitigating Privacy Risks of Contrastive Learning.
Proceedings of the CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, Republic of Korea, November 15, 2021

When Machine Unlearning Jeopardizes Privacy.
Proceedings of the CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, Republic of Korea, November 15, 2021

BadNL: Backdoor Attacks against NLP Models with Semantic-preserving Improvements.
Proceedings of the ACSAC '21: Annual Computer Security Applications Conference, Virtual Event, USA, December 6, 2021

2020
Don't Trigger Me! A Triggerless Backdoor Attack Against Deep Neural Networks.
CoRR, 2020

BAAAN: Backdoor Attacks Against Autoencoder and GAN-Based Machine Learning Models.
CoRR, 2020

Privacy Analysis of Deep Learning in the Wild: Membership Inference Attacks against Transfer Learning.
CoRR, 2020

Label-Leaks: Membership Inference Attack with Label.
CoRR, 2020

BadNL: Backdoor Attacks Against NLP Models.
CoRR, 2020

"Go eat a bat, Chang!": An Early Look on the Emergence of Sinophobic Behavior on Web Communities in the Face of COVID-19.
CoRR, 2020

Everything About You: A Multimodal Approach towards Friendship Inference in Online Social Networks.
CoRR, 2020

NeuLP: An End-to-End Deep-Learning Model for Link Prediction.
Proceedings of the Web Information Systems Engineering - WISE 2020, 2020

Updates-Leak: Data Set Inference and Reconstruction Attacks in Online Learning.
Proceedings of the 29th USENIX Security Symposium, 2020

Towards Plausible Graph Anonymization.
Proceedings of the 27th Annual Network and Distributed System Security Symposium, 2020

Membership Inference Against DNA Methylation Databases.
Proceedings of the IEEE European Symposium on Security and Privacy, 2020

LEAF: A Faster Secure Search Algorithm via Localization, Extraction, and Reconstruction.
Proceedings of the CCS '20: 2020 ACM SIGSAC Conference on Computer and Communications Security, 2020

GAN-Leaks: A Taxonomy of Membership Inference Attacks against Generative Models.
Proceedings of the CCS '20: 2020 ACM SIGSAC Conference on Computer and Communications Security, 2020

2019
An active learning-based approach for location-aware acquaintance inference.
Knowl. Inf. Syst., 2019

GAN-Leaks: A Taxonomy of Membership Inference Attacks against GANs.
CoRR, 2019

Language in Our Time: An Empirical Analysis of Hashtags.
Proceedings of the World Wide Web Conference, 2019

A Graph-Based Approach to Explore Relationship Between Hashtags and Images.
Proceedings of the Web Information Systems Engineering - WISE 2019, 2019

ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models.
Proceedings of the 26th Annual Network and Distributed System Security Symposium, 2019

MBeacon: Privacy-Preserving Beacons for DNA Methylation Data.
Proceedings of the 26th Annual Network and Distributed System Security Symposium, 2019

Fairwalk: Towards Fair Graph Embedding.
Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, 2019

MemGuard: Defending against Black-Box Membership Inference Attacks via Adversarial Examples.
Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, 2019

How to prove your model belongs to you: a blind-watermark based framework to protect intellectual property of DNN.
Proceedings of the 35th Annual Computer Security Applications Conference, 2019

2018
ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models.
CoRR, 2018

Tagvisor: A Privacy Advisor for Sharing Hashtags.
Proceedings of the 2018 World Wide Web Conference on World Wide Web, 2018

You Are Where You App: An Assessment on Location Privacy of Social Applications.
Proceedings of the 29th IEEE International Symposium on Software Reliability Engineering, 2018

Dissecting Privacy Risks in Biomedical Data.
Proceedings of the 2018 IEEE European Symposium on Security and Privacy, 2018

2017
CTRL+Z: Recovering Anonymized Social Graphs.
CoRR, 2017

Does #like4like indeed provoke more likes?
Proceedings of the International Conference on Web Intelligence, 2017

DeepCity: A Feature Learning Framework for Mining Location Check-Ins.
Proceedings of the Eleventh International Conference on Web and Social Media, 2017

Quantifying Location Sociality.
Proceedings of the 28th ACM Conference on Hypertext and Social Media, 2017

Semantic Annotation for Places in LBSN through Graph Embedding.
Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, 2017

walk2friends: Inferring Social Links from Mobility Profiles.
Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, 2017

2016
On Impact of Weather on Human Mobility in Cities.
Proceedings of the Web Information Systems Engineering - WISE 2016, 2016

An Empirical Study on User Access Control in Online Social Networks.
Proceedings of the 21st ACM on Symposium on Access Control Models and Technologies, 2016

Modeling City Locations as Complex Networks: An initial study.
Proceedings of the Complex Networks & Their Applications V - Proceedings of the 5th International Workshop on Complex Networks and their Applications (COMPLEX NETWORKS 2016), Milan, Italy, November 30, 2016

2015
A new access control scheme for Facebook-style social networks.
Comput. Secur., 2015

Exploring Communities for Effective Location Prediction.
Proceedings of the 24th International Conference on World Wide Web Companion, 2015

A Logical Approach to Restricting Access in Online Social Networks.
Proceedings of the 20th ACM Symposium on Access Control Models and Technologies, 2015

Community-Driven Social Influence Analysis and Applications.
Proceedings of the Engineering the Web in the Big Data Era - 15th International Conference, 2015

Location Prediction: Communities Speak Louder than Friends.
Proceedings of the 2015 ACM on Conference on Online Social Networks, 2015

Cryptographic Protocols for Enforcing Relationship-Based Access Control Policies.
Proceedings of the 39th IEEE Annual Computer Software and Applications Conference, 2015

Inferring Friendship from Check-in Data of Location-Based Social Networks.
Proceedings of the 2015 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, 2015

Distance and Friendship: A Distance-Based Model for Link Prediction in Social Networks.
Proceedings of the Web Technologies and Applications - 17th Asia-PacificWeb Conference, 2015

Event Prediction with Community Leaders.
Proceedings of the 10th International Conference on Availability, Reliability and Security, 2015

2013
Twisted edwards-form elliptic curve cryptography for 8-bit AVR-based sensor nodes.
Proceedings of the first ACM workshop on Asia public-key cryptography, 2013


  Loading...