Ashish Hooda

Orcid: 0000-0002-2928-919X

According to our database1, Ashish Hooda authored at least 19 papers between 2021 and 2025.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
What Really is a Member? Discrediting Membership Inference via Poisoning.
CoRR, June, 2025

Through the Stealth Lens: Rethinking Attacks and Defenses in RAG.
CoRR, June, 2025

Computing Optimization-Based Prompt Injections Against Closed-Weights Models By Misusing a Fine-Tuning API.
CoRR, January, 2025

Fun-tuning: Characterizing the Vulnerability of Proprietary LLMs to Optimization-Based Prompt Injection Attacks via the Fine-Tuning Interface.
Proceedings of the IEEE Symposium on Security and Privacy, 2025

Functional Homotopy: Smoothing Discrete Optimization via Continuous Parameters for LLM Jailbreak Attacks.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

2024
PolicyLR: A Logic Representation For Privacy Policies.
CoRR, 2024

Synthetic Counterfactual Faces.
CoRR, 2024

Do Large Code Models Understand Programming Concepts? A Black-box Approach.
CoRR, 2024

D4: Detection of Adversarial Diffusion Deepfakes Using Disjoint Ensembles.
Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2024

Experimental Analyses of the Physical Surveillance Risks in Client-Side Content Scanning.
Proceedings of the 31st Annual Network and Distributed System Security Symposium, 2024

Do Large Code Models Understand Programming Concepts? Counterfactual Analysis for Code Predicates.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

PRP: Propagating Universal Perturbations to Attack Large Language Model Guard-Rails.
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2024

2023
Theoretically Principled Trade-off for Stateful Defenses against Query-Based Black-Box Attacks.
CoRR, 2023

Investigating Stateful Defenses Against Black-Box Adversarial Examples.
CoRR, 2023

Stateful Defenses for Machine Learning Models Are Not Yet Secure Against Black-box Attacks.
Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, 2023

2022
SkillFence: A Systems Approach to Practically Mitigating Voice-Based Confusion Attacks.
Proc. ACM Interact. Mob. Wearable Ubiquitous Technol., 2022

Re-purposing Perceptual Hashing based Client Side Scanning for Physical Surveillance.
CoRR, 2022

Towards Adversarially Robust Deepfake Detection: An Ensemble Approach.
CoRR, 2022

2021
Invisible Perturbations: Physical Adversarial Examples Exploiting the Rolling Shutter Effect.
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2021


  Loading...