|
Ensemble-based Blackbox Attacks on Dense Prediction
Zikui Cai*, Yaoteng Tan*, M. Salman Asif
(* Equal contribution)
CVPR, 2023
arXiv /
open access /
code /
poster
We propose a query-efficient approach for adversarial attacks on dense prediction models. Our proposed method can generate a single perturbation that can fool multiple blackbox detection and segmentation models simultaneously, demonstrate generalizability across different tasks.
|
|
Targeted Unlearning with Single Layer Unlearning Gradient
Zikui Cai, Yaoteng Tan, M. Salman Asif
Preprint, 2024 (New)
arXiv /
code /
project page
We propose a highly efficient machine unlearning method for fundation models (e.g., CLIP, Stable Diffusion, VLMs) that requires only one-time gradient calculation and one-step update on one model layer that are selected based on introduced metrics, layer importance and gradient alignment.
|
|
Transform-Dependent Adversarial Attacks
Yaoteng Tan, Zikui Cai, M. Salman Asif
Preprint, 2024 (New)
arXiv
Many properties of adversarial attacks are well-studied today (e.g., optimization, transferability, physical implementa-
tion, etc.). In this work, we explore an under-researched transform-dependent property of adversarial attacks, which
the optimization process of additive adversarial perturbations can be combined with various image transformations to
produce versatile, transform-dependent attack effects.
|
Service
Conference review:
Teaching Assistant:
- UCR EE240 Pattern Recognition, 2023, 2024 Spring
- UCR CS171/EE142 Intro. to Machine Learning, 2023 Fall
Acknowledgement: template from Jon Barron
|
|