packages of attack and defense
attack Methods
Badnets: Identifying vulnerabilities in the machine learning model supply chain. |
|
Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning |
|
Blind Backdoors in Deep Learning Models |
|
BppAttack: Stealthy and Efficient Trojan Attacks Against Deep Neural Networks via Image Quantization and Contrastive Adversarial Learning |
|
An Embarrassingly Simple Backdoor Attack on Self-supervised Learning |
|
Input-aware dynamic backdoor attack |
|
Label-Consistent Backdoor Attacks |
|
Rethinking the backdoor attacks' triggers: A frequency perspective |
|
LIRA: Learnable, Imperceptible and Robust Backdoor Attacks |
|
Poison ink: Robust and invisible backdoor attack |
|
Normal training case (Train a clean model with clean data) |
|
Reflection Backdoor: A Natural Backdoor Attack on Deep Neural Networks |
|
A new backdoor attack in CNNs by training set corruption without label poisoning |
|
Invisible backdoor attack with sample-specific triggers |
|
Trojaning Attack on Neural Networks |
|
WaNet - Imperceptible Warping-based Backdoor Attack |
defense Methods
Anti-backdoor learning: Training clean models on poisoned data. |
|
Detecting Backdoor Attacks on Deep Neural Networks by Activation Clustering |
|
Adversarial Neuron Pruning Purifies Backdoored Deep Models |
|
Pre-activation Distributions Expose Backdoor Neurons |
|
Data-free backdoor removal based on channel lipschitzness |
|
Effective backdoor defense by exploiting sensitivity of poisoned samples |
|
Effective backdoor defense by exploiting sensitivity of poisoned samples |
|
Backdoor Defense Via Decoupling The Training Process |
|
Pre-activation Distributions Expose Backdoor Neurons |
|
Fine-Pruning: Defending Against Backdooring Attacks on Deep Neural Networks |
|
Basic class for ft defense method. |
|
Enhancing Fine-Tuning Based Backdoor Defense with Sharpness-Aware Minimization |
|
Adversarial unlearning of backdoors via implicit hypergradient |
|
Pre-activation Distributions Expose Backdoor Neurons |
|
Bridging mode connectivity in loss landscapes and adversarial robustness |
|
Beating Backdoor Attack at Its Own Game |
|
Neural Attention Distillation: Erasing Backdoor Triggers From Deep Neural Networks |
|
Neural Cleanse: Identifying And Mitigating Backdoor Attacks In Neural Networks |
|
Neural polarizer: A lightweight and effective backdoor defense via purifying poisoned features |
|
Reconstructive Neuron Pruning for Backdoor Defense |
|
Shared adversarial unlearning: Backdoor mitigation by unlearning shared adversarial examples |
|
Spectral Signatures in Backdoor Attacks |
inference-time detection Methods
STRIP: A Defence Against Trojan Attacks on Deep Neural Networks |
|
Detecting Backdoors During the Inference Stage Based on Corruption Robustness Consistency |
pretraining detection Methods
STRIP: A Defence Against Trojan Attacks on Deep Neural Networks |
|
The Beatrix Resurrections: Robust Backdoor Detection via Gram Matrices |
|
Demon in the Variant: Statistical Analysis of DNNs for Robust Backdoor Contamination Detection |
|
SPECTRE: Defending Against Backdoor Attacks Using Robust Statistics |
|
Spectral Signatures in Backdoor Attacks |
|
Detecting Backdoor Attacks on Deep Neural Networks by Activation Clustering |