Skip to content

sonal-ssj/selfstudy-adversarial-robustness

 
 

Repository files navigation

Self-study course in evaluating adversarial robustness

This repository contains a collection of defenses aimed at researchers who wish to learn how to properly evaluate the robustness of adversarial example defenses.

While there is a vast literature of published techniques that help to attack adversarial example defenses, few researchers have practical experience actually running these. This project is designed to give researchers that experience, so that when they develop their own defenses, they can perform a thorough evaluation.

The primary purpose of these defenses is therefore to be pedagogically useful. We aim to make our defenses as simple as possible---even sacrificing accuracy or robustness if we can make a defense simpler.

This simplicity-above-all-else has one cost: it possible for one to be able to break all of our defenses, but still not properly evaluate the robustness of their own (often more complicated codebase) codebase. Researchers who aim to build strong defenses would be well served by first creating a defense as simple as we have here and then analyzing it.

This is currently a preliminary code release. We expect that we will make changes to the defenses, training algorithms, and possibly framework, as we receive feedback. We have no immediate timeline for when it will become stable.

A whitepaper describing this project will be coming in the future. More details on getting started to use this project are available in the following three documents.

List of all defenses

The list below goes (roughly) from the easiest to easier to harder defenses. The first defenses require very little overall knowledge, and by the end we hope to cover all modern defense techniques. However, feel free to study them in any convenient order.

  1. Baseline (defense_baseline)
    • This is naively trained model without any hardening against adversarial examples.
  2. Bluring (defense_blur)
    • This model blurs its input images in an attempt to remove adversarial perturbation.
  3. Softmax temperature (defense_temperature)
    • This model trains a neural network with increased softmax temperature.
  4. Ensemble of binary classifiers (defense_mergebinary)
    • The model works by merging predictions of independent per-class binary classfiers.
  5. Label smoothing (defense_labelsmooth)
    • This model is trained with a label smoothing objective.
  6. Jump (defense_jump)
    • This model applies a "jump" activation function instead of ReLU.
  7. Majority vote (defense_majority)
    • This model takes a best-of-three majority vote among separate classifiers.
  8. Discretization (defense_discretize)
    • This model encodes the input into a more sophisticated discretization of the input.
  9. Random neuron perturbations (defense_randomneuron)
    • This model adds test-time randomness to the activations of a neural network.
  10. Transform (defense_transform)
    • This model randomly modifies the input image before classifying it.
  11. Injection (defense_injection)
    • This model injects backdoors into a model so that inputs can be fingerprinted.
  12. K-nearest neighbours (defense_knn)
    • This model embeds inputs into representation space and then uses a nearest neighbor classifier.
  13. Simple adversarial training (defense_advtrain)
    • This model trains on adversarial examples to become robust to them.

Repository structure

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 94.0%
  • Shell 6.0%