# convex_adversarial
**Repository Path**: xieck13/convex_adversarial
## Basic Information
- **Project Name**: convex_adversarial
- **Description**: No description available
- **Primary Language**: Unknown
- **License**: MIT
- **Default Branch**: master
- **Homepage**: None
- **GVP Project**: No
## Statistics
- **Stars**: 0
- **Forks**: 0
- **Created**: 2021-02-17
- **Last Updated**: 2021-02-17
## Categories & Tags
**Categories**: Uncategorized
**Tags**: None
## README
# Provably robust neural networks
*A repository for training provably robust neural networks by optimizing convex outer bounds on the adversarial polytope. Created by [Eric Wong](https://riceric22.github.io) and [Zico Kolter](http://zicokolter.com). [Link to the original arXiv paper][paper]. The method has been further extended to be fully modular, scalable, and use cascades to improve robust error. Check out our new paper on arXiv: [Scaling provable adversarial defenses][scalable_paper].*
[paper]: https://arxiv.org/abs/1711.00851
[scalable_paper]: https://arxiv.org/abs/1805.12514
## News
+ 10/30/2018 - Version 0.4 released to reflect the NIPS final copy. Added
model weights for all models described in the paper.
+ 7/26/2018 - Version 0.3.3 code refactor
+ 6/2/2018 - Version 0.3.1 released to reflect the new paper.
+ 5/31/2018 - New paper on a scalable version for models with skip connections
and a fully modular implementation for simple extension. Code base with these
improvements with a port to PyTorch 0.4 will be released shortly.
+ 4/26/2018 - Added robust models from the paper to the `models/` folder in the
repository.
+ 3/4/2018 - Updated paper with more experiments. Code migrated to work with
PyTorch 0.3.0+. Real mini-batching implemented, with a 5x speedup over the
old codebase, and several NaN bugfixes.
+ 12/8/2017 - Best defense paper at the NIPS 2017 ML & Security Workshop
+ 11/2/2017 - Initial preprint and release of codebase.
## Installation & Usage
You can install this repository with
`pip install convex_adversarial`
If you wish to have the version of code that reflects the first paper, use
`pip install convex_adversal=0.2`, or clone the [0.2 release on Github](https://github.com/locuslab/convex_adversarial/tree/v0.2).
The package contains the following functions:
+ `robust_loss(net, epsilon, X, y, l1_proj=None,
l1_type='exact', bounded_input=False, size_average=True)`
computes a robust loss function for a given ReLU network `net` and l1
radius `epsilon` for examples `X` and their labels `y`. You can use
this as a drop in replacement for, say, `nn.CrossEntropyLoss()`, and is
equivalent to the objective of Equation 14 in the original paper.
To use the scalable version, specify a projection dimension with `l1_proj`
and set `l1_type` to `median`.
+ `robust_loss_parallel` computes the same objective as `robust_loss`, but
only for a *single* example and using
data parallelism. This is useful for exact evaluation if a single
example doesn't fit in memory.
+ `dual_net = DualNetwork(net, X, epsilon, l1_proj=None, l1_type='exact', bounded_input=False)`
is a PyTorch module that computes the layer-wise upper and lower bounds for
all activations in the network. This is useful if you are only interested
in the bounds and not the robust loss, and corresponds to Algorithm
1 in the paper.
+ `dual_net(c)` is the module's forward pass which computes the lower
bound on the primal problem described in the paper for a given
objective vector c. This corresponds to computing objective of Theorem 1 in
the paper (Equation 5).
## Why do we need robust networks?
While networks are capable of representing highly complex functions. For
example, with today's networks it is an easy task to achieve 99% accuracy on
the MNIST digit recognition dataset, and we can quickly train a small network
that can accurately predict that the following image is a 7.
However, the versatility of neural networks comes at a cost: these networks
are highly susceptible to small perturbations, or adversarial attacks (e.g. the [fast gradient sign method](https://arxiv.org/abs/1412.6572) and [projected gradient descent](https://arxiv.org/abs/1706.06083))! While
most of us can recognize that the following image is still a 7, the same
network that could correctly classify the above image instead classifies
the following image as a 3.
While this is a relatively harmless example, one can easily think of
situations where such adversarial perturbations can be dangerous and costly
(e.g. autonomous driving).
## What are robust networks?
Robust networks are networks that are trained to protect against any sort of
adversarial perturbation. Specifically, for any seen training example, the
network is robust if it is impossible to cause the network to incorrectly
classify the example by adding a small perturbation.
## How do we do this?
The short version: we use the dual of a convex relaxation of the network over
the adversarial polytope to lower bound the output. This lower bound can be
expressed as another deep network with the same model parameters, and
optimizing this lower bound allows us to guarantee robustness of the network.
The long version: see our original paper,
[Provable defenses against adversarial examples via the convex outer adversarial polytope][paper].
For our updated version which is scalable, modular, and achieves even better
robust performance, see our new paper,
[Scaling provable adversarial defenses][scalable_paper].
## What difference does this make?
We illustrate the power of training robust networks in the following two scenarios: 2D toy case for a visualization, and on the MNIST dataset. More experiments are in the paper.
### 2D toy example
To illustrate the difference, consider a binary classification task on 2D
space, separating red dots from blue dots. Optimizing a neural network in the
usual fashion gives us the following classifier on the left, and our robust
method gives the classifier on the right. The squares around each example
represent the adversarial region of perturbations.