Adversarial examples are special inputs to deep learning models, maliciously crafted to fool them into incorrect outputs. Even the state-of-the-art models are vulnerable to adversarial attacks, thus a lot of issues arise in many security fields of artificial intelligence. In this repo we aim at investigating techniques for training adversarially robust models.
Examples of adversarial perturbations:
data/
training data and adversarial perturbationsnotebooks/
results/
collected results and plotsimages/
src/
implementationsRandomProjections/
methods based on random projectionsBayesianSGD/
implementation of Bayesian SGD from Blei et al. (2017)BayesianInference/
BNN training using VI and HMC
trained_models/
baseline/
randens/
randreg/
bnn/
tensorboard/
Scripts should be executed from src/
directory.