This repository contains code of algorithm AdvLC and pre-trained models from the paper (published in Pattern Recognition 2022) "Understanding and Combating Robust Overfitting via Input Loss Landscape Anlaysis and Regularization".
Please find the pre-trained models through this link.
data
: datasetmodel
: model checkpointstrained
: saved model checkpoints
output
: experiment logssrc
: source codetrain.py
: training modelsadversary.py
: evaluating adversarial robustnessutils
: shared utilities such training, evaluation, log, printing, adversary, multiprocessing distributionmodel
: model architecturesdata
: datasetsconfig
: configurations for training and adversarial evaluation
The development environment is:
- Python 3.8.13
- PyTorch 1.11.0 + torchvision 0.12.0
The remaining dependencies are specified in the file requirements.txt
and can be easily installed via the command:
pip install -r requirements.txt
To prepare the involved dataset (pre-trained model), an optional parameter --download
(--ptrained
) should be specified in the running command. The program will download the required files automatically. This functionality currently doesn't support the dataset Tiny ImageNet.
- The training script is based on the PyTorch official example
- the code of Wide ResNet is a revised version of wide-resnet.pytorch.
- the code of PreAct ResNet is from Alleviate-Robust_Overfitting
- Stochastic Weight Averaging (SWA): Alleviate-Robust_Overfitting
- Hessian spectrum computation: PyHessian
To adversarially train a PreAct ResNet18 on CIFAR10 using PGD10, run:
python src/train.py -a paresnet --depth 18 --max_iter 10
To adversarially train a PreAct ResNet18 on CIFAR10 using FGSM-N, run:
python src/train.py -a paresnet --depth 18 -ri --eps_step 8
To adversarially train a PreAct ResNet18 on CIFAR10 using PGD10 with the proposed regularization, run:
python src/train.py -a paresnet --depth 18 --max_iter 10 --reg_lam 0.3 --reg_top 0.1
To adversarially train a PreAct ResNet18 on CIFAR10 using PGD10 with the proposed regularization with SWA, run:
python src/train.py -a paresnet --depth 18 --max_iter 10 --reg_lam 0.4 --reg_top 0.1 --swa 50 n 500
There are also a lot of hyper-parameters allowed to be specified in the running command in order to control the training. The common hyper-parameters, shared between src/train.py
and src/adversary.py
are stored in the src/config/config.py
and the task-specific hyper-parameters are defined in the corresponding configuration file in the src/config
folder. Please refer to the specific configuration file for the details of the default and the available options.
For each training, the checkpoints will be saved in model/trained/{log}
where {log} is the name of the experiment logbook (by default, is log
). Each instance of training is tagged with a unique identifier, found in the logbook output/log/{log}.json
, and that id is later used to load the well-trained model for the evaluation.
To evaluate the robustness of the "best" checkpoint against PGD50, run:
python src/adversary.py 0000 -v pgd -a PGD --max_iter 50
Similarly against AutoAttack (AA), run:
python src/adversary.py 0000 -v pgd -a AA
where "0000" should be replaced the real identifier to be evaluated.