Skip to content

Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness (Neurips 2023)

License

Notifications You must be signed in to change notification settings

tml-tuebingen/pags

Repository files navigation

Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness (NeurIPS 2023)

Conceptual Overview

This is the code repository for the NeurIPS'23 paper "Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness" by Suraj Srinivas*, Sebastian Bordt* and Himabindu Lakkaraju.

The repository contains the code to train regularized models on the different dataset, estimate on- and off-manifold robustness, and produce the various figures in the paper.

Overview

Here we briefly describe the structure of the code.

run_experiments.py: Entry point for the different experiments.

experiments/measure_manifold_robustness.py: Script to measure the on- and off-manifold robustness of models on CIFAR-10.

experiments/measure_score_alignment.py: Script to measure the alignment of models on CIFAR-10 with the score.

notebooks/*: Jupyter notebooks to perform analysis and generate the figures in the paper.

train_models.py: Train robust models on CIFAR-10.

train_robust_imagenet.py: Train robust models with projected gradient descent on ImageNet64x64.

utils/regularized_loss.py: Defines the different regularized losses for training robust models.

utils/edm_score.py: Wrapping the diffusion models from Karas et al. (2022) to estimate the score.

Citation

If you find this code useful in your research, please consider citing the paper.

@inproceedings{srinivas2023pags,
  author    = {Suraj Srinivas and Sebastian Bordt and Himabindu Lakkaraju},
  title     = {Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness},
  booktitle = {NeurIPS},
  year      = {2023}
}

Related Research

Acknowledgements

  • We use the diffusion models from Karas et al. "Elucidating the Design Space of Diffusion-Based Generative Models (EDM)", NeurIPS'22.
  • We use the robustness library and pre-trained models from Salman et al. "Do Adversarially Robust ImageNet Models Transfer Better?", NeurIPS'20.
  • We use the LPIPS metric from Zhang et al. "The Unreasonable Effectiveness of Deep Features as a Perceptual Metric", CVPR'18.
  • We train an autoencoder on CIFAR-10 using pythae from Chadebec et al. "Pythae: Unifying Generative Autoencoders in Python - A Benchmarking Use Case", NeurIPS'22.

About

Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness (Neurips 2023)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published