Skip to content

Releases: RobustBench/robustbench

Add Tian2022Deeper

09 Apr 09:33
2317b19
Compare
Choose a tag to compare
Add Tian2022Deeper Pre-release
Pre-release
Merge pull request #133 from Hollen0318/Hollen

notes for dataset and threat model

timm integration and new models

28 Sep 17:13
Compare
Choose a tag to compare

What's Changed

  • Added new model ID's by @nmndeep in #52
  • Modas2021_PRIME models added by @nmndeep in #57
  • readme correction by @nmndeep in #60
  • Remove all .cuda() calls by @dedeswim in #61
  • Enable using ImageNet benchmarking for new models by @dedeswim in #64
  • add Kang2021Stable model by @fra31 in #66
  • sort models including external evaluations by @fra31 in #69
  • add results from Erichson2022NoisyMix by @fra31 in #70
  • Explicit encoding in setup.py to fix #72 by @dedeswim in #74
  • Updated results for Sehwag2021Proxy by @VSehwag in #76
  • Pang2022Robustness models added by @nmndeep in #77
  • additional data flags corrected by @nmndeep in #78
  • add models from Jia2022LAS-AT by @fra31 in #80
  • remove unnecessary file when loading custom ImageNet by @CNOCycle in #75
  • update info of Sridhar2021Robust models by @fra31 in #82
  • ImageNet-3DCC and corruption updates by @ofkar in #85
  • add models from Addepalli2022Efficient by @fra31 in #91
  • throw error when too many images are requested by @fra31 in #96
  • Add Debenedetti2022Light and support for timm by @dedeswim in #100

New Contributors

Full Changelog: v1.0...v1.1

v1.0 - Code corresponding to NeurIPS'21 Datasets and Benchmarks version of the whitepaper

29 Nov 14:10
d3795a0
Compare
Choose a tag to compare

Updates:

  • New ImageNet leaderboards (Linf and common corruption) and relevant benchmarking functions (including ImageNet evaluation on a fixed subset of 10% of the test set).
  • New models and leaderboard entries (now in total: 120+ evaluations, 80+ models).

v0.2.1 - Various fixes

07 May 14:46
4af5687
Compare
Choose a tag to compare

This minor release improves some internals and fixes some bugs in the model zoo.

Internals improvements:

  • Now, when no normalization is applied, the models in the model zoo are loaded with anonymous lambda functions instead of full-fledged classes to keep the code more concise and cleaner.

Bug fixes

  • The CIFAR-100 version of Heyndrycks2019Using was missing rescaling in [-1, 1], hence leading to poor accuracy.
  • The CIFAR-100 version of Rice2020Overfitting was missing a slightly different step from the other PreActResNets in the forward method of the PreActBlock.

v0.2 - CIFAR-100 and new models

05 May 09:22
Compare
Choose a tag to compare

This release adds the following new leaderboards:

  • CIFAR-100 Linf
  • CIFAR-100-C (common corruptions)

Moreover, it adds to the model zoo:

  • 7 new models for CIFAR-10 Linf (28 in total)
  • 2 new models for CIFAR-10 L2 (11 in total)
  • 10 new models for CIFAR-100 Linf (10 in total)
  • 2 new models for CIFAR-100-C (2 in total)

It also fixes some bugs and improves some internals:

  • The the common corruptions datasets are now downloaded from the original Zenodo repositories, instead of Google Drive
  • The benchmark function now raises a working if it is run on a model which is not in .eval() mode

v0.1

11 Mar 08:29
Compare
Choose a tag to compare

This is the first stable release of RobustBench. It includes the following features:

  • A Model Zoo and a leaderboard containing up-to-date and State of the Art models trained on CIFAR-10 for robustness to L2 and Linf adversarial perturbations, and common corruptions.
  • An updated API that makes it easier to add new datasets and threat models.
  • A function to benchmark the robustness of new models on a given dataset and threat model.
  • Functions to automatically generate the leaderboards of our website, and to generate LaTeX tables.