Skip to content

We propose a theoretically motivated method, Adversarial Training with informative Outlier Mining (ATOM), which improves the robustness of OOD detection to various types of adversarial OOD inputs and establishes state-of-the-art performance.

License

jfc43/informative-outlier-mining

Repository files navigation

ATOM: Robustifying Out-of-distribution Detection Using Outlier Mining

This project is for the paper: ATOM: Robustifying Out-of-distribution Detection Using Outlier Mining.

Illustration

ATOM banner

Experimental Results

Main Results

Preliminaries

It is tested under Ubuntu Linux 16.04.1 and Python 3.6 environment, and requires some packages to be installed:

Downloading In-distribution Dataset

  • SVHN: need to download Format 2 data.
  • CIFAR: included in PyTorch.

Downloading Auxiliary OOD Datasets

cd datasets/unlabeled_datasets/80M_Tiny_Images
wget http://horatio.cs.nyu.edu/mit/tiny/data/tiny_images.bin

Downloading Out-of-distribution Test Datasets

We provide links and instructions to download each dataset:

  • SVHN: download it and place it in the folder of datasets/ood_datasets/svhn. Then run python select_svhn_data.py to generate test subset.
  • Textures: download it and place it in the folder of datasets/ood_datasets/dtd.
  • Places365: download it and place it in the folder of datasets/ood_datasets/places365/test_subset. We randomly sample 10,000 images from the original test dataset. We provide the file names for the images that we sample in datasets/ood_datasets/places365/test_subset/places365_test_list.txt.
  • LSUN-C: download it and place it in the folder of datasets/ood_datasets/LSUN.
  • LSUN-R: download it and place it in the folder of datasets/ood_datasets/LSUN_resize.
  • iSUN: download it and place it in the folder of datasets/ood_datasets/iSUN.

For example, run the following commands in the root directory to download LSUN-C:

cd datasets/ood_datasets
wget https://www.dropbox.com/s/fhtsw1m3qxlwj6h/LSUN.tar.gz
tar -xvzf LSUN.tar.gz

Downloading Pre-trained Models

We provide pre-trained DenseNet models for vanilla, SOFL, OE, ACET, CCU, ROWL, NTOM, and ATOM methods on SVHN, CIFAR-10, and CIFAR-100 datasets using TinyImages as auxiliary OOD dataset. We also provide pre-trained WideResNet models for vanilla, NTOM, and ATOM methods on SVHN, CIFAR-10, and CIFAR-100 datasets using TinyImages as auxiliary OOD dataset. They can be downloaded from Google Drive.

Overview of the Code

Running Experiments

  • select_svhn_data.py: select SVHN test data.
  • eval_ood_detection.py: evaluate OOD detection performance of models.
  • compute_metrics.py: compute metrics of evaluation results.
  • gen_rowl_train_data.py: generate ROWL training data.
  • gen_validation_data.py: generate validation data used to select the best q.
  • train_acet.py: train ACET model.
  • train_atom.py: train ATOM model.
  • train_ccu.py: train CCU model.
  • train_oe.py: train OE model.
  • train_rowl.py: train ROWL model.
  • train_sofl.py: train SOFL model.
  • train.py: train vanilla model.
  • tune_mahalanobis_hyperparams.py: tune hyperparameters of Mahalanobis detector.
  • tune_odin_hyperparams.py: tune hyperparameters of ODIN detector.

Example

For CIFAR-10 experiments, you can run the following script to get the results:

./run_all_cifar10_exps.sh

To evaluate an OOD detection method, you can use the following command:

python eval_ood_detection.py --in-dataset {in-distribution dataset} --name {model name} --method {scoring function} [--adv or --corrupt or --adv-corrupt]

The argument --adv is to evaluate L_inf OOD, --corrupt is to evaluate Corrupted OOD, --adv-corrupt is to evaluate Comp. OOD. You can only specify one of them. Without them, it will evaluate natural OOD.

{in-distribution dataset} can be SVHN, CIFAR-10 or CIFAR-100.

{scoring function} can be msp, odin, mahalanobis, sofl, rowl, ntom or atom.

{model name} is the name of the model that you have trained. The model type should match the corresponding scoring function. See the following table for the matchings:

Model Type Scoring Function
vanilla msp
vanilla odin
vanilla mahalanobis
SOFL sofl
OE msp
ACET msp
CCU msp
ROWL rowl
NTOM ntom
ATOM atom

Acknowledgements

Part of this code is inspired by ODIN, Outlier Exposure, Deep Mahalanobis Detector, Corruption and Perturbation Robustness and Certified Certain Uncertainty.

Citation

Please cite our work if you use the codebase:

@article{chen2021atom,
title={ATOM: Robustifying Out-of-distribution Detection Using Outlier Mining},
author={Chen, Jiefeng and Li, Yixuan and Wu, Xi and Liang, Yingyu and Jha, Somesh},
journal={In Proceedings of European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD)},
year={2021}
}

License

Please refer to the LICENSE.

About

We propose a theoretically motivated method, Adversarial Training with informative Outlier Mining (ATOM), which improves the robustness of OOD detection to various types of adversarial OOD inputs and establishes state-of-the-art performance.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published