Skip to content

Code for the paper "A Geometric Explanation of the Likelihood OOD Detection Paradox"

License

Notifications You must be signed in to change notification settings

layer6ai-labs/dgm_ood_detection

Repository files navigation

OOD Detection with Likelihood-based Deep Generative Models

The codebase used to produce the results for "A Geometric Explanation of the Likelihood OOD Detection Paradox", accepted to ICML 2024.

News: We have revamped the codebase in our new website

Explanation of OOD failure

Overview

In many likelihood-based generative models, the model specifies a density function over all the possible datapoints, and the training process typically aims to increase the density for training (in-distribution) data. We know that densities integrate to one over all the possible datapoints, and as a consequence, one can reasonably assume that this would also decrease the density regions that are far apart from the training data, i.e., out-of-distribution (OOD) regions. Paradoxically, based on research first presented in Nalisnick et al.'s work titled "Do deep generative models know what they don't know?", likelihood values (or probability densities) alone are not a reliable indicator for whether a datapoint is OOD or not, and in many cases, the OOD data consistently gets higher likelihoods assigned (see the figure above).

On the flip side, the generative models that exhibit such pathological behavior are simultaneously capable of generating high-quality in-distribution data. Thus, the information required for OOD detection likely exists within these models, it just might not be the likelihood values alone. One important observation is that since OOD data is never generated from a generative model, the area around them (i.e., local probability mass around the OOD region) should have densities that integrate to a small probability. Thus using local probability masses instead of using mere probability densities seems like a more reasonable approach for OOD detection using generative models. To unravel this, we explore methods that take the entire high-dimensional density landscape induced by the generative model into consideration for OOD detection.

This repository contains our ideas to tackle this problem alongside implementation of some recent relevant OOD detection baselines on a large set of models and datasets. The methods that are included are:

  1. Likelihood ratio method for normalizing flows by Ren et al.
  2. Complexity method for flow models by Serrà et al.
  3. Likelihood ratio method for diffusion models by Goodier et al.
  4. Reconstruction-based OOD detection with diffusions by Graham et al..

Setup

Environment

The codebase uses functional autodiff features that were recently added to pytorch. Make sure that your Python is 3.9 or higher; otherwise, some of the new autodiff functionalities we use might break. You should also install the nflows package from here which enables functional autodiff for the flow models such as the neural spline flows that we are using.

Create the following conda environment to handle dependencies.

# This creates an environment called 'ood-detection':
conda env create -f env.yml 
# You can activate it as follows:
conda activate ood-detection

Reproducibility Statement

We have an old guide for reproducing our results. However, we have refactored the entire codebase and now included it in our new repository. We recommend using that version of our codebase in case you want to run our OOD detection approach.

About

Code for the paper "A Geometric Explanation of the Likelihood OOD Detection Paradox"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published