28 Jan 2019 | Dan Hendrycks, Mantas Mazeika, Thomas Dietterich
The paper "Deep Anomaly Detection with Outlier Exposure" by Dan Hendrycks, Mantas Mazeika, and Thomas Dietterich proposes a method called Outlier Exposure (OE) to improve the detection of anomalous inputs in machine learning systems. The authors argue that larger and more complex inputs in deep learning make it more challenging to distinguish between anomalous and in-distribution examples. To address this, OE leverages diverse and large datasets of out-of-distribution (OOD) examples to train anomaly detectors, enabling them to generalize and detect unseen anomalies.
The method involves training anomaly detectors against an auxiliary dataset of outliers, which helps the detectors learn heuristics for detecting novel forms of anomalies. Extensive experiments on natural language processing and vision tasks demonstrate that OE significantly improves detection performance. The authors also show that cutting-edge generative models trained on CIFAR-10 may assign higher likelihoods to SVHN images than to CIFAR-10 images, and OE can mitigate this issue.
The paper evaluates OE's flexibility and robustness, identifying characteristics of the auxiliary dataset that enhance performance. It also discusses the method's applicability to different types of data and tasks, including multiclass classification, confidence branch experiments, and density estimation. The results show that OE can improve calibration and enhance existing anomaly detection techniques. Overall, OE is presented as a simple and effective approach to improving out-of-distribution detection in deep learning systems.The paper "Deep Anomaly Detection with Outlier Exposure" by Dan Hendrycks, Mantas Mazeika, and Thomas Dietterich proposes a method called Outlier Exposure (OE) to improve the detection of anomalous inputs in machine learning systems. The authors argue that larger and more complex inputs in deep learning make it more challenging to distinguish between anomalous and in-distribution examples. To address this, OE leverages diverse and large datasets of out-of-distribution (OOD) examples to train anomaly detectors, enabling them to generalize and detect unseen anomalies.
The method involves training anomaly detectors against an auxiliary dataset of outliers, which helps the detectors learn heuristics for detecting novel forms of anomalies. Extensive experiments on natural language processing and vision tasks demonstrate that OE significantly improves detection performance. The authors also show that cutting-edge generative models trained on CIFAR-10 may assign higher likelihoods to SVHN images than to CIFAR-10 images, and OE can mitigate this issue.
The paper evaluates OE's flexibility and robustness, identifying characteristics of the auxiliary dataset that enhance performance. It also discusses the method's applicability to different types of data and tasks, including multiclass classification, confidence branch experiments, and density estimation. The results show that OE can improve calibration and enhance existing anomaly detection techniques. Overall, OE is presented as a simple and effective approach to improving out-of-distribution detection in deep learning systems.