We call our approach seamless because it neither presents special requirements (e.g., hyperparameter tuning, additional data) nor produces side effects (e.g., inefficient inference or detection, classification accuracy drop).
Our approach consists of a loss that works as a drop-in replacement to the SoftMax loss (i.e., the combination of the output linear layer, the SoftMax activation, and the cross-entropy loss). The out-of-distribution detection is performed using a zero computational cost score.
Besides all the above, the IsoMax+ loss (the most recent version) produces state-of-the-art out-of-distribution detection performance.
class Model(nn.Module):
def __init__(self):
(...)
#self.classifier = nn.Linear(num_features, num_classes)
self.classifier = losses.IsoMaxPlusLossFirstPart(num_features, num_classes)
#criterion = nn.CrossEntropyLoss()
criterion = losses.IsoMaxPlusLossSecondPart()
outputs = model(inputs)
# outputs are equal to logits, which in turn are equivalent to negative distances
score = outputs.max(dim=1)[0] # this is the minimum distance score for detection
# the minimum distance score is the best option for the IsoMax+ loss
python example.py
Much code reused from "deep_Mahalanobis_detector" and odin-pytorch.
pip install -r requirements.txt
./prepare-cifar.sh # downloads out-of-distrbution data for CIFAR datasets.
./run_cifar10.sh
./run_cifar100.sh
./run_svhn.sh
./analyze.sh
The paper "Entropic Out-of-Distribution Detection" (arXiv) (IJCNN 2021) proposes the IsoMax loss, which works as a drop-in replacement of the SoftMax loss. The solution is seamless. Hence, it does not produce classification accuracy drop while significantly improves the OOD detection performance. Additionally, unlike other OOD detection approaches, IsoMax loss delivers inferences as fast and energy-efficient as SoftMax loss-trained neural networks. Our solution does not require hyperparameter tuning. No additional data are needed. Finally, the out-of-distribution detection is performed using the speedy entropic score.
The paper "Entropic Out-of-Distribution Detection: Seamless Detection of Unknown Examples" (arXiv) (IEEE TNNLS 2021) is the journal version of the paper "Entropic Out-of-Distribution Detection". It presents additional explanations, experiments, and analyses regarding the IsoMax loss.
Enhanced Isotropy Maximization Loss: Seamless and High-Performance Out-of-Distribution Detection Simply Replacing the SoftMax Loss
The paper "Enhanced Isotropy Maximization Loss: Seamless and High-Performance Out-of-Distribution Detection Simply Replacing the SoftMax Loss" (arXiv) proposes the Enhanced IsoMax (IsoMax+) loss that significantly improved the out-of-distribution detection performance relative to IsoMax loss while also keeping the solution seamless. The minimum distance score is used rather than the entropic score. Besides being seamless and easy to implement and use, this solution produces competitive state-of-the-art out-of-distribution detection results.
From now on, use the IsoMax+ loss combined with the minimum distance score rather than the IsoMax loss combined with the entropic score.
Please, cite our papers if you use our losses in your works:
@INPROCEEDINGS{9533899,
author={Macêdo, David and Ren, Tsang Ing and Zanchettin, Cleber and Oliveira,
Adriano L. I. and Ludermir, Teresa},
booktitle={2021 International Joint Conference on Neural Networks (IJCNN)},
title={Entropic Out-of-Distribution Detection},
year={2021},
pages={1-8},
doi={10.1109/IJCNN52387.2021.9533899}
}
@ARTICLE{9556483,
author={Macêdo, David and Ren, Tsang Ing and Zanchettin, Cleber and Oliveira,
Adriano L. I. and Ludermir, Teresa},
journal={IEEE Transactions on Neural Networks and Learning Systems},
title={Entropic Out-of-Distribution Detection: Seamless Detection of Unknown Examples},
year={2022},
volume={33},
number={6},
pages={2350-2364},
doi={10.1109/TNNLS.2021.3112897}
}
@article{DBLP:journals/corr/abs-2105-14399,
author = {David Mac{\^{e}}do and
Teresa Bernarda Ludermir},
title = {Enhanced Isotropy Maximization Loss:
Seamless and High-Performance Out-of-Distribution Detection
Simply Replacing the SoftMax Loss},
journal = {CoRR},
volume = {abs/2105.14399},
year = {2021},
url = {https://arxiv.org/abs/2105.14399},
eprinttype = {arXiv},
eprint = {2105.14399},
timestamp = {Wed, 02 Jun 2021 11:46:42 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-2105-14399.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
@article{DBLP:journals/corr/abs-2208-03566,
author = {David Mac{\^{e}}do},
title = {Towards Robust Deep Learning using Entropic Losses},
journal = {CoRR},
volume = {abs/2208.03566},
year = {2022},
url = {https://doi.org/10.48550/arXiv.2208.03566},
doi = {10.48550/arXiv.2208.03566},
eprinttype = {arXiv},
eprint = {2208.03566},
timestamp = {Wed, 10 Aug 2022 14:49:54 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-2208-03566.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}