File size: 1,346 Bytes
974587a
 
 
 
 
 
 
 
 
 
 
 
 
9aec45b
974587a
 
 
 
 
 
1da6f37
 
eee57ab
 
9aec45b
974587a
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
---
license: mit
datasets:
- ILSVRC/imagenet-1k
tags:
- uncertainty quantification
- model robustness
- selective classification
- label-smoothing
---

[![arXiv](https://img.shields.io/badge/arXiv-2403.14715-b31b1b.svg)](https://arxiv.org/abs/2403.14715)

This repository contains the models trained as experimental support for the paper "Towards Understanding Why Label Smoothing Degrades Selective Classification and How to Fix It" published at ICLR 2025.

The code is based on [TorchUncertainty](https://github.com/ENSTA-U2IS-AI/torch-uncertainty) and available on [GitHub](https://github.com/o-laurent/Label-smoothing-Selective-classification).

## List of models

This repository contains:
- for classification on ImageNet with ViTs: 4 ViTs-S/16 trained with label-smoothing coefficients in [0, 0.1, 0.2, 0.3]
- for classification on ImageNet with ResNets: 4 ResNet-50 trained with label-smoothing coefficients in [0, 0.1, 0.2, 0.3]
- for classification on CIFAR-100: 4 DenseNet-BC trained with label-smoothing coefficients in [0, 0.1, 0.2, 0.3]
- for segmentation: 4 DeepLabv3+ Resnet-101 trained with label-smoothing coefficients in [0, 0.1, 0.2, 0.3]
- for nlp: one CE-based and one LS-based (LS coefficient 0.6) LSTM-MLP

The rest of the models (notably on tabular data) used in the paper are trainable on CPU in the dedicated notebooks.