TResNet

Last updated on Feb 14, 2021

tresnet_l

Parameters 53 Million
FLOPs 11 Billion
File Size 214.04 MB
Training Data <h2>oi</h2>
Training Resources 8x NVIDIA 100 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_l
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 224
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
tresnet_l_448

Parameters 53 Million
FLOPs 43 Billion
File Size 214.04 MB
Training Data <h2>oi</h2>
Training Resources 8x NVIDIA 100 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_l_448
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 448
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
tresnet_m

Parameters 41 Million
FLOPs 6 Billion
File Size 120.03 MB
Training Data <h2>oi</h2>
Training Resources 8x NVIDIA 100 GPUs
Training Time < 24 hours

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_m
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 224
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
tresnet_m_448

Parameters 29 Million
FLOPs 23 Billion
File Size 120.03 MB
Training Data <h2>oi</h2>
Training Resources 8x NVIDIA 100 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_m_448
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 448
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
tresnet_xl

Parameters 76 Million
FLOPs 15 Billion
File Size 299.82 MB
Training Data <h2>oi</h2>
Training Resources 8x NVIDIA 100 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_xl
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 224
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
tresnet_xl_448

Parameters 76 Million
FLOPs 61 Billion
File Size 214.04 MB
Training Data <h2>oi</h2>
Training Resources 8x NVIDIA 100 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_xl_448
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 448
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
README.md

Summary

A TResNet is a variant on a ResNet that aim to boost accuracy while maintaining GPU training and inference efficiency. They contain several design tricks including a SpaceToDepth stem, Anti-Alias downsampling, In-Place Activated BatchNorm, Blocks selection and squeeze-and-excitation layers.

How do I load this model?

To load a pretrained model:

import timm
m = timm.create_model('tresnet_m', pretrained=True)
m.eval()

Replace the model name with the variant you want to use, e.g. tresnet_m. You can find the IDs in the model summaries at the top of this page.

How do I train this model?

You can follow the timm recipe scripts for training a new model afresh.

Citation

@misc{ridnik2020tresnet,
      title={TResNet: High Performance GPU-Dedicated Architecture}, 
      author={Tal Ridnik and Hussam Lawen and Asaf Noy and Emanuel Ben Baruch and Gilad Sharir and Itamar Friedman},
      year={2020},
      eprint={2003.13630},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Results

Image Classification on ImageNet

Image Classification on ImageNet
MODEL TOP 1 ACCURACY TOP 5 ACCURACY
tresnet_xl_448 83.06% 96.19%
tresnet_l_448 82.26% 95.98%
tresnet_xl 82.05% 95.93%
tresnet_m_448 81.72% 95.57%
tresnet_l 81.49% 95.62%
tresnet_m 80.8% 94.86%