Packaged TResNet based on Official PyTorch Implementation [paper] [github]
Install with pip:
pip install torch_tresnet
or directly:
pip install git+https://github.com/tczhangzhi/torch-tresnet
Follow the grammatical conventions of torchvision
from torch_tresnet import tresnet_m, tresnet_l, tresnet_xl, tresnet_m_448, tresnet_l_448, tresnet_xl_448
# pretrianed on 224*224
model = tresnet_m(pretrained=True)
model = tresnet_m(pretrained=True, num_classes=10)
model = tresnet_m(pretrained=True, num_classes=10, in_chans=3)
# pretrianed on 448*448
model = tresnet_m_448(pretrained=True)
model = tresnet_m_448(pretrained=True, num_classes=10)
model = tresnet_m_448(pretrained=True, num_classes=10, in_chans=3)
For a demonstration of training and testing, please see the jupyter notebook, which is an unofficial implementation, please do not use this implementation as the final evaluation standard
TResNet models accuracy and GPU throughput on ImageNet, compared to ResNet50. All measurements were done on Nvidia V100 GPU, with mixed precision. All models are trained on input resolution of 224.
Models | Top Training Speed (img/sec) | Top Inference Speed (img/sec) | Max Train Batch Size | Top-1 Acc. |
---|---|---|---|---|
ResNet50 | 805 | 2830 | 288 | 79.0 |
EfficientNetB1 | 440 | 2740 | 196 | 79.2 |
TResNet-M | 730 | 2930 | 512 | 80.7 |
TResNet-L | 345 | 1390 | 316 | 81.4 |
TResNet-XL | 250 | 1060 | 240 | 82.0 |
Comparison of ResNet50 to top modern networks, with similar top-1 ImageNet accuracy. All measurements were done on Nvidia V100 GPU with mixed precision. For gaining optimal speeds, training and inference were measured on 90% of maximal possible batch size. Except TResNet-M, all the models' ImageNet scores were taken from the public repository, which specialized in providing top implementations for modern networks. Except EfficientNet-B1, which has input resolution of 240, all other models have input resolution of 224.
Model | Top Training Speed (img/sec) | Top Inference Speed (img/sec) | Top-1 Acc. | Flops[G] |
---|---|---|---|---|
ResNet50 | 805 | 2830 | 79.0 | 4.1 |
ResNet50-D | 600 | 2670 | 79.3 | 4.4 |
ResNeXt50 | 490 | 1940 | 78.5 | 4.3 |
EfficientNetB1 | 440 | 2740 | 79.2 | 0.6 |
SEResNeXt50 | 400 | 1770 | 79.0 | 4.3 |
MixNet-L | 400 | 1400 | 79.0 | 0.5 |
TResNet-M | 730 | 2930 | 80.7 | 5.5 |
Comparison of TResNet to state-of-the-art models on transfer learning datasets (only ImageNet-based transfer learning results). Models inference speed is measured on a mixed precision V100 GPU. Since no official implementation of Gpipe was provided, its inference speed is unknown.
Dataset | Model | Top-1 Acc. | Speed img/sec | Input |
---|---|---|---|---|
CIFAR-10 | Gpipe | 99.0 | - | 480 |
CIFAR-10 | TResNet-XL | 99.0 | 1060 | 224 |
CIFAR-100 | EfficientNet-B7 | 91.7 | 70 | 600 |
CIFAR-100 | TResNet-XL | 91.5 | 1060 | 224 |
Stanford Cars | EfficientNet-B7 | 94.7 | 70 | 600 |
Stanford Cars | TResNet-L | 96.0 | 500 | 368 |
Oxford-Flowers | EfficientNet-B7 | 98.8 | 70 | 600 |
Oxford-Flowers | TResNet-L | 99.1 | 500 | 368 |