# TResNet
**Repository Path**: qianyiz/TResNet
## Basic Information
- **Project Name**: TResNet
- **Description**: No description available
- **Primary Language**: Unknown
- **License**: Apache-2.0
- **Default Branch**: master
- **Homepage**: None
- **GVP Project**: No
## Statistics
- **Stars**: 0
- **Forks**: 0
- **Created**: 2021-03-03
- **Last Updated**: 2021-03-03
## Categories & Tags
**Categories**: Uncategorized
**Tags**: None
## README
# TResNet: High Performance GPU-Dedicated Architecture
[](https://paperswithcode.com/sota/image-classification-on-cifar-10?p=tresnet-high-performance-gpu-dedicated)
[](https://paperswithcode.com/sota/image-classification-on-cifar-100?p=tresnet-high-performance-gpu-dedicated)
[](https://paperswithcode.com/sota/fine-grained-image-classification-on-stanford?p=tresnet-high-performance-gpu-dedicated)
[](https://paperswithcode.com/sota/fine-grained-image-classification-on-oxford?p=tresnet-high-performance-gpu-dedicated)
[](https://paperswithcode.com/sota/multi-label-classification-on-ms-coco?p=asymmetric-loss-for-multi-label)
[](https://paperswithcode.com/sota/multi-label-classification-on-nus-wide?p=asymmetric-loss-for-multi-label)
[](https://paperswithcode.com/sota/multi-label-classification-on-pascal-voc-2007?p=asymmetric-loss-for-multi-label)
[paperV2](https://arxiv.org/pdf/2003.13630.pdf) |
[pretrained models](MODEL_ZOO.md)
Official PyTorch Implementation
> Tal Ridnik, Hussam Lawen, Asaf Noy, Itamar Friedman, Emanuel Ben Baruch, Gilad Sharir
> DAMO Academy, Alibaba Group
**Abstract**
> Many deep learning models, developed in recent years, reach higher
> ImageNet accuracy than ResNet50, with fewer or comparable FLOPS count.
> While FLOPs are often seen as a proxy for network efficiency, when
> measuring actual GPU training and inference throughput, vanilla
> ResNet50 is usually significantly faster than its recent competitors,
> offering better throughput-accuracy trade-off. In this work, we
> introduce a series of architecture modifications that aim to boost
> neural networks' accuracy, while retaining their GPU training and
> inference efficiency. We first demonstrate and discuss the bottlenecks
> induced by FLOPs-optimizations. We then suggest alternative designs
> that better utilize GPU structure and assets. Finally, we introduce a
> new family of GPU-dedicated models, called TResNet, which achieve
> better accuracy and efficiency than previous ConvNets. Using a TResNet
> model, with similar GPU throughput to ResNet50, we reach 80.7\%
> top-1 accuracy on ImageNet. Our TResNet models also transfer well and
> achieve state-of-the-art accuracy on competitive datasets such as
> Stanford cars (96.0\%), CIFAR-10 (99.0\%), CIFAR-100 (91.5\%) and
> Oxford-Flowers (99.1\%). They also perform well on multi-label classification and object detection tasks.
## 07/01/2021: New SOTA Results
With better pretraining, Our medium-size newtork, TResNet-L-V2, achieved
SOTA result on stanford-cars dataset, and 3rd place on CIFAR100, while
being significantly faster and smaller than the competitors.
## 30/9/2020: New Paper Released
We released a new paper, [Asymmetric Loss For Multi-Label
Classification](https://arxiv.org/abs/2009.14119).
The paper
presents a new novel loss function, that is applicable for tasks with
inherent data imbalancing - first and foremost multi-label
classification, and also object detection and fine-grain single-label
classification. The new paper also describe and expand in details
preliminary results provided in TResNet paper on MS-COCO multi-label
dataset.
[Github link](https://github.com/Alibaba-MIIL/ASL)
## 28/8/2020: V2 of TResNet Article Released
## Sotabench Comparisons
Comparative results from
[sotabench benchamrk](https://sotabench.com/benchmarks/image-classification-on-imagenet#code),
demonstartaing that TReNset models give excellent speed-accuracy tradoff:
![]() |
| Bacbkone | mAP |
|---|---|
| KSSNet (previous SOTA) | 83.7 |
| TResNet-L | 86.4 |
![]() |
| Models | Top Training Speed (img/sec) |
Top Inference Speed (img/sec) |
Max Train Batch Size | Top-1 Acc. |
|---|---|---|---|---|
| ResNet50 | 805 | 2830 | 288 | 79.0 |
| EfficientNetB1 | 440 | 2740 | 196 | 79.2 |
| TResNet-M | 730 | 2930 | 512 | 80.8 |
| TResNet-L | 345 | 1390 | 316 | 81.5 |
| TResNet-XL | 250 | 1060 | 240 | 82.0 |
| Model | Top Training Speed (img/sec) |
Top Inference Speed (img/sec) |
Top-1 Acc. | Flops[G] |
|---|---|---|---|---|
| ResNet50 | 805 | 2830 | 79.0 | 4.1 |
| ResNet50-D | 600 | 2670 | 79.3 | 4.4 |
| ResNeXt50 | 490 | 1940 | 79.4 | 4.3 |
| EfficientNetB1 | 440 | 2740 | 79.2 | 0.6 |
| SEResNeXt50 | 400 | 1770 | 79.9 | 4.3 |
| MixNet-L | 400 | 1400 | 79.0 | 0.5 |
| TResNet-M | 730 | 2930 | 80.8 | 5.5 |
![]() |
![]() |
| Dataset | Model |
Top-1
Acc. |
Speed
img/sec |
Input |
| CIFAR-10 | Gpipe | 99.0 | - | 480 |
| TResNet-XL | 99.0 | 1060 | 224 | |
| CIFAR-100 | EfficientNet-B7 | 91.7 | 70 | 600 |
| TResNet-XL | 91.5 | 1060 | 224 | |
| Stanford Cars | EfficientNet-B7 | 94.7 | 70 | 600 |
| TResNet-L | 96.0 | 500 | 368 | |
| Oxford-Flowers | EfficientNet-B7 | 98.8 | 70 | 600 |
| TResNet-L | 99.1 | 500 | 368 |