DJW c16313bb6a 第一次提交 | преди 9 месеца | |
---|---|---|
.. | ||
README.md | преди 9 месеца | |
cascade-mask-rcnn_r50-rsb-pre_fpn_1x_coco.py | преди 9 месеца | |
faster-rcnn_r50-rsb-pre_fpn_1x_coco.py | преди 9 месеца | |
mask-rcnn_r50-rsb-pre_fpn_1x_coco.py | преди 9 месеца | |
metafile.yml | преди 9 месеца | |
retinanet_r50-rsb-pre_fpn_1x_coco.py | преди 9 месеца |
The influential Residual Networks designed by He et al. remain the gold-standard architecture in numerous scientific publications. They typically serve as the default architecture in studies, or as baselines when new architectures are proposed. Yet there has been significant progress on best practices for training neural networks since the inception of the ResNet architecture in 2015. Novel optimization & dataaugmentation have increased the effectiveness of the training recipes.
In this paper, we re-evaluate the performance of the vanilla ResNet-50 when trained with a procedure that integrates such advances. We share competitive training settings and pre-trained models in the timm open-source library, with the hope that they will serve as better baselines for future work. For instance, with our more demanding training setting, a vanilla ResNet-50 reaches 80.4% top-1 accuracy at resolution 224×224 on ImageNet-val without extra data or distillation. We also report the performance achieved with popular models with our training procedure.
Method | Backbone | Lr schd | Mem (GB) | Inf time (fps) | box AP | mask AP | Config | Download |
---|---|---|---|---|---|---|---|---|
Faster R-CNN | R-50 rsb | 1x | 3.9 | - | 40.8 (+3.4) | - | Config | model | log |
Mask R-CNN | R-50 rsb | 1x | 4.5 | - | 41.2 (+3.0) | 38.2 (+3.0) | Config | model | log |
Cascade Mask R-CNN | R-50 rsb | 1x | 6.2 | - | 44.8 (+3.6) | 39.9 (+3.6) | Config | model | log |
RetinaNet | R-50 rsb | 1x | 3.8 | - | 39.0 (+2.5) | - | Config | model | log |
Notes:
@article{wightman2021resnet,
title={Resnet strikes back: An improved training procedure in timm},
author={Ross Wightman, Hugo Touvron, Hervé Jégou},
journal={arXiv preprint arXiv:2110.00476},
year={2021}
}