<aid="ft1">1</a>: Here we comapre models that have identical training & inference cost between the two implementation. However their numbers are different due to many small implementation details.
<aid="ft2">2</a>: Numbers taken from [Group Normalization](https://arxiv.org/abs/1803.08494)
<aid="ft3">3</a>: Our mAP is __10+ point__ better than the official model in [matterport/Mask_RCNN](https://github.com/matterport/Mask_RCNN/releases) with the same R101-FPN backbone.
## Notes
[NOTES.md](NOTES.md) has some notes about implementation details & speed.