- Notifications
You must be signed in to change notification settings - Fork10
Implementation of Tracking Every Thing in the Wild, ECCV 2022
License
SysCV/tet
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
This is the offical implementation of paperTracking Every Thing in the Wild.
Our project website contains more information:vis.xyz/pub/tet.
Current multi-category Multiple Object Tracking (MOT) metrics use class labels to group tracking results for per-class evaluation. Similarly, MOT methods typically only associate objects with the same class predictions.These two prevalent strategies in MOT implicitly assume that the classification performance is near-perfect.However, this is far from the case in recent large-scale MOT datasets, which contain large numbers of classes with many rare or semantically similar categories. Therefore, the resulting inaccurate classification leads to sub-optimal tracking and inadequate benchmarking of trackers.We address these issues by disentangling classification from tracking.We introduce a new metric, Track Every Thing Accuracy (TETA), breaking tracking measurement into three sub-factors: localization, association, and classification, allowing comprehensive benchmarking of tracking performance even under inaccurate classification. TETA also deals with the challenging incomplete annotation problem in large-scale tracking datasets. We further introduce a Track Every Thing tracker (TETer), that performs association using Class Exemplar Matching (CEM). Our experiments show that TETA evaluates trackers more comprehensively, and TETer achieves significant improvements on the challenging large-scale datasets BDD100K and TAO compared to the state-of-the-art.
New Repo: we update the TETA metric to a new repo:TETA. It is more user-friendly and has more features.
The new repo directly supports evaluation forTAO TETA benchmark
,Open-vocabulary MOT benchmark
,BDD100K MOT benchmark
, andBDD100K MOTS benchmark
.Please refer to the new repo for the latest updates.
TETA builds upon the HOTA metric, while extending it to better deal withmultiple categories and incomplete annotations. TETA evaluate trackers based on a novel local cluster design. TETA consists of three parts: alocalization score, an association score, and a classification score, which enableus to evaluate the different aspects of each tracker properly.
Install TETA metric:
python -m pip install git+https://github.com/SysCV/tet.git/#subdirectory=teta
TETer follows an Associate-Every-Thing (AET) strategy.Instead of only associating objects in the same class, we associate every object in neighboring frames.We introduce Class Exemplar Matching (CEM), where the learned class exemplars incorporate valuable class information in a soft manner.In this way, we effectively exploit semantic supervision on large-scale detection datasets while not relying on the often incorrect classification output.
Our method outperforms the states of the art on BDD100K, and TAO benchmarks.
Method | backbone | pretrain | mMOTA | mIDF1 | TETA | LocA | AssocA | ClsA | IDF1 | MOTA | config | model |
---|---|---|---|---|---|---|---|---|---|---|---|---|
QDTrack(CVPR21) | ResNet-50 | ImageNet-1K | 36.6 | 51.6 | 47.8 | 45.9 | 48.5 | 49.2 | 71.5 | 63.5 | - | - |
TETer (Ours) | ResNet-50 | ImageNet-1K | 39.1 | 53.3 | 50.8 | 47.2 | 52.9 | 52.4 | 71.1 | 64.4 | cfg | google drive |
Method | backbone | pretrain | mMOTA | mIDF1 | TETA | LocA | AssocA | ClsA | IDF1 | MOTA | config | model |
---|---|---|---|---|---|---|---|---|---|---|---|---|
QDTrack(CVPR21) | ResNet-50 | ImageNet-1K | 35.7 | 52.3 | 49.2 | 47.2 | 50.9 | 49.2 | 72.3 | 64.3 | - | - |
TETer (Ours) | ResNet-50 | ImageNet-1K | 37.4 | 53.3 | 50.8 | 47.0 | 53.6 | 50.7 | 71.8 | 64.8 | cfg | google drive |
Method | backbone | pretrain | TETA | LocA | AssocA | ClsA | config | model | preds |
---|---|---|---|---|---|---|---|---|---|
QDTrack(CVPR21) | ResNet-101 | ImageNet-1K | 30.0 | 50.5 | 27.4 | 12.1 | - | - | - |
TETer (Ours) | ResNet-101 | ImageNet-1K | 33.3 | 51.6 | 35.0 | 13.2 | cfg | google drive | coming soon |
TETer-HTC (Ours) | ResNeXt-101-64x4d | ImageNet-1K | 36.9 | 57.5 | 37.5 | 15.7 | - | - | - |
TETer-SwinT (Ours) | SwinT | ImageNet-1K | 34.6 | 52.1 | 36.7 | 15.0 | cfg | google drive | coming soon |
TETer-SwinS (Ours) | SwinS | ImageNet-1K | 36.7 | 54.2 | 38.4 | 17.4 | cfg | google drive | coming soon |
TETer-SwinB (Ours) | SwinB | ImageNet-22K | 38.8 | 55.6 | 40.1 | 20.8 | cfg | google drive | coming soon |
TETer-SwinL (Ours) | SwinL | ImageNet-22K | 40.1 | 56.3 | 39.9 | 24.1 | cfg | google drive | google drive |
Please refer toINSTALL.md for installation instructions.
Please refer toGET_STARTED.md for dataset preparation and running instructions.
@InProceedings{trackeverything, title = {Tracking Every Thing in the Wild}, author = {Li, Siyuan and Danelljan, Martin and Ding, Henghui and Huang, Thomas E. and Yu, Fisher}, booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)}, month = {Oct}, year = {2022}}