作者:Xingyi Zhou, Tianwei Yin,Vladlen Koltun,Philipp Krahenb ¨ uhl

作者单位:The University of Texas at Austin,Apple

发布时间:2022

发布期刊/会议:CVPR

出版商:IEEE

论文全称:Global Tracking Transformers

论文地址:

CVPR 2022 Open Access Repository

论文代码:

https://github.com/xingyizhou/GTR

地位:

个人理解

一、摘要

We present a novel transformer-based architecture for global multi-object tracking. Our network takes a short sequence of frames as input and produces global trajectories for all objects. The core component is a global tracking transformer that operates on objects from all frames in the sequence. The transformer encodes object features from all frames, and uses trajectory queries to group them into trajectories. The trajectory queries are object features from a single frame and naturally produce unique trajectories. Our global tracking transformer does not require intermediate pairwise grouping or combinatorial association, and can be jointly trained with an object detector. It achieves competitive performance on the popular MOT17 benchmark, with 75.3 MOTA and 59.1 HOTA. More importantly, our framework seamlessly integrates into stateof-the-art large-vocabulary detectors to track any objects. Experiments on the challenging TAO dataset show that our framework consistently improves upon baselines that are based on pairwise association, outperforming published work by a significant 7.7 tracking mAP.