作者:Fangao ZengBin DongYuang ZhangTiancai WangXiangyu ZhangYichen Wei

**作者单位:**旷视科技,上海交通大学

发布时间:2021

发布期刊/会议:2022ECCV

出版商:

**论文全称:**MOTR: End-to-End Multiple-Object Tracking with Transformer

论文地址:

MOTR: End-to-End Multiple-Object Tracking with Transformer

论文代码:

https://github.com/megvii-research/MOTR

地位:

个人理解

一、摘要

   Temporal modeling of objects is a key challenge in multipleobject tracking (MOT). Existing methods track by associating detections through motion-based and appearance-based similarity heuristics. The post-processing nature of association prevents end-to-end exploitation of temporal variations in video sequence.
   In this paper, we propose MOTR, which extends DETR [6] and introduces “track query” to model the tracked instances in the entire video. Track query is transferred and updated frame-by-frame to perform iterative prediction over time. We propose tracklet-aware label assignment to train track queries and newborn object queries. We further propose temporal aggregation network and collective average loss to enhance temporal relation modeling. Experimental results on DanceTrack show that MOTR significantly outperforms state-of-the-art method, ByteTrack [42] by 6.5% on HOTA metric. On MOT17, MOTR outperforms our concurrent works, TrackFormer [18] and TransTrack [29], on association performance. MOTR can serve as a stronger baseline for future research on temporal modeling and Transformer-based trackers.

二、受DETR启发

Untitled