作者:Ruopeng Gao,Limin Wang

作者单位:南京大学新型软件技术国家重点实验室,上海人工智能实验

发布时间:2023

发布期刊/会议:Arxiv

出版商:

**论文全称:**MeMOTR: Long-Term Memory-Augmented Transformer for Multi-Object Tracking

论文地址:

MeMOTR: Long-Term Memory-Augmented Transformer for Multi-Object Tracking

论文代码:

https://github.com/MCG-NJU/MeMOTR

地位:

个人理解

一、摘要

As a video task, Multi-Object Tracking (MOT) is expected to capture temporal information of targets effectively. Unfortunately, most existing methods only explicitly exploit the object features between adjacent frames, while lacking the capacity to model long-term temporal information. In this paper, we propose MeMOTR, a long-term memory-augmented Transformer for multi-object tracking. Our method is able to make the same object’s track embedding more stable and distinguishable by leveraging long term memory injection with a customized memory-attention layer. This significantly improves the target association ability of our model. Experimental results on DanceTrack show that MeMOTR impressively surpasses the state-of-theart method by 7.9% and 13.0% on HOTA and AssA metrics, respectively. Furthermore, our model also outperforms other Transformer-based methods on association performance on MOT17 and generalizes well on BDD100K. Code is available at https://github.com/MCG-NJU/MeMOTR

二、MeMOTR