Overview of Deep Learning-Based Visual Multi-Object Tracking: Comparison
Please note this is a comparison between Version 1 by Shichang Wang and Version 4 by Sirius Huang.

Multi-target tracking is an advanced visual work in computer vision, which is essential for understanding the autonomous driving environment. Due to the excellent performance of deep learning in visual object tracking, many state-of-the-art multi-target tracking algorithms have been developed.多目标跟踪是计算机视觉中的高级视觉工作,对于理解自动驾驶环境至关重要。由于深度学习在视觉对象跟踪领域的出色表现,许多一流的多目标跟踪算法已经发展起来。

  • autonomous driving
  • deep learning
  • visual multi-object tracking
  • transformer

1. Introduction

The primary area of intelligent and networked development in the vehicle and transportation industries is autonomous driving. AVs have the potential to fundamentally alter transportation systems by averting deadly crashes, providing critical mobility to the elderly and disabled, increasing road capacity, saving fuel, and lowering emissions [1][2][1,2]. The vehicle perception system’s accurate perception of the environment is essential for safe autonomous driving. The perception of autonomous driving settings depends heavily on object tracking, a high-level vision job in the discipline of computer vision. As a result, the development of an object tracking algorithm ensures the development of an automatic driving system that is both safer and more effective.
This text focuses on multi-object tracking in autonomous driving systems. Multi-object tracking is crucial to ensuring the effectiveness and safety of autonomous driving because it is the fundamental component of the technology. Rarely do objects in traffic situations appear alone. Autonomous driving frequently involves recognizing and tracking many things at once, some of which may be moving in relation to the vehicle or to one another. The majority of techniques in the related literature therefore deal with many objects and attempt to address the multi-object tracking issue. In essence, the MOT algorithm can be summarized as: given the data collected by one or more sensors, how to identify multiple objects in each frame of data and assign an identity to each object, and match those IDs in subsequent data frames [3]. An example of the output of the MOT algorithm is shown in Figure 1 below.
Figure 1. An illustration of the output of an MOT algorithm. Each output bounding box has a number that identifies a specific object in the video.
The research on multi-object tracking for autonomous driving has advanced significantly in recent years, but it is still challenging to use the current multi-object tracking techniques for autonomous driving to their full potential because of issues including the varied shapes of cars and pedestrians in traffic scenes, motion blur, and background interference. There are still several difficulties with the existing visual multi-object tracking technology. Visual multi-object tracking must first address more challenging problems such as: an unpredictable number of objects, frequent object occlusion, challenging object differentiation, etc. In particular, the frequent entry and exit of objects from the field of view is a typical and expected behavior in autonomous driving applications, which results in the uncertainty of the number of objects faced by multi-object tracking and necessitates real-time detection of multi-object tracking algorithms. The method must extract robust object features and keep the object-specific ID after occlusion in complicated situations, since the occlusion of an object by other objects or backgrounds will lead to object ID switches (IDs). The high degree of similarity in object appearance also adds to the difficulty of maintaining the right object ID over the long term. The algorithm must be able to extract the characteristics of comparable items that make them separable. Finally, the challenges that multi-object tracking in autonomous vehicles face can be broken down into two categories: the tracking object factor and the backdrop factor. Shape change, scale change, motion blur, etc. are some of the issues brought on by the object’s factors. The impact of backdrop elements is also substantial, particularly the blurring of background interference, occlusion and disappearance of objects, changes in weather, comparable background interference, etc. [4].

2. Overview of Deep Learning-Based Visual Multi-Object Tracking

Deep learning-based visual multi-object tracking systems have several overview techniques from various angles. The methods for visual multi-object tracking based on deep learning are outlined here in terms of algorithm classification, related data sets, and algorithm assessment.

2.1. Visual Multi-Object Tracking Algorithm Based on Deep Learning

The tracking algorithm based on detection results has evolved and has quickly taken over as the standard framework for multi-object tracking due to the rapid advancement of object detection algorithm performance [5][7]. The TBD sub-modules, such as feature extraction, can be included in the object detection network, though, from the standpoint of the deep neural network’s structure. Joint detection and tracking, or JDT, using a deep network framework to perform visual multi-object tracking, has emerged as a new development trend based on TBD neutron module fusion [6][7][11,12]. The attention mechanism has been incorporated into computer vision systems recently because it has the benefit of efficiently capturing the region of interest in the image, enhancing the performance of the entire network [8][9][10][11][13,14,15,16]. It is used to solve various vision problems, including multi-object tracking. The specific classification structure for the three types of tracking frameworks is shown in Figure 23.
Figure 23. Classification and algorithm of visual multi-object tracking based on deep learning. Overall, the development trend of visual multi-object tracking algorithm is from TBD, to JDT, to Transformer-based tracking algorithm.
At the same time, the characteristics, advantages, and disadvantages of the tracking algorithms of the three types of frameworks are organized as shown in Table 12 [12][13][14][15][16][17][18][19][17,18,19,20,21,22,23,24].
Table 12.
Comparison of characteristics of three types of visual multi-object tracking algorithms.

2.2. MOT Datasets

2.2. MOT 数据集

Deep learning has the benefit over more conventional machine learning techniques in that it can automatically identify data attributes that are pertinent to a specific task. For deep learning-based computer vision algorithms, data sets are crucial. The datasets and traits that are frequently utilized in the field of automatic driving tracking are outlined in the following. Due to their frequent updates and closer resemblance to the actual scene, the 与传统的机器学习技术相比,深度学习的优势在于它可以自动识别与特定任务相关的数据属性。对于基于深度学习的计算机视觉算法,数据集至关重要。下面概述了自动驾驶跟踪领域经常使用的数据集和特征。由于MOT datasets [20][21] raise the most concerns in the field of visual multi-object tracking. On the 数据集[2526]经常更新并且与实际场景更加相似,因此在视觉多目标跟踪领域引起了最大的关注。在MOT dataset, other cutting-edge tracking methods are also tested.数据集上,还测试了其他尖端跟踪方法。 The MOT16数据集[25]专门用于跟踪行人。共有 dataset [20] is used exclusively for tracking pedestrians. There are a total of 14 videos, 个视频、7 practice sets, and 7 test sets. These videos were created using a variety of techniques, including fixed and moving cameras, as well as various shooting perspectives. Additionally, the shooting circumstances vary, depending on whether it is day or night and the weather. The MOT16 detector, called DPM, performs better in the area of detecting pedestrians.个练习集和 7 个测试集。这些视频是使用各种技术创建的,包括固定和移动相机,以及各种拍摄视角。此外,拍摄环境会有所不同,具体取决于是白天还是黑夜以及天气。MOT16探测器称为DPM,在检测行人方面表现更好。 The video content of the MOT17 dataset [20] is the same as that of 数据集[25]的视频内容与MOT16, but it also provides two detector detection results, namely 相同,但它也提供了两个探测器检测结果,即SDP and Faster R-CNN, which have more accurate ground-truth annotations.,它们具有更准确的地面真相注释。 The MOT20 dataset [21] has 数据集[26]有8 video sequences, 个视频序列、4 training sets and 4 testing sets, and the pedestrian density is further increased, with an average of 246 pedestrians per frame.个训练集和4个测试集,行人密度进一步提高,平均每帧246个行人。 The KITTI dataset [22][23] is数据集[2728]是目前用于评估自动驾驶场景中计算机视觉算法的最大数据集。这些数据用于评估 currently the largest dataset for evaluating computer vision algorithms in autonomous driving scenarios. These data are used to evaluate 3D object detection and tracking, visual odometry, evaluation of stereo images, and optical flow images.物体检测和跟踪、视觉里程计、立体图像评估和光流图像。 The NuScenes dataset [24] provides a large dataset of full sensor data for autonomous vehicles, including six cameras, one lidar, five radars, as well as 数据集[29]为自动驾驶汽车提供了一个完整的传感器数据的大型数据集,包括六个摄像头,一个激光雷达,五个雷达,以及GPS and IMU. Compared with the KITTI dataset, it includes more than seven times more object annotations. For each scene, its key frames are selected for annotation, and the annotation rate is 2 Hz. However, it is worth noting that since 23 types of objects are marked, the class- The imbalance problem will be more serious.。与KITTI数据集相比,它包含的对象注释多了七倍以上。对于每个场景,选择其关键帧进行注释,注释速率为 2 Hz。但是,值得注意的是,由于标记了23种类型的对象,因此类不平衡问题将更加严重。 The Waymo dataset [25] is collected with five 数据集[30]是用五个LiDAR和五个高分辨率针孔摄像头收集的。整个数据集包含 and five high-resolution pinhole cameras. The entire dataset contains 1150 scenes, which are divided into 1000 training sets and 150 test sets, with a total of about 12 million1150 个场景,分为 1000 个训练集和 150 个测试集,共有约 1200 万个 LiDAR annotation boxes and approx. 12 million image annotation boxes.标注框和约 1200 万个图像标注框。 The Mapillary Traffic sign dataset [26] is the largest and most diverse traffic sign dataset in the world, which can be used for research on the automatic detection and classification of traffic signs in autonomous driving.交通标志数据集[31]是世界上最大、最多样化的交通标志数据集,可用于自动驾驶中交通标志的自动检测和分类研究。 为了执行视觉多对象跟踪任务,研究人员收集并介绍了表3中列出的数据集。数据收集中的检测和跟踪元素大多与汽车和行人有关,这有助于增强自动驾驶。

To perform visual multi‐object tracking tasks, the datasets are listed in Table 2. Most detection and tracking elements in data collection are related to autos and pedestrians, which helps enhance autonomous driving.

Table 23. Summary of Visual Multi‐object Tracking Datasets.
可视化多对象跟踪数据集摘要。
Ref.裁判。 Datasets数据 Year Feature特征 DOI/URL数字对象标识符/网址
[20][21][27][25,2632 ] MOT15, 16,17,,17,20 2016–2020 Sub datasets containing multiple different camera angles and scenes包含多个不同相机角度和场景的子数据集 https://doi.org/10.48550/arXiv.1504.01942

https://motchallenge.net/

https://doi.org/10.48550/arXiv.1504.01942 https://motchallenge.net/
[22][23][27,28 ] KITTI-Tracking基蒂跟踪 2012 Provides annotations for cars and pedestrians, scene objects are sparse为汽车和行人提供注释,场景对象稀疏 https://doi.org/10.1177/0278364913491297

https://www.cvlibs.net/datasets/kitti/eval_tracking.php

https://doi.org/10.1177/0278364913491297 https://www.cvlibs.net/datasets/kitti/eval_tracking.php
[24][29] NuScenes新场景 2019 Dense traffic and challenging driving conditions交通繁忙,驾驶条件恶劣 https://doi.org/10.48550/arXiv.1903.11027

https://www.nuscenes.org/

https://doi.org/10.48550/arXiv.1903.11027 https://www.nuscenes.org/
[25][30] Waymo韦莫 2020 Diversified driving environment, dense label information多元化的驾驶环境,密集的标签信息 https://doi.org/10.48550/arXiv.1912.04838

https://waymo.com/open/data/motion/tfexample

https://doi.org/10.48550/arXiv.1912.04838 https://waymo.com/open/data/motion/tfexample

2.3. MOT 评估指标

Setting realistic and accurate evaluation metrics is essential for comparing the effectiveness of visual multi-object tracking algorithms in an unbiased and fair manner. The three criteria that make up the multi-object tracking assessment indicators are if the object detection is real-time, whether the predicted position matches the actual position, and whether each object maintains a distinct ID [28]. MOT Challenge offers recognized MOT evaluation metrics.

设置真实准确的评估指标对于以公正和公平的方式比较视觉多目标跟踪算法的有效性至关重要。构成多目标跟踪评估指标的三个标准是,目标检测是否实时,预测位置是否与实际位置匹配,以及每个对象是否保持不同的ID[33]。MOT 挑战赛提供公认的 MOT 评估指标。 MOTA(多目标跟踪精度):多目标跟踪的精度用于计算跟踪中累积的误差,包括跟踪对象的数量以及它们是否匹配:

MOTA (Multi‐Object‐Tracking Accuracy): the accuracy of multi‐object tracking is used to count the accumulation of errors in tracking, including the number of tracking objects and whether they match:

 
MOTP = ( FN + FP + IDSW ) G T

where FN (False Negative) is the number of detection frames that do not match the prediction frame; FP (False positive) is the number of prediction frames that do not match the detection frame; IDSW (ID Switch) is the object ID change the number of times; GT(Ground Truth) is the number of tracking objects.

其中 FN(假阴性)是与预测帧不匹配的检测帧数;FP(误报)是与检测帧不匹配的预测帧数;IDSW(ID开关)是对象ID变化的次数;GT(地面实况)是跟踪对象的数量。

MOTP (Multi‐Object‐Tracking Precision): the accuracy of multi‐object tracking, which is used to evaluate whether the object position is accurately positioned.

MOTP(多目标跟踪精度):多目标跟踪的精度,用于评估目标位置是否准确定位。
 
MOTP = Bt ( i ) C t

where Ct is the number of matches between the object and the predicted object in the t‐th frame; Bt(i) is the distance between the corresponding position of the object in the t‐th frame and the predicted position, also known as the matching error.

AMOTA (Average Multiple Object Tracking Accuracy): summarize MOTA overall object confidence thresholds instead of using a single threshold. Similar to mAP for object detection, it is used to evaluate the overall accurate performance of the tracking algorithm under all thresholds to improve algorithm robustness. AMOTA can be calculated by integrating MOTA under the recall curve, using interpolation to approximate the integral in order to simplify the calculation.

其中 C t 是第t帧中对象与预测对象之间的匹配数;Bt(i) 是物体在第 t 帧中的相应位置与预测位置之间的距离,也称为匹配误差。 AMOTA(平均多对象跟踪精度):汇总所有对象置信度阈值的 MOTA,而不是使用单个阈值。与用于目标检测的mAP类似,它用于评估跟踪算法在所有阈值下的整体准确性能,以提高算法的鲁棒性。AMOTA可以通过在召回率曲线下积分MOTA来计算,使用插值来近似积分以简化计算。
 
AMOTA = 1 L r { 1 L , 2 L , ... , 1 } ( 1 M O T A r )

where L represents the number of recall values (integration confidence threshold), the higher the L, the more accurate the approximate integral. AMOTA represents the multi‐object tracking accuracy at a specific recall value r.

其中 L 表示召回率值的数量(积分置信阈值),L 越高,近似积分越准确。AMOTA 表示特定召回值 r 下的多目标跟踪精度。

AMOTP (Average Multi‐object Tracking Precision): The same calculation method as AMOTA, with recall as the abscissa and MOTP as the ordinate, use the interpolation method to obtain AMOTP.

AMOTP(平均多目标跟踪精度):与AMOTA相同的计算方法,以召回为横坐标,MOTP为纵坐标,采用插值法得到AMOTP。
 
AMOTP = 1 L r { 1 L , 2 L , ... , 1 } ( 1 M O T P r )
IDF1(ID F1 分数):衡量预测 ID 与正确 ID 之间的差异。

IDF1 (ID F1 score): measures the difference between the predicted ID and the correct ID.

MT(大部分跟踪):在 80% 的时间内成功跟踪的对象数占所有跟踪对象的百分比。

MT (Mostly Tracked): the number of objects that are successfully tracked 80% of the time as a percentage of all tracked objects.

ML(大部分丢失):满足跟踪成功的对象数量的百分比 20% 的时间在所有跟踪的对象中。

ML (Mostly Lost): the percentage of the number of objects that satisfy the tracking success 20% of the time out of all the objects tracked.

FM(分段):评估跟踪完整性,定义为 FM,每当轨迹将其状态从跟踪更改为未跟踪时计数,并在以后的时间点跟踪相同的轨迹。

FM (Fragmentation): evaluate tracking integrity, defined as FM, counted whenever a trajectory changes its state from tracked to untracked, and the same trajectory is tracked at a later point in time.

HOTA(高阶度量):由 [34] 提出的用于评估 MOT 的高阶度量。以前的指标过分强调了检测或关联的重要性。此评估指标明确地平衡了执行准确检测、关联和定位的效果,以比较跟踪器。HOTA评分与人类视觉评估更一致。

HOTA (Higher Order Metric): A higher order metric for evaluating MOT proposed by [29]. Previous metrics overemphasized the importance of detection or association. This evaluation metric explicitly balances the effects of performing accurate detection, association, and localization into a unified metric for comparing trackers. HOTA scores are more consistent with human visual evaluations.

 
H O T A = 0 1 H O T A α d α
 
H O T A α = c A ( c ) | T P | + | F N | + | F P |
 
A ( c ) = | T P A ( c ) | | T P A ( c ) | + | F N A ( c ) | + | F P A ( c ) |

where

其中

α is the IoU threshold, and

是 IoU 阈值,

c is the number of positive sample trajectories. In the object tracking experiment, there are predicted detection trajectories and ground truth trajectories. The intersection between the two trajectories is called true positive association(TPA), and the trajectory outside the intersection in the predicted trajectory is called false positive association (FPA). Detections outside the intersection in ground truth trajectories are false negative associations (FNA).

是正样本轨迹的数量。在目标跟踪实验中,有预测的检测轨迹和地面真实轨迹。两条轨迹之间的交点称为真阳性关联(TPA),预测轨迹中交点外的轨迹称为假阳性关联(FPA)。地面真实轨迹中交叉点外的检测是假阴性关联 (FNA)。
ScholarVision Creations