PaperHub
5.8
/10
Rejected4 位审稿人
最低3最高8标准差1.8
3
6
8
6
4.0
置信度
ICLR 2024

CO-MOT: Boosting End-to-end Transformer-based Multi-Object Tracking via Coopetition Label Assignment and Shadow Sets

OpenReviewPDF
提交: 2023-09-22更新: 2024-02-11

摘要

Existing end-to-end Multi-Object Tracking (e2e-MOT) methods have not surpassed non-end-to-end tracking-by-detection methods. One potential reason is its label assignment strategy during training that consistently binds the tracked objects with tracking queries and then assigns the few newborns to detection queries. With one-to-one bipartite matching, such an assignment will yield an unbalanced training, i.e., scarce positive samples for detection queries, especially for an enclosed scene, as the majority of the newborns come on stage at the beginning of videos. Thus, e2e-MOT will be easier to yield a tracking terminal without renewal or re-initialization, compared to other tracking-by-detection methods. To alleviate this problem, we present Co-MOT, a simple and effective method to facilitate e2e-MOT by a novel coopetition label assignment with a shadow concept. Specifically, we add tracked objects to the matching targets for detection queries when performing the label assignment for training the intermediate decoders. For query initialization, we expand each query by a set of shadow counterparts with limited disturbance to itself. With extensive ablations, Co-MOT achieves superior performance without extra costs, e.g., 69.4% HOTA on DanceTrack and 52.8% TETA on BDD100K. Impressively, Co-MOT only requires 38% FLOPs of MOTRv2 to attain a similar performance, resulting in the 1.4$\times$ faster inference speed. Codes are attached for re-implementation.
关键词
End-to-End Multi-Object TrackingTransformer

评审与讨论

审稿意见
3

This paper is about using Transformer to perform multi-object tracking task. Authors introduce "coopetition alternative" strategy to train Transformer's decoders, also a "shadow set" is proposed to maintain a set of queries when tracking each object. Experiments are conducted on DanceTrack, BDD100K, MOT17 to compare with prior works.

优点

  1. This paper comes with a good motivation by first analyzing the drawbacks transformer based trackers, such as MOTR and MOTRv2, then try proposing some improvements based on that.

  2. The idea of using a set of queries to track each object, is interesting.

  3. Results on DanceTrack seems to be strong, e.g. it is comparable to MOTRv2 from Table 2(a) and it is better than a recent work MeMOTR, which is also based on Transformer.

缺点

  1. The performance is not stable across different datasets. For instance, the tracker cannot even achieve comparable performance with existing trackers such as GTR/PA3Former/GRTU on MOT17, which is by far the most widely used and trusted benchmark within MOT community. While authors claim that MOT17 is rather a small dataset to train transformer-based trackers, other ways such as using synthetic dataset[a] could be one option.

Also the proposed tracker performs worse than the major baseline MOTRv2, which is also transformer-based, in terms of IDF1 and HOTA. In this case, it is a bit hard to justify that authors are making a direction towards improving MOTRv2.

In fact, I would stay away with a rather heavier tracker (Transformer with multiple layers and many parameters) that gives even worse performance on MOT17 than light-weight tracking-by-detection/regression based trackers that require way less data to train, while still work better on major tracking benchmarks.

  1. The current approach lacks some form of interpretability. For example, from this claim "COLA and TALA on the different decoders,", why different label assignment should be used in different decoders? What exactly happened inside the decoder part for Transformer-based MOT, is it solving data association exactly or approximately? It would be more advocated if authors could come up with a more insightful explanation for this, and it will also benefit readers.

  2. For the idea of using shadow set, too many engineering tricks such as min, max, operations are used, it would be much better if authors could come up with some mathematical formulations as a technical sound approach to replace these simple heuristics, so as to better justify the use of this method. For example, IMHO, how about borrowing ideas from particle filter, to use importance sampling to maintain a set of queries to track each object?

References: [a] Motsynth: How can synthetic data help pedestrian detection and tracking? ICCV2021

问题

I see in Table 3(a) that adding COLA and shadow set both improves performance than not having them, but this is only tested on DanceTrack, would the COLA and shadow set also work better than not having them at all, on MOT17 validation set?

评论

We thank the reviewer gEi4 for the insightful comments and suggestions. We detailedly respond to each comment in the following.

Response to Weaknesses 1:

1.1 Due to the limited time of the rebuttal, we will supplement the results adding synthetic dataset.

1.2 MOTRv2 is transformer-based, but due to the addition of YOLOX preprocessing, it has a larger computational load and is not a pure End-to-end method. The original intention of this paper is to maintain the End-to-end structure and computational load of MOTR and achieve performance comparable to MOTRv2. The strong performance in Table 2 has well supported our motivation on multiple datasets.

1.3 The table below lists the performance of known End-to-end methods on the MOT17 test. It can be seen that End-to-end tracking has been prevalent in the past several years, and its performance has gradually improved. Currently, End-to-end MOT methods are still in the early stages of development, and we believe that there will be significant improvements in the future. The value of a method should not be denied just because it has not achieved state-of-the-art performance on a particular dataset. For example, although DETR and Deformable DETR have lower performance on the COCO dataset compared to other non-Transformer detection methods, their contributions to subsequent methods cannot be ignored.

SourceHOTADetAAssAMOTAIDF1
MeMOTCVPR’2256.9-55.272.569.0
TrackFormerCVPR’22---65.063.9
MQTarXiv’22---66.565.2
DNMOTarXiv’2358.0--75.668.1
MOTRECCV’2257.860.355.773.468.6
MeMOTRICCV’2358.859.658.472.871.5
MOTRv3 (ConvNeXt-B)arXiv’2360.158.762.175.972.4
CO-MOT-60.159.560.672.672.7

Table 2 (c): Comparison to End-to-end methods on the MOT17 test dataset. The methods with bold italic fonts are newly introduced. CO-MOT has achieved the highest HOTA(60.1). Although the HOTA performance of MOTRv3 is the same as CO-MOT, it uses a more powerful backbone (ConvNeXt-B).

Response to Weaknesses 2:

The following table studies the performance impact of COLA inserting different decoder layers. l<0 means that COLA is not used, and all 6 layers of the decoder use TALA. l<2 means that the first layer of the 6-layer decoder uses COLA. It can be seen that deploying COLA in more decoder layers leads to better HOTA. As shown in Figure 3, during training in MOTR, TALA forcibly matches the tracking query with the corresponding track object, even if the current prediction of the tracking query is far from the target. This strategy makes it more difficult for the network to converge compared to the optimal solution of the Hungarian matching, which in turn makes it difficult for the track query to reach the optimal solution. This can also be seen in Figure 1 in the motivation section. Introducing COLA can improve this convergence. The detection query recalls the tracked targets through Hungarian matching and then passes the relevant information to the track query through the attention mechanism of the Transformer. With the prior knowledge from the detection query, the convergence of the track query becomes easier, and the target will be consistently assigned the same identity.

lHOTADetAAssAMOTAIDF1
l<058.672.347.780.158.9
l<259.072.948.181.259.8
l<459.674.348.082.860.5
l<659.973.249.381.360.9

Table 8: The performance metrics of COLA inserting different numbers of decoder layers.

Response to Weaknesses 3:

Thank the reviewer for the provided advice to make this work more technically solid. "min" and "max" used to select the samples for training are heuristic but effective. More exploration such as importance sampling for sample selection will be conducted in our future work.

Response to Question 1:

The following two tables respectively show the ablation experiments of inserting COLA and Shadow set into TrackFormer and MOTR. As components are added, the tracking performance gradually improves.

COLAShadowHOTA
(a) TrackFormer41.4
(b) ours47.8(+6.4)
(c) ours50.7(+9.3)

Table 4: The ablation experiments of applying CO-MOT to TrackFormer on the DanceTrack validation set.

COLAShadowHOTA of DancetrackHOTA of MOT17
(a) MOTR56.457.8
(b)60.2(+3.8)58.5(+0.7)
(c)61.8(+5.4)60.1(+2.3)

Table 9: The ablation experiments of applying CO-MOT to MOTR on the DanceTrack validation set and MOT17 test.

评论

Thanks for the author for the feedback. I have carefully read the rebuttals.

  1. Regarding "End-to-end MOT methods are still in the early stages of development, and we believe that there will be significant improvements in the future. The value of a method should not be denied just because it has not achieved state-of-the-art performance on a particular dataset..." I do agree that not achieving SOTA should not be the reason to reject a submission. But, DETR is the first work to make Transformers work for object detection and thus is a 01**0-1** work, so no one will criticize its inferior performance compared to non-Transformer based detectors and it has influenced lots of follow-up works. However, the current submission is nothing like DETR, as lots of Transformer based trackers have already been published, e.g. TrackFormer/MOTR/MOTRv2, therefore the current work is not 01**0-1** work like DETR but 12,23**1-2, 2-3** incremental**incremental** work, it is just about "climbing mountains" yet the performance is not even better than non-Transformer based trackers, though I'm not criticizing it not being SOTA. Making it impossible to justify the motivation as, sadly I didn't see novel insights nor strong results.

  2. I do agree with the argument that " More exploration such as importance sampling for sample selection will be conducted in our future work". I hope the author could improve towards a better formulation for this part in the future work.

I think for conferences like NeurIPs/ICLR, we should have more expectations for submissions. I therefore continue with my rating and I hope author would continue working on improving it towards a stronger submission.

审稿意见
6

This paper proposes an end-to-end Transformer-based MOT method. It introduces a coopetition alternative for training the intermediate decoders, and also develops a shadow set as units to augment the queries. Experimental results on some public datasets demonstrate the effectiveness of the proposed method.

优点

  1. This paper is well written and organized. The S-COLA is also clearly explained.
  2. Experimental results on some public datasets demonstrate the effectiveness of the proposed method.

缺点

  1. For the tracking queries and the detection queries, they both self-attend each other and then cross-attend the image feature tokens. Intuitively, all the queries can perceive the whole image, and it is not good enough to split them into two parts, which may degrade the detection results.

  2. Why not evaluate the proposed method on the crowded dataset MOT20? In crowd scenes, the tracking query has the similar semantic information with the detection query, and may fail to assign correct label in crowd scenes.

  3. In Table 1, the authors point out that the detection result of MOTR can be improved by removing the tracking queries. How about the proposed CO-MOT? If CO-MOT is only used as a detection task, and whether the tracking performance can be further improved?

  4. It is better to evaluate the effect of the number of decoders L on the tracking performance.

  5. Pease reorganize the formats of the references, and make them consistent.

问题

See the Weaknesses.

评论

We thank the reviewer oLii for the insightful comments and suggestions. We detailedly respond to each comment in the following.

Response to Weaknesses 1:

In the multi-object tracking (MOT) task, it is necessary to keep tracking the targets that are tracked in the previous frame, while also detecting newborns that appear in the current frame. The tracking queries and detection queries play different roles in these tasks. During training, they are supervised by different supervision, where the detection queries use Hungarian matching, and the tracking queries use TALA. During inference, the tracking queries and detection queries are concatenated together and indiscriminately passed through multiple Transformer decoders. Thus, all the queries can perceive the whole image, as the reviewer points out, but both tracking and detection queries must exist for MOT.

Response to Weaknesses 2:

As the End-to-End solution has just emerged in the past year, there are not many methods evaluated on MOT20 that we could find. Here are the ones in Table 2 (d). Notably, our approach achieves 57.5% HOTA, which is the state-of-the-art in End-to-end tracking methods.

VenueHOTADetAAssAMOTAIDF1
MeMOTCVPR’2254.1\55.063.766.1
TrackFormerCVPR’2254.7\\68.665.7
CO-MOT-57.550.565.760.170.5

Table 2 (d): Comparison to existing methods on the MOT20 test dataset.

Response to Weaknesses 3:

The following table lists the mAP indicators, which are consistent with the DetA indicators in Table 2(a). The detection performance of CO-MOT is far superior to MOTR but slightly inferior to MOTRv2.

Tracking tasks not only include detecting new targets but also need to track targets that have already been tracked, maintaining the identity of the targets between consecutive frames. Therefore, CO-MOT cannot be used only as a detection task, otherwise, it would not be able to associate targets between previous and subsequent frames.

methodmAP
(a) MOTR42.5
(b) MOTRv270.7
(c) CO-MOT69.1

Table 6: The detection performance (mAP) of MOTR (v2) and CO-MOT on DanceTrack validation dataset.

Response to Weaknesses 4:

We ablate the number of decoders and illustrate the results in Table 7. The results show that the HOTA consistently improves as the number of decoders increases, while other metrics significantly increase or are comparable. To make a fair comparison with MOTR and consider the efficiency, we choose six decoders by default in all the experiments.

LHOTADetAAssAMOTAIDF1
L=249.051.946.653.853.2
L=458.271.747.580.860.7
L=661.073.950.682.762.6
L=861.577.049.386.662.6

Table 7: Ablation study on the number of decoders L. As L increases, the computational load increases, and the HOTA tracking performance improves, but the growth rate of HOTA significantly decreases.

Response to Weaknesses 5:

We reorganize the formats of the references according to the official regulations in the manuscript.

审稿意见
8

The paper works on an end-to-end Multi-Object Tracking problem via a query-based Transformer, which assumes that the existing label assignment leads to scarce positive samples for detecting newborns. With this limitation in training, this paper observes a tracking terminal without renewal or re-initialization compared to other tracking-by-detection methods. To remedy this issue, this work presents Co-MOT by matching detection queries to all the tracked and untracked targets. Additionally, each query is augmented by a set with limited disturbance to itself. CO-MOT, with extra costs, achieves 69.4 HOTA on DanceTrack and 52.8 TETA on BDD100K, making it 1.4 times faster than the previous approach MOTRv2.

优点

Deep observation of the existing approaches: This paper provides qualitative results in Figure 1 and quantitative results in Table 1 to bring the readers to what has been done to the existing e2e-MOT via Transformer. The bipartite matching label assignment in the previous approaches has a limitation regarding the video task, compared to the image task, such as object detection. From this perspective, this work involves the inherent property of the e2e-MOT via Transformer and offers a promising solution to the limitation. The superior performance of the proposed method is inspiring, which only requires fewer FLOPS of MOTRv2 to achieve a similar performance. I reckon this work reveals the core problem of the Transformer-based e2e-MOT.

Simple and effective solutions: The paper reveals that the target assignment on detection queries is scarce due to the exclusive tracking queries, leading to unbalanced training. The proposed coopetition label assignment solves this problem by allowing the detection queries to match all the targets in the intermediate decoders except the last decoder. This simple solution can effectively recall newborn detection in the intermediate decoders while keeping standard MOT in the final outputs. In addition, the shadow concept is introduced to augment each query in a feature-augmentation manner. A bulk of ablations are provided to demonstrate the effectiveness of each proposed solution.

缺点

Missing experiments on more models: In Section 3.2, this paper mentions that we revisit the Tracking Aware Label Assignment (TALA) used to train end-to-end Transformers such as MOTR and TrackFormer for MOT. However, all the ablations are conducted on MOTR. I cannot find evidence that COLA and shadow work on other models like TrackFormer. The authors should provide results of TrackFormer on two datasets at least to make the proposed solutions more solid.

Comparison of failure cases: The failure cases shown in Figure 5 are the common cases of all the Transformer-based approaches. The authors should provide the failure cases of previous approaches and denote which kind of case has been solved by CO-MOT.

问题

In Table 2, the authors categorize all the methods into Non-End-to-end and End-to-end. What are the exact definitions of these two categories? Also, it will be better to list all the methods using Transformer.

In Figure 4, do the FLOPs of the MOTR include the YOLOX detector? What are the specific numbers of parameters for different approaches? The specific number of parameters should be indicated in Figure 4. For example, the number can be placed in circles.

伦理问题详情

n/a

评论

We thank the reviewer NAsG for the insightful comments and suggestions. We detailedly respond to each comment in the following.

Response to Weaknesses 1:

In the supplementary materials, we provide the results of COLA and Shadow Set on Trackformer. We do not make any modifications to the TrackFormer framework and used the default hyperparameters provided by TrackFormer. The result shown as follows demonstrates the effectiveness of COLA (+6.4%) and Shadow Set (+9.3%) when applied to TrackFormer. Codes will be released upon acceptance.

COLAShadowHOTA
(a) TrackFormer41.4
(b) ours47.8 (+6.4)
(c) ours50.7 (+9.3)

Table 4: The ablation experiments of applying CO-MOT to TrackFormer on the DanceTrack validation set. As components are added, the tracking performance improves gradually.

Response to Weaknesses 2:

The Figure 7 illustrates the failure cases of MOTR and denotes which kind of case has been solved by CO-MOT. MOTR has poor detection and tracking performance. First, as shown in Figure 7 (a), it fails to detect the person in time under the extreme case of bending over. Second, as shown in Figure 7 (b), due to the tiny visual features when a person stretches out their hand, the detection box is inaccurate, and the model misidentifies it as multiple people. Third, as shown in Figure 7 (c), the tracking identity switches after the human body is obscured. However, all of the above cases can be solved by CO-MOT, showing the extraordinary performance of MOT.

Response to Question 1:

End-to-end tracking [1,2,3,4] is defined as inputting contiguous video frames and directly outputting the boxes and association information without any pre- or post-processing (such as separate detectors, Reid feature extraction, or IOU match). So methods such as MOTRv2, ByteTrack, TransTrack, and OC-SORT belong to Non-End-to-end tracking. On the contrary, MOTR, MeMOTR, MeMOT does indeed belong to an end-to-end approach. We append more Transformer-based methods for comparison on MOT17 in Table 2 (c). The methods with bold italic fonts are newly introduced.

SourceHOTADetAAssAMOTAIDF1
TransTrackarXiv’2054.161.647.974.563.9
P3AFormerECCV’22---81.278.1
DiffusionTrackarXiv’2360.863.258.877.973.8
MOTRv2CVPR’2362.063.860.678.675.0
---------------------
MeMOTCVPR’2256.9-55.272.569.0
TrackFormerCVPR’22---65.063.9
MQTarXiv’22---66.565.2
DNMOTarXiv’2358.0--75.668.1
MOTRECCV’2257.860.355.773.468.6
MeMOTRICCV’2358.859.658.472.871.5
MOTRv3 (ConvNeXt-B)arXiv’2360.158.762.175.972.4
CO-MOT-60.159.560.672.672.7

Table 2 (c): Comparison to existing methods using Transformer on the MOT17 test dataset. The last few lines are all end-to-end tracking methods, among which CO-MOT has achieved the highest HOTA. Although the HOTA performance of MOTRv3 is the same as CO-MOT, it uses a more powerful backbone (ConvNeXt-B).

Response to Question 2:

We show the specific numbers in the following table and indicate them in Figure 4.

paramsFLOPsInference FPSHOTA
MOTR40M173G1954.2
MOTRv2139.1M454.9G1469.9
CO-MOT40M173G1969.4

Table 5: The parameters, model size, and FLOPs of MOTR, MOTRv2, and CO-MOT.

References:

[1] F. Zeng et.al. Motr: End-to-end multiple-object tracking with transformer. ECCV 2022

[2] Y. Zhang et.al. Motrv2: Bootstrapping end-to-end multi-object tracking by pretrained object detectors. CVPR2023

[3] J. Cai et.al. Memot: multi-object tracking with memory. CVPR2022

[4] T. Meinhardt et.al. Trackformer: Multi-object tracking with transformers. CVPR2022

[5] Yu et.al. MOTRv3: Release-Fetch Supervision for End-to-End Multi-Object Tracking. https://arxiv.org/abs/2305.14298

审稿意见
6

The paper introduces CO-MOT, a method that enhances the performance of end-to-end Transformer-based MOT. It addresses issues in the existing framework where label assignment fails to fully utilize detection queries due to the exclusive nature of detection and tracking queries. To overcome this, a coopetition alternative is proposed for training intermediate decoders. Additionally, a shadow set is developed to augment queries and mitigate training imbalances caused by one-to-one matching. The experiments are conducted on MOT17, BDD100k and DanceTrack.

优点

  • This paper mentioned that there is a large amount of post-processing in non-end-to-end methods in the MOT field, which leads to difficulties in parameter tuning in application. In this paper, some improvements have been made based on existing end-to-end methods, resulting in certain speed and accuracy improvements.

  • This paper analyzes the shortcomings of existing e2e-MOT models and proposes a new network design strategy based on the analyses.

缺点

  • The layout of the article is quite chaotic. Figure 1 is used as an example in the method introduction, but it is too far from the text information. For another example, in Section 3.2 of the introduction of strategies and methods, formulas are mixed in a large paragraph without a comprehensive formula or image description, making it difficult for readers to understand. The reviewer must refer to the paper of the baseline tracker MOTR for a comprehensive understanding. The authors should re-organize the method introduction and include a detailed figure to highlight the technical improvements of the new approach, including both the label assignment and shadow set.

  • The comparisons with more tracking approaches on the BDD100k or DanceTrack datasets are required to demonstrate the performance improvement.

问题

Please see the Weaknesses section.

评论

We thank the reviewer kcLe for the insightful comments and suggestions. We detailedly respond to each comment in the following.

Response to Weaknesses 1:

Thank you for your constructive feedback on the paper organization. We have updated the manuscript according to your suggestions to improve its readability and comprehensibility. For example, we move Figure 1 near the method section so that readers can refer to it more conveniently. Additionally, we demonstrate the symbols of Section 3.2 in Figure 2. To highlight the technical improvements of the new approach, we add a detailed figure of MOTR (v2) and CO-MOT in Figure 6. Note that the stacked circles and squares denote the shadow sets.

Response to Weaknesses 2:

We demonstrate tracking methods as many as possible in the supplementary materials, which are evaluated on DanceTrack. The methods with bold italic fonts are newly introduced. For BDD100K, we have already shown all the methods evaluated in Table 2(b) so that no more methods are introduced.

SourceHOTADetAAssAMOTAIDF1
SORTICIP’1647.972.031.291.850.8
CenterTrackECCV’2041.878.122.686.835.7
TransTrackarXiv’2045.575.927.588.445.2
FairMOTIJCV’2139.766.723.882.240.8
QDTrackCVPR’2154.280.136.887.750.4
TraDeSCVPR’2143.374.525.486.241.2
ByteTrackECCV’2247.771.032.189.653.9
OC-SORTCVPR’2355.180.338.392.054.6
GTRCVPR’2248.072.531.984.750.3
C-BIoUWACV’2360.681.345.491.661.6
MT IoTarXiv’2266.784.153.094.070.6
DiffusionTrackarXiv’2352.482.233.589.347.5
MOTRv2+CVPR’2369.983.059.091.971.7
---------------------
MOTRECCV’2254.273.540.279.751.5
DNMOTarXiv’2353.5--89.149.7
MeMOTRICCV’2363.477.052.385.465.5
MeMOTR*ICCV’2368.580.558.489.971.2
MOTRv3 (Res50)arXiv’2368.3--91.770.1
CO-MOT-65.380.153.589.366.5
CO-MOT+-69.482.158.991.271.9

Table 2 (a): Comparison to existing methods on the DanceTrack test set. The last few lines are all end-to-end tracking methods, among which CO-MOT has achieved the highest HOTA.

AC 元评审

The paper initially had diverging reviews (3, 6, 6, 8). The major concerns of the reviewers were:

  1. presentation issues, need to refer to baseline MOTR paper [kcLe]
  2. Needs more comparison on BDD100k and DanceTrack [kcLe]
  3. Missing evidence that COLA/Shadows works on other models like Trackformer. [NAsG]
  4. Provide failures cases specific to previous approaches that were solved by CO-MOT [NAsG]
  5. Table 2 presentation issues [NAsG]
  6. Fig 4 - show the number of parameters, FLOPS of YOLOX included in MOTR? [NAsG]
  7. Is it necessary to split tracking/detection queries? [oLii]
  8. Missing evaluation on MOT20. [oLii]
  9. What if the tracking queries are removed in CO-MOT. Does detection result improve similar to MOTR? [oLii]
  10. effect of the number of decoders L on tracking performance [oLii]
  11. Performance improvement not consistent across different datasets [gEi4]
  12. Performance worse than MOTRv2 in terms of IDF1 and HOTA [gEi4]
  13. Lack of interpretability of proposed method [gEi4]
  14. Lack of mathematical formulation for shadow sets -- too many engineering tricks or heuristics [gEi4]
  15. missing ablation study on w/ and w/o COLA/shadow set on MOT17 validation set. [gEi4]

The authors wrote a response to address these concerns. During the discussion, Rev gEi4 was still concerned about the novelty being incremental, since the proposed method mainly works to enhance the detector for MOT and not the data association, and the proposed methods used are heuristic and without strong mathematical formulations or justifications. Rev oLii agreed with Rev gEi4 about the novelty. Meanwhile, Rev oLii was not convinced with the MOT20 results, since the proposed method seems worse than two-stage methods. Rev gEi4 also agreed that the MOT20. In contrast, Rev NAsG thought that the method was able to rectify the shortcomings of MOTRv2 by introducing efficient label assignment and shadow sets, achieving efficiency and competitive performance. Rev NAsG also thought that the inferior performance on MOT17/20 was due to the small scale of the datasets, and the proposed method showed its advantage on BDD100K and DanceTrack, which are larger datasets.

Overall, the AC agrees with the concerns of Rev gEi4 and oLii. The methodology is heuristic and not strongly mathematically motivated or analyzed. Thus, the proposed work lacks methodological depth or broad applicability. Rev gEi4 offers a good suggestion to improve the formulation of ShadowSets by borrowing from e.g., particle filters or importance sampling. Overall the results are promising, though. Thus, the AC recommends reject.

为何不给更高分

While the results are promising (better than previous end-to-end trackers, but not better than transformer trackers), the methodology is incremental and lacks mathematical formulations/justifications. It will be a very narrow audience in ICLR.

为何不给更低分

n/a

最终决定

Reject