PaperHub
7.5
/10
Spotlight4 位审稿人
最低6最高8标准差0.9
8
8
8
6
3.3
置信度
正确性3.0
贡献度2.8
表达3.3
ICLR 2025

Test-time Adaptation for Cross-modal Retrieval with Query Shift

OpenReviewPDF
提交: 2024-09-27更新: 2025-02-28
TL;DR

We propose a novel test-time adaptation method for achieving robust cross-modal retrieval against query shift.

摘要

关键词
Test-time adaptationCross-modal retrievalQuery shift

评审与讨论

审稿意见
8

This paper introduces a novel method named TCR for addressing the query shift problem in cross-modal retrieval. TCR employs a test-time adaptation approach that leverages a multi-scale adaptive convolutional neural network and a hybrid transformer module to refine query predictions and adapt to shifts in query distribution without additional training data. The method is designed to enhance the uniformity of the query modality and reduce the gap between query and gallery modalities, thereby improving retrieval performance. The study demonstrates TCR's effectiveness on image-text retrieval tasks using standard benchmarks and various corruption types.

优点

  1. The paper proposes a novel test-time adaptation method (TCR) to address the query shift problem in cross-modal retrieval. This method achieves robustness against query shift by adjusting query predictions and designing a joint objective function, which is an interesting and potentially influential direction for research.
  2. The authors have conducted extensive experiments on multiple datasets, including COCO-C and Flickr-C, to verify the effectiveness of the proposed method. The experiments cover comparisons across different model types and sizes, as well as varying severity levels of query shift, demonstrating the robustness of the method.
  3. The paper not only introduces a new method but also provides an in-depth analysis of the impact of query shift on cross-modal retrieval, revealing how query shift can reduce the uniformity of the query modality and increase the gap between the query and gallery modalities. These theoretical analyses offer valuable insights for future research.

缺点

  1. The TCR method proposed in the paper performs model adaptation at test time, which may increase additional computational costs. It is recommended that the authors analyze the computational complexity of the model and the additional cost incurred.
  2. Are the COCO-C and Flickr-C datasets constructed by the authors themselves? It seems that the paper does not explain whether the results of the baseline methods for comparison were obtained by the authors' own experiments or cited from their respective articles. If they were obtained through their own experiments, it should be clarified whether such comparisons are fair (whether they were trained on the new baselines), which is quite confusing for readers.
  3. The authors are advised to provide more explanation on the baselines and to include a few actual examples of query shifts to help readers intuitively feel the task.

问题

Please address my concerns proposed in Weakness.

评论

Thanks for the detailed comments. In the following, we will answer your questions one by one.

Q1: The TCR method proposed in the paper performs model adaptation at test time, which may increase additional computational costs. It is recommended that the authors analyze the computational complexity of the model and the additional cost incurred.

A1: Thanks for your comments. In response to your insightful suggestion, we conduct additional experiments to analyze the efficiency of TCR. To this end, we choose the pre-trained model BLIP as the source model and perform zero-shot retrieval on the COCO dataset. We measure the GPU time during the test-time adaptation phase. Results are summarized in Table 19 within the revised manuscript. For your convenience, we attach the corresponding results in the following table.

MethodTRIRAvg.
Tent285.5 seconds189.7 seconds237.6 seconds
EATA276.3 seconds190.4 seconds233.3 seconds
DeYO391.6 seconds254.2 seconds322.9 seconds
Ours291.1 seconds193.6 seconds242.4 seconds

Note that the learnable parameters of all the methods are the same for a fair comparison. The results underscore that TCR achieves adaptation more efficiently than the augmentation-based method DeYO[A]. Compared to the vanilla Tent[B] and EATA[C] (only low-entropy samples are employed for optimization), TCR requires only a negligible additional time cost, primarily due to the nearest neighbor selection in the query prediction refinement module.

Q2: Are the COCO-C and Flickr-C datasets constructed by the authors themselves? It seems that the paper does not explain whether the results of the baseline methods for comparison were obtained by the authors' own experiments or cited from their respective articles. If they were obtained through their own experiments, it should be clarified whether such comparisons are fair (whether they were trained on the new baselines), which is quite confusing for readers.

A2: We apologize for the confusion arising from the initial presentation. Following the setting in [D], we have introduced 31 types of corruptions to establish the Query Shift benchmarks (i.e., COCO-C and Flickr-C datasets). The details of the benchmarks are provided in Appendix B.1 and the data processing code is available at https://github.com/Jielin-Qiu/MM_Robustness. Specifically, for image corruptions, we employ image_perturbation/perturb_COCO_IP.py to construct image corruptions for the COCO dataset and image_perturbation/perturb_Flickr30K_IP.py for the Flickr dataset. For text corruptions, we utilize text_perturbation/perturb_COCO_TP.py to construct text corruptions for the COCO dataset and text_perturbation/perturb_Flickr30K_TP.py for the Flickr dataset.

In the paper, we aim to achieve online adaptation for cross-modal retrieval models under query shift, which is a novel challenge. Unfortunately, existing test-time adaptation (TTA) methods overlook the query shift in cross-modal settings and do not address this challenge. Moreover, most existing TTA methods are specifically designed for the recognition task and cannot be directly employed for the cross-modal retrieval task. To solve the problem, we propose a simple baseline so that the recognition-oriented TTA methods could be employed for the cross-modal retrieval task. For a fair comparison, the optimizer, learning rate, and training parameters across all the methods are the same. Besides, we carefully set the temperature for baselines on various datasets, as detailed in Appendix B.2. For your convenience, we attach the corresponding content as follows.

To guarantee the performance of the baselines, we select the optimal temperature (Eq. 1) for the TTA baselines upon each dataset. According to Fig. 4(a), the temperature is fixed as 0.010.01 for COCO-C, Flickr-C, COCO, Flickr, and Nocaps datatsets, 0.0010.001 for Fashion-Gen dataset, and 0.00010.0001 for CUHK-PEDE and ICFG-PEDES datasets.

The code will be released upon acceptance of the paper.

评论

Q3: The authors are advised to provide more explanation on the baselines and to include a few actual examples of query shifts to help readers intuitively feel the task.

A3: Thanks for your valuable suggestions. In Appendix B.1, we have provided examples of query shift on the COCO-C dataset, which includes 16 types of image corruption and 15 types of text corruption.

More examples of query shifts. In response to your insightful suggestion, we offer more examples of query shifts in the ReID domain. The examples are depicted in Fig. 6 within the revised manuscript. For your convenience, we summarize the difference about the distribution shifts between the CUHK-PEDES dataset and ICFG-PEDES datasets:

CUHK-PEDES is a dataset designed for text-to-image person re-identification, and the test set consists of 3,074 images and 6,156 textual descriptions associated with 1,000 identities. The images are sourced from five re-identification datasets, CUHK03 [E], Market-1501 [F], SSM [G], VIPER [H], and CUHK01 [I]. These images mainly capture outdoor scenes in diverse public spaces, such as markets and campuses. The textual descriptions often contain details not directly relevant to identity (e.g., actions and backgrounds), with an average of 23.5 words per description.

ICFG-PEDES is a large-scale text-to-image person re-identification dataset, and the test set contains 19,848 image-text pairs of 1,000 identities. The images are sourced from the MSMT17 dataset [J] and depict scenes within a campus environment, with a mix of indoor and outdoor settings. Textual descriptions are more identity-focused and fine-grained, averaging 37.2 words per description.

Notably, images in the ICFG-PEDES dataset are collected over multiple days at different times (morning, noon, and afternoon), which introduces considerable illumination variation. In contrast, images in the CUHK-PEDES dataset are sourced from short-duration surveillance videos, leading to minimal lighting variation.

评论

More explanation on the baselines. In response to your constructive feedback, we provide more detailed introdution of the baselines, , which can be found in Appendix B.3 due to space limitations. For your convenience, we attach the added statement as follows.

Test-time Adaptation (TTA) aims to reconcile the distribution shifts in an online manner. Towards achieving this goal, Fully TTA ([B]) has been proposed, which fine-tunes the BatchNorm layers by minimizing entropy during the test phase. EATA ([C]) employs a Fisher regularizer to limit excessive model parameter changes and filter out high-entropy samples via selection strategy. SAR [K] removes high-gradient samples and promotes flat minimum weights, enhancing robustness against more challenging TTA scenarios such as mixed domain shifts, single-sample adaptation, and imbalanced label shifts. READ ([L]) proposes a noise-robust adaptation loss and reliable fusion module to tackle the reliability bias challenge in the multi-modal setting. DeYO ([A]) reveals the unreliability of treating entropy as the confidence metric and establishes a novel metric by measuring the difference between predictions before and after applying an object-destructive transformation.

Reference:

[A] Jonghyun Lee, Dahuin Jung, Saehyung Lee, Junsung Park, Juhyeon Shin, Uiwon Hwang, and Sungroh Yoon. Entropy is not enough for test-time adaptation. In ICLR, 2024.

[B] Dequan Wang, Evan Shelhamer, Shaoteng Liu, Bruno Olshausen, and Trevor Darrell. Tent: Fully test-time adaptation by entropy minimization. In ICLR, 2021.

[C] Shuaicheng Niu, Jiaxiang Wu, Yifan Zhang, Yaofo Chen, Shijian Zheng, Peilin Zhao, and Mingkui Tan. Efficient test-time model adaptation without forgetting. In ICML, 2022.

[D] Jielin Qiu, Yi Zhu, Xingjian Shi, Florian Wenzel, Zhiqiang Tang, Ding Zhao, Bo Li, and Mu Li. Benchmarking Robustness of Multimodal Image-Text Models under Distribution Shift. Journal of Data-centric Machine Learning Research, 2023.

[E] Wei Li, Rui Zhao, Tong Xiao, and Xiaogang Wang. DeepReID: Deep Filter Pairing Neural Network for Person Re-Identification. In CVPR, 2014.

[F] Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jiahao Bu, and Qi Tian. Person Re-identification Meets Image Search. In arXiv, 2015.

[G] Tong Xiao, Shuang Li, Bochao Wang, Liang Lin, and Xiaogang Wang. Joint Detection and Identification Feature Learning for Person Search. In arXiv, 2016.

[H] Douglas Gray, Shane Brennan, and Hai Tao. Evaluating appearance models for recognition, reacquisition, and tracking. In Proc. IEEE International Workshop on Performance Evaluation for Tracking and Surveillance (PETS), 2007.

[I] Wei Li, Rui Zhao, and Xiaogang Wang. Human Reidentification with Transferred Metric Learning. In ACCV, 2012.

[J] Longhui Wei, Shiliang Zhang, Wen Gao, and Qi Tian. Person Transfer GAN to Bridge Domain Gap for Person Re-Identification. In CVPR, 2018.

[K] Shuaicheng Niu, Jiaxiang Wu, Yifan Zhang, Zhiquan Wen, Yaofo Chen, Peilin Zhao, and Mingkui Tan. Towards stable test-time adaptation in dynamic wild world. In ICLR, 2023.

[L] Mouxing Yang, Yunfan Li, Changqing Zhang, Peng Hu, and Xi Peng. Test-time Adaption against Multi-modal Reliability Bias. In ICLR, 2024.

评论

Thanks for the response, I choose to keep my score.

评论

We greatly appreciate the time and effort you invested in reviewing our work. Your feedback has been helpful in improving the paper. We would be happy to discuss further if needed.

审稿意见
8

This paper introduces a novel setting, cross-modal retrieval under query shift. To address this challenge, it introduces a test-time adaptation method called TCR, which includes a query prediction refinement module to produce retrieval-optimized predictions for incoming queries. Additionally, it employs a joint objective function for online adaptation, effectively handling the query shift and noise.

优点

  1. The research question, cross-modal retrieval under query shift, is challenging and holds significant practical relevance.
  2. Although this method builds on the principles of TTA, it also reveals TTA’s limitations in cross-modal retrieval and effectively overcomes these challenges.
  3. Extensive experiments demonstrate the effectiveness of the proposed TCR method.
  4. The paper is well-organized and well-written, enhancing the clarity and impact of its findings.

缺点

  1. This research setting is limited by the assumption that each query batch contains i.i.d. samples. However, in real scenarios, query shift may occur unpredictably, introducing non-i.i.d. data within the same batch. This raises concerns about the method’s applicability under such conditions.
  2. Regarding the emergence of query shift, I am curious whether temporal issues, such as temporal shifts or concept drift discussed in [1-3], are present in real-world scenarios. Could the authors provide relevant discussion on this aspect?
    [1] Evolving standardization for continual domain generalization over temporal drift. NIPS 2023. [2] Temporal domain generalization with drift-aware dynamic neural networks. arXiv preprint arXiv:2205.10664 (2022) [3]Online Boosting Adaptive Learning under Concept Drift for Multistream Classification, AAAI 2024
  3. In Section 3.2.1 on candidate selection, it would be valuable to address two points: first, whether gallery shift affects the outcomes of nearest neighbor selection; and second, how the number of selected candidates impacts the results. Additional experiments should be conducted to clarify these aspects.
  4. In Section 3.2.2, given the shift between the source and target domains, it is unclear why source-domain-like data can be directly selected based on centers. Could the authors provide further analysis and explanation on this approach?
  5. In section 3.5the definition of S(x_{i}^Q) in Equation (11) lacks corresponding theoretical analysis.
  6. In the experiments, the authors employed various methods to generate image or text query shifts. I believe the results may depend on the specific shift generation techniques used. Therefore, it is crucial to provide access to the data processing methods and code to ensure the reproducibility of the experimental results.
  7. I recommend that the authors discuss the limitations of the proposed method and outline specific future research directions. This would provide readers with additional insights and considerations for further exploration.

问题

discussed in Weaknesses section.

评论

How the number of selected candidates impacts the results. We conduct more experiments to investigate how the number of selected candidates affects performance. To this end, we directly perform zero-shot retrieval experiment on the COCO dataset with pre-trained BLIP as the source model. In the paper, we retrieve the most similar sample from the gallery set for each query, thus the number of selected candidates is equal to the batch size BB. In the additional experiment, we vary the number of selected candidates at different values, i.e., [0.2B,0.5B,B,2B,5B,10B,50B][0.2B, 0.5B, B, 2B, 5B, 10B, 50B]. Specifically, assume that the number of selected candidates is λB\lambda B, where λ\lambda is an integer. When λ<1\lambda < 1, we randomly select λB\lambda B candidates from the original BB selected candidates. When λ1\lambda \geq 1, we retrieve the most similar λ\lambda candidates from the gallery set for each query, forming a new set of λB\lambda B selected candidates. The results are depicted in Table 18 within the revised manuscript. For your convenience, we attach the corresponding numerical results in the following tables.

Number0.2B0.5BB (Default)2B5B10B50B
TR@167.268.568.965.364.964.764.6
IR@147.348.348.948.348.248.047.5

From the results, one could observe that enlarging the number of selected candidates would significantly degrade the performance. Such a phenomenon indicates that an excessively large number of candidates may lead to the underfitting issue, which highlights the necessity and effectiveness of the query refinement module.

Q4: In Section 3.2.2, given the shift between the source and target domains, it is unclear why source-domain-like data can be directly selected based on centers. Could the authors provide further analysis and explanation on this approach?

A4: We appreciate your feedback. As illustrated in Fig. 1(c), we observe that distribution shift would diminish the modality uniformity, defined as the average distance between all samples and the modality center (Eq. 14 in the manuscript). For your convenience, we have attached the corresponding equation below.

Uniformity=1NQi=1NQziQZQ. \text{Uniformity}=\frac{1}{N^{Q}}\sum_{i=1}^{N^{Q}}\|\mathbf{z}_{i}^{Q}-\overline{\mathbf{Z}}^{Q}\|.

In other words, the distribution shift would narrow the distance between samples and their modality centers. Therefore, we conclude that data from the source domain should exhibit higher modality uniformity. Based on the conclusion, we select samples farther from their modality centers as the source-domain-like data, since these samples enjoy higher modality uniformity.

Q5: In section 3.5,the definition of S(xiQ)S(x_{i}^Q) in Equation (11) lacks corresponding theoretical analysis.

A5: We apologize for the initial oversight regarding the theoretical analysis in Section 3.5. The proposed noise-robust adaptation loss (Eq. 11) aims to achieve robustness against heavy noise by excluding high-entropy query predictions from adaptation and assigning higher weights to query predictions with lower uncertainties. Specifically, for a given query sample xiQx_{i}^{Q}, let its entropy be denoted as E(xiQ)E(x_{i}^Q).

When E(xiQ)EmE(x_i^Q) \geq E_m, the weight S(xiQ)S(x_i^Q) is defined as

S(xiQ)=0.S(x_i^Q) = 0.

In this case, xiQx_{i}^{Q} is excluded from optimization. Such a design prevents high-entropy (i.e., noisy) query predictions from degrading performance, as their gradients produced by entropy loss might be biased and unreliable.

When 0E(xiQ)<Em0 \leq E(x_i^Q) < E_m, the weight S(xiQ)S(x_i^Q) is positive and defined as

S(xiQ)=1E(xiQ)Em.S(x_i^Q) = 1 - \frac{E(x_i^Q)}{E_m}.

The weight is inversely proportional to entropy, i.e., the weight decreases as entropy increases. Formally,

S(xiQ)1E(xiQ).S(x_i^Q) \propto \frac{1}{E(x_i^Q)}.

The adaptive weighting strategy enjoys two advantages. On the one hand, query predictions with lower entropy (i.e., reliable predictions) are assigned with higher weights, thus guiding the optimization. On the other hand, query predictions with higher entropy (i.e., uncertain predictions) are assigned with lower weights, thereby preventing overfitting on noisy query predictions.

评论

Q6: It is crucial to provide access to the data processing methods and code to ensure the reproducibility of the experimental results.

A6: Following the setting in [G], we have introduced 31 types of corruption to establish the Query Shift benchmarks (i.e., COCO-C and Flickr-C datasets). The details of the benchmarks are provided in Appendix B.1 and the data processing code is available at https://github.com/Jielin-Qiu/MM_Robustness. Specifically, for image corruptions, we employ image_perturbation/perturb_COCO_IP.py to construct image corruptions for the COCO dataset and image_perturbation/perturb_Flickr30K_IP.py for the Flickr dataset. For text corruptions, we utilize text_perturbation/perturb_COCO_TP.py to construct text corruptions for the COCO dataset and text_perturbation/perturb_Flickr30K_TP.py for the Flickr dataset.

Q7: I recommend that the authors discuss the limitations of the proposed method and outline specific future research directions. This would provide readers with additional insights and considerations for further exploration.

A7: Thanks for your valuable suggestions. In response to your concern, we summarize the following limitations and potential directions for future work of TCR.

Limitations. The proposed TCR might have the following two limitations. On the one hand, although TCR achieves significant performance improvement in the cross-modal retrieval task, it remains uncertain whether TCR could achieve similar success in other cross-modal tasks, such as image captioning and visual question answering (VQA). On the other hand, the robustness of TCR against more challenging TTA scenarios (e.g., single-sample adaptation and continuous adaptation) is worth further investigation.

Future research. In the future, we plan to extend TCR to more applications and more challenging scenarios. Specifically, we would like to extend TCR for a broader range of cross-modal tasks, such as image captioning and visual question answering (VQA). Besides, although we have demonstrated that TCR could achieve robustness against the query shift evaluated in the paper, further work is needed to verify whether TCR could address more challenging scenarios, such as temporal shift and concept drift.

Reference:

[A] Dequan Wang, Evan Shelhamer, Shaoteng Liu, Bruno Olshausen, and Trevor Darrell. Tent: Fully test-time adaptation by entropy minimization. In ICLR, 2021.

[B] Shuaicheng Niu, Jiaxiang Wu, Yifan Zhang, Yaofo Chen, Shijian Zheng, Peilin Zhao, and Mingkui Tan. Efficient test-time model adaptation without forgetting. In ICML, 2022.

[C] Jonghyun Lee, Dahuin Jung, Saehyung Lee, Junsung Park, Juhyeon Shin, Uiwon Hwang, and Sungroh Yoon. Entropy is not enough for test-time adaptation. In ICLR, 2024.

[D] En Yu, Jie Lu, Bin Zhang, and Guangquan Zhang. Online Boosting Adaptive Learning under Concept Drift for Multistream Classification. In AAAI, 2024.

[E] Guangji Bai, Chen Ling, and Liang Zhao. Temporal domain generalization with drift-aware dynamic neural networks. In arXiv, 2022.

[F] Mixue Xie, Shuang Li, Longhui Yuan, Chi Harold Liu, and Zehui Dai. Evolving standardization for continual domain generalization over temporal drift. In NIPS, 2023.

[G] Jielin Qiu, Yi Zhu, Xingjian Shi, Florian Wenzel, Zhiqiang Tang, Ding Zhao, Bo Li, and Mu Li. Benchmarking Robustness of Multimodal Image-Text Models under Distribution Shift. Journal of Data-centric Machine Learning Research, 2023.

评论

Thank you very much for your responses. I have carefully reviewed each of them. You have addressed all my concerns, and I think this is a very interesting and meaningful work that provides valuable insights into cross-modal retrieval research. Therefore, I will increase my score.

评论

Thank you for your prompt response and for upgrading your score! We deeply appreciate the time and effort you dedicated to reviewing our work. Your constructive feedback has been invaluable in helping us refine and improve the paper.

评论

Thanks for the insightful reviews. We will answer your questions one by one in the following.

Q1: This research setting is limited by the assumption that each query batch contains i.i.d. samples. However, in real scenarios, query shift may occur unpredictably, introducing non-i.i.d. data within the same batch. This raises concerns about the method’s applicability under such conditions.

A1: In order to address your concerns, we conduct more experiments on the COCO-C benchmark, investigating the robustness of the proposed TCR under non-i.i.d. settings (i.e., Mixed Severity Levels and Mixed Corruption Types). Specifically,

  • Mixed Severity Levels: For each corruption, we create the test pairs by selecting 1/m1/m of the data from each severity level, resulting in a total of NN test pairs, where mm is the number of severity levels and m=5 / 7 / 2 for the image / character-level / word-level corruptions.

  • Mixed Corruption Types: For the text retrieval, we construct the test pairs by selecting 1/16 of the data from each image corruption (1 through 16), resulting in a total of NN test pairs. For the image retrieval, we create test pairs by selecting 1/15 of the data from each text corruption (1 through 15), resulting in a total of NN test pairs.

To verify the effectiveness of TCR under the Mixed Severity and Mixed Corruption Types settings, we choose the typical TTA method Tent ([A]) and the SOTA TTA methods EATA ([B]), DeYO ([C]) as baselines for comparisons. In the experiment, we carry out the experiments on the COCO-C benchmark, and the corresponding results are depicted in Tables 13-15 within the revised manuscript. For your convenience, we attach the corresponding numerical results (regarding Recall@1) in the following tables.

Mixed Severity LevelsGauss.ShotImpul.SpeckleDefoc.GlassMotionZoomSnowFrostFogBrit.Contr.ElasticPixelJPEGAvg.
BLIP ViT-B/1661.861.859.766.458.570.756.722.542.660.466.070.861.061.247.869.858.6
Tent65.364.959.969.431.674.135.71.910.763.370.473.864.465.847.871.554.4
EATA64.965.664.670.062.074.361.728.155.363.971.174.465.566.053.772.763.4
DeYO64.066.063.069.864.674.663.05.856.165.771.474.565.767.852.572.762.3
Ours67.268.166.670.767.075.865.845.761.268.474.275.270.470.458.673.567.4
Mixed Severity LevelsOCRCICRCSCDSRRIRSRDIPAvg.
BLIP ViT-B/1642.129.728.033.930.044.851.750.550.856.841.8
Tent42.428.523.633.826.945.552.451.450.957.041.2
EATA43.430.829.434.830.746.053.251.851.457.642.9
DeYO43.430.729.335.030.946.253.451.951.457.743.0
Ours44.432.230.635.731.746.353.852.151.557.443.6

Note that for the Mixed Corruption Types setting, there are five levels of the mixed corruptions in text retrieval, corresponding to the image corruptions with five severity levels. For image retrieval, the severity levels of character-level / word-level / sentence-level text corruptions are 7 / 2 / 1. Thus, we select the two highest severity levels for character-level and word-level corruptions, and combine them with sentence-level corruptions, resulting in two levels of the mixed corruptions.

评论
Mixed Corruption TypesTR@1 Level 1TR@1 Level 2TR@1 Level 3TR@1 Level 4TR@1 Level 5IR@1 Level 1IR@1 Level 2Avg.
BLIP ViT-B/1668.864.561.454.544.942.541.153.9
Tent70.067.064.456.433.242.238.553.1
EATA71.768.264.758.948.043.241.956.7
DeYO71.568.465.159.948.342.941.856.8
Ours73.370.466.961.553.643.842.358.8

The performance superiority of TCR over all baselines under both Mixed Severity and Mixed Corruption Types settings demonstrates its robustness against non-i.i.d query shift.

Q2: Regarding the emergence of query shift, I am curious whether temporal issues, such as temporal shifts or concept drift discussed in [1-3], are present in real-world scenarios. Could the authors provide relevant discussion on this aspect? [1] Evolving standardization for continual domain generalization over temporal drift. NIPS 2023. [2] Temporal domain generalization with drift-aware dynamic neural networks. arXiv preprint arXiv:2205.10664 (2022). [3] Online Boosting Adaptive Learning under Concept Drift for Multistream Classification. AAAI 2024.

A2: Thanks for your constructive comments. We completely agree with your insightful opinion that the temporal issues would lead to query shift. We have cited these related works and established some connections with them in Appendix E of the revised manuscript. To be specific, as discussed in [D], the underlying distributions of data are distinct at different times, leading to concept drift. For instance, in weather forecasting, data is collected across diverse distributions (e.g., sunny, frost, snow). It is noteworthy that we have evaluated TCR under query shift caused by various weather conditions (e.g., frost, snow, fog). The corresponding results from COCO-C (Table 1 in the manuscript) and Flickr-C (Table 7 in the manuscript) benchmarks demonstrate the robustness of TCR against concept drift to some extent. For your convenience, we attach the corresponding numerical results (regarding TR@1) in the following tables.

COCO-C BenchmarkSnowFrostFogBrightAvg.
BLIP ViT-B/1632.352.257.066.852.1
Tent31.948.756.366.550.9
EATA45.656.762.571.459.0
SAR38.056.259.170.656.0
READ39.949.958.470.354.6
DeYO37.559.766.471.258.7
Ours56.564.171.073.466.3
Flickr-C BenchmarkSnowFrostFogBrightAvg.
BLIP ViT-B/1666.480.479.585.578.0
Tent67.280.979.686.878.6
EATA72.083.782.587.981.5
SAR71.983.182.287.981.3
READ71.783.881.987.781.3
DeYO73.184.183.288.682.3
Ours78.285.285.789.584.7
评论

As noted in [E] and [F], the collected data would continuously vary over time, resulting in temporal shift. For example, changes in lighting conditions throughout the day could impact the distribution of the collected data. For evaluation, we have conducted experiments under the CUHK2ICFG setting (Table 4 in the manuscript). Specifically, the ICFG-PEDES dataset is gathered at different times of the day (i.e., morning, noon, and afternoon), while the CUHK-PEDES dataset is derived from short-duration surveillance videos. Therefore, compared to CUHK-PEDES, the data in the ICFG-PEDES dataset exhibit distribution shifts due to time changes, such as illumination variation. The corresponding results from the CUHK2ICFG setting demonstrate that TCR could achieve robustness against temporal shift to some extent. For your convenience, we attach the corresponding numerical results (regarding IR@1) in the following tables.

MethodCUHK2ICFG
CLIP ViT-B/1641.0
Tent41.9
EATA42.2
SAR42.2
READ42.3
DeYO42.2
Ours42.4

The three works mentioned are primarily designed for the classification task and focus on addressing distribution shifts caused by temporal issues. In contrast, TCR aims to tackle the query shift challenge in the cross-modal retrieval task. Notably, any distribution shifts in the query modality would lead to query shift, not limited to temporal issues. For instance, personalized issues such as writing habits and styles, as well as real-world corruptions like noise and blur, would lead to query shift.

Q3: In Section 3.2.1 on candidate selection, it would be valuable to address two points: first, whether gallery shift affects the outcomes of nearest neighbor selection; and second, how the number of selected candidates impacts the results. Additional experiments should be conducted to clarify these aspects.

A3: Thanks for your comments. In the submission, we have conducted experiments under the Query-Gallery Shift setting (Table 3 in the manuscript), which demonstrates that TCR could improve retrieval performance even when the gallery modality occurs distribution shift. To address your concerns, we conduct more experiments during the rebuttal and present the results and analysis as follows.

Whether gallery shift affects the outcomes of nearest neighbor selection. In response to your insightful suggestion, we conduct additional experiments to investigate whether gallery shift would affect the outcomes of nearest neighbor selection. Specifically, we carry out experiments on the COCO-C benchmark under two gallery shift settings, i.e., only gallery shift setting, both query and gallery shift setting. The corresponding results are depicted in Tables 16-17 within the revised manuscript. For your convenience, we attach the numerical results regarding Recall@1 and neighbor ACC (i.e., the cross-modal nearest neighbor of the query is correct) in the following tables. Notably, for the baseline BLIP ViT-B/16, the neighbor ACC and R@1 are the same since both are computed using cosine similarity for ranking.

  • Only Gallery Shift: In this setting, there is no distribution shift in the query modality. For the baseline BLIP ViT-B/16, the IR@1 and TR@1 without any query or gallery shift are 57.1% and 74.0%, respectively.
IR/Gallery Shift TypesGauss.ShotImpul.SpeckleDefoc.GlassMotionZoomSnowFrostFogBrit.Contr.ElasticPixelJPEGAvg.
BLIP ViT-B/16 (ACC & R@1)35.437.435.944.438.553.235.815.236.543.047.751.732.536.220.348.938.3
TCR (ACC)36.338.036.645.239.153.737.416.438.444.249.252.533.138.521.849.539.4
TCR (R@1)36.538.637.145.239.953.837.416.838.344.549.352.433.538.521.849.639.6
评论
TR/Gallery Shift TypesOCRCICRCSCDSRRIRSRDIPFormalCasualPassiveActiveBacktransAvg.
BLIP ViT-B/16 (ACC & R@1)49.723.120.134.522.859.764.865.266.973.173.272.571.573.671.156.1
TCR (ACC)55.526.823.539.527.566.270.970.471.377.176.576.275.177.375.460.6
TCR (R@1)55.526.823.539.027.765.970.970.671.677.276.776.275.277.375.360.6
  • Both Query and Gallery Shift: In this setting, we choose the OCR / Gaussian corruptions as the query shift for image / text retrieval, respectively. For the baseline BLIP ViT-B/16, the IR@1 / TR@1 with OCR / Gaussian corruptions is 31.4% / 43.4%.
IR/Gallery Shift TypesGauss.ShotImpul.SpeckleDefoc.GlassMotionZoomSnowFrostFogBrit.Contr.ElasticPixelJPEGAvg.
BLIP ViT-B/16 (ACC & R@1)18.519.818.623.420.128.819.08.218.922.525.527.617.118.610.426.020.2
TCR (ACC)20.621.820.626.022.631.621.59.321.625.528.830.518.921.812.128.422.6
TCR (R@1)20.621.820.626.022.631.721.59.421.625.628.830.519.021.812.228.422.6
TR/Gallery Shift TypesOCRCICRCSCDSRRIRSRDIPFormalCasualPassiveActiveBacktransAvg.
BLIP ViT-B/16 (ACC & R@1)27.213.211.118.427.232.536.337.238.942.242.842.741.043.241.033.0
TCR (ACC)35.716.015.224.735.742.147.146.048.253.053.853.051.953.151.741.8
TCR (R@1)35.216.515.324.635.242.247.745.448.352.852.452.752.153.552.141.7

From the results, one could observe that gallery shift degrades both retrieval performance and nearest neighbor selection accuracy, whether in only gallery shift or both query and gallery shift settings. However, the proposed TCR improves the retrieval performance under gallery shift, with the selected nearest neighbors more likely to be correct. Besides, even under gallery shift setting, TCR could enhance retrieval performance surpassing the baseline performance without gallery shift. For example, in the both query and gallery shift setting, the text retrieval performance of TCR under RI (47.7%), RS (45.4%), Formal (52.4%), and Passive (52.1%) gallery shift exceeds the baseline performance without gallery shift (43.4%). It’s worth noting that in real-world scenarios, data with only gallery shift is rare, as the data in the gallery is often extensive and curated. In contrast, the queries of the users are more diverse, which might lead to the distribution shift challenge.

审稿意见
8

The paper presents a Test-time adaptation for Cross-modal Retrieval (TCR) method to address query shift, which is a critical and understudied problem in cross-modal retrieval tasks. Query shift occurs when the distribution of online query streams differs from the source domain, leading to performance degradation in existing models. TCR introduces a query prediction refinement module and a joint objective function to refine query predictions and prevent query shift from disturbing the common space. It improves the existing test-time adaptation (TTA) methods with the capacity to manipulate both the modality uniformity and modality gap. Overall speaking, this paper is well-organized and of practical value.

优点

The proposed TCR method addresses an important problem and is supported by strong experimental results.

It provides extensive experiments demonstrating the effectiveness of TCR against query shift. The comparisons with existing TTA methods show convincing improvements, with is a strong validation of the ablation study .

缺点

In Section 4.2, it is said that “We compare TCR with five SOTA TTA methods (Tent (Wang et al., 2021), EATA(Niu et al.,2022), SAR(Niu et al.,2023), READ(Yang et al.,2024), and DeYO...”. These methods should be introduced in Section 2.2 of the related work part.

Line 212, ,where Q and G denotes as query modality and gallery modality for clarity in the following. Change to “denote””

Tables 1 and 2 appear too early. They should not be on Page 7 but on the page where they are referred for the first time,

问题

see above comment

伦理问题详情

n/a

评论

Thanks for your valuable reviews. We would like to address your concerns one by one in the following.

Q1: In Section 4.2, it is said that “We compare TCR with five SOTA TTA methods (Tent (Wang et al., 2021), EATA(Niu et al.,2022), SAR(Niu et al.,2023), READ(Yang et al.,2024), and DeYO...”. These methods should be introduced in Section 2.2 of the related work part.

A1: Thanks for your valuable suggestions. We apologize for the missing details on the baselines. In response to your constructive feedback, we provide more detailed introdution of the baselines, which can be found in Appendix B.3 due to space limitations. For your convenience, we attach the added statement as follows.

Test-time Adaptation (TTA) aims to reconcile the distribution shifts in an online manner. Towards achieving this goal, fully TTA ([A]) has been proposed, which fine-tunes the BatchNorm layers by minimizing entropy during the test phase. EATA ([B]) employs a Fisher regularizer to limit excessive model parameter changes and filter out high-entropy samples via the selection strategy. SAR [C] removes high-gradient samples and promotes flat minimum weights, enhancing robustness against more challenging TTA scenarios such as mixed domain shifts, single-sample adaptation, and imbalanced label shifts. READ ([D]) proposes a noise-robust adaptation loss and reliable fusion module to tackle the reliability bias challenge in the multi-modal setting. DeYO ([E]) reveals the unreliability of treating entropy as the confidence metric and establishes a novel metric by measuring the difference between predictions before and after applying an object-destructive transformation.

Q2: Line 212, ,where Q and G denotes as query modality and gallery modality for clarity in the following. Change to “denote”

A2: Thanks for your careful reading. We apologize for the typos and have revised them in the updated manuscript.

Q3: Tables 1 and 2 appear too early. They should not be on Page 7 but on the page where they are referred for the first time.

A3: Thanks for your valuable suggestions. We apologize for the misplacement of Table 1 and Table 2 in the submission and have revised them in the updated manuscript.

Reference:

[A] Dequan Wang, Evan Shelhamer, Shaoteng Liu, Bruno Olshausen, and Trevor Darrell. Tent: Fully test-time adaptation by entropy minimization. In ICLR, 2021.

[B] Shuaicheng Niu, Jiaxiang Wu, Yifan Zhang, Zhiquan Wen, Yaofo Chen, Peilin Zhao, and Mingkui Tan. Towards stable test-time adaptation in dynamic wild world. In ICLR, 2023.

[C] Shuaicheng Niu, Jiaxiang Wu, Yifan Zhang, Yaofo Chen, Shijian Zheng, Peilin Zhao, and Mingkui Tan. Efficient test-time model adaptation without forgetting. In ICML, 2022.

[D] Mouxing Yang, Yunfan Li, Changqing Zhang, Peng Hu, and Xi Peng. Test-time Adaption against Multi-modal Reliability Bias. In ICLR, 2024.

[E] Jonghyun Lee, Dahuin Jung, Saehyung Lee, Junsung Park, Juhyeon Shin, Uiwon Hwang, and Sungroh Yoon. Entropy is not enough for test-time adaptation. In ICLR, 2024.

评论

Thanks for the authors's feedback.

Compared methods in the experiments should be most relevant to the proposed work, but even their names did not appear in the related work part in the main text. I can understand due to page limit their details can be moved to appendix, but they should be at least mentioned in related work.

评论

Thanks for your valuable suggestions. The five baselines compared in the manuscript (Tent, EATA, SAR, READ, and DeYO) are the most relevant works to the proposed TCR. Specifically, Tent is the most classic TTA method; EATA and SAR are robust TTA methods designed to handle noisy predictions; DeYO is the state-of-the-art method for unimodal recognition, while READ is the state-of-the-art for multimodal recognition. Since the proposed TCR might be one of the first TTA works for cross-modal retrieval, we select the above baselines for comparison.

In response to your insightful suggestion, we have cited these methods in Related Work (Sec. 2.2 in the manuscript). For your convenience, we attach the revised content as follows.

To avoid the reduplicated training cost of the source model, Fully Test-Time Adaptation paradigm (Tent) has been proposed, which could be coarsely divided into the following three categories: i) online TTA methods (e.g., DeYO), which continually update the normalization layers by resorting to the unsupervised objectives, such as entropy minimization or its variants. ii) robust TTA methods (e.g., EATA, SAR), which strive to improve the robustness against noisy predictions, mixed distribution shifts, label shifts, and so on. iii) TTA beyond recognition, which focuses on the tasks including but not limited to image restoration, multimodal recognition (e.g., READ), and multimodal segmentation.

We would be happy to discuss further if needed.

审稿意见
6

The paper addresses the challenge of cross-modal retrieval in scenarios where the query data distribution deviates from the source domain, a phenomenon known as "query shift." This deviation often leads to a performance decline in cross-modal retrieval systems. The authors propose a novel approach called TCR: Test-time adaptation for Cross-modal Retrieval, which adapts cross-modal retrieval models during inference to account for query shift. The proposed method includes a query prediction refinement module and a joint objective function to prevent the disturbances caused by the query shift, enhancing the uniformity within the query modality and minimizing the gap between query and gallery modalities. The model is designed to operate effectively in real time by adapting to changing online queries. The approach was tested on six popular image-text datasets and demonstrated superior performance against existing test-time adaptation (TTA) techniques.

优点

  1. The paper tackles the underexplored problem of query shift in cross-modal retrieval, providing a comprehensive analysis of its effects on retrieval performance. The method's unique combination of query prediction refinement and multiple loss functions sets it apart from traditional TTA approaches.
  2. The authors conducted extensive experiments across six datasets and compared their method against several state-of-the-art TTA models. The amount of experiments is fair and convincing.
  3. The paper proposes a joint objective consisting of three loss functions—uniformity learning, gap minimization, and noise-robust adaptation—that each address specific challenges introduced by query shift. This is a novel design for this problem.

缺点

  1. The authors provide only limited discussion regarding the sensitivity of the various hyperparameters involved, such as the temperature and trade-off parameters. A more detailed analysis would improve understanding of the model's adaptability to different scenarios.
  2. The approach heavily relies on pre-trained models and assumes the existence of a well-aligned common space. In cases where the source domain model lacks robust representations, the effectiveness of TCR may be diminished. This could limit the generalizability of the approach to pre-trained models of different quality. More discussions and insights need to be given in order to make the paper more readable.

问题

  1. The proposed TCR method aims to enhance retrieval robustness under query shift by manipulating modality uniformity and the modality gap. Given the variety of potential shifts in real-world data (e.g., subtle cultural variations, extreme distortions, rare domain-specific content), how does TCR perform across these different types of shifts?

  2. Could the model's performance degrade if it encounters shifts it was not explicitly evaluated against? A thorough breakdown of the model’s robustness to a diverse set of query shifts would strengthen the understanding of its general applicability.

  3. This paper introduces several hyperparameters, such as the temperature parameter (τ) for controlling the trade-off between smoothness and sharpness, and others for balancing the different loss terms. How sensitive is TCR to these hyperparameters, and how easy is it to tune them for new domains? Some more results from these ablations studies will be very beneficial.

评论

Q.3: This paper introduces several hyperparameters, such as the temperature parameter (τ) for controlling the trade-off between smoothness and sharpness, and others for balancing the different loss terms. How sensitive is TCR to these hyperparameters, and how easy is it to tune them for new domains? Some more results from these ablations studies will be very beneficial.

A.3: Thanks for your comments. There are two hyperparameters in the paper, i.e., the temperature τ\tau for controlling the trade-off between smoothness and sharpness, the trade-off parameter tt for controlling the intra-modality uniformity. For parameter sensitivity analysis, we have conducted ablation studies about τ\tau and tt in Fig. 4(a) and Fig. 4(b), respectively. For your convenience, we attach the corresponding numerical results of Fig. 4(a) in the following tables.

τ\tau1e-51e-41e-35e-30.010.020.050.1
Base2COCO52.453.052.955.756.457.857.055.2
ICFG2CUHK32.933.934.534.635.836.235.934.7
Flickr-C58.862.364.066.067.369.165.264.7
Base2Fashion21.121.523.125.025.826.424.713.7

The results denote that TCR demonstrates stable performance within the range of [0.001,0.05] and achieves the best performance when τ=0.02\tau=0.02.

In response to your concern, we conduct more ablations studies on the trade-off parameter tt under "Flickr-C", "ICFG2CUHK", "Base2Fashion" settings. The results are summarized in Fig. 4(b) within the updated manuscript. For your convenience, we attach the corresponding numerical results (regarding Recall@1) in the following tables.

t0.11.02.05.010.020.0100.0
Base2COCO58.458.458.858.959.058.558.0
ICFG2CUHK37.037.137.337.337.337.337.0
Flickr-C67.868.268.268.468.468.267.9
Base2Fashion26.927.027.027.127.427.127.0

The results indicate that TCR is not sensitive to the choice of the parameter tt.

Besides, we apologize for the omission of the sensitivity analysis for these hyperparameters. In the revised manuscript, we have supplemented the detailed parameter analysis. For your convenience, we attach the added statement as follows.

As shown in Fig. 4(a), we observe that: i) selecting an appropriate temperature for the existing TTA approach across various datasets is challenging; ii) even a low temperature (e.g., 1e41e-4) is a better setting across all datasets, the performance degrades as a low temperature tends to make model overfitting on noisy query prediction. In contrast, the query prediction refinement module not only stabilizes the temperature setting for all the datasets but also prevents the model from either underfitting or overfitting by excluding some irrelevant samples in the gallery. Besides, TCR demonstrates stable performance within the range of [0.001,0.05] and achieves the best performance when τ=0.02\tau=0.02. As depicted in Fig. 4(b), one could observe that TCR is not sensitive to the choice of tt.

评论

Q4: The approach heavily relies on pre-trained models and assumes the existence of a well-aligned common space. In cases where the source domain model lacks robust representations, the effectiveness of TCR may be diminished. This could limit the generalizability of the approach to pre-trained models of different quality. More discussions and insights need to be given in order to make the paper more readable.

A4: Thanks for your comments. We acknowledge that TCR relies on the well-aligned common space, which is essential for achieving good intra-modality uniformity and inter-modality gap. In other words, if the source model is suboptimal, the performance improvements of TCR might be less pronounced. However, we highlight that TCR outperforms all baselines and achieves a stable performance improvement across various pre-trained model types and sizes, including BLIP ViT-B/16, BLIP ViT-L/16, CLIP ViT-B/16, and CLIP ViT-B/32. Indeed, any test-time adaptation paradigm heavily depends on the high-quality source model [A] [B] [F], i.e., the high-quality source model would provide more reliable predictions that support performance improvement.

Reference:

[A] Dequan Wang, Evan Shelhamer, Shaoteng Liu, Bruno Olshausen, and Trevor Darrell. Tent: Fully test-time adaptation by entropy minimization. In ICLR, 2021.

[B] Shuaicheng Niu, Jiaxiang Wu, Yifan Zhang, Zhiquan Wen, Yaofo Chen, Peilin Zhao, and Mingkui Tan. Towards stable test-time adaptation in dynamic wild world. In ICLR, 2023.

[C] Jielin Qiu, Yi Zhu, Xingjian Shi, Florian Wenzel, Zhiqiang Tang, Ding Zhao, Bo Li, and Mu Li. Benchmarking Robustness of Multimodal Image-Text Models under Distribution Shift. Journal of Data-centric Machine Learning Research, 2023.

[D] Xiaoqiang Lu, Binqiang Wang, Xiangtao Zheng, and Xuelong Li. Exploring models and data for remote sensing image caption generation. IEEE Transactions on Geoscience and Remote Sensing, 2017.

[E] Zhiqiang Yuan, Wenkai Zhang, Kun Fu, Xuan Li, Chubo Deng, Hongqi Wang, and Xian Sun. Exploring a fine-grained multiscale method for cross-modal remote sensing image retrieval. IEEE Transactions on Geoscience and Remote Sensing, 2021.

[F] Shuaicheng Niu, Jiaxiang Wu, Yifan Zhang, Yaofo Chen, Shijian Zheng, Peilin Zhao, and Mingkui Tan. Efficient test-time model adaptation without forgetting. In ICML, 2022.

[G] Jonghyun Lee, Dahuin Jung, Saehyung Lee, Junsung Park, Juhyeon Shin, Uiwon Hwang, and Sungroh Yoon. Entropy is not enough for test-time adaptation. In ICLR, 2024.

评论

Dear reviewer Wks8,

We would like to know if our response has addressed your concerns and questions. If you have any further concerns or suggestions for the paper or our rebuttal, please let us know. We would be happy to engage in further discussion and manuscript improvement.

Thank you again for the time and effort you dedicated to reviewing this work.

评论

Thanks for your constructive reviews and suggestions. In the following, we will answer your questions one by one.

Q1: Given the variety of potential shifts in real-world data (e.g., subtle cultural variations, extreme distortions, rare domain-specific content), how does TCR perform across these different types of shifts?

A1: Thanks for your constructive comments. In the submission, we have reported the experiment results on a variety of real-world query shift, including noise, blur, and weather (e.g., frost, snow, and fog) distortions. Besides, we have evaluated TCR on the rare domains, such as ReID domain, e-commerce domain and natural image domain. Notably, we have carried out the experiments to verify that TCR could handle the maximum severity level of the corruptions (Tables 1-2 and Tables 7-8). In other words, TCR is able to address the most extreme distortions, which are highlighted and widely acknowledged in previous works ([A] [B] [C]).

In response to your insightful suggestion, we conduct additional experiments in the even rarer remote sensing domain. Specifically, we choose the BLIP as the source model and perform zero-shot retrieval on the remote sensing datasets RSICD ([D]) and RSITMD ([E]). To verify the effectiveness of TCR, we choose the typical TTA method Tent ([A]), the SOTA TTA methods EATA ([F]) and DeYO ([G]) as the baselines for comparisons. Results are summarized in Table 20 within the revised manuscript. For your convenience, we attach the corresponding results in the following tables.

Base2RSICDTR@1IR@1
BLIP ViT-B/166.46.8
Tent5.75.4
EATA6.96.7
DeYO6.46.5
Ours8.57.1
Base2RSITMDTR@1IR@1
BLIP ViT-B/167.610.4
Tent7.99.3
EATA8.010.4
DeYO7.710.0
Ours8.410.7

The results indicate that TCR could also achieve the best performance in even rarer remote sensing domain.

Q.2: Could the model's performance degrade if it encounters shifts it was not explicitly evaluated against? A thorough breakdown of the model’s robustness to a diverse set of query shifts would strengthen the understanding of its general applicability.

A.2: Thanks for your comments. In this paper, we have evaluated TCR on the Flickr and COCO datasets with image and text corruptions, which simulate real-world query shift. It is worth noting that we have introduced a total of 130 perturbations across various severity levels, comprising 80 for the image modality and 50 for the text modality. Specifically, the image modality includes 16 types of corruption, each with 5 levels of severity, while the text modality comprises 15 corruption types with 7/2/1 severity levels for character-level/word-level/sentence-level corruptions. Besides, we conduct experiments on the datasets with real-world query shift, including the ReID, e-commerce, and natural image domains.

In Appendix D.3 of the manuscript, we have reported the experiment results on the 130 perturbations in Fig. 7 and Fig. 8. Specifically, we carry out the experiments on the COCO-C benchmark and report the average performance across various severity levels for each corruption. For your convenience, we have included a summary of these results (regarding Recall@1) in the following tables.

评论
BLIP ViT-B/16Gauss.ShotImpul.SpeckleDefoc.GlassMotionZoomSnowFrostFogBrit.Contr.ElasticPixelJPEGAvg.
Tent64.364.458.870.250.374.045.820.629.863.468.673.959.766.942.772.557.9
EATA64.165.562.170.364.274.962.626.955.564.470.874.666.467.552.572.863.4
SAR63.965.361.869.860.674.258.222.947.764.069.374.063.067.551.772.261.6
READ64.663.959.169.058.574.061.022.249.962.069.073.663.065.848.671.961.0
DeYO65.466.364.270.262.974.661.222.552.865.571.974.366.067.750.772.763.1
Ours67.667.867.171.267.875.866.246.462.168.674.075.671.171.859.373.267.8
BLIP ViT-B/16OCRCICRCSCDSRRIRSRDIPFormalCasualPassiveActiveBacktransAvg.
Tent42.027.124.531.127.845.552.851.351.157.056.656.455.357.254.046.0
EATA43.330.529.034.630.746.253.251.651.356.656.656.355.657.254.247.1
SAR42.029.728.334.030.145.051.850.550.956.856.556.254.956.854.246.5
READ42.930.028.634.230.245.853.251.751.456.756.556.255.257.053.946.9
DeYO43.429.628.234.530.446.253.651.751.356.756.656.355.657.154.247.0
Ours44.031.730.335.231.546.654.052.051.657.357.156.856.057.354.747.7

The results demonstrate that TCR outperforms all baselines across various severities and corruptions, showcasing its robustness against different distribution shifts.

Besides, we conduct more experiments on the remote sensing datasets RSICD ([D]) and RSITMD ([E]), which might encounter the query shift not explicitly evaluated in the paper. For your convenience, we have included a summary of these results in the following tables.

Base2RSICDTR@1IR@1
BLIP ViT-B/166.46.8
Tent5.75.4
EATA6.96.7
DeYO6.46.5
Ours8.57.1
Base2RSITMDTR@1IR@1
BLIP ViT-B/167.610.4
Tent7.99.3
EATA8.010.4
DeYO7.710.0
Ours8.410.7

The results underscore that TCR achieves superior performance on the RSICD, RSITMD datasets from the remote sensing domain, further validating its effectiveness on unevaluated query shifts in the paper.

AC 元评审

This paper tackles the challenging 'query shift' problem in cross-modal retrieval, in which the distribution of query data deviates from the source domain. They formulate their algorithm as a test-time adaptation for cross-modal retrieval, which includes a query prediction refinement module to refine the query predictions and a joint objective function to prevent the disturbances caused by the query shift. Sufficient experimental results on multiple datasets show the effectiveness of the proposed method in such different cases.

In the first round, the main concerns lie in the types of query shifts, robustness to a diverse set of query shifts, parameters, and other minor issues. After rebuttal, all these concerns were eliminated, and all the reviewers voted to accept this work.

After carefully reading the main document and supplementary docs, the AC has the following comments about this paper:
Advantages
This paper addresses the query shift problem in cross-modal retrieval during test time. The problem itself is practical due to the diverse and unpredictable nature of open-world queries. Through extensive experimental analyses, the authors identify two key challenges posed by query shift and propose an effective method to address them. The manuscript and appendix include thorough empirical studies validating the significance of the addressed problem and the effectiveness of the proposed approach.
Future Direction
Although the authors pointed out the query shift issue in cross-modal retrieval, how to effectively handle more potential scenarios contaminated with query shift should be further discussed and highlighted.

In summary, due to novel problem definition, technical context and experimental results, this paper is a good work to be presented in ICLR 2025.

审稿人讨论附加意见

All four reviewers recommend acceptance, with three giving clear acceptance and one providing a borderline acceptance. The reviewers recognize the contributions of this work. Overall, I believe this paper provides valuable insights for the test-time adaptation (TTA) and cross-modal retrieval communities and strongly recommend its acceptance.

最终决定

Accept (Spotlight)