PaperHub
5.3
/10
Rejected3 位审稿人
最低5最高6标准差0.5
6
5
5
4.0
置信度
正确性2.7
贡献度2.0
表达3.0
ICLR 2025

Harnessing Shallow Features in Pre-Trained Models for Out-of-Distribution Detection

OpenReviewPDF
提交: 2024-09-24更新: 2025-02-05

摘要

关键词
out-of-distribution detectionlong-tail learning

评审与讨论

审稿意见
6

With the favor of transformer architecture, this work proposes an out-of-distribution (OOD) score based on the fused feature entailed from each layer. Further, it is empirically observed that parameter-efficient fine-tuning (PEFT) is more robust than fully fine-tuning when using different learning rates. Finally, it demonstrates the effectiveness of the proposed OOD score along with PEFT on the OOD benchmark and also long-tailed OOD benchmark.

优点

  • The proposed OOD score based on the fused feature from each layer is novel.
  • This work aims to tackle the OOD detection on both class-balanced benchmark and long-tailed benchmark.

缺点

  • The title appears somewhat overstated, as the proposed method specifically requires transformer architectures. A more precise title, such as "... in pre-trained transformers for ...", would better reflect the scope of the method.
  • Line 103-104: The mention of "measuring the degree of neural collapse of each layer" is unclear. Could the authors elaborate on this point and clarify how it relates to determining the importance weights for each layer.
  • The proposed OOD score aligns with post-hoc OOD scores that require access to training data. To assess its effectiveness, the authors should consider evaluating it on a standard OOD benchmark, as outlined in [2]. Comparisons with other training-data-dependent OOD scores—such as KL matching, Residual, and ViM—would provide a more comprehensive evaluation (see Table 1 in [1] for reference).
  • It is unclear why the authors chose to fine-tune a network pre-trained on ImageNet-21k. Why not instead use models pre-trained directly on ImageNet-1k or CIFAR-100 and then perform OOD detection on a standard benchmark[2]?
  • Dependency on Fine-tuning: Could the authors discuss whether the proposed OOD score's effectiveness relies on the pre-trained models with extensive datasets? For instance, the selected models in this work is pre-trained ViT models on ImageNet-21k and CLIP-ViT-B/16. Have you tested or considered testing with smaller pre-trained models or datasets to assess its effectiveness across different scales of pre-training.
  • Comparison with MCM (Line 394): The comparison with MCM, in which the authors claim that SFM outperforms MCM by ~2%, may not be fair. The original MCM model was not fine-tuned on ImageNet-1k, which could impact the validity of this comparison.

Reference

  • [1] GEN: Pushing the Limits of Softmax-Based Out-of-Distribution Detection. In CVPR, 2023.
  • [2] OpenOOD v1.5: Enhanced Benchmark for Out-of-Distribution Detection. In NeurIPS 2023 Workshop on Distribution Shifts.

问题

See weakness.

评论

[W4] Why chose to fine-tune a network pre-trained on ImageNet-21k?

[A4] The main reason is that our downstream tasks are ImageNet-1k and CIFAR-100. We aim to avoid overfitting of the model to the downstream datasets, as an overly fitted model would render the tasks overly simplistic and might not provide a comprehensive and accurate assessment of the method's performance and generalization ability. To ensure a fair comparison, we have chosen to follow the practices of existing works such as NeCo and ViM. Moreover, to further guarantee that the pre-training dataset and the downstream task datasets do not have excessive overlap, we have also conducted experiments using CLIP. This multi-faceted approach allows us to better evaluate the effectiveness and robustness of our proposed method across different datasets and pre-training setups, and we believe it contributes to a more reliable and meaningful research outcome.

评论

Dear Reviewer cJWd, We are extremely grateful for the comprehensive and detailed review you have provided. Your insights and suggestions are invaluable in helping us improve our work.

[W1] The title appears overstated.

[A1] We are truly and sincerely grateful for your astute and perceptive advice regarding the title of our work. After meticulous and in-depth consideration, we have come to recognize the validity of your observation. Indeed, a more precise and accurate title is essential to aptly reflect the core scope and nature of our proposed method. We wholeheartedly concur with your suggestion and will accordingly modify the title to "Harnessing Shallow Features in Pre-Trained Transfomers for Out-of-Distribution Detection". This revised title, we believe, will not only enhance the clarity and specificity of our research's representation but also align more closely with the technical details and contributions of our work. Once again, we express our profound appreciation for your invaluable feedback, which has significantly contributed to the refinement and improvement of our manuscript.

[W2] An explanation for "measuring the degree of neural collapse of each layer".

[A2] We are extremely grateful for your incisive query regarding the mention of "measuring the degree of neural collapse of each layer" and its relation to determining the importance weights. Our fundamental motivation was grounded in the convergence behavior of the last layer. In machine learning, the typical aspiration is for the class center, classifier, and samples to converge. However, this poses a detriment to OOD detection as OOD samples may also converge to the same class-center, and this is equally true for OOD detection scores based on Mahalanobis distance. To address this, we aimed to introduce less convergent information into the sample features, thereby enabling them to escape the neural collapse condition, while still retaining a certain degree of convergent information. Through comprehensive statistical analysis and comparisons, we devised a method to dissolve the weights of different layers. We posit that the weights we currently employ can, to a certain extent, represent the convergence degree of each layer. For a more detailed understanding, please refer to the definition of alpha in Equation 5. To illustrate, if all samples within the same layer achieve optimal convergence, known as neural collapse, the variance of feature samples in that layer should be zero, meaning that the covariance matrix is zero except along the diagonal. Hence, moving towards neural collapse requires minimizing the off-diagonal elements while maximizing the diagonal elements, or the trace. Consequently, we assert that the definition of alpha provided by Eq5 effectively quantifies the extent of neurocollapse. We also acknowledge that the expression in lines 103 - 104 was suboptimal and lacked clarity. We will undertake a thorough revision of this description in the subsequent iteration to enhance the overall readability and comprehensibility of the paper.

评论

[W3] To assess effectiveness.

[A3] We greatly appreciate your valuable feedback. We have seriously considered how to more suitably assess the proposed OOD score. We've carefully analyzed the dataset in OpenOOD v1.5. In terms of using CIFAR-10 as the in-distribution (ID) dataset, we found it to be relatively straightforward. As you might be aware, most methods we investigated achieve nearly perfect results, with the AUROC almost reaching 100 and the FPR95 being 0.00 when utilizing the ViT model. Currently, our CIFAR100 test set diverges only from Mnist. Therefore, for the sake of simplicity and also due to the time constraints during the rebuttal phase, we've opted to repeat the experiment with Imagenet-1k as the ID dataset in the OpenOOD v1.5 benchmark. Furthermore, we chose Imagenet-1K-LT as an ID as well since earlier experiments showed it is quite similar to Imagenet-1k but saves time. The outcomes of these experiments are shown in the following table. We sincerely hope that this strategy provides a more meaningful and challenging benchmark to evaluate our proposed OOD score while acknowledging practical constraints such as limited time.

Results for ImageNet-1k-LT in-distribution dataset on OpenOOD v1.5

MethodNinco (AUROC)Ninco (FPR95)Openimage-o (AUROC)Openimage-o (FPR95)Ssb-hard (AUROC)Ssb-hard (FPR95)Imagenet_c (AUROC)Imagenet_c (FPR95)Imagenet_es (AUROC)Imagenet_es (FPR95)Imagenet_r (AUROC)Imagenet_r (FPR95)Imagenet_v2 (AUROC)Imagenet_v2 (FPR95)Average (AUROC)Average (FPR95)
MSP87.8150.0293.7227.5176.7268.4367.9178.5869.3569.2679.7359.1557.5789.9276.1263.27
MLS91.5942.8096.2818.6581.2563.8670.5476.7772.1166.7983.6553.2257.8690.0979.0458.88
Energy92.1239.6296.8016.0281.8761.5370.8076.0272.4466.0684.2550.5257.7990.1979.4457.14
Mahalanobis94.0032.5197.5812.6185.0152.1773.9372.6473.0467.0885.3248.9558.0290.8180.9953.83
residual83.8762.4592.4133.8884.8756.1974.9678.0365.2582.8775.0576.4653.0394.3875.6369.18
Vim92.2938.6596.9415.3282.3660.4071.1975.3972.4765.9884.3750.2057.7990.1579.6356.58
NECO91.9738.0996.9015.1984.8154.9670.5575.4872.0167.6182.4353.8656.8690.4479.3656.52
KL-matching90.5341.6395.9518.1579.5263.0270.0375.9171.5466.3582.5652.6058.3389.8578.3558.22
SFM (ours)94.9826.7498.219.7286.3449.4683.9653.5776.7863.4588.4942.8558.3691.3683.8848.16
--------------------------------------------------------------------
MSP80.1168.9488.2246.7268.0683.6673.4470.4470.3168.4677.2764.0057.1290.7873.5070.43
MLS84.1767.1192.9335.1771.9982.4477.6667.6575.6664.2184.6155.9658.2490.3877.8966.13
Energy84.1567.7893.7330.5972.2582.4378.0267.1376.5162.5185.8752.2658.2290.2878.3964.71
Mahalanobis75.1383.2886.9563.8266.1189.4982.6862.6484.2752.4690.0247.3358.1890.3177.6269.90
residual61.5691.5470.4381.3561.0092.6382.8668.2886.3054.6586.2857.4156.4792.0672.1376.85
Vim83.9168.2093.5731.1572.2882.5778.8565.0677.5260.3486.7349.7958.3589.8878.7463.86
NECO80.9671.9092.4837.1369.2285.1077.8468.1778.6961.7686.0054.9258.1889.8377.6366.97
KL-matching83.2166.8892.1334.0870.6381.7576.0967.0772.9264.2581.7055.4657.8390.5176.3665.71
SFM (ours)78.3074.6792.9036.9371.6481.7985.4950.9487.8443.4988.7650.3158.2289.5080.4561.09
评论

Thank you for conducting the comparative experiments. If I understand correctly, the results presented in the first half of the block use ImageNet-1k as the ID dataset, while the second half of the block uses ImageNet-1k-LT as the ID dataset. Additionally, all post-hoc scores are calculated using the fine-tuned checkpoints obtained with the proposed loss. This indicates that the effectiveness of the proposed post-hoc method depends on the additional fine-tuning step.

  • I am curious, if you directly use the pre-trained checkpoints with ImageNet-1k as the ID dataset, what would the performance look like?
  • If the proposed post-hoc score depends on fine-tuning, it would be beneficial to include a comparison of classification accuracy before and after fine-tuning. This would help demonstrate that the improvement in OOD detection performance does not come at the expense of classification accuracy.
评论

Dear Reviewer cJWd,

Thank you for your thoughtful comments and for your accurate understanding of our proposed method. Below, we provide detailed responses to address your additional concerns.

[Q1] Performance for directly use the pre-trained checkpoints with ImageNet-1k as the ID dataset.

[A1] Thank you for your suggestion! We have included the comparison results for applying our proposed method to both the pre-trained model and the fine-tuned model in the tables below. As expected, the fine-tuned model demonstrates superior performance due to its ability to learn more discriminative features specific to the ID datasets. Interestingly, even when applied directly to the pre-trained model, our method achieves competitive performance, highlighting its effectiveness.

Table 1: Results for using ImageNet-1k as ID dataset

Texture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
Pre-trained model97.199.8882.7265.1486.4154.6498.804.7593.4031.4591.7033.17
Fine-tuned model96.6511.7089.6446.0092.0437.7899.402.2693.7629.8094.3025.51


Table 2: Results for using ImageNet-1k-LT as ID dataset

Texture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
Pre-trained model97.1510.2081.5167.2085.4956.4998.695.1393.1532.1591.2034.23
Fine-tuned model96.9211.7989.8245.3692.1836.1699.332.5193.4631.1094.3425.38

[Q2] Comparison of classification accuracy before and after fine-tuning.

[A2] Thanks for your suggestion again. We present the accuracy of the pre-trained model and after the fine-tuning in the table below. More detailed performance comparisons are reported in Table 7 (Appendix A) of our paper. The results demonstrate that fine-tuning is essential for boosting ID classification accuracy. This justify that the improvement in OOD detection performance does not come at the expense of classification accuracy.

Table 3: Comparison of classification accuracy for the pre-trained model and the fine-tuned model.

Pre-trained modelFine-tuned model
ImageNet-1k80.0383.50
ImageNet-1k-LT79.2581.79
CIFAR-10077.6293.47
CIFAR-100-LT75.5489.99

Thank you for your valuable comments! We will include the experimental analysis in the paper and further discuss the implications.

评论

Thanks for conducting the suggested experiments!

  • It would be beneficial to include the results for other post-hoc methods such as ViM when utilizing the pre-trained checkpoints as the backbone.
  • I double-checked Table 7 in Appendix A and confirmed that the numbers presented represent the accuracy for the original classification tasks. However, it appears that the accuracy shown in Table 7 reflects the results after fine-tuning, while the accuracy before fine-tuning is missing. Could you please clarify this?
评论

We are highly appreciative of your valuable suggestion regarding the comparison with other training-data-dependent OOD scores. We have indeed made efforts in this regard. As you can see, we have presented the detailed results of Residual and ViM within our experimental results section. Additionally, to further address your concern and provide a more comprehensive evaluation, we have also includ the KL-matching results in the table below for direct comparison. This will enable a more in-depth analysis and a better understanding of the performance and characteristics of our proposed method in relation to these benchmarks. We hope that this additional information will enhance the overall quality and persuasiveness of our research. The results of ImageNet-21k pre-trained ViT are displayed in the top half of each table, while the results on CLIP are displayed in the bottom half.

Results for CIFAR-100 in-distribution dataset

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
KL-matching98.606.1096.6614.9396.3417.1290.0534.1788.1549.3493.6728.2193.9124.98
SFM (ours)100.00.0099.501.9196.4719.5299.801.11100.00.00100.00.0099.293.76
------------------------------------------------------------
KL-matching94.3225.1290.6938.2590.6938.5284.1652.9477.8570.9686.9947.8087.4545.60
SFM (ours)99.950.0298.318.6288.5653.9797.5412.9199.930.0699.950.0297.3712.60

Results for CIFAR-100-LT in-distribution dataset

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
KL-matching98.486.4097.4412.1194.0026.8887.5638.9186.6552.7892.9431.0192.8428.02
SFM (ours)100.00.0099.750.4394.2229.8699.751.1299.990.0099.990.0198.955.24
------------------------------------------------------------
KL-matching95.0121.7690.7631.6988.8744.1781.6857.9379.3163.6587.6443.2187.2143.73
SFM (ours)99.920.0297.5015.6585.2060.4195.0526.0599.920.0499.930.0196.2517.03
评论

Results for ImageNet-1k in-distribution dataset

MethodTexture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
KL-matching87.8540.9286.7653.0287.8949.1997.848.8486.2549.2089.3240.23
SFM (ours)96.6511.7089.6446.0092.0437.7899.402.2693.7629.8094.3025.51
----------------------------------------------------
KL-matching86.6446.4583.2859.2583.2161.2394.1824.9983.1962.4586.1050.87
SFM (ours)89.1648.4491.8836.4693.2434.8795.4723.4982.2770.3590.4142.72

Results for ImageNet-1k-LT in-distribution dataset

MethodTexture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
KL-matching88.7238.7187.4150.0389.1445.8398.446.1987.2447.5090.1937.65
SFM (ours)96.9211.7989.8245.3692.1836.1699.332.5193.4631.1094.3425.38
----------------------------------------------------
KL-matching85.3551.5682.8457.0082.5157.5694.5423.3682.5264.0085.5550.70
SFM (ours)90.9543.1092.4334.8092.6239.0794.5928.5483.6268.9090.8442.88
评论

[W6] May not be fair to compare with MCM.

[A6] The MCM method is naturally better suited for zero-shot OOD tasks compared to fine-tuning tasks. The prevalent fine-tuning approach, which mainly targets the visual encoder, tends to disrupt the initial alignment between the visual and text components after fine-tuning, resulting in less effective outcomes. Our goal in including the MCM method in our experiment was not to make a direct comparison but to empirically showcase that our proposed method enhances OOD detection performance. Conversely, methods like ViM and NECO are methodologically and conceptually more similar to our approach and, therefore, require a more thorough comparison. We will elaborate on this in the revised manuscript to provide readers with a clearer understanding of the comparative framework and the uniqueness of our contributions. Moreover, we present the results of MCM on the fine-tuned model (i.e. MCM-tuned) in the table below for comparison.

Results for CIFAR-100 in-distribution dataset

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
MCM-tuned75.3391.3891.5560.9675.6091.0364.0795.4055.1498.9363.7197.6770.9089.23
SFM99.950.0298.318.6288.5653.9797.5412.9199.930.0699.950.0297.3712.60

Results for CIFAR-100-LT in-distribution dataset

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
MCM-tuned75.3391.3891.5560.9675.6091.0364.0795.4055.1498.9363.7197.6770.9089.23
SFM99.920.0297.5015.6585.2060.4195.0526.0599.920.0499.930.0196.2517.03

Results for ImageNet-1k-LT in-distribution dataset

MethodTexture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
MCM-tuned85.6460.1189.8244.3292.9236.2594.2632.0179.2676.1088.3849.76
SFM90.9543.1092.4334.8092.6239.0794.5928.5483.6268.9090.8442.88
评论

[W5] Whether effectiveness relies on the pre-trained models and test with smaller pre-trained models.

[A5] Thank you so much for raising this important point regarding whether the effectiveness of our method relies on the pre-trained models and suggesting to test with smaller pre-trained models. Firstly, we wish to convey that our Out-of-Distribution (OOD) score's effectiveness is not contingent upon pre-trained models. We employ pre-trained models primarily due to their practical benefits. It's significantly simpler and requires fewer resources to fine-tune existing pre-trained models rather than developing a new model from scratch. Moreover, many contemporary state-of-the-art methods also rely on fine-tuning pre-trained models. This prevalent approach not only streamlines the experimental process but also facilitates more straightforward and fair comparisons across different approaches, which is essential for an objective assessment of our method's performance and contributions. Additionally, in light of your insightful suggestion, we have indeed verified the validity of our OOD score with smaller pre-trained models. As depicted in the table below, we have included models like vit_tiny_patch16_224 and vit_small_patch16_224, shown in the upper and lower sections of each table. The outcomes from these smaller models provide further confirmation that our OOD score remains robust and effective across various model scales, thereby enhancing the generalizability and reliability of our proposed approach.

Results for CIFAR-100 in-distribution dataset

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
MSP92.0935.3483.2861.2883.3063.7379.8969.7372.8684.0782.1365.5582.2663.28
MLS98.626.1392.0935.8587.3954.9287.7152.7988.9257.5794.2730.3291.5039.60
Energy99.034.2692.7832.4187.2855.9688.1851.2790.3251.0495.2524.9492.1436.65
Mahalanobis99.900.3596.2815.7887.7856.6792.4833.8198.209.1098.776.2795.5720.33
residual99.710.8586.2452.7076.8672.7290.8942.5597.4614.0297.2513.6691.4032.75
Vim99.193.6292.9931.1887.4954.9088.7048.9691.1447.6395.6722.6392.5334.82
NECO99.173.8392.3434.2487.8553.4789.4746.4592.3843.0596.0621.1192.8833.69
KL-matching95.4118.4087.5845.7186.3253.7582.4760.4575.2879.2485.6552.8285.4551.73
SFM (ours)100.00.0296.8514.1386.4860.5697.0015.3499.980.0199.960.1096.7115.03
------------------------------------------------------------
MSP95.9819.1792.2938.1890.8239.0185.9552.3682.8468.9289.3147.8789.5344.25
MLS99.283.1696.3518.1695.2224.9092.1832.7296.2125.4097.7113.4496.1619.63
Energy99.482.2996.5416.5595.4223.2492.5929.9997.1218.5798.2510.0996.5716.79
Mahalanobis99.910.5999.054.7294.6528.6597.5311.3699.601.7899.542.5298.388.27
residual99.960.1198.607.0688.6652.2798.099.6899.650.7599.671.1497.4411.83
Vim99.561.9996.8814.6395.4623.0693.1727.6897.5216.2898.478.8296.8415.41
NECO99.502.1696.7615.9195.3324.3393.4926.6297.2617.0498.299.7596.7715.97
KL-matching97.669.2494.7522.2193.1828.0288.0440.2685.2755.4791.7333.4191.7731.43
SFM (ours)100.00.0099.363.1694.0931.3599.472.6999.990.0199.990.0198.826.20
评论

Results for ImageNet-1k-LT in-distribution dataset

MethodTexture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
MSP78.0173.3575.5078.4575.3079.0787.2154.5267.9587.7076.8074.62
MLS84.4465.7378.5075.8479.4475.4491.8346.4976.8384.5582.2169.61
Energy85.8560.0478.7674.9080.0274.1192.7242.1078.7381.5583.2266.54
Mahalanobis89.6141.8679.2767.2682.4463.8897.6211.8380.0977.3085.8152.43
residual84.8656.2168.2186.0369.9684.5588.6349.7973.9477.9577.1270.91
Vim86.4957.0678.9774.3580.2773.0293.2538.6679.2280.8583.6464.79
NECO86.8456.4479.0373.4780.6172.2694.7830.0179.5479.7584.1652.39
KL-matching81.9767.2277.8074.2778.0673.7291.5941.1872.7884.3580.4468.15
SFM (ours)92.2129.8478.4068.8781.2566.8697.7211.3082.4369.9086.4049.35
----------------------------------------------------
MSP82.6060.1181.4166.5981.9763.8494.3125.6977.7473.6083.6157.97
MLS87.9450.5184.9760.7386.4656.4496.6317.0484.6465.4088.1350.02
Energy88.9646.0385.4558.1087.2053.2797.0914.0685.9160.8588.9246.46
Mahalanobis91.1336.0686.3054.8789.5446.8199.034.4987.7455.4590.7539.54
residual88.6645.1179.4370.3084.2960.8996.1820.3082.0765.8586.1252.49
Vim89.3844.0885.7256.9787.5752.0697.3912.3886.2959.2589.2744.95
NECO89.7243.4085.8656.6388.1851.5098.079.3687.0058.4089.7743.86
KL-matching86.0150.6783.6360.6884.8056.7196.6814.6281.9065.3586.6049.61
SFM (ours)93.3525.9486.1855.2189.5246.7499.133.9989.1350.2091.4636.42
评论

Results for CIFAR-100-LT in-distribution dataset

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
MSP90.0843.3781.7067.6579.4871.6275.9675.2471.3684.2679.4171.4079.6668.92
MLS99.123.6093.3333.9379.8774.8685.8156.6893.8233.0396.5818.9791.4236.85
Energy99.382.1393.8331.2578.3378.8686.2456.1095.3224.8997.5013.2691.7734.42
Mahalanobis99.850.5397.2912.8085.0863.7891.2635.4598.448.1098.676.6795.1021.22
residual99.362.7085.0963.9963.9286.0886.9256.5095.5523.3895.8423.6888.6842.72
Vim99.481.8694.0230.0678.5978.2986.7553.9595.6722.9697.7012.1792.0333.21
NECO99.432.1693.7131.7880.4273.0987.2650.5295.1924.2297.4214.0192.2432.63
KL-matching94.5023.4886.5455.2182.7462.8179.1666.6674.5680.3083.5360.5983.5158.18
SFM (ours)99.990.0497.7710.8183.6766.9996.3116.6799.970.0199.950.1496.2815.78
------------------------------------------------------------
MSP96.3916.7292.7237.3987.5849.6082.3957.1080.5468.0987.5249.6487.8546.42
MLS99.691.4495.9721.8491.9641.6792.6229.4597.6614.6198.776.7496.1119.29
Energy99.801.1395.2927.2891.5545.2093.3725.7598.588.8699.283.7896.3118.67
Mahalanobis99.910.5399.432.3593.0235.9897.1512.9399.592.3599.651.6798.129.30
residual99.930.2596.2224.4983.2864.1395.9621.2099.263.7899.372.7895.6719.44
Vim99.840.9695.7425.1291.6944.6893.7824.6098.777.6599.383.2996.5317.72
NECO99.771.1396.3020.9291.6441.4992.8627.0397.6714.0198.925.9096.1918.41
KL-matching98.127.7395.6320.2790.1637.6285.2546.4083.9456.0090.7736.0690.6534.01
SFM (ours)100.00.0099.681.0592.2439.1199.482.44100.00.00100.00.0098.577.10
评论

Dear Reviewer cJWd,

Thank you for your follow-up questions! We address your concerns below.

[Q1] Performance comparision with ViM based on pre-trained checkpoints as the backbone.

[A1] Thanks for the suggestion! We compare the performance of our method with ViM in the table below. The results are based on Imagenet-1K-LT as the ID dataset. Given that running experiments on the ImageNet-1k dataset is time-consuming, we'll update the results as promptly as possible once the testing phase finishes.

From the results, we can see that our approach still outperforms ViM by over 3% w.r.t. FPR95 averaged over 5 OOD datasets, also showing its effectiveness.

Table 1: Performance comparison between ViM and our approach using ImageNet-1k-LT in-distribution dataset. In this experiment, we utilize the pre-trained model as the backbone without fine-tuning.

MethodTexture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
ViM90.7834.8887.5652.8089.0348.0998.615.6289.3647.5091.0737.78
Ours97.1510.2081.5167.2085.4956.4998.695.1393.1532.1591.2034.23

[Q2] Clearification on the results in Table 7 of the paper

[A2] Sorry for the confusion! Table 7 in the paper reports results for both pre-trained models (denoted by zero-shot) and fine-tuned models (denoted by long-tailed or balanced). We will revise the description to improve the clearity.

评论

Thank you for the prompt responses. Since all my concerns have been thoroughly addressed, I will be increasing my score.

评论

Thank you for all your suggestions which help a lot to improve our work!

评论

As mentioned in the previous response, we present the comparison results for ViM and SFM (ours) on the ImageNet-1k dataset when utilizing the pre-trained checkpoints as the backbone. From the results, we draw a similar conclusion that our proposed SFM achieves superior performance than ViM, showing its robustness to different backbones.

Results for ViM and SFM on ImageNet-1k in-distribution dataset based on pre-trained model without fine-tuning.

MethodTexture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
ViM90.5735.1686.1356.5487.6753.8297.909.1689.9044.9590.4339.93
SFM (ours)97.199.8882.7265.1486.4154.6498.804.7593.4031.4591.7033.17
审稿意见
5

This paper works on out-of-distribution detection (OOD) questions and proposes to use the rich information contained in the shallow features to improve the detection of the OOD samples. Experiments on both class-balanced and long-tailed datasets show the effectiveness of using such shallow features and the proposed method.

优点

  1. The writing is good and the idea is easy to follow.
  2. The observation of existing methods neglecting the shallow features makes sense to me and Fig.1 well illustrates the comparison of using the fused features from all layers than only using the last layer.
  3. The achieved results are promising showing the proposed method outperforms other competitors.

缺点

  1. The contribution of using all features for OOD looks over-claimed given the fact that there are already two related works (Mahalanobis and TRUSTED) explored that for OOD. Emphasis on the novel multiple-feature fuse method might be more suitable.
  2. The novelty part is not that new to me. Though the adaptive fuse for multiple features is not explored for OOD, what I understand of the key idea is to improve the discriminates of different image features, from this perspective, the idea of using multilayer features simply or adaptively is quite explored as in [ref1-4] for example.
  3. The proposed fusion method is not comprehensively studied. For example, what about the comparison results with some simple fuse method e.g., average or fixed ratios, and other proposed fusion methods e.g. [ref1-4]?
  4. Open questions about the task motivation: is it definitely better to detect all out-of-distribution data for the model training, or will the OOD data help with the model's generalization ability for the cross-domain testing which we can't ensure is always in the same domain with the training.
  5. Minor: the bold font is recommended to apply for tables to make it easier to read the results.

[ref1] Attentional Feature Fusion [ref2] Enhanced Blind Face Restoration with Multi-Exemplar Images and Adaptive Spatial Feature Fusion [ref3] Shallow Feature Matters for Weakly Supervised Object Localization [ref4] Multilayer Feature Fusion Network for Scene Classification in Remote Sensing

问题

please see weakness

评论

Dear Reviewer 5FtG,

We are truly and deeply grateful for the substantial amount of time and painstaking effort you have committed to meticulously reviewing our paper and furnishing such highly detailed and exceptionally valuable feedback.

[W1] Emphasis on the novel multiple-feature fuse method might be more suitable.

[A1] We truly value your valuable feedback and suggestions regarding the focus of our paper. We will modify our statement in the contribution section to "proposed a new adaptive feature fusion strategy". We think better showcases the unique characteristics and advantages of our research and appreciate your help in improving and strengthening our manuscript.

[W2,W3] Innovation declaration and the comparison results with some simple fuse method.

[A2,A3] We appreciate your suggestions greatly. After carefully reviewing [ref1-4], we agree that these articles are pioneering in the use of multi-layer feature fusion. Consequently, contingent upon the acceptance of our paper, we shall ensure an accurate citation and comprehensive analysis of these works within the related work section. Regarding the novelty of our paper, we wish to clarify that we have proposed a novel adaptive feature fusion algorithm applicable to OOD detection. Unlike [ref1-3], which modify network architecture by utilizing residuals or attention mechanisms to combine multi-layer features, [ref4] uses different enhancements of the same sample. Our approach of adaptive feature fusion distinctly differs from these methodologies. To thoroughly investigate our algorithm's application in OOD detection, we have also examined other weight configurations, such as uniform or fixed ratio weights, and conducted ablation experiments detailed in the following table. Due to constraints on rebuttal time, we assessed CIFAR-100, CIFAR-100-LT, and ImageNet-1k-LT utilizing the pre-trained ImageNet21k model alongside CLIP. The findings demonstrate that the SFM approach consistently achieves either the top (bold) or the second-best (italic) position across all contexts. Remarkably, when ranked second, our method's performance closely rivals that of the first-place finisher. Specifically, our approach attains 96.17 and 17.82, whereas W_0.5 achieves 95.82 and 19.04, w.r.t. AUROC and FPR95 respectively, which shows the effectiveness of our SFM approach.

Table 1: Results on CIFAR-100 in-distribution dataset based on ImageNet-21k pre-trained ViT and CLIP, respectively.

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
In21k W_0.5100.00.0099.432.3696.7618.2699.621.81100.00.0099.990.0199.303.74
In21k W_0.7599.990.0499.273.3396.9916.8298.835.2299.890.0699.890.2699.144.29
In21k W_1.099.970.1299.163.9297.0916.4997.998.9699.611.0799.671.3398.925.32
In21k SFM100.00.0099.501.9196.4719.5299.801.11100.00.00100.00.0099.293.76
------------------------------------------------------------
CLIP W_0.599.920.0798.0010.2789.1551.0996.4917.3699.880.2199.910.1297.2213.19
CLIP W_0.7599.610.8597.8412.3489.3650.6694.3826.5699.243.8599.551.6896.6615.99
CLIP W_1.099.231.6896.8923.2789.0152.2693.7532.2898.816.4499.293.1396.1619.84
CLIP SFM99.950.0298.318.6288.5653.9797.5412.9199.930.0699.950.0297.3712.60
评论

Table 2: Results on CIFAR-100-LT in-distribution dataset based on ImageNet-21k pre-trained ViT and CLIP, respectively.

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
In21k W_0.5100.00.0099.680.6894.5627.9099.581.6899.990.0099.980.0198.975.05
In21k W_0.7599.990.0599.491.6694.9326.4298.675.5799.810.3999.830.5398.795.77
In21k W_1.099.960.2099.332.5195.0925.9897.639.2699.482.2699.571.7198.516.99
In21k SFM100.00.0099.750.4394.2229.8699.751.1299.990.0099.990.0198.955.24
------------------------------------------------------------
CLIP W_0.599.880.0497.3716.9085.5258.9794.3628.6599.870.1199.890.0596.1517.45
CLIP W_0.7599.520.5597.1519.6685.7358.9991.5839.2299.362.6399.521.2095.4820.38
CLIP W_1.099.111.0395.9229.8784.7660.5890.9743.8399.084.0799.281.9994.8523.56
CLIP SFM99.920.0297.5015.6585.2060.4195.0526.0599.920.0499.930.0196.2517.03
评论

[W4] Open questions about the task motivation.

[A4] Regarding the initial point about whether detecting every out-of-distribution (OOD) sample during model training is always advantageous, the answer isn't straightforward. Many semi-supervised or label noise learning techniques recommend using the most accurate data possible during training. However, there are instances where including a small number of errors or OOD samples can be beneficial to the learning process. This inclusion might mimic the effects seen in self-supervised or contrastive learning approaches. Hence, we assert that identifying all OOD samples may not always be the best strategy.

For the second part of your inquiry, this paper primarily focuses on semantic OOD. It proposes that in-distribution (ID) and OOD classes are distinct, and training is conducted solely with ID data, yet OOD detection is necessary during testing. As pertains to your question, a covariate shift is expected in OOD detection. In such cases, cross-domain OOD becomes relevant during training, particularly when cross-domain test samples are involved. This is due to its potential to improve the model's ability to generalize and withstand varied data distributions.

We hope this provides a clearer and more satisfactory explanation. We appreciate your insightful feedback once more.

[W5] Minor: the bold font

[A4] We're sincerely thankful for your careful review and the suggestion about bold font. We'll use it appropriately in the table to enhance result readability. Appreciate your help in improving our work.

评论

Table 3: Results on ImageNet-1k-LT in-distribution dataset based on ImageNet-21k pre-trained ViT and CLIP, respectively.

MethodTexture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
In21k W_0.595.0217.9689.7644.7992.0836.6099.362.6392.6834.3093.7827.26
In21k W_0.7593.7823.3989.6245.5491.8837.2299.322.7492.0636.7593.3329.13
In21k W_1.092.9926.9589.4846.3491.7138.3599.282.8491.6638.8593.0230.67
In21k SFM96.9211.7989.8245.3692.1836.1699.332.5193.4631.1094.3425.38
----------------------------------------------------
CLIP W_0.588.5252.2389.8745.4792.2140.5894.5428.4582.5671.0089.5447.55
CLIP W_0.7587.6855.3989.6946.3792.0841.2394.4129.3082.2372.4089.2248.94
CLIP W_1.083.8167.6484.4466.8585.5069.5887.4972.5778.8280.2084.0171.37
CLIP SFM90.9543.1092.4334.8092.6239.0794.5928.5483.6268.9090.8442.88
评论

Dear authors,

thanks for the responses, I have read them. However, my first and second concerns still hold, and I will keep my score unchanged.

评论

We sincerely apologize for any confusion caused by the original writing. We acknowledge that several existing works explore multi-layer or all-layer feature integration for OOD detection, and we do not claim to be the first to address this. Instead, as emphasized in our revised manuscript, our primary contribution lies in the development of an adaptive weighting module that dynamically adjusts based on the data. We believe this approach provides a novel perspective in addressing the challenges of OOD detection.

Additionally, as the reviewer has correctly pointed out, the adaptive fusion of multiple features remains underexplored in the OOD detection domain. After carefully reviewing the papers referenced by the reviewer, we found that our key idea is substantially different from those works. In our understanding, many existing methods focus on improving the discrimination of image features, while our work proposes a distinct mechanism for adaptive feature fusion, specifically tailored to OOD detection.

Therefore, we respectfully disagree with the assertion that the originality of our method is significantly diminished. We sincerely hope the reviewer can reconsider the evaluation of our work. Thank you.

审稿意见
5

This paper presents a weighted feature fusion method from the top to bottom layers of a network for OOD detection. It argues that, instead of relying solely on the features of the penultimate layer in a deep network for OOD detection, incorporating features from shallow layers provides diversity in feature spaces, aiding in the separation of OOD examples from ID examples. The method uses the total variance of features across different neural network layers to estimate feature weights for OOD detection. Experiments are conducted on CIFAR-100 and Imagenet, two popular benchmarks for OOD detection, including a setup with a long-tail distribution in in-domain training examples. The experimental results show that the proposed method outperforms the compared method.

优点

The paper is well-written and easy to follow.

Extensive experiments are presented to support the claims.

There are sufficient ablation studies.

缺点

I find the novelty of the method modest. There is a remark section in the paper highlighting that the idea of estimating layerwise importance is the contribution compared to the inspiring methods, Mahalanobis and TRUSTED. While I can see a performance comparison with Mahalanobis in the paper, I could not find one for TRUSTED. Moreover, based on Figures 4 and 5 from the ablation studies, it hardly supports any contribution of layerwise feature weighting in OOD detection. The last layer has significant weight, while the other layers have more or less similar weights. A baseline with uniform weighting across all feature layers is also important.

Although not directly related, a paper on the variance of gradients for estimating the difficulty of images for OOD detection [ref1] is relevant and would support the argument.

[ref1] Agarwal, C., D'souza, D., & Hooker, S. (2022). Estimating example difficulty using variance of gradients. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 10368-10378).

问题

I am not fully convinced that there is a significant role of layerwise weightage in OOD detection. Giving more weights for the penultimate layer and the same weight to other layers can still perform well. A clarification with experiments may change my rating.

评论

Dear Reviewer GG5e,

We sincerely appreciate the time and effort you have dedicated to reviewing our paper. We address your concerns below.

[W1] Experimental results for TRUSTED.

[A1] Thank you for your reminder. We previously overlooked including the comparison results with the Trusted method, but these have now been added to the revised manuscript in Table 1, 2, 3, 4. For your convenience, we have summarized the comparison between the Trusted method and our approach on CIFAR-100 dataset in tables below. As shown in the results, our method consistently outperforms Trusted.

Table 1: Results on CIFAR-100 in-distribution dataset for Trusted and SFM. The top two rows of the table present results based on the ImageNet-21k pre-trained ViT, while the bottom two rows show results based on CLIP.

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
Trusted100.00.0098.785.7793.3533.5198.099.76100.00.01100.00.0098.378.17
SFM (ours)100.00.0099.501.9196.4719.5299.801.11100.00.00100.00.0099.293.76
------------------------------------------------------------
Trusted99.980.0497.2117.8086.3261.4597.1315.6899.950.0399.960.0896.7615.85
SFM (ours)99.950.0298.318.6288.5653.9797.5412.9199.930.0699.950.0297.3712.60


Table 2: Results on CIFAR-100-LT in-distribution dataset for Trusted and SFM. The top two rows of the table present results based on the ImageNet-21k pre-trained ViT, while the bottom two rows show results based on CLIP.

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
Trusted100.00.0099.123.6087.3452.8497.6710.3799.970.0099.980.0097.3511.13
SFM (ours)100.00.0099.750.4394.2229.8699.751.1299.990.0099.990.0198.955.24
------------------------------------------------------------
Trusted99.970.1193.5743.8080.7670.3695.4625.5899.950.1099.950.0894.9423.34
SFM (ours)99.920.0297.5015.6585.2060.4195.0526.0599.920.0499.930.0196.2517.03
评论

[W2,Q1] Contribution of layerwise feature weighting in OOD detection.

[A2] Thank you very much for your valuable suggestion! We agree with the reviewer that the penultimate layer plays a crucial role in OOD detection. To further investigate this, we conducted comparative experiments by assigning different weights to the penultimate layer. For example, in the case of W_0.5, the penultimate layer was given a weight of 0.5, while the remaining layers equally shared the remaining 0.5 weight.

The results are summarized in the tables below. From these results, it is evident that our method achieves superior performance on most datasets compared to assigning fixed weights across all layers. Notably, the improvement is particularly significant on the ImageNet dataset.

Thank you again for your insightful comment, which helped us further validate the effectiveness of our approach.

Table 1: Results on CIFAR-100 in-distribution dataset based on ImageNet-21k pre-trained ViT and CLIP, respectively.

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
In21k W_0.5100.00.0099.432.3696.7618.2699.621.81100.00.0099.990.0199.303.74
In21k W_0.7599.990.0499.273.3396.9916.8298.835.2299.890.0699.890.2699.144.29
In21k W_1.099.970.1299.163.9297.0916.4997.998.9699.611.0799.671.3398.925.32
In21k SFM100.00.0099.501.9196.4719.5299.801.11100.00.00100.00.0099.293.76
------------------------------------------------------------
CLIP W_0.599.920.0798.0010.2789.1551.0996.4917.3699.880.2199.910.1297.2213.19
CLIP W_0.7599.610.8597.8412.3489.3650.6694.3826.5699.243.8599.551.6896.6615.99
CLIP W_1.099.231.6896.8923.2789.0152.2693.7532.2898.816.4499.293.1396.1619.84
CLIP SFM99.950.0298.318.6288.5653.9797.5412.9199.930.0699.950.0297.3712.60
评论

Table 2: Results on CIFAR-100-LT in-distribution dataset based on ImageNet-21k pre-trained ViT and CLIP, respectively.

MethodTexture (AUROC)Texture (FPR95)SVHN (AUROC)SVHN (FPR95)CIFAR10 (AUROC)CIFAR10 (FPR95)Tiny ImageNet (AUROC)Tiny ImageNet (FPR95)LSUN (AUROC)LSUN (FPR95)Places365 (AUROC)Places365 (FPR95)Average (AUROC)Average (FPR95)
In21k W_0.5100.00.0099.680.6894.5627.9099.581.6899.990.0099.980.0198.975.05
In21k W_0.7599.990.0599.491.6694.9326.4298.675.5799.810.3999.830.5398.795.77
In21k W_1.099.960.2099.332.5195.0925.9897.639.2699.482.2699.571.7198.516.99
In21k SFM100.00.0099.750.4394.2229.8699.751.1299.990.0099.990.0198.955.24
------------------------------------------------------------
CLIP W_0.599.880.0497.3716.9085.5258.9794.3628.6599.870.1199.890.0596.1517.45
CLIP W_0.7599.520.5597.1519.6685.7358.9991.5839.2299.362.6399.521.2095.4820.38
CLIP W_1.099.111.0395.9229.8784.7660.5890.9743.8399.084.0799.281.9994.8523.56
CLIP SFM99.920.0297.5015.6585.2060.4195.0526.0599.920.0499.930.0196.2517.03


Table 3: Results on ImageNet-1k-LT in-distribution dataset based on ImageNet-21k pre-trained ViT and CLIP, respectively.

MethodTexture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
In21k W_0.595.0217.9689.7644.7992.0836.6099.362.6392.6834.3093.7827.26
In21k W_0.7593.7823.3989.6245.5491.8837.2299.322.7492.0636.7593.3329.13
In21k W_1.092.9926.9589.4846.3491.7138.3599.282.8491.6638.8593.0230.67
In21k SFM96.9211.7989.8245.3692.1836.1699.332.5193.4631.1094.3425.38
----------------------------------------------------
CLIP W_0.588.5252.2389.8745.4792.2140.5894.5428.4582.5671.0089.5447.55
CLIP W_0.7587.6855.3989.6946.3792.0841.2394.4129.3082.2372.4089.2248.94
CLIP W_1.083.8167.6484.4466.8585.5069.5887.4972.5778.8280.2084.0171.37
CLIP SFM90.9543.1092.4334.8092.6239.0794.5928.5483.6268.9090.8442.88
评论

[W3] Relevant [ref1] to support the argument.

[A3] Upon thorough examination of [ref1], we have found it to be highly pertinent to the task of OOD detection. We are truly grateful for this valuable reference. We intend to discuss it comprehensively and accurately cite it within the related work section. This will fortify the theoretical basis and context of our research, while also augmenting the thoroughness and persuasiveness of our arguments concerning OOD detection. We believe that by incorporating the insights from [ref1], our work will make a more significant contribution to the field of OOD detection and related research areas.

评论

Results for ImageNet-1k in-distribution dataset on Trusted and SFM

MethodTexture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
Trusted43.5686.4546.8296.9550.9594.7549.3691.4839.1595.4545.9793.02
SFM (ours)96.6511.7089.6446.0092.0437.7899.402.2693.7629.8094.3025.51
----------------------------------------------------
Trusted95.8719.8074.5978.0676.7176.4284.6172.7784.1262.4083.1861.89
SFM (ours)89.1648.4491.8836.4693.2434.8795.4723.4982.2770.3590.4142.72

Results for ImageNet-1k-LT in-distribution dataset on Trusted and SFM

MethodTexture (AUROC)Texture (FPR95)Places (AUROC)Places (FPR95)SUN (AUROC)SUN (FPR95)iNaturalist (AUROC)iNaturalist (FPR95)ImageNet-O (AUROC)ImageNet-O (FPR95)Average (AUROC)Average (FPR95)
Trusted91.9832.3682.1166.3185.7258.3498.099.2990.9140.1589.7641.29
SFM (ours)96.9211.7989.8245.3692.1836.1699.332.5193.4631.1094.3425.38
----------------------------------------------------
Trusted71.9670.4644.5197.8949.7897.7749.4498.5948.7989.0552.9090.75
SFM (ours)90.9543.1092.4334.8092.6239.0794.5928.5483.6268.9090.8442.88
评论

Dear Reviewer GG5e,

We sincerely thank you for taking the time to review our paper and providing us with your valuable feedback. Your insights are greatly appreciated and have been invaluable in improving our work.

May we kindly ask if our responses have sufficiently addressed your concerns? If there are any remaining questions or issues, we are fully committed to addressing them.

Once again, we deeply appreciate your time and effort in reviewing our work.

Best regards,

The Authors

评论

Dear PC memebers,

Thank you for your valuable comments during the review period, which raised many interesting and insightful questions. Now the discussion period is coming to a close, please take a moment to review the authors’ responses if you haven’t done so already. Even if you decide not to update your evaluation, kindly confirm that you have reviewed the responses and that they do not change your assessment.

Timeline: As a reminder, the review timeline is as follows:

November 26: Last day for reviewers to ask questions to authors. November 27: Last day for authors to respond to reviewers. November 28 - December 10: Reviewer and area chair discussion phase. December 20: Meta reviews and initial decisions are due.

Thank you for your time and effort!

Best regards, AC

评论

We sincerely thank each reviewer for their valuable comments. We have addressed them in our responses and updated the manuscript accordingly. If the reviewers have any further questions, we are always ready to provide additional clarifications.

The Authors

AC 元评审

This paper focuses on layer weights and proposes an adaptive fusion method for addressing OOD scenarios. While the idea is straightforward and easy to follow, as the reviewers noted, it lacks novelty. Although there isn’t an identical method in the literature, similar ideas have been explored previously. The proposed approach, while showing some improvement in numerical experiments through careful design, does not offer a fundamental advancement in OOD.

The reviewers' opinions are split, with two negative and one positive score. The positive reviewer did not appreaciate fundamental or technical contributions neither and did not defend his/her score during the internal discussion. Considering the overall feedback, I recommend rejection.

审稿人讨论附加意见

The three reviewers generally agreed that the method lacks novelty, noting that similar ideas (of course, not the same) have been explored previously. In the rebuttal, the authors were unable to overturn this impression but provided additional experiments to strengthen the numerical comparison with existing methods. As a result, Reviewer 5FtG increased their score to 6, while the other two reviewers kept their scores unchanged.

Unfortunately, Reviewer GG5e did not respond to reminders and did not participate in the discussion, which is regrettable. However, based on the feedback from the other two reviewers and Reviewer GG5e’s initial comments, it is clear that the paper does not present an exciting contribution, and the numerical improvements is not fully convincing. Therefore, I recommend rejection.

最终决定

Reject