PaperHub
7.0
/10
Poster3 位审稿人
最低4最高5标准差0.5
4
4
5
2.7
置信度
创新性2.7
质量3.0
清晰度2.3
重要性2.7
NeurIPS 2025

Redundancy-Aware Test-Time Graph Out-of-Distribution Detection

OpenReviewPDF
提交: 2025-05-07更新: 2025-10-29

摘要

关键词
out-of-distribution detectionstructural entropy

评审与讨论

审稿意见
4

This paper proposes to construct a coding tree based on a well-trained GNN to reduce the redundancy in the graph structure and use the construction loss as a criterion for OOD detection.

优缺点分析

Strength:

  • The proposed method seems novel. This paper proposes to detect OOD samples by constructing a coding tree that minimizes the structural entropy.

  • The paper provides sufficient experimental results to verify the effectiveness and sensitivity to hyperparameter settings of the proposed method.

Weakness:

  • Some parts of the paper are hard to follow. For example, Figure 2 contains "?", which makes it difficult to understand. Sec. 3 introduces the ReGIB and the objective for coding tree construction, but lacks a straightforward explanation on the choice of using the construction loss as the OOD detection score.

  • The details of how the OOD samples are detected are vague. While constructing the coding tree provides a loss that is used as an OOD detection score, the method for separating OOD samples from IID samples seems unclear. If the OOD samples are detected when the loss exceeds a certain threshold, the good performance of the proposed method may stem from a well-tuned threshold on the benchmark datasets.

  • The paper could benefit from more empirical details on the coding tree construction and a time consumption comparison between the proposed method and other baseline methods.

问题

  • Since the proposed method uses the optimized loss as an OOD detection score, how does the method identify a sample as an OOD sample? Is there a manually set threshold that separates IID and OOD samples?

  • While the paper provides time complexity analysis, could the comparison between the proposed method and other baseline methods be provided? Constructing and optimizing a coding tree can be a time-consuming process. It would be beneficial to have more details about optimizing the coding tree on a sample provided.

I look forward to the authors' response and will adjust my rating based on the rebuttal.

局限性

The authors have discussed the limitation that the proposed method mainly focuses on graph-level detection. However, I encourage the authors to further discuss the limitations of the proposed method.

最终评判理由

This paper proposes a new OOD detection method using a coding tree. The proposed method seems novel. My initial concerns has been addressed in the rebuttal, therefore I am increasing my rating.

格式问题

I do not notice any major formatting issues in this paper.

作者回复

We sincerely appreciate your valuable comments and will reply to all these concerns individually.

W1.1: Figure 2 contains "?", which makes it difficult to understand

We would like to clarify that Figure 2 in our paper does not contain "?" symbols. The issue you mentioned may result from incomplete image rendering in the browser. We recommend downloading the paper and viewing it locally using Adobe Reader to ensure correct visualization.


W1.2 & W2.1: Section 3 introduces ReGIB and the objective for coding tree construction, but lacks a straightforward explanation on the choice of using the construction loss as the OOD detection score

Thank you for the question. The construction of coding tree does not introduce any new loss function. Instead, we use structural entropy as the optimization metric for coding tree construction (see Eq. (12) and (14)). Moreover, we would like to clarify that we do not use structural entropy as the OOD score, and the OOD score is actually computed based on Eq. (18).


W2.2: If the OOD samples are detected when the loss exceeds a certain threshold, the good performance of the proposed method may stem from a well-tuned threshold on the benchmark datasets

Thank you for the comment. The strong performance of our method does not stem from a well-tuned threshold. We follow the same evaluation protocol as GOOD-D and GOODAT, where the AUC metric is computed using auc = sklearn.metrics.roc_auc_score(y, ood_score). Specifically, y denotes the binary ground-truth labels, and ood_score represents the confidence score for predicting a sample as OOD. This score is a continuous value that can be ranked across samples. AUC is computed by ranking all samples according to their ood_score, and then calculating the corresponding true positive rate (TPR) and false positive rate (FPR) over the full range of score distributions to construct the ROC curve (with FPR on the x-axis and TPR on the y-axis). The AUC is the area under this ROC curve, with a larger value indicating stronger discrimination ability. The effectiveness of these methods, including ours, is thus evaluated independently of any manually selected threshold.


W3 & Q2.2: The paper could benefit from more empirical details on the coding tree construction and a time consumption comparison between the proposed method and other baseline methods

Thank you for pointing this out. We provide detailed illustrations and explanations of the coding tree construction process in Appendix D and Figures 10 to 12 of the paper. Moreover, coding tree construction, as well as the proposed RedOUT method, is not time-consuming. In Appendix E7, Table 10 compares the time consumption of pretraining, coding tree construction, and test-time OOD detection.

We additionally provide a comparison of time consumption with other baselines in the table below. For deep learning methods under other settings, we select the best-performing methods, GOODD and GraphCL-MD, for reference.

IDBZRPTC-MRENZYMESIMDB-MFreeSolv
OODCOX2MUTAGPROTEINIMDB-BToxCast
Coding Tree Construction (s)0.140.040.150.090.38
GraphCL-MD (s)27.9421.7114.4310.6440.02
GOODD (s)323.36320.77137.533.59356.44
GOODAT (s)6.863.555.240.874.81
RedOUT (s)7.515.652.880.287.51

Note that the official implementation of AAGOD could not run successfully in our experimental environment. Several issues have also been reported by other users on its GitHub repository and remains unresolved.


Q1: Since the proposed method uses the optimized loss as an OOD detection score, how does the method identify a sample as an OOD sample? Is there a manually set threshold that separates IID and OOD samples?

Thank you for the insightful question. A global evaluation metric AUC is used to assess a model’s ability to distinguish between ID and OOD samples overall. In practical applications, we can estimate a decision threshold based on the ID distribution observed in the pretraining dataset. For example, a score threshold that corresponds to a 5% false positive rate (FPR) on ID samples can be chosen. This threshold can then be used at test time to decide whether a new input is OOD. Notably, the selection of threshold does not affect the evaluation of the model’s OOD detection performance, that is, the threshold selection does not affect the computation of AUC metric.


Q2.1: While the paper provides time complexity analysis, could the comparison between the proposed method and other baseline methods be provided?

Thank you for the valuable question. A detailed analysis and comparison of computational complexity between our proposed RedOUT and GOODAT is provided on page 31 in Appendix E7. In addition, we provide a comparative analysis of time complexity for our proposed method and several other representative baselines as follows:

  • RedOUT only optimizes a tree encoder with LL layers, each being a two-layer MLP with hidden dimension hh. The total time complexity is O(L(mh+nh2))O(L(mh + nh^2)) per graph.
  • GOOD-D updates two GIN encoders with LL layers and hidden size hh. The forward pass complexity is O(L(mh+nh2))O(L(mh + nh^2)) per graph. An additional O(nh)O(nh) cost arises from the group-space clustering, resulting in O(L(mh+nh2)+nh)O(L(mh + nh^2)+nh) overall.
  • GTrans performs optimization over node features and edge structure based on a fixed backbone. The per-graph time complexity is O(L(mh+nh2)+md+nd2+BlogB)O(L(mh + nh^2)+md + nd^2 + B'\log B'), where BB' is the number of candidate edge flips.
  • AAGOD introduces an auxiliary augmentation scorer alongside the main GNN model with LL layers and hidden dimension hh that scoring KK augmentations per graph. The total time complexity is O(L(mh+nh2)+K(nd+m))O(L(mh + nh^2)+K(nd+m)).
  • GOODAT introduces two masking modules (feature and edge maskers). The time complexity per graph is O(L(mh+nh2)+nd+m)O(L(mh + nh^2)+nd + m), as it applies learnable masks over node features and edge connections.

Overall, the time complexity of our proposed RedOUT is comparable to that of other baselines.


Limitation: The authors have discussed the limitation that the proposed method mainly focuses on graph-level detection. However, I encourage the authors to further discuss the limitations of the proposed method.

Thank you for the suggestion. The current coding tree optimization algorithm relies on the fixed height applied to the whole dataset, without considering the diversity among samples. One potential direction for improving our method is to incorporate an adaptive coding tree height to better extract essential structures in graphs, which we leave as our future work. We will expand the discussion of limitations in the revised version.

We once again thank you for your constructive feedback and valuable suggestions. We will carefully address all the raised points in the revised version.

评论

I appreciate the rebuttal provided by the authors, which has addressed most of my concerns. I will increase my rating.

评论

Dear Reviewer tiWp,

We would like to express our sincere gratitude to you for endorsing our work and providing constructive suggestions. Thanks again for the time and effort in reviewing our work!

审稿意见
4

The paper proposes RedOUT, a framework for test-time graph out-of-distribution (OOD) detection that addresses the challenge of structural redundancy in graph data. It introduces the Redundancy-aware Graph Information Bottleneck (ReGIB) to separate essential information from irrelevant redundancy by minimizing structural entropy, which helps in effectively distinguishing between in-distribution and out-of-distribution graph samples. The framework demonstrates superior performance over state-of-the-art baselines through extensive experiments on real-world datasets, highlighting its capability to capture essential structural information in graphs.

优缺点分析

Strength:

  1. By minimizing structural entropy, it effectively extracts essential structural information from graphs while eliminating redundancy, offering a new perspective for test-time graph OOD detection.
  2. Extensive experiments on multiple real-world graph datasets demonstrate that RedOUT outperforms state-of-the-art baselines, achieving an average improvement of 6.7% in unsupervised OOD detection tasks.
  3. The paper provides theoretical analysis and proofs for ReGIB, such as deriving upper and lower bounds for essential and redundant information optimization.
  4. The Paper is well written and easy to follow.

Weakness:

  1. It lacks clear introduction of the baseline to analyze the contribution of this work.
  2. The performance of RedOUT is sensitive to hyperparameters such as the coding tree height k and the trade-off parameter λ. While the paper provides a reasonable range for these parameters, it lacks detailed analysis on how to select optimal values for different datasets, which may affect the method's practical usability.
  3. The coding tree construction process involves structural entropy minimization, which may incur higher computational costs for large-scale graphs.

问题

  1. It lacks clear introduction of the baseline to analyze the contribution of this work.
  2. The performance of RedOUT is sensitive to hyperparameters such as the coding tree height k and the trade-off parameter λ. While the paper provides a reasonable range for these parameters, it lacks detailed analysis on how to select optimal values for different datasets, which may affect the method's practical usability.
  3. The coding tree construction process involves structural entropy minimization, which may incur higher computational costs for large-scale graphs.

局限性

  1. The coding tree construction process involves structural entropy minimization, which may incur higher computational costs for large-scale graphs.

最终评判理由

The paper proposes RedOUT, a framework for test-time graph out-of-distribution (OOD) detection that addresses the challenge of structural redundancy in graph data. It introduces the Redundancy-aware Graph Information Bottleneck (ReGIB) to separate essential information from irrelevant redundancy by minimizing structural entropy, which helps in effectively distinguishing between in-distribution and out-of-distribution graph samples. The framework demonstrates superior performance over state-of-the-art baselines through extensive experiments on real-world datasets, highlighting its capability to capture essential structural information in graphs. By minimizing structural entropy, it effectively extracts essential structural information from graphs while eliminating redundancy, offering a new perspective for test-time graph OOD detection. Extensive experiments on multiple real-world graph datasets demonstrate that RedOUT outperforms state-of-the-art baselines, achieving an average improvement of 6.7% in unsupervised OOD detection tasks. The paper provides theoretical analysis and proofs for ReGIB, such as deriving upper and lower bounds for essential and redundant information optimization. The Paper is well written and easy to follow.

格式问题

N/A

作者回复

We sincerely appreciate your thoughtful comments and address your comments and suggestions point by point below.

Q1: Further analysis of baselines and the contribution of this work

Thank you for the suggestion. In Appendix E2, we provide a detailed categorization and discussion of the 18 baseline methods considered in our comparison. These baselines are grouped into 4 categories: Non-deep Two-step Methods, Deep Two-step Methods, End-to-end Training Methods, and Test-time and Data-centric Methods. Additional discussion and comparison with related works are provided in Appendix B. We summarize the key differences and contributions below.

Non-deep two-step methods, deep two-step methods, and end-to-end training methods all rely on training data to modify the parameters of GNN models. Data-centric methods such as AAGOD do not alter pre-trained GNN parameters but still require access to ID training data for post-hoc data augmentation. In contrast, our method is designed for the test-time setting, aiming to improve OOD detection performance without modifying pre-trained models or requiring any access to training data. Although Gtrans and GOODAT are also test-time methods, they requires additional training of graph transformation functions or graph maskers. In our case, the extraction of essential structure is achieved via structure entropy minimization, which is a training-free optimization procedure. We would like to further emphasize that the core contribution of this work lies in the first to extend GIB framework to redundancy-aware disentanglement. Specifically, we propose ReGIB, which effectively removes redundant information and extracts essential structure by jointly optimizing theoretical upper and lower bounds.

Thanks again and we will include further analysis in the revised version.


Q2: How to select optimal values for different datasets

Thank you for the insightful question. For the hyperparameter kk, while the performance may exhibit slight variations, we observe that RedOUT consistently achieves strong results, as illustrated in Figure 4. Specifically, on 5 out of 6 datasets, RedOUT outperforms all baselines regardless of the choice of kk. Furthermore, compared with GOODAT (which also operates in test-time setting), our method achieves superior performance on all 6 datasets. These results demonstrate that ReGIB can effectively extract essential structural information and eliminate irrelevant redundancy, regardless of the depth of coding tree. For the hyperparameter λ\lambda, we empirically find that values in the range from 1e-4 to 1e-2 lead to consistently effective performance. This robustness reduces the need for extensive hyperparameter tuning, making our method well-suited for practical deployment in real-world scenarios.


Q3: Analysis of the computational overhead of coding tree construction on large-scale graphs

Thank you for the suggestion. We provide a detailed analysis of the scaling efficiency of our method in Appendix E7. As shown in Figure 13, the results indicate that from small-scale graphs to larger ones such as the OGB datasets (e.g., ogbg-code2), the runtime and memory consumption of coding tree construction increase nearly linearly with the number of nodes.

We once again thank you for your constructive feedback and valuable suggestions. We will carefully address all the raised points in the revised version.

评论

I appreciate the rebuttal provided by the authors, which has addressed some of my concerns. I believe a Borderline Accept score remains fair for this paper. I will therefore maintain my original score.

审稿意见
5

This paper proposes RedOUT, a redundancy-aware framework for test-time graph OOD detection. The method introduces a Redundancy-aware Graph Information Bottleneck , which decomposes graph information into essential components and redundant noise. By minimizing structural entropy on a coding tree constructed from test graphs, RedOUT extracts distinctive substructures critical for OOD detection while preserving semantic consistency. The framework operates as a post-hoc, unsupervised test-time algorithm without retraining, achieving substantial improvements on several graph benchmarks. Experimental results demonstrate sota performance on both OOD detection and anomaly detection tasks.

优缺点分析

Strengths:

  1. Novel formulation of redundancy-aware test-time OOD detection for graphs, filling a gap in current data-centric, post-hoc graph OOD detection literature.
  2. The introduction of ReGIB with theoretical bounds and entropy minimization provides a well-motivated, principled approach for isolating essential information.
  3. The ablation studies and parameter sensitivity analyses are thorough and help clarify the contribution of each component.
  4. The framework does not require model retraining, which makes it practical for real-world applications.

Weaknesses:

  1. The performance improvement of RedOUT on the IMDB datasets appears to be relatively limited. Could the authors provide an explanation for this phenomenon?
  2. The reliance on pseudo-labels during test-time introduces potential error accumulation that is not deeply analyzed
  3. The time complexity of the coding tree construction is relatively high (O(h|E|log|V|)), which may affect scalability on large graphs.

问题

  1. How robust is RedOUT to incorrect pseudo-labels at test time, especially on highly uncertain OOD samples? Could this reliance on pseudo-labels amplify noise or bias in the essential view extraction?
  2. Since RedOUT mainly improves on molecular graph datasets, what specific characteristics of social network graphs limit its performance? Could incorporating node/edge attributes (besides structure) help mitigate this?
  3. Given the strong reliance on structural entropy minimization, how sensitive is the framework to the choice of coding tree height k in unseen datasets or new application scenarios? Could an adaptive or data-driven selection of k improve stability?

局限性

N/A

最终评判理由

The authors have addressed most of my concerns, after reviewing the other reviewers' comments, I have decided to maintain my original score.

格式问题

no

作者回复

We sincerely appreciate your valuable comments and are happy to address your concerns as follows:

W1 & Q2: An explanation for the relatively limited performance improvement on the IMDB dataset. Could incorporating node/edge attributes (besides structure) help?

Thank you for the insightful suggestion. In Section 4.5, through a case study, we investigate that the suboptimal performance mainly stems from the highly similar structural semantic information and the shared data source, which makes it challenging even for existing methods to effectively distinguish OOD samples.

For graphs in social networks, which typically exhibit high connectivity and edge density, incorporating node and edge attributes implies huge potential to enhance the model’s representational capacity, beyond the essential structural information we currently extract. However, the current IMDB-B/M datasets do not incorporate node or edge attributes. This is a promising direction that inspires us to take a deep exploration in our future work. We will include further analysis in the revision.


W2 & Q1.1: The impact of incorrect pseudo-labels at test time on RedOUT, especially for highly uncertain OOD samples

Thank you for the valuable question. While the pseudo-labels predicted by pre-trained model may contain errors when encountering highly uncertain OOD samples, such errors do not compromise the effectiveness of our redundancy removal mechanism. Specifically, we leverage ReGIB to disentangle the conditional redundant information associated with pseudo-labels in Eq. (6), and optimize the upper bound in Eq. (8) to minimize I(f(G);f(G)Y~)I(f(G ^ *); f(G) \mid \tilde{Y}).

In the general case, we decompose the learned representation as f(G)R1+YGf(G) \rightarrow{R_1 + Y_G} and f(G)R2+YGf(G ^ *) \rightarrow{R_2 + Y_G}, where R1/2R_{1/2} denotes label-irrelevant redundant information, and YGY_G denotes label-related information. Minimizing I(f(G);f(G)Y~)I(f(G ^ *); f(G) \mid \tilde{Y}) is equivalent to minimizing I((R1+YG);(R2+YG)Y~)minI(R1;R2)I((R_1+Y_G);(R_2+Y_G) \mid \tilde{Y}) \rightarrow \min I(R_1; R_2).

If the pseudo-label Y~\tilde{Y} predicted by the pre-trained model is incorrect, then f(G)Y~Rˉ1+YˉGf(G) \mid \tilde{Y} \rightarrow \bar{R}_1 + \bar{Y}_G and f(G)Y~Rˉ_2+Yˉ_Gf(G ^ *) \mid \tilde{Y} \rightarrow \bar{R}\_2 + \bar{Y}\_G, where Rˉ_1/2 \bar{R}\_{1/2} denotes the residual irrelevant information after removing the wrongly associated label part, and Yˉ_G\bar{Y}\_G similarly denotes the remaining label-related information. In this case, minimizing I(f(G);f(G)Y~)I(f(G ^ *); f(G) \mid \tilde{Y}) becomes minI(Rˉ1+YˉG;Rˉ2+YˉG)minI(Rˉ1;Rˉ2)\min I(\bar{R}_1 + \bar{Y}_G ; \bar{R}_2 + \bar{Y}_G) \rightarrow \min I(\bar{R}_1 ; \bar{R}_2).

Hence, the optimization still targets the irrelevant redundancy and the direction of our objective remains unchanged. Therefore, the effectiveness of our method in removing irrelevant redundancy is not compromised.


W3: Scalability of coding tree construction on large graphs

Thanks for your suggestions. We provide a scalability efficiency analysis of our method in Appendix E7. As shown in Figure 13, the results illustrate that from smaller scales to the OGB datasets (e.g., ogbg-code2), the runtime and memory usage scale up nearly linearly with the number of nodes. Furthermore, as shown in Table 10, the time overhead introduced by coding tree construction is negligible compared to the cost of model pretraining and test-time detection.


Q1.2: Could the reliance on pseudo-labels amplify noise or bias in the essential view extraction?

Thanks for the question. The extraction of essential information does not depend on pseudo-labels and thus is not subject to such bias. It is derived independently by minimizing the structure entropy in Eq. (12). We will include the additional explanation in the revision.


Q3: How sensitive is the framework to the choice of coding tree height kk in unseen datasets or new application scenarios? Could an adaptive or data-driven selection of k improve stability?

Thanks for the question. The impact of the coding tree height kk on performance is analyzed in Figure 4. We observe that across 5 out of 6 datasets, our method outperforms the baselines regardless of the specific choice of kk. Notably, compared with GOODAT, which also operates in a test-time setting, our method achieves superior performance across all 6 datasets. This further confirms the effectiveness of our proposed ReGIB in extracting essential structures and removing redundancy, irrespective of the coding tree depth.

Additionally, in light of the variability of test-time samples, an adaptive or data-driven selection of kk shows great potential for improving stability. We leave this exploration for future work.

We once again thank you for your constructive feedback and valuable suggestions. We will carefully address all the raised points in the revised version.

最终决定

The paper proposes RedOUT, a redundancy-aware framework for test-time graph Out-of-Distribution (OOD) detection. The core contribution is the introduction of the Redundancy-aware Graph Information Bottleneck (ReGIB), which decomposes graph information into essential components and redundant noise by minimizing structural entropy on a coding tree constructed from test graphs. This approach enables the extraction of distinctive substructures important for OOD detection while preserving semantic consistency. The framework operates as a post-hoc, unsupervised test-time algorithm without retraining and demonstrates substantial improvements over state-of-the-art baselines on several graph benchmarks.

Strengths

The strengths include its novel formulation of redundancy-aware test-time OOD detection for graphs, the introduction of ReGIB with theoretical bounds and entropy minimization, thorough ablation studies and parameter sensitivity analyses, and the framework's practicality due to not requiring model retraining. Additionally, the paper offers a new perspective on test-time graph OOD detection by addressing structural redundancy in graph data.

Weaknesses

The weaknesses include limited performance improvement on certain datasets (e.g., IMDB), potential error accumulation from relying on pseudo-labels during test-time, and high time complexity of coding tree construction (O(h|E|log|V|)), which may affect scalability on large graphs. Furthermore, some reviewers noted poor clarity in parts of the paper, lack of straightforward explanation for using construction loss as an OOD detection score, and vagueness in details of OOD sample detection.

Summary

During the rebuttal period, the authors addressed several concerns raised by reviewers, including explanations for limited performance improvement on certain datasets, impact of incorrect pseudo-labels at test time, and scalability of coding tree construction on large graphs. The authors also provided additional analyses, such as comparisons with baseline methods and discussions on selecting optimal values for different datasets. Reviewers appreciated the rebuttal, with one reviewer increasing their rating after finding that the authors had addressed most of their concerns. While some reviewers maintained their original scores, they acknowledged the authors' efforts to address concerns and recognized the paper's technical solidity despite some limitations.

Overall, while not working on an exciting new topic, the paper presents a decent contribution. Based on the reviews and rebuttal, I recommend accepting this paper. The authors have addressed most of the reviewers' concerns, providing additional explanations and analyses to clarify the methodology and results. The paper's strengths, including its novelty, thorough experiments, and good performance on benchmark datasets, outweigh its weaknesses. While some limitations remain, such as sensitivity to hyperparameters and potential computational costs, the authors have acknowledged these and plan to discuss them further in a revised version.