PaperHub
6.6
/10
Poster4 位审稿人
最低3最高4标准差0.5
4
3
4
3
ICML 2025

GPEN: Global Position Encoding Network for Enhanced Subgraph Representation Learning

OpenReviewPDF
提交: 2025-01-24更新: 2025-07-24

摘要

关键词
Subgraph Representation Learning

评审与讨论

审稿意见
4

The paper introduces GPEN (Global Position Encoding Network), a novel algorithm for subgraph representation learning. The algorithm addresses the task of predicting labels for subgraphs within a large input graph, given a set of labeled subgraphs. GPEN innovates by moving beyond the limitations of existing methods that primarily focus on local neighborhood structures. It achieves this by incorporating global structural information through a novel tree-based global position encoding for each node. This encoding is combined with a boundary-aware convolution module and an optional tree perturbation technique. The authors support their approach with both empirical evaluations demonstrating effectiveness and theoretical analyses to justify the method.

给作者的问题

The current work primarily addresses subgraph representation learning on undirected graphs. Could the authors discuss the applicability of GPEN to directed graphs (e.g., follow relationships in social networks, citation links in academic networks)? Specifically, we are interested in how edge directionality would be handled in the tree construction phase and what modifications to the maximum spanning tree algorithm would be necessary to accommodate directed scenarios. The answer would help understand GPEN's potential for broader applications in directed graph domains.

论据与证据

The paper's main claims are well supported by convincing evidence. Experimental results show that GPEN outperforms existing methods across various datasets, while theoretical analysis validates the method's soundness from multiple angles. In particular, Figure 1 clearly demonstrates the limitations of relying solely on local structural information and how GPEN addresses this issue by capturing global structural information.

方法与评估标准

The proposed methods are innovative and well-justified. GPEN cleverly utilizes tree structures to encode global position information, while the boundary-aware convolution module effectively balances local and global information. The experimental evaluation employs widely recognized datasets and baseline methods in the field, with a well-designed and comprehensive evaluation protocol.

理论论述

I carefully checked the correctness of the theoretical claims, including the theorems and proofs related to bounded representation discrepancy, global position encoding distinctness, noise robustness, and the generalization bound. They are robust and clearly presented, providing a strong theoretical foundation for the GPEN algorithm. The inclusion of these formal analyses significantly strengthens the paper.

实验设计与分析

The experimental design is sound and effective. Extensive experiments on eight public datasets, with results averaged over 10 different random seeds, provide reliable statistical significance. The hyperparameter analysis and ablation studies effectively validate the contribution of each module.

补充材料

I reviewed the supplementary material, which consists of the code for the proposed method. Providing the code allows for reproducibility and further exploration of the method's implementation.

与现有文献的关系

The work is well-connected with existing research. The paper clearly identifies limitations in existing methods and proposes effective solutions. GPEN's main innovations - global position encoding and boundary-aware convolution - represent significant improvements and additions to prior work.

遗漏的重要参考文献

I did not find any essential references that were missing or overlooked in the paper's discussion of related work.

其他优缺点

Strengths:

The core idea of tree-based global position encoding is novel and effective. Robust theoretical analysis strengthens the paper. The paper is well-structured and easy to follow.

Weaknesses:

Visualizations of learned embeddings could further illustrate the effectiveness of GPEN.

其他意见或建议

  1. Although the paper includes hyperparameter analysis, a sentence or two summarizing the general robustness or sensitivity of the key hyperparameters within the main body (perhaps in the experimental section) would be helpful for readers. This would provide a concise summary without requiring a deep dive into the appendix.

  2. Adding visualizations of the final learned embeddings could highlight the effectiveness of the approach to a reader.

作者回复

We deeply appreciate your thoughtful review and positive assessment of our work. Below we will respond to the raised questions.

W1 and S2: Visualization of Learned Embeddings

We appreciate your suggestion about visualization. We agree that adding visualizations of the learned embeddings would enhance the interpretability of our approach. In the revised manuscript, we will add t-SNE visualizations of subgraph representations learned by GPEN compared with baseline methods on selected datasets. These visualizations will help readers intuitively understand how our method better separates subgraphs of different classes in the embedding space.

S1: Hyperparameter Analysis Summary

Thank you for this valuable suggestion. We will add a concise summary of our hyperparameter analysis in Section 5.2.5 of the main experimental section. The key findings include:

  • The balance factor b (controlling the trade-off between local and global structural information) shows optimal performance in the range of 0.6-0.8 across all datasets, indicating that moderately emphasizing local structural information while maintaining global context leads to better subgraph representations.

  • The tree-based data augmentation threshold c achieves optimal results with moderate values (4-6), suggesting that this range provides an effective balance between generating sufficient augmented samples and maintaining structural integrity.

  • Batch size exhibits relatively stable performance for values between 5 and 15, with gradual decline for larger values (>20), likely due to reduced gradient update frequency and less effective early stopping with larger batches.

  • GPEN demonstrates reasonable robustness to hyperparameter changes within these ranges, with different datasets exhibiting varying sensitivities based on their inherent structural properties.

Q1: Applicability to Directed Graphs

GPEN can be naturally extended to directed graphs. Below we elaborate on how each module can be adapted:

Our importance calculation already accommodates directed graphs, as the transition matrix MM inherently represents directional flow where MijM_{ij} is defined based on the out-degree of nodes. This formulation naturally captures the directional influence of nodes in the graph without requiring fundamental changes.

For tree construction, we would maintain the same edge weight assignment while selecting an appropriate spanning arborescence algorithm (directed version of spanning tree) such as Edmonds' algorithm to construct the tree structure from the weighted directed graph. The root node selection would remain based on the highest PageRank score, preserving our hierarchical encoding approach. Once the directed tree (arborescence) is constructed, the global position encoding process remains identical to the undirected case. We would still compute each node's position based on its path distance to the root node, enabling the same systematic way to capture relationships between distant nodes in directed scenarios. For boundary-aware convolution, the primary change would be in the adjacency matrix representation, which would need to reflect the directionality of edges. This modification preserves our boundary-aware convolution's ability to control information flow during the process.

These adaptations would enable GPEN to effectively capture hierarchical relationships in directed network scenarios such as citation networks, social influence networks, and information flow systems, while preserving the theoretical guarantees established in our paper.

审稿意见
3

This paper presents GPEN, a novel method for subgraph representation learning that addresses two key challenges: capturing structural relationships between distant nodes and preventing excessive aggregation of global structural information.

给作者的问题

Please answer W1-W3

论据与证据

yes, the submission is supported by clear and convincing evidence

方法与评估标准

yes, the proposed method makes sense for the graph learning challenges.

理论论述

yes, the theoretical proof of the paper seems correct.

实验设计与分析

yes, the experiment section is well-designed.

补充材料

yes, I reviewed additional experimental and theoretical analyses

与现有文献的关系

This paper addresses existing challenges in a more novel way

遗漏的重要参考文献

No

其他优缺点

Strengths:

S1. The paper provides rigorous theoretical analysis (Theorems 4.1–4.4) to validate its claims, such as bounded representation discrepancy and noise robustness, which adds intellectual depth to the innovation.

S2. The introduction provides a clear overview of the limitations inherent in existing methods, which highlights the key challenges capturing distant relationships, avoiding over-aggregation. Figure 1 enhances this by visually contrasting fraudulent and legitimate subgraphs, making the motivation intuitive.

Weaknesses:

W1. Although the tree-based encoding approach introduces an innovative perspective, its fundamental assumption may limit its applicability, such as a tree structure can sufficiently capture the intricacies of complex graph topologies. Specifically, in graphs characterized by high cyclicity or dense interconnectivity, the hierarchical simplification imposed by tree-based representations could lead to an oversimplification of relational patterns, potentially compromising the model's ability to generalize effectively.

W2. While it is appropriate to provide the detailed theoretical proofs in appendix, it would be beneficial to provide a concise summary of the key insights in the main text, such as the mechanism by which boundary-aware convolution enhances the signal-to-noise ratio. This approach would significantly improve the accessibility and clarity of the work for readers.

W3. There is little discussion of scenarios where it might underperform (e.g., sparse graphs, small subgraphs). This could highlight limitations and guide future work.

Typo: Line 804 'real-work datasets' -> 'real-world datasets'

其他意见或建议

Please answer W1-W3

作者回复

We deeply appreciate your kind words regarding the clarity of our presentation. Thank you for acknowledging the thoroughness of our theoretical and experimental sections. Below, we have responded to the weaknesses raised by the reviewer:

W1: Concerns about tree-based encoding and complex graph topologies

We appreciate this thoughtful observation. Our extensive experiments were conducted on eight public benchmark datasets that have been widely used in previous subgraph representation learning studies [1-4]. We evaluated GPEN on eight public datasets comprising various graph structures, including dense networks and small subgraphs. For clearer perspective, here's a summary of these datasets:

DatasetNodesEdgesAvg. DegreeSubgraphsAvg. Nodes per Subgraph
Density5,00029,5215.9025020.0 ± 0.0
Cut Ratio5,00083,96916.7925020.0 ± 0.0
Coreness5,000118,78523.7622120.0 ± 0.0
Component19,55543,7012.2325074.2 ± 52.8
PPI-BP17,080316,95118.561,59110.2 ± 10.5
HPO-METAB14,5873,238,174222.01,40014.4 ± 6.2
HPO-NEURO14,5873,238,174222.04,00014.8 ± 6.5
EM-USER57,3334,573,41779.77324155.4 ± 100.2

Notably, GPEN achieves superior performance (0.912 micro-F1) on EM-USER, which contains over 4.5 million edges, demonstrating that our approach effectively captures complex structural information in densely connected graphs.

This effectiveness primarily stems from the Maximum Spanning Tree (MaxST) methodology, which preserves critical structural information during tree construction by leveraging node importance scores derived from PageRank. The weighting scheme ensures connections between high-importance nodes are prioritized in the resulting tree structure, effectively capturing the backbone of the graph's hierarchical organization. Our tree analysis (Table 6) shows MaxST consistently outperforms other tree construction methods across all datasets, confirming its ability to capture essential graph topology even when simplifying complex structures. As other reviewers noted, our approach "effectively clarifies the practical significance of utilizing a hierarchical tree in the study."

[1] Alsentzer et al., "Subgraph Neural Networks", NeurIPS 2020

[2] Wang & Zhang, "Glass: GNN with labeling tricks for subgraph representation learning", ICLR 2022

[3] Jacob et al., "Stochastic subgraph neighborhood pooling for subgraph classification", CIKM 2023

[4] Kim & Oh, "Translating subgraphs to nodes makes simple GNNs strong and efficient for subgraph representation learning", ICML 2024

W2: Request for concise theoretical summaries in main text

We appreciate this constructive suggestion. Our theoretical analysis establishes four key properties:

  1. Theorem 4.1 proves the controllability of GPEN's representations
  2. Theorem 4.2 proves the distinctiveness of global position encoding
  3. Theorem 4.3 proves that the boundary aware convolution module can suppress noise propagation
  4. Theorem 4.4 proves that the empirical distribution covers a wider range after perturbation.

In our revised manuscript, we will add concise summaries after each theorem in Section 4, highlighting key insights in accessible language and demonstrating how these theoretical properties address the challenges outlined in our introduction.

W3: Discussion of potential underperformance scenarios

We thank the reviewer for this valuable suggestion. Regarding synthetic datasets with small subgraphs (Table 3), these were specifically designed to test different capabilities. For density and component tasks, where GNNs already perform well, our model complements these strengths. As shown in Figure 2, our position encoding supplements original features rather than replacing them, while our boundary-aware convolution prevents excessive aggregation. For cut-ratio and coreness datasets, which require more sophisticated structural understanding, our approach effectively captures the necessary information.

Our experiments do reveal some insights about potential limitations. In the Component dataset (Table 3), which has the lowest average degree (2.23) among all datasets, multiple methods including GPEN achieve perfect scores, suggesting that for very sparse graphs with clear component structures, simpler methods may be equally effective. Additionally, our hyperparameter analysis (Figure 4) shows varying sensitivities across datasets, with Coreness exhibiting more pronounced performance fluctuations with changes in the balance factor b, indicating that optimal parameter tuning may be more critical for certain graph structures.

Typo correction

Thank you for pointing out the typo on line 804. We will correct "real-work datasets" to "real-world datasets" in the revised manuscript.

审稿意见
4

This paper presents GPEN (Global Position Encoding Network), a novel approach for subgraph representation learning that addresses the limitation of existing methods which primarily focus on local neighborhood structures while overlooking global structural information. GPEN implements two key modules: (1) global position encoding, which leverages hierarchical tree structure to encode each node's global position, enabling the capture of structural relationships between distant nodes; and (2) boundary-aware convolution, which computes difference vectors between nodes to control information flow, selectively integrating global structural information while maintaining the unique structural patterns of each subgraph. Experiments show that GPEN achieves competitive or superior performance compared to state-of-the-art methods on eight public datasets.

给作者的问题

Among the different tree construction algorithms, the Maximum Spanning Tree performs relatively well. Is this advantage consistent across all types of graph structures, or is it more significant for certain types of graphs (such as sparse or dense graphs)?

论据与证据

The main claims of the paper are well-supported by theoretical analysis and experimental results. The authors claim that GPEN can effectively capture global structural information while preserving the unique features of subgraphs, which is supported by multiple theoretical analyses and relatively comprehensive experimental validation. Theoretically, the authors provide four theorems with proofs that analyze the bounded representation discrepancy, global position encoding distinctness, noise robustness of boundary-aware convolution, and generalization guarantees for tree perturbation. Experimentally, results on four real-world datasets and four synthetic datasets show that GPEN performs well on subgraph representation learning tasks. The ablation study (Table 4) demonstrates the contribution of each component, validating the complementary nature of the three modules: Global Position Encoding (GPE), Boundary-Aware Convolution (BWC), and Optional Tree Perturbation (OTP).

方法与评估标准

The proposed methods and evaluation criteria are appropriate for the research problem. The authors use the same datasets and data splits as previous work, ensuring comparability of results. The evaluation uses micro-F1 scores as metrics, which is a common measure for subgraph classification tasks. Experiments are conducted on various datasets with different characteristics, including molecular biology, clinical diagnostics, and user profiling datasets from real-world scenarios, as well as synthetic datasets designed to test the ability to recognize different structural features. Notably, the authors explore the impact of different tree construction algorithms (Table 6), demonstrating the superiority of the Maximum Spanning Tree algorithm, which enhances the credibility of their method selection.

理论论述

The theoretical claims in the paper are supported by sound mathematical proofs. Section 4 provides four core theorems: Theorem 4.1 proves that the discrepancy between GPEN representations and standard GNN representations is bounded; Theorem 4.2 establishes that global position encoding effectively distinguishes nodes with different connectivity patterns; Theorem 4.3 analyzes the noise robustness of boundary-aware convolution, proving it achieves higher signal-to-noise ratio compared to standard GNN aggregation; and Theorem 4.4 provides generalization bounds for the tree perturbation technique. The proofs are elaborated in Appendix A.1 with rigorous derivations using appropriate mathematical tools such as Lipschitz continuity, Perron-Frobenius theorem, and PAC-Bayes framework.

实验设计与分析

The experimental design is reasonably sound. The authors adopt the same datasets and data splits as the baselines, facilitating fair comparison. The experiments cover multiple aspects: performance comparison on real-world datasets (Table 1), controlled experiments on synthetic datasets (Table 3), ablation studies (Table 4), and hyperparameter analysis. The authors explore the impact of different tree construction algorithms (Table 6), comparing Breadth-First Search Tree, Depth-First Search Tree, Minimum Spanning Tree, and Maximum Spanning Tree algorithms.

补充材料

The authors submitted the code and configuration files. I checked the convolution process in the code and it is consistent with the paper.

与现有文献的关系

The paper has connections to the scientific literature. In Section 6, the authors review related work, including the development of Graph Neural Networks and advances in subgraph representation learning. The paper discusses existing methods such as SubGNN, GLASS, SSNP, and S2N, and explains how GPEN attempts to address some of the challenges faced by these methods from the perspectives of global structural information and selective information integration. This literature connection helps to understand the research background and contributions of GPEN.

遗漏的重要参考文献

The paper discusses most of the relevant important references. The authors' review of related work covers the progression from fundamental work on Graph Neural Networks to advances in subgraph representation learning methods.

其他优缺点

Strengths:

  1. The global position encoding proposed in GPEN is a novel approach that systematically captures multi-hop relationships between nodes in a graph.
  2. The paper provides comprehensive and rigorous theoretical analysis, proving the effectiveness and stability of the method.
  3. Experiments are conducted on multiple datasets, validating various aspects of the method through ablation studies and hyperparameter analysis.
  4. The paper mentions using COO-formatted sparse adjacency matrices to calculate difference vectors of node representations, demonstrating consideration for memory efficiency.

Weaknesses:

  1. Some figures' labels and color choices could be clearer, particularly the visualizations in Figure 4.
  2. The paper discusses the impact of threshold c but lacks methods for automatically determining the optimal threshold.

其他意见或建议

Some figures in the paper could be improved, such as the y-axis labels and line colors in Figure 4. Consider adding grid lines or using different line types to distinguish different data series, which would help readers better interpret the results.

作者回复

We thank the reviewer for their detailed comments. We appreciate that they found our "theoretical claims in the paper are supported by sound mathematical proofs" and acknowledged that our "experimental design is reasonably sound." Below, we have responded to the weaknesses raised by the reviewer:

Weakness 1 & Other Comments/Suggestions: Improving Figure Clarity

Thank you for your suggestion regarding the color coding in the captions of tables. For the revised manuscript, we will:

  • Enhance the clarity of y-axis labels and line colors in Figure 4.
  • Add grid lines to facilitate easier interpretation of results in Figure 4.

These visual improvements will make our results more accessible and interpretable for readers.

Weakness 2: Methods for Automatic Threshold Determination

We agree with the reviewer that such methods can be very insightful. However, we would like to note that since the tree perturbation module is optional (as not all datasets suffer from insufficient samples), and as reviewer FTVY13 noted, we don't generate too many samples, making manual parameter tuning sufficient for most applications.

In Figure 3 of our manuscript, we presented a comprehensive analysis of how different threshold values cc affect model performance. For enhanced clarity, we have reformatted these experimental results in tabular form below:

Threshold c value2468
ppi_bp0.6110.6220.6440.633
hpo_metab0.6380.6100.6030.598
hpo_neuro0.6710.6910.6810.681
em_user0.8760.9030.9120.902

This table reveals that while threshold selection influences performance, the variation is generally moderate. GPEN demonstrates relatively stable performance across a reasonable range of threshold values (c=4 to c=8), suggesting it is not overly sensitive to the exact choice of c. This stability reduces the need for precise automatic tuning. For the hpo_neuro dataset, we observe identical performance at thresholds 6 and 8 because there are no remaining connected components larger than these thresholds, which inherently limits the number of samples our method generates. This natural upper bound on generated samples further reduces the necessity for complex automatic threshold determination methods. In addition, the tree perturbation module is optional and primarily benefits datasets with insufficient samples. For datasets with abundant samples, this module may not be necessary.

Question: Maximum Spanning Tree Performance Across Graph Types

Thank you for this insightful question about the consistency of Maximum Spanning Tree (MaxST) performance across different graph structures. In Appendix A.2.4, we conducted a comprehensive analysis of different tree construction algorithms and their impact on GPEN's performance. As shown in Table 6 of our appendix, we evaluated four representative tree construction methods: Breadth-First Search Tree (BFS), Depth-First Search Tree (DFS), Minimum Spanning Tree (MST), and Maximum Spanning Tree (MaxST). Our analysis shows that while MaxST consistently outperforms other tree construction algorithms across all datasets, its advantage is indeed more pronounced in certain graph types.

The experimental results show that the Maximum Spanning Tree algorithm consistently achieves superior performance across all datasets. This is primarily because MaxST effectively preserves critical structural information during the tree construction process by utilizing node importance scores derived from PageRank. The weighting scheme ensures that connections between high-importance nodes are prioritized in the resulting tree structure, effectively capturing the backbone of the graph's hierarchical organization. As reviewer FTVY13 noted, this "effectively clarifies the practical significance of utilizing a hierarchical tree in the study."

The advantage of Maximum Spanning Tree is particularly pronounced in dense graphs with higher average node degrees. For instance, in dense networks such as hpo-metab and hpo-neuro, MaxST achieves significantly better results (0.638 ± 0.009 and 0.691 ± 0.006 respectively) compared to BFS (0.515 ± 0.031 and 0.606 ± 0.037) and DFS (0.494 ± 0.030 and 0.605 ± 0.034). This performance advantage stems from MaxST's ability to identify and preserve important pathways in complex network topologies, resulting in more meaningful global position encodings that better capture the structural relationships between distant nodes.

审稿意见
3

The paper introduces a method called GPEN for Subgraph Representation Learning. It proposes the construction of a hierarchical tree to compute the Global Position Encoding (GPE) and introduces Boundary-aware Convolution (BWC) and tree-based Optional Tree Perturbation (OTP). These strategies aim to address two major challenges in graph representation learning: capturing structural relationships between distant nodes and preventing excessive aggregation of global structural information.

给作者的问题

See above.

论据与证据

The paper provides detailed theoretical analysis and validation for the GPE, BWC, and OTP in GPEN.

方法与评估标准

Experiments show that GPEN has better average performance and lower standard deviation. However, the experimental results are not significant, as detailed in [W4].

理论论述

The paper proposes several theorems as theoretical support for the effectiveness of the modules in GPEN, and provides comprehensive proofs in the appendix.

实验设计与分析

The paper demonstrates through experiments that the standard deviation of GPEN is significantly lower than other methods, indicating better robustness. However, several conclusions from the experiments are not significant, as detailed in comment W4.

补充材料

The paper supplements a large number of experiments in the appendix and provides comprehensive proofs for the theorems proposed in the main text. Particularly, the experiments and analyses carried out on tree construction algorithms suggest that the use of the maximum spanning tree "ensures that connections between high-importance nodes are prioritized in the resulting tree structure." This effectively clarifies the practical significance of utilizing a hierarchical tree in the study, which addresses the concerns and questions I had while reading the main text.

与现有文献的关系

The paper identifies two major challenges in existing Subgraph Representation Learning methods: capturing structural relationships between distant nodes and preventing excessive aggregation of global structural information. It proposes to construct a hierarchical tree and designs the GPE, BWC, and OTP modules to address these challenges.

遗漏的重要参考文献

N/A

其他优缺点

Strengths:

The paper provides detailed theoretical analysis and validation for each module.

Weaknesses:

[W1] In section 3.1.1, as the transition matrix of PageRank, why MM is Mij=1di\mathbf M_{ij}=\frac{1}{d_i} instead of 1dj\frac{1}{d_j}? Also, what is the initial value of RR in equation 3?

[W2] Section 3.1.3 mentions that "The insufficient number of subgraphs can affect the model’s stability." However, the tree perturbation module generates at most one new sample for each subgraph. Therefore, the total number of subgraphs is at most twice the original number. Is this sufficient to address the issue of insufficient subgraphs?

[W3] In section 3.2, the notion in equation 12 is unclear:

  • What is A\mathbf A? It seems to be undefined in the paper.
  • What does the subscript nn in Hn(l1)\mathbf H_n^{(l-1)} stand for? Is it a different matrix from H(l1)\mathbf H^{(l-1)}?
  • Is W(l1)\mathbf W^{(l-1)} the edge weight of the weighted graph in section 3.1.1? If so, what does the superscript (l1)(l-1) represent? The edge weight in section 3.1.1 seems to be a constant independent of the layer.

[W4] In Table 1, the pp-values comparing GPEN with S2N on all datasets are greater than 0.05, indicating non-significant differences. The same applies to the experiments in Table 3 and the ablation study in Table 4. The advantage of GPEN seems to lie only in its more stable performance and smaller variance.

其他意见或建议

In section 3.1.2, the global positional encoding groups nodes according to their depth in the tree, then encodes them directly into a 01 vector. It loses the information of its value. Especially considering that the edge weights and trees here are artificially calculated and constructed, the shortest path length on the tree may not represent real and accurate global information. Would it be better to use an encoding that retains the depth value of the nodes, such as the sin/cos function encoding designed for sequences in the classic "Attention is all you need"?

作者回复

We sincerely appreciate your positive feedback and constructive remarks on our paper. Below, we provide a detailed response to your questions and comments.

[W1 and W3]

We thank the reviewer for pointing out these notation issues throughout the paper. To clarify:

  • MijM_{ij} : Thank you for bringing this to our attention. You are correct; Mij=1djM_{ij} = \frac{1}{d_j} if there is an edge between nodes ii and jj, where djd_j is the out-degree of node jj. The incorrect notation in the paper was a typographical error.
  • RR: The initial value of RR in equation 3 is a uniform distribution where each element equals 1V\frac{1}{|V|}, with V|V| being the number of nodes in the graph.
  • A\mathbf{A}: This represents the adjacency matrix of the graph, where Aij=1A_{ij} = 1 if there is an edge between nodes ii and jj, and Aij=0A_{ij} = 0 otherwise.
  • Hn(l1)\mathbf{H}_n^{(l-1)}: The subscript nn was a typographical error. It should be H(l1)\mathbf{H}^{(l-1)}, representing the node embeddings from layer l1l-1.
  • W(l1)\mathbf{W}^{(l-1)}: This refers to the trainable parameter matrix in the graph convolution operation, not the edge weights wijw_{ij} mentioned in section 3.1.1. The superscript (l1)(l-1) indicates that this parameter matrix is associated with the transformation from layer l1l-1 to layer ll.

We will correct these notations in the revised manuscript and use more distinctive symbols to avoid confusion.

[W2]

Our experimental results show that tree perturbation is effective with modest sample increases, though excessive samples may introduce noise. As shown in Figure 3 in our paper, we conducted detailed experiments on the impact of different threshold values c on model performance. Due to space limitations, we present the numerical results in our response to Reviewer 2fVD under "Weakness 2: Methods for Automatic Threshold Determination." These results clearly identify that even a relatively small number of additional samples can significantly improve the model's performance, while excessive samples may actually harm performance.

We agree that this is an important consideration. However, our primary goal is to explore the potential of tree structures in subgraph representation learning. While our results show promising improvements, we do not claim this approach "completely solves the issue of insufficient subgraphs" in our paper. We recognize that developing high-quality data augmentation methods requires consideration of many factors, which is beyond the scope of our current work.

[W4]

We thank the reviewer for the detailed statistical analysis. However, we respectfully argue that a holistic evaluation reveals significant advantages for GPEN.

Crucially, GPEN shows superior performance over S2N on most datasets, particularly on synthetic datasets specifically designed to evaluate distinct structural learning capabilities. On these challenging tasks, GPEN outperforms all baseline methods, showcasing its robust structural understanding. In stark contrast, S2N not only fails to match GPEN but actually exhibits performance degradation compared to other established baselines. For example, on cut-ratio, S2N scores 0.892 vs. GLASS's 0.935 vs. GPEN's 0.936, and on coreness, S2N scores 0.726 vs. GLASS's 0.840 vs. GPEN's 0.876.

While S2N achieves comparable mean results on a few datasets, it does so with markedly high standard deviations compared to other baselines (averaging a 43.2% increase over SubGNN and a 127.7% increase over GLASS). GPEN, conversely, achieves its strong performance with significantly lower standard deviations across the board (averaging a 59.8% reduction compared to SubGNN and 37.5% compared to GLASS).

[S1]

Thank you for this insightful suggestion regarding alternative encoding methods for the global positional information. Our choice of one-hot encoding for tree depths was based on several practical considerations:

First, the inherent message-passing mechanisms in GNNs enable the model to naturally learn relationships between different depth levels during convolution operations, making explicit continuous encoding less critical in this context. Second, node depth primarily functions as a categorical feature that distinguishes nodes. The one-hot representation provides clear separation between these structural groups. Third, unlike transformer models that should generalize across unseen positions, our model operates on specific graph structures where the positional relationships are fixed, reducing the need for the interpolation capabilities that make sin/cos function encodings valuable in sequence modeling.

We conducted additional experiments comparing one-hot encoding with sin/cos function encodings, but due to space limitations, we apologize that we couldn't present these results. Consistent with our reasoning, these experiments showed no significant performance improvements, as both encoding methods capture the same fundamental structural information.

最终决定

This paper introduces GPEN (Global Position Encoding Network), a novel method aimed at enhancing subgraph representation learning (learning the representation of a subgraph in a graph) by effectively capturing global structural information through hierarchical tree-based position encoding and a boundary-aware convolution module. The reviewers generally appreciate the clear motivation, rigorous theoretical analysis, and thorough empirical validation across diverse datasets. Two reviewers recommended acceptance (score 4), while two give weak accept (score 3). Given the important problem of subgraph representation learning, high-quality theoretical contributions, comprehensive empirical validation, clarity of writing, and thorough responses addressing reviewer concerns, the consensus is to recommend acceptance of the paper, subject to the inclusion of the promised clarifications and improvements in the final manuscript.