PaperHub
6.0
/10
Poster4 位审稿人
最低5最高7标准差0.7
5
6
6
7
3.8
置信度
正确性2.8
贡献度2.5
表达2.5
NeurIPS 2024

UGC: Universal Graph Coarsening

OpenReviewPDF
提交: 2024-05-15更新: 2025-01-13
TL;DR

UGC is a graph coarsening framework which is extremely fast, has lower eigen-error, and yields superior performance on downstream processing tasks.

摘要

In the era of big data, graphs have emerged as a natural representation of intricate relationships. However, graph sizes often become unwieldy, leading to storage, computation, and analysis challenges. A crucial demand arises for methods that can effectively downsize large graphs while retaining vital insights. Graph coarsening seeks to simplify large graphs while maintaining the basic statistics of the graphs, such as spectral properties and $\epsilon$-similarity in the coarsened graph. This ensures that downstream processes are more efficient and effective. Most published methods are suitable for homophilic datasets, limiting their universal use. We propose **U**niversal **G**raph **C**oarsening (UGC), a framework equally suitable for homophilic and heterophilic datasets. UGC integrates node attributes and adjacency information, leveraging the dataset's heterophily factor. Results on benchmark datasets demonstrate that UGC preserves spectral similarity while coarsening. In comparison to existing methods, UGC is 4x to 15x faster, has lower eigen-error, and yields superior performance on downstream processing tasks even at 70% coarsening ratios.
关键词
Graph CoarseningGraph Neural NetworksLocality sensitive hashingHeterophilic GraphScaling Graph Learning

评审与讨论

审稿意见
5

Graph coarsening aims at scaling an original large graph into a small graph. This paper proposes a graph coarsening method which was designed to be equally suitable for homophilic and heterophilic datasets, specifically, aggregating node clusters identified by a hash function. This paper is one of the pioneering works exploring graph coarsening in heterophilic graph datasets, and the proposed solution seems promising. However, the current formulation of the proposed method and some of the claims in the paper are not rigorous and solid enough.

优点

  • This paper is one of the pioneering works exploring graph coarsening in heterophilic graph datasets.
  • Aggregating hypernodes through a hash function instead of neighbors is a promising solution to perform heterophilic graph coarsening.

缺点

  • The presentation of this paper is unclear. Check out `Questions’.
  • The current formulation of the proposed Hash functions’ is ambiguous and biased. Check out Questions’.
  • Some claims were in fact wrong. Check out `Questions’.

问题

  • Line 4. vital insights’, essential features’ in line 5. and ‘vital information’ in line 20. are too vague. I suggest being more specific and keeping the phrasing unchanged.
  • Line 67-68. `The Graph Coarsening(GC) problem \bf{may be} thought of as learning a coarsening matrix….’ Is it or is it not? A formal definition of the Graph Coarsening problem is required whether cited from others or by you.
  • Line 128-129. `Due to this, these methods are extremely computationally demanding and defeat the purpose of reducing the original graph.’ This statement suggests that ‘GNN-based Graph Condensation’ is useless. I suggest being more objective.
  • Line 143-144.The concepts of homophily and heterophily in graphs are primarily concerned with whether the edges between two nodes align with their respective features.’ is not correct. Recall that homophily and heterophily are defined by whether edges align with labels, not features. You need to state a hypothesis or experimentally verify that:the same labels imply similar features’.
  • Line 155-160. The current formulation hash function is biased to node degree distribution since you concat the adj with node feature. Consider this example that a graph has 1 node whose degree is 999 and other 999 nodes’ degree is 1 (a radial graph). In this case, the hash value of the degree 999 node will be significantly larger than the other nodes, e.g. 90-99 to 1-10, and setting `r’ parameter only makes it worse. Meanwhile, the 11-89’th hypernode is empty. This is an extreme case of long-tail distribution of node degrees, however, simple concatenation inevitably causes the number of aggregated nodes in the hypernode to be uneven, specifically strongly correlated with the node distribution.
  • Again, concat the adj with node feature is ambiguous. Considering that the hash result will change w.r.t. the change of index order of nodes.
  • What is ‘|||’ in the proof of Theorem 3.1 in line 575? Moreover, the derivation in the proof is speculative. You need to explain clearly why each step of the inequality holds.
  • Line 180-181. `It means that the adjacency matrix A~ has a substantially smaller number of non-zero elements than A.’ is not correct. Consider a chain graph with N nodes, N-1 edges, and if the head node and the tail node were aggregated, it became a circle with N-1 nodes and N-1 edges, where the number of edges remains unchanged.
  • Where is the introduction of baselines from Var.Neigh.’ to Kron’? If somewhere else, I suggest moving to Chap experiments.
  • Figure 7 is confusing. Present `actual time saved’ (time by downstream task on original graph – time by downstream task on coarsened graph + time by coarsening the graph) together with downstream task performance loss/gain by your method and baselines will be more meaningful.

局限性

The authors have adequately addressed the limitations.

作者回复

We thank the reviewer for their valuable comments and insights and for taking the time to go through our paper.

Ques 1) Regarding .. *Line 4. vital insights’ *

Ans 1) We thank the reviewer for the suggestion. By vital insights and vital information, we mean retaining the basic statistics of the graphs, such as spectral properties and ϵ\epsilon-similarity in the coarsened graph. This ensures that downstream processes are more efficient and effective. Given the opportunity, we can integrate these clarifications into the manuscript.

Ques 2) Regarding Line 67-68. The Graph Coarsening(GC) problem ..

Ans 2) The Graph Coarsening (GC) problem is indeed about learning a coarsening matrix. This definition is well-established in the literature [1, Section 2], [2, Section 2.2]. The objective and problem formulation are also discussed in Section 2.

[1] Hashemi, Mohammad, et al. "A comprehensive survey on graph reduction: Sparsification, coarsening, and condensation." arXiv preprint arXiv:2402.03358 (2024).

[2] Kumar, Manoj, et al. "Featured graph coarsening with similarity guarantees." International Conference on Machine Learning. PMLR, 2023.

Ques 3) Regarding Line 128-129. Due to this, these methods ....

Ans 3) We thank the reviewer for the suggestion. The main reason behind this statement was the following results, where we compared the coarsening time using GCond with the whole graph training time. As it can be seen from the table, the time it takes to coarsen the graph is multiple times higher than the training time. However, the above-mentioned statement can be rephrased as :

...these methods are extremely computationally demanding and may not be suitable for the scalability of GNN models. However, these methods can be beneficial for other tasks, like solving storage and visualization issues.

GCond accuracy and time

DataGCond AccGCond Coarsen TimeGCN training time on original graphUGC Coarsening Time(x Fast compared to GCond)
Cora80.43264025.770.41(x6440)
Pubmed76.981620114.551.62(x1000)
PhysicsOOM-1195.566.4
DBLP82.6325500174.101.86(x13710)
Squirrel59.647860228.522.14(x3673)
Chameleon52.29774054.340.49(x15796)

Ques 4) Regarding Line 143-144.The concepts of homophily and heterophily .....

Ans 4) We understand the statement mentioned can be misleading and therefore we have corrected the line to the following:

The concepts of homophily and heterophily in graphs are primarily concerned with whether the edges between two nodes align with their respective labels.

Ques 5) Regarding.. Line 155-160. The current formulation hash function ... Consider this example that a graph has 1 node whose degree is 999 and......

Ans 5) We appreciate your concern. To clarify, in the example cited by the reviewer, the node with a degree of 999 will indeed produce a higher hash value and will be collapsed into a super-node. This is reasonable because it is an important node and should be assigned to a different super-node. It is important to note that we ensure no super-node is empty during the formulation of the super-nodes. If there is no hash value between the range of 11-89, we will not create a super-node for this range. For the other nodes with a degree of 1, as these nodes are connected to the same high-degree node, the distinction between the hash values will be governed by their node feature vectors. The super-node index is determined by setting the appropriate bin-width value "r".

Ques 6) Regarding.. Again, concat the adj with node feature is ambiguous...

Ans 6) We appreciate your concern. To clarify, the order of the nodes is fixed and consistent across all l projectors at the start of the coarsening and does not change during the process. Once the node order is established, the LSH framework's locality-preserving property ensures that the hashing process remains stable.

Ques 7) Regarding. What is ‘|||’ in the proof of Theorem 3.1 in l......

Ans 7) We thank the reviewer for pointing out the typo. .p||.||_p denotes pthp^{th} norm. We have corrected the Proof of Theorem 3.1 here is the revised version:

Proof: Let S be defined such that L = STSS^TS, by Cholesky’s decomposition:

xLxcLc=Sx2SP+Px2| \lVert x\rVert_{L} - \lVert x_c\rVert_{L_c} | = | \lVert Sx\rVert_{2} - \lVert SP^+Px\rVert_{2}|

From Modulus inequality property,

$

\leq ||Sx - SP^+Px||_2 = ||x - \widetilde{x}||_L \leq ||x||_L

$

The conversion from LnormL-norm to 2ndnorm2^{nd}-norm or vice-versa is as follows:

xL=xTLx=xTSTSx=Sx2\lVert x\lVert_L = \sqrt{x^T L x} = \sqrt{x^T S^T S x} = \lVert Sx\lVert_2

Ques 8) Regarding. Line 180-181. "It means that the adjacency matrix A~ has a substantially smaller number of non-zero elements .....

Ans 8) As mentioned in the Line 177 coarsened graph, adjacency (AcA_c) is directly calculated from CTACC^T A C. In general, AcA_c has a substantially smaller number of non-zero elements than A. However, such extreme cases may exist where the number of edges remains the same and we will make a mention of such cases in the text if given an opportunity.

Ques 9) Regarding Where is the introduction of baselines .....

Ans 9) Section 2 includes the introduction to baseline(line 111-122)

  • UGC: Universal Graph Coarsening, our proposed method
  • VAN: Variation Neighborhood
  • VAE: Variation Edge
  • VC: Variation Clique
  • HE: Heavy Edge
  • aJC: Algebraic Distance
  • aGS: Affinity
  • Kron: Kron

Ques 10) Regarding Figure 7 is confusing. Present actual time saved’.....

Ans 10) The reason this figure is being presented is because the downstream tasks take substantially longer time as compared to the coarsening time in our case, whereas it is not true for the other methods. Adding a large number (downstream time) to a smaller one (coarsening time) could diminish the emphasis on coarsening, which is the focus of this work.

评论

I have read your replies and the replies corresponding to those of other fellow reviewers. My questions Q1, Q2, Q3, Q4, Q5, Q7, Q8, Q9, Q10 are settled; Q6 remains a shortcoming of the proposed method for the need of fixed node ordering, which should not be the concern of a Graph model'; Consider the progresses been made, I recommend to consider Borderline Accept' this paper where reasons to accept outweigh reasons to reject.

审稿意见
6

This paper proposes a new Universal Graph Coarsening (UGC) framework designed to handle both homophilic and heterophilic graphs. The UGC framework is capable of retaining important spectral properties, including eigenvalue error, hyperbolic error, and 𝜖-similarity measure. Experimental results demonstrate significant improvements in both performance and effectiveness with the UGC framework.

优点

  • Derive the coarsening matrix over graphs with heterophily property is an interesting idea, and Locality Sensitive Hashing (LSH) technique significantly reduces the algorithm's complexity.
  • UGC offers significant speed improvements and is capable of handling both homogeneous and heterogeneous graphs.
  • The experiments has demonstrated UGC’s limitations and possibility to improve the effectiveness across datasets.
  • The paper is also very clear with thorough experiments and analysis.

缺点

  1. Although the LSH strategy is much faster than other methods, the memory space overhead is non-negligible. Could you please measure the space complexity and provide a further empirical evaluation of the method? Additionally, how is the number of hash functions (𝑙) configured?
  2. The paper lacks a detailed description of the experimental setup, including software and hardware environments, as well as the parameters used for different algorithms.
  3. From Figure 5, we can see that UGC does not exhibit lower errors (RRE and HE) at lower Coarsening Ratios. I think it questions the claimed advantage of preserving spectral properties. Can the authors give a detail explanation?
  4. Figure 5 illustrates that the methods (e.g., VAE, VC) do not maintain a monotonic relationship between the Coarsening Ratio and RRE/HE Error. Additionally, UGC does not exhibit a clear advantage across the overall Coarsening Ratio. Can the authors make a further clarify?
  5. Table 3 does not prove the claim of “UGC (features) method achieves substantial accuracy enhancements” over GCN, GraphSAGE, GIN and GAT. Can the authors utilize UGC to more recent models, e.g., 3WL-GNNs, heterogeneous graph neural networks?
  6. Minor issues:
    • In Figure 2, the specific meanings of a, b, and c are not labeled in the diagram.

问题

Please see the weakness section.

局限性

The authors have adequately addressed the limitation of the potential negative societal impact of their work.

作者回复

We thank the reviewer for their valuable comments and insights and for taking the time to go through our paper.

Ques 1) Although the LSH strategy is much faster than other methods, the memory space overhead is non-negligible. Could you please measure the space complexity and provide a further empirical evaluation of the method? Additionally, how is the number of hash functions (𝑙) configured?

Ans 1) The memory space overhead by UGC arises from two steps

a) Random sampling of lRdl \in R^d different projectors,

b) Storing the hash values of each of 'n' nodes across these l projectors.

Hence, the additional space complexity is bounded by O(l*d + n*l). l is a hyperparameter, for all our experiments, and all datasets, we have used 3000 different projectors(l). It is worth noting that for some datasets, the value of l can be as low as 1000.

Additionally, due to its construction, UGC is suitable for online or streaming data as it does not require the entire feature matrix to be present simultaneously. In scenarios with limited memory space, the feature matrix can be split into different chunks, and UGC can be applied to each chunk separately, ensuring that the same memory is reused for all chunks. In this case, the required additional space complexity is bounded by O(l*d + n'*l) where n' is the size of the chunk and it can vary from 1 to n.

Ques 2) The paper lacks a detailed description of the experimental setup, including software and hardware environments, as well as the parameters used for different algorithms.

Ans 2) The details are included in the Appendix due to limited space in the main manuscript. For convenience, we also provide the details here:

Appendix H contains a detailed discussion about the experimental setup as mentioned in Line 301. Figure 9 provides an overview of the GCN training pipeline, while Table 7 includes information on the hyper-parameters used for training.

Appendix F contains details about the hardware specifications.

"All experiments conducted for this work were performed on a desktop with an Intel Xeon W-295 CPU and 64GB of RAM using the Python environment."

Ques 3) and Ques4)

  • From Figure 5, we can see that UGC does not exhibit lower errors (RRE and HE) at lower Coarsening Ratios. I think it questions the claimed advantage of preserving spectral properties. Can the authors give a detail explanation?

  • Figure 5 illustrates that the methods (e.g., VAE, VC) do not maintain a monotonic relationship between the Coarsening Ratio and RRE/HE Error. Additionally, UGC does not exhibit a clear advantage across the overall Coarsening Ratio. Can the authors make a further clarify?

Ans 3) and Ans4) In Figure 5, lower coarsening ratios indicate that the graph is significantly reduced. We observe that the RRE error for UGC is optimal when the graph is reduced by approximately 0-65%. For HE error, we acknowledge that UGC is not the best-performing algorithm, but it is comparable to existing methods. However, for the node classification task, we can see that UGC gives the best results. How these properties are related to downstream tasks is not well understood yet and warrants further investigation.

We have noticed the monotonic relationship that the reviewer mentioned but at the moment we are trying to analyze the data from additional experiments so that we can make an objective statement after a thorough assessment.

Ques 5) Table 3 does not prove the claim of “UGC (features) method achieves substantial accuracy enhancements” over GCN, GraphSAGE, GIN and GAT. Can the authors utilize UGC to more recent models, e.g., 3WL-GNNs, heterogeneous graph neural networks?

Ans 5)

We have added the results of 3WL-GNN methods. It is to be noted that originally 3WL-GNN was a graph classification method, we have made the necessary changes to adapt it to node classification task.

Data\ModelVar.NeighVar.EdgesVar.CliqueHeavy EdgeAlg. Dis.Aff. GSKronUGCFull Dataset
Cora57.5543.3053.1260.3160.0152.9356.7555.3263.45
DBLP48.6049.8351.7651.9551.7452.1352.1953.1061.58
Physics85.9984.8387.0283.4982.1285.8987.8888.5692.87
Pubmed69.9464.2970.4263.1153.5527.0462.3084.1586.68
Squirrel19.8319.2819.7120.0320.8620.0420.8261.9231.73
Film17.1326.1010.9322.7226.3327.1618.3124.8231.64
Chameleon18.4022.5920.0323.7823.0416.6323.2469.0144.61

In 5 out of 7 datasets UGC gives the best results for 3WL-GNN.

Ques 6) In Figure 2, the specific meanings of a, b, and c are not labeled in the diagram.

Ans 6) We thank the reviewer for pointing it out, Given the opportunity, we will improve the diagram to avoid any confusion.

评论

Thank you very much for your reply. My assessment of the paper remains positive.

审稿意见
6

The authors propose a novel Universal Graph Coarsening (UGC) framework, which is suitable for both homophilic and heterophilic datasets. UGC integrates node attributes and adjacency information to leverage dataset heterogeneity effectively. The results demonstrate that UGC is significantly faster (4x to 15x), maintains spectral similarity, and outperforms existing methods in terms of computational efficiency, eigen-error, and downstream processing tasks, especially at 70% coarsening ratios. The key contributions highlight UGC's universal applicability, efficiency, and information preservation.

优点

  1. The approach is intuitive and easy to understand.
  2. The approach has a lower computational cost than most common method (Var.Neigh. etc,.)

缺点

For ScalableTrainingof Graph Neural Networks section, there are no detailed discussion on GNN models except GCN.

问题

For Table 4, why only GCN results are shown? I also expected results for GIN GAT and GraphSage.

局限性

Limitation is discussed in detail in papers.

作者回复

We thank the reviewer for their valuable comments and insights and for taking the time to go through our paper.

Ques 1) For ScalableTrainingof Graph Neural Networks section, there are no detailed discussion on GNN models except GCN.

Ans 1) Due to the limited space of the manuscript, we have only added a discussion about GCN to the manuscript. If the reviewer suggests, the following detailed discussion can be added to the paper:

GraphSAGE [1] is a scalable inductive framework for generating node embeddings in graphs. It leverages a neighborhood sampling and aggregation approach, allowing it to generalize to unseen nodes. This makes GraphSAGE particularly effective for large-scale graphs where retraining the model for new nodes would be computationally prohibitive. The Graph Isomorphism Network (GIN) [2] takes a different approach, designed to be as powerful as the Weisfeiler-Lehman graph isomorphism test. GIN uses a sum aggregation function, ensuring that different graph structures produce distinct embeddings. This ability to distinguish graph structures makes GIN a robust choice for tasks requiring high discriminative power. Graph Attention Networks (GAT) [3], on the other hand, introduce attention mechanisms to graph neural networks. GATs assign different importance to nodes in a neighborhood, which enhances the model's capability to focus on the most relevant parts of the graph. This attention mechanism allows GATs to achieve state-of-the-art performance on various node classification tasks by effectively capturing the underlying structure of the graph.

[1] Hamilton, Will, Zhitao Ying, and Jure Leskovec. "Inductive representation learning on large graphs." Advances in neural information processing systems 30 (2017).

[2] Xu, K., Hu, W., Leskovec, J., & Jegelka, S. (2018). How powerful are graph neural networks?. arXiv preprint arXiv:1810.00826.

[3] Veličković, Petar, et al. "Graph attention networks." arXiv preprint arXiv:1710.10903 (2017).

Ques 2) For Table 4, why only GCN results are shown? I also expected results for GIN GAT and GraphSage.

Ans 2) Due to space limitations in the manuscript, we included only GCN results in the main manuscript. However, we had already conducted experiments with GraphSage, GIN, and GAT for two homophilic and two heterophilic datasets, as shown in Table 3, to demonstrate the model-agnostic behavior of UGC.

As suggested by the reviewer, we have now conducted experiments using GIN, GAT, and GraphSage models for Table 4. These results further demonstrate that UGC is not restricted to any specific model.

DatasetModelVar.NeighVar.EdgesVar.CliqueHeavy EdgeAlg. Dis.Aff. GSKronUGC
gcn20.0329.9531.9233.328.8127.5829.1048.7
Cham.graphSage20.0320.0222.0523.0319.8820.0227.6258.86
gin20.2219.5325.2519.9818.2018.0621.5054.92
gat22.9419.3326.4421.9523.7218.0621.9555.58
gcn19.6720.2219.5420.3619.9620.0018.0331.62
Squ.graphSage19.8720.0020.0320.0319.9320.0019.9857.60
gin18.5419.6518.9821.6519.4718.2920.5635.64
gat20.9018.5620.6819.9320.4620.0520.0832.28
gcn15.6721.8020.3519.1619.2320.3417.4125.40
FilmgraphSage22.3226.0524.0121.4921.8821.5023.7321.12
gin24.2023.5117.5111.4913.9021.9318.0421.12
gat17.5021.7317.8221.1817.9417.4024.1521.71
gcn77.8778.3473.3274.6674.5980.5374.8984.77
pubmedgraphSage78.8562.7367.1860.1163.0971.2562.0083.76
gin74.7739.2946.1935.9732.1349.6339.2976.36
gat75.2272.6374.8160.0469.4759.7671.9283.56
gcn93.7493.8692.9493.0393.9493.0692.2696.12
physicsgraphSageOOMOOMOOMOOMOOMOOMOOMOOM
ginOOMOOMOOMOOMOOMOOMOOMOOM
gat92.0491.8091.4891.8092.9493.3391.6093.80
gcn77.0579.9379.1577.4674.5178.1577.7975.50
dblpgraphSage68.5460.1774.1772.7072.1971.8171.7668.25
gin35.8433.9335.1224.1651.4747.3042.2455.28
gat70.2074.0772.8271.3571.1776.1272.2773.49
gcn79.7581.5780.9279.9079.8380.2080.7186.30
coragraphSage70.4968.4870.1669.1772.2667.7773.2069.39
gin47.6535.0352.9134.0063.0523.4948.5667.23
gat69.2674.0275.9268.9573.0973.8373.2474.21

If the reviewer suggests, we can include these additional results in the Appendix and refer to them in the caption of Table 4 as follows:

Table 4: This table illustrates the accuracy of the GCN model when trained with a 50% coarsened graph. UGC demonstrated superior performance compared to existing methods in 7 out of the 9 datasets. Please refer to Appendix H for results with GraphSage, GIN, and GAT models.

评论

Thanks for your reply. It resolved my concern, and I decided to increase my rating.

评论

Dear Reviewer,

Since we are only a day away from the completion of the discussion phase, we are eagerly awaiting your feedback on the rebuttal.

Your review pointed out important empirical studies that further enhanced our work. We have incorporated all of them and we thank the reviewer again for the deep insightful comments on our work. We would love to discuss more if any concern remains unaddressed. Otherwise, we would really appreciate it if you could support the paper by increasing the score.

regards,

Authors

审稿意见
7

This paper present a framework UGC for graph coarsening to reduce a larger graph to a smaller graph. It uses Locality Sensitive Hashing (LSH) of augmented node features, and works on both homophily and heterophilic graphs. Experiments could verify its effectiveness in original graph property perservation and efficiency in coarsening speed.

优点

S1: This work proposes to use hashing method to graph coarsening and works well on universal homophilic and heterophilic graphs, which is interesting and rational to me.

S2: The overall logic, problem definition, and the solution, are well described and clearly illustrated. The presentation is good and not hard to follow.

S3: UGC is faster compared to existing methods. It achieves a reduction in graph size with lower computational time, making it suitable for large datasets.

缺点

Here are some questions need to be further addressed:

W1: In lines 149-150, it claims that the augmented feature vector is calculated by dot product and concatenation, it is not very clear that how to concate A with node features X?

W2: Some figures, e.g., figure 1 and figure 4, is not very clear for visualization.

W3: In table 4, for comparisons in GNN classification accuracy, it would be better to involve some other graph size reduction methods, like GCOND, SFGC to show the performance.

W4: Is there any ablation study to show the effectiveness of the proposed augmented feature vector?

W5: More detailed explanations and analysis of whether the proposed method could adapt heterophilic graph well are expected. Is there any specific design targeting for heterophily?

问题

See Weaknesses.

局限性

Yes.

作者回复

We thank the reviewer for their valuable comments and insights and for taking the time to go through our paper.

Ques 1) it claims that the augmented feature vector is calculated by dot product and concatenation, it is not very clear that how to concate A with node features X

Ans 1) We thank the reviewer for bringing this typo to our attention. The augmented feature vector is calculated by scaling and concatenation operation instead of dot product and concatenation.

The augmented feature vector of node viv_i is calculated by scaling the adjacency vector AiA_i by (α\alpha) and feature vector XiX_i by (1 - α\alpha), followed by the concatenation operations of the two vectors as (1 - α\alpha)*XiX_i || (α\alpha) * AiA_i, where α\alpha is the heterophily factor.

A detailed illustration of this process is given in Figure 11 in Appendix K. This figure provides a toy example demonstrating how the augmentation matrix is formulated. It is also mentioned in the main manuscript, specifically on lines 150-151.

Ques 2) Some figures, e.g., figure 1 and figure 4, is not very clear for visualization.

Ans 2) Thank you for bringing this to our attention. We will address this issue by increasing the size of the axis labels in Figure 1 and Figure 4 to enhance their clarity. We appreciate your feedback and will make these adjustments in the updated manuscript.

Ques 3) In table 4, for comparisons in GNN classification accuracy, it would be better to involve some other graph size reduction methods, like GCOND, SFGC to show the performance.

Ans 3) We thank the reviewer for the suggestion. We have added the accuracy of GCond in the following table.

GCond accuracy and time

DataGCond AccuracyGCond Coarsening TimeGCN training time on original graphUGC Coarsening Time(x Fast compared to GCond)UGC accuracy
Cora80.43264025.770.41(x6440)86.30
Pubmed76.981620114.551.62(x1000)84.77
PhysicsOOM-1195.566.496.12
DBLP82.6325500174.101.86(x13710)75.50
Squirrel59.647860228.522.14(x3673)31.62
Chameleon52.29774054.340.49(x15796)48.7

As mentioned in the paper, these methods are computationally demanding, which is also evident from the table above. Specifically, the time required to coarsen the graph exceeds the time needed to train the GNN on the original graphs.

These results can be included in Table 1 and Table 4 if suggested by the reviewer.

Ques 4) Is there any ablation study to show the effectiveness of the proposed augmented feature vector?

Ans 4) Yes, the ablation study to demonstrate the effectiveness of the proposed augmented feature vector is included in Table 4. In this table, "UGC feat." denotes the scenario where α\alpha is set to zero, meaning only the feature matrix is considered, while "UGC-feat. + adj." denotes the case where α\alpha is set to heterophily factor, thereby incorporating the adjacency vector.

For heterophilic datasets, node classification accuracy improves significantly when using the augmented feature vector. This highlights the importance of the adjacency vector in the augmented feature vector.

Ques 5) More detailed explanations and analysis of whether the proposed method could adapt heterophilic graph well are expected. Is there any specific design targeting for heterophily?

Ans 5) We thank the reviewer for bringing this up. We have observed that the heterophily factor can be directly utilized as the α\alpha value. When we extrapolated the results for different α\alpha values, it was observed that setting α\alpha around the heterophily factor yielded the best results.

The results are shown in the table below for two heterophilic datasets, Squirrel and Chameleon. For both datasets, the heterophily factor is approximately 0.78. We observed that the best results for these datasets are obtained when α\alpha is set around heterophily factor.

α\alpha valueGCN accuracy for SquirrelGCN accuracy for Chameleon
020.7129.90
0.124.1438.02
0.226.7142.86
0.327.0343.74
0.428.1241.98
0.527.8940.00
0.628.9347.91
0.729.8249.49
0.831.6249.45
0.929.9346.59
1.028.4646.15
评论

Dear Reviewer,

We thank you for the insightful comments on our work. Your suggestions have now been incorporated in our revision and we are eagerly waiting for your feedback. As the author-reviewer discussion phase is approaching its conclusion in just a few hours, we are reaching out to inquire if there are any remaining concerns or points that require clarification. Your feedback is crucial to ensure the completeness and quality of our work.

We are pleased to share that the responses from other reviewers also indicate a positive inclination toward acceptance. Your support in this final phase, would be immensely appreciated.

regards,

Authors

作者回复

We thank the reviewers for their insights and constructive suggestions. A comprehensive point-by-point response to the reviewers' comments is presented below. The major additional changes are listed below.

Additional experiments: We have incorporated all of the additional experiments requested by the reviewers spanning

  • Adding the GCond node classification accuracies and computational time in the rebuttal.
  • Extending Table 4 results to include GraphSage, GIN, and GAT models.
  • Conducting experiments with varying values of the α\alpha hyperparameter from [0,1] to justify the UGC design for handling heterophily datasets.
  • Including node classification experiments with the 3WL-GNNs model.

We hope these revisions will satisfactorily address the concerns raised by the reviewers and elevate the overall quality of our work.

评论

Dear Reviewers,

Thank you once again for all of your constructive comments, which have helped us significantly improve the paper! As detailed below, we have performed several additional experiments and analyses to address the comments and concerns raised by the reviewers.

Since we are into the last two days of the discussion phase, we are eagerly looking forward to your post-rebuttal responses.

Please do let us know if there are any additional clarifications or experiments that we can offer. We would love to discuss more if any concern still remains. Otherwise, we would appreciate it if you could support the paper by increasing the score.

Thank you!

Authors

最终决定

This paper present a framework for graph coarsening to reduce a larger graph to a smaller graph that works on both homophilic and heterophilic graphs. Empirical results verify the effectiveness of the method in preserving graph properties as well as in in coarsening speed.