PaperHub
7.3
/10
Spotlight3 位审稿人
最低6最高8标准差0.9
6
8
8
3.7
置信度
ICLR 2024

NetInfoF Framework: Measuring and Exploiting Network Usable Information

OpenReviewPDF
提交: 2023-09-17更新: 2024-03-20
TL;DR

We propose a framework that measures the usable information of graphs and exploits it to solve the graph tasks, namely link prediction and node classification.

摘要

关键词
Graph Neural NetworksInformation TheoryHeterophily Graphs

评审与讨论

审稿意见
6

The paper discusses the NETINFOF framework, which focuses on measuring and utilizing the usable information in node-attributed graphs for graph tasks such as link prediction and node classification. The authors propose two components of the framework: NETINFOF PROBE, which measures the amount of information present in the graph structure and node features without any model training, and NETINFOF ACT, which uses the measured usable information to solve the graph tasks.

优点

Lots of experiments

缺点

See below

问题

  1. "How to connect the graph information with the performance metric on the task?" Here are some missed related works [1,2,3,4,5].

  2. What is ff in C4? I suggest not using ff, because you already use it as the number of features.

  3. In theorem 1, "Given a discrete random variable Y , we have..." What is "accuracy(Y)"? Accuracy of a a discrete random variable? This looks weird and I suggest giving a more strict and accurate statement. What is pyp_y?

  4. "Predicting links by GNNs relies on measuring node similarity, which is incorrect if the neighbors have heterophily embeddings." The embedding will be similar even with heterophily connections if the nodes share similar neighborhood patterns [6].

  5. "the node embeddings of linear GNNs require no model training“ I don't understand this. Why linear GNNs don't require training? Do you mean SGC [7] don't require training?

  6. "It results in low value when node i and node j have heterophily embeddings, even if they are connected by an edge." This is not correct, see [2,3,6].

[1] Characterizing graph datasets for node classification: Beyond homophily-heterophily dichotomy. arXiv preprint arXiv:2209.06177.

[2] Revisiting heterophily for graph neural networks. Advances in neural information processing systems, 35, 1362-1375.

[3] When do graph neural networks help with node classification: Investigating the homophily principle on node distinguishability. arXiv preprint arXiv:2304.14274.

[4] Demystifying Structural Disparity in Graph Neural Networks: Can One Size Fit All?. arXiv preprint arXiv:2306.01323.

[5] Exploiting Neighbor Effect: Conv-Agnostic GNN Framework for Graphs With Heterophily. IEEE Transactions on Neural Networks and Learning Systems.

[6] Is Homophily a Necessity for Graph Neural Networks?. In International Conference on Learning Representations 2022.

[7] Simplifying graph convolutional networks. In International conference on machine learning, pp. 6861–6871. PMLR, 2019.

评论

We thank the reviewer for the comments, and apologize for a few unclear writings that make the reviewer misunderstand the idea that we tried to express. We sincerely hope that the reviewer can re-evaluate our paper and raise the rating after reviewing our answers and revised paper. We address the questions one by one.

W1. "How to connect the graph information with the performance metric on the task?" Here are some missed related works [1,2,3,4,5].

Thank you for your supplement on related works. We have added them to Section 2 (Page 2) in our revised paper. Noting that our focus has two major differences from [1, 2, 3, 4, 5]:

  1. These works focus on node classification, while our work majorly focuses on link prediction.
  2. These works focus on graphs that are heterophily defined by the node labels, while our work focuses on cases such as heterophily w.r.t. node features ("talkative people have silent friends"), as well as heterophily w.r.t. graph structure (forming bipartite cores). We did not consider node labels in link prediction.

W2. What is ff in C4?

As mentioned in the last sentence in C4 (Section 3.2, Page 3), ff is the column-wise L2 normalization. We replace it with ll in our revised paper for increased clarity.

W3. In theorem 1, "Given a discrete random variable Y , we have..." What is "accuracy(Y)"?

As shown in Equation 1 (Page 4), given a discrete random variable, its accuracy is the probability of its most probable outcome. This is achieved by majority voting, which is the best we can do without any other information. It is also described in Appendix A.1 (Page 13).

W4. "Predicting links by GNNs relies on measuring node similarity, which is incorrect if the neighbors have heterophily embeddings." The embedding will be similar even with heterophily connections if the nodes share similar neighborhood patterns.

The main point is that we use structure embeddings (C1, Section 3.2, Page 3).

We have clarified our wording as follows (Page 5, 4 lines from the top): “Predicting links by GNNs relies on measuring node similarity, which is incorrect if the neighbors are expected to have dissimilar embeddings; for example, in a bipartite graph, …” in our revised paper.

We agree with the reviewer that in the (label-)heterophily graphs, the propagated embeddings (C4 and C5) of connected nodes will be proximity embeddings, and are very likely to be similar. However, for a (structure-)heterophily graph (e.g., consisting of bipartite cores), our proposed structure embeddings (C1), will not be similar even if they share similar neighborhood patterns.

W5. "the node embeddings of linear GNNs require no model training“. Do you mean SGC [7] don't require training?

The node embeddings of linear GNNs in Equations 6 and 8 (Page 6 and 7, respectively) require no training, as they do not have any learnable parameters and can be computed with the closed-form formula. By “linear GNNs”, we mean GNNs that have linear transfer functions (instead of non-linear ones, e.g. ReLU) and do not have learnable weight matrices during message-passing.

In our revised paper, we clarified it with the sentence (Page 5, 2 lines from the top): “Compared to general GNNs, the node embeddings of linear GNNs are given by closed-form formula.”

W6. "It results in low value when node i and node j have heterophily embeddings, even if they are connected by an edge." This is not correct, see [2,3,6].

The statement in the paper is correct, when we use structure embeddings (C1, Section 3.2, Page 3) in a (structure-)heterophily graph.

In our revised paper, we clarified it with the sentence (Page 5, Section 4.1): “However, even if node ii and node jj are connected by an edge, it may result in low value if they are expected to have dissimilar embeddings (e.g. structure embeddings in a bipartite graph).”

In more detail, in the cited papers [2, 3, 6], these are node features propagated through graph structure, which are similar to our derived propagated embeddings (C4 and C5). As discussed in W4, the propagated embeddings of connected nodes can be similar in a (label-)heterophily graph, while the structure embeddings will not be similar in a (structure-)heterophily graph.

评论

Dear Reviewer ANgE,

Thank you again for your thoughtful review. Based on your suggestions, we added the citations into the related works and clarified our wording. Please let us know your thoughts or if there is anything else we can do to address your comments.

Best,

Authors

评论

Thanks for the response. The authors have addressed most of my concerns and I will raise my rating to 6. Remember to elaborate the statement and the proof of your theorem. Good luck.

评论

We truly appreciate your response and suggestion!

审稿意见
8

The paper proposes a framework called NETINFOF for measuring and exploiting network usable information (NUI) in node-attributed graphs. The authors aim to determine if a graph neural network (GNN) will perform well on a given task by assessing the information present in the graph structure and node features. NETINFOF consists of two components: NETINFOF PROBE, which measures NUI without model training, and NETINFOF ACT, which solves link prediction and node classification tasks. The framework offers several advantages, including generality, principled approach, effectiveness, and scalability. The authors demonstrate the superiority of NETINFOF in identifying NUI and its performance in real-world datasets compared to general GNN baselines.

优点

  1. The paper introduces the novel NETINFOF framework, which addresses the problem of measuring and exploiting network usable information in graph for GNNs.
  2. The paper demonstrates sound technical claims, supported by theoretical guarantees and empirical evaluations on synthetic and real-world datasets.
  3. The writing style is clear, and the paper is well-organized, making it easy to understand the proposed framework and its contributions.
  4. The paper's contributions are promising as they provide a practical tool (NETINFOF) for assessing the usefulness of graph structure and node features in GNN tasks. The framework shows promising results and outperforms existing baselines in link prediction.

缺点

  1. How does the NETINFOF framework handle noisy or incomplete graph data? Can it effectively measure and exploit network usable information in such scenarios?

  2. Are there specific types of graph structures or node features for which NETINFOF may not perform well? How robust is the framework in diverse graph settings?

  3. I have concerns regarding the sensitivity of NETINFOF to the estimated compatibility matrix (H) used in the framework. It would be beneficial if the authors could provide additional empirical results that examine the performance of NETINFOF under different label rates, as label rates can significantly impact the correctness of H.

问题

see weaknesses

评论

We thank the reviewer for the positive and detailed comments, and give us valuable suggestions. We address the questions one by one.

W1. How does the NETINFOF framework handle noisy or incomplete graph data? Can it effectively measure and exploit network usable information in such scenarios?

Yes, in such cases, NetInfoF will give low NetInfoF_Score to the corresponding noisy components in Section 3.2 (Page 3), and down-weigh them for the upcoming predictions.

Different graph scenarios are studied in Table 6 in Appendix C.1 (Page 16): For example, if the given graph has useful structure but very noisy features, NetInfoF gives high scores to the structural embeddings (C1 and C2), and a low score to the feature embedding (C3).

W2. Are there specific types of graph structures or node features for which NETINFOF may not perform well? How robust is the framework in diverse graph settings?

NetInfoF makes no assumption for either the graph structure or the node feature. In fact, as shown in Table 2 (Page 8) and Table 9 (Page 17), NetInfoF successfully tackles multiple graph scenarios on link prediction and node classification, respectively.

W3. I have concerns regarding the sensitivity of NETINFOF to the estimated compatibility matrix (H) used in the framework. It would be beneficial if the authors could provide additional empirical results that examine the performance of NETINFOF under different label rates.

Thank you for your valuable suggestion. For link prediction, even a small fraction of existing links (=”labels”) is enough for estimating the compatibility matrix, as we show in the following table. The table shows two medium size datasets “Computers” and “Actor”, where we vary the ratio of positive edges used for estimating the compatibility matrix. NetInfoF still works well even with 10% of the edges for the compatibility matrix estimation, and it also performs better than without using the compatibility matrix.

For large graphs, we do not expect NetInfoF to be sensitive to the ratio of edge labels, since it requires only a few edge labels to well estimate the compatibility matrix. For example, in “ogbn-Products”, there are 43M (= 62M * 70%) edges in the training set; as we mentioned in T3 in Section 4.1, we sample only 200K for estimating the compatibility matrix, which is 0.5% (= 200K / 43M) edges in the training set.

Edge Label Ratiow/o CM10%30%50%70%
Comp.27.9±0.227.9\pm0.229.4±2.729.4\pm2.733.0±1.833.0\pm1.831.7±3.431.7\pm3.431.1±1.931.1\pm1.9
Actor29.5±1.829.5\pm1.833.3±1.133.3\pm1.133.9±1.133.9\pm1.134.3±0.634.3\pm0.636.2±1.236.2\pm1.2
评论

Dear Reviewer WXaL,

Thank you again for your thoughtful review. Based on your suggestions, we conduct an additional experiment on the sensitivity of NetInfoF on the edge label ratio. Please let us know your thoughts or if there is anything else we can do to address your comments.

Best,

Authors

审稿意见
8

The paper presents a novel framework, NETINFOF, designed to quantify and leverage Network Usable Information (NUI) for graph tasks like link prediction and node classification. The approach tackles the need for extensive model training by using NETINFOF_PROBE to measure NUI directly from graph data, which is a significant improvement from traditional GNNs that rely on trained low-dimensional representations. The NETINFOF_ACT component then uses this measured information to enhance the performance of graph tasks. The framework's robustness is tested on synthetic datasets designed to include various graph scenarios and validated on real-world datasets, showing superior results in link prediction tasks over standard GNN models. This paper's primary contribution is a method that can quickly assess a graph's usefulness for a task, offering a theoretical lower bound for accuracy without the computational overhead of model training​. The empirical study of this paper is very strong, surpassing most GNN methods both in accuracy and scalability.

优点

  • Methodological Innovation: NetInfoF introduces a novel approach to measure NUI directly from the graph structure and node features, which is a different from traditional methods that rely heavily on model training. This innovation could have a broad impact, particularly in scenarios where computational efficiency is paramount​. Also, the adjustments of node similarity using compatibility matrix w/ negative edges is very interesting.
  • Teoretical Foundation and Empirical Validation: The paper provides a theoretical analysis for the NetInfoF_Score, presenting it as a lower bound for the accuracy of a GNN on a given task. This theoretical contribution is well-supported by empirical evidence on both synthetic datasets and real-life datasets.
  • Scalability: The demonstrated scalability of NetInfoF, especially its linear scaling with input size (Fig. 6), and the use of significantly fewer parameters compared to GAE methods (1280 vs 279k) demonstrates its potential for application in large-scale graph tasks, presenting a substantial advancement in the practical deployment of GNNs​. Though NetInfoF is slower than SlimG [1], NetInfoF has better accuracy.

Combining 2nd and 3rd points, NetInfoF is better than GAE methods both in accuracy and scalability empirically, which is a significant contribution in link prediction tasks.

[1] Yoo, Jaemin, Meng-Chieh Lee, Shubhranshu Shekhar, and Christos Faloutsos. "Slendergnn: Accurate, robust, and interpretable gnn, and the reasons for its success." arXiv preprint arXiv:2210.04081 (2022).

缺点

  1. The proof of Theorem 2 is problematic due to the incorrect use of conditional entropy. The proof incorrectly states the conditional entropy in Appendix A.2 as:
H(YX)=i=1mpi(j=1npijlog2pij),H(Y|X) = \sum_{i=1}^{m} p_i \left(-\sum_{j=1}^{n} p_{ij} \log_2 p_{ij}\right),

which is incorrect because it uses the joint probabilities pijp_{ij} instead of the conditional probabilities. The correct expression for conditional entropy, which is based on the conditional probabilities, is:

H(YX)=i=1mj=1npijlog2P(Y=yjX=xi),H(Y|X) = -\sum_{i=1}^{m} \sum_{j=1}^{n} p_{ij} \log_2 P(Y=y_j|X=x_i),

where P(Y=yjX=xi)=pijpiP(Y=y_j|X=x_i) = \frac{p_{ij}}{p_i} is the conditional probability of YY given X=xiX=x_i. And pi=j=1npijp_i = \sum_{j=1}^{n} p_{ij} represents the marginal probability of XX taking the value xix_i. This definition adheres to the fundamental property that the conditional probabilities for a given xix_i should sum to 1, i.e., j=1nP(Y=yjX=xi)=1\sum_{j=1}^{n} P(Y=y_j|X=x_i) = 1.

The misuse of conditional entropy in the proof leads to an erroneous application of Jensen's Inequality and subsequently invalidates the derived bound on the NetInfoF score.

  1. (Minor) Given the paper is mostly dealing with link prediction tasks. It will be beneficial to include some subgraph-based methods as baselines, such as SEAL [1]. I understand that NetInfoF is intended as an advancement beyond traditional GNN approaches, but including subgraph-based methods will give the community a more comprehensive evaluation of NetInfoF's capabilities. It's evident NetInfoF is way more scalable than subgraph-based methods, but subgraph-based methods are still SOTA on some datasets, e.g. Ogbl-collab.

[1] Zhang, Muhan, and Yixin Chen. "Link prediction based on graph neural networks." Advances in neural information processing systems 31 (2018).

问题

  1. Can the authors provide a more detailed justification for Theorem 2, particularly in light of the concerns highlighted in the first weakness? I will raise my score if the authors can give a reasonable update.
  2. How are the "Our bound" lines in Figures 3 and 4 derived? Additional details on this calculation would aid in understanding these figures.
  3. How NetInfoF_Probe estimates and calculates the discretizer in section 4.2? Maybe some examples or explanations will help reader understand this.
  4. An ablation study detailing the individual contributions of the five different node embedding components mentioned in Section 3.2 would be beneficial. Are these components equally influential in terms of accuracy, or do some weigh more significantly than others?
评论

W2. It will be beneficial to include some subgraph-based methods as baselines, such as SEAL.

We add the citations in Section 2 and run the experiments of SEAL with the default settings provided by OGB and set the hidden size to 128 as the same as other baselines, please find additional results in the following table.

The experiments are run by an AWS EC2 g4dn.metal instance, with 384GB RAM and Tesla T4 GPUs. Noting that SEAL runs out of memory (O.O.M.) in the preprocessing step for several graphs, and exceeds time limit (T.L.E.) on “Photo”, taking more than 11 hours to run one (out of five) split. Furthermore, as shown by Table 1 (Page 2), subgraph GNNs miss several desired properties focused by this study, i.e., SEAL is not scalable and can not be used on node classification, and NetInfoF is the only one that matches all properties. Due to the time constraints, we are not able to conduct hyperparameter search and will conduct a more rigorous experiment in the next version.

DatasetCoraCiteSeerPubMedComp.PhotoArXivProductsCham.SquirrelActorTwitchPokec
SEAL65.2±0.565.2\pm0.560.0±1.160.0\pm1.160.9±2.1\mathbf{60.9\pm2.1}O.O.M.\color{red}{O.O.M.}T.L.E.\color{red}{T.L.E.}O.O.M.\color{red}{O.O.M.}O.O.M.\color{red}{O.O.M.}87.6±0.9\mathbf{87.6\pm0.9}O.O.M.\color{red}{O.O.M.}43.8±0.8\mathbf{43.8\pm0.8}O.O.M.\color{red}{O.O.M.}O.O.M.\color{red}{O.O.M.}
NetInfoF81.3±0.6\mathbf{81.3\pm0.6}87.3±1.3\mathbf{87.3\pm1.3}59.7±1.159.7\pm1.131.1±1.9\mathbf{31.1\pm1.9}46.8±2.2\mathbf{46.8\pm2.2}39.2±1.8\mathbf{39.2\pm1.8}35.2±1.1\mathbf{35.2\pm1.1}86.9±2.386.9\pm2.324.2±2.0\mathbf{24.2\pm2.0}36.2±1.236.2\pm1.219.6±0.7\mathbf{19.6\pm0.7}31.3±0.5\mathbf{31.3\pm0.5}
评论

We thank the reviewer for the positive and detailed comments, and truly appreciate that you even check the proof carefully. We address the questions one by one.

W1 & Q1. Can the authors provide a more detailed justification for Theorem 2, particularly in light of the concerns highlighted in the first weakness?

Thank you for pointing out the problem, which is caused by the mistyping of the notations. We update the proof of Theorem 2 in Appendix A.2 (Page 13), where our conclusion remains the same.

Q2. How are the "Our bound" lines in Figures 3 and 4 derived?

We update the annotations in Figure 3 (Page 4) to make it more clear. In more detail, according to Theorem 2, the training accuracy can not be lower than NetInfoF_Score, and thus our bound is the dash x=yx=y line in Figure 3. We show that even for validation accuracy, the results still obey Theorem 2. In Figure 4, the fitting lines have high R2R^2 values, indicating that NetInfoF_Score is highly correlated with the test performance. Since it is difficult to connect a metric to the test performance, the ideal metric is expected to be highly correlated with the test performance.

Q3. How NetInfoF_Probe estimates and calculates the discretizer in section 4.2?

In Section 4.2, for link prediction, we did the quantile bucketization with k=8k=8 bins on the adjusted node similarity. Due to space limitation, the detailed algorithm was provided in Algorithm 2 in the appendix (Page 15).

Moreover, we conduct a sensitivity analysis on a medium size dataset “Computers" (see table below) and show that the result is insensitive to the exact value of kk, forming a plateau when kk increases. We put these results in the appendix (Table 8 and Figure 7, Page 16) in our revised paper.

Number of Binsk=4k=8k=16k=32k=64
C1: U74.2±0.374.2\pm0.379.7±0.279.7\pm0.280.4±0.380.4\pm0.380.8±0.380.8\pm0.381.0±0.381.0\pm0.3
C2: R76.5±0.276.5\pm0.280.4±0.080.4\pm0.081.2±0.381.2\pm0.381.4±0.381.4\pm0.381.5±0.381.5\pm0.3
C3: F63.8±0.163.8\pm0.164.9±0.264.9\pm0.265.0±0.165.0\pm0.165.0±0.165.0\pm0.165.1±0.165.1\pm0.1
C4: P77.3±0.277.3\pm0.282.1±0.182.1\pm0.183.0±0.383.0\pm0.383.3±0.383.3\pm0.383.4±0.383.4\pm0.3
C5: S77.3±0.177.3\pm0.182.5±0.182.5\pm0.183.3±0.183.3\pm0.183.6±0.283.6\pm0.283.7±0.283.7\pm0.2

Q4. An ablation study detailing the individual contributions of the five different node embedding components mentioned in Section 3.2 would be beneficial. Are these components equally influential in terms of accuracy, or do some weigh more significantly than others?

In short, different components have variable impact depending on the input graph. It is an advantage of NetInfoF that it can pinpoint which component is useful.

In more detail, we studied the information of individual components in Appendix C.1 (Page 16) on the synthetic datasets. In addition, we further conduct this experiment on real-world datasets, please find additional results in the following table (major component: in bold).

This conclusion is similar to the one we have in Appendix C.1 (Page 16), where different input graphs may have different preferences on the components depending on the underlying scenarios. Therefore, by combining all five components of node embeddings, NetInfoF is effective as well as robust. We put this table in the appendix (Table 7, Page 16) in our revised paper.

DatasetCoraCiteSeerPubMedComp.PhotoArXivProductsCham.SquirrelActorTwitchPokec
C1: U48.3±0.848.3\pm0.836.9±2.536.9\pm2.543.7±1.843.7\pm1.824.0±1.824.0\pm1.838.5±2.638.5\pm2.618.1±0.618.1\pm0.613.2±0.313.2\pm0.375.0±5.275.0\pm5.212.3±1.812.3\pm1.823.2±2.223.2\pm2.215.8±0.2\mathbf{15.8\pm0.2}16.2±0.6\mathbf{16.2\pm0.6}
C2: R61.5±1.261.5\pm1.250.0±1.950.0\pm1.947.9±0.847.9\pm0.819.4±0.819.4\pm0.836.8±3.336.8\pm3.312.3±1.012.3\pm1.009.4±0.709.4\pm0.764.6±3.264.6\pm3.208.2±1.408.2\pm1.434.2±2.3\mathbf{34.2\pm2.3}01.5±0.201.5\pm0.205.2±0.205.2\pm0.2
C3: F58.3±3.258.3\pm3.271.5±2.471.5\pm2.441.6±0.441.6\pm0.407.1±0.407.1\pm0.415.6±1.215.6\pm1.204.9±0.204.9\pm0.200.4±0.100.4\pm0.110.7±1.010.7\pm1.000.6±0.100.6\pm0.102.6±0.502.6\pm0.501.7±0.601.7\pm0.600.5±0.200.5\pm0.2
C4: P67.3±1.867.3\pm1.860.9±2.560.9\pm2.548.1±1.848.1\pm1.828.6±2.4\mathbf{28.6\pm2.4}42.4±1.4\mathbf{42.4\pm1.4}34.2±1.0\mathbf{34.2\pm1.0}27.6±0.527.6\pm0.584.3±1.8\mathbf{84.3\pm1.8}20.9±2.3\mathbf{20.9\pm2.3}13.1±1.113.1\pm1.107.3±1.007.3\pm1.012.4±0.812.4\pm0.8
C5: S82.4±1.0\mathbf{82.4\pm1.0}88.7±1.5\mathbf{88.7\pm1.5}63.3±1.1\mathbf{63.3\pm1.1}29.0±2.329.0\pm2.340.3±1.140.3\pm1.133.6±1.133.6\pm1.128.1±0.7\mathbf{28.1\pm0.7}80.5±2.780.5\pm2.719.2±1.219.2\pm1.222.3±0.922.3\pm0.902.7±1.802.7\pm1.816.2±0.6\mathbf{16.2\pm0.6}
评论

Dear Reviewer 96be,

Thank you again for your thoughtful review. Based on your suggestions, we corrected the proof of Theorem 2 and conducted an additional experiment on SEAL. Please let us know your thoughts or if there is anything else we can do to address your comments.

Best,

Authors

评论

Thank the authors for their detailed reply and experiments and sorry for my late reply. In general, I think the authors deal with my concern effectively. I have raised my score to 8 for now. And I am looking forward to discussing with other reviewers about the paper to see if I have missed any points.

W1 & Q1: Proof of theorem looks good to me now.

Q2: Thanks for the clarification.

Q3: It is interesting to see that NetInfoF_Probe is insensitive to kk.

Q4: The ablation study is great and reveals interesting patterns. Can the authors provide some insight about why certain components perform better on homophily/heterophily datasets?

W2: It’s great to see NetInfoF outperform SEAL on certain datasets and also turns out to be scalable on large datasets. It’s a great supplement to this work.

In general, I am satisfied with the current paper. And it seems that the authors have made significant efforts on responding to reviewers’ feedback.

评论

We truly appreciate your response and are grateful that you noticed our effort on responding!

Q4: …Can the authors provide some insight about why certain components perform better on homophily/heterophily datasets?

Thank you for your valuable suggestion, we plan to add the following explanations:

“For all homophily graphs (on the left of the vertical line in Table 7), the propagated embeddings (C4 and C5) are useful for link prediction; for non-homophily graphs (on the right of the vertical line in Table 7), it depends on specifics of the dataset.”

评论

We thank all reviewers for the insightful and constructive comments on our work. We have carefully addressed the comments and incorporated the changes into our revised paper, highlighted with the red font.

AC 元评审

This paper investigates an interesting problem, network usable information (NUI), which aims to measure the extent to which the graph structure and node features carry useful information for a link prediction or node classification task. The proposed method seems technically sound, and could be practically very useful. There were questions and concerns regarding the proofs and clarity. But the authors have successfully addressed most of them and all reviewers ended up being positive about this paper.

为何不给更高分

Overall, there is room to improve the writing and clarity of this paper.

为何不给更低分

The proposed method is shown to be effective and could be practically very useful.

最终决定

Accept (spotlight)