PaperHub
4.7
/10
Rejected3 位审稿人
最低3最高6标准差1.2
5
6
3
3.3
置信度
ICLR 2024

Rethinking Independent Cross-Entropy Loss For Graph-Structured Data

OpenReviewPDF
提交: 2023-09-16更新: 2024-02-11

摘要

关键词
Supervised LearningCross-Entropy LossNode ClassificationRobustness

评审与讨论

审稿意见
5

The paper proposes joint-cluster supervised learning for graph neural networks. Instead of adopting the cross-entropy loss for each node independently, the paper models the joint distribution of node and cluster labels, given their respective representations. Extensive experiments are conducted across multiple benchmarks, demonstrating the proposed loss can boost the performance of different backbone GNNs and robustness against adversarial attack.

优点

  1. The proposed joint modeling of node and cluster is novel and sound.

  2. The scope of the experimental evaluation is broad, including small graphs and large graphs.

  3. The empirical analyses are comprehensive in terms of necessary discussions, comparisons, and visualizations.

缺点

  1. The backbones adopted for experiments are mostly not those that perform the best on these benchmarks. It would be more convincing to see how the proposed loss boost the performance of strong GNN models, e.g., GCNII on Cora, that may give rise to sota performance.

  2. Similar concern to 1 also exists for analyses like Table 6 (with GCN) and Table 7 (with MLP).

[1] Chen et al. Simple and deep graph convolutional networks. In ICML.

问题

  1. Could more results with stronger backbones be provided on these datasets? e.g., GCNII on Cora.

  2. It is vague how this technique helps improve the best models that prevail on different tasks. For example, at least a comprehensive table which enumerates most recent or best performing methods on several datasets should be presented to give the readers an overview how this approach situate in the rich literatures in GNN-based node classification.

审稿意见
6

The paper studies the problem of discrepancy between the non-i.i.d. property of GNN and the MLE learning. It proposed a new loss function to address the problem and its performance is demonstrated by extensive experiments.

优点

  1. The studied problem is important.
  2. The idea of the proposed method is novel.
  3. The method is effective in comparison to the baseline cross-entropy loss.

缺点

  1. Some notations or definitions haven't been clearly explained. See the questions.
  2. The discussion about the connection between (5) and (4d) is missing. This makes it difficult to follow (5).

问题

  1. More explanation about the equivalance between (4c) and (4d) should be provided.
  2. In the definitions of zˉm\bar{z}_m and yˉm\bar{y}_m, there are two indices kk and ii that are confusing.
  3. In (5), yiyˉmy _i\bar{y} _m^\top is a matrix, which is not consistent with the shape of output of gϕg _\phi.
  4. What are the labeling rates for the datasets in Table 1? How does labeling rate influence the classification accuracy?
  5. How did the authors determine the hyperparameters of the compared methods?
  6. It is not clear why the improvement on balanced data is higher than imbalanced data.
评论

We really appreciate the reviewer for the agreements on our technical contributions and presentation. We would like to address the concerns one by one in the following.

W1 and W2: Thanks for your constructional suggestions. We will revise the paper according to the questions and clarify the connection between (5) and (4d) to facilitate readers' understanding of the relationship between maximum likelihood estimation and cross-entropy loss.

Q1: We will introduce the transition from (4c) to (4d). We aim to learn the parameters θ\theta to maximize the likelihood of the observed conditional distribution p(y1,,yLz1,,zL;θ)p\left(y_{1}, \ldots, y_{L} \mid z_{1}, \ldots, z_{L} ; \theta\right), known as maximum likelihood estimation. To simplify the computation of the likelihood function, it is common practice to use the logarithm of the likelihood function, which transforms the product of probabilities into a sum (taking the logarithm does not alter the optimal solution of the likelihood).

Q2: Thanks for your pointing out. We will fix this mistask and use ii as the indice uniformly.

Q3: We flatten the matrix yiymRC×Cy_i y_m^{\top} \in \mathbb{R}^{C\times C} into shape 1×C2{1\times C^2}, and the shape of output of gϕg_\phi is n×C2{n\times C^2}, where nn represents the number of nodes and cc represents the number of classes.

Q4: In Table 1, for citation networks(Cora, CiteSeer and PubMed), following the widely used standard split proposed by [1], we use 20 labeled nodes per class for training, 500 nodes for validation and 1000 nodes for testing. For DBLP and Facebook, we use 20 labeled nodes per class as the training set, 30 nodes per class for validation, and the rest for testing. To verify the the influence of labeling rate, we compare with vanilla cross-entropy (CE) loss on three citation networks. We randomly select five and ten labeled nodes per class as the training set, leaving the validation and test sets unchanged. Considering each backbone model (i.e., GCN, SGC and MLP), our joint-cluster (JC) loss consistently diliver a much higher node classification accuracy.

CoraGCN+CEGCN+JCSGC+CESGC+JCMLP+CEMLP+JC
571.80±2.4872.38±2.7372.61±2.5973.67±2.7244.37±3.5551.25±2.76
1078.08±1.1378.86±1.4579.34±0.7079.49±1.7353.66±1.5559.19±3.28
CiteSeerGCN+CEGCN+JCSGC+CESGC+JCMLP+CEMLP+JC
557.91±6.1860.98±5.0856.67±5.3261.34±6.5645.52±3.8346.84±5.00
1068.22±1.3669.92±1.2868.62±1.1369.88±1.1453.47±2.5255.85±2.22
PubMedGCN+CEGCN+JCSGC+CESGC+JCMLP+CEMLP+JC
570.36±4.7972.00±4.9670.77±3.7771.85±4.6563.45±2.2767.84±3.76
1074.94±2.6175.65±1.7175.38±2.8375.79±2.1767.28±1.4171.72±2.14

Q5: Following the experimental settings of original papers, for GAT, we choose the model parameters by utilizing an early stopping strategy with a patience of 100 epochs on classification cross-entropy loss. For other models(GCN, SGC, MLP and GrapgSAGE), we utilize the model parameters which perform best on the validation set for testing. The remaining hyper-parameters including learning rate, dropout and weight decay are tuned for different models.

Q6: In Table 2 of manuscript, we have already demonstrated the advantages on two extremely imbalanced datasets, LastFMAsia and ogbn-arxiv. We attribute this result to the referential ability of joint-cluster distribution modeling, which refers to a cluster of nodes when making decision on target node label during training and inference. The joint distribution weakens the over-confident prediction on the majority classes by assigning prediction confidence on other related minority classes at the cluster, and thus ameliorates the generalization performance. However, in the extremely imbalanced datasets, there are a large number of majority class nodes in the reference cluster of minority class nodes, making the reference information of nodes in the imbalanced datasets not as rich as in the balanced datasets. We think this leads to the improvement on balanced data is higher than imbalanced data.


Reference:

[1] Thomas N. Kipf and Max Welling. Semi-Supervised Classification with Graph Convolutional Networks. ICLR, 2017.

审稿意见
3

This paper argues that existing approaches for graph learning assume independent cross-entropy loss and ignores the inter-dependence induced by observed graph structures. In light of this, the authors propose a new objective that incorporates the inter-dependence among node points into the loss computation. This ensures that the loss for each node is dependent on other nodes during the training. Experiments on many benchmark datasets and using various GNNs as backbones verify the effectiveness of the new loss over the traditional loss function.

优点

  1. The paper is well written and easy to follow

  2. The proposed method seems reasonable and sound

  3. Experiments entail a lot of datasets and different GNNs as backbones

缺点

The major concern on this work is the potential over-claiming. The authors argued that existing approaches ignore the inter-dependence among node points for loss computation, which is incorrect. There are in fact quite a few existing works that already considered designing inter-dependent loss for graph learning tasks.

For example, [1] proposes a new objective based on conditional random field for node classification, and [2] harnesses label propagation as a re-weighted loss. Besides, there are also recent works proposing self-supervised loss that considers enforcing the consistency between connected nodes [3]. These approaches integrate the inter-dependence of nodes into the loss function for training.

Another weakness lies in the comparison in experiments. The current experiment only compares with the traditional cross-entropy loss, which is a very weak baseline. More comparison with other advanced methods, particularly the above-mentioned models are needed to well justify the efficacy of the new design.

[1] Meng Qu, et al., Neural Structured Prediction for Inductive Node Classification, ICLR 2022

[2] Hande Dong, et al., On the Equivalence of Decoupled Graph Convolution Network and Label Propagation, WWW 2021

[3] Hengrui Zhang, et al., Localized Contrastive Learning on Graphs

问题

See weakness above

评论

We really appreciate the reviewer for the agreements on our technical contributions and presentation. We would like to address the concerns one by one in the following.

W1: Appreciate for the pointing out of these related work. It should be clarified that we are not overemphasizing the contribution of our method, and we will revise the paper to make our contribution clearly understood. We agree that our method only has the same motivation with the above-mentioned models, but this does not overshadow the multifold contributions of this work. The motivation of node labels are dependent is a common sense in graph community (e.g. GNNs utilize message passing mechanism, label propagation[2] distributes labels along edge, CRF-based methods[1] focus on modeling the local label correlation of every linked node pair and contrastive methods[3] treat neighbors as positive samples), but it is more important in how we can incorporate this knowledge. Comparing with each mentioned work, we list the detailed differences below:

  • SPN[1] for inductive setting models the local label correlation of each linked node pair via nodeGNN and edgeGNN and takes all edges in whole graph as input to propagate all the pairwise label correlations along edges. In contrast, we train and infer the joint distribution of the target node only with one reference signal (i.e., cluster), which allows the batch training on large graphs (e.g., Amazon with millions of nodes).
  • PTA[2] analyzes the relationship between label propagation and decoupling GNN based on feature propagation. We will supplement this excellent work in the label propagation section of related work. We believe that this method still belongs to the label propagation paradigm, which utilizes label propagation to assign pseudo labels to unlabeled nodes and then performs training.
  • Local_GCL[3] proposes a graph contrastive learning method, which creates positive pairs from first-order neighbors. Different from self-supervised contrastive learning, which measures the feature similarity of a pair of data, instead, we extend to measure the joint label distribution of the target node and its cluster.

We introduce a new paradigm of joint-cluster supervised learning for graph data, which breaks the i.i.d assumption in independent cross-entropy loss computation and model the joint distribution between the target node and its located cluster. This perspective on viewing and handling label dependencies between nodes has never appeared before, and we believe it is novel enough.

W2: Thank you for your suggestion. We compare with each mentioned work and list the node classification accuracy on citation networks below. We follow the standard data split. Our method is orthogonal to label propagation and contrastive learning, and they can be effectively combined with our joint-cluster loss, such as using the pseudo-labels assigned by label propagation to make the cluster information richer or introducing node-cluster level contrastive objectives in the representation learning process. We intentionally keep our method simple to illustrate the effectiveness of joint-cluster loss and do not utilize unlabeled nodes during the training process, unlike PTA[2] which assigns pseudo labels to unlabeled nodes and Local_GCL[3] which constrains the representation learning process of unlabeled nodes. It can be found that our method achieves results second only to Local_GCL on small graph data, and delivers the best performance on larger graph arxiv. The SPN[1] requires the edges between labeled nodes for training due to its edge_GNN component. Under the transductive setting, the extremely small number of edges makes it difficult for SPN to model the information of node pairs. For arxiv, SPN cannot be trained on titan RTX24G GPU due to high memory requirements for training each node-pair, which reflects the advantage of utilizing cluster as the reference signal. In addition, our method demonstrates good scalability on millions of large-scale multi-class graphs (Yelp and Amazon), where multi-class tasks are a scenario not covered by the above mentioned works.

MethodsCoraCiteSeerPubMedArxiv
GCN81.7071.4379.0671.74
SPN [1]80.9069.9776.66OOM
PTA [2]83.5572.8979.5262.79
Local-GCL [3]84.573.682.171.34
GCN_JC83.5172.9779.8072.17

Reference:

[1] Meng Qu, et al., Neural Structured Prediction for Inductive Node Classification, ICLR 2022

[2] Hande Dong, et al., On the Equivalence of Decoupled Graph Convolution Network and Label Propagation, WWW 2021

[3] Hengrui Zhang, et al., Localized Contrastive Learning on Graphs

AC 元评审

Paper argues that the average cross entropy loss used for node classification on graphs using GNNs does not take into account the graph structure of the data. As an alternative, the paper proposes joint cluster which also tries to predict the cluster membership of the node in the graph. These clusters are computed using standard graph cluster algorithms. It is shown that the new loss function can improve performance and robustness of some GNNs on classification problems. While this is novel idea with promising results, paper does not provide enough comparison to state-of-the-art GNNs and other competing graph-level loss ideas.

为何不给更高分

Experiments do not compare to prior state of the art results.

为何不给更低分

N/A

最终决定

Reject