Low-Rank Robust Graph Contrastive Learning
We propose Low-Rank Robust Graph Contrastive Learning, which performs transductive node classification with a robust GCL encoder and a novel low-rank transductive algorithm inspired by sharp generalization bound for transductive classification.
摘要
评审与讨论
This paper proposes a robust graph contrastive learning framework based on prototype learning and low-rank decomposition. The key ideas are to explore the high-confident prototypical nodes for contrastive learning and utilize the low-rank structure of the kernel matrix for denoising. Bayesian nonparametric prototype learning is adopted to identify high-confident nodes. SVD decomposition is performed on the kernel matrix generated from the RGCL embeddings, based on which a linear classifier is trained. Experiments on real world datasets validate the effectiveness and robustness of the proposed method.
优点
- The idea of utilizing BPL and low-rank for robust graph ML is interesting.
- Theoretical analysis on the generalization bound is provided to assess the effectiveness of the proposed method.
- Experiments on real-world graphs seem solid. Source code and hyperparameter settings are provided for reproducibility check.
缺点
- More ablation studies needed. Since the model contains two modules (BPL based prototype learning and low-rank classifier), which part contributes more to the effectiveness and robustness?
- Presentation. Overall, I think further improvements on presentation are needed. The current writing spent too much space for less important preliminaries and related works, while too little space for the method (e.g., for DPMM, label propagation, BPL, etc.). See more details in the Question part.
问题
- For Eq. (2), can you explain why KL divergence is adopted here? Since is a one-hot vector, the calculation for KL divergence will be 0 for all the zero elements in .
- Current framework utilizes the BPL and prototype CL separately. Since both BPL and low-rank aims to deal with the noise/robustness, will it be possible, and potentially beneficial, to incorporate both BPL and low-rank property simultaneously?
- Theorem 4.1 provide a upper bound on the Frobenius distance between F(W,r,t) and . It would be beneficial to derive a distance between F(w,r,t) and to measure the difference between prediction and the noise-free label.
- Based on the definition of at the end of page 5, is there a lot of overlap between different ? According to my understanding, if , then ?
- From Figure 4, it seems that high-confident prototype nodes are actually quite far away from the cluster center. My concern is whether they can well represent the cluster? If not, the objective function encouraging node embeddings to be similar to the prototype nodes may push the clusters to the underrepresented prototype nodes.
- Presentation:
- Eq.(1): What is c_k in the nominator? Consider use another subscript to distinguish c_k in the numerator from c_k in the denominator.
- For defined at the end of Section 4.2, For nodes in , they are not necessarily connected (and even be distant), I’m wondering whether such mean aggregation is valid? E.g., what if the node features are positional embeddings? Such mean aggregation of distant and disconnected nodes will corrupt the information.
- Shrink the caption for Fig.1, and move some technical parts, e.g., how to compute p, etc, to the main paragraphs.
- Font sizes are too small for Table 1,4,5.
- Typo in abstract: Robst -> Robust; missing superscript for the definition of in section 4.3.
This paper tends to enhance the robustness of the GCL with Bayesian non-parametric prototype learning and low-rank transductive classification. The paper is based on prototypical contrastive learning. Both two components are from existing papers. The evaluations demonstrate its robustness to noises.
优点
- The robustness of graph contrastive learning is an interesting problem.
- The experiments basically demonstrate its robustness.
缺点
- The writing and organization are too poor to get the contribution. Most parts are seriously mixed, including preliminary, motivation, notations, and methods. For example, the second paragraph in section 4.3 should be placed in preliminaries.
- Many key component details are missing, such as the Bayesian non-parametric prototype learning.
- The citation is not correct, such as (Zhang & Chen, 2018) in section 4.2 is for link prediction.
- The novelty is very limited. Low-rank-based transductive classification is not novel, such as [1]. The Bayesian non-parametric prototype learning is also investigated.
- Some statement is not serious, including the problem description.
- Some sections are confusing. The title of section 4.1 is RGCL: ROBUST GRAPH CONTRASTIVE LEARNING WITH BAYESIAN NONPARAMETRIC PROTOTYPE LEARNING (BPL). However, the BAYESIAN NONPARAMETRIC PROTOTYPE LEARNING is in section 4.2.
- Figure 1 needs more explanations in the main body.
[1] Andrew B. Goldberg, Xiaojin Zhu, Ben Recht, Jun-Ming Xu, Robert D. Nowak: Transduction with Matrix Completion: Three Birds with One Stone. NIPS 2010: 757-765
问题
See weaknesses.
The paper studies the problem of graph classification with noisy features and labels. The proposed methods are compared with a few baselines.
优点
- The preliminary knowledge has been clearly explained.
- The experiments are sufficient, to some extent.
缺点
-
The noisy classification problem is not interesting since there have been many studies about the noisy classification of images and tabular datasets. The problem considered in the paper is not related to the graph structure. For graph data, the noisy adjacency matrix would be more worthy of study.
-
The paper is poorly organized. For instance, the connection between Section 4.3 and Sections 4.1 and 4.2 is not clear. It used four pages to introduce the background and related work but gave much fewer details about the proposed methods and experimental setting.
-
The low-rank technique is a very classical method for handling noise. Introducing low-rank factorization to the feature matrix H does not make a significant contribution.
-
The optimization details about the low-rank factorization is missing. The factorization is not a continuous operation and will break off the backpropagation.
问题
-
What is the connection between (2) and (1)?
-
What is the impact of in (3)? Is always nonnegative? According the analysis , when is larger, the bound is looser. But in practice, when is too small, the prediction will not be accurate.
-
The definition of noise level is missing.
伦理问题详情
n/a
The paper introduces a node classification method for noisy graphs (noise may be in the labels or the node attributes). The proposed LR-RGCL first learns robust node embeddings using Bayesian Prototype Learning for estimating clusters and using contrastive learning. It then uses the learned embeddings for low-rank transductive node classification. On comparing with competing baselines, LR-RGCL performs well on multiple benchmark graph datasets and across multiple noise levels.
优点
- The problem of noisy node classification addressed in this paper is an important one that has been studied in the community for a long time and has significant value.
- The proposed LR-RGCL relies on contrastive learning which is a well-founded field of study.
- The experiments included in the paper are extensive and they show that RL-RGCL consistent improves the noisy node classification performance compared to several baselines on multiple datasets across all noise levels.
- The overall presentation of the paper is good.
缺点
- The paper lacked a comprehensive section covering works related to graph contrastive learning.
- The idea of using clustering/cluster labels in GCL is not completely novel. See [1, 2] but I believe that it is still used differently in the proposed work. Even so, it would be nice to compare the proposed method with these, especially CGC[1].
- Similarly, the idea of using cluster labels as pseudo labels has been explored [3, 4].
- Clarification on a few things would be useful (see Questions).
All that being said, although the individual components are not necessarily novel, there is still some novelty and effectiveness (as shown by experiment results) in combining them in the way this paper does for the given task. The comment is to potentially include comparisons of the proposed work with these similar ideas.
[1] Park, Namyong, et al. "Cgc: Contrastive graph clustering for community detection and tracking." Proceedings of the ACM Web Conference 2022. 2022.
[2] Zhao, Han, et al. "Graph Debiased Contrastive Learning with Joint Representation Clustering." IJCAI. 2021.
[3] Yang, Fan, et al. "Class-aware contrastive semi-supervised learning." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
[4] Chakraborty, Souradip, Aritra Roy Gosthipaty, and Sayak Paul. "G-SimCLR: Self-supervised contrastive learning with guided projection via pseudo labelling." 2020 International Conference on Data Mining Workshops (ICDMW). IEEE, 2020.
问题
- No related works on GCL/prototypical CL?
- What is the effect of the estimated ? Does the clustering really need to be non-parametric? How does this compare with simple clustering methods like k-means? Any reason they were not included as baselines?
- For a particular epoch, the is the same irrespective of ? So how are the nodes assigned to clusters? Are only the nodes with considered for identifying the confident nodes within that cluster?
- How does the proposed method compare with CGC [1]?
- How does an ablation study perform based on vs ? What is the significance of each of the components?
[1] Park, Namyong, et al. "Cgc: Contrastive graph clustering forcommunity detection and tracking." Proceedings of the ACM Web Conference 2022. 2022.
伦理问题详情
N/A