Adversarial Robust Generalization of Graph Neural Networks
We establish an adversarial generalization bound of general GNNs via covering number analysis.
摘要
评审与讨论
The paper investigates the adversarial robustness of Graph Neural Networks (GNNs) in node classification tasks. The authors propose a high-probability generalization bound for GNNs under adversarial attacks using covering number analysis. They derive bounds for several popular GNN models (GCN, GCNII, APPNP) and analyze the impact of architectural factors on adversarial generalization. The paper also provides experimental results on benchmark datasets to validate the theoretical findings, showing that factors like model architecture, graph filters, and regularization parameters influence the generalization gap under adversarial attacks.
给作者的问题
While the paper presents an interesting theoretical framework for analyzing the adversarial robustness of GNNs, the empirical evaluation is insufficient to support the broad claims made by the authors. The lack of comparison with state-of-the-art methods and the limited scope of the experiments weaken the paper's contribution to the field. Additionally, the assumptions made in the theoretical analysis are not thoroughly discussed, and their practical implications are not explored in depth. For these reasons, I recommend rejecting the paper in its current form.
论据与证据
The paper makes several claims regarding the generalization bounds of GNNs under adversarial attacks. While the theoretical framework is well-structured, the evidence supporting these claims is not entirely convincing.
The experimental results, though consistent with the theoretical predictions, are limited in scope and do not fully validate the broad applicability of the proposed bounds.
The authors rely heavily on synthetic or controlled settings, and the generalization to real-world scenarios remains unclear.
Additionally, the paper lacks a thorough comparison with state-of-the-art adversarial training methods, which weakens the claim of providing a comprehensive understanding of adversarial robustness in GNNs.
方法与评估标准
The methods proposed in the paper, particularly the covering number analysis, are theoretically sound and appropriate for analyzing the adversarial robustness of GNNs. However, the evaluation criteria are somewhat limited. The experiments are conducted on standard benchmark datasets, but the adversarial attacks used (e.g., PGD) are relatively simple and do not cover the full spectrum of possible adversarial perturbations. The paper would benefit from evaluating the proposed bounds against more diverse and challenging attack scenarios, as well as comparing with other adversarial training techniques.
理论论述
The theoretical claims are based on covering number analysis, which is a well-established tool in statistical learning theory. The proofs provided in the appendix appear to be correct, but the paper lacks a detailed discussion of the assumptions made (e.g., Lipschitz continuity of the loss function and model architecture).
These assumptions may not hold in practice, especially for more complex GNN architectures or non-smooth loss functions. For instance, though Assumption4.1 maybe satisfied in standard non-graph neural networks, it could be violated in graph neural networks due to the message passing on the interdependent graph data.
The paper would benefit from a more thorough exploration of the limitations of these assumptions and their impact on the generalization bounds.
实验设计与分析
The experimental design is reasonable but lacks depth. The authors evaluate the generalization gap under adversarial attacks for different GNN models, but the experiments are limited to a few datasets and attack methods. The results, while consistent with the theoretical predictions, do not provide strong empirical evidence for the robustness of the proposed bounds. The paper would benefit from more extensive experiments, including comparisons with state-of-the-art adversarial training methods and evaluations on larger, more diverse datasets.
补充材料
N/A
与现有文献的关系
The paper is well-situated within the broader literature on adversarial robustness and GNNs. It builds on prior work in adversarial training and generalization analysis, particularly in the context of graph-structured data. However, the paper does not sufficiently highlight how its contributions advance the state-of-the-art.
While the theoretical bounds are novel, the practical implications and applications of these bounds are not clearly articulated. The paper would benefit from a more detailed discussion of how the proposed bounds compare to existing methods and what new insights they provide.
遗漏的重要参考文献
The referred papers (Szegedyetal.,2013; Goodfellowetal.,2014) are irrelevant to GNN applications
The paper misses many relevant papers on GNN attacks and defenses
其他优缺点
Strengths:
- The paper addresses an important and timely problem in the field of adversarial robustness for GNNs.
- The theoretical framework is well-structured and provides a solid foundation for analyzing the generalization properties of GNNs under adversarial attacks.
Weaknesses:
-
The empirical evaluation is limited in scope and does not fully validate the broad applicability of the proposed bounds.
-
The paper lacks a thorough comparison with state-of-the-art adversarial training methods, which weakens its claim of providing a comprehensive understanding of adversarial robustness in GNNs.
-
The assumptions made in the theoretical analysis (e.g., Lipschitz continuity) are not thoroughly discussed, and their practical implications are not explored in depth.
-
The paper does not clearly articulate how its contributions advance the state-of-the-art or provide new insights beyond existing work.
-
The theoretical results is only for node feature perturbation, while graph structure perturbation is more common against GNNs.
-
The paper misses many relevant work on GNN attacks and defenses
其他意见或建议
Whats the key challenges/difficulties of the derived theoretical results, compared with the existing theoretical results on non-graph data?
Can the proposed theoretical result be applied to graph structure perturbation?
What type of GNN architecture is suited to the derived theoretical results?
How to calculate the generalization gap in the evaluations? The paper uses many assumptions which makes me doubtful on calculating the values of variables in the theoretical gap.
We sincerely appreciate your thorough review and valuable suggestions. However, we would like to clarify the first misunderstanding below:
- Lack of new insights, comparison with existing methods, and application to real-world scenarios.
A1: Generally speaking, our focus does not lie in proposing a competitive algorithm tailored for a real-world application scenario (including real-world datasets and attack scenarios). Instead, this paper aims at a broader theoretical exploration of the robust overfitting in a general adversarial scenario. Our work not only develops a novel analytical framework for general GNNs (Theorem 4.8), but also provides helpful insights into model construction and algorithm designs (Proposition 4.14~4.18).
So, it is imperative for us to further clarify that " The lack of comparison with state-of-the-art methods and the limited scope of the experiments weaken the paper's contribution to the field" is not valid.
To be specific, this paper focuses on the robust overfitting phenomenon of GNN and provides theoretical guidance for improving their robust generalization in a general adversarial scenario. Based on our theoretical results, our empirical evaluation focuses on the influencing factors (some model architecture-related factors, like graph filter, weight norm, hyperparameters, etc.) and demonstrates their important roles in improving (or deteriorating) the adversarial generalization.
- Analytical challenges introduced by graph data.
A2: Challenge 1: The information interaction of nodes leads to the correlation of perturbations between different nodes, making the adversarial perturbation set of graph data different from that of non-graph data. Solvement 1: In adversarial settings, we search for the worst perturbation vector from all node features that consists of a perturbation matrix. Then, by incorporating the worst perturbation vector into coverage analysis, Lemma 4.6 reveals an additional term influencing generalization caused by the interaction between perturbed nodes.
Challenge 2: Each node in GNN aggregates messages from its neighbor nodes through the message-passing mechanism, making the complexity measure of GNN model class different from NN. Solvement 2: In the decomposition of the propagation process (Proposition 4.14~4.18), we pay attention to the information interaction of graph data in the propagation process, which is reflected in the graph filter .
- Lack of discussion of the assumptions made.
A3: Our analytical framework doesn't require smoothness assumptions of the loss function. This paper only needs the Lipschitz continuity assumption of loss function (Assumption 4.2) and activation function (Assumption 4.10), which can be easily satisfied by some commonly used functions (eg, cross-entropy and hinge loss; Sigmoid and ELU). Other assumptions about the norm constraint of input feature and weight matrix are also usually used in literature [1, 2].
In particular, for Assumption 4.1, we give the specific Lipschitz constant of each GNN model. For example, for a two-layer GCN, and , we have
.
Given the GCN with ELU activation and normalized filter (, ), conducting adversarial training with Algorithm 1 ( is controlled) could lead to a controllable and small Lipschitz constant.
- Irrelevance and misses of some referred papers.
A4: Thanks for pointing out the inaccuracy and misses. We will rectify the incorrect references (see [3,4]). As this paper focuses more on the theoretical analysis for adversarial training of GNNs, we will add additional related works about GNN attacks and defenses in the appendix for reference.
- Applicable GNN types
- Extension to structure perturbations.
A5/A6: Please refer to A2/A4 to the first reviewer (jE3b).
[1] Zhou, X. and Wang, H. The generalization error of graph convolutional networks may enlarge with more layers. Neurocomputing, 424:97–106, 2021.
[2] Tang, H. and Liu, Y. Towards understanding generalization of graph neural networks. ICML, pp. 33674–33719. PMLR, 2023.
[3] Ben, F., et al. Single-node attacks for fooling graph neural networks. Neurocomputing, 513:1–12, 2022.
[4] Liu, T., et al. Nt-gnn: Network traffic graph for 5g mobile iot android malware detection. Electronics, 12(4):789, 2023.
This paper investigates the generalization ability of graph neural networks (GNNs) under adversarial training, which is an important and widely interested research direction. The paper first proposes a high probability generalization limit and analyzes the generalization ability of GNN under adversarial training through covering number analysis. This provides theoretical support for understanding the behavior of GNN under adversarial learning. The paper selected three representative GNN variants for experiments and proposed a new adversarial training algorithm, which proved its effectiveness in improving the stability of GNN training through experiments.
给作者的问题
1.You established a high probability generalization limit for GNNs in adversarial learning in your paper. Can the derivation process of this boundary be further simplified? 2.You have proposed an adversarial training algorithm to learn a robust GNNs, but it seems that there is no clear description of the algorithm. 3. Why does the experimental part of this paper seem to lack a comparative study with previous work?
论据与证据
This paper has rigorous formula derivation, which can theoretically support the method proposed in this article. Clear logic in experimental procedures and methods.
方法与评估标准
The proposed methods and evaluation criteria mainly focus on analyzing the adversarial generalization ability of GNNs, which can be applied to the robustness research of GNNs and improving their generalization ability.
理论论述
This paper has rigorous and extensive formula reasoning, and in my reading, I did not see any obvious errors.
实验设计与分析
In the main text and appendix of the paper, the author conducted extensive experiments on six benchmark datasets to support the claims of this article. However, the description of Algorithm 1 seems somewhat vague.
补充材料
I did not review the supplementary materials.
与现有文献的关系
I'm not clear enough.
遗漏的重要参考文献
I'm not clear enough.
其他优缺点
Advantages 1.The paper has a clear structure, rigorous logic, and standardized use of symbols. 2.This paper establishes the high probability generalization limit of GNNs in adversarial learning, providing theoretical guidance for the design and training of GNNs. Weekness The derivation process of the coverage analysis method is relatively complex and may be difficult to understand and apply.
其他意见或建议
nothing
We deeply thank you for acknowledging the rigorous logic, clear structure, and extensive formula reasoning of our work. Below are our detailed responses.
- Why does the experimental part of this paper seem to lack a comparative study with previous work?
A1: Generally speaking, our focus does not lie in proposing a competitive algorithm tailored for a specific application scenario. Instead, this paper aims at a broader theoretical exploration of robust overfitting in a general adversarial scenario.
To be specific, this paper focuses on the robust overfitting phenomenon of GNNs and provides theoretical guidance for improving their robust generalization in a general adversarial scenario. Our work not only develops a novel analytical framework for general GNNs (Theorem 4.8), but also provides helpful insights into model construction and algorithm designs (Proposition 4.14~4.18). Based on our theoretical results, our empirical evaluation focuses on the influencing factors (some model architecture-related factors, like graph filter, weight norm, number of layers, hyperparameters, etc.) and demonstrates their important roles in improving (or deteriorating) the adversarial generalization.
- The description of Algorithm 1 seems somewhat vague.
A2: Thanks for your helpful suggestion. Let be a gradient-based attack algorithm (e.g., PGD, BIM, Mettack), the updated version is provided below.
Input: Graph , dataset , perturbed dataset , perturbation budget , regularization parameter , initialization , learning rate , number of iterations .
while do
.
for do
For the input matrix , perturb .
For each perturbed node in , append it to and choose samples randomly to the training set .
end for
Define a new objective .
For all , update using SGD: .
end while
- Can the derivation process of coverage analysis be simplified? It seems to be difficult to understand and apply.
A3: Covering number is a commonly used tool in (adversarial) generalization analysis [1,2]. Please let me briefly introduce our analysis technology first.
1. For the maximization over adversarial loss ( )(Lemma 4.4). We construct new function classes ( and ) and use their covering number to control the covering number of the adversarial loss class .
2. For the interplay between perturbed nodes (Lemma 4.6). We cover the perturbation set and transform the cover of the loss class to that of the perturbed model function class .
3 (Theorem 4.8). Now we can obtain the relation between the adversarial generalization and the covering number of GNN model class!
4. Covering number derivation (Proposition 4.14~4.18). We utilize the relation between the model function and its weight matrix to derive the covering number of the perturbed GNN class by that of the weight matrix set {}.
Actually, the main step that requires complex calculations is step 4, which is due to the propagation rules of GNN.
Therefore, given the Lipschitz continuity assumption about the adjacency matrix, our work can be applied to topology attack scenarios (by step 1-3). Since our covering number-based framework is applicable to general GNNs, given a specific GNN model with the relation between model function and weight matrix, it can be applied to other types of GNN models (by step 4).
[1] Tang, H. and Liu, Y. Towards understanding generalization of graph neural networks. ICML, pp. 33674–33719. PMLR, 2023.
[2] Tu, Z. et al. Theoretical Analysis of Adversarial Learning: A Minimax Approach. Neurips, 32, 2019.
This paper establishes an adversarial generalization bound of various GNNs, such as GCN, APPNP, GCNII, in the context of transductive learning. The authors provide some guidlines for adversarial generalization based on the theoretical results. The guidlines based on theoretical results are all validated in experimental results.
update after rebuttal
I had no specific concerns about this paper during my initial review. Therefore, I am maintaining my original score after the rebuttal.
给作者的问题
No
论据与证据
Yes, the logical flow of the paper is sound, and all claims may be well supported through both theoretical analysis and empirical evidence.
方法与评估标准
While the paper does not propose a specific method, it offers valuable hyperparameter guidelines for improving robustness across different GNN backbones. The evaluations are conducted on a variety of well-established datasets, yielding consistent and reasonable results. Additionally, the inclusion of diverse GNN backbones strengthens the validity and generalizability of the theoretical findings.
理论论述
I am not confident in assessing the correctness of the theoretical claims or proofs. I did not verify the detailed steps of the proofs, so I cannot confirm their validity. Please take this into consideration.
实验设计与分析
The theoretical analyses appear sound and comprehensive. Moreover, the empirical results are thoughtfully designed to support and validate the theoretical findings.
补充材料
No
与现有文献的关系
Contributing to Learning Theory of GNN adversarial generalization.
遗漏的重要参考文献
None
其他优缺点
- Important topic and solid theoretical analyses.
- Writing quality is very good.
- All theoretical findings are supported by the empirical results.
其他意见或建议
- Fig 2(a) caption might be "Experiments of adversarial training for APPNP."
We deeply appreciate your acknowledgment of the solid and comprehensive theoretical analysis and the thoughtfully designed empirical results presented in our paper. Thanks for pointing out the typo, and we will fix it in the future version.
The paper investigates the adversarial robust generalization of GNNs through a theoretical lens. It derives high-probability generalization bounds for general GNNs in adversarial settings using covering number analysis. The key insight is modeling the adversarial loss class’s complexity by constructing a perturbation cover and analyzing GNN architectures (e.g., GCN, APPNP, GCNII). Theoretical results reveal that adversarial generalization depends on factors like perturbation budget, model depth, weight norms, and graph filters. Experiments on benchmark datasets validate these findings, showing that normalized graph filters, shallower architectures, and regularization reduce generalization gaps.
给作者的问题
Please see above.
论据与证据
The claims are supported by theoretical proofs.
方法与评估标准
The methods and evaluation make sense.
理论论述
The proofs are logically sound.
实验设计与分析
Experiments vary layers, filters, and attack budget on standard datasets, but there are some weaknesses:
- The accuracy difference metric varies depending on the dataset split and does not appear to have been tested for a sufficient number of randomized splits.
- The datasets used are small and large-scale datasets are missing.
- Experimental validation was performed on only three GNNs.
补充材料
Appendix includes proofs of key lemmas (C.1–C.4), additional experiments (Figures 7–27), and setup details.
与现有文献的关系
The work extends adversarial generalization theory to GNNs, addressing unique challenges like transductive learning.
遗漏的重要参考文献
There are no essential related works that are not cited.
其他优缺点
There are some weaknesses:
- The work seems to discuss only the case of counter-attacks against node attributes. However, for graph learning, attacks against structures are more extensively studied.
- Can the proposed theoretical framework be adapted to other commonly used GNNs such as GAT, GraphSAGE, GIN, etc.?
- The experiments conducted were limited, as described earlier.
其他意见或建议
I don't have other comments or suggestions.
Thank you very much for your valuable comments! Please refer to our response below.
- Lack of testing for randomized splits of datasets.
A1: Thanks for pointing out the lack of considering the impact of dataset splitting. Taking two-layer GCN and two datasets for example, we show the generalization gap under different random split rates of training data (including 0.1, 0.3, and 0.5) in the table below.
| Cora | CoraFull | |||||
|---|---|---|---|---|---|---|
| 0.1 | 0.3 | 0.5 | 0.1 | 0.3 | 0.5 | |
| 0.4780.013 | 0.2960.012 | 0.2550.014 | 0.3960.003 | 0.2840.002 | 0.2080.003 | |
| 0.5160.014 | 0.3060.012 | 0.2610.012 | 0.4180.003 | 0.3010.002 | 0.2230.002 | |
| 0.5510.016 | 0.3090.014 | 0.2690.015 | 0.4340.002 | 0.3110.001 | 0.2340.002 |
The results show that they have consistent trends under increasing perturbation budgets. We will include a comprehensive version of the experiments in our future version.
- Adaptation to other GNNs.
A2: Our results provide a general analytical framework for GNN, and give three classical examples of spectral GNN. Although specific results are not presented in the paper, spectral GNN like SGC, AGCN and GPR-GNN are feasible.
Moreover, we are reasonably confident in extending our results to other types of GNN (spatial-based GNN). Taking a single-head GAT for example, as for the model function , and , we can get the relation between the covering number of the model function class and that of the weight matrix set (i.e. and ), which can be applied to our analytical framework.
- Large-scale datasets are missing.
A3: Thanks for your valuable suggestions. Large-scale datasets like Nell and ogbn-arxiv will be included in our future versions.
- Extension to attacks against structures.
A4: Given the similar adversarial settings for topology attacks and node attacks, we suggest that the methodology (e.g., Lemma 4.4 and Theorem 4.8) developed in this paper could be expanded upon the topology attack.
To be more specific, let the adversarial graph be generated from , where denotes the perturbation matrix added to the original adjacency matrix. The adversarial loss w.r.t. adversarial graph is defined by .
Analyzing analogously to the node attacks. For each function and a fixed , we construct a new function as . The adversarial loss is denoted by for any .
From the definition of covering number, we construct the cover of the class of function and obtain the cover of the class of function . Thus, the following inequality holds
Next, we construct a cover to control the infinite class by {}. Similarly, we can obtain the relation between and , which needs an assumption that
,
where the constant can be obtained if given specific GNN models.
This allows us to solve the measurement difficulty caused by the graph structure perturbations and apply it to our main results (Theorem 4.8). The remaining analysis will be left to future work.
Thank you for your response, it has addressed some of my concerns, but the limited experiments and the utility of the theory are still my concerns. I have raised my score accordingly but am leaning towards a borderline acceptance.
Thank you so much for increasing the score. We understand the limitations you mentioned and will take them as guidance for further improvement. Your recognition means a great deal to us.
The paper derive adversarial generalisation bounds for GNNs via covering number analysis. The topic is timely and important. Three out of four reviewers positively evaluated the paper with a "weak accept" and found the theoretical analyses to be solid. Issues regarding the comprehensiveness of the experimental validation were raised by several reviewers (small datasets, few models, limited scope). In my opinion this is not a major issues since the main contribution of the paper is theoretical. I disagree with Reviewer 5KXc that "thorough comparison with state-of-the-art adversarial training methods" is necessary, although it would of course strengthen the paper.
In general, I believe that the authors sufficiently addressed all the issues raised by Reviewer 5KXc in the rebuttal, even though the reviewer did not update their score. Specifically, the authors explained that their goal is to theoretically explain (adversarially) robust overfitting. The authors also articulated how the paper's contributions advance the state-of-the-art, including how they address the analytical challenges introduced by graph data. The authors also clarified the assumptions on the loss functions, which are standard and reasonable.
Nonetheless, I suggest the authors to carefully revise their writing to avoid that impression of "broad claims made by the authors". I also encourage the authors to update the paper to reflect the discussion, e.g. the comments regarding an extension to topology attacks. I recommend the paper to be accepted.