PaperHub
4.5
/10
Rejected4 位审稿人
最低3最高6标准差1.5
3
6
3
6
3.5
置信度
ICLR 2024

Long-distance Targeted Poisoning Attacks on Graph Neural Networks

OpenReviewPDF
提交: 2023-09-22更新: 2024-02-11

摘要

GNNs are vulnerable to targeted poisoning in which an attacker manipulates the graph to cause a target node to be mis-classified to a label chosen by the attacker. However, most existing targeted attacks inject or modify nodes within the target node's $k$-hop neighborhood to poison a $k$-layer GNN model. In this paper, we investigate the feasibility of {\em long-distance} attacks, i.e., attacks where the injected nodes lie outside the target node's $k$-hop neighborhood. We show such attacks are feasible by developing a bilevel optimization-based approach, inspired by meta-learning. While this principled approach can successfully attack small graphs, scaling it to large graphs requires significant memory and computation resources, and is thus impractical. Therefore, we develop a much less expensive, but approximate, heuristic-based approach that can attack much larger graphs, albeit with lower attack success rate. Our evaluation shows that long-distance targeted poisoning is effective and difficult to detect by existing GNN defense mechanisms. To the best of our knowledge, our work is the first to study long-distance targeted poisoning attacks.
关键词
Graph Neural NetworkAdversarial AttacksNode Classification

评审与讨论

审稿意见
3

This paper investigates targeted poisoning attacks on GNNs, in which an attacker injects nodes in a graph to cause a target node to be incorrectly classified to a label of the attacker’s choosing and considers that the attacked nodes are not within the targeted node’s k-neighborhood. The proposed attack is then evaluated and tested against some empirical defenses.

优点

+The studied problem is interesting

缺点

-Threat model is strong

  • Novelty is limited -Missing many references

问题

My major concern is that the problem has been extensively studied, and the novelty is not sufficient.

The authors claim that most of existing attacks on GNNs modify the target node’s k-hop neighborhood, but this is not accurate. For instance, most of the cited poisoning attacks focus on global structure attack, where the entire graph structure can be modified.

The threat model assumes that the attacker has access to the training data, (including the original graph G, node features, and labels, and also knows the training procedure), which is a rather strong assumption. There exist (restricted) black-box attacks to GNNs, while the authors do not compare and discuss with them

The evaluated empirical defenses are easy to be broken by stronger attacks, as demonstrated in [a]. Hence, it is not surprising that these defense cannot defend against the proposed attack.

[a] Felix Mujkanovic, Simon Geisler, Stephan Günnemann, and Aleksandar Bojchevski. Are defenses for graph neural networks robust? Advances in Neural Information Processing Systems 35 (NeurIPS2022), 2022.

In fact, there exist many certified defenses against graph structure attacks, but the authors do not test them against the proposed attack.

评论

Thank you for the valuable feedback. We address your concerns and questions below.

[Q.1.] The authors claim that most of existing attacks on GNNs modify the target node’s k-hop neighborhood, but this is not accurate. For instance, most of the cited poisoning attacks focus on global structure attack, where the entire graph structure can be modified.
[A.1.] Although existing poisoning attacks can modify any part of the entire graph structure, the results of their optimization end up exclusively changing the target’s k-hop neighborhood. Furthermore, existing attacks based on meta learning can not scale to large graphs and thus are not practical.


[Q.2.] There exist (restricted) black-box attacks to GNNs, while the authors do not compare and discuss with them
[A.2.] To the best of our knowledge, no existing black-box GNN attacks can be applied to our setting, i.e., long-distance node-injection targeted poisoning. While we do not discount the possibility that potential black box attacks for this setting could be developed, doing so would require significant further research. We will add a discussion about black box attacks to a revised version of our paper.


[Q.3.] There exist many certified defenses against graph structure attacks, but the authors do not test them against the proposed attack.
[A.3.] We did not evaluate certified defenses because we do not know of any existing implementations that can be applied to our setting. Many of the certified defenses for GNNs, such as Wang et. al KDD’21 [1] and Tao et. al WSDM’21 [2], cannot handle node injection attacks. The only work that handles node injection is by Tao et al [3], however, no public implementation is available.

[1] Wang, B., Jia, J., Cao, X., & Gong, N. Z. (2021, August). Certified robustness of graph neural networks against adversarial structural perturbation. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (pp. 1645-1653).
[2] Tao, S., Shen, H., Cao, Q., Hou, L., & Cheng, X. (2021, March). Adversarial immunization for certifiable robustness on graphs. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining (pp. 698-706).
[3] Tao, S., Cao, Q., Shen, H., Wu, Y., Hou, L., & Cheng, X. (2023). Graph adversarial immunization for certifiable robustness. IEEE Transactions on Knowledge and Data Engineering.


审稿意见
6

This paper discusses the vulnerability of Graph Neural Networks (GNNs) to targeted poisoning attacks, where an attacker manipulates the graph to misclassify a specific node. Most existing attacks focus on manipulating nodes within the node's neighborhood, but this paper explores "long-distance" attacks, where the manipulated nodes are outside this neighborhood. The paper presents a principled optimization-based approach for small graphs but also offers a more cost-effective heuristic-based approach for larger graphs. The findings indicate that long-distance targeted poisoning is effective and challenging to detect by existing GNN defense mechanisms.

优点

  • The paper is well-written, offering a clear and easily understandable presentation of the research.
  • The approach and contributions made by the paper are noteworthy, particularly the exploration of long-distance targeted poisoning attacks in GNNs, even though the proposed method is primarily heuristic in nature.

缺点

  • The MetaLDT method, while promising, appears to demand significant time and computational resources, which may limit its practicality for larger graphs.
  • The MimicLDT approach, while addressing the cost concerns, seems to compromise on the effectiveness of the attack. This trade-off between efficiency and success rate should be discussed better. Some aspects of the paper's approach require further clarification. Additional details and explanations could help the reader better understand the methodology and its intricacies, enhancing the overall quality of the paper.

问题

  • The definition of "long distance" and the specific distance of the injected malicious nodes remain unclear in the current version of the paper.

  • Have you taken into account the influence of robust methods during the poisoning process? If so, what are the results regarding the method's effectiveness when attackers lack knowledge of defense mechanisms, which is more practical in real-world scenarios?

  • The rationale behind why MimicLDT is more efficient is not clearly articulated. Further elaboration on this aspect would be beneficial. Is it possible to discuss the trade-off between efficiency and effectiveness, such as exploring adjustments to hyperparameters to strike a balance between MimicLDT and MetaLDT?

评论

Thank you for the detailed review and valuable feedback. We address your comments and questions one by one below.

[W.1.] This trade-off between efficiency and success rate should be discussed better.
[A.1.] We clarify the tradeoff in our answer to your Question.3. Thank you for the suggestion.


[Q.1.] The definition of "long distance" and the specific distance of the injected malicious nodes remain unclear in the current version of the paper.
[A.1.] We define long distance as following, for any attack point vav_a, the shortest path length from vav_a to the target node vtv_t should be larger than the number of layers used by the victim GNN model. In our experiments, we set k=3k=3. We will emphasize the definition more clearly in revision.


[Q.2.] Have you taken into account the influence of robust methods during the poisoning process? If so, what are the results regarding the method's effectiveness when attackers lack knowledge of defense mechanisms.
[A.2.] In the experiments reported in the paper, MetaLDT’s inner training loops and MimicLDT’s surrogate models account for what GNN defenses are used. In the table below, we include results in a setting where MimicLDT’s surrogate model does not account for GNN defenses, and we instead use vanilla GCN as the surrogate model.

SurrogateGCNGraphSAGEGATGNNGuardSoftMedianGDCJaccardGCNSVDGCNProGNN
CoraAware the defense0.670.630.600.700.550.660.740.59
Vanilla GCN0.670.630.590.680.290.650.350.53
arXivAware the defense0.740.730.700.640.590.630.620.58
Vanilla GCN0.740.720.670.640.420.610.490.55

We observe that for some defenses (e.g., SAGE, GNNGuard), not knowing about the defenses has little or no impact on poison success rate, but for others (e.g., SVDGCN and SoftMedian) the effects are more noticeable. Table 9 in the paper’s appendix evaluates MetaLDT under a similar scenario, i.e., in a case where its inner-training loop does not account for defenses, and leads to similar conclusions.


[Q.3.] The rationale behind why MimicLDT is more efficient is not clearly articulated. Further elaboration on this aspect would be beneficial. Is it possible to discuss the trade-off between efficiency and effectiveness, such as exploring adjustments to hyperparameters to strike a balance between MimicLDT and MetaLDT?
[A.3.] MimicLDT uses certain design heuristics to restrict the optimization to a much smaller search space than MetaLDT, and is thus more efficient. In particular, MimicLDT does not optimize for which attack points to use nor the graph structure among injected nodes. Furthermore, MimicLDT uses a surrogate optimization goal (trying to use injected nodes to cause an attack point and the target to collide in the embedding space) to determine injected node features. This is in contrast to MetaLDT whose optimization considers the entire space of attack node choices and injected node/graph structures and uses meta-learning (which requires very expensive inner training) to achieve direct end-to-end optimization of the label flipping probability.

In terms of the trade-off between MetaLDT and MimicLDT, we note that MetaLDT always has significantly higher computational cost than MimicLDT in order to achieve any decent poisoning rate, e.g., in Figure 5 of the Appendix, we show that, for Cora and GCN, MetaLDT requires a minimum of 8 inner training rounds and 9670.46s to reach the same poison success rate that MimicLDT can reach in about 43.17 seconds. On the other hand, MetaLDT’s larger search space means that it can often achieve higher poison success rates than MimicLDT can if time is not a concern.


评论

Thank you for clarifying my concerns in your rebuttal. I will maintain my original score.

审稿意见
3

This paper studies targeted poisoning attacks on graph neural networks, which aims to cause misclassification of a single victim node. In order to increase the stealthiness of the attack, the injected poisoning points do not belong to the top-k neighbors of the target victim. The edges and node features of the fake injected nodes are optimized through meta-learning for small graphs and through feature-collision for larger graphs. Empirically, the proposed attack performs better compared to existing short-distance attacks, when the manipulatable nodes are far from the target nodes.

优点

  1. The attack performance is good compared to other baselines when the attackers can only manipulate nodes that are far from the target victim.
  2. The approach of summarizing the attack patterns from expensive attacks (on small graphs) to design efficient attacks scalable for larger graphs is good.

缺点

  1. The motivation of considering nodes that are outside the top-k neighbors of the target victim is unclear. The authors argued in the appendix that, using some graph explanation tools, the attached nodes can be retried relatively well in some settings. However, the authors made an implicit assumption that such a tool can be directly treated as a detection method, without distinguishing the differences between the influential nodes for the target victim and other nodes. Can we use some threshold to filter out suspicious looking influence nodes for the node under examination? Will this filtering step can be evaded by some adaptive attacks so that short-distance attacks can still survive without sacrificing the effective much?
  2. From the technical perspective, I did not find a significant (inherent) difference from the previously proposed Meta-Attack, as the major the differences are on optimizing a different loss function to encode the targeted attack objective and also to avoid making connections with nodes of top-k neighbors of the target victim during optimization.

问题

  1. What is the value of kk to determine if an attack is short-distance or long distance.
  2. This is not a question, but rather a comment for the authors on proposing potentially stronger attacks. There is some interesting analogy between poisoning attacks on graphs and on images. The meta-learning approach is used to design poisoning attacks for both graphs (cited in the paper) and the images [1], the common drawback is the lack of scalability. The feature collusion attack is similar to the Shafahi et al.'s PoisonFrog paper cited in the paper. There might be some chance to rely on using the gradient alignment [2] technique to design stronger attacks in graph domains.

[1] Huang et al., "MetaPoison: Practical General-purpose Clean-label Data Poisoning", ICML 2019. [2] Geiping et al., "Witches' Brew: Industrial Scale Data Poisoning via Gradient Matching", ICLR 2021.

post-rebuttal: my concerns still remain and hence will maintain the current rating.

评论

Thank you for the valuable feedback and insightful comments. We address your concerns and questions below.

[W.1.] The motivation of considering nodes that are outside the top-k neighbors of the target victim is unclear. Can we use some threshold to filter out suspicious looking influence nodes for the node under examination? Will this filtering step can be evaded by some adaptive attacks so that short-distance attacks can still survive without sacrificing the effective much?
[A.1.] You made a good point that attackers could try to strengthen their attacks if they knew that GNN explanation tools are being used to detect short distance attacks. In addition to being less conspicuous, we think the bigger advantage of a long distance attack like MimicLDT is that it allows the attacker to choose from a much larger population of potential attack points. In particular, any node whose label is the same as the target poison label can function as an attack point, and there are many more of those at further distance away from the target node.


[W.2.] From the technical perspective, I did not find a significant (inherent) difference from the previously proposed Meta-Attack, as the major the differences are on optimizing a different loss function to encode the targeted attack objective and also to avoid making connections with nodes of top-k neighbors of the target victim during optimization.
[A.2.] Our main contribution is the practical long distance poisoning attack MimicLDT, whose design heuristics are derived from our observations of MetaLDT’s behavior. Compared to existing attacks, MimicLDT can scale to much larger graphs and is easier for the attacker to launch because it is flexible with which attack points to use. We design and evaluate MetaLDT to motivate the heuristics of MimicLDT and to quantify how much MimicLDT gives up in terms of poison effectiveness to its design heuristics.

We also want to clarify that the differences between MetaLDT and Meta-Attack extend beyond changes to the loss function, and require changing the optimization procedure. Specifically, MetaLDT (a) requires the optimization procedure to alternate between optimizing the adjacency matrix and optimizing node features; (b) uses a gradient descent based feature optimizer, rather than one based on meta-scores, thus allowing us to change the feature of all injected nodes in a single feature-optimization iteration (rather than requiring changes to a single node at a time). These changes to the optimization procedure are crucial to making MetaAttack work in our setting.


[Q.1.] What is the value of k to determine if an attack is short-distance or long distance.
[A.1.] kk is the number of GNN layers used by the victim model. A long distance attack is one in which injected nodes are further away from the target node than the number of GNN layers.


[Q.2.] There might be some chance to rely on using the gradient alignment technique to design stronger attacks in graph domains.
[A.2.] Thank you very much for pointing out the connection to gradient alignment. We agree that it might be feasible to apply gradient alignment for attacks on GNNs and will investigate this further.


评论

I appreciate the feedback from the authors. However, the concerns still remain. In particular: W1. The argument of many potential attack points for the attacker to choose is not a very well-motivated. Ideally, if the authors can show that a complete pipeline where the short distance nodes are infeasible to achieve attack success empirically (with concrete numbers), then the argument of using nodes outside top-k neighbors are strongly motivated. W2. Thanks for clarifying on the difference to the Meta-Attack details. Q1. Thanks for the clarification. Is there a reason to justify why such a way set the value of kk captures the threshold between short and long distance? Providing some explanation in the paper will be helpful.

评论

Thank you so much for the reply and follow-up questions.

For W1: Here is an example that short distance attacks are tough. We know that GNNs are often used to analyze social network or citation graphs, where adding nodes and links to the target is more challenging (e.g., requiring a malicious user to befriend a target) than adding links to a more distant node. Consider that you want to poison a user node on Facebook and you need to be a friend with the user. This is super hard.
Also, much of the prior work on explaining GNN behavior, and we thus expect that mislabeling due to short distance poisoning might be more easily detected by existing work.
In general, our work is motivated by observations about how the graphs that GNNs operate on are generated and how poisoning is investigated.

For Q1: We choose kk since nodes within kk-hop will directly contribute to the feature of the target node. In this way, our long-distance attack has a fundamental difference to short-distance attack.
As stated before, kk is the number of layers of the GNN structure. The feature of the target node is calculated with the nodes within kk-hop. As a result, people will put more effort on examining abnormal nodes within kk-hop. On contract, long-distance attack is inconspicuous.

评论

Thanks for the quick response. For W1, I think this is on a right path to provide better motivation for the problem. However, befriending with the target is a special case of k=1 and I think it might still be fairly easy to connect to the target as kk-hop neighbors. In the original comment, what I meant by a complete pipeline is to show that, with some detectors (e.g., graph explanation tools) with reported TPR and FPR, indeed can detect the short distance attacks. This way, it will make the paper more convincing. If all state-of-the-art short-distance attacks are detected in such a way, designing long distance attack will be very intriguing.

评论

Thanks again for all the insightful suggestions to make the paper more convincing.

We have provided the average recall (and precision) on detecting all the malicious edges attacker added for all baseline short-distance attacks in Appendix Table.13. This significant fraction of the perturbed edges found (i.e., detected >98% maliciously added edges for direct attacks and >74% for indirect, by default GNNexplainer settings) revealed the successful detection of the attacks.
To complete the pipeline further, we examine the poison success rate by removing these detected malicious edges and retraining the models for the target node’s prediction. Below for different short-distance attacks settings, we show the attack success rate before detection (i.e., Figure.8 in paper) and the ASR after the remove and retrain. The poison success rate decreased from 100% to 4%~7% for direct attacks, and decreased from 79% to 3% for indirect attacks, which shows the short-distance attacks are defended against.

AttacksSettingsAttack Success Rate (before detection)ASR (after remove and retrain)
NettackDirect; budget=341.000.04
NettackIndirect; budget=34, n_influencers=100.790.03
FGADirect; budget=340.990.04
IG-FGSMDirect; budget=341.000.07

Thanks so much for pointing this out! This reveals the disadvantages of short-distance attacks, and strengthen the motivation of our work. We are going to add this point to our motivation part.

评论

Thanks for providing additional results on the attack success rates. As mentioned in my first review, if the graph explanation tool is to be used as a complete detector, it should report the detection performance when facing perturbed graphs and also the corresponding benign graphs. A detector is only useful if it can work well against both perturbed and benign graphs. That is what I meant by a complete pipeline, and I should have made this clearer in the last comment.

评论

Seems like we misunderstood the previous comments. We think the pipeline you described involves two points: (p-1) the first one is detecting whether a graph is perturbed or benign; (p-2) the second one is detecting the perturbed part of the graph.
The detection tools mentioned in our paper is to find out the part of the graph that contributes to the prediction result, which is the second point (p-2). By the experiments in our last response, we show that a variety of short-distance attacks can be defended against by the GNNexplainer, while our attacks cannot. This already shows the importance of our long-distance attacks.

For the first point (p-1), we agree the question itself is interesting, but it is a question about defense methods. This is beyond the scope of our paper, which describes attack methods. Also, a more realistic situation is, the first point may be bypassed: regardless of what kind of attack is applied or whether there is an attack, some defenses will always be applied.

Below is not a formal response but some further discussion on the first point (p-1). As far as we know, there is no existing work targeting this. However, we have two potential methods in mind that might work for the first point: one is, using the current detection tool, comparing the distributions of “influence nodes” from perturbed graphs and benign graphs. We expect to see that the distribution of the perturbed graphs is abnormal. For example, some nodes/edges might have significant influence on the predicting result. The second one is that people can try to train a binary classification model.

审稿意见
6

This paper proposes and studies a new type of attack on GNNs that does not modify the target node’s k-hop neighborhood, which is called long-distance poisoning attack. To solve the problem, both a bilevel optimization-based approach inspired by meta-learning and an approximate heuristic-based approach are proposed. Extensive experiments are conducted on both small and large-scale graphs.

优点

  1. Exploring the attack performance of long-range targeted poisoning attacks is valuable and important.
  2. The proposed MimicLDT is well-motivated based on the observation from MetaLDT.
  3. The paper is easy-to-follow.

缺点

  1. The comparison with short-distance attacks is valuable. However, the compared baselines lack more recent node injection attack methods.
  2. Some claims lack further empirical or theoretical support. For example, the authors claim that 'there are many more potential attack points beyond the target’s K-hop neighborhood’. It would be better if authors could offer detailed support data analysis.
  3. Some minor errors: designe-> design heuristicsc)Finally -> heuristics. c) Finally

问题

  1. Whether the proposed method be generalized to unknown victim models?
  2. Is there any data analysis supporting the claim that 'there are many more potential attack points beyond the target’s K-hop neighborhood’?
评论

Thank you very much for the detailed review and valuable feedback. We will address each comment and question below.

[W.1.] The comparison with short-distance attacks is valuable. However, the compared baselines lack more recent node injection attack methods.
[A.1.] We evaluated the efficacy of AFGSM (Wang et al., DMKD’20)[1], a well cited recent node-injection poisoning attack. We used the author’s implementation, but modified it to perform targeted label-flipping poisoning instead of misclassifying the target to be any arbitrary class. The table below shows the Cora results of direct attacks (injected nodes can be direct neighbors of the target node) as well as indirect attacks (injected nodes cannot be direct neighbors but can be k-hop neighbors). The experiments use the same setting as that in our paper (Sec 6, Table 1).

GCNGraphSAGEGATGNNGuardSoftMedianGDCJaccardGCNSVDGCNProGNN
CoraDirect0.520.420.430.500.390.490.430.46
Indirect0.370.240.310.330.250.320.230.34

We can see that both the direct and indirect-attacks of AFGSM have a poison success rate ranging from 23–52%, which is lower than what we achieved (55-74% for MimicLDT, 53-96% for MetaLDT). We will add this comparison to a revised version of the paper.

We would also like to note that most recent injection attacks, including, TDGIA(Zou et al., KDD’21), GIA-HAO (Chen et al., ICLR’22), CANA(Tao et al., 2023), G2A2C(Ju et al., AAAI’23), are test-time evasion attacks, so we cannot directly compare against them.

[1] Jihong Wang, Minnan Luo, Fnu Suya, Jundong Li, Zijiang Yang, and Qinghua Zheng. 2020. Scalable attack on graph data by injecting vicious nodes. Data Min. Knowl. Discov. 34, 5 (Sep 2020), 1363–1389. https://doi.org/10.1007/s10618-020-00696-7


[W.2.] Some claims lack further empirical or theoretical support. For example, the authors claim that 'there are many more potential attack points beyond the target’s K-hop neighborhood’.
[A.2.] Please refer to our answer to Q.2.


[W.3.] Some minor errors: designe-> design heuristicsc)Finally -> heuristics. c) Finally
[A.3.] Thank you for pointing out the typos. We will fix these.


[Q.1.] Whether the proposed method be generalized to unknown victim models?
[A.1.] Our proposed attack assumes knowledge of the victim model. We do not believe it can be generalized to unknown victim models in theory. However, empirically, the situation is a bit more complex: we ran experiments where the attacker who uses a GCN as the surrogate model either did not know what model was being used (e.g. GAT) or was unaware of what defenses were in use, and measured the efficacy of MimicLDT. We show results in the table below (the rows corresponding to “vanilla GCN”):

SurrogateGCNGraphSAGEGATGNNGuardSoftMedianGDCJaccardGCNSVDGCNProGNN
CoraAware the defense0.670.630.600.700.550.660.740.59
Vanilla GCN0.670.630.590.680.290.650.350.53
arXivAware the defense0.740.730.700.640.590.630.620.58
Vanilla GCN0.740.720.670.640.420.610.490.55

Our results show that in many cases (e.g. GAT) using a vanilla GCN based surrogate model can poison as successfully as using a surrogate based on the correct victim model. However, we also see a significant drop in poison success rate in other cases (e.g., SoftMedian, SVDGCN).


[Q.2.] Is there any data analysis supporting the claim that 'there are many more potential attack points beyond the target’s K-hop neighborhood’?
[A.2.] We evaluated this empirically for the Cora dataset. Figure 1 of supplementary material shows the distribution of the distance of potential attack points from a target. We can see that about 95% of the attack points are more than 3-hops away (thus qualifying as long-distance attacks). In Figures 2 and 3, we show the attack points picked by MimicLDT across various datasets and MetaLDT across various models in our experiments are between 4 and 19 hops away.


评论

Thank the authors for responding and answering my questions. I will keep my current evaluation rating for this paper.

AC 元评审

The paper proposes an adversarial attack strategy on graphs, targeting distant nodes beyond the typical k-hop neighborhood. However, reviewers raised several critical concerns that ultimately led to a recommendation of rejection.

  1. The motivation of studying long-distance attack is questionable. It's unclear what benefit attackers would gain from prioritizing distant nodes over k-hop neighbors, and while they pointed out the potential of detectors for k-hop neighbors, the effectiveness of such mechanisms remains unproven. Despite having several round of discussions, the reviewer remains unconvinced by this motivation.

  2. Lack of novelty -- the attack is motivated by Meta-Attack. Although there are some modifications for targeting long distance nodes and some heuristics derived from the observations, the novelty on the method is still somehow limited.

为何不给更高分

Two main concerns are not fully addressed, as pointed out in the meta review.

为何不给更低分

N/A

最终决定

Reject