PaperHub
5.5
/10
Poster4 位审稿人
最低5最高6标准差0.5
5
6
6
5
3.8
置信度
正确性2.5
贡献度2.3
表达2.8
NeurIPS 2024

FedGMark: Certifiably Robust Watermarking for Federated Graph Learning

OpenReviewPDF
提交: 2024-05-13更新: 2024-11-06

摘要

关键词
WatermarkFederated Graph Learning

评审与讨论

审稿意见
5

This paper investigated the problem of watermarking the Federated Graph Learning (FGL) models. This paper proposed the first backdoor-based FGL watermarking framework, called FedGMark. Specifically, to tackle the issues of ineffectiveness and vulnerability of existing methods, FedGMark designed two modules respectively. One is a Customized Watermark Generator (CWG). CWG aimed to generate the watermarked trigger samples (graphs) using each client's secret key. The other is the Robust Model Loader (RML). RML guaranteed that the watermarked models were certifiably robust against layer perturbation attacks.

优点

  • The first attempt to watermark federated graph learning models.
  • The watermarked models are certifiably robust against attacks.
  • Experiments on various datasets and models validate the effectiveness of FedGMark.

缺点

My major concerns are as follows.

  1. Unclear threat model: The threat model and the problem formulation of this paper is unclear. What's the capability of the adversary and the defender? And more importantly, who is the adversary to steal the FGL model? This paper proposed to watermark the FGL model from the client side, which means the clients should be trustworthy. Is the central server an adversary in this paper? To my best knowledge, the typical threat model of various attacks in FL (e.g., backdoor attacks or Byzantine attacks) assumes that some of the clients may be malicious. The author should add a section on the threat model or problem formulation and clarify why they make these assumptions. This may be helpful to better understand the problem the authors tried to solve.
  2. Privacy concern: I also worry that utilizing FedGMark may raise privacy concerns. In Section 3.4, the watermarked client needs to use a subset of its training graphs as the watermarked graphs. However, in FL, the client's graphs are privacy-sensitive, and using them to verify ownership may lead to privacy leakage. This is contrary to the original purpose (preserve privacy) of FL.
  3. Missing experiments on the robustness against backdoor defense: This paper considers three different watermark removal attacks. However, since FedGMark utilizes backdoor-based watermarking methods, it is important to validate whether FedGMark is robust against backdoor defenses.
  4. Missing introduction to ownership verification: This paper lacks an important section to introduce the ownership verification procedure of FedGMark.

问题

  1. Clarify the threat model.
  2. Address the privacy concern.
  3. Analysis or experiments on the robustness against backdoor defenses.
  4. Clarify the procedure of ownership verification in FedGMark.

局限性

This paper does not include a discussion of the limitations. However, I think there is a strong assumption that the clients need to be trustworthy in FedGMark. A discussion on this assumption is necessitated.

作者回复

We thank the reviewer for appreciating the novelty of the studied problem and the proposed certified robust watermarking scheme against attacks.

W1: Clearly define the Threat Model and Problem; Who is the adversary to steal the FedGL model); Are clients and central server an adversary? Make assumptions clear

Thanks for the suggestion! See Response to Comment#1 in the global rebuttal.

W2: Privacy concerns raised by FedGMark

Response: FedGMark does not pose additional data privacy concerns. Like classic FL, all (watermarked) clients locally process their watermark data and train the model, and then submit the trained model to the server. Hence, the server cannot access the private data.

W3: Test backdoor defenses based attacks against FedGMark

See Response to Comment#4 in the global rebuttal.

W4: Introduce the "ownership verification" procedure of FedGMark

Thanks for the suggestion! See Response to Comment#2 in the global rebuttal.

评论

Thank you for the response. I still have two questions on W1 and W2.

  • About W1: I think the assumption that all the clients are benign may be too strong. It is acceptable to assume that the clients (including the adversary) follow the training protocol to get a well-trained model. However, during the ownership verification, some clients may be offline and cannot provide the watermark samples. It is also possible for malicious clients to provide fake watermark samples. The discussion of this issue may help improve the soundness of this work.
  • About W2: In practice, the model owner needs to send the watermark samples to a third-party judge or the adversary to calculate the outputs and verify the ownership. In this case, I think using a subset of the client's training graphs as the watermark samples may lead to privacy leakage. Existing client-side watermarking methods (e.g., [1]) tend to utilize noise-based watermark samples which are not privacy-sensitive. Can FedGMark address this issue?

[1] Watermarking in Secure Federated Learning: A Verification Framework Based on Client-Side Backdooring. TIST 2023.

评论

Thanks for the great comments! Below we provide more clarifications and justifications.

Response to Comment#1: Good point! We deem that our ownership verification is still robust against offline clients and malicious clients (if its number is less than 50%).

During ownership verification, each client provides its own watermark data to the trusted judge. When some clients are offline, the trusted judge can simply neglect them and only use participating clients’ watermark data for verification.

When facing malicious clients, their negative effect can be circumvented through a majority voting-based approach. Specifically, all clients provide their own watermark data to the judge and obtain the watermark accuracy per client. Though the watermark accuracy on malicious clients could be very low, the majority of benign clients can produce more number of high watermark accuracy, compared to the number of low accuracy. When the judge uses the majority-vote strategy, the final watermark accuracy is still high, ensuring the accurate ownership claim for benign clients.

Response to Comment#2: We first clarify that watermark data are not necessarily generated from the training/test samples. Remember the primary goal of backdoor-based watermarking is to force the model to memorize the relationship between the backdoor trigger (in the watermark samples) and the target label, while the samples to inject the trigger do not have constraints, i.e., they can be from training samples or artificially synthesized (which does not contain privacy information of any training/test data). For conveniences, existing methods inject backdoor triggers into the training/test samples (including Ref[1]). To validate this, we synthesize a set of random graphs using the popular Erdős–Rényi model (via the NetworkX toolbox) and the watermark samples are generated by injecting the learnt watermark on the synthesized graphs. Under the default setting, we test on Fed-GIN and show results below, where we observe WAs are very close to those shown in the paper on the four datasets.

WatermarkMUTAGPROTEINSDDCOLLAB
on train/test graphs0.81 / 0.900.72 / 0.860.73 / 0.650.73 / 0.75
on synthesized graphs0.80 / 0.880.71 / 0.840.72 / 0.640.72 / 0.73

Furthermore, since all clients intend to verify model ownership, it is reasonable to believe that these clients are willing to provide their watermark data—whether generated from private training/test data or non-private synthesized data—exclusively to a trusted judge, with informed consent and in accordance with legal and ethical standards. From this perspective, the data is confidential between each client and the trusted judge. We acknowledge it is very interesting future work to design a provably private mechanism for model ownership verification that the verifier cannot access the watermark data but can guarantee the correctness of verification.

评论

Dear Reviewer,

As the interaction period is drawing to a close, we would like to kindly inquire whether our rebuttal has satisfactorily addressed all of your comments. Please let us know if further clarifications are needed.

Best,

Authors

评论

Thank you for the extra response. It addresses my concern. I will raise my rating to 5.

评论

We appreciate the reviewer for raising the score! We are pleased that our additional response has addressed all your suggestive concerns. We assure you that all the results, clarifications, and discussions mentioned above will be included in the next version.

审稿意见
6

This manuscript introduces FedGMark, a backdoor-based watermarking method specifically designed to protect Federated Graph Learning (FedGL) models from illegal copying and model theft. They claim that the proposed FedGMark is the first method to safeguard the intellectual property of FedGL models, offering certified robustness against watermark removal attacks, leveraging unique graph structures and client information to create customized and diverse watermarks. Experiments demonstrate its effectiveness and robustness.

优点

The paper introduces FedGMark to address the overlooked vulnerability of FedGL model ownership and identifies three main challenges in current watermarking techniques: inapplicability to graph data, vulnerability to removal attacks, and lack of formal guarantees. The proposed method, including CWG and RML, is clear and intuitive, and the authors have provided comprehensive experiments to support their approach.

缺点

  1. I strongly recommend setting a "Threat Model" subsection to clarify the potential security threats to FedGL. In my opinion, since the authors consider watermark removal attacks like distillation and finetuning, FedGL operates under a white-box setting.
  2. The paper assumes attackers know the internal information of the target watermarked model, enabling distillation, finetuning, and layer-perturbation attacks. However, I find the white-box setting narrow and trivial. The authors should consider black-box attacks, which are more challenging and meaningful. Many studies on black-box attacks can be found.
  3. In watermarking-related literature, robustness and fidelity are more frequently used terms than watermark accuracy and task accuracy.
  4. In the "Inapplicable or Ineffective" item, the authors state, "For instance, they require input data to have the same size, while graphs can have varying sizes," which is not entirely accurate. For example, some Wavelet and DCT-based watermarking methods can be scalable.

问题

Please refer to Weaknesses part

局限性

Please refer to Weaknesses part

作者回复

We thank the reviewer for appreciating the intuition and motivation of the proposed solution and the comprehensive evaluations to support the solution.

W1: Clearly define the "Threat Model"

Thanks for the suggestion! See Response to Comment#1 in the global rebuttal.

W2: Consider black-box attacks

We emphasize this is a defense paper aiming to design a robust watermarking scheme for FedGL. As many lessons have learnt and to avoid the false sense of security [Carlini et al. 2019], an effective defense should be tested against the strongest white-box attacks. This is because a defense that is effective against (impractical) weaker attacks does not imply it is effective against stronger attacks. To this end, we assume the attacker knows all information of the target watermarked model, and this setting actually makes our defense design the most challenging. Moreover, if the defense can successfully defend against the strongest white-box attack on the watermarked FedGL model, it is naturally effective against any weaker attacks (thus including black-box attacks).

Nicholas Carlini et al., On evaluating adversarial robustness. arXiv, 2019.

W3: Replace watermark accuracy main task accuracy with robustness and fidelity.

Thank you for the suggestion. We will change the terms.

W4: Statement on “... while graphs can have varying sizes” is not entirely accurate

We agree that certain types of graphs (like wavelets) have the same size. However, our statement is for the general graph datasets whose graph sizes are varied.

评论

Dear Reviewer,

As the interaction period is drawing to a close, we would like to kindly inquire whether our rebuttal has satisfactorily addressed all of your comments. Please let us know if further clarifications are needed.

Best,

Authors

评论

Dear Reviewer,

We greatly appreciate your time and effort in reviewing our paper and providing constructive comments. We have dedicated significant efforts to addressing all of your feedback to the best of our abilities. As the interactive discussion period is drawing to a close, we are concerned about whether we have fully addressed your comments, as we have not received your feedback on our response. We would be grateful if you could confirm whether our responses meet your expectations or if more clarifications are needed within the limited discussion time.

Thank you once again for your valuable input.

Best,

Authors

审稿意见
6

This paper addresses the problem of protecting model ownership in the emerging domain of Federated Graph Learning (FedGL) by proposing FedGMark, a backdoor-based watermarking technique. The authors argue that existing watermarking approaches are either inapplicable to graph data or exhibit weaknesses in terms of robustness against removal attacks and lack of formal guarantees. FedGMark aims to overcome these limitations by leveraging graph structure and client information to learn customized watermarks, employing a novel graph learning (GL) architecture that enhances robustness, and providing certified robustness guarantees against layer-perturbation attacks.

优点

  • The paper clearly outlines the limitations of existing watermarking techniques and presents a well-motivated approach to address them. The design of FedGMark, with its CWG and RML modules, is tailored to the specific challenges of watermarking in FedGL.
  • FedGMark demonstrates promising empirical performance in terms of both main task accuracy and watermark accuracy. It outperforms the baseline approach (random graph-based watermarking) significantly, especially under watermark removal attacks.
  • The paper provides theoretical guarantees for the robustness of FedGMark against layer-perturbation attacks, a unique and valuable contribution in the watermarking literature.

缺点

  1. The reliance on pre-defined private keys for watermark generation may not be practical in all scenarios, and alternative key management methods should be explored.
  2. The assumption of limited attacker knowledge about the watermarked model may not hold in practice. Evaluating FedGMark against more knowledgeable adversaries would provide a more realistic assessment.
  3. The focus on FedAvg for model aggregation limits the exploration of other aggregation methods and their impact on watermark robustness.

问题

  1. Could you quantify the communication overhead of FedGMark during federated training, especially compared to random graph-based watermarking (in terms of local training time, size of watermarked data, etc.)?
  2. How do you envision FedGMark being deployed in a real-world FedGL system? What practical challenges might arise during implementation and watermark verification?
  3. How would the certified robustness guarantees be affected by more advanced watermark removal attacks beyond layer perturbation (e.g., those involving trigger reverse engineering)?
  4. How would the effectiveness of FedGMark be affected if the attacker had more knowledge about the watermarking process, such as access to the CWG architecture or the private key generation method?

局限性

  1. FedGMark's evaluation focuses solely on FedAvg for aggregating client models. The impact of alternative aggregation methods (e.g., those prioritizing clients based on data quality or model performance) on both watermark robustness and overall FedGL model performance remains unexplored.
  2. The paper acknowledges the increased computational cost of using more submodels (S) in RML but doesn't fully analyze the scalability of FedGMark. Further investigation is needed to understand how performance scales with different numbers of clients.
  3. FedGMark relies heavily on structural modifications of the graph as the watermark. The effectiveness and robustness of alternative trigger designs, such as feature-based triggers, hybrid triggers, or combinations of different trigger types, have not been explored.
  4. The paper lacks specific details about the hyperparameters used for training the GL models on the client-side. The impact of client training dynamics, particularly the choice of learning rate and the number of local epochs, on the watermarking performance and robustness of FedGMark remains unclear and requires further investigation.
作者回复

We thank the reviewer for appreciating the well-motivated approach, promising performance, and robustness guarantees.

W1: Alternative key management methods

We clarify the predefined key is used by the Watermark Generator to know which local watermark is learnt for which client. It is like an identifier of the client and only needed and known by the client’s watermark generator to generate the customized watermark. This is different from the role of the key management methods in crypto. To avoid confusion, we will not use the term “private key”, but “client ID”.

W2 (Q4): Against more knowledgeable adversaries.

As suggested, we test the attacker that has access to CWG to manipulate the FedGL training. We assume some clients are malicious and test two attacks. First, we consider a passive attack where all malicious clients DO NOT use CWG to generate customized local watermarks. Model (b) in Table 5 shows the maximum WA decrease is 9%, where all clients do not use CWG.

Second, we test an active attack where malicious clients modify their watermark data’s label to obfuscate the training. Specifically, all malicious clients’ watermark data are labeled (e.g., 2) differently from the target label (e.g., 1). The results below show MA/WA is marginally affected even with 20% malicious clients.

0%10%20%30%40%
MU.81/.90.80/.90.80/.89.80/.80.80/.71
PR.72/.86.72/.85.72/.76.71/.67.70/.67
DD.73/.65.72/.63.72/.58.71/.53.71/.52
CO.73/.75.74/.74.73/.68.73/.61.72/.60

W3 (L1): Other aggregation methods

We test M-Krum [Blanchard et al 18] and T-mean [Yin et al 19] that consider data quality (e.g., remove outlier clients). M-Krum filters a fraction pp of clients whose gradients largely deviated from others, while T-mean trims off a fraction qq of highest and lowest values for each parameter in clients’ models. We set p=10%p=10\% and q=10%q=10\% , and the MA/WA results are as follows. We see these aggregators achieve a robustness-utility tradeoff, and MA and WA are not largely different.

AvgM-KrumT-mean
MU.81/.90.78/.92.75/.93
PR.72/.86.73/.85.70/.87
DD.73/.65.72/.63.70/.65
CO.73/.75.73/.74.71/.77

Q1(L2): Scalability

See Response to Comment#3 in the global rebuttal.

Q2: (1) Deployment of FedGMark; (2) Challenges

(1) Deployment: Stage I: Server-client model for FedGMark training

  • Setup: Server and clients make an agreement on, e.g., the GL model, the aggregator, and server initializes the global model.

  • Local training: Clients download the global model, define their watermark data, locally train their watermarked GL model (with the CWG and RML module) using the global model, and upload the updated local model to the server;

  • Server aggregation: Server aggregates local watermarked models of selected clients to update the global watermarked model.

The final global watermarked model is shared by all clients for legal use.

Stage II: Ownership verification. See Response to Comment#2 in the global rebuttal.

(2) Challenges: We consider security and privacy threats, and data quality.

  • Security: How to guarantee all clients and the server are benign or detect / remove malicious ones? How to mitigate more advanced attacks?

  • Privacy: Though FL methods do not access the data, they may still leak data privacy in practice. How to provably protect the data privacy, while keeping the utility?

  • Data quality: Low quality data negatively affects the utility. This could reduce the interest for clients with high-quality data to participate in the FL. How to ensure all data have high quality?

Q3: Attacks use trigger reverse engineering

See Response to Comment#4 in the global rebuttal.

L3: Alternative triggers design

Response: We add some details to adjust FedGMark to the suggested alternative triggers.

To learn feature-based triggers, we first select a set of nodes from a graph as the target nodes, and learn the watermarked features for the target nodes. We use a graph Gi=(Vi,Ei,Xi)G^i = (V^i, E^i, X^i) from client ii for illustration, where XiX^i is the node feature matrix. We then define a feature-mask Mfi[vj]=1M_f^i[v_j]=1 if vjVwiv_j \in V_w^i and 0 otherwise, where VwiV_w^i is the watermark node set described in the paper. Then, we introduce a feature network (FeaNet) that learns watermarked node features as Xwi=FeaNet(Xi)MfiX_w^i = FeaNet(X^i) \odot M_f^i. The FeaNet takes input XiX^i and outputs a matrix having the same size as XiX^i, e.g., it has the same architecture as GatingNet but adjusts the input size. The corresponding watermarked graph is defined as Gwi=(Vi,Ei,Xwi)G_w^i=(V^i, E^i, X_w^i). By generating a set of watermarked graphs {GwiG_w^i} for client ii, we minimize the loss on client ii’s both clean graphs {Gci G_c^i} and {GwiG_w^i}. More details about training refer to Section 3.4.

Further, to learn feature-structure triggers, we combine FeaNet (that gets XwiX_w^i) with GatingNet/KeyNet (that gets EwiE_w^i), and the watermarked graphs are Gwi=(Vi,Ewi,Xwi)G_w^i=(V^i, E_w^i, X_w^i). We then minimize the loss on {GciG_c^i} and {GwiG_w^i}.

We evaluate these triggers and the results are follows. We observe that structure information alone is sufficient to enable designing effective triggers.

fsf-s
MU.81/.78.81/.90.79/.92
PR.72/.77.72/.86.73/.87
DD.72/.53.73/.65.74/.66
CO.73/.67.73/.75.72/.76

L4: Performance on hyperpara.

By default, we set the learning rate (lr) to 0.01 and #local epochs (le) to 5. The results with more lr and le are below:

(le=5)lr=.01.05.1(lr=.01)le=51020
MU.81/.90.84/.89.79/.90.81/.90.84/.9.76/.92
PR.72/.86.71/.79.71/.73.72/.86.72/.87.72/.89
DD.73/.65.71/.57.70/.53.73/.65.73/.66.73/.69
CO.73/.75.72/.71.70/.65.73/.75.74/.75.73/.77

We see that a large lr may reduce WA, and WA slightly increases as le grows, indicating more thorough training makes our method perform better.

评论

Thank you for addressing my concerns. I am satisfied with the authors' responses and explanations. I will raise my rating to 6 and look forward to seeing the final version of the paper.

评论

Thanks for raising the score! We are happy that our response has addressed all your concerns. We will promise to include all the above results, clarifications, and discussions in the next version.

审稿意见
5

This work studies watermarking for federated graph learning (FGL) to protect the ownership of participants. It proposes a customized watermark generator for local clients that can capture the local graph structure and private client information, and a robust model loader consisting of multiple GL submodels and a majority-voting-based ensemble classifier, which can defend against the proposed layer-perturbation attack.

优点

  1. This work claims to be the first to study watermarking for FGL models.

  2. The method can leverage local graph and client information to generate customized watermarks.

  3. The paper introduces a layer-perturbation attack to further demonstrate the certifiably robustness of the proposed backdoor-based watermarking for FGL.

  4. The work is well-motivated with preliminary studies.

缺点

  1. The concept of ownership in FGL can be confusing and is not well-defined in this paper. For example, can every client claim ownership of the federated trained model? Since the watermarks from different clients are different, can any single client claim entire ownership? Additionally, for clients who participate in the FL but do not have watermarks, how can they claim ownership?

  2. The motivation for using local customized watermarks is not clear. The following problems arise: (1) It is unclear how to conduct ownership verification. Should it use the global watermark or the local watermarks? (2) If using a global watermark, what is the necessity of employing customized watermarks, or what is the adequate way to aggregate the global watermark from customized watermarks? If using local watermarks, how can the customized watermarks be used across clients?

  3. The method requires specific GL models (to be split to multiple submodels), which can be hard to adapt to existing FGL methods, especially for advanced FGL methods.

  4. The motivation for incorporating submodels for GL is missing. Why is this design necessary?

  5. (1) What does “layer indexes” for splitting GL models mean? From section 3.3, it is not clear how the submodels are split and how the split submodels are decoupled from each other regarding cascaded structures. (2) Additionally, structural information can be important for graph learning. How would discarding such structural information impact in this setting?

  6. The global model is obtained by simply averaging uploaded clients’ models (not weighted by data size, or applying proxy terms for regularization). Can this method address the potential heterogeneity issue when local watermarks are highly disparate from each other?

  7. The proposed method can introduce efficiency issues, as it significantly increases the number of parameters and computation time.

问题

  1. When the set of selected clients for aggregation is different from the set of watermarked clients, can the method achieve stable convergence?

  2. Is the layer-perturbation attack applied before or after submodel splitting? If it is applied after, does it perturb all submodels or not?

  3. Out of curiosity, is it possible to federated learn the local watermarking? How do you expect this would perform?

局限性

Please see Weaknesses above.

作者回复

We thank the reviewer for appreciating the motivation and novelty of this work (first to study robust watermarking for FedGL models).

W1: Clarify the concept of ownership in FedGL

In typical FL, a server and multiple clients collaboratively train a global model stored in the server, which is used by all clients for their tasks. Accordingly, in our ownership verification problem in FedGL, all clients design their own watermark data and collaboratively train the watermarked global model, which is for joint ownership by all participating clients.

Further, since all clients have devoted computation and data to the training, they have a strong intention to jointly protect their ownership of the model. Hence, we do not consider the case where the clients did NOT participate in watermark training, but claim the ownership of the model (actually these clients do not know how to do so).

W2: (1) Global or local watermarks for ownership verification; (2) Motivation of developing local customized watermarks

(1) We use the global watermark (integration of local watermarks) for ownership verification, in line with the fact that the watermarked global model is collaboratively learnt from all clients with their local watermarks. Figure 6 in Appendix also shows global watermark is more effective than local watermarks.

(2) Learning local customized watermarks is to utilize the unique property of each client, as different clients could have different properties (e.g., distributions of their data) and their optimal watermark could be different. Also, the learnt customized watermarks enhance the ownership verification performance. For instance, Table 5 shows watermark accuracy improves 8% with local customized watermarks.

W3: Require split specific GL models to multiple submodels

Sorry for the confusion! In our design, we do not split the existing GL model into multiple submodels. Instead, we can use any GL model (e.g., a 3-layer GIN) as a submodel and each client’s GL model is an ensemble of a set of submodels (More details in Response to W5). Hence, our approach can be easily adapted to any FedGL, where it can use any aggregator or base GL model. For instance., our experiments conducted on Fed-GIN means we use the average aggregator and GIN as the submodel.

W4: Why incorporate submodels for GL

Our goal is to ensure the provable robustness of our watermarked FedGL model against layer-perturbation attacks. Recall that we propose a majority-voting based ensemble classifier on the submodels of our GL. Based on this, we can guarantee the ownership verification is provably accurate, when the #perturbed layers on GL models satisfies Eqn (1) in Thm 1.

W5: (1) What does “layer indexes” mean? (2) Structural information is important.

(1) A GL model contains multiple layers. E.g., a 8-layer GIN can be represented with layer indexes {l0,,l7l_0, \ldots, l_7}. Splitting this GIN into 4 submodels {GIN1,,GIN4}\{GIN_1, \cdots, GIN_4\} with layer indexes {l0,l1l_0, l_1} … {l6,l7l_6, l_7} means GINiGIN_i contains layers l2il_{2i}, l2i+1l_{2i+1} from the GIN. However, submodels splitted in this way are coupled from each other, making them unable to defend against layer-perturbation attacks. To tackle this problem, we design the novel GL model that is an ensemble of a set of independent submodels, where each submodel is a base GL model, e.g., GIN or GCN.

(2) Yes, it is. All our submodels are GL models that take the whole graph as input, and hence retrain the graph structure information.

W6: Can this method address the potential heterogeneity issue?

Good question! Per the Response to W2, local watermarks are learnt by considering the unique properties in each client. Such unique properties may include the heterogeneity across clients’ data. To validate this, we test our method with non-IID graphs across clients, where each client holds a single label data. The MA/WA results are below:

DatasetMUTAGPROTEINSDDCOLLAB
paper results0.81 / 0.900.72 / 0.860.73 / 0.650.73 / 0.75
non-IID0.80 / 0.890.72 / 0.830.72 / 0.630.72 / 0.75

We can see FedGMark also performs well with non-IID datasets. This implies the learnt customized watermarks indeed capture the heterogeneity of clients’ graphs.

W7: Efficiency issues

See Response to Comment#3 in the global rebuttal.

Q1: Selected clients for aggregation different from watermarked clients

In our experiments, the server randomly selects a fraction (e.g., 50%) of clients for aggregation in each training round. Also, all clients participating in the model ownership verification have watermarked data. Hence, there is no case where the set of clients selected for aggregation differs from the set of watermarked clients.

Q2: Layer-perturbation attack before or after submodel splitting?

It is before submodel splitting (ensemble).

Q3: Federally learn the local watermarking?

In our method, local watermarks are learnt using the global model (see Step 2 in Section 3.4), which is federally trained by all clients’ local models. From this point of view, local watermarks are also federally learnt by all clients.

评论

Dear Reviewer,

As the interaction period is drawing to a close, we would like to kindly inquire whether our rebuttal has satisfactorily addressed all of your comments. Please let us know if further clarifications are needed.

Best,

Authors

评论

Dear Reviewer,

We greatly appreciate your time and effort in reviewing our paper and providing constructive comments. We have dedicated significant efforts to addressing all of your feedback to the best of our abilities. As the interactive discussion period is drawing to a close, we are concerned about whether we have fully addressed your comments, as we have not received your feedback on our response. We would be grateful if you could confirm whether our responses meet your expectations or if more clarifications are needed within the limited time.

Thank you once again for your valuable input.

Best,

Authors

评论

Thank you to the authors for the clarification, which addresses my concerns. I will raise my rating to 5.

评论

We thank the reviewer for raising the score! We are happy that our response has addressed all your concerns. We will promise to include all the clarifications and discussions in the next version.

作者回复

We thank all reviewers for their constructive comments! We first summarize the global response to the common comments raised by the reviewers and then reply to individual reviewers’ comments.

Comment#1:Threat Model (djtK-W1 and yEMY-Q1)

Response: Thanks for the suggestion! We add more details about our motivation, assumption, threat model and problem definition.

Motivation, Adversary and Assumptions: Watermarking is designed to safeguard well-trained models against threats like illegal copying and unauthorized distribution. An adversary could be, e.g.,

  • a business competitor seeking to replicate a model to gain competitive advantages by significantly reducing development costs
  • a malicious user who sells the model for profits
  • a cybercriminal who uses the stolen model for malicious purposes such as conducting large-scale spam campaigns

In the paper, we follow existing methods [Shafieinejad et al., 21, Bansal et al., 22, Xu et al. 23, Jiang et al., 23], where the adversary is assumed to know all details of the pretrained watermarked FedGL model, but does NOT tamper with the training process. This means all clients and the server are benign and follow the federated training protocol, and the attack happens at the testing/inference time. We highlight this is in stark contrast to the training-time Byzantine attack on FL where some clients are malicious and they manipulate the training process.

Threat model:

  • Attacker’s knowledge: The attacker has white-box access to the pretrained watermarked FedGL model. In addition, the attacker may also know some clean (unlabeled or labeled) training data, as well as watermarked data.

  • Attacker’s capability: The attacker can modify the pretrained model via leveraging its white-box access to the trained model and its hold training and watermarked data. For instance, the attacker can finetune the pretrained model via the labeled training data. More details of the capabilities of considered attacks are described in Sec 2.3.

  • Attacker’s goal: To remove the watermark based on its knowledge and capability, while maintaining the model utility. This allows it to illegally use the model without detection.

Problem definition: As the defender, we focus on protecting the ownership of the trained FedGL model from the aforementioned security threats. In particular, we aim to build a certifiably robust watermarking scheme for FedGL against the worst-case layer-perturbation attack, such that the learnt watermarked FedGL model can achieve two goals:

  • High task accuracy (fidelity): predict correct labels as many as possible for clean test graphs

  • High (certified) watermark accuracy (robustness): (provably) predict the target label as many as possible for test graphs with the watermark

Comment#2: Ownership Verification (ZEKF-Q2 and yEMY-W4)

Response: As FedGMark uses backdoor-based watermarking, the ownership verification procedure of FedGMark is similar to that of the standard backdoor-based watermarking (Line 35-40). Specifically, when suspecting the target FedGMark model is illegally used by others, the model owner (all the participating clients or their representative) can recruit a trusted judge for model ownership verification. Typically, the judge requests both the true model owner and the illegal party to provide some test data for verification. Only when the one knows the predictions by the target model for the provided test data by both parties, the judge will confirm this party the model ownership. In particular, besides providing the clean data by both parties that behave normally, the true model owner especially provides the designed watermarked data that only he knows the model behaves on. As a result, both parties know the prediction results on clean data, but the illegal party is hard to predict accurately on the watermarked data provided by the true model owner.

Comment#3: Scalability of FedGMark (Y9Y7-W7 and ZEKF-Q1)

Response: Compared with graph-based or non-robust watermarking methods, the computation overhead of FedGMark is mainly from the introduced submodel models (fixing all the other parameters, such as #clients, #iterations, to be the same). Particularly, the overhead scales linearly with the number of submodels SS. Here we copy the runtime result from Table 9 in Appendix for reference, where S=1S=1 is the runtime of the existing method.

Runtime(s)MUTAGPROTEINSDDCOLLAB
S=10.130.722.9937.65
S=40.462.7911.14161.46
S=80.745.1020.12296.68
S=161.329.4936.74563.51

We believe the computation overhead is not an issue in practice, compared with the importance of designing a provably robust watermark that aids the ownership verification of the FedGL model. Note that as all clients have devoted computation and data to the training, they have a strong intention to jointly protect their ownership of the FedGL model.

Comment#4: More watermark removal attacks (ZEKF-Q3 and yEMY-W3)

Response: Many existing works [Wang et al 20, Zhang et al. 21] show the trigger-reverse based backdoor detection/removal is vulnerable to the ``stealthy” backdoor. This is because the effectiveness of trigger reverse attacks largely depends on the statistical differences between clean data and backdoored data. Since we do not notice any graph backdoor trigger-reverse attack, we instead propose to quantitatively test the structure similarity between the generated watermarked graphs and the clean graphs. Here we use the metrics NetSim and DeltaCon proposed in [Wills and Meyer’20], with the range [0,1] and the higher value the larger similarity. The results are below:

DatasetMUTAGPROTEINSDDCOLLAB
NetSim0.970.980.990.99
DeltaCon0.980.980.990.99

We observe the watermarked graphs and their clean counterparts are structurally very close. This implies that the proposed watermarks are hard to be detected or removed.

最终决定

The paper proposes a watermarking method for federated graph learning (FGL). The method relies on a local watermark generator at each client to capture the local graph structure and private client information; the paper also proposes a novel GL architecture to defend against watermark removal. The authors have addressed all the major concerns and I strongly encourage the authors to include them in the final version of the paper.