PaperHub
6.3
/10
Poster4 位审稿人
最低5最高7标准差0.8
5
6
7
7
4.3
置信度
正确性3.0
贡献度2.8
表达2.5
NeurIPS 2024

DropEdge not Foolproof: Effective Augmentation Method for Signed Graph Neural Networks

OpenReviewPDF
提交: 2024-05-02更新: 2025-01-10
TL;DR

A novel signed graph augmentation method

摘要

Signed graphs can model friendly or antagonistic relations where edges are annotated with a positive or negative sign. The main downstream task in signed graph analysis is $link sign prediction$. Signed Graph Neural Networks (SGNNs) have been widely used for signed graph representation learning. While significant progress has been made in SGNNs research, two issues (i.e., graph sparsity and unbalanced triangles) persist in the current SGNN models. We aim to alleviate these issues through data augmentation ($DA$) techniques which have demonstrated effectiveness in improving the performance of graph neural networks. However, most graph augmentation methods are primarily aimed at graph-level and node-level tasks (e.g., graph classification and node classification) and cannot be directly applied to signed graphs due to the lack of side information (e.g., node features and label information) in available real-world signed graph datasets. Random $DropEdge $is one of the few $DA$ methods that can be directly used for signed graph data augmentation, but its effectiveness is still unknown. In this paper, we first provide the generalization bound for the SGNN model and demonstrate from both experimental and theoretical perspectives that the random $DropEdge$ cannot improve the performance of link sign prediction. Therefore, we propose a novel signed graph augmentation method, $\underline{S}$igned $\underline{G}$raph $\underline{A}$ugmentation framework (SGA). Specifically, SGA first integrates a structure augmentation module to detect candidate edges solely based on network information. Furthermore, SGA incorporates a novel strategy to select beneficial candidates. Finally, SGA introduces a novel data augmentation perspective to enhance the training process of SGNNs. Experiment results on six real-world datasets demonstrate that SGA effectively boosts the performance of diverse SGNN models, achieving improvements of up to 32.3% in F1-micro for SGCN on the Slashdot dataset in the link sign prediction task.
关键词
Graph Neural NetworksSigned Graph Neural NetworksGraph Data AugmentationStabilityGeneralization Guarantees

评审与讨论

审稿意见
5

The paper presents a novel Signed Graph Augmentation (SGA) framework designed to enhance the performance of Signed Graph Neural Networks (SGNNs). The primary focus is on addressing two persistent issues in SGNNs: graph sparsity and unbalanced triangles. The authors demonstrate that the commonly used DropEdge method is ineffective for signed graph augmentation. Instead, they propose SGA, which integrates structure augmentation, candidate edge selection, and a new data augmentation perspective to improve training. Experiments on six real-world datasets show significant performance improvements in link sign prediction tasks.

优点

Originality: The paper introduces a new augmentation framework specifically designed for SGNNs, which addresses the unique challenges of signed graphs.

Quality: The methodology is well-validated through extensive experiments on multiple real-world datasets, demonstrating substantial improvements in performance metrics.

Clarity: The paper is well-organized, with clear explanations of the proposed methods and thorough discussions of experimental results.

Significance: This work provides valuable insights and tools for improving SGNNs, which are crucial for tasks like link sign prediction in social networks.

缺点

Generalization: The effectiveness of SGA on other types of signed graph tasks, such as node classification or community detection, has not been explored.

Complexity: The proposed framework involves multiple steps and parameters, which may complicate its implementation and tuning.

Resource Intensive: The computational requirements for the proposed method may be high, potentially limiting its practical application in resource-constrained environments.

问题

  • Can the proposed SGA framework be adapted for other signed graph tasks, such as node classification or community detection?

  • How does the performance of SGA scale with larger datasets or more complex network structures?

  • What are the computational overheads associated with the different components of the SGA framework, and are there ways to optimize them?

局限性

The authors acknowledge that their data augmentation method is based on balance theory, which may not be applicable to all real-world signed graph datasets. Additionally, the method has only been validated on link sign prediction tasks, and its effectiveness on other tasks remains untested. To improve, the authors could explore the generalization of SGA to other tasks and consider alternative theoretical frameworks for signed graph augmentation. Potential societal impacts, especially in contexts involving negative social interactions, should also be addressed.

作者回复

For Weakness 1 and Q1 on more downstream tasks:

Thanks for your constructive comments. Existing SGNN methods focus primarily on the link sign prediction task, and they ignore the performance on other tasks. We experiment with the performance (metric: Average Accuracy± standard deviation) of SGA on different baselines in node classification and community detection tasks. The datasets were sourced from [16] .

  1. Node Classification
modelrainfallsamspon
SGCN0.72 ± 0.050.76 ± 0.17
SGCN+SGA0.75 ± 0.030.72 ± 0.18
improv.4.2% \uparrow-5.3% \downarrow
GSGNN0.72 ± 0.060.44 ± 0.30
GSGNN+SGA0.68 ± 0.020.44 ± 0.26
improv.-5.6% \downarrow0% -
SiGAT0.64 ± 0.230.60 ± 0.20
SiGAT+SGA0.65 ± 0.130.68 ± 0.18
improv.1.2% \uparrow13.3% \uparrow
  1. Community Detection
modelrainfallppi
SGCN0.49 ± 0.150.26 ± 0.09
SGCN+SGA0.60 ± 0.040.26 ± 0.09
improv.22.4% \uparrow0% -
GSGNN0.39 ± 0.020.36 ± 0.01
GSGNN+SGA0.40 ± 0.010.33 ± 0.06
improv.2.6% \uparrow-8.3% \downarrow
SiGAT0.53 ± 0.150.14 ± 0.01
SiGAT+SGA0.60 ± 0.080.16 ± 0.02
improv.13.2% \uparrow14.3% \uparrow

Based on the experimental results, the SGA shows minimal improvement for the node classification task, but it significantly enhanced the performance for the community detection task.

For Weakness 2:

Please see Common Concern.

For Weakness 3 and Q3 on computational overheads:

  1. Model parameter statistics |Model|#params| |:-:|:-:| | SGCN |14851| |GSGNN |30580| |SiGAT |185600| |SGCN+SGA|29702| |GSGNN+SGA|45431| |SiGAT+SGA|200451|

  2. model training time cost BitcoinOTC: |Model|time(s)| |:-:|:-:| |SGCN|75.4 ± 6.4| |GSGNN|142.8 ± 17.2| |SiGAT|277.0 ± 30.7| |SGCN+SGA|149.6 ± 42.7| |GSGNN+SGA|5217 ± 48.0| |SiGAT+SGA|351.2 ± 41.2|

The method proposed maintains a relatively stable number of additional parameters and training time, even for large models, avoiding increased computational complexity. Training is efficient and can be completed quickly. The additional operations required by the method can be optimized using parallelization, divide-and-conquer strategies, and feature reduction algorithms. More detailed statistical results are available in the referenced PDF.

For Q2 on the performance on large datasets:

We tested the performance of SGA on the Amazon-CD dataset [17]. The dataset contains 895,266 edges and 97,731 nodes, making it one of the larger signed graph datasets we could find.

AUCF1-BinaryF1-MicroF1-Macro
SGCN61.65 ± 0.1469.11 ± 1.4158.56 ± 1.1453.04 ± 0.69
SGCN+SGA58.26 ± 0.3783.08 ± 1.4772.47 ± 1.6557.44 ± 0.39
SIGAT55.08 ± 1.4589.96 ± 0.2981.76 ± 0.4845.02 ± 0.48
SIGAT+SGA64.58 ± 0.2989.51 ± 0.3281.83 ± 0.2053.28 ± 0.17

Based on the experimental results, our method improves the performance of the baselines on the Amazon-cd dataset.

For Limitation on alternative theoretical framework:

Regarding the analysis of SGA's generalization, we have conducted the following brief discussion.

SGA prevents the formation of unbalanced triangles in the input graph while selecting beneficial training samples. This stabilization of the graph structure preserves the eigenvalue distribution, thereby enhancing the quality of the graph embedding. Let LL be the Laplacian matrix with unbalanced triangles and λi\lambda_i its eigenvalues. After eliminating unbalanced triangles, the Laplacian matrix and eigenvalues become LL' and λi\lambda'_i, with ΔL\Delta L representing the matrix perturbation. The corresponding change in eigenvalues can be expressed as

Δλi=λiλi=viTΔLvi,\Delta\lambda_i=\lambda_i^{'}-\lambda_i=\mathbf{v}_i^T\Delta L\mathbf{v}_i,

Where viv_i the eigenvector of LL, ZZ the original embedding vector, and ZZ' the embedding vector after eliminating unbalanced triangles. Removing unbalanced triangles reduces the absolute values of the Laplacian matrix's non-diagonal elements, decreasing the perturbation matrix ΔL\Delta L. Consequently, the change in eigenvalues λi\lambda_i decreases, leading to Var(λ)>Var(λ)\mathrm{Var}(\lambda) > \mathrm{Var}(\lambda'). By Spectral Graph Theory, this implies Var(Z)>Var(Z)\mathrm{Var}(Z) > \mathrm{Var}(Z').

Let HH be the original embedding space and HH^{'} be the embedding space after variance reduction. Since the variance of the embedding vectors becomes smaller, the distribution range of the embedding vectors becomes smaller. We have:ZiZˉ2>ZiZˉ2.\|\|\mathbf{Z}_i-\bar{\mathbf{Z}}\|\|^2>\|\|\mathbf{Z}_i^{'}-\bar{\mathbf{Z}}^{'}\|\|^2.

Thus, coverage number N(ϵ,H,)N(\epsilon,\mathrm{H}^{\prime},\|\cdot\|) will be less than coverage number N(ϵ,H,)N(\epsilon,\mathcal{H},\|\cdot\|) for the same ϵ\epsilon.

In particular, the covering number can be used to define an upper bound on Rademacher complexity [15]. HH is the hypothesis space, N(ϵ,H,)N(\epsilon,\mathcal{H},\|\|\cdot\|\|) is the number of covers, and Rademacher complexity Rn(H)R_n(H) can be defined by the following inequality:

Rn(H)2ϵ+3logN(ϵ,H,)n.\mathrm{R}_n(\mathrm{H})\leq2\epsilon+3\sqrt{\frac{\log N(\epsilon,\mathrm{H},\|\|\cdot\|\|)}n}.

We've got logN(ϵ,H,)<logN(ϵ,H,).\log N(\epsilon,\mathrm{H}^{'},\|\|\cdot\|\|)<\log N(\epsilon,\mathrm{H},\\||\cdot\|\|).

Thus we finally show that SGA can be useful in improving the generalization performance by decreasing the number of coverages in the embedding space while increasing the number of samples, which ultimately shrinks the upper bound of Rn(H)R_n(H).

评论

Thank you for your response. The additional experiments provided are comprehensive and thorough; the updated figure is clearer and more comprehensible. Based on these improvements, I am pleased to upgrade the score.

评论

Thank you for your recognition. If you have any other questions, please feel free to ask, and we will respond promptly.

审稿意见
6

This work addresses the scarcity of effective data augmentation strategies tailored for signed graphs, especially considering the dearth of auxiliary information in real-world datasets. By presenting the generalization error bound for SGNNs and disproving the universal benefit of random DropEdge, this paper introduces Signed Graph Augmentation (SGA). SGA innovates with a structure augmentation module identifying potential edges from network patterns and employs a selective strategy to enrich training data. The proposed method significantly improves SGNN performance.

优点

The paper is well structured and has clear writing. The task of signed graph augmentation is interesting and novel. The experimental results show the effectivity of the proposed method.

缺点

  1. Regarding the experimental results shown in Table 1, the paper only shows the results of solely using the backbone model and with SGA augmentation method. The efficiency of the proposed method will be more convincing if it can be compared with another augmentation method. In addition to the DropEdge method, there are still other augmentation methods, e.g., mixup-based methods [1,2] and spectrum-based methods [3,4].
  2. Though the performance when using SGA is better, compared to do not using any augmentation method, it introduces many modules to train, e.g., the encoder and the MLG classifier, thus introducing many more parameters. A comparison of a number of parameters and the training time will be better.
  3. Minor comments: Figure 3 can be improved to be more clear. The Encoder, MLG classifier, and the classifier loss are not explicitly shown in the figure.

Reference: [1] G-mixup: Graph data augmentation for graph classification. ICLR’22 [2] Graph mixup with soft alignments. ICML’23 [3] Spectral augmentation for self-supervised learning on graphs. ICLR’22 [4] Through the Dual-Prism: A Spectral Perspective on Graph Data Augmentation for Graph Classification. arXiv'24

问题

  1. In Section 4, according to ‘the generalization performance of the model is affected by the number of edges’, will randomly adding edges will be more effective than dropping edges in signed graphs? Is there any empirical evidence about this?
  2. Lack of complexity analysis and cost time comparison. Is SGA efficient in large-size datasets?

After rebuttal

Most of my concerns have been addressed. I raise my score to 6.

局限性

The authors discussed the limitations of the proposed method.

作者回复

For Weakness 1 on other augmentation methods:

Thank you for the reviewers' suggestions. We carefully reviewed the recommended papers [1-4]. Although the data augmentation methods discussed in those articles are not specifically designed for the link sign prediction task, their underlying concepts have been highly inspiring for our work. We will include a discussion of the aforementioned methods in the Related Work section of the final version. We selected one method each from the mixup-based methods and spectrum-based methods as augmentation techniques, namely S-Mixup [2] and DP-noise [4], to be applied in link sign prediction on Bitcoin-otc dataset.

ModelAUCF1MicroMacro
SGCN78.3 ± 1.490.5 ± 2.684.0 ± 3.969.2 ± 4.1
SGCN+S-Mixup76.8 ± 2.388.8 ± 2.981.5 ± 4.167.4 ± 3.5
improv.-1.9% \downarrow-1.9% \downarrow-3.0% \downarrow-2.6% \downarrow
GSGNN89.1 ± 1.296.5 ± 0.793.6 ± 1.280.8 ± 2.0
GSGNN+S-Mixup87.0 ± 2.196.2 ± 0.293.1 ± 0.379.4 ± 1.7
improv.-2.4% \downarrow-0.3% \downarrow-0.5% \downarrow-1.7% \downarrow
SiGAT76.5 ± 2.594.9 ± 0.290.4 ± 0.359.2 ± 1.6
SiGAT+S-Mixup57.1 ± 4.465.9 ± 22.256.2 ± 21.243.5 ± 12.2
improv.-25.4% \downarrow-30.6% \downarrow-37.8% \downarrow-26.5% \downarrow
ModelAUCF1MicroMacro
SGCN84.20 ± 0.1293.69 ± 0.1689.07 ± 0.2676.55 ± 0.35
SGCN+DP-noise79.26 ± 0.4192.74 ± 0.4887.42 ± 0.7672.84 ± 0.67
improv.-5.85% \downarrow-0.99% \downarrow-1.92% \downarrow-4.92% \downarrow
GSGNN71.19 ± 7.7396.28 ± 0.6093.13 ± 1.1975.29 ± 8.70
GSGNN+DP-noise70.36 ± 6.3096.03 ± 0.3992.69 ± 0.8174.36 ± 6.60
improv.-1.16% \downarrow-0.26% \downarrow-0.47% \downarrow-1.23% \downarrow
SiGAT76.52 ± 1.9894.85 ± 0.1791.08 ± 0.3159.24 ± 1.61
SiGAT+DP-noise85.20 ± 0.6595.11 ± 0.4991.10 ± 0.5070.87 ± 2.80
improv.11.34% \uparrow+0.00%+0.00%19.63% \uparrow

Based on the experimental results, S-Mixup did not effectively enhance the baseline, because the fusion operation of the S-Mixup at the feature level loses the structural information of the graph. However, DP-noise demonstrated effectiveness on certain metrics, and we believe that this kind of method of preserving the graph structure is worth further exploring. We will include these experimental results in the final version.

For Weakness 2 and Q2 on parameter number and training time:

  1. Model parameter statistics
Model#params#change
SGCN14851-
GSGNN30580-
SiGAT185600-
SGCN+SGA2970214851 \uparrow
GSGNN+SGA4543114851 \uparrow
SiGAT+SGA20045114851\uparrow
  1. Model training cost time statistics

Bitcoin-alpha:

Modeltime(s)change(s)
SGCN51.0 ± 5.3-
GSGNN186.8 ± 42.5-
SiGAT257.6 ± 25.8-
SGCN+SGA197.4 ± 74.1146.4 \uparrow
GSGNN+SGA295.2 ± 65.9108.4 \uparrow
SiGAT+SGA395 ± 79.6137.4 \uparrow

BitcoinOTC:

Modeltime(s)change(s)
SGCN75.4 ± 6.4-
GSGNN142.8 ± 17.2-
SiGAT277.0 ± 30.7-
SGCN+SGA149.6 ± 42.774.2 \uparrow
GSGNN+SGA5217 ± 48.05074.2 \uparrow
SiGAT+SGA351.2 ± 41.274.2 \uparrow
  1. Complexity analysis please see Common Concern.

Overall, the amount of additional parameters and training time for our method is relatively stable. When the model is large, our method does not bring greater computational complexity. Training can be carried out in a relatively short amount of time.

For Weakness 3 on Figure 3:

Thank you for your suggestion. We have made revisions to Figure 3 and have submitted a revised version, which is included in the uploaded PDF. Due to time constraints, we will continue making further modifications later.

For Q1 on randomly adding edges

Due to space limitations, we only present the performance of different baselines on the bitcoin-otc dataset with a randomly increased proportion of edges. The performance on other datasets is provided in the uploaded PDF file. Please refer to the uploaded PDF for detailed statistical results across more datasets.

SGCN:

add_edgeAUCF1MicroMacro
0%78.3 ± 1.490.5 ± 2.684.0 ± 3.969.2 ± 4.1
5%77.4 ± 2.589.1 ± 1.981.8 ± 2.866.8 ± 3.0
10%75.5 ± 1.585.6 ± 2.976.8 ± 4.062.5 ± 3.0
15%74.7 ± 1.685.4 ± 2.976.4 ± 4.161.9 ± 3.1
20%67.5 ± 9.064.2 ± 32.357.7 ± 24.447.3 ± 19.4

GSGNN:

add_edgeAUCF1MicroMacro
0%89.1 ± 1.296.5 ± 0.793.6 ± 1.280.8 ± 2.0
5%87.1 ± 1.696.1 ± 0.192.9 ± 0.378.7 ± 1.0
10%86.4 ± 1.196.1 ± 0.192.8 ± 0.278.1 ± 1.1
15%84.5 ± 1.195.9 ± 0.292.5 ± 0.477.7 ± 0.6
20%85.2 ± 1.495.7 ± 0.392.2 ± 0.577.3 ± 1.4

SiGAT:

add_edgeAUCF1MicroMacro
0%76.5 ± 2.594.9 ± 0.290.4 ± 0.359.2 ± 1.6
5%72.8 ± 3.794.7 ± 0.290.0 ± 0.455.4 ± 4.3
10%69.8 ± 5.394.6 ± 0.189.8 ± 0.254.7 ± 3.1
15%70.1 ± 4.294.5 ± 0.189.7 ± 0.254.4 ± 2.4
20%70.3 ± 3.094.6 ± 0.189.9 ± 0.255.2 ± 2.4

Overall, randomly adding edges tends to degrade the model's performance, and the decline is more significant compared to randomly removing edges (refer to Fig.2 in the paper).

评论

Thank you for your response. The additional experiments provided are comprehensive and thorough; the updated figure is clearer and more comprehensible. Based on these improvements, I am pleased to upgrade the score.

评论

Thank you for your comments and recognition of this work.

If you have any other questions, please let us know and we will present timely responses.

审稿意见
7

Link sign prediction is a significant downstream task in graph data analysis. Current graph data augmentation methods seldom explore this task. This paper analyzes, both theoretically and experimentally, why existing graph data augmentation methods perform poorly on this task and proposes a new data augmentation method tailored for signed graph analysis. Its main contributions include the following two aspects:

  • It provides a generalization error bound for signed graph neural networks and theoretically verifies that the widely used random edgedrop method is not suitable for the link sign prediction task.
  • It proposes a novel signed graph data augmentation framework to address two major issues in current signed graph neural networks (SGNNs), namely sparsity and imbalanced triads.

优点

  1. Although significant progress has been made in the field of graph augmentation, data augmentation methods specifically for signed graphs and link sign prediction are relatively new research directions. Currently, there are relatively few data augmentation methods specifically designed for edge-level tasks in graph augmentation.
  2. In this data augmentation module, the authors propose an intriguing new approach, which is to treat the training difficulty of training samples as a feature for augmentation, thereby adjusting the training weights of these samples.
  3. The paper is well structured, logically coherent and very easy to understand.

缺点

  1. According to the authors' theoretical analysis, reducing the number of training samples does not contribute to improving model performance. However, in Section 3.2, we observed that some training samples were removed because they belonged to imbalanced triplets. Does this observation contradict the theoretical analysis?
  2. The authors mention in the limitations section that "for real-world datasets that do not strongly conform to balance theory, our data augmentation may be less effective." What is the basis for this statement? Are there possible solutions proposed by the authors for this issue?
  3. Some typographical errors: on line 164, the concatenation operation should use [,] instead of [.]. On line 225, there is a missing space between SGNN and "our results". Also, some formulas have punctuation while some do not.

问题

Please refer to Weaknesses.

局限性

The paper argues that the data augmentation method can alleviate rather than completely solve the current obstacles of SGNN methods. This claim is supported by experimental and theoretical analyses. However, a limitation of this article is that it only addresses a subset of issues within GNN models and their impact on downstream tasks, thus its influence may be somewhat limited.

作者回复

(1) For Weakness 1 on some training samples removed :

Our theoretical analysis indeed demonstrates that reducing the number of edges during training can degrade model performance. However, the SGA method does not reduce the overall number of edges. Specifically, we take a cautious approach to edge removal by setting a high threshold (over 0.9, up to 1) before executing any edge removal operation, as detailed in our published code. As a result, the number of edges in the training set actually increases after the data augmentation process. This is corroborated by Table 2, where the graph density is shown to improve following data augmentation.

(2) For Weakness 2 on limitation part:

Regarding the issues raised in the limitations:

Most existing SGNN methods are built on the GCN architecture, and it has been shown [6] that such models struggle to learn appropriate representations from unbalanced cycles. Our data augmentation method relies heavily on these SGNN models to identify suitable potential candidates (i.e., edges). Therefore, designing a signed graph representation framework based on alternative architectures could better address this issue. For example, leveraging large language models (LLMs) could be a promising approach. However, the integration of LLMs with graph structures is still in its early stages, and, to our knowledge, no studies have specifically targeted signed graphs. The reliability of directly applying existing frameworks to this context remains uncertain and requires further validation. Overall, this is a problem worth exploring in greater depth.

(3) For Weakness 3 on typos:

Thank you for the reviewer’s comments. We will correct the typos in the final version of the paper.

(4) For Limitations on influence:

Signed graphs, which assign positive or negative signs to edges, are powerful tools for modeling complex relationships across various fields. They offer valuable insights into social network dynamics, biological and chemical interactions, recommendation systems, and international relations. By capturing both positive and negative interactions, signed graphs facilitate a deeper understanding of intricate systems, providing sophisticated analytical methods to address challenges in multiple disciplines.

评论

The author's response basically addressed my concerns. I also carefully read the author's responses to other reviewers and found them to be very comprehensive. I will stand with my rating.

评论

Thank you for your feedback and appreciation; if you have any further questions, please don't hesitate to reach out, and we'll respond promptly.

审稿意见
7

This paper proposes a new research subfield focusing on data augmentation methods for signed graphs. Unlike the widely studied unsigned graph augmentation, this method targets the downstream task of link sign prediction rather than the mainstream node classification [1] or graph classification [2]. As far as I know, most current graph data augmentation methods mainly address node classification tasks, while there is relatively little enhancement work for edge tasks, or their effectiveness for link prediction is not significant [3]. Additionally, this article provides the first generalization error bound for signed graph neural networks, which is used to analyze why current commonly used data augmentation methods yield unstable results for signed graphs. In designing data augmentation methods, the article introduces a new perspective from curriculum learning.

[1] Kazi, Anees, et al. "Differentiable graph module (dgm) for graph convolutional networks." IEEE Transactions on Pattern Analysis and Machine Intelligence 45.2 (2022): 1606-1617. [2] Chen, Yu, Lingfei Wu, and Mohammed Zaki. "Iterative deep graph learning for graph neural networks: Better and robust node embeddings." Advances in neural information processing systems 33 (2020): 19314-19326. [3] Gasteiger, Johannes, Stefan Weißenberger, and Stephan Günnemann. "Diffusion improves graph learning." Advances in neural information processing systems 32 (2019).

优点

  • This paper proposes a new subfield of research with a wide range of applications. According to my investigation, this article is indeed the first paper on this topic.
  • This paper introduces a novel approach to data augmentation, which involves using curriculum learning to adjust the training weights of challenging edges. This perspective is quite insightful.
  • This paper presents the first generalization error bound for signed graph neural networks and, based on this, analyzes the reasons why current commonly used graph data augmentation methods (such as random edge deletion) yield unstable results for link sign prediction.

缺点

  • Although the article introduces a new problem within a subfield, its application scope is relatively narrow compared to more generalized graph structures.
  • Why can't current graph augmentation methods be directly applied to signed graph representation learning? In the limitations, the authors mention that for datasets with poor balance, the effectiveness of this data augmentation method decreases. Are there any good solutions to this problem?
  • Can the authors provide a more detailed analysis of the factors affecting SGA's performance? Specifically, why do some unsigned and signed GNNs show significant performance enhancements on certain datasets, while others exhibit only marginal improvements or none at all?
  • Some typos: In line 231, there is a missing space after "algorithm 2". In line 236, there is also a missing space after "the main result".

问题

See weaknesses.

局限性

See weaknesses.

作者回复

(1) For weakness 1 on application scope:

Compared to more generalized graphs, signed graph analysis has its exclusive downstream task (i.e., link sign prediction) which are important and very interesting, such as product reviews [10], bill votes [11], paper reviews, polarization study [14], echo chambers [12]. In addition to social networks, the field of bioinformatics also utilizes this approach, for example, to predict upregulation and downregulation relationships between diseases and genes [13].

(2) For weakness 2 on other graph augmentation methods and limitation issues:

Regarding the current graph data augmentation methods that cannot be directly applied to signed graphs, we primarily discuss this in the 3rd paragraph of the introduction. In summary, it mainly includes two aspects:

  1. Most methods are designed for node classification [5], graph classification [1-2], and link prediction tasks, with no existing approaches specifically targeting link sign prediction. Moreover, these methods rely on side information, such as node features and labels, which are often missing in most real-world signed graph datasets that contain only structural information [7-8].

  2. Random structural perturbation based augmentation methods [15] cannot improve SGNN performance. We have verified this from both experimental and theoretical perspectives. (see Fig.2 and Sec.4)

Regarding the issues raised in the limitations:

Given that most current SGNN methods are designed based on the GCN architecture, and This paper [6] has demonstrated that current SGNN models based on such architectures are unable to learn suitable representations from unbalanced cycles. Our data augmentation method heavily relies on this kind of SGNN models to mine suitable potential candidates (i.e., edges), so designing a signed graph representation encoding framework based on alternative architectures could better address the current issue. For instance, using large language models to tackle this problem could be a promising approach. However, the integration of large language models with graphs is still relatively limited. According to our research, there is currently no studies specifically targeting signed graphs. Whether it is reliable to directly use some of the existing frameworks also needs further validation. Overall, this is a very worthwhile problem to explore.

(3) For weakness 3 on Experimental Results:

Here are two observations regarding the experimental results:

  1. The performance improvement of the methods based on balance theory architectures, namely SGCN [7] and SiGAT [8], is more significant compared to GS-GNN [9]. In other words, the improvements achieved by SGCN and SiGAT are more noticeable, whereas the improvements with GS-GNN are relatively smaller.

  2. The overall performance on the first four datasets (Bitcoin-alpha, Bitcoin-otc, Epinion, Slashdot) is better than on the last two datasets (Wiki-elec, Wiki-Rfa).

Regarding Observation 1, from the perspective of method design, the information fusion mechanism of the model itself influences the performance of SGA. We believe the reasons might include two aspects. First, as indicated by the analysis in RSGNN [6], the current SGNN methods based on balanced theory [7] (i.e., SGCN, SiGAT) fail to learn appropriate representations from unbalanced cycles, whereas SGA effectively reduces the proportion of unbalanced cycles (see Table 3), leading to better enhancement for these two methods. Second, since GS-GNN [9] is not limited to the balanced theory assumption, it can handle unbalanced cycles well, thus the enhancement effect from SGA is less significant.

Regarding Observation 2, from the perspective of dataset balance, we believe that the performance of SGA is related to the balance of the dataset. As shown in Table 3, the initial balance degree (BD %) of the first four datasets is better than that of the last two datasets. Considering that we use SGCN as the link sign prediction model to identify potential candidates, its prediction performance is poorer for datasets with low balance degree. As evidenced in Table 1, this conclusion holds true. Therefore, for datasets with low balance, it is more challenging to identify suitable potential candidates, leading to a decrease in the overall data augmentation effectiveness.

(4) For weakness 4 on typos:

Thank you for the reviewer’s comments. We will carefully review the paper and correct typos in the final version.

评论

Thanks for the author's response. It has largely addressed my questions and provided a better understanding of the details of the paper. Therefore, I will keep my score unchanged.

作者回复

Thank all reviewers for their valuable and constructive comments. We address the common concern here and believe the quality of the paper has been improved following the reviewers' suggestions.

Common Concern: Time and Space Complexity of SGA

Suppose we are given a signed graph G=(V,E+,E)\mathcal{G}=(\mathcal{V},\mathcal{E}^+,\mathcal{E}^- ), where V=(v1,,vV)\mathcal{V} = ({v_1, \ldots, v_{|\mathcal{V}|}}) represent the set of nodes, E+\mathcal{E}^+ and E\mathcal{E}^- respectively denote the positive and negative edges, and the structure of G\mathcal{G} is represented by the adjacency matrix ARV×VA∈\mathbb{R}^{|\mathcal{V}|×|\mathcal{V}|}.

Background

The computation of the l-th layer of a SGNN network is:

XL+1=σ(AXlWl)=σ(A+XlWl+AXlWl)X^{L+1}=σ(AX^l W^l) =σ(A^+ X^l W^l+A^- X^l W^l)

Where σ()σ() is a non-linear activation function, WlW^l is a feature transformation matrix RFl×Fl+1∈\mathbb{R}^{F_l×F_{l+1}}, and A+A^+ and AA^- are the adjacency matrix of positive edges and negative edges, respectively. For simplicity, we assume the node features at every layer are size-F. As such, WlW^l is an F×FF×F matrix.

Time Complexity Analysis:

  1. SGNN

    We analyze the complexity of the SGNN by three high-level operations:

    i. feature transformation (ZL=XlWlZ^L=X^l W^l),

    ii. neighborhood aggregation (A+ZL,AZLA^+ Z^L,A^- Z^L),

    iii. activation function (σ()σ()).

Part i. is a dense matrix multiplication between matrices of size N×FlN×F_l and Fl×Fl+1F_l×F_{l+1}. As our assumption Fl=Fl+1=FF_l=F_{l+1}=F, the complexity is O(NF2)O(NF^2).

Part ii. is a multiplication between matrices of size N×NN×N and N×FN×F, yielding O(N2F)O(N^2 F). We use a sparse operator, neighborhood aggregation for each node therefore requires O(d+F+dF)O(d^+ F+d^- F), where d+d^+ and dd^- respectively represent the positive and negative neighbors of each node on average. Finally, we have a total complexity of O(NdF)O(NdF), where d=d++dd=d^++d^-.

Part iii. is simply an element-wise function, so its cost is O(N)O(N).

Over LL layers, the complexity is O(L×(NF2+NdF+N)O(L×(NF^2+NdF+N) for one forward propagation. The time complexity of backpropagation is usually the same as that of forward propagation because it requires the gradient to be computed and the parameters to be updated is the same as forward propagation.

  1. SGA

    We decompose the complexity of the SGA by two operations:

    i. Updating training data by similarity calculation,

    ii. Training model by curriculum learning.

Part i. involves a dense matrix multiplication between two matrices of size N×FN×F, yielding O(N2F)O(N^2 F). Note that, this part runs only once and does not participate in forward and backward propagation. In our experiments, this process can be optimized in parallel, and its time complexity is lower than that of SGN.

Part ii. is an upgraded version of the original SGNN. Similar to the analysis in section 1, we can have a complexity of O(L×(NF2+NdF+N))O(L×(NF^2+NdF+N)) in this part.

Finally, we conclude that the complexity of SGA is the same as that of SGNN, which is O(L×(NF2+NdF+N))O(L×(NF^2+NdF+N)).

Space Complexity Analysis:

  1. SGNN

    i. The space complexity of the input feature matrix X is O(NF)O(NF).

    ii. The adjacency matrix A: O(N2)O(N^2) which can be optimized by sparse operator O(Nd)O(Nd).

    iii. The space complexity of the output feature matrix Hl+1H^{l+1} is O(NF)O(NF), where ll represents the current layer and l+1l+1 represents the next layer.

    iV. The space complexity of the weight matrix W is O(F2)O(F^2 ).

The space complexity of SGNN is O(Nd+NF+F2)O(Nd+NF+F^2 ) for one layer network.

  1. SGA

    i. The space complexity of the input feature matrix X is O(NF)O(NF).

    ii. The adjacency matrix A: O(Nd+M)O(Nd+M) where M is the augmented edges and M<NdM<Nd.

    iii. The space complexity of the output feature matrix Hl+1H^{l+1} is O(NF)O(NF), where ll represents the current layer and l+1l+1 represents the next layer.

    iV. The space complexity of the weight matrix W is O(F2)O(F^2 ).

In the case of using the same activation function, the space complexity of SGA is also O(Nd+NF+F2)O(Nd+NF+F^2 ).

Experimental Setup: Our experimental results are reported as the mean and standard deviation calculated over 5 runs. The uploaded PDF includes supplementary experimental results and an updated framework figure.

References throughout the entire rebuttal:

[1] Han, Xiaotian, et al. G-mixup: Graph data augmentation for graph classification. ICML'22.

[2] Ling, Hongyi, et al. Graph mixup with soft alignments. ICML'23.

[3] Lin, Lu, et al. Spectral augmentation for self-supervised learning on graphs. ICLR'22

[4] Xia, Yutong, et al. Through the Dual-Prism: A Spectral Perspective on Graph Data Augmentation for Graph Classification. arXiv'2024.

[5] Zhao, Tong, et al. Data augmentation for graph neural networks. AAAI'21.

[6] Zhang, Zeyu, et al. Rsgnn: A model-agnostic approach for enhancing the robustness of signed graph neural networks. WWW'23.

[7] Derr, Tyler, Signed graph convolutional networks. ICDM'18.

[8] Huang, Junjie, et al. Signed graph attention networks.ICANN'19

[9] Liu, Haoxin, et al. Signed graph neural network with latent groups. KDD'21.

[10] Seo, Changwon, et al. SiReN: Sign-aware recommendation using graph neural networks. TNNLS'23.

[11] Huang, Junjie, et al. Signed bipartite graph neural networks. CIKM'21.

[12] Tzeng, Ruo-Chun, et al. Discovering conflicting groups in signed networks. Neurips'20.

[13] Zhang, Guangzhan, et al. SGNNMD: signed graph neural network for predicting deregulation types of miRNA-disease associations. BIBM'22.

[14] Xiao, Han, et al. Searching for polarization in signed graphs: a local spectral approach. WWW'20.

[15] Tang, Huayi, et al. Towards understanding generalization of graph neural networks. ICML'23.

[16] He, Yixuan, et al. SSSNET: semi-supervised signed network clustering. SDM'22.

[17] Chen, Sirui, et al. SIGformer: Sign-aware Graph Transformer for Recommendation. SIGIR'24.

最终决定

The paper studies the problem of data augmentation in signed GNNs (SGNNs). Specifically, it first derives generalization bounds for SGNNs, while it also introduces a new augmentation strategy to improve upon DropEdge's performance on link sign prediction. The proposed approach has merits that the reviewers have highlighted. These mostly include the novelty of the methodology as well as the theoretical and empirical support. The authors have adequately addressed most of the reviewers' concerns. Therefore, I recommend accepting the paper.