On the Benefits of Attribute-Driven Graph Domain Adaptation
摘要
评审与讨论
In this paper, we focus on the graph domain adaptation (GDA) problem and propose a novel cross-channel graph attribute-driven Alignment (GAA) algorithm, which aims to address the distribution differences between the source and target graphs, especially the differences in node attributes and topology.
优点
- Through theoretical and experimental analysis, the difference of eigenvalues distribution between graph topology and node attributes is analyzed, which reveals that the difference of node attributes is more significant than the difference of topology structure, and further emphasizes the importance of node attribute alignment.
- The Cross-Channel Module is proposed to align the graph topology and node attributes simultaneously. This module effectively reduces the domain differences by fusing and aligning the two views of the source and target images.
缺点
- The whole setting of this paper is to perform domain adaptation (GDA) under supervised conditions, that is, the annotation of the target domain is involved in the training process. However, for most practical problems, the target domain is extremely lack of data annotation or even no annotation, which lacks certain considerations for practical applications. If the target domain is annotated why not design the model for the target domain instead of introducing the source domain.
- In the experiments section, we see that the latest method you currently cite is the result of 2023, and the comparison test of the relevant latest method in 2024 is missing. And on the types of downstream tasks, there is a lack of evaluation of other tasks, such as graph classification tasks.
- This paper lacks the time complexity analysis of the model, and does not analyze the overall operating efficiency of the model. In particular, GAA algorithm introduces additional calculation steps (such as cross-view similarity matrix), which may lead to increased computational complexity. On large-scale graph data, the efficiency and scalability of the algorithm may become an issue.
- In the model graph part of this paper, in the target domain part, the legend labels of the attribute graph and the topology graph are wrong. The text annotation at the bottom of Figure 1 is ambiguous from the data in the bar graph (the difference in the distribution of eigenvalues is greater for topological graphs than for attributed graphs?).
问题
- This paper does not analyze the time complexity of the model, whether it is universal for larger scale node classification problems, and whether it can ensure the operation efficiency of the model on large-scale problems. For example, in the construction of attributed graph using KNN, is it too time-consuming to successively calculate the similarity score between nodes to connect edges?
- The article is generally on the supervised domain adaptation problem. Personally, I think this application scenario is relatively limited in the general scenario.
- The article does not give detailed hyperparameter Settings, for example, in the weight design of the loss function, although the experiment carries out the analysis of different weights, the final setting value is not given. In addition, for the weight setting of the loss function, will the weight of a loss function be too low to cause it to not work in the model training process?
- Can it be compared with the latest GDA method?
Part 2/3
Regarding NN, we follow existing works in the literature e.g., [4, 5, 6], which adopt it as a pre-processing step to construct the attribute view. Specifically, we note that the construction of attribute view using NN is tied to the graph dataset itself, but does not rely on any specific algorithms or downstream tasks (i.e., we can obtain the attribute view prior to implementing our algorithm). In other words, it does not affect the computational complexity of the algorithm.
Q2
We would like to clarify that our paper specifically addresses the unsupervised domain adaptation (UDA) problem, where the model adapts knowledge from a source domain (with labeled data) to a target domain (without labeled data in line 161 target graph ) . Additionally, our target graph classification loss (Section 4.4, Eq. (4), line 339) does not rely on labels. is a pseudo-label that is obtained through dimensionality reduction. GAA focuses on aligning the node attributes and topological structures between the source and target graphs in the absence of labeled target data, which is a critical challenge in many real-world scenarios.
We believe that this unsupervised setting is highly important, as labeled data in the target domain is often unavailable or expensive to obtain. While our method does not rely on target labels, it employs novel techniques such as attribute alignment to achieve effective adaptation.
Q3
To address your concern, we have provided a detailed explanation of our hyperparameter selection strategy. Additionally, we discuss the optimal parameter settings in the context of our study. Furthermore, a comprehensive overview of the hyperparameter setting value for different datasets is provided in the revised manuscript Appendix D, Table 5.
, , and are chosen from the set . These values provide flexibility for adjusting the relative importance of different loss terms. (the number of neighbors for -NN graph construction) is typically 1, \cdots, 10 . The optimal value for k depends on the density and connectivity of the graph. The extremely large could also introduce noise that will deteriorate the performance [5]. Usually, our largest will be [4].
Airport Dataset: Often contains transportation networks with fewer nodes but complex edge relationships. Given the complexity of this dataset, , , and should be set relatively higher to emphasize topology alignment and capture key structural relationships. , and is selected from . A wide could be more effective due to the nature of these networks (containing both density and sparse). We select from .
Citation Dataset: This dataset often has a higher node count and diverse structural characteristics. In such datasets, balancing the impact of node attributes and topology is important. , and is selected from . A middle value of to capture relevant local structures could work well for this dataset. We select from .
Social Network Dataset (Blog and Twitch): Social networks often contain a large number of nodes with rich attribute information but high variance in topology structure. Emphasize attribute consistency since social networks have distinctive attributes. Thus, attribute discrepancy is more sensitive to the values of , , and , which are selected from the set . Small is recommended due to the dense connections in social networks. We select from .
MAG Dataset: The MAG dataset is large and diverse, containing lots of classes with various relationships and rich metadata. Structural and attribute alignment are key factors. In this context, attribute shifts are both important to the values of , , and , which are selected from the set . The parameter works well when enabling the model to capture high-level local and global structural information within the graph. We select from .
In summary, the weight of the loss function remains stable between 0.1 and 0.5, indicating that each loss function works well.
Part 3/3
Q4
To address your concern, we have included a comparison with the two latest state-of-the-art (SOTA) methods from 2024( Table 1-1, Table 1-2 and Table 3). Additionally, detailed descriptions of the experimental settings and an expanded comparison with other SOTA methods can be found in the revised manuscript ( Section 5.4, Table 2, Table 3, Table 4).
[1] Liu, Shikun, et al. "Pairwise Alignment Improves Graph Domain Adaptation." International Conference on Machine Learning, 2024.
[2] Man, Wu, et al. "Unsupervised Domain Adaptive Graph Convolutional Networks." The Web Conference, 2020.
[3] Shi, Boshen, et al. "Improving Graph Domain Adaptation with Network Hierarchy." Conference on Information and Knowledge Management, 2023.
[4] Chen, Jie, et al. "Fast Approximate kNN Graph Construction for High Dimensional Data via Recursive Lanczos Bisection." Journal of Machine Learning Research, 2009.
[5] Wang, Xiao, et al. "AM-GCN: Adaptive Multi-channel Graph Convolutional Networks." SIGKDD International conference on knowledge discovery & data mining, 2020.
[6] Fang, Ruiyi, et al. "Structure-Preserving Graph Representation Learning." International Conference on Data Mining, 2022.
Thank you for your feedback, which effectively addressed my concern and I am willing to improve my score.
Part 1/3
Thank you for your constructive feedback! Below, we address the concerns and questions raised in the weaknesses section. Please feel free to reach out if further clarification is required.
Q1
We introduce Cross-view Similarity Matrix and Attention-based Attribute components to further boost the performance of GAA. This is essentially a tradeoff between algorithm performance and computational complexity. In light of your comments, we conducted additional experiments by removing the Cross-view Similarity Matrix and Attention-based Attribute, and the results are shown in Table 1-1 and Table 1-2. It can be observed that there is a slight performance drop after removing the Cross-view Similarity Matrix and Attention-based Attribute component. Specifically, we replace with , with , with , and with . However, we note that it still outperforms other SOTA methods, such as recent 2024 work PA[1], which highlights the importance of attribute alignment in GAA.
| Methods | U B | U E | B U | B E | E U | E B | DE EN | EN DE |
|---|---|---|---|---|---|---|---|---|
| PA | 0.679 | 0.528 | 0.562 | 0.532 | 0.529 | 0.677 | 0.739 | |
| GAA_o | 0.552 | |||||||
| GAA | 0.704 | 0.563 | 0.542 | 0.573 | 0.546 | 0.691 | 0.779 | 0.751 |
Table 1-1. Cross-network node classification on the airport network and social network.
| Methods | A D | D A | A C | C A | C D | D C | B1 B2 | B2 B1 |
|---|---|---|---|---|---|---|---|---|
| PA | 0.752 | 0.740 | 0.768 | 0.755 | 0.780 | 0.662 | 0.654 | |
| GAA_o | 0.815 | |||||||
| GAA | 0.789 | 0.754 | 0.824 | 0.782 | 0.771 | 0.798 | 0.681 | 0.679 |
Table1-2. Cross-network node classification on the citation, blog network.
To further investigate the efficiency of GAAo, Table 2 reports the running time comparison across various algorithms. We also compared the training time and GPU memory usage of common baselines UDAGCN [2] and a recent SOTA method, JHGDA, which aligns graph domain discrepancy hierarchical levels [3]. As shown in Table 2, the evaluation results on U B further demonstrate that, with tolerable computational and storage overhead, our method achieves superior performance.
| Methods | Training Time (Normalized w.r.t. UDAGCN) | Memory Usage (Normalized w.r.t. UDAGCN) | Accuracy (%) |
|---|---|---|---|
| UDAGCN | 1 | 1 | 0.607 |
| JHGDA | 1.314 | 1.414 | 0.695 |
| PA | 0.498 | 0.679 | |
| GAA_o | 0.514 | ||
| GAA | 1.063 | 1.113 | 0.704 |
Table2. Comparison of Training Time, Memory Usage, and Accuracy on U B.
We also conduct one additional experiment on large-scale data MAG, and the results are shown in Table 3. It can be observed that GAA achieves state-of-the-art (SOTA) performance on large-scale datasets. Details of other baselines and dataset information are provided in the revised version (Table 4 and Table 1).
| Methods | US CN | US DE | US JP | US RU | US FR | CN US | CN DE | CN JP | CN RU | CNFR |
|---|---|---|---|---|---|---|---|---|---|---|
| PA | 0.400 | 0.389 | 0.474 | 0.252 | 0.262 | 0.383 | 0.333 | 0.242 | ||
| GAA_o | 0.363 | 0.445 | ||||||||
| GAA | 0.410 | 0.401 | 0.492 | 0.372 | 0.288 | 0.453 | 0.302 | 0.400 | 0.351 | 0.293 |
Table 3. Cross-network node classification on MAG datasets.
Dear Reviewer,
Thank you for your thoughtful feedback and for increasing the score of our manuscript. We sincerely appreciate your insightful questions, as addressing them has significantly strengthened our paper. We welcome any additional questions or suggestions you may have. Sincerely,
The Authors
This paper studies the problem of Graph Domain Adaptation and shows the significance of the graph node attribute for graph domain alignment. Then, they propose a novel cross-channel module to fuse and align both views between the source and target graphs for GDA. Experimental results on a variety of benchmarks verify the effectiveness of the method.
优点
- The studied problem is interesting and important.
- The paper is well-organized and clearly written.
- The experiments are extensive and helpful to validate the effectiveness of the model.
缺点
- Some citation styles are not correct, e.g., Line 37.
- The paper lacks some of the latest baselines such as "Information filtering and interpolating for semi-supervised graph domain adaptation (2024)"
- Masking node attributes could have the risk of losing crucial information. How about alignment in the embedding space?
- I suggest that the authors should use some tSNE to show the effectiveness of the method.
- The parameter sensitivity needs more analysis. How to apply your method to a new dataset.
问题
See above.
Part 2/2
Q5
Following your suggestion, we also conduct one additional experiment on the hyperparameter analysis experiment in our revised manuscript Appendix D, Figure 5
To address your concern further, we have provided a detailed explanation of our hyperparameter selection strategy and discussed the optimal parameter settings in the context of our study. Furthermore, the revised manuscript includes a comprehensive overview of the hyperparameter selection process for different datasets, as outlined in Appendix D, Table 5.
, , and are chosen from the set . These values provide flexibility for adjusting the relative importance of different loss terms. (the number of neighbors for -NN graph construction) is typically 1, \cdots, 10 . The optimal value for depends on the density and connectivity of the graph. The extremely large could also introduce noise that will deteriorate the performance [5]. Usually, our largest will be [4].
Airport Dataset: Often contains transportation networks with fewer nodes but complex edge relationships. Given the complexity of this dataset, , , and should be set relatively higher to emphasize topology alignment and capture key structural relationships. , and is selected from . A wide could be more effective due to the nature of these networks (containing both density and sparse). We select from .
Citation Dataset: This dataset often has a higher node count and diverse structural characteristics. Balancing the impact of node attributes and topology is important in such datasets. , and is selected from . A middle value of to capture relevant local structures could work well for this dataset. We select k from .
Social Network Dataset (Blog and Twitch): Social networks often contain a large number of nodes with rich attribute information but high variance in topology structure. Emphasize attribute consistency since social networks have distinctive attributes. Thus, attribute discrepancy is more sensitive to the values of , , and , which are selected from the set . Small is recommended due to the dense connections in social networks. We select from .
MAG Dataset: The MAG dataset is large and diverse, containing lots of classes with various relationships and rich metadata. Structural and attribute alignment are key factors. In this context, attribute shifts are both important to the values of , , and , which are selected from the set . The parameter k works well when enabling the model to capture high-level local and global structural information within the graph. We select from .
[1] Qiao, Ziyue, et al. "Information filtering and interpolating for semi-supervised graph domain adaptation." Pattern Recognition, 2024
Thanks for your response and I support the paper.
Dear Reviewer,
We thank the reviewer for their positive and constructive feedback. We are glad to know that the reviewer liked our paper. We provide answers to the reviewers' comments below and refer to them in red in our revised manuscript. We hope we were able to provide satisfactory answers. Please feel free to ask if the reviewer requires any further clarification.
Sincerely,
The Authors
Part 1/2
Thank you for your appreciated feedback. Below, we address the concerns and questions raised in the weaknesses section. Please feel free to reach out if further clarification is required.
Q1
Thank you for your comment regarding the citation style. We apologize for the inconsistencies in our citations and have carefully reviewed and corrected them throughout the paper.
Q2
We acknowledge the importance of comparing our approach with recent advancements, including GIFI [1]. Although our method focuses on unsupervised graph domain adaptation, we recognize the value of benchmarking against semi-supervised methods for broader context. In our revised submission, we have incorporated this baseline into our experiments (see Section 5.4, Table 3 and Table 4). Our results demonstrate that while their method performs well in semi-supervised settings, our approach exhibits outperformance in unsupervised scenarios, highlighting its broader applicability. Additionally, we have clarified the distinctions between the two settings in Section 5.2.
| Methods | U B | U E | B U | B E | E U | E B | DE EN | EN DE |
|---|---|---|---|---|---|---|---|---|
| GIFI | 0.636 | 0.521 | 0.493 | 0.535 | 0.501 | 0.623 | 0.719 | 0.705 |
| GAA | 0.704 | 0.563 | 0.542 | 0.573 | 0.546 | 0.691 | 0.779 | 0.751 |
Table1-1. Cross-network node classification on the airport network and social network.
| Methods | A D | D A | A C | C A | C D | D C | B1 B2 | B2 B1 |
|---|---|---|---|---|---|---|---|---|
| GIFI | 0.751 | 0.737 | 0.793 | 0.755 | 0.739 | 0.751 | 0.653 | 0.642 |
| GAA | 0.789 | 0.754 | 0.824 | 0.782 | 0.771 | 0.798 | 0.681 | 0.679 |
Table1-1. Cross-network node classification on the Citation, Blog network.
Q3
Regarding the concern about masking node attributes and potential information loss, we would like to clarify that our method does not utilize node attributes masking. Our approach leverages full node attributes to preserve all available information. Although GAA is not designed to align in the embedding space, our approach achieves alignment by technically aligning graph node attributes and topology discrepancy in the embedding space. This ensures that representations in the embedding space are well-suited for domain adaptation without masking. We acknowledge the importance of exploring alignment in the embedding space as an independent strategy. This aligns with your suggestion and provides a promising avenue for future research. We will consider incorporating dedicated embedding alignment techniques in follow-up studies to further validate and improve our method.
Q4
We are pleased to provide additional T-SNE visualizations of citation datasets to demonstrate the effectiveness of our method. Detailed visualization results are in Appendix F of our updated manuscript. We observe from Appendix F compared with other SOTA methods. Firstly, nodes belonging to different classes are well-separated, demonstrating that GAA effectively distinguishes between classes in the embedding space. Secondly, nodes from the same class across different domains largely overlap, indicating that GAA significantly reduces domain discrepancies. The first observation highlights strong classification performance, while the second suggests effective domain adaptation.
This paper aims to address the Graph Domain Adaptation problem by incorporating both structural information and node attribute values, the latter of which is often neglected by previous approaches. The paper first presents the significance of considering node attribute shifts in GDA. It then derives a PAC-Bayesian analysis, which provides an upper bound on the overall discrepancy in expected loss. This bound relies solely on the divergence between the topologies and node attributes of the source and target graphs. The proposed method combines the topological and node feature representation through an attention module and performs domain adaptive learning between the source and target representations. Experiments showed good performance of the proposed algorithm, compared with multiple baselines.
优点
- This paper makes a good observation that node attributes are important for domain adaptation tasks and then follows up with a theoretical analysis.
- The authors conducted comprehensive experiments to validate the algorithm's performance with a sufficient number of baseline methods compared.
- The ablation study shows the effectiveness of the proposed method and its key components.
缺点
- Writing: The writing of the paper can be improved. In particular, citation errors appear on the very first page of the paper. The author should proofread the paper more carefully before submission. Some sentences contain grammatical errors.
- Theoretical novelty: The theoretical analysis of the paper is a direct application of the PAC-Bayesian results derived from [1], including the problem formulation and Theorem 1. The discrepancy between subgroups in [1] can be naturally applied to graphs in the source/target domain.
- Tightness of bounds: The upper bound derived in Proposition 1 is novel, yet I am curious if the upper bound is tight enough to be useful. It would be great if the author could provide some numerical evidence showing the tightness of the bound or experiments that show how the improvement of performance relates to the divergence of both topological structure and data features over training.
- Scalability: The scalability of the proposed algorithm is a concern, especially when applied to large graphs. The GAA algorithm requires constructing and processing both topology and attribute views, with a cross-view similarity matrix that grows with the size of the graph. This can lead to high computational costs, particularly when both views are combined through an attention module. Running time analysis and comparison with existing baselines are needed.
References [1] Ma, J et al. Subgroup generalization and fairness of graph neural networks.
问题
Please refer to the weakness section.
Part 2/2
Q4
We introduce Cross-view Similarity Matrix and Attention-based Attribute components to further boost the performance of GAA. This is essentially a tradeoff between algorithm performance and computional complextity. In light of your comments, we conduct additional experiments by removing Cross-view Similarity Matrix and Attention-based Attribute and the results are shown in Table 1-1 and Table 1-2. It can be observed that there is a slight performacne drop after removing the Cross-view Similarity Matrix and Attention-based Attribute component. Specifically, we replace with , with , with , and with . However, we note that it still outperforms other SOTA methods, such as recent 2024 work PA[1], which highlights the importance of attribute alignment in GAA.
| Methods | U B | U E | B U | B E | E U | E B | DE EN | EN DE |
|---|---|---|---|---|---|---|---|---|
| PA | 0.679 | 0.528 | 0.562 | 0.532 | 0.529 | 0.677 | 0.739 | |
| GAA | 0.552 | |||||||
| GAA | 0.704 | 0.563 | 0.542 | 0.573 | 0.546 | 0.691 | 0.779 | 0.751 |
Table1-1. Cross-network node classification on the airport network and social network.
| Methods | A D | D A | A C | C A | C D | D C | B1 B2 | B2 B1 |
|---|---|---|---|---|---|---|---|---|
| PA | 0.752 | 0.740 | 0.768 | 0.755 | 0.780 | 0.662 | 0.654 | |
| GAA | 0.815 | |||||||
| GAA | 0.789 | 0.754 | 0.824 | 0.782 | 0.771 | 0.798 | 0.681 | 0.679 |
Table1-2. Cross-network node classification on the citation, blog network.
To further investigate the efficiency of GAAo, Table 2 reports the running time comparison across various algorithms. We also compared the training time and GPU memory usage of common baselines UDAGCN [2] and a recent SOTA method, JHGDA, which align graph domain discrepancy hierarchical levels [3]. As shown in Table2, the evaluation results on U B further demonstrate that, with tolerable computational and storage overhead, our method achieves superior performance.
| Methods | Training Time (Normalized w.r.t. UDAGCN) | Memory Usage (Normalized w.r.t. UDAGCN) | Accuracy (%) |
|---|---|---|---|
| UDAGCN | 1 | 1 | 0.607 |
| JHGDA | 1.314 | 1.414 | 0.695 |
| PA | 0.498 | 0.679 | |
| GAA | 0.514 | ||
| GAA | 1.063 | 1.113 | 0.704 |
Table2. Comparison of Training Time, Memory Usage, and Accuracy on U B.
Following your suggestion, we also conduct one additional experiment on large-scale data MAG, and the resutls are shown in Table 3. It can be observed that GAA achieves state-of-the-art (SOTA) performance on large-scale datasets. Details of other baselines and dataset information are provided in the revised version (Table 4 and Table 1).
| Methods | US CN | US DE | US JP | US RU | US FR | CN US | CN DE | CN JP | CN RU | CN FR |
|---|---|---|---|---|---|---|---|---|---|---|
| PA | 0.400 | 0.389 | 0.474 | 0.252 | 0.262 | 0.383 | 0.333 | 0.242 | ||
| GAA | 0.363 | 0.445 | ||||||||
| GAA | 0.410 | 0.401 | 0.492 | 0.372 | 0.288 | 0.453 | 0.302 | 0.400 | 0.351 | 0.293 |
Table3. Cross-network node classification on MAG datasets.
[1] Liu, Shikun, et al. "Pairwise Alignment Improves Graph Domain Adaptation." International Conference on Machine Learning, 2024.
[2] Man, Wu, et al. "Unsupervised Domain Adaptive Graph Convolutional Networks." The Web Conference, 2020.
[3] Shi, Boshen, et al. "Improving Graph Domain Adaptation with Network Hierarchy." Conference on Information and Knowledge Management, 2023.
I appreciate the thorough rebuttal, and Appendix F addresses my concerns regarding the usefulness of the theoretical results. After seeing the comprehensive discussion in the rebuttal, I am now leaning toward accepting this paper. Therefore, I've raised the score to 6. Please ensure the efficient GAA results are also included in the manuscript.
Dear Reviewer,
Thank you for your thoughtful feedback and for increasing the score of our manuscript. We greatly appreciate your insightful questions, as addressing them has helped strengthen our paper. The efficient GAA results will be included in the revised submission, along with the other results. We welcome any further questions or suggestions you may have.
Sincerely,
The Authors
Part 1/2
Thank you for your constructive feedback! Below, we address the concerns and questions raised in the weaknesses section. Please feel free to reach out if further clarification is required.
Q1
Thank you for your comment regarding the citation style. We apologize for the inconsistencies in our citations and have carefully reviewed and corrected this throughout the paper.
Q2
As mentioned in Line 187 of our manuscript, Theorem 1 indeed follows (Ma et al.,2021). On the other hand, we would like to note that Theorem 1 serves merely as a stepping stone for our theoretical analysis; our ultimate goal is to introduce Proposition 1, which highlights the role of attribute and topology divergence in GDA. On the algorithmic side, GAA minimizes both attribute and topology distribution shifts based on intrinsic graph property.
Q3
We appreciate your question, which gives us the opportunity to further validate the rationale behind our theoretical analysis and algorithm. Specifically, to investigate how the performance relates to the divergence, we have conducted an additional experiment, where we randomly generated 1000 pairs of graphs (source and target) varying attribute and topology values. At each time, the divergence of attribute and topology values between the source and target graphs was adjusted by utilizing make-blobs and stochastic block model (Details in Appendix F Section F1, Section F2). Figure 7 in Appendix F presents the normalized bound value divergence (calculated using Eq. (4) from Proposition 1) and loss value divergence (defined in Eq. (12) on page 6) for both attribute and topology features as a function value divergence. It is evident that the theoretical values (blue dots) and actual loss values (red dots) exhibit highly consistent trends.
Furthermore, in each simulation, we first train a model using source values (either attribute or topology) and then evaluate its accuracy on the target graph (target accuracy). As shown in Figure 6, we observe that regardless of whether attributes or topology values are used, the target accuracy (green dots) consistently decreases as the divergence between the source and target domains increases. These empirical results highlight that both factors significantly influence the model's performance on the target domain, while the importance of attributes has long been overlooked in existing GDA studies. This further validates the rationale behind our work.
This paper addresses the Graph Domain Adaptation (GDA) problem by investigating the critical role of node attributes in domain shift. This study reveals that node attributes contribute significantly to domain discrepancy. Through theoretical analysis using the PAC-Bayes framework, the authors prove that the domain shift is bounded by both topological structure differences and node attribute differences. Based on this insight, they propose Graph Attribute-driven Alignment (GAA), a novel method that aligns both structural and attribute information between domains. Experiments are conducted on two datasets: the Airpot network, Citation, Blog, and Social network.
优点
- Unlike previous methods that primarily focused on graph structural information, this approach provides a novel perspective by emphasizing the importance of node attributes in GDA.
- A solid theoretical analysis proved that domain discrepancy comes from both graph structure and node attributes.
- Strong experimental results.
缺点
- The motivation for using an attention-like strategy is unclear.
- L_A contains both and . There is no clear demonstration of which component is more important for performance. To highlight the effectiveness of aligning attributes, we need to check the effectiveness of (with/without Coss-view Matrix Refinement).
- The effect of is unclear. It is necessary to verify the contribution of compared to , as is also a domain alignment strategy.
问题
- Please provide more information about the motivation behind the attention-like strategy (Eq. 8).
- How does the contribution of compare to ?
- Which is more important for the performance? or ?
Part 2/2
Q3
is a widely adopted unsupervised domain adaptation (UDA) loss in many graph domain adaptation (GDA) works [1, 2, 3], we include it to ensure a fair comparison. In Section 5.5 ablation study of our manuscript, the variant GAA_3 relies on the , which only leverages the topological information from the source and target graphs. Following your suggestion, we also conduct one additional experiment on the ablation experiment, including the performance of the model variant GAA_D, where is excluded. We can observed that we observe that the model's performance is only slightly affected without , confirming that plays a more important role.
| Methods | U B | U E | B U | B E | E U | E B | DE EN | EN DE |
|---|---|---|---|---|---|---|---|---|
| GAA | 0.696 | 0.557 | 0.538 | 0.567 | 0.539 | 0.682 | 0.772 | 0.744 |
| GAA | 0.704 | 0.563 | 0.542 | 0.573 | 0.546 | 0.691 | 0.779 | 0.751 |
Table3. Cross-network node classification on the airport network and social network.
| Methods | A D | D A | A C | C A | C D | D C | B1 B2 | B2 B1 |
|---|---|---|---|---|---|---|---|---|
| GAA | 0.782 | 0.747 | 0.820 | 0.778 | 0.768 | 0.786 | 0.674 | 0.671 |
| GAA | 0.789 | 0.754 | 0.824 | 0.782 | 0.771 | 0.798 | 0.681 | 0.679 |
Table4. Cross-network node classification on the Citation, Blog network.
[1] Wu, Jun, et al. "Non-IID Transfer Learning on Graphs." Annual AAAI Conference on Artificial Intelligence, 2023
[2] Man, Wu, et al. "Unsupervised Domain Adaptive Graph Convolutional Networks." The Web Conference, 2020.
[3] Shi, Boshen, et al. "Improving Graph Domain Adaptation with Network Hierarchy." Conference on Information and Knowledge Management, 2023.
Dear authors,
Thanks for your reply.
It is noteworthy that both and perform well. This indicates that the proposed method is indeed superior to the baseline. However, the margin between and is also small, which makes it difficult to demonstrate that the node attribute-concentrated alignment is more effective than the topology-concentrated alignment. The performance of and is also similar, especially on the second dataset.
I will keep my scores.
Part 1/2
Dear Reviewer,
Thank you for your follow-up comments! We are happy to hear your thoughtful feedback and being supportive. There might be miscommuncations regarding the motivation of our work. We would like to note that the focus of this paper is NOT to demonstrate that attribute-concentrated alignment is more effective than the topology-concentrated alignment. Instead, the key insight of our work is both attribute and topology information plays crucial role for GDA, while former has been long overlooked in existing methods.
Firstly, by utilizing both attribute and topology of information, GAA performs better than the existing SOTA, as shown in Table 2, Table 3 and Table 4. In addition, to further validate the motivation of our work, we conducted an additional experiment to investigate the impact of attribute and topology information for GDA, respectively. Specifically, we andomly generated 1000 pairs of graphs (source and target) varying attribute and topology values. At each time, we first train a model using source values (either attribute or topology) and then evaluate its accuracy on the target graph (target accuracy). Figure 7 in Appendix F presents the normalized bound value divergence (calculated using Eq. (4) from Proposition 1) and loss value divergence (defined in Eq. (12) on page 6) for both attribute and topology features as a function of value divergence. It can be observed that regardless of whether attributes or topology values are used, the target accuracy (green dots) consistently decreases as the divergence between the source and target domains increases. These empirical results further highlight that both factors significantly influence the model's performance on the target domain, while the importance of attributes has long been overlooked in existing GDA studies.
Secondly, the performance of and cannot reflect the impact of attribute or topology information -- only removing (or ) cannot completely eliminate the effect of attribute (or topology) information in , since both types of information intertwine to each other through Cross-view Similarity Matrix Refinement (Eq. 9, Eq. 10).
Thirdly, in order to eliminate the impact of attribute information, we have conducted an ablation study in Section 5.5, where we construct by removing the whole attribute graph channel and only utilize the topology information through . Table 5 and Table 6 below compare the performances of and , and it is evident that the accuracy of has dropped significantly, which further validates the rationale behind .
Regarding , it is actually removing rather than only utlizing . The reason that removing only slightly affects the performance is that we already align topology information through (hence adding does not provide additional performance gain). Here, we include only for fair comparison purpose, since the other GDA algorithms also employ this term.
In summary, our work emphasizes the importance of both topology and attribute for GDA, which has not been proposed in previous work. Moreover, our work also opens venues for future work to further leverage the attribute information for GDA.
In the meanwhile, we really appreciate your question, which gives us the opportunity to further improve our manuscript. Specifically, in light of your comments, we have revised Section 3 (line 203-208) and Section 6 (line 512-514) to further clarify our motivation and avoid potential confusion.
Part 2/2
| Methods | U B | U E | B U | B E | E U | E B | DE EN | EN DE |
|---|---|---|---|---|---|---|---|---|
| 0.609 | 0.498 | 0.512 | 0.517 | 0.441 | 0.479 | 0.688 | 0.654 | |
| 0.704 | 0.563 | 0.542 | 0.573 | 0.546 | 0.691 | 0.779 | 0.751 |
Table 5. Cross-network node classification on the airport network and social network.
| Methods | A D | D A | A C | C A | C D | D C | B1 B2 | B2 B1 |
|---|---|---|---|---|---|---|---|---|
| 0.685 | 0.631 | 0.731 | 0.681 | 0.719 | 0.655 | 0.623 | 0.611 | |
| 0.789 | 0.754 | 0.824 | 0.782 | 0.771 | 0.798 | 0.681 | 0.679 |
Table 6. Cross-network node classification on the Citation, Blog network.
Part 1/2
Thank you for your feedback. We would also appreciate your agreement on our method's novelty and effectiveness. Below, we address the concerns and questions raised in the weaknesses section. Please feel free to reach out if further clarification is required.
Q1
Our motivation is to make the attribute view matrix learnable. It selects essential attribute information to enable the model to obtain learnable attribute view information rather than relying on fixed attributes. This is essentially a tradeoff between algorithm performance and computional complextity to provide better performance.
Q2
The results from the NN-GCN experiment (Tables 2 and 3) show that the Graph Convolutional Network (GCN) performs better when only relying on node attribute information. This finding highlights influence of attributes on model performance. We conducted ablation experiments focusing on this two specific terms: and . The other loss functions used in model training are the same as those in GAA. Specifically, in the function, GAA employs but removes , while GAA utilizes but removes . According to Table1. and Table2., we observe that GAA generally outperforms GAA, suggesting that addressing attribute discrepancy is more beneficial for the performance of GDA. This is consistent with our observation that attributes discrepancy significantly larger than topology feature value discrepancy in our benchmarks. In addition, we discuss the impact of these two items separately in more detail in our revised manuscript Appendix F. In summary, while contributes more significantly to performance improvement, also plays a important role in enhancing the overall model.
| Methods | U B | U E | B U | B E | E U | E B | DE EN | EN DE |
|---|---|---|---|---|---|---|---|---|
| GAA | 0.672 | 0.544 | 0.526 | 0.534 | 0.678 | 0.762 | 0.738 | |
| GAA | 0.561 | |||||||
| GAA | 0.704 | 0.563 | 0.542 | 0.573 | 0.546 | 0.691 | 0.779 | 0.751 |
Table1. Cross-network node classification on the airport network and social network.
| Methods | A D | D A | A C | C A | C D | D C | B1 B2 | B2 B1 |
|---|---|---|---|---|---|---|---|---|
| GAA | 0.761 | 0.738 | 0.804 | 0.756 | 0.783 | 0.660 | 0.659 | |
| GAA | 0.771 | |||||||
| GAA | 0.789 | 0.754 | 0.824 | 0.782 | 0.771 | 0.798 | 0.681 | 0.679 |
Table2. Cross-network node classification on the Citation, Blog network.
This paper focuses on addressing challenges in graph domain adaption, where labeled data in graph domains is often scarce. The authors highlight that most existing methods neglect the importance of node attributes and primarily focus on aligning structural features of graphs. They demonstrate that node attribute discrepancies significantly impact graph domain adaption performance, often more than topology shifts. They introduce Graph Attribute-Driven Alignment (GAA) algorithm, which incorporates both topology and attribute views for effective alignment. The experimental results demonstrate the effectiveness of the proposed method.
The overall quality of this paper is good. This paper has both theoretical and experimental analysis, which can which reveal that the difference of node attributes is more significant than the difference of topology structure and further emphasize the importance of node attribute alignment. The cross-channel module is proposed to align the graph topology and node attributes simultaneously, which is also demonstrated to be effective.
审稿人讨论附加意见
This paper finally receives the scores of 6, 6, 6, 6, which means, all the reviewers vote for weak acceptance. After the rebuttal process, the authors have addressed the key concerns raised from the reviewers, and thus all the reviewers finally gave a positive score to this paper. I would suggest that the authors could incorporate the suggestions and other minor issues pointed out by the reviewers into the final version of the paper.
Accept (Poster)