PaperHub
7.0
/10
Poster3 位审稿人
最低4最高5标准差0.5
4
5
4
3.7
置信度
创新性2.7
质量2.7
清晰度2.7
重要性2.3
NeurIPS 2025

How Does Topology Bias Distort Message Passing in Graph Recommender? A Dirichlet Energy Perspective

OpenReviewPDF
提交: 2025-04-04更新: 2025-10-29

摘要

关键词
Simplicial ComplexRecommender SystemDirichlet Energy

评审与讨论

审稿意见
4

Graph based recommender systems suffer from popularity bias. That is, the input graph (interaction graph between users and items) gets dominated by highly connected nodes which results in over-represenation of certain items. This results in reinforcing biases and fairness issues through the user-system feedback loop. This work analyses this topology bias through the lens of Dirichlet energy and show that message passing inherently minimises a global smoothing objective. Higher degree nodes accumulate larger embedding norms, a phenomenon called norm concentration. As a resolution, a test time simplicial propagation method is proposed which performs message passing over simplicial complexes with intra- and inter-simplex smoothing to avoid the dominance of certain nodes or substructures.

优缺点分析

Strengths

  1. The Dirichlet energy perspective to analyse graph based recommender systems is interesting.

  2. The formalization of norm concentration and how message passing favors highly connected nodes in recommender systems explaining over-representation of certain items is also fascinating.

  3. The paper is well-written and easy to follow.

Weaknessess

Major weaknesses

  1. I think the title of the paper should accurately reflect the contributions of the work. Currently, the title might prompt readers to think that the study is about topology bias in message passing GNNs covering more general form of GNNs such as GCNs, GATs, but the empirical analysis focuses specifically on LightGCN-style architectures applied to graph based recommender systems, which are simplified frameworks lacking feature transformations and non-linearities found in widely used GNNs. While I understand the creative liberty to have a catchy paper title, I think the title can be revised to better reflect the scope of the contributions.

  2. The idea that message passing (MP) is inherently minimizing Dirichlet energy is not a new insight. Extensive work studying over-smoothing in GNNs (see some of the references listed below [1], [2], [3]) have already established this fact. The way the work is currently presented doesn't adequately acknowledge this foundation or clearly articulate what's novel beyond existing theoretical understanding. It would make sense to properly position this work’s contributions in light of these related works.

  3. Although the norm concentration analysis is interesting, the idea that higher degree nodes receive larger message passing updates is obvious. For instance [4] also observes this phenomenon empirically (the current work also acknowledges this which is great). This leads again to my main issue which is properly discussing the contributions of the work in light of contemporary works. For instance works such as [9] and [10] treat this degree bias/topology bias more generally for GNNs and one can consider norm concentration in LightGCN like architectures used specifically for recommender systems as an application/special case of mitigating degree bias in GNNs ([4],[5],[6],[7],[8],[9],[10]).

  4. The resolution proposed to mitigate this topology bias is test-time simplicial complex propagation approach which although seems to work based on presented experimental validation, it is usually expensive to compute. Currently, the work compares with only few debiasing methods but what about works such as [5],[6],[7],[8],[9] and [10] which might provide a cheaper way to mitigate degree bias in the general framework of GNNs and not restricted to recommender systems? It would be interesting to check if such proposed methods already address the identified issue of topology bias. And why a more computationally expensive approach of lifting graphs to simplices is a better/preferred approach?

Minor weaknesses

  1. L26-27 maybe explain in 1-2 lines what exactly is Matthew effect?

  2. The computational complexity for lifting the graphs into simplicial complex is missing.

References

  1. A Note on Over-Smoothing for Graph Neural Networks. Cai et al, ICML 2020.

  2. PairNorm: Tackling Oversmoothing in GNNs. Zhao et al, ICLR 2020

  3. Not too little, not too much: a theoretical analysis of graph (over)smoothing. Keriven et al, 2022.

  4. Test-Time Embedding Normalization for Popularity Bias Mitigation. Kim et al, CIKM 2023.

  5. On Generalized Degree Fairness in Graph Neural Networks. Liu et al, AAAI 2023.

  6. GRAPHPATCHER: Mitigating Degree Bias for Graph Neural Networks via Test-time Augmentation. Ju et al, NeurIPS 2023.

  7. Tail-GNN: Tail-Node Graph Neural Networks. Liu et al, KDD, 2021.

  8. Investigating and Mitigating Degree-Related Biases in Graph Convolutional Networks. Tang et al. CIKM 2020.

  9. Theoretical and Empirical Insights into the Origins of Degree Bias in Graph Neural Networks. Subramonian et al, 2024.

  10. Degree-based stratification of nodes in Graph Neural Networks. Ali et al, ACML 2023.

问题

See weaknessess above.

局限性

Yes

最终评判理由

The authors have addressed my concerns and the experimental results strongly support their claims, I believe the current presentation does not adequately acknowledge existing work on over-smoothing and degree bias in the broader GNN literature. As I said in my review I believe the present work can be viewed as a special case of these established approaches applied to recommender systems. The theoretical contributions presented do not clearly articulate what is novel beyond the existing body of work on Dirichlet energy and over-smoothing. Given these concerns about theoretical positioning alongside the solid experimental validation, I have increased my score to 4.

格式问题

None

评论

W1: "I think the title of the paper should accurately reflect the contributions ..."

A1: Thanks for raising this point. We agree that the scope of our paper is focused on graph-based recommender systems (RS) and this research focus could be better reflected in the title. Since the LightGCN-style architecture is widely adopted by graph-based RS and state-of-the-art models, we focus primarily on presenting the bias issue presented in such models in empirical results. However, our theoretical framework and the induced method can be applied to general graph-based models. We are willing to modify our title to "How Does Topology Bias Distort Message Passing in Graph Recommenders? A Dirichlet Energy Perspective".

W2: "The idea that message passing (MP) is inherently minimizing Dirichlet energy is not a new insight ..."

A2: First, we want to clarify the theoretical differences of our paper with existing oversmoothing research papers. Dirichlet energy has long been used to study the oversmoothing problem in graph representation learning. while our paper shares the definition of aligning Dirichlet energy optimization with graph message passing, we present completely distinct theoretical analysis and methodology in our paper. Existing oversmoothing papers have established that the Dirichlet energy is exponentially minimized over the stacking of message passing layers and eventually result in the difference of node representations vanishing. On the contrary, our theoretical results show that the norm updates of node embeddings not only relate to their degrees (Corollary 3.1) but also are bounded by their Dirichlet energy (Lemma 3.2). Moreover, we extend the Dirchlet energy analysis to higher order simplicial complexes (SC) and present how message on SC can reduce the norm concentration phenomenon under such theoretical framework. We believe that our theoretical contributions are distinct from existing oversmoothing papers in both problem studied and the resulting methodology, showing originality and novelty in the study of topology bias. Nevertheless, we agree that clearly positioning our work in relation to existing studies on Dirichlet energy is important. We will add these details in the related work section.

W3: "Although the norm concentration analysis is interesting, the idea that higher degree nodes receive larger message ..."

A3: We partially agree that the degree bias problem shares some similar aspects to our paper, but there are still some key differences we want to clarify. The degree bias often describes how nodes with high degrees in graph affect the classification results of other nodes, under the assumption of homophily, i.e., nodes closely connected tend to share the same class label. The debiasing process mainly focuses on improving the label accuracy for low degree nodes. However, we study how the topology bias in a graph distorts the message passing dynamics in graph-based recommenders, leading to popular items being dispropotionately recommended to users. The recommendation process can be regarded as a link prediction task in a heterogeneous user-item bipartite graph. We want our debiasing method to help niche items (low degree nodes) have an appropriate portion in final recommendation lists and thus provide results fit users' real interests. Moreover, the papers studying degree bias often deal with the node classification task in a semi-supervised manner, where labels are only available for a small subset of nodes. In contrast, nodes in graph-based RS are supervised by all their interactions, providing a supervision signal even for low-degree nodes. In summary, while they are both bias issues originated from unbalanced node degree distribution, they have distinct tasks, debiasing goals, bias formation mechanism and learning dynamics. Therefore, we believe our contributions are distinct and do not overlap with this line of research on degree bias

W4: "The resolution proposed to mitigate this topology bias is test-time simplicial complex ..."

A4: Regarding the reviewer's concern about complexity, we want to further stress that our proposed TSP is designed as a test-time module that operates on a pre-trained model. The computational overhead of our TSP method, including the Hodge Laplacian calculation, does not affect the training efficiency of the underlying recommendation model. This one-time computation per test phase is a key advantage of our approach. As for the degree bias solutions, specially [1-6] mentioned by reviewer, as in our answers to Weakness 3, we want to kindly remind the reviewer that these methods all focus on node classification tasks. Due to distinct task, debiasing goals, learning dynamics and bias formation mechanism, we believe they are not directly applicable to topology bias in RS and thus can not serve as baselines for comparison with our proposed method.

评论

W5: "L26-27 maybe explain in 1-2 lines what exactly is Matthew effect?"

A5: The Matthew effect describes the rich-get-richer phenomenon and is widely used in recommendation fairness works to represent the bias reinforcement loop. More specifically in topology bias, the popular items tend to have larger embeddings norms (Corollary 3.1) and updates (Lemma 3.2), which results in potentially higher prediction scores. This leads to more frequent recommendations, making these items even more popular. In this reinforcement loop, popular items can eventually dominate the recommendation lists and become disproportionately recommended to all users.

W6: "The computational complexity for lifting the graphs into simplicial complex is missing."

A6: The process of lifting a graph to a k-simplicial complex (e.g., finding all 3-cliques for 2-simplices) is a one-time, offline preprocessing step performed at test time. We apply the widely used clique mining Bron–Kerbosch [7]. For a graph with nn vertices, the worst-case complexity is O(3(n/3))O(3^\wedge{(n/3)}) [8, 9].

References

[1] On Generalized Degree Fairness in Graph Neural Networks. Liu et al, AAAI 2023.

[2] GRAPHPATCHER: Mitigating Degree Bias for Graph Neural Networks via Test-time Augmentation. Ju et al, NeurIPS 2023.

[3] Tail-GNN: Tail-Node Graph Neural Networks. Liu et al, KDD, 2021.

[4] Investigating and Mitigating Degree-Related Biases in Graph Convolutional Networks. Tang et al. CIKM 2020.

[5] Theoretical and Empirical Insights into the Origins of Degree Bias in Graph Neural Networks. Subramonian et al, 2024.

[6] Degree-based stratification of nodes in Graph Neural Networks. Ali et al, ACML 2023.

[7] Algorithm 457: finding all cliques of an undirected graph. Bron, C. et al, Communications of the ACM 1973.

[8] On cliques in graphs. Moon J. W. et al, Israel Journal of 1965.

[9] The worst-case time complexity for generating all maximal cliques and computational experiments. Tomita Etsuji et al, Theoretical Computer Science 2006.

评论

Thank you for taking time to respond to my questions. Some clarifications about my concerns -

  1. I still think the paper does not convincingly clarify the distinction between degree-bias studied in regular GNNs vs topology-bias in LightGCN like architectures studied in the context of recommender systems. The authors' rebuttal (A3) argues that the tasks (link prediction vs. node classification), debiasing goals, and supervision schemes are different. However, I think these distinctions are superficial since the underlying mechanism is identical: a message-passing scheme where update magnitude is directly influenced by node degree. The fact that this mechanism leads to over-recommending popular items in a bipartite graph (your "topology bias") is simply a different manifestation of the same fundamental issue that leads to poor performance on low-degree nodes in a classification task (the "degree bias").

  2. Further, the method uses a pre-trained model and then constructs the semantic graph based on the learned representations and then proceeds to lift this semantic graph to simplicial complexes which is great that it works but kind of presents a circular problem - if we want to tackle topology bias in recommender graphs, do we not want to start at the beginning, for instance during training itself [1] or normalization techniques [4] or virtual nodes [2]? I think the paper fails to convince me - Why is it preferable to first learn biased representations and then apply a computationally expensive post-processing step (lifting to simplicial complexes), rather than addressing the bias at its source?

评论

Thank you for your engagement during the discussion period. We appreciate your time and thorough assessment, and your valuable feedback will help us improve the manuscript. We hope the following points address your concerns.

Q1: "I still think the paper does not convincingly clarify the distinction between ..."

A1: We agree that the degree bias in node classification and the topology bias we study in recommender systems can be viewed collectively as the fundamental graph structure imbalance in different downstream graph tasks. We will include this similar line of work in the related work section to discuss the similarities and differences, which can indeed help to position our work. Additionally, we respectfully request that the reviewer gives greater consideration to the contributions presented in our manuscript. We would like to emphasize again that we are the first to study graph and simplicial complexes message passing under unified Dirichlet energy framework and introduce message passing on simplicial complexes to mitigate bias in recommender systems guided by our theoretical analysis.

Q2: "Further, the method uses a pre-trained model and then ..."

A2: We understand that unbiased training is an intuitive solution. Indeed, most existing methods adopt this approach, typically incorporating debiasing strategies based on data distribution, user preference analysis, and similar considerations. However, we argue that training a model from scratch is often prohibitively expensive. In practical recommendation scenarios, the recommender system may already be trained on large-scale datasets, some encompassing billions of interactions. Therefore, debiasing methods that require full retraining are frequently impractical, and training-free test-time solutions are significantly more practical and appealing. This is the starting point of our work and a key distinction from existing approaches.
Compared with training-based methods, our proposed TSP is indeed efficient. As shown in Table 3 of our paper, the one-time simplex lifting preprocessing for the ML10M dataset (with ~70k nodes and 5M edges) takes only 90 seconds. This is a small fraction of the 7-hour training time. Despite the worst-case complexity of simplex lifting, our TSP method achieves up to a 66% performance gain for tail nodes (Table 1) at less than 17% of the training time cost (Table 3), demonstrating its practical efficiency. In addition, another big advantage of our proposed method is that it is a plug-and-play module, demonstrating high flexibility and scalability. TSP can be integrated with various pre-trained graph recommenders without requiring further training or fine-tuning. We believe this model-agnostic nature makes our method widely applicable, allowing for efficient and flexible deployment across existing recommender systems.

评论

I would like to thank the authors for their prompt answers. This clarifies all of my doubts. I will be increasing my score. I would request the authors to kindly change the title to accurately reflect the paper's contributions and also discuss the related work on over-smoothing and discuss how fundamentally different/similar the current proposed approach is with respect to existing works as highlighted in my review.

评论

Thank you for your quick response and for your helpful suggestions. We are glad that we were able to answer your questions. We will modify the title and enhance the discussions of related work to better position our work.

审稿意见
5

The paper provides, for the first time in the literature, a theoretical analysis of topology bias in message passing for graph-based recommender systems. Unlike similar previous analyses, mainly focusing on the node embeddings and gradients levels, the authors investigate this phenomenon under the lens of the Dirichlet energy.

To begin with, the paper theoretically demonstrates that performing the message passing is conceptually equivalent to setting a Dirichlet energy minimization problem. Secondly, they demonstrate how message passing leads to having a larger accumulation of embedding norms for high-degree nodes in the graph, and this phenomenon keeps amplifying the deeper we go in the message passing.

On such a basis, the authors propose a novel approach to tackle this issue, which interestingly runs in the inference step of the model's pipeline, thus it can be applied to any graph-based recommender systems during inference. The proposed pipeline creates a novel graph whose node connections are not necessarily based upon the topological structure of the original graph (which brings the bias information), but rather on the similarity of node embeddings conditioned on a hyperparameter θ\theta. After that, the framework transitions to a simplicial complex (intra-simplex smoothing module) and a propagation function to propagate embeddings across simplicials (inter-simplex propagation module). Finally, a fusion strategy is adopted to combine all node representations from different orders of simplices.

Extensive experimental results when applying the proposed solution during the inference process of three graph-based recommender models, against other debiasing baselines from the literature, largely demonstrate the efficacy of the proposed approach. This is true also across some recommendation datasets. Moreover, an ablation study on the main modules shows the goodness of the architectural choices made by the authors. Finally, an embedding distribution analysis empirically proves how the distribution is less concentrated in the case of the proposed solution, while a scalability study analyzes how the introduction of the methodology adds a very limited overhead on the simple inference of the original models.

优缺点分析

Strengths

\bullet The paper has the merit to not providing yet another graph-based recommendation model, but rather to analyze a critical issue in the current literature, topological bias, and disentangle it from a different theoretical perspective than other analyses provided in the recent literature.

\bullet The paper is well-written and easy to follow.

\bullet The methodology is quite clear and sound, as it is supported quite nicely by the background section. All theorems seem adequately sound and are rigorously proven in the Supplementary Material section.

\bullet The proposed approach, which tackles the outlined issues, is nicely applied at inference time, thus not hurting the training of the models in any ways. The inference is also very limitedly hurt.

\bullet Experimental settings are extensive with many evaluated dimensions. The code is released at review time, thus ensuring the complete reproducibility of the code and results presented in the paper.

Weaknesses

\bullet The authors might have considered to include a further baseline (RP3Beta [*]) which comes from a less-recent literature but works quite well on the user-item graph through random-walk and it is especially designed to not penalize less popular items. I think including this baseline as a standalone model to be compared against the other would provide a very nice complementary evaluation to the paper.

\bullet In relation to the above, I think the authors might also consider to measure other recommendation measures which are usually adopted to assess the diversity in the recommendation lists [**] and the presence of items from the long-tail [***].

References

[*] Christoffel et al., Blockbusters and Wallflowers: Accurate, Diverse, and Scalable Recommendations with Random Walks. RecSys 2015.

[**] S. Vargas, P. Castells, Rank and relevance in novelty and diversity metrics for recommender systems, in: RecSys, ACM, 2011, pp. 109–116.

[***] Himan Abdollahpouri, Robin Burke, and Bamshad Mobasher. 2017. Controlling Popularity Bias in Learning-to-Rank Recommendation. In RecSys. ACM, 42–46.

问题

My main questions are related to the two weaknesses I outlined above.

Q1) Could the authors provide additional results when running the model RP3beta on the tested datasets?

Q2) How are the recommendation performance when considering additional standard bias and fairness metrics in the literature, such as the expected free discovery (EFD) and the average percentage of long-tail items (APLT)?

Q3) Do the authors have any intuition on how their theoretical analysis and subsequent methodology could be differently applied to the user side and the item side? In principle, the two sides are different and show different properties, also in the topological bias (i.e., low-high activeness on the platform for the user vs. popular/niche products for the items).

局限性

Yes.

最终评判理由

After the rebuttal phase, after reading the other reviews, and authors' responses to those, I decide to confirm my initial score I gave to the paper.

Specifically, all my concerns have been discussed and addressed by the authors.

Whatever the final outcome for this paper, I think it will benefit quite a lot from the fruitful discussion we had along with the authors, the reviewers, and the Area Chair. I believe the paper will improve its quality a lot in any case.

格式问题

I do not see any major formatting issues with this paper.

评论

W1: "The authors might have considered to include a further baseline (RP3Beta ..." Q1: "Could the authors provide additional results when running the model RP3beta ..."

A1: We expand our experiments by adding the new baseline RP3beta compared with our method. The results of tail performance are listed below:

ModelsAdressa R@20Adressa N@20Gowalla R@20Gowalla N@20Yelp2018 R@20Yelp2018 N@20ML10M R@20ML10M N@20Globo R@20Globo N@20
PR3beta0.0260.0090.0130.0090.0060.0040.0050.0020.0130.007
LightGCN+TSP0.0550.0240.0380.0210.0090.0080.0130.0080.0240.013
LightGCL+TSP0.0750.0470.0500.0320.0200.0190.0140.0100.0320.018
SimGCL+TSP0.0850.0460.0550.0470.0150.0140.0160.0120.0250.017

These additional result further present the strong performance of our proposed TSP method.

W2: "In relation to the above, I think the authors might also consider to measure other recommendation measures ..." Q2: "How are the recommendation performance when considering additional standard bias and fairness metrics ..."

A2: Thank you for your suggestions. We agree that the metrics mentioned can further present the debiasing ability of our proposed method. Due to limited rebuttal time we only evaluate methods on LightGCN backbone. We show these additional results of EFD (with exponential discount 0.85k10.85^{k–1}) and APLT of below:

ModelsAdressa EFD@20Adressa APT@20Gowalla EFD@20Gowalla APT@20Yelp2018 EFD@20Yelp2018 APT@20ML10M EFD@20ML10M APT@20Globo EFD@20Globo APT@20
PR3beta0.8750.1250.8400.1420.8040.0950.7410.0720.7550.068
LightGCN0.8230.1240.7910.1310.7520.0850.7150.0720.7240.064
LightGCN+IPS0.8310.1330.8010.1560.7640.1010.7320.0800.7140.073
LightGCN+CausE0.8460.1580.8310.1570.7930.1050.7630.0930.7350.079
LightGCN+MACR0.9010.1470.8560.1500.8140.1180.7940.1070.7560.121
LightGCN+SDRO0.8320.1360.7810.1380.7350.1100.7050.0970.7210.112
LightGCN+TSP0.9230.2170.8750.1640.8130.1330.8020.1160.7730.149

The constant better performance of TSP in fairness metrics further prove its ability to mitigate the topology bias and produce balanced recommendation.

Q3: "Do the authors have any intuition on how their theoretical analysis and subsequent ..."

A3: This is a very insightful question. In our current framework, we treat users and items symmetrically when constructing the semantic graph. However, the framework is flexible enough to handle them asymmetrically. For instance, one could use different similarity thresholds (θ\theta) for user-user and item-item pairs to reflect the different semantics and densities of these relationships. It would also be possible to construct simplices only on the item side to specifically focus on diversifying item representations. We believe this is a fascinating and promising direction for future research.

评论

Dear Authors,

Thanks for your time in answering to my raised weaknesses and questions.

Your additional experiments with RP3Beta, as well as the diversity and popularity bias metrics, provide further empirical justifications to the goodness of your proposed approach.

Moreover, I believe the intuitions you gave about how you could modify the formulation on the two sides (user and item) is quite nice and might represent an important avenue for future directions of your work.

I do not have any other concerns regarding the paper. The answers you gave helped confirming the initial positive judgment I had on the paper. I will keep my score, which I believe is already quite high and mirrors the quality of your submission.

评论

Thanks for acknowledging our response and for your continued support!

审稿意见
4

This paper studied graph-based recommender methods. The authors started by analyzing the common popularity bias issue in the graph-based methods, from a Dirichlet energy perspective. Combined with the theoretical analysis, the authors proposed TSP, a training framework for existing graph-based recommender methods to alleviate the effect of popularity bias. The proposed TSP contains multiple steps, such as the semantic graph reconstruction, which leverages a pretrained model to add in predicted edges, similar to graph structure learning, and a few following steps as detailed in Sec. 4.2.

优缺点分析

Strengths:

  1. Popularity bias is a very important problem in collaborative filtering, which may even be enlarged in graph-based methods due to message passing. I'm glad to see papers trying to solve this critical issue.
  2. This paper is well supported by theoretical analysis.

Weakness:

  1. Graph-based (or message passing-based) recommender methods are often criticized by their limitations on efficiency when it comes to larger graphs [1]. While the authors has shown that the proposed TSP added limited overhead to lightGCN, considering that TSP has steps such as calculating the graph laplacian, I'm concerned on the efficiency of the proposed method.
  2. Additional experiments on larger scale datasets would be appreciated.

[1] UltraGCN: Ultra Simplification of Graph Convolutional Networks for Recommendation, CIKM 2021

问题

  1. The term "semantic graph" is quite confusing. IIUC, the graph-based collaborative filtering methods doesn't use any semantic features, but instead learn ID embeddings for each nodes. The way TSP reconstructs the semantic graph is very similar to existing literature in graph structure learning and graph data augmentation, which were not discussed in this work. Hence I wonder why the reconstructed graph is named "semantic graph" when it's more of a "augmented graph"?

局限性

see weakness

最终评判理由

I appreciate the authors' rebuttal, which is very helpful with additional clarifications. So I raised my score.

格式问题

n/a

评论

W1: "Graph-based (or message passing-based) recommender methods are ..."

A1: We agree that graph recommenders are often considered less efficient when trained on large graphs. However, our proposed TSP is designed as a test-time module that operates on a pre-trained model. The computational overhead of our TSP method, including the Hodge Laplacian calculation, does not affect the training efficiency of the underlying recommendation model. This one-time computation during test phase is a key advantage of our approach.

Furthermore, the additional computational overhead introduced by our method is quite small. We have included a scalability analysis in Section 5.4 of our original paper, which demonstrates that our proposed TSP adds only a limited and acceptable overhead to the backbone models during inference time across datasets of varying sizes, as shown in Table 3.

W2: "Additional experiments on larger scale datasets would be appreciated."

A2: We thank the reviewer for this suggestion and agree that scalability is a crucial aspect of graph-based models. While the definition of a "large-scale" dataset can vary, we have selected datasets that are considered substantial within the context of recent research. For example, the work cited by the reviewer [1] evaluates GCNs on datasets with up to around 100k nodes. Our experiments include datasets of comparable or greater scale, ranging from Gowalla (~14k nodes, ~100k interactions) to Globo (~160k nodes, ~2.5M interactions). We believe these results can already demonstrate our model's scalability across different dataset scales. Nevertheless, we are committed to providing a thorough empirical analysis. To best address the reviewer's concern, we would be grateful for any clarification on the specific scale or type of dataset that would be considered sufficient for a more comprehensive evaluation.

Q: "The term 'semantic graph' is quite confusing. IIUC, the graph-based ..."

A3: We use the term "semantic graph" to emphasize that its structure is derived from the semantic similarity of node embeddings, which capture latent relationships (e.g., user preferences, item attributes) learned by the GNN, as opposed to the raw, explicit user-item interaction topology. This new topology is more reflective of the underlying semantic meaning of the nodes. We agree that this process can be considered as a form of graph augmentation or structure learning. However, our goal is not merely to augment the graph but to construct a new topology specifically guided by our Dirichlet energy analysis to mitigate topology bias.

References

[1] UltraGCN: Ultra Simplification of Graph Convolutional Networks for Recommendation, CIKM 2021

评论

I appreciate the authors' rebuttal, which is very helpful with additional clarifications. Nevertheless, I don't think W2 and Q1 are well-addressed. For example, it doesn't really make sense to me to measure any "semantic similarity" between ID embeddings, which learns from collaborative signals.

评论

Thank you for your feedback.

We agree that the name of the term "semantic graph" can cause certain confusion. In research fields like multi-modal recommendation, "semantic" typically refers to side information (e.g., text or images) rather than collaborative signals. This may lead to the assumption that our "semantic graph" should be built from such external data. In our work, we use this term to differentiate the latent relationships we want to uncover between nodes from the explicit connections in the existing interaction graph. Enhancing graph structures with learned embeddings or networks is a well-established practice in various domains [1-3]. Therefore, given our goal of a flexible, training-free solution, leveraging similar method by using the similarity of pre-trained embeddings as structural hints is an intuitive approach, which is also shown to be effective in our experiments. However, to avoid ambiguity, we will revise the manuscript and adopt the more precise term "latent graph" or "latent connections".

Furthermore, to address your concerns and further validate our method, we have conducted additional experiments on the QK-video dataset from the Tenrec benchmark [4]. This is a large-scale, real-world dataset with 928,562 users, 1,189,341 items, and 37,823,609 interactions. With no specific statistics or dataset reference provided, we choose this industrial dataset which is more than 10 times larger than the largest dataset of our original experiments. The results of model performance and computational overhead are listed below. Due to the time constraints of the rebuttal period, we present results with only LightGCN as the backbone. We will include the full experimental results in the final version of our manuscript.

ModelsTail Recall@20Tail NDCG@20overall Recall@20Overall NDCG@20
LightGCN0.0010.0020.0630.049
LightGCN+IPS0.0080.0110.0690.057
LightGCN+CausE0.0030.0030.0610.050
LightGCN+MACR0.0060.0100.0810.069
LightGCN+SDRO0.0040.0060.0700.053
LightGCN+TSP0.0100.0170.0810.073
LightGCN training timeTSP preprocessing timeLightGCN inference timeTSP inference time
39h4h0.853s0.881s

We believe these additional results align with the conclusions in our original paper and further demonstrate the effectiveness and efficiency of our proposed TSP model on a large-scale, real-world dataset.

References

[1] Hierarchical Graph Convolutional Networks With Latent Structure Learning for Mechanical Fault Diagnosis. Kai Zhong et al, 2023.

[2] Neural Common Neighbor with Completion for Link Prediction. Xiyuan Wang et al, ICLR 2024.

[3] Locality-Aware Graph Rewiring in GNNs. Federico Barbero et al, ICLR 2024.

[4] Tenrec: A Large-scale Multipurpose Benchmark Dataset for Recommender Systems. Guanghu Yuan et al, NeurIPS 2022.

评论

Dear Reviewer HkJG, It would be useful if you could respond to the rebuttal of the authors to indicate if your concerns are satisfactorily answered.

sincerely AC

最终决定

The paper addresses the problem of popularity bias—over-representation of popular items—in graph-based recommendation systems. It offers a fresh perspective by analyzing how graph topology influences message passing through the lens of Dirichlet energy. While the new ideas were appreciated, reviewers raised several concerns, particularly: How the use of Dirichlet energy meaningfully differs from existing work on oversmoothing. Evaluation of effectiveness on the long-tail of items. Scalability of the proposed approach. In response, the authors conducted additional experiments, which significantly strengthened the empirical section of the paper, leading to higher scores from some reviewers. However, during the discussion phase, the paper lacked a strong advocate. The ambivalence stemmed mainly from the perception that the theoretical contributions were modest. This impression was reinforced by the absence, in the main manuscript, of a clear summary of how the Dirichlet energy perspective connects to prior research on Graph Neural Networks. Although the rebuttal partially addressed this, questions about the extent of novelty remained. Since a central motivation of the paper is to mitigate popularity bias, additional empirical evidence demonstrating improvements on long-tail recommendations would have been particularly impactful. Encouragingly, in the rebuttal, the authors presented results on one such dataset (as suggested by a reviewer), and the reviewer appreciated the results.

In summary, it will be very difficult to make a strong recommendation for the paper. All the acceptance was contingent on the fact that substantive revision needs to be done. In case the paper is accepted the authors should mandatorily implement the changes discussed during rebuttal.