PaperHub
4.8
/10
Poster3 位审稿人
最低2最高3标准差0.5
3
3
2
ICML 2025

Pruning for GNNs: Lower Complexity with Comparable Expressiveness

OpenReviewPDF
提交: 2025-01-23更新: 2025-07-24
TL;DR

We prune redundant structures in MP, K-path and K-hop GNNs, achieving lower complexity while preserving expressive power.

摘要

关键词
Graph Neural NetworkPruningExpressivenessComplexity

评审与讨论

审稿意见
3

In this paper, the author proposed pruned MPNNs, K-path GNNs, and K-hop GNNs to reduce the computation redundancy in the original method. The author proved the expressive equivalence between the original version and the pruned version. The proposed method is evaluated on various benchmarking datasets and shows comparable results to the original version.

Update after rebuttal

I would like to thank the authors for the detailed response to my questions and concerns. The author provides both a textual explanation for the complexity analysis of their method and the experimental comparison result. This helps to solve my biggest concern about the proposed method. I would like to increase my original score.

给作者的问题

See above.

论据与证据

  1. In the title, the authors claim that the pruned GNNs have higher expressiveness. However, I think it’s kind of misleading, as the expressive power of pruned GNNs is bounded by the original version based on the proofs provided in the paper. It is only safe to say that given the limited number of layers/iteration, the pruned version is better than the original one.
  2. In Table 5, the authors claim that the pruned version has better complexity than the original version. However, I am not convinced. First, I do not understand why the space complexity is associated with the distance LL. Since for both MPNN, K-path GNNs, and K-hop GNNs or their pruned version, we only need to save embedding for each node in the graph, which results in O(n)O(n). For time complexity, it’s also not correct, as we need to consider the density of the graph. Let’s use the average degree dd to characterize. For MPNNs, it should be O(ndL)O(ndL). For PR MPNNs, it seems to me that when we need to aggregate higher order neighbors, the complexity of the aggregation will go up, which means that the complexity is different for different layers. The first layer is just ndnd. For the second layer, we will aggregate neighbors from the second hop, which results in nd2nd^2. Overall, it should be O(log(L)ind2i1)O(\sum_{log(L)}^{i}nd^{2^{i-1}}) . Similarly, the analysis for K-path and K-hop GNNs is also not correct to me. Given my understanding of the complexity part, I don’t think the proposed pruned framework will result in huge computation improvement over the original version. Even worse, the pruned version requires aggregation of different neighbors for different layers, which introduces additional pre-processing time and is not flexible for adjusting the number of layers.
  3. The author mentioned that fewer layers can reduce the nonlinearity of GNNs and reduce over-smoothing. However, it is a problem primarily for node tasks. But the proposed methods are mainly used in Graph-level tasks.

方法与评估标准

  1. As mentioned above, although the proposed method requires fewer layers to achieve certain expressive power, the complexity is varied and much higher than the pruned version for each layer. Therefore, it is hard to see that the proposed method is more efficient.
  2. The pruning strategy seems totally different for different GNNs. Are there any common criteria for pruning so that we can apply a similar strategy to other GNNs?
  3. Even if the proposed method does introduce efficiency due to the fewer number of layers. The application is limited. Specifically, the advantage is log(L)log(L) over LL. However, real-world graph tasks are usually molecules, whose LL are usually less than 10. Therefore, it seems not a big improvement to me.

理论论述

The complexity analysis seems wrong to me. See above. Given the limited time, I cannot check the proofs of all theorems. All other conclusions seem correct to me.

实验设计与分析

  1. I believe a more comprehensive ablation study on the complexity is necessary for showing the potential of the proposed methods. Specifically, I would like to see the comparison of the original version and the pruned version on both pre-process, train, and inference time under different numbers of layers and datasets (different graph distribution). Meanwhile, the cost of space should also be evaluated to support Table 5.
  2. I would suggest another ablation study between the number of layers and the performance using both the original version and the pruned version using both the simulation data and real-world data to see if the pruned version really has better performance, especially under a imited number of layers.

补充材料

I did not see any supplementary material, and there is no code provided.

与现有文献的关系

If the method really has a computation advantage over the existing method, it can be used in the field of molecular or biomedical research.

遗漏的重要参考文献

The topic is related, and the experiment protocol is almost identical to the existing paper [1], but it is not cited, discussed, and compared.

[1] Jiarui Feng, et al., How powerful are K-hop message passing graph neural networks, NeurIPS22.

其他优缺点

See above.

其他意见或建议

See above.

作者回复

We sincerely thank all the reviewers for their feedback and constructive comments

Claims And Evidence:

(2)Space Complexity:During backpropagation in training, the node representations of intermediate layers (e.g., activation values) are required to compute gradients and update weights. If intermediate node representations are discarded, backpropagation cannot proceed. Therefore node representations from each layer need to be cached, which results in O(nL)O(n·L).

Time Complexity:We speculate the reviewer might draw conclusion:log(L)iO(nd2i1)\sum^i_{log(L)}O(nd^{2^{i-1}}) due to the divergence in computational paradigms. Suppose aggreation method as sum. In the reviewer’s view, the aggregation of ll times is computed as follow:(1)Compute the ll-th power of the adjacency matrix AlA^l.(2)Calculate the product of AlA^l and node representations hh as AlhA^l·h. This computation method involves calculating the product between matrices L times, resulting in extremely high complexity (even if A is a sparse matrix).

However in our approch (mentioned in Equation (15) on page 5), the aggregation is decomposed into ll times of 1-neighborhood aggregation. In other word, AlhA^l·h is computed as:

forfor ii inin [l][l] dodo: hi=Ahi1h_i=A·h_{i-1}

Our approch will only compute the product between matrice and node representations vector. Thus, it achieves lower computational complexity compared to the approach above. Therefore the complexity of aggregation is log(L)iO(2i1nd)=O(ndL)\sum^i_{log(L)}O(2^{i-1}nd)=O(ndL). We provided pruned architectures better visualized in the link.

As for reviewer's concern for complexity, we decompose the computational time complexity into two components: (1) feature aggregation(Oagg(1)O_{agg}(1) refers to one basic aggreation for a node) and (2) the MLP operations(OMLP(1)O_{MLP}(1) refer one layer's MLP) and space complexity (Ospa(1)O_{spa}(1) refers to one node's represetations). Table 5 can be modified into the following table(we assume K<<L):

Training ComplexityMP-GNNPR MP-GNNK-HOPPR K-HOPK-PATHPR K-PATH
TimeOagg(nL)+OMLP(L)O_{agg}(nL)+O_{MLP}(L)Oagg(nL)+OMLP(log(L))O_{agg}(nL)+O_{MLP}(log(L))Oagg(nL)+OMLP(L/K)O_{agg}(nL)+O_{MLP}(L/K)Oagg(nL/K)+OMLP(L/K)O_{agg}(nL/K)+O_{MLP}(L/K)Oagg(nL)+OMLP(L/K)O_{agg}(nL)+O_{MLP}(L/K)Oagg(nL/K)+OMLP(L/K)O_{agg}(nL/K)+O_{MLP}(L/K)
SpaceOspa(nL)O_{spa}(nL) Ospa(nlog(L))O_{spa}(nlog(L)) Ospa(nL/K)O_{spa}(nL/K) Ospa(nL/K)O_{spa}(nL/K) Ospa(nL/K)O_{spa}(nL/K) Ospa(nL/K)O_{spa}(nL/K)

On the other side, in our ablation experiment, the time consumed by aggregation is not significant.

As for adjusting the number of layers, ala_l doesn's have to be 2l12^{l-1}, we point out 2l12^{l-1} because it will reach its maximum expressiveness. Theorem4.1has shown as long as it is viewable, it will be as powerful as WL. Hence we can flexibly adjust ala_l as geometric sequence(al=la_l=l) or other viewable sequence.

(1):The Initial intention of the title derives from the comparasion of MP-GNN and PR K-PATH GNN instead of MP-GNN and PR MP-GNN: from(2), if K<<L, the complexity of PR K-PATH GNN is less than MP-GNN, while PR K-PATH is as powerful as K-PATH and K-PATH is stirctly more powerful than MP-GNN. We sincerely appreciate the reviewer's insightful comment regarding the potential ambiguity in the paper's title. In response to the reviewer's suggestion, we would like to modify the title to "PRUNING For GNNs: LOWER COMPLEXITY WITH COMPARABLE EXPRESSIVENESS" to better reflect the study's focus.

(3)The pruned methods can also be used in Node-level tasks, since the graph representation is a aggreation of node representations. In other word, given a Pruned MP-GNN MM' as powerful as arbitrary MP-GNN MM, if for any two nodes they get same representation MM it will also get same representation in MM'.

Methods And Evaluation Criteria:

(1)It has been mentioned above

(2) The pruning strategy seems different because MP-GNN is one-aggreation while K-HOP\PATH is Multi-aggregation. Our pruned method for MP-GNN is suitable for any one-aggreation GNN, but pruned method for Multi-aggregation is associations between different aggregated neighbor nodes (can not guarantee the expressive equivalence).

(3)The reason why GNNs typically aggregate node information within a distance L≤10 is that larger L values lead to issues such as excessive parameter size (prone to overfitting on small datasets) and high computational/memory costs. However, our pruning method reduces the parameters to a logarithmic scale (log(L)),thereby potentially enabling GNNs to aggregate information from more distant nodes

Experimental Designs Or Analyses:We have supplemented the ablation experiments on pruning efficiency and conducted experiments on the large-scale graph OGB-arXiv. The experimental conclusion can be found in our response to other reviewers. All the experimental materials(code), results, the better visualized pruned architectures and responses to other questions are provided in https://anonymous.4open.science/r/PrunedGNN-AC61/README.md

审稿人评论

I would like to thank the authors for the detailed clarification and response. Most of my concerns have been addressed. I will increase my score accordingly. Please include all the additional results in the future version.

作者评论

Thank you sincerely for your generous feedback and for raising your evaluation of my work. I greatly appreciate the time and thought you invested in reviewing my manuscript and offering such valuable suggestions. Your support is truly meaningful, and I’ve found the revision process very rewarding thanks to your insights.

审稿意见
3

The paper proposes a pruning framework for GNNs aimed at improving computational efficiency while maintaining or even enhancing expressive power. The authors introduce pruned versions of Message Passing GNNs (MP-GNNs), K-Path GNNs, and K-Hop GNNs by identifying and removing redundant structures. Theoretical analysis using MATLANG demonstrates that pruned versions retain the expressive power of their unpruned counterparts, and experiments on multiple datasets validate the efficiency gains. The main contributions include:

  • A theoretical justification for pruning in GNNs without sacrificing expressive power.
  • A pruned message passing framework that maintains equivalent expressiveness while reducing complexity.
  • Empirical validation showing that pruned frameworks outperform or match unpruned models across benchmark datasets with lower computational cost.

给作者的问题

  1. Are there any known cases where pruned K-Hop fails to distinguish graphs that the original K-Hop framework can differentiate? Providing examples would help clarify the limitations.
  2. The claim that pruning redundant structures enhances efficiency without compromising expressiveness is central to the paper. Could the authors conduct an ablation study comparing different pruning strategies to determine whether all identified redundant structures should be removed?
  3. Have the authors tested how pruning affects GNNs trained on very large graphs (e.g., OGB datasets)? If so, do the training and inference times scale as expected?
  4. The experimental results validate the effectiveness of pruned frameworks but do not directly measure expressiveness beyond isomorphism-based tasks. Could the authors evaluate the frameworks on tasks that require high expressiveness, such as molecular property prediction with long-range dependencies?
  5. Could there be datasets where pruning negatively impacts expressiveness? If so, an analysis of when pruning is beneficial versus detrimental would strengthen the conclusions.

论据与证据

The authors claim that:

  • Pruning redundant structures does not reduce expressive power - This is well-supported by theoretical analysis using MATLANG and empirical validation on graph isomorphism tasks.
  • Pruned models have lower computational complexity - The paper provides clear evidence through algorithmic complexity analysis and runtime measurements.
  • Pruned K-Hop and K-Path GNNs distinguish more non-isomorphic graphs than MP-GNNs - While theoretically sound, this claim could benefit from more empirical results on diverse graph structures.
  • Pruning improves training efficiency and scalability - The results support this claim, but more large-scale graph evaluations (e.g., OGB datasets) would strengthen it.

Overall, the claims are mostly well-supported, but additional empirical validation, particularly on large-scale datasets, would reinforce the conclusions.

方法与评估标准

The methodology is well-structured, with clear mathematical definitions and derivations. The pruning strategies are rigorously developed, and the expressiveness analysis is grounded in matrix algebra. The evaluation criteria include:

  • Expressiveness tests: Distinguishing non-isomorphic graphs.
  • Graph property prediction: Evaluating node and graph features.
  • Real-world benchmarks: TU datasets, QM9, and ZINC. The chosen benchmarks and evaluation metrics are appropriate, though additional experiments on larger-scale datasets would provide further validation.

理论论述

The paper provides several key theoretical results:

  • Proof that pruned MP-GNNs retain 1-WL equivalence - The proof is logically structured and appears correct.
  • Equivalence of pruned K-Path GNNs to standard K-Path GNNs - This is rigorously shown via MATLANG.
  • Pruned K-Hop GNNs retain equivalence for distinguishing regular graphs - The proof is incomplete due to lost structural information.
  • Computational complexity analysis - The theoretical derivations appear sound.

实验设计与分析

The experiments are well-designed but could be improved in a few areas:

  • Expressiveness experiments: The graph isomorphism tests effectively demonstrate the theoretical claims.
  • Graph property prediction: Results on the TU datasets and QM9 confirm that pruning does not degrade performance.
  • Computational efficiency: The paper successfully demonstrates reduced parameter count and runtime improvements.

However, experiments on larger graphs (e.g., OGB datasets) and ablation studies on different pruning strategies would strengthen the empirical claims.

补充材料

The supplementary material was not extensively reviewed, but it includes:

  • Detailed proofs for theoretical claims.
  • Extended experiments and hyperparameter details.
  • MATLANG derivations used in expressiveness proofs.

Reviewing the correctness of all proofs would require additional time, but the main arguments seem sound.

与现有文献的关系

The paper is well-grounded in the existing literature on GNN expressiveness and efficiency:

  • Expressiveness limits of MP-GNNs (Xu et al., 2019; Morris et al., 2019) - The pruning approach aligns with previous findings that message-passing GNNs are constrained by 1-WL limitations.
  • Higher-order GNNs (Maron et al., 2019; Azizian & Lelarge, 2021) - The authors position their work as a computationally efficient alternative to higher-order approaches.
  • Subgraph-based methods (Bevilacqua et al., 2022; Zhao et al., 2022) - The paper could better contrast pruning with subgraph-based improvements in expressiveness.

遗漏的重要参考文献

The following studies could be cited, as they present different pruning techniques that can be compared against the proposed approach in the paper:

[1] Dupty et al., PF-GNN: Differentiable particle filtering based approximation of universal graph representations, ICLR 2022.

[2] Fey et al., GNNAutoScale: Scalable and Expressive Graph Neural Networks via Historical Embeddings, ICML 2021.

[3] Müller et al., GraphChef: Decision-Tree Recipes to Explain Graph Neural Networks, ICLR 2024.

其他优缺点

Strengths

  • The paper introduces a novel pruning strategy for GNNs, which reduces computational complexity while preserving expressive power. Unlike existing methods that enhance expressiveness by increasing depth or complexity, this approach optimizes efficiency without sacrificing performance.
  • The paper is well-structured, with clear theoretical justifications, proofs, and empirical results supporting the claims.
  • The extensive experimentation across synthetic and real-world datasets strengthens the findings, particularly for efficiency improvements.
  • If widely adopted, the proposed pruning strategies could make expressive GNNs more computationally feasible for large-scale applications, such as molecular modeling.

Weaknesses

  • While the theoretical analysis supports the expressiveness claims, the experimental validation relies mostly on synthetic datasets and indirect comparisons (e.g., isomorphism tests). Direct comparisons on real-world tasks requiring high expressiveness (e.g., molecular property prediction) would strengthen the argument. A more comprehensive ablation study showing how different pruning levels affect expressiveness would be useful.
  • The efficiency claims are theoretically sound, but the experimental validation on large-scale graphs is somewhat lacking. How well do the pruned frameworks generalize to datasets with millions of nodes and edges?
  • In particular, K-Hop pruning loses information about non-shortest paths. While the authors argue that this is not critical, further empirical evidence would help support this claim.

其他意见或建议

Please refer to the weaknesses mentioned above.

作者回复

We sincerely acknowledge reviewer V1ZW for the constructive criticism and insightful suggestions.

Questions For Authors:

Q1: Unfortulately, we have to admit that find a pair of graphs that pruned K-Hop fails to distinguish graphs and the original K-Hop framework can differentiate involves a tremendous amount of algebraic analytic work. Since the proof of the theorem of the existance of a pair of graphs distinguishable by the (k+1)-dimensional Weisfeiler-Lehman ((k+1)-WL) test but not by the k-dimensional Weisfeiler-Lehman (k-WL) test spans 58 pages. However, even if it exists, we can draw the conclusion that the probability for inconsistent expressiveness between K-Hop and pruned K-Hop goes to 0 as the size of graphs n goes to infinity. As shown in the respond to Q2, we also conducted experiments on this issue real-world and synthetic datasets, all of the results show that they have same expressiveness.

Q2: We have added experiments to verify the consistency of the expressive power between the pruned architecture and the original algorithm on both real and synthetic datasets. The experimental results show that the expressive power of the pruned architecture is identical to that of the original algorithm. We also conducted ablation experiments for GNNs, result shows the pruned version achieves significant reductions in both parameter count and training duration, showing noticeable efficiency improvements.

We have supplemented ablation experiments analyzing the impact of different pruning strategies on model efficiency and accuracy. We summarize the core conclusion:(1)Regarding to the expressiveness(Graph Isomorphism WL test), all identified redundant structures should be removed since it will enhance efficiency.(2)When it comes to the performances(accuracy), sometimes retaining certain redundant structures can slightly improve the model's accuracy. The reason is that if all identified redundant structures are removed, the subsequent layers will need to aggregate a large amount of representations. Since MLP are not strict hash functions, this can degrade the model's performance.

As for pruning strategies for MP-GNNs, we point out that the sequence ala_l can be flexibly adjusted, as long as ala_l is viewable(The subset's sum of {ala_l} is dense in [Sl][S_l]), such as geometric sequence(al=la_l=l) or fibonacci sequence(al=al1+al2a_l=a_{l-1}+a_{l-2}), then it have same expressiveness as MP-GNN.

Q3:We have tested how pruning affects GNNs trained on very large graphs on dataset obg-arvix. The result shows that while maintaining comparable accuracy to the original model, the pruned version achieves significant reductions in both parameter count and training duration, showing noticeable efficiency improvements.

Q4:As shown in Q2, we have conducted WL Test on both real and synthetic datasets. For the pruned WL algorithm, it is considered accurate only when its graph output results are entirely consistent with the original algorithm. This requires high expressiveness. Meanwhile, experimental results demonstrate a significant improvement in the efficiency of the pruned algorithm.

Q5:During the experiments, the performance for 4 layers [1,2,3,4] Pruned GIN often performs worse than 3 layers [1,2,3] Pruned GIN, We find out the reason is, when the layer gets deeper,a substantial number of neighbor representations will be repeatedly aggregated. This significantly impedes the model's ability to extract useful information. Consequently, we refined the model to eliminate redundant representation aggregation, thereby enhancing Pruned GIN's performance.

Weaknesses:In particular, K-Hop pruning loses information about non-shortest paths. While the authors argue that this is not critical, further empirical evidence would help support this claim.

Our original intention was to point out that, compared to the K-Path framework, K-Hop loses information about non-shortest paths. As a result, we cannot prove that the pruned K-Hop is equivalent in expressive power to the original model. However, this does not have a significant impact. Firstly, we demonstrate the equivalence between the pruned K-Hop and the original model for regular graphs and strongly regular graphs—while the K-Hop model was specifically designed to address the inability of MP-GNNs to distinguish regular graphs. Secondly, in our ablation experiments, the pruned K-Hop exhibits equivalent expressive power to the original model for both real-world and Synthetic experiments.

Compared to K-Path, K-Hop indeed is less powerful than K-Path, since K-Path asigned more number of classes during the WL expressiveness experiment. All the experimental materials (code), results and responses to References in provided in https://anonymous.4open.science/r/PrunedGNN-AC61/README.md

Thanks again, for the reviewer's detailed analysis and questions, which allowed me to further refine the model in the ogbn-arxiv experiments.

审稿人评论

Thank you for your considerate answers to my questions. I truly appreciate the time and effort you took to address my concerns. I will be keeping my original score.

作者评论

I’m very grateful for your thoughtful comments. Your feedback significantly contributed to refining the manuscript, and I truly value your support throughout the revision process.

审稿意见
2

This paper proposes pruned versions of Message Passing GNNs, K-Hop GNNs, and K-Path GNNs by eliminating redundant structures. The authors claim that these pruned frameworks maintain or even improve expressive power while reducing computational complexity. Theoretical analysis based on matrix language is used to demonstrate equivalence in expressiveness between the pruned and original frameworks. Additionally, experimental results on benchmark datasets show that pruned GNNs achieve comparable or better performance with improved efficiency.

给作者的问题

Have you tested the approach on large-scale graphs where efficiency improvements might be more noticeable?

论据与证据

The paper claims that pruning redundant structures in MP-GNNs, K-Hop GNNs, and K-Path GNNs maintains expressiveness while reducing complexity. While some theoretical arguments support this claim, the practical significance is not well justified.

方法与评估标准

The methodology is theoretically sound in using matrix language tools to analyze expressiveness. However, the evaluation mainly focuses on standard benchmark datasets without strong ablation studies or comparisons with more sophisticated recent baselines.

理论论述

The theoretical analysis claims that pruned GNNs maintain the expressive power of the original architectures.

实验设计与分析

The experimental setup evaluates pruned GNNs on several benchmark datasets. However, the improvements in efficiency are marginal, and the comparisons do not convincingly demonstrate practical benefits over existing optimized GNN models. I do not think the TUdatasets are enough to evaluate the method. The authors should consider OGB datasets for evaluations.

补充材料

I read the supplementary material.

与现有文献的关系

It improves the expressiveness of GNNs (broader literature) while reducing complexity.

遗漏的重要参考文献

No significant gaps found.

其他优缺点

The theoretical analysis provides an interesting perspective on expressiveness equivalence. But the empirical evaluations are limited.

其他意见或建议

Consider adding experiments on larger, real-world datasets to better illustrate the efficiency gains.

作者回复

We thank reviewer hERx careful evaluation and meaningful suggestions.

Q1:The methodology is theoretically sound in using matrix language tools to analyze expressiveness. However, the evaluation mainly focuses on standard benchmark datasets without strong ablation studies or comparisons with more sophisticated recent baselines.

To verify the expressive power equivalence between the pruned WL test and their original algorithms, We have added experiments to verify the consistency of the expressive power between the pruned architecture and the original algorithm on both real and synthetic datasets. The experimental results show that the expressive power of the pruned architecture is almost identical to that of the original algorithm.

Additionally, we have evaluated the efficiency improvements of the pruned architecture compared to the original architecture in the WL algorithm and as a GNN model. The results show that under the WL algorithm, the pruned architecture maintains the same expressive power as the original algorithm while significantly reducing computation time. As for GNN, We compared the pruned model in terms of accuracy and training time, and the results show that the pruned models achieve comparable accuracy to the original model, while most pruned variants demonstrate better training efficiency than the baseline. Some of the results are shown below.

ModelCOLLAB TimeCOLLAB Acc (%)NCI1 TimeNCI1 Acc (%)IMDB-B TimeIMDB-B Acc (%)IMDB-M TimeIMDB-M Acc (%)MUTAG TimeMUTAG Acc (%)PROTEINS TimePROTEINS Acc (%)
GIN(3)1.10474.8 ± 1.30.48071.9 ± 0.50.25171.9 ± 0.30.30449.9 ± 0.00.88989.4 ± 0.40.26873.7 ± 0.7
PR GIN(1)1.06073.9 ± 0.00.46172.9 ± 1.40.20969.9 ± 2.00.28450.6 ± 0.30.88688.5 ± 0.00.23372.2 ± 1.9
GIN(7)1.63877.4 ± 1.60.74871.5 ± 1.40.57872.6 ± 0.30.53451.1 ± 0.30.90489.4 ± 1.00.52776.3 ± 0.2
PR GIN(124)1.28476.4 ± 0.70.696175.4 ± 0.20.48171.7 ± 1.40.46452.0 ± 0.50.91692.0 ± 0.40.42574.1 ± 1.0
GIN(10)2.14274.7 ± 0.61.12275.9 ± 1.30.94872.1 ± 2.80.89849.7 ± 1.50.87487.7 ± 0.20.86772.3 ± 0.0
PR GIN(1234)1.68975.6 ± 0.30.98174.7 ± 0.20.71071.5 ± 0.50.78051.2 ± 0.90.92990.7 ± 2.10.66772.5 ± 2.2
2-Hop(3)1.35776.8 ± 0.80.60873.6 ± 0.90.41071.0 ± 0.70.41550.1 ± 1.50.91091.0 ± 0.00.39469.5 ± 1.3
PR 2-Hop(3)1.18075.1 ± 1.10.52876.5 ± 1.60.35771.5 ± 0.50.36152.5 ± 0.70.92991.3 ± 1.50.34273.3 ± 0.3
2-Hop(5)1.92774.2 ± 0.50.88070.6 ± 1.70.68068.8 ± 0.80.62849.5 ± 0.60.89488.3 ± 1.00.62171.0 ± 0.8
PR 2-Hop(5)1.60674.5 ± 0.40.73471.1 ± 1.50.56669.7 ± 0.20.52348.0 ± 2.20.87188.7 ± 1.50.51772.0 ± 0.3
2-Path(3)1.38575.6 ± 0.00.62073.0 ± 0.80.41871.4 ± 0.10.42349.5 ± 1.80.904588.7 ± 1.70.40272.7 ± 2.0
PR 2-Path(3)1.38576.1 ± 0.30.63275.5 ± 0.40.48874.2 ± 1.90.45151.6 ± 0.40.91591.6 ± 0.00.44676.9 ± 0.7

Q2:The experimental setup evaluates pruned GNNs on several benchmark datasets. However, the improvements in efficiency are marginal, and the comparisons do not convincingly demonstrate practical benefits over existing optimized GNN models. I do not think the TUdatasets are enough to evaluate the method. The authors should consider OGB datasets for evaluations.

After we made appropriate improvements to the models to make them suitable for large-scale graphs, we conducted large-scale graph experiments on dataset ogb-arvix for GIN, Pruned GIN, and Pruned multiple aggregation GIN, the results show that while maintaining comparable accuracy to the original model, the pruned version achieves significant reductions in both parameter count and training duration, showing noticeable efficiency improvements. Some of the results are shown below.

ModelTest AccuracyVal AccuracyParameterTotal time
GIN0.7012 ±\pm 0.01140.7132±\pm0.00232.29M2.23h
Pruned GIN0.6905 ±\pm 0.02410.7231±\pm0.00410.98M1.52h
Pruned 2-Mul GIN0.7121 ±\pm 0.00920.7341±\pm0.00661.14M1.73h

And all the experimental materials(code) and results in provided in https://anonymous.4open.science/r/PrunedGNN-AC61/README.md

最终决定

The paper proposes pruning techniques for various graph neural network architectures, including Message Passing (MP), K-Path, and K-Hop GNNs. The core contribution is a theoretically grounded framework for pruning redundant structures within these models, which reduces complexity while preserving the expressive power relative to traditional architectures. The authors provide rigorous theoretical justifications using matrix language (MATLANG) and substantiate their claims through experimental validations. The authors should include all the rebuttal efforts into the revision, especially more experiments on large-scale datasets to demonstrate the practical effectiveness and efficiency of the pruned versions. Other than that, it is known that theoretical expressiveness does not always turn into practical distinguishability of non-isomorphic graphs. Thus, I strongly encourage the authors to test their pruned GNNs on the BREC dataset (https://arxiv.org/pdf/2304.07702) to test whether the pruning compromises practical expressiveness.