PaperHub
4.3
/10
Rejected3 位审稿人
最低3最高5标准差0.9
5
3
5
4.0
置信度
ICLR 2024

EPIC: Compressing Deep GNNs via Expressive Power Gap-Induced Knowledge Distillation

OpenReviewPDF
提交: 2023-09-24更新: 2024-02-11

摘要

关键词
deep graph neural networksknowledge distillationexpressive power gap

评审与讨论

审稿意见
5

This paper theoretically and experimentally analyzes the expressive power gap between deep GNN teachers and lightweight students. It formulates the estimation of the expressive power gap to an embedding regression problem. The results on three large-scale datasets demonstrate the effectiveness of the proposed method. Overall, the paper is well written with an interesting topic.

优点

  • The analysis of graph knowledge distillation from the perspective of expressive power is novel.
  • The paper is well written and presented, especially about the notation and background.
  • EPIC outperforms 1001-layer GNNs with a 4-layer GNN.

缺点

  • My main concern is the motivation for the paper. Although deep GNNs have been a popular research topic, deep GNNs are prone to suffer from overfitting and over-smoothing and so far have not been widely used as shallow GNNs. I am confused about the significance of studying distillation for deep GNN.
  • The authors did not conduct their experiments in a transductive setting as GLNN and NOSMOG did.
  • Why does the performance of NOSMOG drop so much in Table 1? According to their original paper, NOSMOG can achieve performance even better than the teacher GNN.
  • Can the authors provide more results on the different layers of teacher GNN in Table 1? In particular, I would like to know how generalizable EPIC is in a shallow teacher GNN setting?
  • As a GNN-to-GNN distillation method, the authors mainly compare EPIC with two GNN-to-MLP methods (GLNN and NOSMOG) in their experiments. Comparisons with some SOTA GNN-to-GNN distillation methods are missing, such as those introduced by the authors in related work.

问题

Please explain (3-5) in the weakness part. How about the time complexity (especially the bound estimation) and running time?

伦理问题详情

No ethics review is needed.

评论

We thank the reviewer for the insightful and valuable comments. We respond to your comments as follows and sincerely hope that our rebuttal could properly address your concerns. If so, we would deeply appreciate it if you could raise your score. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our submission.

1. The motivation for studying the distillation for deep GNNs.

Thanks for your comments.

We summarize the motivation for studying the distillation for deep GNNs as follows.

  • First, on large-scale datasets (e.g., ogbn-arxiv, ogbn-products, and ogbn-proteins), the deep teacher models that we distill in our paper (i.e., RevGCN-Deep, RevGNN-112, and RevGNN-Deep) outperform what existing KD4GNN works aim to distill, such as GCN, GraphSAGE, GAT, and APPNP. Compared with shallow GNNs, deep GNNs with more message passing layers are better at capturing long-range patterns [1, 2], which are critical for many downstream tasks such as search engine [3], recommendation systems [4], and molecular property prediction [5].

  • Second, although some deep GNNs still face challenges such as over-fitting and over-squashing, there have been many efforts to solve these challenges, such as [6, 7, 8, 9, 10, 11]. We believe that the challenges are temporary obstacles to the development of deep GNNs, and we believe that deep GNNs will receive widespread attention in the future.

  • Last but not least, the slow inference speed is also one of the important factors limiting the real applications of deep GNNs. Despite the great success in compressing and accelerating moderate-sized or shallow GNNs of existing KD techniques, it still remains a tough challenge to distill deep GNNs due to the huge expressive power gap between the teachers and students. Therefore, this paper aims to address the challenge of distilling deep GNNs that excel on large-scale graphs and possess numerous layers.

[1] Chen, Tianlong, et al. "Bag of tricks for training deeper graph neural networks: A comprehensive benchmark study." IEEE Transactions on Pattern Analysis and Machine Intelligence 45.3 (2022): 2769-2781.

[2] Cong, Weilin, Morteza Ramezani, and Mehrdad Mahdavi. "On provable benefits of depth in training graph convolutional networks." Advances in Neural Information Processing Systems 34 (2021): 9936-9949.

[3] Brin, Sergey, and Lawrence Page. "The anatomy of a large-scale hypertextual web search engine." Computer networks and ISDN systems 30.1-7 (1998): 107-117.

[4] Fan, Wenqi, et al. "Graph neural networks for social recommendation." The world wide web conference. 2019.

[5] Kearnes, Steven, et al. "Molecular graph convolutions: moving beyond fingerprints." Journal of computer-aided molecular design 30 (2016): 595-608.

[6] Li, Guohao, et al. "Training graph neural networks with 1000 layers." International conference on machine learning. PMLR, 2021.

[7] Chen, Ming, et al. "Simple and deep graph convolutional networks." International conference on machine learning. PMLR, 2020.

[8] Zhao, Lingxiao, and Leman Akoglu. "Pairnorm: Tackling oversmoothing in gnns." arXiv preprint arXiv:1909.12223 (2019).

[9] Guo, Xiaojun, et al. "Contranorm: A contrastive learning perspective on oversmoothing and beyond." arXiv preprint arXiv:2303.06562 (2023).

[10] Jaiswal, Ajay, et al. "Old can be gold: Better gradient flow can make vanilla-gcns great again." Advances in Neural Information Processing Systems 35 (2022): 7561-7574.

2. The experiments in the transductive setting.

Thanks for your comments. We conduct experiments in the setting of transductive training. The results are as follows.

DatasetsTeacherLSPGLNNNOSMOGEPIC (Ours)
Arxiv72.9669.7757.4165.7973.33
Products82.1178.3161.4564.6283.73
Proteins87.42-75.1973.8487.13

3. The drop of the performance of NOSMOG compared to that reported in the original paper.

Thanks for your comments.

We train MLPs using the source code of NOSMOG in the inductive setting. We speculate that the reasons for the drop of performance of NOSMOG compared to that reported in the original paper are as follows.

In the original paper, the authors train NOSMOG in the transductive setting and a loose inductive setting (with only 20% of test nodes invisible to the models). In these settings, the teacher model's predictions on most test nodes are available to students, hence the students can directly fit the predictions of their teachers on test nodes to achieve comparable or even superior performance to the teachers. However, we conduct experiments in the strictly inductive setting, which is more challenging for student MLPs than the experiment settings in the original paper.

评论

4. The distillation results for shallow GNN teachers.

Thanks for your comments.

We conduct experiments of distilling shallow GNN teachers on ogbn-arxiv in the transductive and inductive settings. Specifically, in the two settings, we train teacher models with 7 and 3 layers, and distill them to 3-layer and 1-layer students, and a 1-layer student, respectively. The results are as follows.

SettingTeacher-7layersEPIC-3layersEPIC-1layerTeacher-3layersEPIC-1layer
Transductive70.9372.3170.5970.3570.57
Inductive70.6971.9170.4269.9770.13

5. Comparisons between EPIC and the state-of-the-art GNNs-to-GNNs distillation method.

Thanks for your comments.

Because different GNNs-to-GNNs distillation works use different teacher models and conduct experiments on different datasets, there is currently no widely recognized state-of-the-art GNNs-to-GNNs distillation framework. Therefore, we select a new GNNs-to-GNNs distillation baseline, i.e., BGNN [1] with 2-layer GCNs as students, as the average results of BGNN on the datasets Cora, CiteSeer, and Pubmed are the best among existing GNNs-to-GNNs frameworks to the best of our knowledge. Besides, we implement BGNN ourselves, as BGNN is not open source. If you want us to compare EPIC with other baselines, please let us know and we will add new baselines to our paper. The results are as follows.

DatasetsTeacherBGNN-GCNEPIC (Ours)
Inductive
Arxiv72.9168.8773.06
Products77.8674.0178.58
Proteins85.91-85.85
Transductive
Arxiv72.9669.9473.33
Products82.1179.4583.73
Proteins87.42-87.13

[1] Guo, Zhichun, et al. "Boosting graph neural networks via adaptive knowledge distillation." Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 37. No. 6. 2023.

评论

Dear Reviewer 8ZML,

Thank you again for your careful reading and insightful comments, which are of great significance for improving our work. The deadline for the discussion stage is approaching, and we are looking forward to your feedback and/or questions. We sincerely hope that our rebuttal has properly addressed your concerns. If so, we would deeply appreciate it if you could raise your score. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our submission.

Best,

Authors

审稿意见
3

The author(s) theoretically demonstrate that the shallow student model lacks sufficient expressive power to mimic the teacher; hence, distilling deep GNNs remains a tough challenge. The derived upper bound allows for quantitative analysis of the gap, making it easy to determine an appropriate number of layers for student models.

优点

  1. Distilling very deep GNNs is challenging and important.
  2. This work provides some valuable suggestions on layer number selection.

缺点

  1. The conclusion obtained from Theorem 1 is obvious, and the upper bound derived therein appears difficult to apply in practice. While the hyperparameter γ\gamma can achieve this to some extent, the value of γ\gamma is not directly related to classification performance gap. Therefore, a search is also required to determine the appropriate value.
  2. The complexity of computing EPIC bounds should be analyzed (since SVD for large-scale graphs is expensive).
  3. Feature distillation (namely the EPIC loss presented in this paper) is a well-known technique.

问题

Please refer to Weaknesses.

评论

We thank the reviewer for the insightful and valuable comments. We respond to your comments as follows and sincerely hope that our rebuttal could properly address your concerns. If so, we would deeply appreciate it if you could raise your score. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our submission.

1. The conclusion obtained from Theorem 1 is obvious.

Thanks for your comments.

Theorem 1 guarantees that we can distill a deep GNN into a shallow GNN, as there exist some shallow GNNs that have similar expressive power to the deep GNN. This conclusion is significant for the distillation of deep GNNs and is surprising in a way, as the numbers of neighbor hops included in the computation of each node in deep GNNs and shallow GNNs are very different.

Specifically, Theorem 1 shows that the minimum approximation error of HS(M)\mathbf{H}_S^{(M)} to HT(L)\mathbf{H}_T^{(L)} in terms of the Frobenius norm has an upper bound that decreases monotonically with respect to MM. Furthermore, we empirically demonstrate that the upper bound converges exponentially to zero as MM increases.

2. How to apply the EPIC bound and select the hyperparameter γ\gamma in practice?

Thanks for your comments.

After reading your comments, we remove the hyperparamter γ\gamma and instead use the EPIC bounds to narrow the range of tuning the number of student layers as follows.

  1. We compute the EPIC bounds (uEPIC(M;L))M=1L1(u_{\rm EPIC}(M;L))_{M=1}^{L-1} .
  2. We plot uEPIC(M;L)u_{\rm EPIC}(M;L) as a function of MM, as shown in Figure 1.
  3. By observing the plot, we find the maximum value MmaxM_{\max} such that uEPIC(M;L)u_{\rm EPIC}(M;L) decreases slowly when M>MmaxM>M_{\max}.
  4. We tune the hyperparamter MM in [1,Mmax][1, M_{\max}].

Because uEPIC(M;L)u_{\rm EPIC}(M;L) converges exponentially to zero as MM increases, MmaxM_{\rm max} is usually much smaller than LL. Thus, we can significantly reduce the range of tuning MM.

3. The complexity of computing the EPIC bound is high.

Thanks for your comments.

Please note that we do not need to compute EPIC bounds in the inference stage of GNNs, hence the computation of EPIC bounds does not affect the inference speed of GNNs, which is the focus of this paper and most knowledge distillation methods.

评论

Dear Reviewer G7rf,

Thank you again for your careful reading and insightful comments, which are of great significance for improving our work. The deadline for the discussion stage is approaching, and we are looking forward to your feedback and/or questions. We sincerely hope that our rebuttal has properly addressed your concerns. If so, we would deeply appreciate it if you could raise your score. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our submission.

Best,

Authors

审稿意见
5

This paper focuses on the knowledge distillation of deep GNNs, such as that with over 100 layers. To address the issue of the representation ability gap between teacher and student networks, this paper proposes a framework, EPIC, leveraging the embedding regression based on the theory of polynomial approximation. Specifically, EPIC first shows the EPIC bound exponentially converges as the number of student layers increases with experiments. Then, with this observation, it selects an appropriate layer number of the student network by analyzing the EPIC bound. Furthermore, an expressive power gap-induced loss term is proposed to reduce the gap. In the experiments, it reduces 94% of layers, improves 8x speed, and obtains comparable performance for 1,001-layer RevGNN-Deep.

优点

  1. The field of distilling a very deep GNN model is less explored. By distilling, the inference speed can be improved by a large margin, which benefits the potential applications.
  2. A theoretical framework is proposed to analyze the gap between the student and teacher network from a spectral perspective.
  3. The theoretical results with an EPIC bound are used to select the appropriate layer number. It is interesting and meaningful to reduce the cost of the hyper-parameter tuning.

缺点

The main concerns lie in the method details and experimental settings. The details are shown below:

  1. The hyper-parameter γ\gamma in Eq. (8) weakens (somehow) the theoretical results. The optimal # layers MM^* are dependent on this hyper-parameter, regardless of the EPIC bound. How to choose γ\gamma, is there any guidance? Does it involve the extra hyperparameter tuning cost? When comparing tuning M (layer number) and γ\gamma, what is the advantage of the latter?
  2. The scalability is also a concern, since it requires eigenvalue decomposition to find out all the eigenvalues, incurring high memory and time complexity, especially for large-scale graphs, which is the setting in this work. It is suggested to add some experimental cost to the process of calculating EPIC bounds.
  3. The singular value decomposition (SVD) and Laplacian requires the assumptions of linearity. I understand that this may be easier for derivation. However, it still would be helpful if you could explain why this linear-based theory could be used for non-linear graph models, such as GCN and GAT. It seems that the used teacher model is also non-linear.
  4. Novelty concerns on the EPIC loss. The paper claims the new EPIC loss (expressive power gap-induced loss). However, If I understand correctly, H_S and H_T are representations of student and teacher models respectively. It seems the same as standard knowledge-distilling loss where the embeddings in the student network mimic the ones in the teacher network. It is suggested to mention this and avoid the new terms if they are similar. Otherwise, it is better to clarify the differences and provide corresponding evidence (insights and experimental results.)
  5. There are some concerns in the experiment design. First, the comparison needs to be more fair. As for GLNN and NOSMOG, the student model is MLP, while for EPIC, the student model is a GNN, the performance gain may be caused by the enhanced expressivity of student models. Second, this experiment can't be used to verify the correctness of the proposed theorem (i.e., is this bound tight when deciding the # of layers in student GNNs). Do the results of the real experiments align with the curves in Figure 1? I suggest redesigning this experiment to stress the correctness of the proposed theorem. Third, It is suggested to add baselines that train the GNN with fewer layers from scratch. Again, the cost of hyper-parameter γ\gamma is an issue.
  6. (Minor), It is suggested to refine Figures 1 and 2. For example, it is better to provide more details in Figure 2. Currently, the details of EPIC and Bound and EPIC loss are lost.

问题

  1. Regard the γ\gamma. How to choose γ\gamma in the experiments? Does it involve the extra hyperparameter tuning cost (How much)? When comparing tuning M (layer number) and γ\gamma, what is the advantage of the latter?

  2. What is the experimental cost of the calculating process of EPIC bounds?

  3. Could you please explain why the proposed linear-based theory could be used for non-linear graph models?

  4. What is the unique contribution of the proposed EPIC loss?

评论

We thank the reviewer for the insightful and valuable comments. We respond to your comments as follows and sincerely hope that our rebuttal could properly address your concerns. If so, we would deeply appreciate it if you could raise your score. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our submission.

1. The selection of the hyperparameter γ\gamma and the corresponding cost.

Thanks for your comments.

After reading your comments, we remove the hyperparamter γ\gamma and instead use the EPIC bounds to narrow the range of tuning the number of student layers as follows.

  1. We compute the EPIC bounds (uEPIC(M;L))M=1L1(u_{\rm EPIC}(M;L))_{M=1}^{L-1} .
  2. We plot uEPIC(M;L)u_{\rm EPIC}(M;L) as a function of MM, as shown in Figure 1.
  3. By observing the plot, we find the maximum value MmaxM_{\max} such that uEPIC(M;L)u_{\rm EPIC}(M;L) decreases slowly when M>MmaxM>M_{\max}.
  4. We tune the hyperparamter MM in [1,Mmax][1, M_{\max}].

Because uEPIC(M;L)u_{\rm EPIC}(M;L) converges exponentially to zero as MM increases, MmaxM_{\rm max} is usually much smaller than LL. Thus, we can significantly reduce the range of tuning MM.

2. The complexity of computing the EPIC bound is high.

Thanks for your comments.

Please note that we do not need to compute EPIC bounds in the inference stage of GNNs, hence the computation of EPIC bounds do not affect the inference speed of GNNs, which is the focus of this paper and most knowledge distillation methods.

3. The relationship of our linear-based theory to non-linear graph models.

Thanks for your comments.

The reason why we analyze the expressive power of GNNs from a linear perspective is that this is a standard practice [1, 2], and that existing works [1, 3] have demonstrated that the theoretical analysis under the linear assumption is applicable to general non-linear GNNs.

[1] Xu, Keyulu, et al. "Optimization of graph neural networks: Implicit acceleration by skip connections and more depth." International Conference on Machine Learning. PMLR, 2021.

[2] Wang, Xiyuan, and Muhan Zhang. "How powerful are spectral graph neural networks." International Conference on Machine Learning. PMLR, 2022.

[3] Thekumparampil, Kiran K., et al. "Attention-based graph neural network for semi-supervised learning." arXiv preprint arXiv:1803.03735 (2018).

4. The comparison between EPIC and existing GNNs-to-MLPs distillation frameworks is unfair.

Thanks for your comments.

Because we focus on how to select an appropriate student that is lightweight, while expressive, and MLP is currently a popular student structure for distilling GNNs, we compare our proposed EPIC with state-of-the-art GNNs-to-MLPs distillation frameworks. As mentioned in Section 6.2, we attribute the reason for the unsatisfactory performance of GNNs-to-MLPs frameworks to the weak expressive power of MLPs. The experiments support our claim in Introduction that "MLPs are not 'good students' for distilling deep GNNs".

5. Experiments of baselines that train the GNN with fewer layers from scratch.

Thanks for your comments.

We conduct experiments of training student GNNs with fewer layers from scratch on the three OGB datasets. The results are as follows.

Arxiv234567
w/ KD72.5773.0673.1873.2473.3573.24
w/o KD72.2572.9173.0072.6572.3472.15
Products234567
w/ KD76.8076.9977.7078.3278.2678.58
w/o KD76.6776.8676.8577.2977.5677.71
Proteins102030405060
w/ KD85.6485.6885.3385.8185.7285.85
w/o KD84.1384.2384.5884.7584.8985.00
评论

Thank you for your response, and I apologize for my delayed reply.

Upon review, I find that the current revision does not adequately address my concerns. Specifically:

  1. Clarification on 'h' Removal: The implications of removing 'h' are not clear. Could you elaborate on how this affects the study's results?

  2. Rationale Behind Unfair Comparison: The justification for using an unfair comparison remains unconvincing. A more thorough explanation is needed to understand this choice.

  3. Overlooked Novelty Aspects: The issue of novelty in your study appears to have been overlooked. It's important to distinctly highlight what sets your research apart.

  4. Discrepancies in Revision: I did not find any changes in the revised paper that correspond to the issues raised in the response.

Given these unresolved issues, my evaluation of the paper remains unchanged.

评论

We thank the reviewer again for the constructive comments. We respond to your comments as follows and sincerely hope that our rebuttal could properly address your concerns. If so, we would deeply appreciate it if you could raise your score. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our submission.

1. The implications of removing γ\gamma is unclear. How does this affect the results?

Removing the hyperparameter γ\gamma and directly tuning MM does not affect the results of distillation, but is more efficient. This is because γ\gamma is continuous, while MM is discrete.

Specifically, in the original version of our paper, the mapping γM\gamma \mapsto M^* is not injective, that is, there are many different γ\gamma corresponding to the same MM^*. Hence, we need to first find an appropriate γ\gamma in the continuous (0,1](0, 1], and compute MM^*, which results in the high cost of tuning hyperparameters.

After removing γ\gamma, we only need to tune MM in the discrete set {1,2,,Mmax1,2,\ldots, M_{\max}}, which is more efficient than tuning γ\gamma.

2. The explanation about the comparisons between EPIC and existing GNNs-to-MLPs frameworks.

The explanation is as follows.

First, the selection of student models is a critical factor for knowledge distillation [1, 2]. Existing GNNs-to-MLPs works imply the view that "MLPs are enough for the distillation of GNNs" and have triggered an upsurge in the GNNs-to-MLPs distillation. However, we claim that this view does not apply to the distillation for deep GNNs, and call for more attention to turn to the research on distilling deep GNNs using GNNs as students. Therefore, we select two GNNs-to-MLPs frameworks as baselines to support our claim.

Second, we fully understand that the reason why the reviewer thinks the comparison is unfair, i.e., our student models are more expressive than MLPs, leading to the difference in the performance of distillation. However, this is exactly the claim we want to support through our experiments.

[1] Mirzadeh, Seyed Iman, et al. "Improved knowledge distillation via teacher assistant." Proceedings of the AAAI conference on artificial intelligence. Vol. 34. No. 04. 2020.

[2] Gao, Mengya, Yujun Wang, and Liang Wan. "Residual error based knowledge distillation." Neurocomputing 433 (2021): 154-161.

3. The contributions of our paper.

We summarize our contributions as follows.

  1. The problem we focus on (i.e., distillation for deep GNNs) is more challenging than what existing works focus on (i.e., distillation for moderate-sized GNNs). Specifically, this is the first paper that attempts to distill GNNs with more than 100 layers to the best of our knowledge. It is challenging to distill deep GNNs due to the difficult trade-off between the needs of being "lightweight" and being "expressive" when selecting a student for the deep teacher.
  2. Our theoretical analysis and experiments reveal that it is feasible to distill deep GNNs into shallow GNNs.
    • This is the first paper that quantitatively analyzes the teacher-student expressive power gap from a theoretical perspective. Our theoretical analysis reveals the feasibility of distilling a deep GNN into a shallow GNN, as there exist some shallow GNNs that have similar expressive power to the deep GNN. This conclusion is significant for the distillation of deep GNNs and is surprising in a way, as the size of the neighborhood involved in the computation of each node is very different between deep GNNs and shallow GNNs.
    • We reduce the number of layers of the 1,001-layer RevGNN-Deep by 94% and accelerate its inference by roughly eight times, while achieving comparable performance. The result shows that it is possible to significantly reduce the number of layers of deep GNNs via knowledge distillation, bringing confidence to the future research on knowledge distillation for deep GNNs.

4. The changes in the revised version of our submission.

Thanks for the reminder. We have revised our paper based on your comments. The changes are marked in red. For the revision in appendices, please refer to the supplementary materials. For your convenience, we list the changes as follows.

  1. In Abstract, Introduction, Section 5.1, and Conclusion, we revise the contents about the method to "narrow the range of tuning the number of student layers".
  2. In Page 6, Section 5.1, we revise the content about the computation of EPIC bounds.
  3. In Appendix B.1, we add the relationship of linear-based theory to non-linear GNNs.
  4. In Appendix A.5, we add the explanation about why do we compare EPIC with GNNs-to-MLPs frameworks.
  5. In Appendix C.1, we add the experiments of training GNNs with fewer layers from scratch.
  6. We revise all statements about EPIC loss to "a feature distillation loss" or "an expressive power gap-related loss".
AC 元评审

This paper proposes EPIC, an expressive power-based method to distill knowledge from a deeper graph neural network (GNN) teacher model into a shallower GNN student model. The reviewers raised several critical concerns regarding the novelty, theoretical analysis, experiment design, and evaluation of the proposed method:

First, the key concept behind EPIC seems similar to common knowledge distillation methods that minimize the difference between student and teacher representations. The claim of a "new loss" needs more clarification on how it differs from existing techniques.

Second, the theoretical analysis relies on assumptions of linearity while common GNN models are highly non-linear. The bound may not accurately reflect real model capacities. Also, the bound computation requires expensive SVD, hurting scalability.

Additionally, the experiments compare EPIC against student MLPs, while it uses a GNN student itself. More comparisons with state-of-the-art GNN-to-GNN distillation methods are needed to showcase advantage. Experiments also do not strongly verify the theory.

Finally, motivation behind studying deep GNN distillation is unclear, as shallow GNNs remain dominant in practice. The transductive experimental setting used by existing methods is not followed.

I would encourage the authors to consider a major revision, and the paper could be much stronger if these concerns could be properly addressed.

为何不给更高分

NA

为何不给更低分

NA

最终决定

Reject