PaperHub
4.3
/10
Rejected4 位审稿人
最低3最高6标准差1.3
3
3
5
6
4.3
置信度
ICLR 2024

Evaluating graph generative models with graph kernels: what structural characteristics are captured?

OpenReviewPDF
提交: 2023-09-20更新: 2024-02-11

摘要

关键词
graph kernelrandom graph modeldegree distributioncommunity structuregraph generative model

评审与讨论

审稿意见
3

This paper claims that graph kernels are commonly used to evaluate the performance of graph generative models. The paper then investigates whether some standard graph kernels can capture structural properties of graphs such as the degree distribution, the presence of community structure and others. The empirical results demonstrate that the shortest path and graphlet kernels can better capture the considered properties than the rest of the kernels.

优点

  • This is an interesting paper. Even though graph kernels have been extensively studied for two decades, it is still not clear what graph properties are encoded into the explicit (or implicit) representations created by those methods. This work actually complements [1] which theoretically investigates whether graph kernels can capture properties such as triangle-freeness, connectivity and planarity.

  • The work is useful to practitioners since the reported results suggest that some kernels can better capture some properties than other kernels. Thus, if a practitioner is interested in a specific property, this work can help them choose the right kernel.

  • The presentation is reasonably clear.

[1] Kriege, N. M., Morris, C., Rey, A., & Sohler, C. "A Property Testing Framework for the Theoretical Expressivity of Graph Kernels". In Proceedings of the 27th International Joint Conference on Artificial Intelligence, pp. 2348-2354, 2018.

缺点

  • The paper claims that graph kernels are commonly used to evaluate the performance of graph generative models, however, no references are given. The authors should add some references and also motivate the use of graph kernels for this task. What are the advantages of graph kernels over competing evaluation approaches?

  • The paper focuses on 7 graph properties, but it is not clear why those 7 specific properties were chosen and not others. Are there any applications of graph generative models where graphs that exhibit those properties need to be generated? Please provide more details.

  • Since no theoretical results are provided, I would expect the authors to provide some explanation or intuitions about why some kernels fail to capture some of the considered properties. An explanation is provided in the case of the WL kernel, but not for all the kernels.

  • The experimental evaluation is not fully convincing because all the generated graphs consist of 50 nodes and 190 edges (in expectation). It is not thus clear whether the results generalize to graphs of different sizes. I would suggest the authors construct other datasets where graphs are of different size and different density.

问题

  • In page 9, it is mentioned that "there are only two different graphlets of size k=3: wedges and triangles". This is true if connected graphlets are only considered. However, the standard graphlet kernel also counts disconnected graphlets and there are 4 such graphlets for k=3. Did the authors use an implementation that counts connected graphlets?
评论

Thank you for the thorough review and useful suggestions! We address your concerns below.

W4. The experimental evaluation is not fully convincing because all the generated graphs consist of 50 nodes and 190 edges (in expectation). It is not thus clear whether the results generalize to graphs of different sizes. I would suggest the authors construct other datasets where graphs are of different size and different density.

Thank you for this suggestion, the experiments on larger graphs (1000 nodes) are currently running, we will update the paper when they are ready.

The main reason that we focus on small graphs, is that these methods (evaluating graph generative models using graph kernels) are typically applied to relatively small networks. For example, in molecular modeling, the graphs consist of a relatively small number of atoms. Other applications where small graphs are relevant include modeling ego networks in sociology.

Additionally, let us note that not all kernels are scalable: for instance, the graphlet kernel has a prohibitively high computational cost. We chose these smaller graphs so that we can include all popular graph kernels in our experiment.

Also, as suggested, we conduct additional experiments where we consider density as a characteristic that we vary. The results are the following:

KernelDensity
SP0.993 / 0.996
WL0.996 / 0.996
WL-OA0.996 / 0.996
Graphlet-30.996 / 0.996
NSPDK0.389 / 0.341
PM0.982 / 0.979
NetLSD0.996 / 0.996
RandGIN0.903 / 0.973

We see that detecting density is an easy task for most of the kernels, while there are some exceptions, e.g., NSPDK.

W1. The paper claims that graph kernels are commonly used to evaluate the performance of graph generative models, however, no references are given. The authors should add some references and also motivate the use of graph kernels for this task. What are the advantages of graph kernels over competing evaluation approaches?

We refer to Section 2.2 which discusses the evaluation of graph generative models. In particular, O’Bray et al. (2022) focus on this pipeline of evaluating generative models (but they use very simple kernels). Thompson et al. (2022) go beyond kernel-based evaluation, but their proposed approach can also be fitted to the kernel-based framework as we do in the paper.

W2. The paper focuses on 7 graph properties, but it is not clear why those 7 specific properties were chosen and not others. Are there any applications of graph generative models where graphs that exhibit those properties need to be generated? Please provide more details.

We have chosen high-level properties that are meaningful, diverse, and easy to interpret. Of course, one can expect a good graph kernel to be sensitive to all of them, but there is usually a trade-off. Regarding potential applications, for molecular modeling it can be important that the measure is sensitive to clustering or complementarity as the presence of certain small substructures is crucial. For ego-network modeling, heterogeneity, and community structure are important.

W3. Since no theoretical results are provided, I would expect the authors to provide some explanation or intuitions about why some kernels fail to capture some of the considered properties. An explanation is provided in the case of the WL kernel, but not for all the kernels.

Let us note that most of the kernels are intractable for theoretical analysis. For Graphlet3, we might be able to provide the expected values of kernel values, but getting the variances is already much less tractable. Regarding the intuitive explanations, we will try to improve this part in the revised version.

Q1. In page 9, it is mentioned that "there are only two different graphlets of size k=3: wedges and triangles". This is true if connected graphlets are only considered. However, the standard graphlet kernel also counts disconnected graphlets and there are 4 such graphlets for k=3. Did the authors use an implementation that counts connected graphlets?

Yes, we use the GraKeL implementation that counts connected graphlets.

评论

I would like to thank the authors for their response. However, I feel that my concerns have not been properly addressed.

W4: I think there has been a misunderstanding. I never implied that the size of the graphs is small. The results are not convincing because all the generated graphs consist of the same number of nodes and edges. Thus, there is no diversity in the constructed dataset. I suggest the author experiment with graphs of various sizes and densities.

W2: I actually meant application papers that use graph kernels to evaluate the proposed models. As far as I know, most papers use MMD to evaluate the generated graphs (see [1]) or application-specific metrics such as validity in chemoinformatics (see [2]).

W3: This is really important for understanding the limitations of the different kernels. Such an analysis would greatly strengthen the paper.

[1] Garg, S., Dhamo, H., Farshad, A., Musatian, S., Navab, N., & Tombari, F., Unconditional scene graph generation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 16362-16371, 2021.
[2] Luo, Y., Yan, K., & Ji, S., Graphdf: A discrete flow model for molecular graph generation. In International Conference on Machine Learning, pp. 7192-7203, 2021.

评论

Thank you for the additional clarifications!

W1. We have added the experiment with the varying density to the revised version of the paper and we also conducted experiments with larger graphs to see whether our conclusions generalize. But indeed, it seems that we misunderstood this question. Could please clarify which additional experiments you think are useful? In our current experiments, we deliberately try to fix all graph properties except for one property of interest since our goal is to evaluate the sensitivity to a given property and that is why we fix nn and mm (in expectation) in each experiment (except for the density where we vary mm). We would be very grateful if you could specify a particular experimental setup that you think will be more convincing since it may allow us to strengthen the paper.

W2. We agree that most papers use MMD for evaluating unconditional generation which is why we follow the same procedure, as described in Section 2.2. The computation of MMD requires a graph kernel and the choice of this kernel is the subject of our study. Thus, our evaluation procedure is the same as in [1].

W3. We extended the discussion of the shortest path kernel and the graphlet kernel in Section 4. For this, we also added Figures 3 and 4 in Appendix.

Thank you for your involvement in the discussion! Your feedback is helpful for us.

审稿意见
3

This paper presents a graph similarity technique based on graph kernels and their capacity to capture high-level structural properties. The properties considered by the paper include: degree distribution, community structure, latent geometry, and others. The paper presents a continuous transitions between the models and measures the sensitivity of the graph kernels to the changes. The experiments show this property — that different graph kernels have different sensitivity with Shortest Path kernel and Graphlet kernel showing the best performance in terms of capturing graph properties.

优点

Among the strengths this paper has is that the proposed model is relatively easy to follow and the writing and organization are relatively good. Some aspects are subtle due to the nature of the kernel organization and abstraction of features. The most relevant positive aspect is the number of baselines used, and the transparency of the global architecture comparison. That is, this paper could serve as a reference of graph kernel performance across metrics, which is in itself an interesting idea.

缺点

However, the paper comes with several weaknesses. The most noticeable one is the nature of the graphs used for evaluation. With n = 50 nodes and m = 190 edges, the size of the graphs is hardly what we see in practice, where datasets contain [hundreds of] thousands of nodes and many more edges. While this does not disqualify this paper's potential application, it limits the applicability of the insights provided. Other considerations could be improved in the paper. For instance, the Chung Lu models require the use of very heavy parameterization where every node's weight is used for modeling the degree distribution of a graph. Another weakness is that the concept of dimensionality, in page 5, is introduced in the paper in a very general manner and without sufficient rigor to incorporate it in the framework. Thus, something like a pseudocode could be useful to clarify some of these imprecisions.

问题

What would be your main argument to ensure the method is generalizable to larger graphs?

What is the effect of using Chung Lu models computationally?

Can you describe algorithmically how you incorporated/modeled dimensionality in your framework?

How did you encoded these graph characteristics in the kernels?

伦理问题详情

NA

评论

Thank you for the review! We address your concerns below.

W1. With n = 50 nodes and m = 190 edges, the size of the graphs is hardly what we see in practice, where datasets contain [hundreds of] thousands of nodes and many more edges. While this does not disqualify this paper's potential application, it limits the applicability of the insights provided.

Thank you for this suggestion, the experiments on larger graphs (1000 nodes) are currently running, we will update the paper when they are ready. Let us note that graphs significantly larger than this size are rarely considered in settings we focus on, where one needs to compare two sets of graphs (and not just two graphs).

The main reason that we focus on small graphs, is that these methods (evaluating graph generative models using graph kernels) are typically applied to relatively small networks. For example, in molecular modeling, the graphs consist of a relatively small number of atoms. Other applications where small graphs are relevant include modeling ego networks in sociology.

Additionally, let us note that not all kernels are scalable: for instance, the graphlet kernel has a prohibitively high computational cost. We chose these smaller graphs so that we can include all popular graph kernels in our experiment.

Q1. What would be your main argument to ensure the method is generalizable to larger graphs?

To extend our previous reply, let us also note that our method (framework for comparison of graph kernels) is applicable to graphs of any size. The only bottleneck here is that not all popular kernels are computationally efficient.

W2. For instance, the Chung Lu models require the use of very heavy parameterization where every node's weight is used for modeling the degree distribution of a graph. and Q2. What is the effect of using Chung Lu models computationally?

As we describe in Section 3.2, for the Chung Lu model the weights are drawn from a Pareto distribution with some power-law exponent. Thus, this version of the Chung-Lu model takes only one additional parameter, which is used for the interpolation. Then, a naive implementation for generating a graph requires O(n^2) time (and there exist more efficient approaches). Note that this is faster than computing many graph kernels, thus this part is not a bottleneck of our analysis.

W3. Another weakness is that the concept of dimensionality, in page 5, is introduced in the paper in a very general manner and without sufficient rigor to incorporate it in the framework. Thus, something like a pseudocode could be useful to clarify some of these imprecisions. and Q3. Can you describe algorithmically how you incorporated/modeled dimensionality in your framework?

To model latent geometry, we consider a random geometric graph, as described in “Latent geometry” above. The only modification we make in “Dimensionality” is varying the height of the torus from 1 to 0. We will add more details to the text, but if you have any specific questions - please, tell us as it will help us to improve the readability. Note that the code of our models and transformations is also available.

Q4. How did you encoded these graph characteristics in the kernels?

Let us note that we take popular graph kernels and do not modify them in any way. Our aim is to assess what graph characteristic each kernel is sensitive to.

评论

Thank you for your answers. The comments to Q1, W2, W3, and Q4 do clarify my concerns. However, there is a need for incorporate this to the text of the paper. This will substantially improve the paper. However, I feel that W1 and the overall weakness of the evaluation, i.e., not sufficient evidence that shows applicability, make me hesitant to raise my score. I do appreciate the ideas proposed in the paper and think that it has great potential. However, the paper is still not ready for publication.

评论

Thank you very much for your reply! Following your suggestion, we conducted the experiments on larger graphs, please see this comment. The obtained results are consistent with what we observed for smaller sizes. We also updated the paper to address the comments of the reviewers, we describe what is changed in this comment.

审稿意见
5

The paper generate random graphs with different models and compute different kernels of the graphs. It then show how the kernels are sensitive to the interpolations of parameters between generative models.

优点

  • I think the question is reasonable to ask and the paper is an attempt in the right direction.

缺点

  • I find the paper shows some relationship between generative models of graphs and kernels, it is hard to see what are the applications of these findings in real life. The conclusion is rather qualitative.
  • One of the key step of the paper is to interpolate between generative models, its formula is not clearly stated. It seems to "interpolate" probabilities of different models, and the idea of "interpolating graph characteristics" has to be properly defined as the characteristics are not mutually exclusive. It is difficult to understand what it means by having "100%" of a characteristic and "0%" of another when the two are related.
  • The "performance" of kernels is difficult to grasp and it is related to what is needed to be performed as in also the last point. Overall, I'd suggest to go on this direction with a clear, deeper study of what exactly are the different between generative models. At the moments, it doesn't show concrete conclusion yet for any practical purpose.

问题

N/A

评论

Thank you for the review! We address your concerns below.

W1. I find the paper shows some relationship between generative models of graphs and kernels, it is hard to see what are the applications of these findings in real life. The conclusion is rather qualitative.

Kernels can be used to evaluate graph generative models. However, some kernels are insensitive to certain graph characteristics. For example, our experiments show that the widely-used Weisfeiler-Lehman kernel is insensitive to geometry. In turn, geometry can be especially relevant, e.g., for the application of molecular modeling because atoms have spatial locations. Such findings help practitioners decide which kernels they should and shouldn't use for their evaluations.

W2. One of the key step of the paper is to interpolate between generative models, its formula is not clearly stated. It seems to "interpolate" probabilities of different models, and the idea of "interpolating graph characteristics" has to be properly defined as the characteristics are not mutually exclusive. It is difficult to understand what it means by having "100%" of a characteristic and "0%" of another when the two are related.

We agree that varying one characteristic may also affect other graph properties as it can be hard (or even impossible) to vary only one property in isolation. For each property, we have chosen a graph generator that produces graphs in which this property is strongly present (we try to consider the simplest possible model). Then, we fix a starting point (usually the ER graph) and an ending point (a given model where the property is present to some extent) and make a transition between the two. Section 3.2 states how the parameters of the graph generators depend on the interpolation parameter. If some particular transitions are not clear, we are happy to clarify.

W3. The "performance" of kernels is difficult to grasp and it is related to what is needed to be performed as in also the last point. Overall, I'd suggest to go on this direction with a clear, deeper study of what exactly are the different between generative models. At the moments, it doesn't show concrete conclusion yet for any practical purpose.

Following your suggestions, we will extend the discussion to clearly state some practical conclusions that follow from our analysis. For instance, we observe that random GIN is insensitive to communities, while WL is insensitive to latent geometry. If the generator is supposed to generate graphs that match a certain sample of graphs in terms of having a similar community structure or geometry, then these kernels should be avoided.

评论

Thanks to the author for the reply.

I find that the transition betweens graph properties, which are the target to find suitable kernels, is not justified in the properties it retains or losses, nor in its usefulness. I will keep my current evaluation.

审稿意见
6

This paper focuses on identifying the topological or structural features captured by different graph kernels, enabling them to be compared, in the context of generative graph models.

The authors consider the Erdös-Rényi model as a baseline from which the features under consideration are absent. They propose different ways of departing from this model by randomly adding one of the features under consideration to the graph. For each feature (e.g. degree heterogeneity), they obtain a probabilistic graph model parameterized by θ\theta, where θ=0\theta=0 corresponds to the absence of the feature (Erdös-Rényi model), and θ=1\theta=1 to maximum presence.

The sensitivity of a kernel to each of these features is then numerically evaluated by Spearman's correlation between θ\theta and maximum mean discrepancy along the interval with respect to each of the two endpoints.

Numerical results show that most kernels identify some features easily and others less well, although the Shortest Path kernel and the Graphlet kernel have good performance on all models.

优点

The subject covered in this article is certainly very interesting, and in my opinion goes beyond the question of generative graph models.

The methodology employed seems to me original and could be reapplied to other contexts.

The observations made on the various kernels provide further insight into their ability to detect certain topological features, which may already be valuable for practitioners.

Last but not least, the paper is well written and pleasant to read.

缺点

From my point of view, the paper's main weakness lies in the numerical part: the only results presented concern graphs with 50 nodes and 190 edges (on average). We would like to know whether the results obtained are robust when we vary the size of the graph (in number of nodes and number of edges).

Another limitation is that the results obtained depend on the probabilistic models chosen. For instance, each may introduce random features other than the one targeted, which may bias the results obtained. Figure 2 gives some interesting examples of this phenomenon. This may be an intrinsic limitation of the approach, but it can be overcome by considering different probabilistic models for the same feature.

问题

In addition to the comments above, I think the points below should be addressed.

To have a self-contained paper, the Erdös-Rényi and Chung-Lu models should be described.

For some models, the authors explain precisely the link between the parameters and the average number of edges (triadic closure in Appendix B.1 and dimensionality in Appendix B.2). For others (e.g. latent geometry) there is no explanation at all.

As a reader, we are left a little disappointed by the two correlation measures. A single indicator would greatly simplify the statement of results. The data suggest that, in most cases, they are equal. For features such as the triadic model and for the Graphlet kernel, the link between the two correlations might be studied theoretically, which would provide a valuable piece of information on this question.

On page 2, one reads that computing a distance is as hard as graph isomorphism, while a few lines down, it says that any kernel can be easily transformed to a distance measure. Can the authors address this apparent contradiction?

Typos:

page 2: popular

page 3: Gi\mathcal{G}_i is not defined

page 4: triadic model in Appendix B.1

page 8: colors in Table 1?

page 8: 4.1 Results?

page 13 (B.2): h2hh\geq 2h

page 13 (B.2): Solving... r=r^=

评论

Thank you for the thorough review and useful suggestions! We will improve the paper accordingly. We address the questions and concerns below.

W1. The only results presented concern graphs with 50 nodes and 190 edges (on average). We would like to know whether the results obtained are robust when we vary the size of the graph (in number of nodes and number of edges).

Thank you for this suggestion, the experiments on larger graphs (1000 nodes) are currently running, we will update the paper when they are ready.

The main reason that we focus on small graphs, is that these methods (evaluating graph generative models using graph kernels) are typically applied to relatively small networks. For example, in molecular modeling, the graphs consist of a relatively small number of atoms. Other applications where small graphs are relevant include modeling ego networks in sociology.

Additionally, let us note that not all kernels are scalable: for instance, the graphlet kernel has a prohibitively high computational cost. We chose these smaller graphs so that we can include all popular graph kernels in our experiment.

W2. Another limitation is that the results obtained depend on the probabilistic models chosen. For instance, each may introduce random features other than the one targeted, which may bias the results obtained. Figure 2 gives some interesting examples of this phenomenon. This may be an intrinsic limitation of the approach, but it can be overcome by considering different probabilistic models for the same feature.

We agree with this comment and plan to add a clarification about this limitation in the revised text. Our aim was to model each of the properties in the most straightforward way, and in a way that minimally affects other graph characteristics. For some graph characteristics, it can be challenging (or even impossible) to vary them in isolation from everything else.

The idea of considering different models for the same property is interesting - we already do this for clustering: this property can be modeled via triadic closure, latent geometry, or community structure. Other possible variations we can think of are: modeling complementarity by disassortative community structure and modeling degree inhomogeneity by preferential attachment. We consider adding these experiments in the revised version of the paper. However, we may not have time to finalize them during the discussion period as we also need to conduct experiments on larger graphs, as discussed above.

Q1. To have a self-contained paper, the Erdös-Rényi and Chung-Lu models should be described.

Thank you for pointing this out, we will add the definitions in the revised version.

Q2. For some models, the authors explain precisely the link between the parameters and the average number of edges (triadic closure in Appendix B.1 and dimensionality in Appendix B.2). For others (e.g. latent geometry) there is no explanation at all.

The result for latent geometry also follows from Appendix B.2 by setting h=1. We will add a reference to Appendix B.2 to the paragraph about latent geometry.

Q3. As a reader, we are left a little disappointed by the two correlation measures. A single indicator would greatly simplify the statement of results. The data suggest that, in most cases, they are equal.

Thank you for this comment - we agree that having two numbers can be somewhat inconvenient and that they are in most cases consistent. We decided to simplify the presentation by reporting the average value in the main table (or average±difference/2). We will update the table in the revised version.

Q4. On page 2, one reads that computing a distance is as hard as graph isomorphism, while a few lines down, it says that any kernel can be easily transformed to a distance measure. Can the authors address this apparent contradiction?

Thanks for noticing! As we write above, graph ‘distances’ usually violate the positivity axiom, and if we apply the mentioned transformation - positivity can also be violated. We will clarify this in the text.

We hope that our response addresses your concerns. We are happy to answer any further questions.

评论

I would like to thank the authors for their response.

I am curious to see what additional numerical results they will be able to offer to the paper, both on larger graphs and on the study of one feature from several models.

In my opinion, with such further outcomes, the paper could be published at the conference.

I also agree with Reviewer oaCD: beyond the size of the graphs, it is important to see the results over a wide range of graphs. It could be the case with the new experiments if they do not impose a single number of edges in the simulated graphs.

EDIT: Comments from other reviewers are taken into account.

评论

Thank you for your involvement in the discussion and positive feedback! Following your suggestion, we conducted the experiments on larger graphs, please see this comment. The obtained results are consistent with what we observed for smaller sizes. We also updated the paper: fixed typos, addressed your suggestions, and added more results.

Regarding a wider range of graphs, please see our reply to Reviewer oaCD. If you have any particular suggestions here, we will be happy to conduct additional experiments to further strengthen the paper.

评论

Following the comments of several reviewers, we conducted experiments on larger graphs (1000 nodes). The results are the following:

HeterogeneityTriadic closureCommunitiesGeometryDimensionalityComplementarity
WL0.9960.996-0.0300.7410.2260.561
WL-OA0.9960.9960.0990.9480.2100.871
PM0.9750.9930.9310.9890.2380.981
NetLSD0.9100.9950.5300.9850.6310.958
RandGIN0.9310.943-0.0410.8490.219-0.080

We note that not all the kernels are present in this table since the remaining ones are not scalable to large datasets. In the final version of the paper, we plan to add the complete version of this experiment: we can use sampling to speed up the kernel computation. However, studying the effect of sampling requires additional experiments, which is why we only keep the scalable kernels at this point. We see that these results are consistent with those reported in the paper for smaller graphs.

评论

We thank the reviewers for their valuable comments, we updated the paper accordingly. The following changes have been made:

  • We added a new characteristic to our analysis - density - that is modeled via the ER model with different edge probabilities.
  • In Table 1, we now aggregate two performance measurements (for each kernel/transition pair) in one number.
  • We added a new section Limitations (Appendix C).
  • We describe the Erdös-Rényi and Chung-Lu models in Section 3.
  • We extended our discussion of the experimental results for SP and Graphlet kernels. We added Figures 3 and 4 (Appendix) to better illustrate the results of these two kernels.
  • Conclusion is extended with more practical suggestions.
  • We corrected typos and made other small improvements in the text.
AC 元评审

Summary
Graph comparison is one of the core tasks in graph machine learning. This paper directly addresses this problem and empirically evaluates the effectiveness of various graph kernels for comparing artificially generated graphs with a variety of properties, such as density and degree heterogeneity.

Strengths

  • Graph comparison via graph kernels is an important and relevant problem in graph ML. Therefore, this paper, which directly studies this problem, is well-motivated.
  • Presentation is clear, and the paper is easy to follow.
  • I find the choice of graph kernels covered in this paper to be reasonable.

Weaknesses

  • The empirical evaluation is not entirely convincing due to the limited variety of graphs, as also noted by the reviewers, although the authors included additional results on larger graphs in the revised version. Given that empirical comparison is the main contribution of this paper, addressing this limitation is crucial.
  • More careful consideration of hyperparameters for graph kernels is required. For instance, in the WL kernel, fixing ll to be 5 may not capture the full range of performance, as it is well known that the performance largely depends on ll. Therefore, evaluation on various hyperparameter settings would be necessary.
  • Theoretical insights are lacking in this paper. While I understand that the primary focus is on empirical evaluation, which is deserving of publication if thorough and insightful, theoretical analysis is also important for a more comprehensive understanding of graph kernels.

为何不给更高分

I have carefully reviewed the paper, and the weaknesses of the paper, as outlined above and also highlighted by the reviewers, are crucial and must be addressed for the publication of this paper. While I appreciate the authors' response and the revisions made during the author-reviewer discussion phase, in my opinion, it requires resubmission, and I recommend rejecting the paper.

为何不给更低分

N/A

最终决定

Reject