PaperHub
7.3
/10
Poster4 位审稿人
最低4最高5标准差0.5
5
5
4
4
3.0
置信度
创新性3.5
质量3.0
清晰度3.3
重要性2.8
NeurIPS 2025

Parameter-Free Hypergraph Neural Network for Few-Shot Node Classification

OpenReviewPDF
提交: 2025-05-12更新: 2025-10-29
TL;DR

In this work, we propose ZEN (Zero-Parameter Hypergraph Neural Network), a fully linear and parameter-free model that achieves both expressiveness and efficiency.

摘要

关键词
Hypergraph neural networkFew-shot node classificationLinear model

评审与讨论

审稿意见
5

The paper proposes a novel parameter-free hypergraph neural network for few-shot node classification. Tractable Closed-form Solution (TCS) eliminates iterative training, and Redundancy-Aware Propagation (RAP) explicitly removes self-information redundancy. Comprehensive tests show improved accuracy and efficiency compared to previous methods.

优缺点分析

Strengths

  • The method shows improved accuracy and efficiency compared to previous methods.
  • The paper is technically solid with rigorous theoretical analysis.
  • The paper is well-written and easy to read.

Weakness

  • There is no error bounds for closed-form approximation in Theorem 1.
  • Table 4 presents values without specifying units. If the time for the proposed method is used as the unit of time, please state this clearly and maybe provide the original data in the appendix.
  • Typos exist. For example, WW_* should be WW^* in Proof of lemma 1.

问题

  • Can the authors provide error bounds for closed-form approximation in Theorem 1?
  • Can the authors clarify the unit used in Table 4?

局限性

yes

最终评判理由

The rebuttal answers all my questions, so I keep my score.

格式问题

Formatting instructions are followed.

作者回复

We greatly appreciate your high recognition of our work, and are eager to share our thoughts with you.

Response to Question 1:

Question 1: Can the authors provide error bounds for closed-form approximation in Theorem 1?

Yes, we believe more discussion on the error bounds can make our theoretical claim stronger. We present a summary of our analysis as follows, which will be included in the final version of our paper as well.

In TCS, our approximation is given as follows:

1λ12M1+1λ22M2+1λ32M31ϵ(1λ1M1+1λ2M2+1λ3M3),\frac{1}{\lambda_1^2}M_1+\frac{1}{\lambda_2^2}M_2+\frac{1}{\lambda_3^2}M_3 \approx \frac{1}{\epsilon}\left(\frac{1}{\lambda_1}M_1+\frac{1}{\lambda_2}M_2+\frac{1}{\lambda_3}M_3\right),

where

  • λ1=ϵ\lambda_1 = \epsilon
  • λ2=(12ϵ)k+ϵ\lambda_2 = (1 - 2\epsilon)k + \epsilon
  • λ3=k(12ϵ+ϵc)+ϵ\lambda_3 = k(1 - 2\epsilon + \epsilon c) + \epsilon

and the matrices form an orthogonal decomposition:

  • M1=Ikc1k(IcJk)M_1 = I_{kc} - \frac{1}{k}(I_c \otimes J_k)
  • M2=1k(IcJk)1kcJkcM_2 = \frac{1}{k}(I_c \otimes J_k) - \frac{1}{kc}J_{kc}
  • M3=1kcJkcM_3 = \frac{1}{kc} J_{kc}

Here, JJ denotes an all-one matrix and \otimes is the Kronecker product.

To evaluate the approximation quality, we compute the relative Frobenius norm error:

1λ12M1+1λ22M2+1λ32M31ϵ(1λ1M1+1λ2M2+1λ3M3)F1λ12M1+1λ22M2+1λ32M3F\frac{\left\|\frac{1}{\lambda_1^2}M_1+\frac{1}{\lambda_2^2}M_2+\frac{1}{\lambda_3^2}M_3 - \frac{1}{\epsilon}\left(\frac{1}{\lambda_1}M_1+\frac{1}{\lambda_2}M_2+\frac{1}{\lambda_3}M_3\right)\right\|_F}{\left\|\frac{1}{\lambda_1^2}M_1+\frac{1}{\lambda_2^2}M_2+\frac{1}{\lambda_3^2}M_3\right\|_F}

For a representative case with k=5k = 5, c=10c = 10, the relative errors are as follows:

ϵ\epsilonRelative Error
0.11.14%
0.010.10%
0.0010.01%

These results indicate that the approximation becomes increasingly accurate as ϵ\epsilon decreases. This supports the validity of our approach, particularly in the regime where ϵ\epsilon is sufficiently small.

Response to Question 2:

Question 2: Can the authors clarify the unit used in Table 4?

Thank you for your suggestion. As shown in Table 4, the values represent the running times relative to ZEN, and thus no explicit unit is indicated. The actual running times are reported in the table below, with all values measured in seconds (s). ZEN runs in less than a second across all datasets, achieving as fast as 0.003s on the Senate dataset.

MethodCoraCiteseerPubmedCora_ca20NewsMN40CongressWalmartSenateHouse
HGNN2.3012.8002.4232.5062.4993.9584.95020.2772.5402.543
HNHN2.0231.9962.8772.0243.6762.7203.92514.4352.2782.267
HCHA3.2903.4854.4482.8502.8944.0726.6758.5583.2974.584
UniGCN4.5531.9326.6932.5116.8254.3158.49722.1212.3421.920
UniGCNII4.1933.8122.0694.8264.0253.8317.47224.4522.2772.419
AllDeepSets15.45618.3318.88916.1088.4209.37825.49960.61913.23211.461
AllSetTransformer2.5973.1678.8653.46214.2734.9246.11451.7023.2603.435
ED-HNN4.2674.4584.4076.58811.0725.66439.79131.1372.3362.487
ZEN (ours)0.2660.7360.8070.2800.2370.1970.0930.6720.0030.007
评论

Thanks for answering all my questions. I'll keep my score.

评论

Thank you again for your insightful review. We sincerely appreciate your time and consideration.

审稿意见
5

This paper proposes ZEN, which is a parameter-free hypergraph neural network. They theoretically derive their parameter-free solution. The experimental study shows the effectiveness and efficiency of ZEN compared with 8 baselines in ten datasets.

优缺点分析

Strong points.

S1. It is surprising that the parameter-free method ZEN outperforms hypergraph neural networks with learning parameters.

S2. Their theoretical analysis supports their structural design of ZEN.

S3. The experimental study is sufficient, validating efficiency and effectiveness in ten datasets, analyzing the effectiveness of components in ZEN as an ablation study, and testing interpretability.

Weak points.

W1. It is unclear why the authors focus only on few-shot node classification. ZEN seems distinct from the few-shot node classification. And, it is unclear what the performance is if the size of the training data varies.

W2. They do not show the actual running time in ZEN.

问题

Q1. Why do the authors focus on few-shot classifications? Is it possible to use ZEN for non-few-shot settings?

Q2. Could you add the actual running time in ZEN?

局限性

Yes, but very short.

最终评判理由

This paper provides a theoretical analysis of parameter-free models and empirical experiments that showed their performance. Their rebuttals clarified the issues of presentations. I believe that this paper is ready to be accepted in NeurIPS.

格式问题

nan

作者回复

Thank you for your recognition of our work and providing the valuable suggestions and constructive comments. Your thoughtful feedback is deeply appreciated and will greatly assist us in refining and strengthening the manuscript.

Response to Question 1:

Question 1: Why do the authors focus on few-shot classifications? Is it possible to use ZEN for non-few-shot settings?

ZEN is a parameter-free model, which gives it a particular advantage in settings where labeled data is scarce and training is challenging. While ZEN was orginally designed for few-shot setting, our results in the tables below show it performs strongly with more training samples as well. ZEN achieves top average ranks in both 10-shot and 20-shot settings, suggesting it scales well beyond few-shot scenarios.

10-shot classification (10 training nodes per class)

MethodCoraCiteseerPubmedCora_ca20NewsMN40CongressWalmartSenateHouseAvg. Rank
HGNN59.1±5.849.3±3.160.5±7.064.9±4.475.5±1.294.9±0.287.2±0.946.9±1.258.0±2.564.3±4.46.1
HNHN42.5±5.042.8±2.858.9±3.346.7±3.551.2±4.994.0±0.647.8±1.817.1±2.269.7±6.170.3±7.69.9
HCHA59.0±5.449.7±2.960.0±5.664.3±5.675.6±1.394.9±0.287.2±0.947.1±1.757.9±2.863.9±4.45.8
UniGCN60.9±5.851.1±3.061.2±6.966.9±3.073.7±1.796.2±0.291.4±0.745.4±1.462.6±2.570.3±3.04.1
UniGCNII56.8±5.149.2±2.961.1±6.463.2±3.870.9±2.197.0±0.287.7±1.928.7±2.575.1±4.373.7±2.94.8
AllDeepSets57.3±4.450.6±2.661.4±5.162.8±4.162.2±2.395.3±0.372.6±5.835.5±2.769.5±6.770.0±3.16.9
AllSetTransformer57.5±5.054.1±2.463.4±6.667.6±4.272.7±1.495.9±0.183.1±2.245.4±3.773.1±5.571.0±4.53.9
ED-HNN58.8±4.351.1±2.860.6±7.064.7±3.771.1±2.296.3±0.290.7±1.847.2±2.366.4±9.362.5±7.94.9
ZEN (ours)61.4±4.259.2±2.368.8±3.465.6±3.973.0±2.497.9±0.186.5±2.647.1±3.274.7±1.973.9±6.41.7

20-shot classification (20 training nodes per class)

MethodCoraCiteseerPubmedCora_ca20NewsMN40CongressWalmartSenateHouseAvg. Rank
HGNN62.5±3.156.6±1.969.7±3.970.3±1.976.7±1.195.1±0.287.8±0.549.5±2.460.7±3.264.8±2.25.4
HNHN48.3±3.950.1±2.362.6±1.652.9±2.952.7±6.395.2±0.549.9±3.021.8±1.873.5±3.972.0±4.19.7
HCHA64.2±3.956.8±2.169.4±3.370.0±1.976.7±1.095.1±0.187.9±0.449.4±2.261.1±3.364.9±2.34.9
UniGCN65.4±3.358.5±1.270.8±2.671.3±1.872.6±2.196.6±0.191.6±0.349.0±2.162.0±2.372.7±0.93.2
UniGCNII64.3±2.957.0±1.268.9±3.468.6±2.574.1±1.597.6±0.188.3±1.932.0±1.577.6±2.076.1±0.93.6
AllDeepSets62.4±3.757.8±1.867.1±2.265.9±1.367.6±2.596.7±0.377.7±5.844.0±2.473.1±2.273.3±2.56.6
AllSetTransformer63.7±1.659.5±2.169.7±2.369.8±2.274.2±1.296.0±0.289.2±0.549.3±1.876.1±3.875.6±2.73.4
ED-HNN64.2±3.057.3±1.669.6±3.469.3±1.670.9±1.996.7±0.290.1±2.150.6±3.372.3±6.771.3±3.74.3
ZEN (ours)67.9±1.464.0±1.471.3±1.672.0±2.073.7±1.298.2±0.188.0±2.951.3±1.174.1±5.276.7±0.41.9

Response to Question 2:

Question 2: Could you add the actual running time in ZEN?

Thank you for the suggestion. The actual running times are reported in the table below, with all values measured in seconds (s). ZEN runs in less than a second across all datasets, achieving as fast as 0.003s on the Senate dataset.

MethodCoraCiteseerPubmedCora_ca20NewsMN40CongressWalmartSenateHouse
HGNN2.3012.8002.4232.5062.4993.9584.95020.2772.5402.543
HNHN2.0231.9962.8772.0243.6762.7203.92514.4352.2782.267
HCHA3.2903.4854.4482.8502.8944.0726.6758.5583.2974.584
UniGCN4.5531.9326.6932.5116.8254.3158.49722.1212.3421.920
UniGCNII4.1933.8122.0694.8264.0253.8317.47224.4522.2772.419
AllDeepSets15.45618.3318.88916.1088.4209.37825.49960.61913.23211.461
AllSetTransformer2.5973.1678.8653.46214.2734.9246.11451.7023.2603.435
ED-HNN4.2674.4584.4076.58811.0725.66439.79131.1372.3362.487
ZEN (ours)0.2660.7360.8070.2800.2370.1970.0930.6720.0030.007
评论

I really appreciate your additional experiments. Your response clarifies my concern; the advantage of a parameter-free model is a small training time, but few-shot learning takes a small training time even in conventionally trained models, so the benefit seems small. For example, even if 0.266 sec becomes 4.267, it may not have a very large impact practically.

My score is already high enough, so I will keep my score.

评论

Thank you again for your thoughtful review and helpful suggestions. Please let us know if any further clarification would be helpful.

审稿意见
4

This paper introduces ZEN (Zero-Parameter Hypergraph Neural Network), a linear and training-free model for few-shot node classification on hypergraphs. ZEN is built upon a unified linearized formulation of existing HNNs and introduces two key innovations: (1) a tractable closed-form solution (TCS) for computing the weight matrix without iterative optimization, and (2) a redundancy-aware propagation (RAP) mechanism to eliminate self-information in multi-hop neighborhoods. The method shows consistent performance improvements over eight strong baselines across 11 real-world datasets. The model is interpretable and the paper includes a detailed ablation and case study.

优缺点分析

Strengths:

  1. The formulation of a fully parameter-free HNN using closed-form solutions and a redundancy-aware propagation is both original and impactful, especially for low-resource, few-shot learning scenarios.
  2. The theoretical contributions, including the derivation of the closed-form approximation (Theorem 1) and its assumptions, are carefully motivated and reasonably justified.
  3. The theoretical contributions, including the derivation of the closed-form approximation (Theorem 1) and its assumptions, are carefully motivated and reasonably justified.

Weaknesses:

  1. As acknowledged by the authors, ZEN is specifically designed for node classification. Extensions to other hypergraph tasks (e.g., hyperedge prediction, clustering) are left as future work. This means the scope of applicability of this work is limited. The authors may consider analyzing the effectiveness of the work in other tasks.
  2. The paper discusses the success of linear GNNs such as SGC and APPNP, but does not directly compare ZEN to their hypergraph counterparts or generalizations. A direct comparison with a naïve linearized hypergraph model (beyond HGNN) would provide a stronger baseline.
  3. ZEN is limited to 2-hop propagation. While this is justified by computational cost and redundancy, it may restrict representational power on deeper hypergraph structures. A discussion on possible approximations or sampling strategies for deeper propagation would be a valuable addition.
  4. Lack of baselines. The authors focus on the task of few-shot node classification. However, they only conduct experiments with hypergraph neural networks (HNN) models. There are multiple few-shot node classification works that can act as baselines.

问题

Did the authors consider involving more baselines for evaluation?

局限性

See weaknesses.

最终评判理由

I appreciate the additional experiments and clarifications, especially the extended baseline comparisons. I will keep my original score.

格式问题

NA

作者回复

Thank you for your valuable comments. Your feedback is greatly appreciated and helps us improve the clarity and quality of our work. We will do our best to address them in the planned revision.

Response to Weakness 1:

Weakness 1: As acknowledged by the authors, ZEN is specifically designed for node classification. Extensions to other hypergraph tasks (e.g., hyperedge prediction, clustering) are left as future work. This means the scope of applicability of this work is limited. The authors may consider analyzing the effectiveness of the work in other tasks.

We fully agree on the importance of assessing our method on other tasks. Some tasks, such as hyperedge prediction, can be reformulated as node classification, allowing ZEN to be applied without modification—for example, by treating hyperedges as nodes using Xe=De1HX,He=HX_e=D_e^{-1}H^\top X, H_e=H^\top. For tasks not easily reducible to this setting, ZEN may require further adaptation. We plan to explore such extensions in future work.

Response to Weakness 2:

Weakness 2: The paper discusses the success of linear GNNs such as SGC and APPNP, but does not directly compare ZEN to their hypergraph counterparts or generalizations. A direct comparison with a naïve linearized hypergraph model (beyond HGNN) would provide a stronger baseline.

Thank you for your detailed feedback. We have additionally included three representative linear GNNs (SGC [1], APPNP [2], and SSGC [3]) and three linearized hypergraph models based on UniGCNII [4], AllDeepSets [5], and ED-HNN [6]. For GNNs, we applied clique expansion to convert the hypergraph into a pairwise graph. As shown in the table below, ZEN demonstrates consistently strong performance, outperforming all linear GNN and linearized hypergraph baselines on the majority of datasets.

MethodCoraCiteseerPubmedCora_ca20NewsMN40CongressWalmartSenateHouse
SGC44.1±9.840.4±5.357.2±5.550.8±5.958.1±5.191.6±0.772.5±3.124.1±3.251.3±2.752.3±1.4
APPNP45.8±9.539.6±4.755.4±7.150.0±6.759.9±5.892.8±0.369.1±6.824.7±3.371.2±5.470.0±8.4
SSGC42.6±10.940.7±4.757.1±4.952.6±6.759.9±5.293.8±0.281.1±2.426.6±3.473.0±0.771.7±3.2
Lin HGNN42.7±8.434.3±9.451.7±6.649.6±6.968.0±6.094.5±0.583.7±5.029.0±6.455.8±5.157.8±6.5
Lin UniGCNII39.4±7.440.0±4.056.9±4.949.8±6.159.5±5.296.6±0.473.4±3.117.9±1.771.2±10.868.2±10.6
Lin AllDeepSets40.5±8.240.8±4.556.1±6.051.0±5.866.6±5.096.6±0.390.4±1.526.4±3.061.0±4.765.1±13.1
Lin ED-HNN42.0±5.639.6±4.255.8±5.447.4±5.861.5±6.097.3±0.478.4±2.718.8±1.671.1±9.970.2±9.3
ZEN (ours)51.9±10.149.1±4.862.6±3.960.0±6.268.6±4.897.6±0.387.0±4.843.9±3.170.4±10.073.2±6.3

Response to Weakness 3:

Weakness 3: ZEN is limited to 2-hop propagation. While this is justified by computational cost and redundancy, it may restrict representational power on deeper hypergraph structures. A discussion on possible approximations or sampling strategies for deeper propagation would be a valuable addition.

Thank you for your suggestion. We agree that some datasets may require models with higher-hop propagations for better expressivity. Without approximation, the exact computation of RSI requires high computational cost.

One possible approximation is to estimate the probability that a random walker returns to the starting node after ll steps, where ll is the number of hops that we aim to model. A simple pseudocode for this strategy is as follows:

  • Start at node ii.
  • At each step:
    1. Uniformly sample a hyperedge incident to the current node.
    2. Uniformly sample a node connected to that hyperedge.
  • Repeat this process for ll steps.
  • Record whether the walk ends at the original node ii. The return probability may serve as an approximation of node-wise self-information at a given hop.

Another possible approximation is to use Hutchinson's Estimator, which provides an unbiased stochastic estimate of the diagonal entries. The equation is given by:

diag(A)1mk=1mz(k)(Az(k))\operatorname{diag}(A) \approx \frac{1}{m} \sum_{k=1}^m z^{(k)} \odot \left( A z^{(k)} \right)

where:

  • mm is the number of random probe vectors,
  • each z(k)Rnz^{(k)} \in \mathbb{R}^n is a random vector with entries independently sampled from the Rademacher distribution (i.e., each entry is +1+1 or 1-1 with equal probability),
  • \odot denotes the element-wise (Hadamard) product.

The expectation satisfies

diag(A)=E[z(Az)].\operatorname{diag}(A) = \mathbb{E} \left[ z \odot (A z) \right].

Hutchinson's Estimator does not require explicit storage of the entire matrix AA; instead, it relies solely on matrix-vector multiplications with randomized vectors zz. This characteristic renders Hutchinson's Estimator particularly suitable and computationally efficient for hypergraph structures.

We consider these approaches to be promising directions for extending ZEN to scenarios where deeper propagation may provide additional benefits.

Response to Weakness 4:

Weakness 4: Lack of baselines. The authors focus on the task of few-shot node classification. However, they only conduct experiments with hypergraph neural networks (HNN) models. There are multiple few-shot node classification works that can act as baselines.

We appreciate the suggestion to strengthen our evaluation by including additional baselines. To the best of our knowledge, there has been no hypergraph model explicitly addressing few-shot node classification as a direct keyword. In response, we have included two semi-supervised hypergraph models—LEGCN [7] and HyperND [8]—as additional baselines. LEGCN utilizes a line expansion approach to adapt hypergraphs to conventional GNN architectures, while HyperND introduces a diffusion-based mechanism for improved label propagation in hypergraphs. As shown in the table below, ZEN consistently demonstrates strong performance, outperforming all additional baselines across the majority of datasets.

MethodCoraCiteseerPubmedCora_ca20NewsMN40CongressWalmartSenateHouse
LEGCN37.6±5.237.3±3.858.1±4.037.6±5.249.4±4.393.3±0.772.1±3.4O.O.M71.3±8.572.9±7.0
HyperND39.1±6.235.3±4.856.5±4.541.7±4.654.7±3.891.4±0.573.6±4.913.6±1.773.9±7.572.8±7.1
ZEN (ours)51.9±10.149.1±4.862.6±3.960.0±6.268.6±4.897.6±0.387.0±4.843.9±3.170.4±10.073.2±6.3

References

[1] Felix Wu, Amauri Souza, Tianyi Zhang, Christopher Fifty, Tao Yu, and Kilian Weinberger. Simplifying graph convolutional networks. In PMLR, 2019. URL https://proceedings.mlr.press/v97/wu19e.html.

[2] Johannes Gasteiger, Aleksandar Bojchevski, and Stephan Günnemann. Predict then propagate: Graph neural networks meet personalized pagerank, 2022. URL https://arxiv.org/abs/1810.05997.

[3] Hao Zhu and Piotr Koniusz. Simple spectral graph convolution. In International conference on learning representations, 2021.

[4] Jing Huang and Jie Yang. Unignn: a unified framework for graph and hypergraph neural networks, 2021. URL https://arxiv.org/abs/2105.00956.

[5] Eli Chien, Chao Pan, Jianhao Peng, and Olgica Milenkovic. You are allset: A multiset function framework for hypergraph neural networks, 2022. URL https://arxiv.org/abs/2106.13264.

[6] Peihao Wang, Shenghao Yang, Yunyu Liu, Zhangyang Wang, and Pan Li. Equivariant hypergraph diffusion neural operators, 2023. URL https://arxiv.org/abs/2207.06680.

[7] Yang, Chaoqi and Wang, Ruijie and Yao, Shuochao and Abdelzaher, Tarek. Semi-supervised hypergraph node classification on hypergraph line expansion. CIKM, 2022. URL https://arxiv.org/abs/2005.04843.

[8] Konstantin Prokopchik, Austin R Benson, and Francesco Tudisco. Nonlinear feature diffusion on hypergraphs. PMLR, 2022. URL https://arxiv.org/pdf/2103.14867.

评论

Thank you for the detailed and thoughtful rebuttal. I appreciate the additional experiments and clarifications, especially the extended baseline comparisons. I will keep my original score.

评论

Thank you for your valuable feedback. We sincerely appreciate the time and effort you put into reviewing our work.

审稿意见
4

This paper focuses on the few-shot node classification task on hypergraphs, and proposes a model, ZEN, to capture high-order relationships while maintaining strong generalization ability and scalability. Experiments on 11 real-world datasets demonstrate the effectiveness and efficiency of the proposed method.

优缺点分析

S1. This paper presents a comprehensive study on liner hypergraph neural networks.

S2. Two techniques, tractable closed-form approximation and redundancy-aware propagation, are developed to improve the efficiency and the corresponding time complexity analysis is provided.

S3. Experiments on real-world datasets have been conducted and the results show the superiority of the proposed method.

W1. While the paper addresses the problem of few-shot node classification, it does not follow the conventional nn-way kk-shot learning paradigm in the evaluation setup. This deviation from standard practices should be discussed or justified.

W2. The ablation study shows that removing certain components of ZEN may lead to improved performance, which requires further discussion.

问题

Q1. More experiments are needed to evaluate the effectiveness of the proposed method under the nn-way kk-shot setting.

Q2. The ablation study (Table 6) shows that ZEN underperforms compared to the No TCS variant on the 20News and MN40 datasets, and compared to the No RAP variant on the Senate dataset. Thus, the limitations of ZEN need to be clarified in the revised manuscript.

局限性

yes

最终评判理由

The authors have conducted additional experiments in the n-way k-shot paradigm and provided a more thorough ablation study. Thus, I will maintain my original score.

格式问题

no

作者回复

Thanks for your thoughtful feedback. We appreciate your constructive review, which has helped us improve the clarity and quality of the paper.

Response to Weakness 1:

Weakness 1: While the paper addresses the problem of few-shot node classification, it does not follow the conventional n-way k-shot learning paradigm in the evaluation setup. This deviation from standard practices should be discussed or justified.

To address the concerns, we would like to respectfully provide clarification on our task formulation. We define our few-shot node classification task as a transductive, semi-supervised node classification under severe label constraints. In standard n-way k-shot learning, the entire set of classes is typically partitioned into disjoint training and test sets. From each set, n classes are sampled, and k labeled examples per class are selected to construct support sets for episodic training. The model is trained on a variety of such tasks in order to evaluate its capacity for adaptation and generalization to novel, unseen tasks. In contrast, our setting can be regarded as a more challenging variant, where the class set remains fixed and identical to the test classes, and n corresponds to the total number of classes.

To avoid potential confusion with n-way k-shot learning from the meta-learning literature, we are open to revising our terminology in the revised version of our paper to reflect this distinction more clearly; e.g., hard few-shot node classification or semi-supervised node classification with hard label scarcity.

On the other hand, it would be interesting to see how our approach fits into the standard n-way k-shot setting. To the best of our knowledge, this setup has not been explicitly explored for general hypergraphs. Accordingly, we designed a new evaluation protocol by drawing inspiration from the experimental setup of Meta-GNN [1] and adopting hyperparameter configurations aligned with those used in ED-HNN [2]. The results are reported in the table below.

MethodCora 2-way 3-shotCora 2-way 1-shotCiteseer 2-way 3-shotCiteseer 2-way 1-shot
HGNN64.4±0.155.9±0.160.1±0.254.9±0.2
UniGCNII68.0±0.158.9±0.265.2±0.156.8±1.6
AllDeepSets52.3±0.248.7±0.251.1±0.250.2±0.2
ZEN (ours)73.3±0.162.6±0.171.5±0.162.3±0.1

The strong performance of ZEN in this setting highlights its potential as an effective approach for standard few-shot tasks on hypergraphs. Interestingly, ZEN yields promising results even when applied solely to test episodes, suggesting a potentially valuable and underexplored direction. We appreciate the opportunity to highlight this aspect and will consider it an important avenue for future research.

Response to Weakness 2:

Weakness 2: The ablation study shows that removing certain components of ZEN may lead to improved performance, which requires further discussion.

It is true that removing RAP or TCS leads to performance improvements on some datasets (Congress and Senate for RAP, 20News and MN40 for TCS). To analyze this behavior, we examine the ratio of hyperedges to nodes (i.e., density), which we believe an important factor to understand the characteristic of a dataset. The results are reported in the table below.

DatasetCoraCiteseerPubmedCora-CA20NewsMN40CongressWalmartSenateHouse
Density0.58350.32580.40410.39590.00621.000048.39460.78681.11600.2636

We observe that RAP improves performance on datasets with low density, while it slightly degrades performance on Congress and Senate, which have densities greater than 1. This aligns with the intuition: as the number of neighbors increases, each node's contribution diminishes due to degree normalization, reducing the relative influence of self-information. In such cases, removing self-information via RAP may have limited or even adverse effects.

As for TCS, it assumes good node embeddings. In the ablation study, the impact of removing TCS is more significant on 20News than on MN40. This may due to the extremely low density of 20News, where ZEN, which relies solely on propagation to refine embeddings, may have struggled to produce sufficiently informative representations.

On the other hand, the result on MN40 may have been influenced by its large number of classes — 40, the highest among all datasets. The relative Frobenius norm error of TCS is given by:

1λ12M1+1λ22M2+1λ32M31ϵ(1λ1M1+1λ2M2+1λ3M3)F1λ12M1+1λ22M2+1λ32M3F\frac{\left\|\frac{1}{\lambda_1^2}M_1+\frac{1}{\lambda_2^2}M_2+\frac{1}{\lambda_3^2}M_3 - \frac{1}{\epsilon}\left(\frac{1}{\lambda_1}M_1+\frac{1}{\lambda_2}M_2+\frac{1}{\lambda_3}M_3\right)\right\|_F}{\left\|\frac{1}{\lambda_1^2}M_1+\frac{1}{\lambda_2^2}M_2+\frac{1}{\lambda_3^2}M_3\right\|_F}

As shown in the table below, this approximation error slightly increases with the number of classes:

εεc=4c=4c=40c=40
0.11.1064%1.1753%
0.010.1006%0.1010%
0.0010.0100%0.0100%

Although the error remains small overall, the increase suggests that the approximation may become less accurate as the number of classes grows, which could partially explain the results observed on MN40.

References

[1] Fan Zhou, Chengtai Cao, Kunpeng Zhang, Goce Trajcevski, Ting Zhong, and Ji Geng. Meta-gnn: On few-shot node classification in graph meta-learning. CIKM, 2019. URL https://doi.org/10.1145/3357384.3358106.

[2] Peihao Wang, Shenghao Yang, Yunyu Liu, Zhangyang Wang, and Pan Li. Equivariant hypergraph diffusion neural operators, 2023. URL https://arxiv.org/abs/2207.06680.

评论

Thank you for your response. My concerns have been addressed, so I will not change my score.

评论

We greatly appreciate your comments. Thank you for the time and effort you dedicated to reviewing our paper.

评论

We sincerely appreciate the reviewers’ insightful comments, which have greatly helped improve our submission. We have made every effort to address the concerns raised and would be happy to provide further clarification should any questions remain.

最终决定

This paper introduces ZEN, a parameter-free hypergraph neural network for few-shot node classification. ZEN is built on a unified linearized formulation of HNNs. Two key novelties include a tractable closed-form solution for weight computation and a redundancy-aware propagation mechanism to avoid self-information leakage. Extensive experiments across 11 real-world datasets show consistent improvements over strong baselines.

The main strengths lie in its originality and simplicity: the parameter-free formulation is theoretically grounded and empirically outperforms previous hypergraph neural networks. The paper is also clearly written, with thorough ablations and case studies.

The weaknesses are relatively minor. The method is currently limited to node classification; broader applicability to other hypergraph tasks remains unexplored. The evaluation setup only partially follows the conventional few-shot learning paradigm, which could be better justified. ZEN is restricted to 2-hop propagation, which may limit representation capacity, and some comparisons (e.g., with more few-shot methods or deeper hypergraph variants) are missing. Finally, further clarification on time complexity, error bounds, and reporting conventions would improve the presentation.

Overall, despite these weaknesses, the paper makes a solid and meaningful contribution to hypergraph learning. Its theoretical insights, empirical effectiveness, and efficiency justify acceptance. I recommend acceptance.