PaperHub
7.3
/10
Poster4 位审稿人
最低4最高5标准差0.5
4
4
5
5
3.8
置信度
创新性2.8
质量3.3
清晰度2.5
重要性2.5
NeurIPS 2025

Functional Virtual Adversarial Training for Semi-Supervised Time Series Classification

OpenReviewPDF
提交: 2025-05-07更新: 2025-10-29
TL;DR

We propose functional Virtual Adversarial Training (f-VAT) that incorporates the functional structure of data into perturbations.

摘要

Real-world time series analysis, such as healthcare, autonomous driving, and solar energy, faces unique challenges arising from the scarcity of labeled data, highlighting the need for effective semi-supervised learning methods. While the Virtual Adversarial Training (VAT) method has shown promising performance in leveraging unlabeled data for smoother predictive distributions, straightforward extensions of VAT often fall short on time series tasks as they neglect the temporal structure of the data in the adversarial perturbation. In this paper, we propose the framework of functional Virtual Adversarial Training (f-VAT) that can incorporate the functional structure of the data into perturbations. By theoretically establishing a duality between the perturbation norm and the functional model sensitivity, we propose to use an appropriate Sobolev ($H^{-s}$) norm to generate structured functional adversarial perturbations for semi-supervised time series classification. Our proposed f-VAT method outperforms recent methods and achieves superior performance in extensive semi-supervised time series classification tasks (e.g., up to $ \approx 9$ % performance improvement). We also provide additional visualization studies to offer further insights into the superiority of f-VAT.
关键词
Semi-Supervised classification; Virtual Adversarial Training; Function Space; Sobolev Space

评审与讨论

审稿意见
4

The paper points out that the Virtual Adversarial Training falls short on time series as the added perturbations bounded by Euclidean norm disrupt time series patterns. The paper proposes the Sobolev norm to replace the Euclidean norm to generate structured functional adversarial perturbations.

优缺点分析

Strengths

This paper points out that the Virtual Adversarial Training falls short on time series: the perturbation vector constrained by Euclidean norm disrupts temporal patterns. This insight is very enlightening for designing effective adversarial methods tailored for time series.

Weakness 1

Why is time series considered functional data, while images are not? Time series is a function of some interval TT, and images can also be considered functions in a spatial context.

Despite "Functional" being a key term in the title, I still do not understand its specific meaning in this paper after reading the introduction. Additionally, it's unclear what unique designs are applied to time series (at least not sufficiently articulated in the introduction). From this perspective, I believe the writing of the paper needs improvement.

Weakness 2

The clarity and motivations in Section 3 are confusing.

  • W 2.1: How do Theorem 3.3 and Theorem 3.4 assist in the method design? The condition of Theorem 3.3 is that the model is linear, and the condition of Theorem 3.4 is ϵ0+\epsilon\rightarrow0^+. These two conditions are not satisfied in practical training.
  • W 2.2: Based on Line 163~170, the conclusions of Theorem 3.3 and Theorem 3.4 state that "if we want to impose a stronger norm on the sensitivity of the method, we should relax the norm of the perturbation." How is this related to the introduction of the Sobolev norm?
  • W 2.3: Why does the Sobolev space penalize high-frequency oscillations? Is the benefit of the Sobolev space solely that it penalizes high-frequency noise? This can be achieved through common high-frequency filtering, which has already been widely applied in existing time series analysis methods.
  • W 2.4: Compared with VAT:
    • The VAT only needs a hyperparameter ϵ\epsilon, but f-VAT needs four hyperparameters ϵ\epsilon, ss, LL, and η\eta. VAT only needs to compute once to obtain the optimal perturbation rr, whereas f-VAT requires LL iterations and the hyperparameter η\eta.
    • Do the perturbation vector rr and the model parameters θ\theta share the same learning rate η\eta as shown in Algorithm 1?
    • In addition to the Sobolev norm order, the sensitive analysis of other hyperparameters should be included, along with a comparison of the training time between VAT and f-VAT.
  • W 2.5: In Line 6 of Algorithm 1, why adopt gradient ascent and normalize rir_i? In the original paper of VAT, the optimal rr is obtained by calculating the first dominant eigenvector of the Hessian matrix rD(r,x,θ^)r=0\nabla\nabla_r D(r, x, \hat{\theta})|_{r=0}. Is the approach not suitable for f-VAT? For instance, calculate the optimal rr whose Sobolev norm is ϵ\epsilon in the direction of the first dominant eigenvector?
  • W 2.6: Could the authors provide the formula of Hs||\cdot|| _ {H ^ {-s}} as well, rather than just Hs||\cdot|| _ {H ^ {s}}. It is confusing that Eq 11 only introduces Hs||\cdot|| _ {H ^ {s}}, while Algorithm 1 actually constrains rr using Hs||\cdot|| _ {H ^ {-s}}. Additionally, in Eq 11 and its context, LL and LNL _ N seem to be confused.

Weakness 3

This paper lacks a discussion of related works, particularly in the area of semi-supervised time series. Additionally, the experiments do not compare with the latest semi-supervised methods [1, 2]. More importantly, the paper only includes three datasets from UCR and three from UEA. However, TS-TFC [1] includes 106 datasets from UCR. As far as I know, these time series datasets are relatively small-scale and highly heterogeneous, making the evaluation in the paper seem inadequate.

[1] Temporal-Frequency Co-training for Time Series Semi-supervised Learning. AAAI, 2023.

[2] Self-supervised Contrastive Representation Learning for Semi-supervised Time-Series Classification. TPAMI, 2022.

Weakness 4

In Section 4.3, the authors claim that functional adversarial perturbations add subtle disturbances only in non-critical areas, while keeping them small in important regions. However, I did not observe this phenomenon. Where are the critical regions versus the non-critical regions? It appears that f-VAT applies small perturbations at each time point, while VAT applies larger perturbations at certain time points.

Therefore, a straightforward improvement to VAT could involve not only constraining the norm of the overall perturbation vector but also limiting the magnitude of perturbations at each time point, or constraining the variance of perturbations at each time point. Would this be viable?

问题

Please refer to Weaknesses.

局限性

The evaluation is limited and some statements are confusing.

最终评判理由

I appreciate the thorough experimental results provided by the authors, and I would increase my rating to 4.

格式问题

None.

作者回复

We would like to point out that the contribution of this paper is to propose the theory-inspired f-VAT algorithm that performs well in practice, which is supported by our experiments (and additional ones, see answers below) that f-VAT consistently outperforms other competitive baselines in all settings. While the theory considers simplified settings, it provides us with key insights into designing the f-VAT algorithm, particularly in constructing perturbation that aligns with the time series structure of the data. Therefore, we believe that our paper is novel and significant both empirically and theoretically.

W1.1

We agree that both time series and images can be viewed as functional data, but their intrinsic structures can be different. In this paper, we mainly focus on time series data, but our general f-VAT framework can be applied to broader settings. Exploring f-VAT for functional data with other structures, such as image data, can be one of our future directions.

W1.2

We use the term "Functional" to emphasize the function structure of the data, particularly for time series data. Focusing on the additional function structure of the data, our novel f-VAT method generates adversarial perturbations that align with the key temporal structure, i.e., the dynamic correlations within variables changing over time steps. This is the key improvement of f-VAT over VAT.

W2.1

Our theory aims to provide insights into designing adversarial perturbations when the data have additional structure, particularly for time series data. We believe that our existing theory has already provided much insight for the design of f-VAT with the duality theory, and due to the highly complex nature of neural networks, further theoretical analysis would be beyond the scope of this paper. Moreover, strong empirical results on various time series datasets demonstrate the effectiveness of our method.

Please see our response to the next point for more details on how the theory is used in the method design.

W2.2

In our theory, we first establish the general theory on duality between the perturbation norm and the model sensitivity norm (Theorem 3.3 and Theorem 3.4). However, we still need to choose a norm (or the space for the norm) in the concrete implementation. For time series data, it is natural to consider Sobolev spaces HsH^s (of positive order) to capture the smoothness/regularity for model sensitivity. Then, with the duality theory, we transfer HsH^s to HsH^{-s} for the perturbation norm, where we use the properties of Sobolev spaces that the dual space of HsH^s is (Hs)=Hs(H^s)^* = H^{-s}. In fact, it is one of the key points in our paper to use the Sobolev norm HsH^{-s} of negative order for the perturbation, which is critical for the success of the proposed f-VAT.

We would like to clarify that our point is "The Sobolev norm allows us to properly control the model's sensitivity to high-frequency noises''. This can be seen from duality theory: the Sobolev norm HsH^{-s} imposed on the perturbation effectively serves as a regularization term that penalizes the model sensitivity in terms of the Sobolev space HsH^s.

W2.3

The infinite-dimensional nature of the Sobolev space allows for a richer specification of perturbation constrains that align with the geometry of functional data like periodicity, rather than penalizing high-frequency filtering. Thus, the functional adversarial perturbation generated by f-VAT in Sobolev space can capture the temporal structure of time series data.

W2.4

  • W2.4.1: We would like to clarify that f-VAT and VAT require the same set of hyperparameters: ϵ\epsilon controls the perturbation radius; LL is the number of power-iteration steps (typically set to L=1L=1); and η\eta is the step size for gradient ascent during the power-iteration. The only additional hyperparameter of f-VAT is the order ss of norm Hs\|\cdot\|_{H^s}.

  • W2.4.2: Yes, both the perturbation vector rr and the model parameters share the same learning η\eta, as clarifies in [Takeru et al., 2018].

  • W2.4.3: Compared to original VAT, f-VAT only introduces an additional hyperparameter ss. We evaluate the performance on more datasets, which shows that setting s=2s=2 achieves the best performance. Table shows that the extra computational cost remains in the same order as VAT. Due to the page limit, please refer to our replies to Reviewer #Jy1d Q1 & W2.

W2.5

The original VAT also adopts gradient ascent and normalizes rir_i to approximate the first dominant eigenvector of the Hessian matrix D(r,x,θ^)r=0\nabla\nabla D(r,x,\hat{\theta})|_{r=0}, while the proposed f-VAT adopts the same power-iteration procedure with L=1L=1. The key difference is that the projection into the ϵ\epsilon ball in the L2L_2 space is replaced by the HsH^{-s} space.

W2.6

We apologize for the typo in Eq.(11) and the confusion caused by the notation. In fact, as emphasized in the paragraph before Eq.(11), the formula rHα2rN(IN+LN)αrN\|r\|_{H^\alpha}^2 \approx r_N^\top (I_N + L_N)^\alpha r_N applies to all αR\alpha \in \mathbb{R}, including negative orders, so we do not distinguish ss or s-s. Here, the operator LL should always be replaced by LNL_N which represents the discrete Laplacian operator. We have corrected the notation in the revised version of the paper, and we hope this resolves the confusion.

W3

We have add more detailed discussion of literature review in the revised version. We follow the settings [Haoyi et al., 2021] to compare f-VAT with other competitive baselines and report results on six UCR/UEA datasets. We randomly sample 30 datasets from UCR/UEA datasets and add a state-of-the-art Class-Aware Temporal and Contextual Contrasting (TS-TCC) [Emadeldeen et al., 2023] within available resources. Table shows that tVAT consistently outperforms other baselines in all settings. Please see our additional comments for more experiment details.

Average accuracy (%) and average rank across the dataseets under different label ratios:

Method10%20%40%
AvgAccAvgRankAvgAccAvgRankAvgAccAvgRank
SupL35.316.6736.927.0037.157.33
PI53.093.9355.164.4063.604.47
MTL45.195.7045.725.8746.116.80
SemiTime56.533.5758.933.7769.023.17
TapNet58.673.7060.413.4070.283.20
TS‑TCC58.073.3759.843.6763.274.27
f-VAT65.851.5068.871.5076.241.53

Additionally, we construct empirical on several large-scale China Securities Index (CSI) datasets (i.e., CSI 50 and 500 futures) spanning from 2020 to 2023 for predicting directions (upward or downward) of futures prices [Zihao et al., 2019]. The dataset collects over 42×10342\times 10^{3} records spanning from 2020 to 2022, and each time step contains bid/ask prices and corresponding volumes. Table presents the performance of various semi-supervised methods across different label ratios, and shows that f-VAT significantly outperforms other competitive baselines on futures datasets, especially on more volatile CSI 500 futures.

The performance comparison on domestic futures datasets:

FuturesRatiosupLPIMTLSemiTimeTapNetTS-TCCf-VAT
10%40.05 ±\pm 1.3842.87 ±\pm 1.4553.94 ±\pm 0.1355.19 ±\pm 0.7754.62 ±\pm 0.5455.72 ±\pm 1.1658.64 ±\pm 0.53
5020%45.08 ±\pm 1.4547.26 ±\pm 0.7354.97 ±\pm 0.6156.69 ±\pm 0.5056.93 ±\pm 1.2858.21 ±\pm 0.4462.09 ±\pm 0.26
40%50.69 ±\pm 3.5352.42 ±\pm 1.2156.97 ±\pm 0.5757.34 ±\pm 0.2059.75 ±\pm 1.0959.80 ±\pm 1.3264.78 ±\pm 0.45
10%34.23 ±\pm 0.2444.00 ±\pm 0.0639.53 ±\pm 1.3538.86 ±\pm 2.0440.53 ±\pm 0.8339.79 ±\pm 1.5443.77 ±\pm 0.73
50020%35.38 ±\pm 1.0646.57 ±\pm 0.4845.26 ±\pm 0.2946.04 ±\pm 0.1244.58 ±\pm 1.7045.05 ±\pm 1.7452.14 ±\pm 0.25
40%43.85 ±\pm 1.4949.25 ±\pm 0.5547.66 ±\pm 1.5050.65 ±\pm 1.0551.50 ±\pm 1.3654.29 ±\pm 0.8758.66 ±\pm 0.46

W4.1

The critical areas refer to the input feature essential for prediction. Figure 4 visualizes the essential "shapelets'' in yellow and the irrelevant regions in blue. Our proposed f-VAT can adaptively concentrate its perturbations on these non-critical regions, aligning with the original samples across all key time steps, while the original VAT may introduce large spikes in some critical regions and disrupt the underlying trend structure.

W4.2

The straightforward improvement to VAT may be unsuitable for time series. If we limit the magnitude of perturbations at each time step, the order of time steps can be arbitrarily permutable to destroy the temporal structure. On the other hand, our proposed f-VAT can capture the key trend structure, i.e., the implicit correlations within variables changing over time, by generating theory-inspired functional adversarial perturbations.

  • [Takeru et al., 2018] Takeru Miyato, etc., Virtual adversarial training: a regularization method for supervised and semi-supervised learning. TPAMI, 2018.

  • [Haoyi et al., 2021] Haoyi Fan, etc, Semi-supervised time series classification by temporal relation prediction. ICASSP, 2021.

  • [Emadeldeen et al., 2023] Emadeldeen Eldele, et. al., Self-supervised contrastive representation learning for semi-supervised time-series classification. TPAMI, 2023.

  • [Zihao et al., 2019] Zihao Zhang etc, Deeplob: Deep convolutional neural networks for limit order books, 2019.

评论

Due to character limitations, we split each label-ratio table into two parts and posted them in separate comments for your reference.

DatasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCfVATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankCATCC_rankf-VAT_rank
ACSF110.00 ±\pm 0.1014.06 ±\pm 1.5610.93 ±\pm 2.9029.68 ±\pm 2.8021.10 ±\pm 2.7925.00 ±\pm 2.8224.22 ±\pm 2.8260.42 ±\pm 2.6386725341
Adiac5.98 ±\pm 2.827.45 ±\pm 2.872.54 ±\pm 0.0414.08 ±\pm 2.8920.08 ±\pm 1.4334.39 ±\pm 0.1430.44 ±\pm 2.9038.85 ±\pm 2.6276854231
AllGestureWiimoteX12.68 ±\pm 1.2152.23 ±\pm 2.8423.44 ±\pm 0.2250.00 ±\pm 2.8445.98 ±\pm 0.4540.62 ±\pm 0.9046.43 ±\pm 2.8355.36 ±\pm 0.4582735641
AllGestureWiimoteY12.68 ±\pm 1.2143.08 ±\pm 2.9524.38 ±\pm 2.0514.74 ±\pm 1.7834.60 ±\pm 2.2344.64 ±\pm 1.7858.19 ±\pm 1.2861.50 ±\pm 2.7884675321
AllGestureWiimoteZ13.89 ±\pm 0.1038.84 ±\pm 2.5420.31 ±\pm 1.1220.31 ±\pm 2.7542.18 ±\pm 0.2235.27 ±\pm 2.6840.85 ±\pm 2.9648.96 ±\pm 0.6784662531
ArrowHead38.10 ±\pm 0.1065.16 ±\pm 2.7155.78 ±\pm 0.7825.00 ±\pm 0.1067.50 ±\pm 1.8875.16 ±\pm 2.7669.06 ±\pm 2.5783.44 ±\pm 2.6375684231
BME45.83 ±\pm 2.5182.81 ±\pm 2.5549.22 ±\pm 0.7821.10 ±\pm 2.9365.62 ±\pm 2.8687.89 ±\pm 2.7988.54 ±\pm 2.7789.06 ±\pm 2.7974685321
Beef20.00 ±\pm 0.1025.00 ±\pm 2.6120.00 ±\pm 0.1020.00 ±\pm 0.1035.00 ±\pm 2.8825.00 ±\pm 2.6823.33 ±\pm 2.5240.00 ±\pm 2.9563662351
BeetleFly50.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1037.50 ±\pm 2.5962.50 ±\pm 2.7963.50 ±\pm 2.7033333821
BirdChicken50.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1054.00 ±\pm 0.1050.00 ±\pm 0.1052.00 ±\pm 2.5350.00 ±\pm 0.1053.25 ±\pm 0.1044414342
CBF57.80 ±\pm 2.64100.00 ±\pm 0.0097.55 ±\pm 0.7799.48 ±\pm 0.5298.96 ±\pm 0.52100.00 ±\pm 0.0097.54 ±\pm 0.4399.44 ±\pm 1.9581635174
Car25.00 ±\pm 0.1025.00 ±\pm 0.1029.16 ±\pm 2.8525.00 ±\pm 0.1037.50 ±\pm 2.7952.08 ±\pm 2.0929.17 ±\pm 2.7058.34 ±\pm 2.6166563241
Chinatown71.83 ±\pm 0.1087.43 ±\pm 2.7578.52 ±\pm 2.6097.92 ±\pm 2.9094.70 ±\pm 0.1096.35 ±\pm 0.5297.40 ±\pm 0.8594.96 ±\pm 0.5286715324
ChlorineConcentration55.36 ±\pm 0.1065.07 ±\pm 0.9054.12 ±\pm 0.3477.22 ±\pm 1.8857.29 ±\pm 0.1066.90 ±\pm 0.7653.70 ±\pm 2.6984.80 ±\pm 0.1064725381
CinCECGTorso18.49 ±\pm 0.1063.34 ±\pm 2.5134.00 ±\pm 0.4273.22 ±\pm 2.6774.31 ±\pm 1.1094.00 ±\pm 1.0463.69 ±\pm 2.1086.18 ±\pm 3.0086743152
Coffee50.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1080.00 ±\pm 2.6450.00 ±\pm 0.1058.00 ±\pm 0.1033333132
评论

Thanks for the clarification, which addressed most of my concerns. However, I still have two key questions that I look forward to further explanations from the authors:

  • Why choose Sobolev Space? In the rebuttal, the authors mentioned, “For time series data, it is natural to consider Sobolev spaces to capture the smoothness/regularity for model sensitivity.” Could the authors provide more theoretical explanations to support it?
  • Weakness 4 in my review remains unaddressed. Can the authors provide some experimental comparisons or analyses to explain why limiting the magnitude of perturbations at each time step does not directly improve VAT performance?

Regardless, I appreciate the thorough experimental results provided by the authors, and I would increase my rating to 3.

评论
datasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCf-VATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankTS-TCC_rankf-VAT_rank
Computers50.00±\pm0.1056.64±\pm5.0856.25±\pm6.2539.06±\pm0.1049.22±\pm2.7859.38±\pm3.1257.82±\pm2.7567.97±\pm3.126.004.005.008.007.002.003.001.00
CricketX4.69±\pm0.3928.40±\pm1.7815.62±\pm1.3410.94±\pm1.5630.98±\pm1.9732.82±\pm2.8236.07±\pm3.3734.49±\pm4.878.005.006.007.004.003.001.002.00
CricketY5.08±\pm0.1018.12±\pm3.3010.60±\pm2.9211.25±\pm3.1327.32±\pm0.1835.22±\pm2.6335.09±\pm4.1736.34±\pm4.558.005.007.006.004.002.003.001.00
CricketZ6.48±\pm1.4025.31±\pm0.4011.45±\pm3.2410.94±\pm1.5632.10±\pm2.0132.10±\pm1.3835.51±\pm3.2738.58±\pm0.138.005.006.007.003.003.002.001.00
Crop15.47±\pm3.6755.90±\pm1.0752.96±\pm2.4634.93±\pm1.7654.20±\pm0.1361.14±\pm2.8264.75±\pm0.3863.31±\pm0.358.004.006.007.005.003.001.002.00
DiatomSizeReduction30.65±\pm0.1064.48±\pm2.7930.60±\pm0.0830.32±\pm0.6263.23±\pm2.7694.93±\pm3.5196.81±\pm3.4799.16±\pm0.846.004.007.008.005.003.002.001.00
DistalPhalanxOutlineAgeGroup59.81±\pm0.1082.49±\pm0.3979.87±\pm1.5270.03±\pm3.6277.91±\pm1.9276.67±\pm3.0975.90±\pm2.1982.92±\pm4.588.002.003.007.004.005.006.001.00
DistalPhalanxOutlineCorrect73.83±\pm0.1072.47±\pm0.3067.74±\pm5.9869.53±\pm4.4368.60±\pm0.5969.97±\pm3.5475.92±\pm0.5374.32±\pm3.503.004.008.006.007.005.001.002.00
DistalPhalanxTW48.11±\pm0.1077.97±\pm1.7264.22±\pm7.3466.40±\pm2.8673.04±\pm2.7372.66±\pm2.1977.35±\pm2.2180.08±\pm0.398.002.007.006.004.005.003.001.00
DodgerLoopDay13.79±\pm0.1018.96±\pm2.8620.69±\pm2.8113.79±\pm0.1024.14±\pm3.0027.58±\pm6.8917.24±\pm2.7528.58±\pm3.007.005.004.007.003.002.006.001.00
DodgerLoopGame51.61±\pm0.1051.61±\pm0.1051.61±\pm0.1051.61±\pm0.1069.35±\pm2.8467.74±\pm0.1054.84±\pm2.7974.19±\pm2.865.005.005.005.002.003.004.001.00
DodgerLoopWeekend70.97±\pm0.1070.97±\pm0.1088.71±\pm2.8670.97±\pm0.1077.42±\pm3.2388.71±\pm1.6182.79±\pm2.8295.16±\pm1.616.006.002.006.005.002.004.001.00
ECG20067.95±\pm1.2876.45±\pm3.2468.08±\pm2.3480.80±\pm2.6882.81±\pm2.7581.92±\pm2.7583.60±\pm3.9186.30±\pm3.918.006.007.005.003.004.002.001.00
ECG500056.93±\pm0.1093.02±\pm0.1491.90±\pm0.5391.42±\pm0.2291.48±\pm0.5591.40±\pm0.5991.28±\pm0.3683.81±\pm0.368.001.002.004.003.005.006.007.00
Average35.31±\pm0.1053.09±\pm0.1045.19±\pm0.1042.89±\pm0.1056.53±\pm0.1058.67±\pm0.1058.07±\pm0.1065.85±\pm0.106.673.935.705.933.573.703.371.50
评论

Due to character limitations, we split each label-ratio table into two parts and posted them in separate comments for your reference.

datasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCfVATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankCATCC_rankf-VAT_rank
ACSF113.75±\pm3.7516.40±\pm8.6013.28±\pm2.3435.94±\pm2.8242.97±\pm2.5132.82±\pm3.1243.75±\pm2.6457.29±\pm2.617.006.008.004.003.005.002.001.00
Adiac5.98±\pm4.027.54±\pm1.746.06±\pm0.2616.73±\pm2.6811.19±\pm6.1929.04±\pm1.1129.25±\pm2.8833.94±\pm4.208.006.007.004.005.003.002.001.00
AllGestureWiimoteX10.12±\pm1.3536.61±\pm4.4626.90±\pm0.7823.66±\pm2.6846.20±\pm2.9124.33±\pm4.6942.71±\pm1.6844.86±\pm7.368.004.005.007.001.006.003.002.00
AllGestureWiimoteY12.68±\pm1.2133.93±\pm7.5920.42±\pm0.3420.76±\pm2.9051.49±\pm2.2432.36±\pm3.7944.19±\pm5.5855.36±\pm1.788.004.007.006.002.005.003.001.00
AllGestureWiimoteZ12.68±\pm1.2125.67±\pm6.4718.02±\pm3.7419.20±\pm2.9238.17±\pm3.7923.66±\pm2.6834.38±\pm5.8433.54±\pm4.918.004.007.006.001.005.002.003.00
ArrowHead38.10±\pm0.1052.34±\pm4.2239.61±\pm0.8639.84±\pm2.8060.78±\pm2.3464.84±\pm1.5661.67±\pm3.9671.25±\pm1.418.005.007.006.004.002.003.001.00
BME45.83±\pm2.6775.00±\pm2.5750.78±\pm2.3445.32±\pm2.8264.06±\pm2.7890.24±\pm1.1774.48±\pm2.7079.69±\pm2.697.003.006.008.005.001.004.002.00
Beef20.00±\pm0.1020.00±\pm0.1015.00±\pm5.0020.00±\pm0.1025.00±\pm5.0020.00±\pm2.7520.00±\pm0.1028.00±\pm0.103.003.008.003.002.003.003.001.00
BeetleFly50.00±\pm0.1062.50±\pm2.8650.00±\pm0.1050.00±\pm0.1050.00±\pm0.1043.75±\pm6.2541.67±\pm2.7868.75±\pm5.753.002.003.003.003.007.008.001.00
BirdChicken50.00±\pm0.1056.25±\pm6.2550.00±\pm0.1050.00±\pm0.1062.50±\pm2.8656.25±\pm2.8645.83±\pm5.8962.50±\pm5.155.003.005.005.001.003.008.001.00
CBF24.22±\pm0.1099.46±\pm0.5398.16±\pm0.2475.10±\pm2.8999.74±\pm0.2699.20±\pm0.8098.84±\pm0.7499.92±\pm0.588.003.006.007.002.004.005.001.00
Car25.00±\pm0.1041.66±\pm2.8725.00±\pm0.1025.00±\pm0.1037.50±\pm2.8652.08±\pm2.0925.00±\pm0.1058.34±\pm4.165.003.005.005.004.002.005.001.00
Chinatown71.83±\pm0.1094.49±\pm1.3488.58±\pm0.4179.69±\pm0.5291.56±\pm0.1090.56±\pm2.8296.35±\pm3.7195.88±\pm0.528.003.006.007.004.005.001.002.00
ChlorineConcentration55.36±\pm0.1056.06±\pm1.9654.96±\pm0.0953.70±\pm0.1055.67±\pm0.2352.18±\pm1.1469.43±\pm3.4366.97±\pm0.685.003.006.007.004.008.001.002.00
CinCECGTorso18.49±\pm0.1071.94±\pm2.8338.72±\pm1.9629.84±\pm2.4162.32±\pm2.7586.58±\pm0.3779.68±\pm4.0191.96±\pm2.098.004.006.007.005.002.003.001.00
Coffee50.00±\pm0.1050.00±\pm0.1050.00±\pm0.1050.00±\pm0.1075.00±\pm2.8670.00±\pm2.8650.00±\pm0.1082.00±\pm0.104.004.004.004.002.003.004.001.00
评论
datasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCfVATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankTS-TCC_rankfVAT_rank
Computers50.00 ±\pm 0.1068.36 ±\pm 2.7353.71 ±\pm 2.9966.02 ±\pm 2.7360.94 ±\pm 0.1065.62 ±\pm 2.9966.40 ±\pm 2.9771.09 ±\pm 0.1082746531
CricketX5.08 ±\pm 0.1042.77 ±\pm 2.8218.84 ±\pm 2.0552.68 ±\pm 1.6158.76 ±\pm 0.6263.57 ±\pm 2.9856.49 ±\pm 2.8066.78 ±\pm 0.7186753241
CricketY22.76 ±\pm 2.7831.25 ±\pm 2.7915.14 ±\pm 3.0030.62 ±\pm 2.8754.51 ±\pm 2.8460.31 ±\pm 2.9954.23 ±\pm 2.8464.10 ±\pm 2.9575863241
CricketZ6.48 ±\pm 1.4044.20 ±\pm 2.6818.84 ±\pm 2.9340.62 ±\pm 2.8055.80 ±\pm 2.8660.14 ±\pm 2.9856.13 ±\pm 2.9363.75 ±\pm 2.9485764231
Crop24.52 ±\pm 0.9671.32 ±\pm 2.8049.84 ±\pm 0.6654.92 ±\pm 2.3962.40 ±\pm 0.7860.79 ±\pm 0.26-74.32 ±\pm 1.47726534-1
DiatomSizeReduction31.46 ±\pm 0.8092.16 ±\pm 2.7945.14 ±\pm 2.8382.19 ±\pm 2.8595.32 ±\pm 2.8195.21 ±\pm 2.8259.48 ±\pm 2.8896.21 ±\pm 0.2684752361
DistalPhalanxOutlineAgeGroup59.81 ±\pm 0.1076.08 ±\pm 2.7972.05 ±\pm 2.7978.31 ±\pm 1.5279.76 ±\pm 2.3478.73 ±\pm 2.6678.22 ±\pm 2.3276.76 ±\pm 0.9486731245
DistalPhalanxOutlineCorrect73.83 ±\pm 0.1073.32 ±\pm 2.7859.92 ±\pm 2.5279.91 ±\pm 0.3780.43 ±\pm 0.5265.62 ±\pm 2.7878.08 ±\pm 2.7081.32 ±\pm 2.1456832741
DistalPhalanxTW48.11 ±\pm 0.1071.59 ±\pm 2.7871.48 ±\pm 0.8675.78 ±\pm 0.1076.95 ±\pm 0.3973.23 ±\pm 1.8275.78 ±\pm 0.1078.59 ±\pm 2.7986732531
DodgerLoopDay13.79 ±\pm 0.1042.58 ±\pm 2.7925.86 ±\pm 1.7222.42 ±\pm 2.7729.31 ±\pm 1.7244.82 ±\pm 2.8824.14 ±\pm 2.8943.82 ±\pm 2.9683574162
DodgerLoopGame51.61 ±\pm 0.1071.58 ±\pm 2.5151.61 ±\pm 0.1075.81 ±\pm 1.6270.97 ±\pm 2.9275.81 ±\pm 2.9059.14 ±\pm 2.8678.15 ±\pm 2.6374725261
DodgerLoopWeekend70.97 ±\pm 0.1092.83 ±\pm 2.7190.32 ±\pm 0.1080.64 ±\pm 2.8498.38 ±\pm 1.6296.77 ±\pm 0.1070.97 ±\pm 0.1099.38 ±\pm 0.9674562371
ECG20066.67 ±\pm 0.1076.14 ±\pm 2.9672.72 ±\pm 3.0085.04 ±\pm 2.9283.48 ±\pm 2.8780.24 ±\pm 2.7880.24 ±\pm 2.7883.14 ±\pm 0.1686712443
ECG500056.93 ±\pm 0.1091.38 ±\pm 2.7790.83 ±\pm 0.1992.30 ±\pm 0.3293.04 ±\pm 0.1094.68 ±\pm 0.2494.68 ±\pm 0.2495.68 ±\pm 0.3686754221
-37.1563.6046.1163.8569.0270.2863.2776.247.334.476.804.133.133.174.271.53
评论
datasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCfVATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankTS-TCC_rankfVAT_rank
ACSF111.25 ±\pm 1.2533.60 ±\pm 2.348.59 ±\pm 2.3429.69 ±\pm 2.7035.16 ±\pm 2.6643.75 ±\pm 1.5644.79 ±\pm 2.8846.88 ±\pm 1.5675864321
Adiac4.78 ±\pm 2.873.93 ±\pm 1.252.93 ±\pm 0.2511.06 ±\pm 2.9136.26 ±\pm 0.1455.67 ±\pm 1.2548.67 ±\pm 2.9260.08 ±\pm 1.2167854231
AllGestureWiimoteX13.89 ±\pm 0.1051.56 ±\pm 1.1125.22 ±\pm 0.6755.58 ±\pm 0.6754.28 ±\pm 0.6757.36 ±\pm 2.8055.65 ±\pm 2.7659.15 ±\pm 2.5886745231
AllGestureWiimoteY13.89 ±\pm 0.1052.90 ±\pm 2.4524.22 ±\pm 2.7956.03 ±\pm 0.6769.19 ±\pm 2.2462.50 ±\pm 2.8254.28 ±\pm 0.5661.99 ±\pm 2.8086741253
AllGestureWiimoteZ13.89 ±\pm 0.1049.11 ±\pm 2.6818.52 ±\pm 0.2254.91 ±\pm 1.3450.22 ±\pm 2.6947.99 ±\pm 0.6736.16 ±\pm 2.7354.61 ±\pm 2.9084713562
ArrowHead38.10 ±\pm 0.1067.66 ±\pm 2.6741.64 ±\pm 2.7777.81 ±\pm 2.9573.28 ±\pm 2.8662.82 ±\pm 2.9666.46 ±\pm 2.8681.56 ±\pm 2.7584723651
BME45.84 ±\pm 2.9788.28 ±\pm 2.7054.30 ±\pm 2.7176.56 ±\pm 2.8193.75 ±\pm 2.9095.32 ±\pm 1.5689.06 ±\pm 2.70100.00 ±\pm 0.0085763241
Beef20.00 ±\pm 0.1025.00 ±\pm 2.7920.00 ±\pm 0.1025.00 ±\pm 2.7925.00 ±\pm 2.7930.00 ±\pm 2.8520.00 ±\pm 0.1050.00 ±\pm 2.8663633261
BeetleFly50.00 ±\pm 0.1056.25 ±\pm 2.7550.00 ±\pm 0.1050.00 ±\pm 0.1068.75 ±\pm 2.8862.50 ±\pm 2.9250.00 ±\pm 0.1087.50 ±\pm 2.7754552351
BirdChicken50.00 ±\pm 0.1062.50 ±\pm 2.5550.00 ±\pm 0.1056.25 ±\pm 2.5581.25 ±\pm 2.7750.00 ±\pm 2.7750.00 ±\pm 0.1062.50 ±\pm 2.6952541552
CBF37.89 ±\pm 2.62100.00 ±\pm 0.0097.92 ±\pm 1.0499.48 ±\pm 0.52100.00 ±\pm 0.0099.71 ±\pm 0.29100.00 ±\pm 0.0098.96 ±\pm 0.8481751416
Car22.92 ±\pm 2.0925.00 ±\pm 0.1025.00 ±\pm 0.1045.84 ±\pm 2.8045.83 ±\pm 0.1060.42 ±\pm 0.1041.67 ±\pm 2.7863.67 ±\pm 2.9386634251
Chinatown71.83 ±\pm 0.1086.68 ±\pm 2.5983.08 ±\pm 2.7797.92 ±\pm 1.0491.98 ±\pm 0.1093.98 ±\pm 2.7798.96 ±\pm 0.8599.40 ±\pm 0.5286735421
ChlorineConcentration55.36 ±\pm 0.1080.24 ±\pm 2.6954.18 ±\pm 0.6489.92 ±\pm 2.6559.94 ±\pm 1.6178.43 ±\pm 1.7753.70 ±\pm 0.3992.97 ±\pm 0.1063725481
CinCECGTorso32.94 ±\pm 2.9379.74 ±\pm 2.6741.37 ±\pm 2.0382.10 ±\pm 2.7686.61 ±\pm 1.3993.50 ±\pm 0.9191.57 ±\pm 2.8894.70 ±\pm 0.3586754231
Coffee50.00 ±\pm 0.10100.00 ±\pm 0.0050.00 ±\pm 0.1090.00 ±\pm 0.1099.00 ±\pm 0.1099.00 ±\pm 0.1080.00 ±\pm 2.70100.00 ±\pm 0.0071753361
评论
DatasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCf-VATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankTS-TCC_rankf-VAT_rank
Computers50.00 ±\pm 0.1062.89 ±\pm 2.6063.28 ±\pm 2.6346.09 ±\pm 1.1763.28 ±\pm 1.5655.47 ±\pm 2.7259.38 ±\pm 2.6567.58 ±\pm 2.7574282651
CricketX6.48 ±\pm 1.4033.44 ±\pm 2.8216.77 ±\pm 0.6430.18 ±\pm 2.8041.83 ±\pm 2.2842.54 ±\pm 0.4148.93 ±\pm 2.8343.16 ±\pm 0.4585764312
CricketY6.48 ±\pm 1.4018.30 ±\pm 2.7717.21 ±\pm 2.8215.32 ±\pm 1.5639.91 ±\pm 0.0943.39 ±\pm 2.7341.25 ±\pm 2.5146.49 ±\pm 2.5585674231
CricketZ6.48 ±\pm 1.4030.09 ±\pm 0.0916.05 ±\pm 1.9036.03 ±\pm 0.9346.07 ±\pm 1.1639.02 ±\pm 1.7044.76 ±\pm 2.9441.20 ±\pm 2.5286751423
Crop23.44 ±\pm 2.4659.36 ±\pm 2.7748.42 ±\pm 0.8459.82 ±\pm 2.6158.88 ±\pm 1.6965.79 ±\pm 1.7771.77 ±\pm 0.8768.95 ±\pm 0.4385746312
DiatomSizeReduction35.48 ±\pm 2.6178.03 ±\pm 2.6130.60 ±\pm 0.2368.75 ±\pm 0.7389.84 ±\pm 0.7899.16 ±\pm 0.1071.91 ±\pm 2.75100.00 ±\pm 0.0074863251
DistalPhalanxOutlineAgeGroup59.81 ±\pm 0.1079.37 ±\pm 0.4370.13 ±\pm 2.5579.09 ±\pm 0.0474.78 ±\pm 2.8972.40 ±\pm 0.1078.24 ±\pm 2.2981.74 ±\pm 2.5582735641
DistalPhalanxOutlineCorrect73.83 ±\pm 0.1068.08 ±\pm 0.3755.92 ±\pm 2.7465.10 ±\pm 0.8679.91 ±\pm 0.1573.40 ±\pm 2.7779.57 ±\pm 1.5680.76 ±\pm 0.1046872531
DistalPhalanxTW48.11 ±\pm 0.1075.39 ±\pm 0.0867.26 ±\pm 0.0875.78 ±\pm 0.1074.61 ±\pm 1.170.00 ±\pm 0.1075.78 ±\pm 0.1079.30 ±\pm 0.3974625821
DodgerLoopDay13.79 ±\pm 0.1018.96 ±\pm 2.7920.69 ±\pm 2.9017.24 ±\pm 2.8329.31 ±\pm 2.7834.48 ±\pm 2.8321.84 ±\pm 2.6626.20 ±\pm 1.7386572143
DodgerLoopGame51.61 ±\pm 0.1061.29 ±\pm 2.9053.22 ±\pm 1.6258.06 ±\pm 2.7959.68 ±\pm 2.8475.81 ±\pm 2.7859.14 ±\pm 2.7777.42 ±\pm 2.8183764251
DodgerLoopWeekend70.97 ±\pm 0.1083.87 ±\pm 2.5091.94 ±\pm 1.6282.26 ±\pm 2.8788.71 ±\pm 2.8395.16 ±\pm 1.6184.95 ±\pm 2.9196.77 ±\pm 0.1086374251
ECG20066.67 ±\pm 0.1073.22 ±\pm 0.9078.02 ±\pm 2.9081.59 ±\pm 2.7184.38 ±\pm 2.7976.34 ±\pm 2.8180.36 ±\pm 2.7186.31 ±\pm 2.9087532641
ECG500056.93 ±\pm 0.1093.16 ±\pm 0.1092.28 ±\pm 0.4991.21 ±\pm 0.2991.96 ±\pm 0.3392.43 ±\pm 0.5492.21 ±\pm 0.0894.21 ±\pm 0.0882476351
Average36.9255.1645.7250.9458.9360.4159.8468.877.004.405.874.873.773.403.671.50
评论

Thank you for your time and efforts in timely reviewing our rebuttal and acknowledging the contributions of our work.

Q1

Thank you for your response. Sobolev spaces are a fundamental concept in functional analysis, designed to extend the notion of "smoothness" for functions beyond classical differentiability. The usage of Sobolev spaces is quite common in the context of functional data analysis, see, for example, [Hsing et al., 2015].

Formally, the Sobolev space Hk(R)H^k(\mathbb{R}) consists of functions ff whose first kk weak derivatives exist and are square-integrable, with the Sobolev norm defined as

fHk=_j=0kˆf(j)(x)2dx,|f|_{H^k}=\sqrt{\sum\_{j=0}\^{k}\int|f^{(j)}(x)|^2 dx},

where f(j)f^{(j)} denotes the jj-th weak derivative of ff. Therefore, this norm penalizes rapid oscillations, ensuring smooth approximations.

Reference

  • [Hsing et al., 2015] Hsing T, and Eubank R L. Theoretical foundations of functional data analysis, with an introduction to linear operators. New York: Wiley, 2015.

Q2

To address this concern, we design a variant called VAT-step, which constrains the magnitude of perturbations at each time step, and conduct extensive experiments across multiple datasets with different label ratios. Table shows that the VAT-step only provides marginal performance gains compared to the original VAT in most settings, and even degrades performance on more challenging multivariate datasets, such as SelfReg and NATOPS, containing complex temporal structure. This is because clipping the perturbation at each time step struggles to adaptively scale the magnitude of perturbations on some critical regions for prediction (like ''shapelets'' in Figure 4), except for carefully tuning the per-step magnitude hyperparameters. Additionally, VAT-step easily converges to the ''permutation-invariant'' [George et al., 1976] local optimum, where even arbitrarily reordering time steps can still generate the identical perturbation which satisfies each step-wise constraint. On the other hand, adversarial perturbations generated by f-VAT, incorporating the temporal structure of time series data, efficiently utilize time series data to improve the smoothness of predictive distributions and predictive performance compared to other VAT-based baselines.

MethodsCricketXUWaveInsectWingSelfRegNATOPS
10%20%40%10%20%40%10%20%40%10%20%40%10%20%40%
VAT42.85±3.9749.14±0.5058.63±0.5094.41±0.0995.53±0.3194.76±0.5455.49±1.2861.27±0.1963.48±0.3053.12±4.5155.76±0.3553.47±1.0482.38±0.9682.81±0.5290.15±1.60
VAT‑step40.47±3.8447.85±0.3359.30±0.3693.78±0.6796.09±0.3295.86±0.5557.45±0.2160.92±0.4863.61±0.5848.61±3.2753.67±2.7952.75±1.8081.33±0.9683.02±0.1286.58±1.18
f‑VAT49.18±1.9657.91±3.5868.39±2.2594.82±0.3996.45±0.2797.23±0.4358.01±1.1261.28±1.8664.81±1.1559.31±3.0661.60±1.1364.44±3.1386.04±1.4186.25±1.3893.13±0.15

Reference

  • [George et al., 1976] George EP Box and Gwilym MJenkins. Time series analysis, forecasting and control. Holden-Day, 1976
评论

Thank you for your response. I have reviewed the content, and these replies effectively address my current concerns. I would increase my rating to borderline accept.

评论

We are pleased that the changes you suggested have enhanced the clarity and presentation of our work. We will follow your suggestion to include the experiments in the updated manuscript. Thank you for raising your score!

审稿意见
4

This paper addresses the flaw of a famous semi-supervised learning method, VAT, when adopted in time series classification tasks, which is suggested to be caused by neglecting the temporal structure of the data in the adversarial perturbation. In response, the paper proposes functional-VAT (f-VAT), using an appropriate Sobolev norm to generate structured functional adversarial perturbations to incorporate the functional structure of the data into perturbations.

优缺点分析

[Strengths]

  • Semi-supervised learning for time series classification is an important but highly under-explored topic. I believe it is meaningful to specifically consider the characteristics of time series data, just as this paper did.

  • Utilizing the duality between perturbation norm and gradient sensitivity and introducing the Sobolev norm to generate structured adversarial perturbations is reasonable to me. The theoretical derivation looks elegant and solid.

  • The experimental results look promising, showing clear improvement of the proposed method.

[Weaknesses]

  • My biggest concern is that the motivation is not convincing enough. Specifically, I could understand that adopting VAT in time series classification tasks would face the suggested problem, inspiring the proposed improvement. However, in the first place, why is it necessary to use VAT? VAT is a quite classical method proposed several years ago, with various more advanced semi-supervised learning methods appearing subsequently. Does VAT have any particular advantages over them in time series tasks? From my perspective, building upon VAT is afraid to limit the contribution of the proposed method to the current real-world practices.

  • As UCR and UEA repositories include 100+ time series classification datasets, just using six datasets is not sufficient and representative for me. I would suggest either including the majority of the datasets or pre-setting a specific rule to select the experimental datasets to avoid cherry-picking. For the compared baseline methods, there is the same case. While it is argued that the settings of both datasets and baselines are the same as [17], please note that it is published on ICASSP. The experiment range that is solid enough for ICASSP may not be similarly solid enough for a top conference like NeurIPS.

  • There is a lack of empirical evidence for some minor points. For instance, while the paper provides a theoretical complexity for implementing the Sobolev norm, it would be beneficial to include an empirical comparison of the actual training overhead of f-VAT versus standard VAT.

  • The code is not available for reproduction.

[Recommendation]

Overall, the novelty, insights, and contributions of the paper are borderline for NeurIPS. The suggested method is quite interesting, and the current results seem promising. However, I have certain concerns about the motivation, and the experiments are not solid enough. Therefore, I recommend a borderline rejection at this time, and would like to change my score based on the rebuttal and the discussion with other reviewers later.

问题

Please check the weaknesses.

局限性

Yes.

最终评判理由

I appreciate the authors' efforts in the rebuttal, and apologize for leaving out the code in the supplementary materials.

After carefully reading the rebuttal and other reviewers' comments, I maintain my position that the novelty of this work is borderline for NeurIPS. And to be honest, regarding the motivation, I'm still not quite convinced by the choice of VAT. Specifically, although the suggested f-VAT achieves a non-trivial improvement in performance compared with the existing method, integrating the same "functional" adversarial data augmentation into other more advanced SSL frameworks may still help achieve better improvement. That is, instead, what I'm mainly concerned about.

However, the authors provide a large number of new empirical results as suggested, which look promising and indeed address some points of my concern. Therefore, I increased my rating to borderline accept.

格式问题

No formatting issues.

作者回复

W1: Motivations for selecting VAT

We appreciate the reviewer’s concern about the choice of Virtual Adversarial Training (VAT) [Takeru et al., 2018]. Some recent self-supervised methods like TS-TCC [Emadeldeen et al., 2023] require careful selection of confidence thresholds and remain susceptible to "confirmation bias", while VAT-based methods do not require selecting thresholds and can adaptively construct adversarial perturbations to effectively utilize unlabeled data for better generalization. Additionally, our proposed f-VAT within the VAT-based framework greatly improves performance by effectively leveraging underlying data structure of time series data. Please read our response to the next question for additional empirical results on the performance comparison of different methods across multiple datasets. Additionally, we provide a theoretical understanding of f-VAT, which gives us insights into the selection of the perturbation norm.

W2: More empirical results

Thanks for your suggestions. In the original paper, we follow the settings [Haoyi et al., 2021] to compare f-VAT with other competitive baselines and report results on six UCR/UEA datasets. To further verify the superiority of f-VAT, due to the time limit, we randomly sample 30 datasets from UCR/UEA datasets and add a recent state-of-the-art Class-Aware Temporal and Contextual Contrasting (TS-TCC) [Emadeldeen et al., 2023] within available computing resources. Table shows that tVAT still consistently outperforms other competitive baselines in all settings. Due to the page limit, please refer to details of more empirical results in the comments.

Average accuracy (%) and average rank under different label ratios:

Method10%20%40%
AvgAccAvgRankAvgAccAvgRankAvgAccAvgRank
SupL35.316.6736.927.0037.157.33
PI53.093.9355.164.4063.604.47
MTL45.195.7045.725.8746.116.80
SemiTime56.533.5758.933.7769.023.17
TapNet58.673.7060.413.4070.283.20
TS‑TCC58.073.3759.843.6763.274.27
f-VAT65.851.5068.871.5076.241.53

Additionally, we construct more empirical results on several large-scale China Securities Index (CSI) datasets (i.e., CSI 50 and 500 futures) spanning from 2020 to 2023 for predicting directions (upward or downward) of futures prices [Zihao et al., 2019] [Qingyi et al., 2024]. The dataset collects over 42×10342\times 10^{3} records spanning from 2020 to 2022, and each time step contains bid/ask prices and corresponding volumes. Table presents the performance of various semi-supervised methods across different label ratios, and shows that f-VAT significantly outperforms other competitive baselines on futures datasets, especially on more volatile CSI 500 futures. This is because f-VAT's adversarial perturbations incorporating key temporal structure facilitate deep models to effectively use unlabeled samples to yield smoother predictive distribution with better generalization.

The performance comparison on domestic futures datasets:

FuturesRatiosupLPIMTLSemiTimeTapNetTS-TCCf-VAT
10%40.05 ±\pm 1.3842.87 ±\pm 1.4553.94 ±\pm 0.1355.19 ±\pm 0.7754.62 ±\pm 0.5455.72 ±\pm 1.1658.64 ±\pm 0.53
5020%45.08 ±\pm 1.4547.26 ±\pm 0.7354.97 ±\pm 0.6156.69 ±\pm 0.5056.93 ±\pm 1.2858.21 ±\pm 0.4462.09 ±\pm 0.26
40%50.69 ±\pm 3.5352.42 ±\pm 1.2156.97 ±\pm 0.5757.34 ±\pm 0.2059.75 ±\pm 1.0959.80 ±\pm 1.3264.78 ±\pm 0.45
10%34.23 ±\pm 0.2444.00 ±\pm 0.0639.53 ±\pm 1.3538.86 ±\pm 2.0440.53 ±\pm 0.8339.79 ±\pm 1.5443.77 ±\pm 0.73
50020%35.38 ±\pm 1.0646.57 ±\pm 0.4845.26 ±\pm 0.2946.04 ±\pm 0.1244.58 ±\pm 1.7045.05 ±\pm 1.7452.14 ±\pm 0.25
40%43.85 ±\pm 1.4949.25 ±\pm 0.5547.66 ±\pm 1.5050.65 ±\pm 1.0551.50 ±\pm 1.3654.29 ±\pm 0.8758.66 ±\pm 0.46

W3: Runtime comparison

Thanks for your suggestions. Compared with VAT (O(N)\mathcal{O}(N)), the computational complexity of f-VAT is O(Nlog(N))\mathcal{O}(N\log(N)) with a small constant factor, involving the FFT-based Hs\|\cdot\|_{H^{-s}} normalization. The empirical results show that the extra computational cost remains in the same order as VAT.

Runtime comparison f-VAT and VAT:

MethodCricketXUWaveInsectWingNATOPSSelfReg
VAT15.6751.6635.5828.0430.68
f-VAT20.4562.0545.9238.9844.95
Δ (%)30.5020.1129.0639.0246.51

W4: The available code for reproduction

We provide the training logs and the source code in the supplementary materials for reference. Could you please provide more details while trying to reproduce empirical results, such as error messages or missing packages?

  • [Takeru et al., 2018] Takeru Miyato, Shin-ichi Maeda, Masanori Koyama, and Shin Ishii. Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018.
  • [Emadeldeen et al., 2023] Emadeldeen Eldele, Mohamed Ragab, Zhenghua Chen, Min Wu, Chee-Keong Kwoh, Xiaoli Li, and Cuntai Guan. Self-supervised contrastive representation learning for semi-supervised time-series classification. IEEE Transactions on Pattern Analysis and Machine Intelligence, , 2023.
  • [Haoyi et al., 2021] Haoyi Fan, Fengbin Zhang, Ruidong Wang, Xunhua Huang, and Zuoyong Li. Semi-supervised time series classification by temporal relation prediction. In ICASSP, IEEE, 2021.
  • [Zihao et al., 2019] Zihao Zhang, Stefan Zohren, and Stephen Roberts. Deeplob: Deep convolutional neural networks for limit order books. IEEE Transactions on Signal Processing, 2019.
  • [Qingyi et al., 2024] Qingyi Pan, Suyu Sun, Pei Yang, and Jingyi Zhang. Futuresnet: Capturing patterns of price fluctuations in domestic futures trading. Electronics, 13(22):4482, 2024
评论

Due to character limitations, we split each label-ratio table into two parts and posted them in separate comments for your reference.

DatasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCfVATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankTS-TCC_rankfVAT_rank
ACSF110.00 ±\pm 0.1014.06 ±\pm 1.5610.93 ±\pm 2.9029.68 ±\pm 2.8021.10 ±\pm 2.7925.00 ±\pm 2.8224.22 ±\pm 2.8260.42 ±\pm 2.6386725341
Adiac5.98 ±\pm 2.827.45 ±\pm 2.872.54 ±\pm 0.0414.08 ±\pm 2.8920.08 ±\pm 1.4334.39 ±\pm 0.1430.44 ±\pm 2.9038.85 ±\pm 2.6276854231
AllGestureWiimoteX12.68 ±\pm 1.2152.23 ±\pm 2.8423.44 ±\pm 0.2250.00 ±\pm 2.8445.98 ±\pm 0.4540.62 ±\pm 0.9046.43 ±\pm 2.8355.36 ±\pm 0.4582735641
AllGestureWiimoteY12.68 ±\pm 1.2143.08 ±\pm 2.9524.38 ±\pm 2.0514.74 ±\pm 1.7834.60 ±\pm 2.2344.64 ±\pm 1.7858.19 ±\pm 1.2861.50 ±\pm 2.7884675321
AllGestureWiimoteZ13.89 ±\pm 0.1038.84 ±\pm 2.5420.31 ±\pm 1.1220.31 ±\pm 2.7542.18 ±\pm 0.2235.27 ±\pm 2.6840.85 ±\pm 2.9648.96 ±\pm 0.6784662531
ArrowHead38.10 ±\pm 0.1065.16 ±\pm 2.7155.78 ±\pm 0.7825.00 ±\pm 0.1067.50 ±\pm 1.8875.16 ±\pm 2.7669.06 ±\pm 2.5783.44 ±\pm 2.6375684231
BME45.83 ±\pm 2.5182.81 ±\pm 2.5549.22 ±\pm 0.7821.10 ±\pm 2.9365.62 ±\pm 2.8687.89 ±\pm 2.7988.54 ±\pm 2.7789.06 ±\pm 2.7974685321
Beef20.00 ±\pm 0.1025.00 ±\pm 2.6120.00 ±\pm 0.1020.00 ±\pm 0.1035.00 ±\pm 2.8825.00 ±\pm 2.6823.33 ±\pm 2.5240.00 ±\pm 2.9563662351
BeetleFly50.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1037.50 ±\pm 2.5962.50 ±\pm 2.7963.50 ±\pm 2.7033333821
BirdChicken50.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1054.00 ±\pm 0.1050.00 ±\pm 0.1052.00 ±\pm 2.5350.00 ±\pm 0.1053.25 ±\pm 0.1044414342
CBF57.80 ±\pm 2.64100.00 ±\pm 0.0097.55 ±\pm 0.7799.48 ±\pm 0.5298.96 ±\pm 0.52100.00 ±\pm 0.0097.54 ±\pm 0.4399.44 ±\pm 1.9581635174
Car25.00 ±\pm 0.1025.00 ±\pm 0.1029.16 ±\pm 2.8525.00 ±\pm 0.1037.50 ±\pm 2.7952.08 ±\pm 2.0929.17 ±\pm 2.7058.34 ±\pm 2.6166563241
Chinatown71.83 ±\pm 0.1087.43 ±\pm 2.7578.52 ±\pm 2.6097.92 ±\pm 2.9094.70 ±\pm 0.1096.35 ±\pm 0.5297.40 ±\pm 0.8594.96 ±\pm 0.5286715324
ChlorineConcentration55.36 ±\pm 0.1065.07 ±\pm 0.9054.12 ±\pm 0.3477.22 ±\pm 1.8857.29 ±\pm 0.1066.90 ±\pm 0.7653.70 ±\pm 2.6984.80 ±\pm 0.1064725381
CinCECGTorso18.49 ±\pm 0.1063.34 ±\pm 2.5134.00 ±\pm 0.4273.22 ±\pm 2.6774.31 ±\pm 1.1094.00 ±\pm 1.0463.69 ±\pm 2.1086.18 ±\pm 3.0086743152
Coffee50.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1050.00 ±\pm 0.1080.00 ±\pm 2.6450.00 ±\pm 0.1058.00 ±\pm 0.1033333132
评论

Due to character limitations, we split each label-ratio table into two parts and posted them in separate comments for your reference.

datasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCfVATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankTS-TCC_rankf-VAT_rank
ACSF113.75±\pm3.7516.40±\pm8.6013.28±\pm2.3435.94±\pm2.8242.97±\pm2.5132.82±\pm3.1243.75±\pm2.6457.29±\pm2.617.006.008.004.003.005.002.001.00
Adiac5.98±\pm4.027.54±\pm1.746.06±\pm0.2616.73±\pm2.6811.19±\pm6.1929.04±\pm1.1129.25±\pm2.8833.94±\pm4.208.006.007.004.005.003.002.001.00
AllGestureWiimoteX10.12±\pm1.3536.61±\pm4.4626.90±\pm0.7823.66±\pm2.6846.20±\pm2.9124.33±\pm4.6942.71±\pm1.6844.86±\pm7.368.004.005.007.001.006.003.002.00
AllGestureWiimoteY12.68±\pm1.2133.93±\pm7.5920.42±\pm0.3420.76±\pm2.9051.49±\pm2.2432.36±\pm3.7944.19±\pm5.5855.36±\pm1.788.004.007.006.002.005.003.001.00
AllGestureWiimoteZ12.68±\pm1.2125.67±\pm6.4718.02±\pm3.7419.20±\pm2.9238.17±\pm3.7923.66±\pm2.6834.38±\pm5.8433.54±\pm4.918.004.007.006.001.005.002.003.00
ArrowHead38.10±\pm0.1052.34±\pm4.2239.61±\pm0.8639.84±\pm2.8060.78±\pm2.3464.84±\pm1.5661.67±\pm3.9671.25±\pm1.418.005.007.006.004.002.003.001.00
BME45.83±\pm2.6775.00±\pm2.5750.78±\pm2.3445.32±\pm2.8264.06±\pm2.7890.24±\pm1.1774.48±\pm2.7079.69±\pm2.697.003.006.008.005.001.004.002.00
Beef20.00±\pm0.1020.00±\pm0.1015.00±\pm5.0020.00±\pm0.1025.00±\pm5.0020.00±\pm2.7520.00±\pm0.1028.00±\pm0.103.003.008.003.002.003.003.001.00
BeetleFly50.00±\pm0.1062.50±\pm2.8650.00±\pm0.1050.00±\pm0.1050.00±\pm0.1043.75±\pm6.2541.67±\pm2.7868.75±\pm5.753.002.003.003.003.007.008.001.00
BirdChicken50.00±\pm0.1056.25±\pm6.2550.00±\pm0.1050.00±\pm0.1062.50±\pm2.8656.25±\pm2.8645.83±\pm5.8962.50±\pm5.155.003.005.005.001.003.008.001.00
CBF24.22±\pm0.1099.46±\pm0.5398.16±\pm0.2475.10±\pm2.8999.74±\pm0.2699.20±\pm0.8098.84±\pm0.7499.92±\pm0.588.003.006.007.002.004.005.001.00
Car25.00±\pm0.1041.66±\pm2.8725.00±\pm0.1025.00±\pm0.1037.50±\pm2.8652.08±\pm2.0925.00±\pm0.1058.34±\pm4.165.003.005.005.004.002.005.001.00
Chinatown71.83±\pm0.1094.49±\pm1.3488.58±\pm0.4179.69±\pm0.5291.56±\pm0.1090.56±\pm2.8296.35±\pm3.7195.88±\pm0.528.003.006.007.004.005.001.002.00
ChlorineConcentration55.36±\pm0.1056.06±\pm1.9654.96±\pm0.0953.70±\pm0.1055.67±\pm0.2352.18±\pm1.1469.43±\pm3.4366.97±\pm0.685.003.006.007.004.008.001.002.00
CinCECGTorso18.49±\pm0.1071.94±\pm2.8338.72±\pm1.9629.84±\pm2.4162.32±\pm2.7586.58±\pm0.3779.68±\pm4.0191.96±\pm2.098.004.006.007.005.002.003.001.00
Coffee50.00±\pm0.1050.00±\pm0.1050.00±\pm0.1050.00±\pm0.1075.00±\pm2.8670.00±\pm2.8650.00±\pm0.1082.00±\pm0.104.004.004.004.002.003.004.001.00
评论
datasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCf-VATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankTS-TCC_rankf-VAT_rank
Computers50.00±\pm0.1056.64±\pm5.0856.25±\pm6.2539.06±\pm0.1049.22±\pm2.7859.38±\pm3.1257.82±\pm2.7567.97±\pm3.126.004.005.008.007.002.003.001.00
CricketX4.69±\pm0.3928.40±\pm1.7815.62±\pm1.3410.94±\pm1.5630.98±\pm1.9732.82±\pm2.8236.07±\pm3.3734.49±\pm4.878.005.006.007.004.003.001.002.00
CricketY5.08±\pm0.1018.12±\pm3.3010.60±\pm2.9211.25±\pm3.1327.32±\pm0.1835.22±\pm2.6335.09±\pm4.1736.34±\pm4.558.005.007.006.004.002.003.001.00
CricketZ6.48±\pm1.4025.31±\pm0.4011.45±\pm3.2410.94±\pm1.5632.10±\pm2.0132.10±\pm1.3835.51±\pm3.2738.58±\pm0.138.005.006.007.003.003.002.001.00
Crop15.47±\pm3.6755.90±\pm1.0752.96±\pm2.4634.93±\pm1.7654.20±\pm0.1361.14±\pm2.8264.75±\pm0.3863.31±\pm0.358.004.006.007.005.003.001.002.00
DiatomSizeReduction30.65±\pm0.1064.48±\pm2.7930.60±\pm0.0830.32±\pm0.6263.23±\pm2.7694.93±\pm3.5196.81±\pm3.4799.16±\pm0.846.004.007.008.005.003.002.001.00
DistalPhalanxOutlineAgeGroup59.81±\pm0.1082.49±\pm0.3979.87±\pm1.5270.03±\pm3.6277.91±\pm1.9276.67±\pm3.0975.90±\pm2.1982.92±\pm4.588.002.003.007.004.005.006.001.00
DistalPhalanxOutlineCorrect73.83±\pm0.1072.47±\pm0.3067.74±\pm5.9869.53±\pm4.4368.60±\pm0.5969.97±\pm3.5475.92±\pm0.5374.32±\pm3.503.004.008.006.007.005.001.002.00
DistalPhalanxTW48.11±\pm0.1077.97±\pm1.7264.22±\pm7.3466.40±\pm2.8673.04±\pm2.7372.66±\pm2.1977.35±\pm2.2180.08±\pm0.398.002.007.006.004.005.003.001.00
DodgerLoopDay13.79±\pm0.1018.96±\pm2.8620.69±\pm2.8113.79±\pm0.1024.14±\pm3.0027.58±\pm6.8917.24±\pm2.7528.58±\pm3.007.005.004.007.003.002.006.001.00
DodgerLoopGame51.61±\pm0.1051.61±\pm0.1051.61±\pm0.1051.61±\pm0.1069.35±\pm2.8467.74±\pm0.1054.84±\pm2.7974.19±\pm2.865.005.005.005.002.003.004.001.00
DodgerLoopWeekend70.97±\pm0.1070.97±\pm0.1088.71±\pm2.8670.97±\pm0.1077.42±\pm3.2388.71±\pm1.6182.79±\pm2.8295.16±\pm1.616.006.002.006.005.002.004.001.00
ECG20067.95±\pm1.2876.45±\pm3.2468.08±\pm2.3480.80±\pm2.6882.81±\pm2.7581.92±\pm2.7583.60±\pm3.9186.30±\pm3.918.006.007.005.003.004.002.001.00
ECG500056.93±\pm0.1093.02±\pm0.1491.90±\pm0.5391.42±\pm0.2291.48±\pm0.5591.40±\pm0.5991.28±\pm0.3683.81±\pm0.368.001.002.004.003.005.006.007.00
Average35.31±\pm0.1053.09±\pm0.1045.19±\pm0.1042.89±\pm0.1056.53±\pm0.1058.67±\pm0.1058.07±\pm0.1065.85±\pm0.106.673.935.705.933.573.703.371.50
评论
datasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCfVATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankTS-TCC_rankfVAT_rank
Computers50.00 ±\pm 0.1068.36 ±\pm 2.7353.71 ±\pm 2.9966.02 ±\pm 2.7360.94 ±\pm 0.1065.62 ±\pm 2.9966.40 ±\pm 2.9771.09 ±\pm 0.1082746531
CricketX5.08 ±\pm 0.1042.77 ±\pm 2.8218.84 ±\pm 2.0552.68 ±\pm 1.6158.76 ±\pm 0.6263.57 ±\pm 2.9856.49 ±\pm 2.8066.78 ±\pm 0.7186753241
CricketY22.76 ±\pm 2.7831.25 ±\pm 2.7915.14 ±\pm 3.0030.62 ±\pm 2.8754.51 ±\pm 2.8460.31 ±\pm 2.9954.23 ±\pm 2.8464.10 ±\pm 2.9575863241
CricketZ6.48 ±\pm 1.4044.20 ±\pm 2.6818.84 ±\pm 2.9340.62 ±\pm 2.8055.80 ±\pm 2.8660.14 ±\pm 2.9856.13 ±\pm 2.9363.75 ±\pm 2.9485764231
Crop24.52 ±\pm 0.9671.32 ±\pm 2.8049.84 ±\pm 0.6654.92 ±\pm 2.3962.40 ±\pm 0.7860.79 ±\pm 0.26-74.32 ±\pm 1.47726534-1
DiatomSizeReduction31.46 ±\pm 0.8092.16 ±\pm 2.7945.14 ±\pm 2.8382.19 ±\pm 2.8595.32 ±\pm 2.8195.21 ±\pm 2.8259.48 ±\pm 2.8896.21 ±\pm 0.2684752361
DistalPhalanxOutlineAgeGroup59.81 ±\pm 0.1076.08 ±\pm 2.7972.05 ±\pm 2.7978.31 ±\pm 1.5279.76 ±\pm 2.3478.73 ±\pm 2.6678.22 ±\pm 2.3276.76 ±\pm 0.9486731245
DistalPhalanxOutlineCorrect73.83 ±\pm 0.1073.32 ±\pm 2.7859.92 ±\pm 2.5279.91 ±\pm 0.3780.43 ±\pm 0.5265.62 ±\pm 2.7878.08 ±\pm 2.7081.32 ±\pm 2.1456832741
DistalPhalanxTW48.11 ±\pm 0.1071.59 ±\pm 2.7871.48 ±\pm 0.8675.78 ±\pm 0.1076.95 ±\pm 0.3973.23 ±\pm 1.8275.78 ±\pm 0.1078.59 ±\pm 2.7986732531
DodgerLoopDay13.79 ±\pm 0.1042.58 ±\pm 2.7925.86 ±\pm 1.7222.42 ±\pm 2.7729.31 ±\pm 1.7244.82 ±\pm 2.8824.14 ±\pm 2.8943.82 ±\pm 2.9683574162
DodgerLoopGame51.61 ±\pm 0.1071.58 ±\pm 2.5151.61 ±\pm 0.1075.81 ±\pm 1.6270.97 ±\pm 2.9275.81 ±\pm 2.9059.14 ±\pm 2.8678.15 ±\pm 2.6374725261
DodgerLoopWeekend70.97 ±\pm 0.1092.83 ±\pm 2.7190.32 ±\pm 0.1080.64 ±\pm 2.8498.38 ±\pm 1.6296.77 ±\pm 0.1070.97 ±\pm 0.1099.38 ±\pm 0.9674562371
ECG20066.67 ±\pm 0.1076.14 ±\pm 2.9672.72 ±\pm 3.0085.04 ±\pm 2.9283.48 ±\pm 2.8780.24 ±\pm 2.7880.24 ±\pm 2.7883.14 ±\pm 0.1686712443
ECG500056.93 ±\pm 0.1091.38 ±\pm 2.7790.83 ±\pm 0.1992.30 ±\pm 0.3293.04 ±\pm 0.1094.68 ±\pm 0.2494.68 ±\pm 0.2495.68 ±\pm 0.3686754221
-37.1563.6046.1163.85 ±\pm -69.0270.2863.2776.247.334.476.804.133.133.174.271.53
评论

I appreciate the authors' efforts in the rebuttal, and apologize for leaving out the code in the supplementary materials. After carefully reading the rebuttal and other reviewers' comments, I maintain my position that the novelty of this work is borderline for NeurIPS. And to be honest, regarding the motivation, I'm still not quite convinced by the choice of VAT. Specifically, although the suggested f-VAT achieves a non-trivial improvement in performance compared with the existing method, integrating the same "functional" adversarial data augmentation into other more advanced SSL frameworks may still help achieve better improvement. That is, instead, what I'm mainly concerned about. However, the authors provide a large number of new empirical results as suggested, which look promising and indeed address some points of my concern. Therefore, I would increase my rating to borderline accept.

评论

Thank you for your efforts in timely reviewing our rebuttal and recognizing the contributions of our work. Additionally, we would like to emphasize that we theoretically establish the duality between the perturbation and the smoothness (non-)linear functional models, which inspires our proposed method, is yet another key contribution of our paper. We would appreciate it if you can recognize our theory contribution.

评论
datasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCfVATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankTS-TCC_rankfVAT_rank
ACSF111.25 ±\pm 1.2533.60 ±\pm 2.348.59 ±\pm 2.3429.69 ±\pm 2.7035.16 ±\pm 2.6643.75 ±\pm 1.5644.79 ±\pm 2.8846.88 ±\pm 1.5675864321
Adiac4.78 ±\pm 2.873.93 ±\pm 1.252.93 ±\pm 0.2511.06 ±\pm 2.9136.26 ±\pm 0.1455.67 ±\pm 1.2548.67 ±\pm 2.9260.08 ±\pm 1.2167854231
AllGestureWiimoteX13.89 ±\pm 0.1051.56 ±\pm 1.1125.22 ±\pm 0.6755.58 ±\pm 0.6754.28 ±\pm 0.6757.36 ±\pm 2.8055.65 ±\pm 2.7659.15 ±\pm 2.5886745231
AllGestureWiimoteY13.89 ±\pm 0.1052.90 ±\pm 2.4524.22 ±\pm 2.7956.03 ±\pm 0.6769.19 ±\pm 2.2462.50 ±\pm 2.8254.28 ±\pm 0.5661.99 ±\pm 2.8086741253
AllGestureWiimoteZ13.89 ±\pm 0.1049.11 ±\pm 2.6818.52 ±\pm 0.2254.91 ±\pm 1.3450.22 ±\pm 2.6947.99 ±\pm 0.6736.16 ±\pm 2.7354.61 ±\pm 2.9084713562
ArrowHead38.10 ±\pm 0.1067.66 ±\pm 2.6741.64 ±\pm 2.7777.81 ±\pm 2.9573.28 ±\pm 2.8662.82 ±\pm 2.9666.46 ±\pm 2.8681.56 ±\pm 2.7584723651
BME45.84 ±\pm 2.9788.28 ±\pm 2.7054.30 ±\pm 2.7176.56 ±\pm 2.8193.75 ±\pm 2.9095.32 ±\pm 1.5689.06 ±\pm 2.70100.00 ±\pm 0.0085763241
Beef20.00 ±\pm 0.1025.00 ±\pm 2.7920.00 ±\pm 0.1025.00 ±\pm 2.7925.00 ±\pm 2.7930.00 ±\pm 2.8520.00 ±\pm 0.1050.00 ±\pm 2.8663633261
BeetleFly50.00 ±\pm 0.1056.25 ±\pm 2.7550.00 ±\pm 0.1050.00 ±\pm 0.1068.75 ±\pm 2.8862.50 ±\pm 2.9250.00 ±\pm 0.1087.50 ±\pm 2.7754552351
BirdChicken50.00 ±\pm 0.1062.50 ±\pm 2.5550.00 ±\pm 0.1056.25 ±\pm 2.5581.25 ±\pm 2.7750.00 ±\pm 2.7750.00 ±\pm 0.1062.50 ±\pm 2.6952541552
CBF37.89 ±\pm 2.62100.00 ±\pm 0.0097.92 ±\pm 1.0499.48 ±\pm 0.52100.00 ±\pm 0.0099.71 ±\pm 0.29100.00 ±\pm 0.0098.96 ±\pm 0.8481751416
Car22.92 ±\pm 2.0925.00 ±\pm 0.1025.00 ±\pm 0.1045.84 ±\pm 2.8045.83 ±\pm 0.1060.42 ±\pm 0.1041.67 ±\pm 2.7863.67 ±\pm 2.9386634251
Chinatown71.83 ±\pm 0.1086.68 ±\pm 2.5983.08 ±\pm 2.7797.92 ±\pm 1.0491.98 ±\pm 0.1093.98 ±\pm 2.7798.96 ±\pm 0.8599.40 ±\pm 0.5286735421
ChlorineConcentration55.36 ±\pm 0.1080.24 ±\pm 2.6954.18 ±\pm 0.6489.92 ±\pm 2.6559.94 ±\pm 1.6178.43 ±\pm 1.7753.70 ±\pm 0.3992.97 ±\pm 0.1063725481
CinCECGTorso32.94 ±\pm 2.9379.74 ±\pm 2.6741.37 ±\pm 2.0382.10 ±\pm 2.7686.61 ±\pm 1.3993.50 ±\pm 0.9191.57 ±\pm 2.8894.70 ±\pm 0.3586754231
Coffee50.00 ±\pm 0.10100.00 ±\pm 0.0050.00 ±\pm 0.1090.00 ±\pm 0.1099.00 ±\pm 0.1099.00 ±\pm 0.1080.00 ±\pm 2.70100.00 ±\pm 0.0071753361
评论
DatasetSupCEPIMTLmeanTeacherSemiTimeTapNetTS-TCCf-VATSupCE_rankPI_rankMTL_rankmeanTeacher_rankSemiTime_rankTapNet_rankTS-TCC_rankf-VAT_rank
Computers50.00 ±\pm 0.1062.89 ±\pm 2.6063.28 ±\pm 2.6346.09 ±\pm 1.1763.28 ±\pm 1.5655.47 ±\pm 2.7259.38 ±\pm 2.6567.58 ±\pm 2.7574282651
CricketX6.48 ±\pm 1.4033.44 ±\pm 2.8216.77 ±\pm 0.6430.18 ±\pm 2.8041.83 ±\pm 2.2842.54 ±\pm 0.4148.93 ±\pm 2.8343.16 ±\pm 0.4585764312
CricketY6.48 ±\pm 1.4018.30 ±\pm 2.7717.21 ±\pm 2.8215.32 ±\pm 1.5639.91 ±\pm 0.0943.39 ±\pm 2.7341.25 ±\pm 2.5146.49 ±\pm 2.5585674231
CricketZ6.48 ±\pm 1.4030.09 ±\pm 0.0916.05 ±\pm 1.9036.03 ±\pm 0.9346.07 ±\pm 1.1639.02 ±\pm 1.7044.76 ±\pm 2.9441.20 ±\pm 2.5286751423
Crop23.44 ±\pm 2.4659.36 ±\pm 2.7748.42 ±\pm 0.8459.82 ±\pm 2.6158.88 ±\pm 1.6965.79 ±\pm 1.7771.77 ±\pm 0.8768.95 ±\pm 0.4385746312
DiatomSizeReduction35.48 ±\pm 2.6178.03 ±\pm 2.6130.60 ±\pm 0.2368.75 ±\pm 0.7389.84 ±\pm 0.7899.16 ±\pm 0.1071.91 ±\pm 2.75100.00 ±\pm 0.0074863251
DistalPhalanxOutlineAgeGroup59.81 ±\pm 0.1079.37 ±\pm 0.4370.13 ±\pm 2.5579.09 ±\pm 0.0474.78 ±\pm 2.8972.40 ±\pm 0.1078.24 ±\pm 2.2981.74 ±\pm 2.5582735641
DistalPhalanxOutlineCorrect73.83 ±\pm 0.1068.08 ±\pm 0.3755.92 ±\pm 2.7465.10 ±\pm 0.8679.91 ±\pm 0.1573.40 ±\pm 2.7779.57 ±\pm 1.5680.76 ±\pm 0.1046872531
DistalPhalanxTW48.11 ±\pm 0.1075.39 ±\pm 0.0867.26 ±\pm 0.0875.78 ±\pm 0.1074.61 ±\pm 1.170.00 ±\pm 0.1075.78 ±\pm 0.1079.30 ±\pm 0.3974625821
DodgerLoopDay13.79 ±\pm 0.1018.96 ±\pm 2.7920.69 ±\pm 2.9017.24 ±\pm 2.8329.31 ±\pm 2.7834.48 ±\pm 2.8321.84 ±\pm 2.6626.20 ±\pm 1.7386572143
DodgerLoopGame51.61 ±\pm 0.1061.29 ±\pm 2.9053.22 ±\pm 1.6258.06 ±\pm 2.7959.68 ±\pm 2.8475.81 ±\pm 2.7859.14 ±\pm 2.7777.42 ±\pm 2.8183764251
DodgerLoopWeekend70.97 ±\pm 0.1083.87 ±\pm 2.5091.94 ±\pm 1.6282.26 ±\pm 2.8788.71 ±\pm 2.8395.16 ±\pm 1.6184.95 ±\pm 2.9196.77 ±\pm 0.1086374251
ECG20066.67 ±\pm 0.1073.22 ±\pm 0.9078.02 ±\pm 2.9081.59 ±\pm 2.7184.38 ±\pm 2.7976.34 ±\pm 2.8180.36 ±\pm 2.7186.31 ±\pm 2.9087532641
ECG500056.93 ±\pm 0.1093.16 ±\pm 0.1092.28 ±\pm 0.4991.21 ±\pm 0.2991.96 ±\pm 0.3392.43 ±\pm 0.5492.21 ±\pm 0.0894.21 ±\pm 0.0882476351
Average36.9255.1645.7250.9458.9360.4159.8468.877.004.405.874.873.773.403.671.50
审稿意见
5

This paper proposes a framework of functional Virtual Adversarial Training (f-VAT) for semi-supervised time series classification. The theoretical analysis and foundation are established for the proposed algorithm. The experimental results demonatre its effectiveness.

优缺点分析

Strengths: This paper tries to utlize the Virtual Adversarial Training (VAT) method on leveraging the unlabeled data of semi-supervised time series classification for smoothing the predictive distributions and proposes the framework of functional Virtual Adversarial Training (f-VAT) to incorporate the functional structure of the data into perturbations. It theoretically establishes a duality between the perturbation norm and the functional model sensitivity, and then use an appropriate Sobolev norm to generate structured functional adversarial perturbations for semi-supervised time series classification. The proposed algorithm is reasonable and novel for semi-supervised time series classification. The experiments strongly support the theoretical analysis.

Weeknesses: The description of theoretical results is not so logic or explainable. It is a big jump to make the derivation from the linear model to the continously differentiiable case. The authors should give a good explain for this generalization in the paper.

问题

Accoding to Eq.(11), the H^s norm can be only approximated, how about the approximation error here? Is there the other method to compute the norm? How to select the proper s for a practical problem or task?

局限性

Yes

最终评判理由

The authors addressed my concerns and questions and I still kept my previous scores.

格式问题

No.

作者回复

W1: Description of theoretical results

We apologize for not making the generalization clear in the original version. The case of linear model (as also functional linear model in the infinite-dimensional case) provides main insights into the duality between the adversarial perturbation norm and the model sensitivity norm, while the generalization to the continuously differentiable case serves mainly as additional theoretical supports for more complex models, though, due to the highly complex nature of neural networks, further theoretical analysis would be beyond the scope of this paper. Nevertheless, our existing theory has already provided much insight for the design of f-VAT with the duality theory--we should choose the proper Sobolev HsH^{-s} norm (but not HsH^s) for adversarial perturbations, which is key for the success of our method.

We hope that this explanation clarifies your doubts, and we will make it more clear in the revised version.

Q1: Approximation error

Thank you for your insightful question. The discrete approximation of the HsH^s norm is known as the spectral method [Canuto et al., 2006], which is the most common way to approximate the Sobolev norm with discrete observations. The approximation error is determined by multiple factors, including the order ss, the distribution of the samples, and the smoothness of the function. Roughly speaking, assuming that uHru \in H^r for some r>sRr > s \in \mathbb{R}, the approximation error of the HsH^s norm with NN equidistant samples would be O(Nrs)\mathcal{O}(N^{r-s}).

There are indeed other methods to compute the Sobolev norm, such as finite element methods and wavelet methods. However, they are more complicated and less efficient. Since the approximation error is not our main focus, we think the spectral method is sufficient for our purpose.

References:

  • [Canuto et al., 2006] Claudio Canuto, M. Youssuff Hussaini, Alfio Quarteroni, and Thomas A. Zang. Spectral Methods: Fundamentals in Single Domains. Scientific Computation. Springer Berlin Heidelberg,2006.

Q2: Selection of ss

Thanks for your suggestions. We evaluate the performance on more datasets in Table, where s=0s=0 reduces to the original VAT[Takeru et al., 2018]. Table shows that setting s=2s=2 achieves the best performance in almost all settings. For high volatile time series data, setting high order s=3s=3 generates adversarial perturbations that capture low-frequency temporal structure, while for relatively stable time series data, the small order s=1s=1 allows the perturbation to flexibly explore the input space. In practice, it suffices to evaluate s1,2,3,4s\in{1,2,3,4} based on the validation set and report the average performance over several runs.

The performance of f-VAT with different Sobolev norm order ss:

ssCricketXUWaveInsectWingNATOPSSelfReg
058.63 ±\pm 0.5094.76 ±\pm 0.5463.48 ±\pm 0.3090.15 ±\pm 1.6053.47 ±\pm 1.04
159.91 ±\pm 2.3296.54 ±\pm 0.6766.70 ±\pm 0.5089.58 ±\pm 0.1256.16 ±\pm 1.73
261.66 ±\pm 2.3397.16 ±\pm 0.2867.08 ±\pm 0.8693.13 ±\pm 0.1558.86 ±\pm 0.35
360.44 ±\pm 0.2396.82 ±\pm 0.1664.10 ±\pm 0.8690.10 ±\pm 0.6551.39 ±\pm 1.21
458.22 ±\pm 3.3996.71 ±\pm 0.6166.11 ±\pm 0.8287.51 ±\pm 0.5250.93 ±\pm 0.12

Reference

  • [Takeru et al., 2018] Takeru Miyato, Shin-ichi Maeda, Masanori Koyama, and Shin Ishii. Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018.
审稿意见
5

This paper proposes functional Virtual Adversarial Training (f-VAT), a novel method for semi-supervised time series classification. The key insight is that standard Virtual Adversarial Training (VAT), when applied to time series, uses Euclidean norm-bounded perturbations, which ignores the temporal structure of the data and leads to unrealistic "jagged" adversarial examples. To address this, f-VAT incorporates the structure of time series by constraining the adversarial perturbations using a Sobolev norm. Empirically, the authors demonstrate that f-VAT significantly outperforms standard VAT and other state-of-the-art methods on several semi-supervised and fully-supervised time series classification benchmarks.

优缺点分析

Strengths
  • The paper introduces a novel and well-motivated approach, where the authors use functional analysis tools (Sobolev norms) to generate structured adversarial perturbations for time series data
  • The work is technically strong, where it provides a theoretical foundation for the proposed framework, establishing a duality between the perturbation norm and the model's sensitivity.
  • The empirical evaluation is comprehensive, where the proposed method is tested on six datasets, consistently showing significant performance improvements over multiple baselines, including standard VAT. The results hold across different ratios of labeled data and in a fully-supervised setting.
  • The paper is well-written. The qualitative analyses, including visualizations of perturbations and loss landscapes, provide good intuition and support the paper's claims.
Weaknesses
  • The paper's empirical validation of the Sobolev order s is based on a single dataset (NATOPS), where s=2 was found to be optimal. It is not entirely clear if this choice is universally optimal or if s is a sensitive hyperparameter that would require tuning for each new dataset. A brief discussion on the generalization of this choice would strengthen the paper.
  • While the paper mentions the efficacy of the norm computation (O(N log N)), I could not find an explicit comparison of the total training time between f-VAT and standard VAT. A discussion on the practical computational overhead would be a valuable addition.

问题

  • Regarding the choice of the Sobolev order s, do the authors have any practical guidance on selecting s for a new task without performing an exhaustive search?
  • Could you elaborate on how one might design other norms for different types of time series? For instance, for data with known seasonality or periodic patterns, could a norm be designed in the Fourier domain to specifically encourage or penalize perturbations in certain frequency bands?

局限性

The authors have adequately addressed the limitations of their work in Section 5

最终评判理由

The author response provides the justification for s and the empirical runtime data, which address my initial concern. As a result, I'm increasing my score.

格式问题

NA

作者回复

W1 & Q1: Selection of ss

Thanks for your suggestions. We evaluate the performance on more datasets in Table, where s=0s=0 reduces to the original VAT[Takeru et al., 2018]. Table shows that setting s=2s=2 achieves the best performance in almost all settings. For high volatile time series data, setting high order s=3s=3 generates adversarial perturbations that preserve low-frequency temporal information, while for relatively stable time series data, the small order s=1s=1 allows the perturbation to flexibly explore the input space. In practice, it suffices to evaluate s[1,2,3,4]s\in[1,2,3,4] based on the validation set and report the average performance over several runs.

The performance of f-VAT with different Sobolev norm order ss:

ssCricketXUWaveInsectWingNATOPSSelfReg
058.63 ±\pm 0.5094.76 ±\pm 0.5463.48 ±\pm 0.3090.15 ±\pm 1.6053.47 ±\pm 1.04
159.91 ±\pm 2.3296.54 ±\pm 0.6766.70 ±\pm 0.5089.58 ±\pm 0.1256.16 ±\pm 1.73
261.66 ±\pm 2.3397.16 ±\pm 0.2867.08 ±\pm 0.8693.13 ±\pm 0.1558.86 ±\pm 0.35
360.44 ±\pm 0.2396.82 ±\pm 0.1664.10 ±\pm 0.8690.10 ±\pm 0.6551.39 ±\pm 1.21
458.22 ±\pm 3.3996.71 ±\pm 0.6166.11 ±\pm 0.8287.51 ±\pm 0.5250.93 ±\pm 0.12

Reference

  • [Takeru et al., 2018] Takeru Miyato, Shin-ichi Maeda, Masanori Koyama, and Shin Ishii. Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018.

W2: Runtime comparison

Thanks for your suggestions. Compared with VAT (O(N)\mathcal{O}(N)), the computational complexity of f-VAT is O(Nlog(N))\mathcal{O}(N\log(N)) with a small constant factor, involving the FFT-based Hs\|\cdot\|_{H^{-s}} normalization. The empirical results show that the extra computational cost remains in the same order as VAT.

Runtime comparison f-VAT and VAT:

MethodCricketXUWaveInsectWingNATOPSSelfReg
VAT15.6751.6635.5828.0430.68
f-VAT20.4562.0545.9238.9844.95
Δ (%)30.5020.1129.0639.0246.51

Q2: Norms for different types of time series

Thanks for your insightful comments. Indeed, we should design other norms to align with different types of time series. For example, for data with known seasonality or periodic patterns, we can design a norm that additionally penalizes the non-periodicity of the perturbation to align with the periodic patterns, which could be done using the Fourier domain decomposition. We believe that investigating additional structure of the time series would further enhance the performance of the f-VAT method, and it can be one of our future directions.

评论

Thank you for providing the justification for s and the empirical runtime data, which address my initial concern. As a result, I'm increasing my score.

评论

Thank you as well! We will follow your suggestion to provide a more detailed justification for ss and include additional runtime analyses in our revised manuscript. We sincerely appreciate your positive assessment of our paper.

评论

Dear Reviewers,

Thank you for your time and effort in reviewing submissions for NeurIPS 2025. As we approach the final stage of the review process, we kindly remind you to submit your responses to the author rebuttals by August 6.

Your engagement in this discussion phase is crucial to ensuring a fair and thorough evaluation of each submission. Please:

  • Carefully consider the authors’ rebuttal and any additional evidence they provide.

  • Update your review (if applicable) to reflect your revised perspective.

The authors provided a thorough experimental analysis in their response to Reviewer QUWH and Reviewer RR58, Could you please have a look at the rebuttal and reply to the authors as soon as possible?

Your AC

最终决定

This paper proposes functional Virtual Adversarial Training (f-VAT), a new framework for semi-supervised time series classification. Unlike standard VAT, which relies on Euclidean norm perturbations that fail to capture temporal structures, f-VAT introduces structural functional perturbations. By establishing a duality between perturbation norms and model sensitivity, the authors design structured adversarial perturbations that preserve low-frequency information while suppressing high-frequency noise. Extensive experiments on multiple datasets demonstrate consistent and significant performance improvements over baselines.

The strengths of this paper include:

  1. The work introduces a novel and well-motivated approach by applying Sobolev norms to construct structured adversarial perturbations that are more appropriate for time series data.

  2. The paper is theoretically solid, providing a clear foundation and duality between perturbation norm and model sensitivity.

  3. The empirical evaluation is comprehensive, demonstrating improvements across multiple semi-supervised datasets and settings.

  4. The paper is also clearly written, with qualitative visualizations that support the claims.

The authors further clarified their theory contributions during the rebuttal period. They have also provided additional experiments on more datasets, runtime comparisons, and clarifications on the selection of the Sobolev order, which resolved most of the reviewers’ concerns. Reviewers acknowledged that their initial issues were addressed.

I recommend acceptance. This is a technically strong paper with clear theoretical contributions and thorough empirical validation. That said, the revision should