PaperHub
6.0
/10
Rejected4 位审稿人
最低5最高8标准差1.2
8
5
6
5
3.8
置信度
ICLR 2024

Resource Efficient Test-Time Training with Slimmable Network

OpenReviewPDF
提交: 2023-09-22更新: 2024-02-11

摘要

关键词
Test-Time TrainingResource EfficientSlimmable Neural Network

评审与讨论

审稿意见
8

This paper introduces SlimTTT, a resource-efficient approach for test-time training. The author proposed a practical scenario of TTT under resource constraints. To address such a challenge, the method utilizes slimmable network that can flexibly switch between different widths of sub-networks for adaptive inference and requires no retraining. The method includes width-enhanced contrastive learning, logit consistency regularization, and global feature alignment to promote representation consistency among sub-networks, in both feature and prediction spaces. The method is demonstrated by the experiment against other TTT methods by the adaption to corrupt samples on several datasets. It can achieve superior performance in the same resource constraint and can be generalized with several different backbones.

优点

  1. Proposed a new challenge of resource constraint in such a field, which fits in with realistic demands as a result.

  2. Proposed a test-time training method for slimmable neural network.

  3. Comprehensive experiments that support the methodology and each sub-module of it.

  4. The framework considered comprehensively every part of an image classification pipeline.

缺点

  1. The definition "Efficient" is vague in the title, that the model result of your method in inference is efficient, but not during the training process(it actually costs more time as claimed in your experiment). Does the efficiency come from utilizing the slimmable models and you proposed a TTT method for them?

  2. There are some basic assumptions that should be explained.

  3. The purpose of the TTT is to generalize the model when a distribution shift occurs, as your experiment proves by using corrupt samples. This logical correlation can be straightly pointed out.

问题

  1. In some works about explainable networks, the grad-cam result shows what the model is focusing on and indicates the semantical feature that is used to make decisions. Does the bigger-width network contain the attention information of smaller networks well in all these backbones? If not, is the bigger network always better?

  2. In some cases of classification, one class can include several visual patterns (eg. some datasets have coarse classes like “vehicle ” including items that do not look exactly the same.) Will these have an impact if you align all networks to the supernet, as it may focus on several instance-level features? I'm also curious if this will have a bigger impact on the contrastive learning that expands the positive data with the different representations from 2 sizes of model.

  3. There seems to be a paradox between the purpose of consistency and the ensemble of results in 3.4, if the difference between all subnets to the super net is lowered, will they intend to focus on the same feature? Though it improves the performance, I'm not fully understanding the reason.

评论

Q1-1: The definition "Efficient" is vague in the title, that the model result of your method in inference is efficient, but not during the training process(it actually costs more time as claimed in your experiment).

A1-1: Good comment. The term "Efficient" in our paper encompasses two key aspects:

Firstly, efficiency is achieved through the ability to seamlessly switch to smaller and more efficient sub-networks. Despite the test-time training process of our slimTTT taking slightly more time than other methods, we can utilize smaller sub-networks as if they were the largest networks (e.g., employing the R-50 (0.75×) sub-network as the largest network). Remarkably, we observe that the results obtained from the R-50 (0.75×) sub-network demonstrate comparable performance to TTAC (R-50) and even surpass TTT++ (R-50), offering a more efficient alternative.

Secondly, efficiency is also evident in the streamlined approach of deploying and training only one model containing multiple sub-networks for adaptive switching. In contrast, employing previous TTT methods on real world scenarios necessitates deploying and training several models independently to accommodate varying resource constraints. The uploading, downloading, and training processes associated with multiple models of other methods contribute significantly to resource consumption, further underscoring the efficiency of our approach.

Q1-2: Does the efficiency come from utilizing the slimmable models and you proposed a TTT method for them?

A1-2: We indeed address the resource-efficient test-time training challenges by employing slimmable networks. However, simply incorporating slimmable networks into test-time training poses several challenges: 1. Different widths of networks may interfere with each other, leading to inconsistent update directions, preventing models of various widths from achieving optimal optimization. This issue is more pronounced in test-time training due to the lack of ground truth labels as well as the distribution shift. Therefore, we propose WCL and LCR to solve this challenge, which promotes the consistency between different width of networks in both feature space and prediction space. 2. The GFA is applied to ensure that the test features do not deviate far from the source feature distribution during the whole adaptation process and it can apply a restriction over the multi-view features for each width of sub-networks and ensure the correctness of the multi-view representations.

In the following analytical experiment, we attempt a straightforward application of slimmable networks in test-time training without employing our consistency method (referred to as naive SlimTTT). Unfortunately, this approach yields unsatisfactory results, particularly when applied in an online manner. This is attributed to the fact that our consistency loss functions—WCL, LCR, and GFA—leverage the diverse views captured by sub-networks of varying widths. This facilitates knowledge interaction among slimmable networks, a crucial aspect in the test-time training setting.

MethodmannerwidthCifar10-C Err. Avg.mannerwidthCifar10-C Err. Avg.
naive SlimTTToffline1.010.80online1.014.78
0.7511.540.7515.64
0.512.590.516.72
0.2515.990.2518.83
SlimTTT (ours)offline1.08.33online1.010.17
0.758.630.7510.55
0.59.120.511.41
0.2510.680.2514.01
评论

Q2: There are some basic assumptions that should be explained.

A2: Thank you for your suggestion. In the test-time training phase, we follow the approach of TTT++ and TTAC, utilizing an offline training manner. After deploying our slimmable networks to different devices according to their maximum resource constrains, we assume that the test-time training is conducting when the devices have its maximum resources (e.g., the devices are charging).

However, motivated by Reviewer 1Qhs, now we find that even though the available compute changes frequently during test-time training in an online manner, our SlimTTT exhibits low sensitivity to changes in resources as shown below. We monitored the sensitivity of the performance in such settings on CIFAR10-C (Gaussian Noise), as different model parameters are updated based on the available compute. We implemented 12 time steps with dynamic resources throughout the period. For example, the resources change after time step 2, so we need to update the 0.75 parameters of the model rather than the whole model.

Comparing the performance on dynamically changing resources (the upper half of the table) to that on constant resources (the lower half of the table), the results below reveal that our SlimTTT exhibits low sensitivity to changes in resources. We believe that this dynamic evaluation you mentioned is both important and practical, and we intend to incorporate these findings into our new paper version to encourage future research.

Dynamic resourcesTime-step0123456789101112
FLOPs (G)4.14.14.12.32.31.11.10.30.31.11.12.32.3
max width1110.750.750.50.50.250.250.50.50.750.75
Err. (1.0)19.5320.1217.8416.3114.9214.7814.7314.3613.8513.5514.0613.9913.82
Err. (0.75)22.2721.2918.7517.1916.0215.7615.7415.3314.8414.4914.9514.8114.78
Err. (0.5)25.0023.6320.5719.1417.9717.5116.9616.4115.8915.3515.5915.4315.38
Err. (0.25)28.1328.5225.2623.9322.2721.8121.0420.4119.7519.1819.3218.9518.96
Constant resourcesTime-step0123456789101112
FLOPs (G)4.14.14.14.14.14.14.14.14.14.14.14.14.1
max width1111111111111
Err. (1.0)19.5320.1217.8416.4115.0814.8414.7314.2613.7613.3613.3913.3813.37
Err. (0.75)22.2721.2918.7517.0916.0215.7615.6815.1414.5414.0614.2414.1614.24
Err. (0.5)25.0023.6320.5718.8517.7317.5117.0216.3615.7615.1615.3415.1015.08
Err. (0.25)28.1328.5225.2624.2222.2721.6820.8720.1219.3118.7918.9318.5918.51

Q3: The purpose of the TTT is to generalize the model when a distribution shift occurs, as your experiment proves by using corrupt samples. This logical correlation can be straightly pointed out.

A3: Thank you for your suggestion. We will add “our conducted experiments use corrupt samples to imitate the distribution shift.” in the new version of our paper.

Q4: In some works about explainable networks, the grad-cam result shows what the model is focusing on and indicates the semantical feature that is used to make decisions. Does the bigger-width network contain the attention information of smaller networks well in all these backbones? If not, is the bigger network always better?

A4: Good comment. Bigger-width networks encompass more parameters, potentially containing attention information from smaller networks in certain instances. However, as illustrated in Fig.1 in our introduction, our key observation shows that bigger-width networks can capture entirely distinct perspectives compared to smaller networks. Hence, it is not accurate to claim that larger-width networks always include the attention information of smaller networks.

Given that larger-width networks have more parameters, they typically exhibit better overall performance than the smaller networks. Nevertheless, in scenarios where larger-width networks capture entirely different perspectives compared to smaller networks, the smaller networks can complement the knowledge of larger networks. This mutual supplementation leads to direct knowledge exchange and mutual promotion, elucidating why our method proves effective.

评论

Q5: In some cases of classification, one class can include several visual patterns (e.g., some datasets have coarse classes like “vehicle ” including items that do not look exactly the same.) Will these have an impact if you align all networks to the supernet, as it may focus on several instance-level features? I'm also curious if this will have a bigger impact on the contrastive learning that expands the positive data with the different representations from 2 sizes of model.

A5: Thanks for the insightful comment. In our study, a single image (not images belonging to the same class) undergoes two different data augmentations and is then processed by four networks of varying widths, resulting in eight features. These features collectively form the positive data for that particular image. Consequently, the positive data for an image doesn't only involve the largest model size but also encompasses features from other model sizes.

Furthermore, the multi-views we emphasize in our paper refer to distinct discriminative features that different widths of sub-networks can capture (e.g., the wheel or the car window, both classifiable as a car). Despite datasets having broad classes like "vehicle" that include items with dissimilar appearances, our approach leverages knowledge interaction during test-time training that enables different widths of networks to capture a wide range of views (discriminative features) associated with coarse classes like "vehicle". Therefore, though the fact that the class "vehicle" may encompass items with varying appearances, our networks are capable of capturing the discriminative features specific to each, ultimately leading to enhanced performance compared to individual training.

Additionally, we conducted an experiment on CIFAR10-C where we don’t align smaller networks to the supernet but to the neighbor width network (e.g., 0.75× width align to the 1.0× width, 0.5× width align to the 0.75× width…). The results indicate that aligning smaller networks to different sizes of the model does not significantly impact contrastive learning.

MethodwidthErr. Avg.
neighbor SlimTTT1.08.55
0.758.94
0.59.49
0.2511.04
SlimTTT1.08.33
0.758.63
0.59.12
0.2510.68

Q6: There seems to be a paradox between the purpose of consistency and the ensemble of results in 3.4, if the difference between all subnets to the super net is lowered, will they intend to focus on the same feature? Though it improves the performance, I'm not fully understanding the reason.

A6: Good comment. The primary objective of enforcing consistency is to facilitate knowledge interaction among sub-networks of varying widths. Our aim is to encourage the multi-views captured by different widths of sub-networks to complement each other. Specifically, the supernet guides the smaller networks to capture accurate views for improved image classification, while the smaller networks supplement views that the supernet may have overlooked, enhancing overall robustness.

However, owing to differences in parameter quantities, networks of various widths cannot capture all perspectives of an object. The supernet tends to cover most views, while the smaller networks retain unique views by aligning not only to the supernet but also to two augmented features of themselves. Consequently, despite that their consistency is promoted, networks of different widths may still capture diverse views. Therefore, we propose an ensemble strategy to guarantee the utilization of all views captured by the networks to improve generalization. This approach has empirically proven to be effective in enhancing the overall performance.

评论

Thanks to the authors for your detailed responses. Some of my concerns have been addressed. Therefore, I will raise my score accordingly.

评论

We thank you for your valuable time and feedback.

审稿意见
5

The authors identify two main shortcomings of existing TTT methods: resource inefficiency during deployment across devices and an inability to handle computational budget variations during testing. To address these, the authors propose SlimTTT, a resource-adaptive test-time training framework. They introduce techniques like Width-enhance Contrastive Learning (WCL), Logits Consistency Regularization (LCR), and Global Feature Alignment (GFA) to promote representation consistency. SlimTTT is reported to achieve state-of-the-art results across various adaptation methods and datasets, with a significant reduction in inference complexity. My main concerns lie on the motivation of SlimTTT and its evaluation protocols.

优点

The proposed method properly integrates Width-enhance Contrastive Learning, Logits Consistency Regularization, and Global Feature Alignment and thus achieves significant performance improvement compared with prior source-free unsupervised domain adaptation methods, such as SHOT.

The findings of “different width of sub-networks can capture multiple views of the same input data, possessing a talent for enhancing TTT performance” is interesting and provides new insights.

缺点

The motivation for slimmable TTA does not convince me. As all TTA methods are inference-time methods and the adapttion+inference is often conducted on the same device, in Table 1, the computational resource consumption comparison should include both training (GPU memory, wall-clock time) and inference, rather than only #Model Params and Flops during inference. I mean, for a resource-limited device, whether a TTA method can be run is determined by the training resource request, rather than inference.

In Table 1, ResNet-50(1.0×\times)+TENT+ImageNet-C@56.89 is evaluated in an online manner (to my knowledge). So if SilmTTT is evaluated in an offline manner, the comparison would be unfair.

Does the proposed SlimTTT work well in the online setting? I am curious about the online performance and TTA efficiency (training+inference) comparisons with compared methods.

Comparisons with more resource-efficient TTA methods (perfectly under the online setting) are preferred, such as EATA [Efficient Test-Time Model Adaptation without Forgetting. ICML 2022].

问题

For all compared methods, are they evaluated in online or offline setting?

评论

Q1: The motivation for slimmable TTA does not convince me. As all TTA methods are inference-time methods and the adapttion+inference is often conducted on the same device, in Table 1, the computational resource consumption comparison should include both training (GPU memory, wall-clock time) and inference, rather than only #Model Params and Flops during inference. I mean, for a resource-limited device, whether a TTA method can be run is determined by the training resource request, rather than inference.

A1: Thank you for your suggestion. We follow the offline TTT setting from TTT++ and TTAC. In this offline manner, test-time training and the inference phase are separate and independent. Therefore, in this scenario, it is reasonable to consider the resources during training and inference separately.

The motivation for our SlimTTT actually accounts for both the inference cost and the test-time training cost. As depicted in Fig.1 (left) in our introduction, the blue segment indicates the deployment of networks of various widths after training-time training, where we already take into account the maximum resource constraints of the device during test-time training (e.g., if the maximum resource constraints during test-time training of the device can only support R-50 (0.75×), we deploy the R-50 (0.75×) sub-network utilize it as the largest network). After it meets the resource constraint of each device during test-time training, SlimTTT has the ability to further switch between sub-networks during test-time inference according to the device’s status, as demonstrated in the orange segment.

To assess the efficacy of our SlimTTT under varying resource constraints during test-time training, we utilize smaller sub-networks as if they were the largest networks (e.g., employing the R-50 (0.75×) sub-network as the largest network). Remarkably, we observe that the results shown in Fig.3 in our experiment chapter obtained from the R-50 (0.75×) sub-network demonstrate comparable performance to TTAC (R-50) and even surpass TTT++ (R-50), offering a more efficient alternative during test-time training phase.

The computational resource consumption comparison and results on ImageNet-C in test-time training phase are shown in the table below.

MethodBackboneGPU memory (peak) (G)wall-clock time (per batch) (s)FLOPs (G)Param. (M)Err. Avg.
TTACR-101>242.8427.844.643.73
TTACR-5023.091.8854.125.645.71
SlimTTTR-5022.592.2537.825.643.98
TTACR-3415.980.8913.721.847.57
SlimTTTR-50 (0.75)19.811.2293.714.745.41
TTACR-1810.990.5591.811.951.34
SlimTTTR-50 (0.5)12.490.7601.36.948.73
SlimTTTR-50 (0.25)6.040.4200.22.058.08

Remarkably, when the available compute changes frequently during test-time training phase, different width of our slimmable network will be updated based on the available compute, while other TTT methods can not conduct test-time training under such dynamic training resource conditions.

We monitored the sensitivity of the performance in such settings on CIFAR10-C (Gaussian_Noise), as different model parameters are updated based on the available compute. We implemented 12 time steps with dynamic resources throughout the period. For example, the resources change after time step 2, so we need to update the 0.75 parameters of the model rather than the whole model.

Comparing the performance on dynamically changing resources (the upper half of the table) to that on constant resources (the lower half of the table), the results below reveal that our SlimTTT exhibits low sensitivity to changes in resources. We believe that this dynamic evaluation you mentioned is both important and practical, and we intend to incorporate these findings into our new paper version to encourage future research.

Dynamic resourcesTime-step0123456789101112
FLOPs (G)4.14.14.12.32.31.11.10.30.31.11.12.32.3
max width1110.750.750.50.50.250.250.50.50.750.75
Err. (1.0)19.5320.1217.8416.3114.9214.7814.7314.3613.8513.5514.0613.9913.82
Err. (0.75)22.2721.2918.7517.1916.0215.7615.7415.3314.8414.4914.9514.8114.78
Err. (0.5)25.0023.6320.5719.1417.9717.5116.9616.4115.8915.3515.5915.4315.38
Err. (0.25)28.1328.5225.2623.9322.2721.8121.0420.4119.7519.1819.3218.9518.96
Constant resourcesTime-step0123456789101112
FLOPs (G)4.14.14.14.14.14.14.14.14.14.14.14.14.1
max width1111111111111
Err. (1.0)19.5320.1217.8416.4115.0814.8414.7314.2613.7613.3613.3913.3813.37
Err. (0.75)22.2721.2918.7517.0916.0215.7615.6815.1414.5414.0614.2414.1614.24
Err. (0.5)25.0023.6320.5718.8517.7317.5117.0216.3615.7615.1615.3415.1015.08
Err. (0.25)28.1328.5225.2624.2222.2721.6820.8720.1219.3118.7918.9318.5918.51
评论

Q2: In Table 1, ResNet-50(1.0×)+TENT+ImageNet-C@56.89 is evaluated in an online manner (to my knowledge). So if SilmTTT is evaluated in an offline manner, the comparison would be unfair.

A2: In our paper, all of results of the compared methods are implemented to be evaluated in an offline manner following the TTT++ and TTAC setting, including TENT (TTT++ includes the implementation of TENT for test-time training in offline manner in its source code, we replicated TENT in offline manner base on the source code of TTT++ across different datasets). So, the comparison is fair.

Q3: Does the proposed SlimTTT work well in the online setting? I am curious about the online performance and TTA efficiency (training+inference) comparisons with compared methods.

A3: Good comment. As you suggested, here we further compare SlimTTT to the baseline methods in the online setting. We follow the TTAC (YO) online setting to compare our SlimTTT with other methods on CIFAR10-C. From the results below, we find that our approach still managed to surpass other methods, despite removing the queues present in other approaches (TTAC), resulting in a significant reduction in training time, and we will append the online results in our new version of paper.

CIFAR10-C (online performance):

MethodwidthBirtContrDefocElastFogFrostGaussGlassImpulJpegMotnPixelShotSnowZoomAvg
TENT1.08.228.079.9318.2915.6514.1416.6024.1025.8013.3912.3411.0614.7513.877.8714.27
TTT++ (w/ q)1.07.707.919.2417.5516.3912.7415.4922.5722.8613.0212.5211.4614.4513.907.5113.69
TTAC (w/ q)1.06.418.057.8514.8110.2810.5113.0618.3617.3510.808.979.3411.6110.016.6810.94
SlimTTT(w/ q)1.05.676.247.9213.028.889.6413.2416.0615.1811.128.898.5112.199.446.4910.17
0.755.956.428.3213.529.5610.2113.3816.6015.5811.279.388.7112.5110.086.8210.55
0.56.557.339.0514.5410.0111.0214.6817.3717.2212.2510.059.6013.5310.637.2711.41
0.258.249.3610.7917.2412.2013.5917.8120.6921.8315.1412.2412.1216.2613.509.1714.01
SlimTTT(w/o q)1.05.516.177.9013.149.189.4312.6616.6517.6310.819.078.6611.879.415.8810.26
0.755.886.697.8713.739.759.6413.5517.5218.1811.359.549.0512.0810.026.1610.73
0.56.627.528.7514.4110.6110.6614.8618.8719.8712.7410.6510.0713.2810.807.0111.78
0.258.3610.0310.9818.8914.4313.9619.3123.8126.8016.1113.2413.1017.1014.459.1815.32

CIFAR10-C (online efficiency: TENT, TTAC (with queue), SlimTTT (w/o queue)):

training time (s)/ itertraining GPU memoryinference time (s) / batchinference GPU memory
TENT0.14612.14G0.05743.59G
TTAC (w/ queue)40.08919.80G0.05743.59G
SlimTTT (w/o queue)4.15616.86G0.11333.59G

Q4: Comparisons with more resource-efficient TTA methods (perfectly under the online setting) are preferred, such as EATA.

A4: Thank you for your suggestion. The comparison with EATA under the online setting on ImageNet-C using ResNet50 as the backbone is shown below, the results show that our SlimTTT can outperform EATA in most of the corruption tasks under online setting.

MethodwidthBirtContrDefocElastFogFrostGaussGlassImpulJpegMotnPixelShotSnowZoomAvg
TTAC1.030.3638.8469.0639.6736.0150.2066.1870.1764.3645.5951.7739.7262.4344.5642.8050.11
EATA1.032.0055.7066.3042.4040.1054.3065.0066.6063.2045.0052.9039.3063.1048.6047.2052.11
SlimTTT1.031.4434.1561.1839.3234.1149.5361.3765.0859.9545.0444.9640.7558.7342.0541.1547.25
0.7533.0335.8562.2940.8935.6151.1862.5565.9861.2246.2546.4642.0760.0343.7842.6348.65
0.536.1739.1264.0643.838.7454.5364.9368.0563.7148.8949.3844.7862.4747.0745.5951.42
0.2542.8946.9770.0350.7346.4763.0270.9573.2469.9155.0156.5051.9568.8554.8952.5658.26

Q5: For all compared methods, are they evaluated in online or offline setting?

A5: All compared methods are evaluated in offline setting follow TTT++ and TTAC to guarantee a fair comparison. It is worth noting that our method can be used in an online manner, even in situations where training resources are dynamically changing, showcasing strong versatility of SlimTTT.

审稿意见
6

The paper presents a varying resource constraint setting for Test-Time training approaches and proposes an architecture based on the Slimmable Network (Yu et al., 2018). The primary assumption made for the targeted varying resource constraints setting is the dynamically changing computational budget (hardware-based computational constraints or dynamic resource allocation to the algorithm) during inference. Slimmable neural network architectures are specifically designed to provide adaptive accuracy-efficiency trade-offs during model inference on various devices with different budgets of computational resources. In this paper, the authors utilize Width-enhanced Contrastive loss and Logit Consistency Regularization for maintaining consistency between the subnetworks in both features and logits space. Further, the paper also introduces an ensembling strategy based on dynamically changing resources to boost the performance of the architecture. The pipeline for the proposed framework can be summarized in 3 modules 1) Source Training Phase: The pretrained backbone is trained on the source dataset with the primary training objective with an additional auxiliary self-supervised learning task. 2) Target Training Phase: for this phase, the paper proposes maintaining feature consistency between differently augmented views obtained by both slimmable network structure and data augmentation and adds another learning objective of Logits Consistency Regularization where an augmented version of the sample is sent to the largest network for obtaining a pseudo label which is further used to maintain prediction consistency among all the sub-networks. 3) In the last phase, the paper makes use of the predictions available from different slimmable networks to create an ensemble version of the predictions for boosting prediction performance.

The paper provides empirical results on four widely used benchmarks (ImageNet-C, CIFAR10-C, CIFAR10.1, and CIFAR100-C) with different backbones (ResNet, Mobilienet, and ViT) for four settings of switchable widths in Slimmable Network (1.0×, 0.75×, 0.5×, 0.25×). The reported results highlight an improvement over the compared baselines and show the computation cost comparison via reporting the FLOPs with prediction accuracies. The paper also provides ablation experiments for highlighting the impact of various components in the design of the proposed framework.

优点

  • The paper highlights an important issue of dynamically changing resource constraints in test-time training settings for deploying models in the real world. The paper utilizes the architecture of slimmable networks to address this issue, making test-time training approaches to incorporate adaptive accuracy-efficiency trade-offs during model inference on various devices with different budgets of computational resources.

  • The paper proposes the WCL, LCR, and GFA for exploiting the slimmable network to ensure consistency between multiple views captured by the architecture. The proposed design choices are sufficiently backed up by suitable ablation experiments. Moreover, as an additional advantage, the paper reports the slimmable network for TTT to be effective when compared with other baselines on the same computation budget.

  • The paper presents a detailed empirical study with various backbones (3 backbones) and datasets (4 benchmarks) along with different settings of Switches in Slimmable Network (4 in number), with required performance comparison over the computation cost for fair comparison making the results more reliable.

缺点

  • One of the primary claims of the paper is the varying inference budget during the inference (also highlighted as Challenge II in Figure 1). Since the paper targets a practical setting of dynamic resource allocation during inference, it is imperative to consider experiments where the available compute changes frequently during inference (test time training). It would be interesting to monitor the sensitivity of the performance in those settings since different parameters of the model will be updated based on the available compute. Low-performance sensitivity on dynamically changing resources will make the method more reliable for practical use cases. Moreover, it’ll promote future research to address the challenges faced in such a setting.

  • The proposed framework highly depends on the source domain training however, the training cost and the convergence rate comparison for various architecture is missing from the paper. It would be good to add a comparison of training times/ convergence rates of the proposed architecture to make the approach more transparent for real-world deployment use cases.

问题

Minor suggestions:

  • It would be better to update the caption of Table 3 to highlight the explanation of numbers presented in red for easier readability.
  • While making the comparison with existing methods, it would be good to highlight the dependency on the availability of the source dataset. Adding another column for a clear distinction between TTA and TTT approaches will make the paper more transparent.
评论

Q1: It would be interesting to monitor the sensitivity of the performance in those settings since different parameters of the model will be updated based on the available compute. Low-performance sensitivity on dynamically changing resources will make the method more reliable for practical use cases. Moreover, it’ll promote future research to address the challenges faced in such a setting.

A1: Good comment. As your suggestion, we conducted experiments on CIFAR10-C (Gaussian_Noise) and CIFAR100-C (Gaussian_Noise) wherein the available compute (e.g., FLOPs) varies frequently during the test-time training phase. In this experiment, we monitored the sensitivity of the performance in such settings, as different model parameters are updated based on the available compute. Specifically, our evaluation spans over 12 time-steps with dynamic resources throughout the period. For example, the resources change after time-step 2 (from 4.1G to 2.3G), so we need to update the 0.75 parameters of the model rather than the whole model.

Comparing the performance on dynamically changing resources (the upper half of the table) to that on constant resources (the lower half of the table), the results below surprisingly reveal that our SlimTTT exhibits low sensitivity to changes in resources (i.e., as of time step 12, the performance of the network trained with dynamic resources is comparable to the performance of the network trained with constant resources), while other TTT methods can not conduct test-time training under such dynamic training resource conditions. We believe that this dynamic evaluation you mentioned is both important and practical, and we intend to incorporate these findings into our new paper version to encourage future research.

CIFAR10-C (Gaussian_Noise):

Dynamic resourcesTime-step0123456789101112
FLOPs4.1G4.1G4.1G2.3G2.3G1.1G1.1G0.3G0.3G1.1G1.1G2.3G2.3G
max width1110.750.750.50.50.250.250.50.50.750.75
Err. (1.0)19.5320.1217.8416.3114.9214.7814.7314.3613.8513.5514.0613.9913.82
Err. (0.75)22.2721.2918.7517.1916.0215.7615.7415.3314.8414.4914.9514.8114.78
Err. (0.5)25.0023.6320.5719.1417.9717.5116.9616.4115.8915.3515.5915.4315.38
Err. (0.25)28.1328.5225.2623.9322.2721.8121.0420.4119.7519.1819.3218.9518.96
Constant resourcesTime-step0123456789101112
FLOPs4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G
max width1111111111111
Err. (1.0)19.5320.1217.8416.4115.0814.8414.7314.2613.7613.3613.3913.3813.37
Err. (0.75)22.2721.2918.7517.0916.0215.7615.6815.1414.5414.0614.2414.1614.24
Err. (0.5)25.0023.6320.5718.8517.7317.5117.0216.3615.7615.1615.3415.115.08
Err. (0.25)28.1328.5225.2624.2222.2721.6820.8720.1219.3118.7918.9318.5918.51

CIFAR100-C (Gaussian_Noise):

Dynamic resourcesTime-step0123456789101112
FLOPs2.3G2.3G2.3G4.1G4.1G0.3G0.3G1.1G1.1G2.3G2.3G1.1G1.1G
max width0.750.750.75110.250.250.50.50.750.750.50.5
Err. (1.0)64.0660.9459.9057.8155.4753.5251.7951.5650.9549.8449.6448.7648.62
Err. (0.75)62.5058.5955.7355.4753.7552.3450.4550.2949.6548.5948.2247.2747.12
Err. (0.5)63.2859.7758.3358.4056.5654.8252.3452.2551.2250.3950.5049.5449.70
Err. (0.25)75.7873.8372.1469.1466.8865.2362.6161.9160.9459.7759.7358.8559.07
Constant resourcesTime-step0123456789101112
FLOPs4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G
max width1111111111111
Err. (1.0)62.5058.256.5154.6952.5050.7848.6648.1447.3146.5646.0945.3145.13
Err. (0.75)62.5058.5955.7355.4753.7552.2150.2249.7148.9648.1247.7346.8146.57
Err. (0.5)63.2859.7758.3358.4056.5654.8252.2351.7650.5249.4549.4348.5048.62
Err. (0.25)75.7873.8372.1469.1466.8865.2362.7262.0161.0259.8459.8759.0559.25
评论

Q2: The proposed framework highly depends on the source domain training however, the training cost and the convergence rate comparison for various architecture is missing from the paper. It would be good to add a comparison of training times/ convergence rates of the proposed architecture to make the approach more transparent for real-world deployment use cases.

A2: Thank you for your suggestions. We provide the results below to show the training times of the proposed architecture (ResNet50, MobileNetV1, MobileNetV2 and ViT-Tiny) and we provide the convergence rate comparison (2000 epochs) for various architecture (ResNet50, MobileNetV1, MobileNetV2) in our new version of supplementary (highlighted in blue).

Training Times per epoch:

BackboneTraining Time/epoch
ResNet50173.48s
MobileNetV128.31s
MobileNetV243.07s
ViT-Tiny202.7s

S1: It would be better to update the caption of Table 3 to highlight the explanation of numbers presented in red for easier readability.

A1: Thank you for your suggestion. We have modified the caption of Table 3 to “The average error rate (%) results of SlimTTT with different backbones on CIFAR10-C. The numbers presented in red indicate the improvement of our method compared to TTAC”, and have updated it in our paper (highlighted in blue).

S2: While making the comparison with existing methods, it would be good to highlight the dependency on the availability of the source dataset. Adding another column for a clear distinction between TTA and TTT approaches will make the paper more transparent.

A2: Thank you for your suggestion. We will add another column like we show below to highlight the dependency on the availability of the source dataset in Table1 in the future version. “SSL head” in the table means whether we need to train an auxiliary branch in source training phase, “Availability of the source” means whether we can access the source data during test-time training.

BN (TTA)TTT-R (TTT)SHOT (Source-free)TENT (TTA)TTT++ (TTT)TTAC (TTT)SlimTTT (TTT)
SSL head×××
Availability of the source data×××××××
评论

Thank you for sharing the results. The additional details, like the requirement for retraining of SSL on source dataset, help improve the paper's clarity.

评论

Thank you for adding the additional results.

The results of dynamically changing computational budget do reveal a low sensitivity to the dynamic changes for CIFAR10C, and it is somewhat surprising that for a few cases, the model performs better over low dynamic resources. As a suggestion, an average over multiple runs column would make it easier for comparison.

CIFAR10C

Constant ResourceDynamic ResourceError Difference (Constant - Dynamic)
15.3915.52769231-0.1376923077
16.4030769216.63230769-0.2292307692
17.9315384618.06384615-0.1323076923
21.9384615422.11769231-0.1792307692

CIFAR100C

Constant ResourceDynamic ResourceError Difference (Constant - Dynamic)
50.9523076954.06615385-3.113846154
52.0284615452.30538462-0.2769230769
53.9746153854.39230769-0.4176923077
65.1353846265.067692310.06769230769

For CIFAR100C, the average performance difference seems to be significant. A brief discussion regarding the observation would also be helpful if there are cases where, on a longer range of dynamic resource changes, the model collapses due to massive error accumulation over time. Mentioning these observations clearly will improve the transparency of the work.

评论

Q: The results of dynamically changing computational budget do reveal a low sensitivity to the dynamic changes for CIFAR10C, and it is somewhat surprising that for a few cases, the model performs better over low dynamic resources. As a suggestion, an average over multiple runs column would make it easier for comparison. For CIFAR100C, the average performance difference seems to be significant. A brief discussion regarding the observation would also be helpful if there are cases where, on a longer range of dynamic resource changes, the model collapses due to massive error accumulation over time. Mentioning these observations clearly will improve the transparency of the work.

A:

Thank you very much for your suggestion! The results we previously reported at each time-step were already averages across all previous steps. Namely, the last column in table shows the average error rates over twelve time-steps and you can calculate the difference between constant and dynamic cases directly from it. For a more comprehensive results, as you suggested, we have now adjusted the table presentation and displayed the instantaneous results on CIFAR10-C (Guassian_Noise) and CIAFR100-C (Guassian_Noise) at each time-step on a longer range of dynamic resource changes. We report their averages across the whole process at the last column, where the values in parentheses represent the error difference (constant - dynamic).

CIFAR10-C (batchsize=256, total steps=40):

Dynamic resourcesTime-step1212132021383940Avg. 40 steps
FLOPs2.3G2.3G1.1G1.1G0.3G0.3G2.3G2.3G2.3G
max width0.750.750.50.50.250.250.750.750.75
Err. (1.0)21.8821.8813.2811.7112.8912.8916.4113.2813.6712.96
Err. (0.75)22.6520.7113.2815.6312.5012.5017.5812.5014.4513.73
Err. (0.5)25.0022.6613.2815.2313.2812.1117.9715.6313.6714.63
Err. (0.25)28.5229.3014.8417.9717.9714.4520.7018.3615.6317.80
Constant resourcesTime-step1212132021383940Avg. 40 steps (error difference)
FLOPs4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G
max width111111111
Err. (1.0)19.5320.7112.8913.2812.8912.1116.4113.6713.6712.81 (-0.15)
Err. (0.75)22.2720.3113.2814.8412.1112.5015.6313.2815.2313.51 (-0.22)
Err. (0.5)25.0022.2712.5014.8413.6711.7217.5814.0615.2314.55 (-0.08)
Err. (0.25)28.1328.9115.2317.5817.1914.4519.9218.3617.5817.62 (-0.18)

CIFAR100-C (batchsize=128, total steps=80):

Dynamic resourcesTime-step1212132021787980Avg. 80 steps
FLOPs2.3G2.3G1.1G1.1G0.3G0.3G2.3G2.3G2.3G
max width0.750.750.50.50.250.250.750.750.75
Err. (1.0)64.0657.8139.0646.8838.2839.8440.6349.2237.5041.08
Err. (0.75)62.5054.6936.7245.3144.5341.4136.7251.5639.0641.96
Err. (0.5)63.2856.2539.0651.5651.5641.4137.5049.2241.4145.15
Err. (0.25)75.7871.8849.2261.7256.2553.1346.8855.4756.2553.57
Constant resourcesTime-step1212132021787980Avg. 80 steps (error difference)
FLOPs4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G4.1G
max width111111111
Err. (1.0)62.5053.9136.7242.9737.5036.7239.8450.7835.9439.94 (-1.14)
Err. (0.75)62.5054.6936.7242.9745.3140.6336.7250.7838.2841.63 (-0.33)
Err. (0.5)63.2856.2538.2850.0052.3441.4138.2849.2241.4144.97 (-0.18)
Err. (0.25)75.7871.8850.0061.7255.4753.9146.8855.4756.2553.71 (+0.14)
评论

A: Based on the average results of networks with various widths trained for an extended period, we observed that the average performance of networks under dynamic resource conditions is inferior to that under static resource conditions with a small gap. Our updated supplementary material provides a comparison figure (highlighted in blue) of the convergence rates for the supernet under dynamic and static resource conditions. The figure shows that as training time increases, the average performance of networks under dynamic resource conditions gradually converges with that under static resource conditions.

According to this observation, we conclude that the optimal performance achievable by networks under dynamic resource conditions is not as high as that under static resource conditions. Additionally, the initial convergence rate of the networks under dynamic resource conditions is slower, explaining the significant performance difference observed after 12 steps in our initial response, which means the dynamic resources condition does have a certain impact on optimization of the slimmable networks. However, with prolonged training, the performance gap diminishes, reaching their respective optima. The results from longer training durations indicate that our approach does not suffer from massive error accumulation that could lead to model crashes. Our intuitive explanation for this phenomenon is that the consistency in multi-view at the feature and prediction space, as proposed in our approach, assists the optimization objectives of each network to gradually become more consistent as training progresses.

Furthermore, from the results of each step in the table, we were pleasantly surprised to find that, at certain steps, the model's performance under dynamic resource changes is better than its performance under static conditions. Our intuitive explanation for this phenomenon is that when adjusting the width based on dynamic resource changes at a particular step, the calculated gradient direction may be more favorable for the update of a specific width network. Consequently, the performance of that width network tends to be better than under static conditions at that step. This inspires our future research on the selection of updating parameters for networks of different widths.

审稿意见
5

This paper proposed to leverage slimmable networks in the scenario of test-time training to allow the model to meet different resource budgets during test-time training. The paper proposed width-enhanced contrastive learning which is to conduct contrastive learning among different network width to learn different representations. The proposed method shows better performance than previous works at different resource budgets

优点

  1. The paper has a clear explanation about the background and objectives of this work.
  2. The proposed method shows better performance than previous works at different resource budgets.

缺点

  1. It seems that this work is pretty much slimmable network, but just in the setting of test-time training. I don’t see what are the unique challenges in applying slimmable networks in the test-time training? The test-time training seems to be the same as training-time training, but just without labels. Then the problem seems to be how to apply slimmable networks in the un-supervised setting, which has been studied in previous works [1].
  2. The motivation that different sub-networks could capture different image features has been studied in [2].
  3. In Table 1, the other methods should also use ResNet-50 with different widths to have a fair comparison.
  4. This work finetuned the pre-trained slimmable networks on ImageNet and ImageNet-C, what about other works?
  5. In Table 5, why is the training cost is comparable or even faster than TTT++ and TTAC? The proposed method needs to forward and backward multiple times, I am assuming it should be more training expensive.

[1] Slimmable Networks for Contrastive Self-supervised Learning. [2] MutualNet: Adaptive ConvNet via Mutual Learning from Different Model Configurations. TPAMI.

问题

Please see the weakness part

评论

Q1-1: The test-time training seems to be the same as training-time training, but just without labels.

A1-1: Good comment. Test-time training (TTT) improves the model’s generalization on specific test data distribution in a way that is beyond the capabilities of training-time training. Since the distribution of the test data can be diverse and different from training data distribution, it is highly difficult for training-time training alone to train a model that can generalize to all possible test data distribution. TTT, on the other hand, provides a solution to enhance the model’s generalization on specific test data distribution. Particularly, TTT takes a training-time trained model, and then updates the model using unlabeled test data to directly adapt it towards the test distribution. The adapted model is finally used for inference. Therefore, TTT is quite different to training-time training, regardless of whether it has training labels or not.

Q1-2: Then the problem seems to be how to apply slimmable networks in the un-supervised setting, which has been studied in previous works [1].

A1-2: In spite of both using unlabeled data, TTT and unsupervised learning are quite different. The goal of TTT is to improve the supervised learning model’s generalization performance on test data by directly adapting the model towards them. By contrast, unsupervised learning aims to learn robust representation that serves as a decent starting point for downstream tasks. Beyond this fundamental discrepancy, SlimTTT differs from [1] in methods. Firstly, there are distinctions between our contrastive learning approach and [1]. SlimTTT aggregates the multi-views captured by networks of varying widths and employs a multi-positive NCE loss, whereas [1] follows the standard SimCLR-MoCo paradigm to train each sub-networks. Our SlimTTT better facilitates knowledge interaction among networks of different widths. Secondly, we employ LCR and GFA to additionally promote the consistency of these multi-views, enhancing the networks' ability to handle distribution shifts.

To further demonstrate the superiority of TTT compared to unsupervised/self-supervised pretraining in the scenario of adapting to specific test distribution, we consider the following alternative involving [1]. We replace our supervised training-time training on CIFAR10 (C10) by unsupervised learning on both C10 and C10-C (Gaussian Noise (G_N)) using [1] and then finetune on C10. We do not conduct test-time training for this alternative and directly test on C10-C (G_N). Nevertheless, the results reveal that the method proposed in [1] struggles to handle distribution shifts in this process, leading to subpar performance on C10-C (G_N). Additionally, it's worth noting that our SlimTTT is adaptable to an online TTT setting, a capability not present in the approach proposed in [1].

MethodProtocolTraining-time trainingTest-time trainingInferenceWidthErr. (G_N)
[1]Self-supervisedunsupervisedNoneon C10-C1.016.06
on C10 + C10-C0.7515.44
then supervised0.515.69
on C100.2516.91
SlimTTTTest-time trainingsupervisedunsupervisedon C10-C1.010.41
on C10on C10-C0.7511.04
0.511.80
0.2513.96

Q1-3: what are the unique challenges in applying slimmable networks in the test-time training?

A1-3: Simply incorporating slimmable networks into TTT poses several challenges: 1. Different widths of networks may interfere with each other, leading to inconsistent update directions, preventing models from achieving optimal optimization. This issue is more pronounced in TTT due to the lack of labels as well as the distribution shift. Therefore, we propose WCL and LCR to solve this challenge, which promotes the consistency between different width of networks. 2. The GFA is applied to ensure that the test features do not deviate far from the source feature distribution and it can apply a restriction over the multi-view features for each width of networks.

In the following analytical experiment on C10-C, we attempt a straightforward application of slimmable networks in test-time training without employing our consistency method (referred to as naive SlimTTT). Unfortunately, this approach yields unsatisfactory results, particularly when applied in an online manner. This is attributed to the fact that our consistency loss functions—WCL, LCR, and GFA—leverage the diverse views captured by networks of varying widths. This facilitates knowledge interaction among slimmable networks, a crucial aspect in the test-time training setting.

MethodwidthmannerErr. Avg.mannerErr. Avg.
naive SlimTTT1.0offline10.80online14.78
0.7511.5415.64
0.512.5916.72
0.2515.9918.83
SlimTTT1.0offline8.33online10.17
0.758.6310.55
0.59.1211.41
0.2510.6814.01
评论

Q1-2: I still didn't see the fundamental difference between the proposed method and unsupervised slimmable networks such as [1]. The WCL is combining the multi-positive NCE loss in contrastive learning with slimmable network. LRC is the same as the loss in slimmable network (largest network to supervise sub-networks). GFA is also proposed in previous works. Prediction ensemble is also a widely used tricks for improved performance. And the method is the combination of these techniques. I don't think there are many technical contributions. And for comparison, I think the author could apply some unsupervised slimmable networks works to the test-time training stage in their to see the difference.

Another question: Table 2 shows that GFA gives most of the improvements. It is proposed in previous works and not clearly explained how it is applied in the proposed method.

评论

Q1: I still didn't see the fundamental difference between the proposed method and unsupervised slimmable networks such as [1]. The WCL is combining the multi-positive NCE loss in contrastive learning with slimmable network. LRC is the same as the loss in slimmable network (largest network to supervise sub-networks). GFA is also proposed in previous works. Prediction ensemble is also a widely used tricks for improved performance. And the method is the combination of these techniques. I don't think there are many technical contributions. And for comparison, I think the author could apply some unsupervised slimmable networks works to the test-time training stage in their to see the difference.

A1: We will describe the fundamental difference between the proposed method and unsupervised slimmable networks such as [1] from three aspects: 1. the differences between the TTT setting and unsupervised setting, 2. the differences of the techniques proposed in our method and [1], 3. the differences of the motivation of our method and unsupervised method as [1].

  1. The differences of the TTT setting and unsupervised learning setting.

In unsupervised learning, methods first utilize unlabeled training data to train the feature encoder. To perform classification, they subsequently learn a classifier in the supervised manner (usually with the feature encoder fixed) for downstream tasks. In contrast, our approach in the TTT setting trains the feature encoder, the classifier and an auxiliary projection head, (altogether referred to as the Y-structure in TTT) using labeled training data in a supervised manner. Then at test time, we fix the classifier while improving the feature encoder in a self-supervised manner. Such fundamental procedure difference prevents the direct application of unsupervised slimmable network methods in test-time training.

SlimTTT builds upon and improves the Y-structure in TTT. We update the feature encoder through the self-supervised branch, mitigating the impact of distribution shift. This allows the slimmable network to robustly adapt using unlabeled test data with distribution shift. Additionally, we update the feature encoder through the main branch (classifier branch) using LCR while keeping the classifier fixed. This approach prevents the classifier from overfitting to the test data.

Since that the unsupervised slimmable network method cannot be directly applied to TTT setting, methods like [1] have to be modified in the first place. In our initial response, the comparison with [1] illustrated the superiority of TTT over unsupervised learning in terms of generalization performance. Here, we further demonstrate the advantages of our approach over the unsupervised method as presented in [1], when adapted for TTT settings, with a comparative experiment on CIFAR10-C (Gaussian_Noise). The unsupervised techniques in [1] involve contrastive learning and online distillation using KL-Divergence (KLD). From the results we can find that, even with modification, [1] alone cannot address TTT due to the distribution shift. After applying GFA module, the performance of [1] greatly improves, yet our SlimTTT still performs better thus highlighting the overall effectiveness of our approach.

MethodErr. (1.0)Err. (0.75)Err. (0.5)Err. (0.25)
[1]23.4724.0724.4227.53
[1] (w/ GFA)11.7212.4613.2915.86
SlimTTT10.4111.0411.8013.96

Additionally, we conduct ablation experiments to individually analyze the effectiveness of WCL vs. contrastive learning in [1] and LCR vs. KLD in [1]. The results demonstrate that both components of WCL and LCR outperform [1]. Further analysis of the specific advantages of each component will be presented in the next section.

MethodErr. (1.0)Err. (0.75)Err. (0.5)Err. (0.25)
[1] (w/ GFA, w/o KLD)11.9612.2113.8816.65
SlimTTT (w/o LCR)10.9611.7513.3416.31
[1] (w/ GFA, w/o contrastive)12.5813.4114.7716.11
SlimTTT (w/o WCL)11.0311.3812.4715.23
评论
  1. The differences between the techniques proposed in our method and in related methods.

Firstly, the WCL is different from the existing contrastive learning methods that are leveraged to training a slimmable network. It is a modification more suitable for slimmable network in TTT setting. First of all, WCL does not require a queue, whereas the traditional contrastive loss like SimCLR-MoCo used in [1] relies on a queue, incurring additional overhead, making it unsuitable for the TTT setting. Moreover, in the construction of positive samples, our WCL differs from the approach used in [1]. Based on our discovery that different networks can capture different views, we aim for a more diverse interaction of views during contrastive learning. In contrast to [1], where one of the augmented images are fed into the supernet while another augmented images are directed to networks of different widths, all our networks can observe both two of distinct data augmentations to construct the positive samples. By jointly combining them for contrastive learning, we enhance the diversity of views, fostering knowledge interaction among networks of different widths. Finally, in the results of previous contrastive learning based slimmable networks, the supernet in the slimmable network may be influenced by the smaller networks, resulting in its performance degradation compared to individual training. By contrast, WCL enables smaller networks to complement the knowledge of larger networks, ultimately promoting the improvement of the performance of the larger networks.

Secondly, the LCR is different from the loss in slimmable network. Notably, we are the first to use hard labels to replace the soft labels commonly used in related methods like [1] for supervision in slimmable network. We generate pseudo-labels by passing weakly augmented images through the largest network, and supervise all width of networks (including supernet itself) on the predictions obtained from strongly augmented images. This is because, in test-time training, only noisy pseudo-labels can be utilized to regularize prediction consistency between supernet and sub-nets, which is in contrast to the previous works that possess ground-truth labels. As a consequence, the entropy minimization effect brought by hard labels becomes more favored than the dark knowledge contained in the soft ones. Through comparison of our SlimTTT using soft labels and hard labels below on CIFAR10-C, we empirically demonstrate the effectiveness of our LCR using hard labels.

MethodR-50 (1.0×)R-50 (0.75×)R-50 (0.5×)R-50 (0.25×)
Soft Labels8.789.259.9811.91
Hard Labels8.338.639.1210.68

Thirdly, although GFA is a method adopted in previous works, we extended it to multiple width networks. Due to the effectiveness of our consistency constraints, we do not need to perform alignment at the classwise or self-supervised head levels, as done in previous methods. This improves the efficiency of our approach, making it simple yet effective. In addition, ensemble techniques are not universally applicable in all cases, as demonstrated in [3]. Therefore, we do not treat ensemble as a widely used trick. Inspired by [3], we theoretically and visually prove that networks of different widths have the ability to capture multi-view features of images. This finding ensures the effectiveness of using ensemble techniques in our approach.

  1. The differences of the motivation of our method and unsupervised method as [1].

Our proposed SlimTTT is not a simple application of Slimmable network on Test-time training. Instead, it is well-motivated to improve the practicability of TTT, with all its modules carefully designed to adapt for the setting. First, through an analysis of the TTT application scenarios, we identified its limitations when resources are constrained or dynamically changing. Additionally, we theoretically and visually demonstrated that sub-networks of different widths in slimmable networks can capture distinct views. These views naturally adapt during the TTT phase through contrastive learning. With the assistance of our designed techniques, WCL, LCR, and GFA, the model learns more robust features in both online and offline TTT processes. The coexistence of multi-width networks effectively addresses challenges faced by previous TTT methods in resource-constrained and dynamically changing environments, expanding the applicability of TTT.

评论

Q2: Table 2 shows that GFA gives most of the improvements. It is proposed in previous works and not clearly explained how it is applied in the proposed method.

A2: GFA is a commonly used technique in prior TTT methods, however, they only apply it on a single network. We extended it to multiple width networks by aligning the global test data distribution of different width to the source one of the same width using KL-Divergence. The GFA in our method is applied to ensure that the test features do not deviate far from the source feature distribution during the whole adaptation process and it can apply a restriction over the multi-view features for each width of sub-networks and ensure the correctness of the multi-view representations. The reason we align the test distribution of each width with the corresponding source distribution, rather than imposing consistency constraints, is because we want the networks to retain their unique multi-view characteristics without excessively converging. This ensures the effectiveness of our WCL knowledge interaction and ensemble technique. We conduct the following experiment on CIFAR10-C (Guassian_Noise) to compare SlimTTT using our GFA with the GFA applied on a single network (GFA (single)) and imposing GFA consistency on self-supervised head (GFA (ssh)). The results show that our GFA can outperform the other forms of GFA, which fully leveraging the characteristics of each width of networks.

MethodR-50 (1.0×)R-50 (0.75×)R-50 (0.5×)R-50 (0.25×)
GFA (single)12.6313.3913.5815.55
GFA (ssh)11.7912.3913.0315.04
GFA (ours)10.4111.0411.8013.96
评论

Q2: The motivation that different sub-networks could capture different image features has been studied in [2].

A2: Thank you for bringing this to our attention. MutualNet [2] observes that distinct model-resolution configurations can capture varying image features. However, it's essential to note that the observations and the underlying motivation differ between MutualNet and our approach.

MutualNet claims that due to the shared parameters, the sub-networks can learn multi-scale representations, in terms of pairing images of different resolutions with networks of various widths. They supervise the sub-networks by the supernet prediction with largest image resolution. Thus, the knowledge transfer direction is only from supernet to the sub-networks. Unlike MutualNet, our SlimTTT proposes consistency loss function WCL to explicitly combine supernet and other sub-networks to do contrastive learning in feature space, which enables explicit knowledge transfer for both direction.

Moreover, MutualNet identifies that distinct model-resolution configurations, such as 1.0×-224 and 0.75×-128, can emphasize different image features. Specifically, larger networks with higher resolutions tend to focus more on details, while smaller networks with lower resolutions prioritize global structures. It's noteworthy that the image features captured by various networks are influenced by the resolution of the image itself, while the multi-view in our SlimTTT rely solely on the width of sub-networks.

Furthermore, the concept of multi-view captured by our SlimTTT differs from the image features identified by MutualNet. Drawing on the principles outlined in [3], we demonstrate that sub-networks of varying widths can capture distinct discriminative multi-views of an object, which is supported by visualization experiments and a theoretical toy example. For instance, certain widths of sub-networks might focus on the head of a dog, while others capture the tail of the same dog. These discriminative features are entirely distinct and can both contribute effectively to the classification of a dog. These multi-views are different from the detailed or global features proposed by MutualNet, which hinge on the resolution of the image.

Finally, we conducted the following experiment to demonstrate that the distinct image features captured by MutualNet using different image resolutions are orthogonal to the multi-views we identified. We treat images of different resolutions as diverse forms of data augmentation. In contrast to MutualNet randomly inputting images of a certain resolution into a network of a particular width, each width of network can observe images of all resolutions once a time in our method. This approach increases the views we claimed in our paper, and through our designed WCL and LCR, we promote consistency between these views. As shown in the table below, we achieve better results on ImageNet-C in online manner after adding images of different resolutions. This indicates that MutualNet and our SlimTTT are orthogonal. It's important to note that these results are preliminary, and we plan to delve deeper into the benefits that images of different resolutions can bring to our method in the future.

MethodwidthGaussian_NoiseSnowBrightness
SlimTTT1.061.3742.0531.44
0.7562.5543.7833.03
0.564.9347.0736.17
0.2570.9554.8942.89
SlimTTT + resolution1.061.1741.7431.50
0.7563.2743.0332.77
0.564.5249.0335.83
0.2570.0554.3742.77

Therefore, both the concept of multi-views and the rationale for their inclusion in our SlimTTT differ from the approach taken by MutualNet.

[3] Towards Understanding Ensemble, Knowledge Distillation and Self-Distillation in Deep Learning.

Q3: In Table 1, the other methods should also use ResNet-50 with different widths to have a fair comparison.

A3: Thank you for your suggestions. We conducted a comparison of our method with other TTT methods using ResNet-50 with varying widths on Cifar10-C and Cifar100-C, and the results are presented in the table below. The findings demonstrate that our method consistently outperforms other TTT methods, attributed to the unique loss we designed to guarantee the consistency of networks with different widths.

MethodwidthCifar10-C Err. Avg.Cifar100-C Err. Avg.
TTT++1.09.6832.71
TTAC1.09.2431.60
SlimTTT (ours)1.08.3329.36
TTT++0.7510.6234.72
TTAC0.7510.0833.90
SlimTTT (ours)0.758.6330.34
TTT++0.511.4736.42
TTAC0.511.3535.73
SlimTTT (ours)0.59.1232.07
TTT++0.2512.8840.32
TTAC0.2513.4640.06
SlimTTT (ours)0.2510.6837.71
评论

Q4: This work finetuned the pre-trained slimmable networks on ImageNet and ImageNet-C, what about other works?

A4: Thank you for pointing it out. In our study, we initially conduct standard training-time training procedure for slimmable networks on training dataset ImageNet. Subsequently, we update/finetune the pre-trained slimmable network on test dataset ImageNet-C (without labels) without accessing ImageNet anymore. The full procedure follows the previous works like TTT++ and TTAC. It's important to note that all the methods compared in our paper follow this standardized process to ensure a fair and equitable comparison. We will rephrase this process in our new version of paper to avoid confusions.

Q5: In Table 5, why is the training cost is comparable or even faster than TTT++ and TTAC? The proposed method needs to forward and backward multiple times, I am assuming it should be more training expensive.

A5: Good comment. In TTT++ and TTAC, maintaining a queue is necessary for projection alignment to update the self-supervised head (SSH). Additionally, TTAC involves classwise alignment and filtering. Although our SlimTTT needs to forward and backward multiple times, our method doesn’t need to use class alignment or ssh alignment like TTT++ and TTAC, thus the total training cost of our SlimTTT can be comparable with TTT++ and TTAC.

To fairly compare the training cost of our method with TTT++ and TTAC, we conduct experiment on CIFAR10-C (batchsize=8) in NVIDIA Jetson TX2 to veirfy the training cost of different module. The results are shown in the table below, the total training costs of our SlimTTT are comparable with TTT++ and TTAC, which means our SlimTTT is simple but effective.

ssl_conpre_regclass_alignext_alignssh_alignoptimizetotalErr. Avg.
TTT++ (ResNet50)0.7906s000.4964s0.7467s0.0820s2.1157s9.80
TTAC (ResNet50)0.6773s00.6042s0.3993s0.6173s0.0938s2.3919s8.52
SlimTTT (ResNet50)1.1958s0.1461s00.6794s00.1207s2.1420s8.33
评论

Dear AC and Reviewers,

We extend our heartfelt appreciation to to all reviewers for their insightful questions, valuable comments, and constructive suggestions. We are encouraged by the recognition of our efficient test-time training setting (noted by 1QhS, aCQv) and clear explanation (highlighted by by2E, aCQv). Our paper provide new insight of the multi-view findings (noted by yJHx) and achieve good performance on several benchmarks (noted by all reviewers).

In response to the thoughtful feedback and suggestions from the reviewers, we have delved deeper into our investigation, making significant enhancements. We have incorporated the following discussions and experiments:

  • The comparison with un-superivsed/self-supervised setting (A1 for by2E).
  • The comparison with MutualNet about multi-view (A1 for by2E).
  • Considering experiments where the available compute changes frequently during test-time training (A1 for 1QhS).
  • Explaining the motivation for SlimTTT (A1 for yJHx).
  • Applying SlimTTT to online test-time training setting (A3 for yJHx).
  • Explaining the “Efficient” and basic assumptions of our SlimTTT (A1 and A2 for aCQv).

Enthusiastically inspired by the insightful remarks and suggestions from the reviewers, we have uncovered further insights into the ability of online dynamic adaptation during test-time training phase of our SlimTTT approach. We hope to have been able to address all of the reviewers' questions and concerns and remain available to provide answers for any follow-up inquiries.

Warm regards,

Authors

AC 元评审

This paper proposes a method for test-time training (TTT) which is the offline fine-training of a source model using unlabeled data from a new domain, followed by inference on actual test inputs from the new domain.

The proposed method is an extension of the slimmable networks idea adapted for the TTT setting which is unsupervised. Albeit not in the TTT setting, unsupervised extensions of slimmable networks have been proposed in prior work as well, as pointed out by one of the reviewers.

The reviewers also expressed concerns about the idea being basically a combination of unsupervised source-free domain adaptation and slimmable networks. There are also concerns regarding efficiency of SlimTTT as compared to methods like TENT. The reviewers were not entirely convinced why slimmable networks would be appropriate.

The authors' response was considered and discussed but the reviewers' concerns remained. Therefore, in the end, it was felt that the paper in its current form is not ready to be published. The authors are advised to address the concerns raised by the reviewers and incorporate the suggested experiments, and resubmit the work to another venue.

为何不给更高分

Several issues pointed out by the reviewers regarding the novelty and the motivation behind using slimmable networks

为何不给更低分

N/A

最终决定

Reject