Mixture of Adversarial LoRAs: Boosting Robust Generalization in Meta-Tuning
A new adversarial meta-tuning method for boosting the performance of pre-trained models across domains in few-shot image classification
摘要
评审与讨论
The paper introduces Adversarial Meta-Tuning (AMT) to enhance the robust generalization of pre-trained models for out-of-domain few-shot learning by constructing a robust LoRAPool through meta-tuning Low-rank Adapters (LoRAs). The approach significantly outperforms previous methods in both clean and adversarial generalization across various benchmarks.
优点
-
this study is novel and significant.
-
the proposed method AMT achieves superior performance.
-
The writing and presentation are good and easy to follow.
缺点
-
Lack of evaluation on unseen adversarial attacks. The authors claimed that they boosted the adversarial generalization. Test the generalization of robustness on unseen attacks would be necessary.
-
Evaluation of adversarial robustness based on PGD-10 is insufficient. [2] points out that the PGD-10 attack is a weak attack and may suffer from a 'gradient mask' problem. It is necessary to present robustness based on AA attack[1].
[1] Croce, Francesco, and Matthias Hein. "Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks." International conference on machine learning. PMLR, 2020.
[2] Athalye, Anish, Nicholas Carlini, and David Wagner. "Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples." International conference on machine learning. PMLR, 2018.
问题
- Could you show whether the robustness achieved by AMT can be generalized to unseen attacks?
-Could you show the robustness under AA attack?
- How does the performance change with varying the rank R in each Lora adaptor? and with varying the number of Lora adaptors?
局限性
The authors adequately addressed the limitations.
We extend our appreciation for your constructive feedback on our manuscript. Below, we address each of your points comprehensively. Should there be any additional queries or clarifications needed, please feel free to let us know.
Q1. Adversarial robustness evaluation of unseen attacks
- We appreciate the reviewer's invaluable suggestion.
- Our claim that "AMT boosts the adversarial generalization" is mainly grounded in our results in Table 3 where adversarial robustness is evaluated under distribution shifts using the same meta-tuned LoRAPool on the source domain ImageNet.
- We conducted additional experiments to measure the adversarial robustness generalization to unseen threat models, including - and unseen -bounded attacks with different perturbation budgets . For each dataset, we sample 600 -way -shot tasks and generate adversarial examples using 10 steps of PGD with the step size for and for attacks, respectively. The results, shown in Table 10 of the rebuttal PDF, demonstrate our method AMT significantly enhance adversarial robustness against unseen attacks under distribution shifts for pre-trained vision transformer. Also, compared to previous adversarial few-shot learning methods StyleAdv, our AMT does not sacrifice in-domain generalization.
Q2. Adversarial robustness evaluation of AutoAttack under distribution shifts
- We appreciate this great suggestion, and in response, we have incorporated additional experiments to measure adversarial robustness against AutoAttack [1] under distribution shifts. Specifically, we ground our method and the baseline on the adversarially pre-trained ViT-Small [2] and use APGD with cross-entropy and targeted DLR loss, FAB-attack and the Square Attack to generate adversarial examples on the 100 sampled -way -shot tasks on each dataset. The -bounded perturbation at the radius is adopted. The results, shown in the below table, demonstrate our method AMT consistently boosts adversarial generalization even under stronger AutoAttack in terms of both in-domain and out-of-domain robust accuracy.
Method ImageNet Omniglot Aircraft CUB DTD QuickDraw Fungi Flower Traffic Sign MSCOCO Avg. PM 29.36 36.52 3.88 15.06 14.20 29.80 3.61 20.48 10.26 8.26 17.14 AMT 39.96 61.48 8.88 24.04 23.12 51.48 11.09 44.76 23.20 22.00 31.00 [1] Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In ICML, 2020
[2] Revisiting Adversarial Training for ImageNet: Architectures, Training and Generalization across Threat Models. In NeurIPS, 2023.
Q3: Hyper-parameter Study of the LoRA rank and the pool size
- We conduct the ablation analysis following the reviewer's valuable suggestion to investigate the impact of the rank of LoRA and the size of LoRAPool. We also report the mean and variance of perturbation budget candidates during adversarial meta-tuning. The results, as illustrated in Tables 5 and 7 of the rebuttal PDF, indicate that (1) our model is not very sensitive to the rank of LoRA; (2) a sufficiently diverse but the largest pool leads to improvements in performance.
This paper proposes to tackle the problem of improving the generalization of pre-trained models to data drawn from a different distribution than the training data. To achieve this, the authors propose using meta-learning to train the models. Since they consider a single source domain, they adversarially generate query samples while the support samples directly come from the source domain data. The inner gradient is computed on the source, while the meta-gradients are computed on adversarial examples.
They ground their approach by citing references that suggest an adversarially robust model also generalizes better on OOD data. Unlike conventional meta-learning approaches, which update the entire model parameters, the authors propose updating models through multiple LoRA steps, similar to FLUTE, which only updates the batch norm parameters. Each LoRA is computed using a query set of adversarial examples generated through a fixed attack budget. At test time, they identify the right combination of LoRAs based on simple cosine similarity measures between features of prototypes for each class (computed during training) and the OOD sample and its corresponding labels. They propose to look at intra-class similarity and inter-class diversity to this end.
Furthermore, instead of perturbing only the images, they also perturb the singular values/vectors of the model weights (gradients) to strengthen the principal components (based on the observation that singular vectors undergo significant change during training). They evaluated their method on multiple datasets and demonstrated good performance gains compared to other methods such as StyleAdv, which alters the style of query samples using AdaIn instead of adversarial attacks.
They also conducted ablation studies and demonstrated the benefits of various design choices.
优点
The paper is very well motivated, and the authors ground their approach appropriately, citing a large collection of work. Each piece is well-motivated and clearly written. Perturbing singular vectors when performing LoRA is a novel idea, and the results indicate the benefits of this approach.
缺点
A primary weakness of the approach is its lack of comparison to another popular method for fine-tuning adaptation of foundation models—adapters. Adapters introduce new learnable parameters in each transformer block and update only them using the new data while freezing everything else. Some references include MiMi and AdaptFormer. There are many more references within these papers and in the scholar link. The current paper does not consider this PEFT technique at all. It would be beneficial to see a comparison or discussion section with adapters. Importantly, adapters are closely related to the approach in FLUTE too and instead of a LoRA pool, one could have an adapter pool.
Secondly, the adversarial attack for images is the standard attack model. If all that is needed is better and harder augmentations, papers like Improving Diversity with Adversarially Learned Transformations for Domain Generalization offer a better alternative and they also work with a single source domain. Other options to consider include augmentation techniques such as PixMix and Rand Conv, where the tradeoff between ID and OOD accuracies is well established, and they don’t require additional steps such as adversarial attacks.
Additionally, as an improvement, from the OOD detection literature this paper could use virtual outlier synthesis (VOS). VOS uses a per-class GMM in the latent space to synthesize feature space outliers. These outliers can then be used in training instead of relying solely on adversarial attacks.
问题
Please see weaknesses section
局限性
Provided limitations are appropriate.
We appreciate your constructive comments on our paper. Please kindly find our response to your comments below. We hope that our response satisfactorily addresses the issues you raised. Please feel free to let us know if you have any additional concerns or questions.
Q1. More discussions concerning other PEFT techniques during meta-tuning
- We sincerely thank the reviewer for the constructive comments and will incorporate the following discussion into our revised manuscript.
- We have conducted additional experiments to adversarially meta-tune full parameters [3], FiLM [1] (after LN layers since there are no BN layers in ViT), Adapter [2], and LoRA. The attack budget is randomly sampled for each training task from the candidate pool which is the same as AMT. The results, shown in the table below, demonstrate that the performance of Adapter and LoRA is comparable in the context of adversarial meta-tuning and outperforms full or FiLM-based meta-tuning.
- We would like to highlight that compared with FLUTE which combines multiple FiLMs with a parametric classifier as the initialization for FiLM-based test-time fine-tuning, our LoRAPool with non-parametric merging mechanism adaptively integrates the LoRAPool into pre-trained weights, and thus provides better flexibility and compatibility with advanced test-time fine-tuning techniques to further improve the few-shot learning performances. For example, Tabel 1 in the work [4] has demonstrated that both LoRA- and Adapter-based test-time fine-tuning are not optimal choices for vision transformers.
- We are still pending for the results of FLUTE-style test-time fine-tuning, which we will supplement as soon as possible during the discussion period.
Adversarial Meta-tuning Test-time merge Test-time fine-tuning ImageNet Omniglot Aircraft CUB DTD QuickDraw Fungi Flower Traffic Sign MSCOCO Avg. Full - - 64.31 62.81 38.46 76.23 60.42 57.99 56.31 81.80 57.31 54.22 60.98 Single FiLM - - 63.23 63.41 37.67 74.41 59.29 57.60 55.23 80.05 58.86 54.57 60.43 Single Adapter - - 64.68 65.32 38.43 75.37 59.68 58.35 55.90 81.69 58.31 54.05 61.18 Single LoRA - - 63.91 65.05 39.44 76.95 58.46 58.35 56.39 82.29 59.56 53.69 61.41 FiLM Pool classifier FiLM Adapter Pool classifier LoRA LoRAPool classifier - 67.22 64.60 37.99 77.96 62.65 57.11 56.62 80.23 58.36 56.10 61.89 LoRAPool criteria - 68.80 71.95 42.90 79.95 62.99 59.62 59.06 85.37 63.78 57.14 65.16 LoRAPool criteria LoRA LoRAPool criteria PMF [3] 68.80 77.83 42.90 79.95 63.77 63.72 59.06 85.37 63.87 57.37 66.26 LoRAPool criteria ATTNSCALE [4] 68.80 79.43 42.90 79.95 63.08 65.66 59.06 85.37 64.13 58.24 66.66 [1] Learning a Universal Template for Few-shot Dataset Generalization. In ICML, 2021
[2] Adaptformer: Adapting vision transformers for scalable visual recognition. In NeurIPS, 2022
[3] Pushing the limits of simple pipelines for few-shot learning: External data and fine-tuning make a difference. In CVPR, 2022
[4] Strong baselines for parameter-efficient few-shot fine-tuning. In AAAI, 2024
Q2. More discussions concerning other data augmentation techniques
- Thank you for your insightful suggestion. We have conducted additional comparisons for AMT against other data augmentation methods, including ALT [1] and Rand Conv [2], as suggested.
- Our experiments were conducted under fair conditions using a single LoRA (pool size ) during meta-tuning for all methods. Specifically, following ALT [1], we employed a learnable adversarial transformation network consisting of 5 convolutional layers with a kernel size of 3 and LeakyReLU activation. The adversarial learning rate was set to , with 10 adversarial steps. For the method employing an attack candidate pool, we randomly select the attack budget from candidates for each training task, with values of {8/255, 6/255, 0.1/255, 0.01/255} for our method, and step number of {1, 3, 5, 10} for ALT.
- The results in Table 9 of the rebuttal PDF demonstrate that static data augmentation cannot effectively simulate the large domain shift required for robust generalization across diverse datasets (e.g., Omniglot). Our AMT with standard pixel-level adversarial attack achieves comparable or superior generalization improvements for pre-trained vision transformers across OOD tasks.
Q3. Other data augmentation techniques deserve future work
We sincerely appreciate the reviewer's invaluable suggestion. We would like to highlight that we present a framework by constructing the robust LoRAPool with test-time merging to significantly boost the robust generalization of the pre-trained vision transformer. In this context, we use adversarial attacks, characterized by the size of the perturbation budget, as an example to mimic different distributional shifts to construct different LoRAs. Our experiments demonstrate the effectiveness of using adversarial training. The focus and contribution of this paper is the whole framework of the algorithm instead of the optimal data augmentations. In addition, we believe the data augmentation to achieve the optimal performance also depends on the OOD test set we use. Nevertheless, we agree data augmentation technique is important for this problem and leave investigating better training data augment techniques (e.g., Virtual Outlier Synthesis (VOS)) as future works.
I thank the authors for addressing my concerns. I wanted to know if the experiments have finished with adapter pool? I commend the authors for running experiments in such a short time. I would also recommend to discuss other PEFTs in the main paper. Given that most of my concerns are addressed, I am going to increase the score. Please do comment when you have adapter pool results too.
- We sincerely thank the reviewer for your prompt response and raising the score.
- We will definitely follow the reviewer's suggestion to include discussions about other PEFTs in our related work.
- Besides, please find the FiLM/adapter pool results below, which we will also include in our revision.
- Regarding the FiLM pool [1] and Adapter pool [2], we have conducted additional experiments by setting the pool size to 4 and adopting the same attack candidate pool used in AMT during adversarial meta-tuning. To estimate the combination coefficients, we follow the method outlined in FLUTE [1]. Specifically, a classifier is trained in a separate stage to predict which FiLM or Adapter the input belongs to, taking as input a batch of adversarial examples generated by attacking different FiLMs or Adapters in the pool.
- In the following table:
- The superiority of FiLM/Adapter Pool over FiLM/Adapter signifies that our adversarial pool design indeed contributes to the out-of-distribution performance without the compromise of in-domain accuracy.
- Ours with the additional (1) perturbation in singular values/vectors and (2) non-parametric test-time merging mechanism utilizing the criteria (i.e., Line 204-208) enjoys significant performance improvement over FiLM/Adapter Pool.
- Compared with the FLUTE-style test-time fine-tuning strategy that requires further tuning of pool components (either a FiLM or an adapter), our framework shows better compatibility with different test-time fine-tuning approaches, including LoRA tuning, full fine-tuning [3], and attention scaling [4].
Adversarial Meta-tuning Test-time merge Test-time fine-tuning ImageNet Omniglot Aircraft CUB DTD QuickDraw Fungi Flower Traffic Sign MSCOCO Avg. Full - - 64.31 62.81 38.46 76.23 60.42 57.99 56.31 81.80 57.31 54.22 60.98 Single FiLM [1] - - 63.23 63.41 37.67 74.41 59.29 57.60 55.23 80.05 58.86 54.57 60.43 Single Adapter [2] - - 64.68 65.32 38.43 75.37 59.68 58.35 55.90 81.69 58.31 54.05 61.18 Single LoRA - - 63.91 65.05 39.44 76.95 58.46 58.35 56.39 82.29 59.56 53.69 61.41 FiLM Pool classifier FiLM 67.45 65.42 37.58 75.02 62.63 59.22 55.09 79.00 60.40 55.69 61.75 Adapter Pool classifier Adapter 67.48 65.33 38.58 80.16 62.76 58.09 57.63 75.23 57.41 54.32 61.70 LoRAPool criteria - 68.80 71.95 42.90 79.95 62.99 59.62 59.06 85.37 63.78 57.14 65.16 LoRAPool criteria LoRA 68.80 80.00 43.49 79.95 62.99 59.62 59.06 85.37 66.42 57.14 66.28 LoRAPool criteria PMF [3] 68.80 77.83 42.90 79.95 63.77 63.72 59.06 85.37 63.87 57.37 66.26 LoRAPool criteria ATTNSCALE [4] 68.80 79.43 42.90 79.95 63.08 65.66 59.06 85.37 64.13 58.24 66.66 [1] Learning a Universal Template for Few-shot Dataset Generalization. In ICML, 2021
[2] Adaptformer: Adapting vision transformers for scalable visual recognition. In NeurIPS, 2022
[3] Pushing the limits of simple pipelines for few-shot learning: External data and fine-tuning make a difference. In CVPR, 2022
[4] Strong baselines for parameter-efficient few-shot fine-tuning. In AAAI, 2024
This paper deals with how to effective adapt a pretrained model to cross-domain few-shot learning task. It focus on both adversarial robustness and clean accuracy of trained model. To realize the goal, it utilizes adversarially trained LoRA to adapt the pretrained model. Specifically, it utilizes SAM to determine the worst-case perturbations to update the matrices A and B of LoRA which are initialized to approximate the modification of principle singular value and vectors of original weight matrix in pretrained model. Further, it designs LoRAPool. LoRAPool consists a series of LoRA modules, each corresponding to a different robustness level controlled by the size of the adversarial budget. Based on that, the authors design a test-time merging strategy to adaptively combine all the LoRAs for optimizing test-time task performance. Extensive experiments verify the effectiveness of proposed method.
优点
-
This paper is generally easy to follow and understand.
-
The idea is reasonable.
-
The experiment results is good.
缺点
-
It is unclear that what key factors make the proposed framework outperforms previous works. From technical level, the operations or strategies used in the paper is not that new, e.g., using LoRA to finetune the pretrained model, using SAM to improve the generalization ability of the model, test-time ensemble, etc. I just wonder what makes the proposed framework most distinct from previous works.The authors should more clearly discuss this to give the readers some insights.
-
Some technical details are not clear. For example, in Eq. (7), what does the top_k means? Or to say, what is the operation detail of top_k? In experiment part, the authors compare their method with previous works under two settings, i.e., tuning-free setting and test-time fine-tuning setting. What are the difference between those two settings? In Line 259-260, what is the exact form of the variant which removes the adversarial perturbations on singular values and vectors? Does it mean not using SAM to update A and B of LoRA?
-
The authors compare the adversarial robustness and the clean accuracy separately in two tables. I just wonder if those two numbers are achieved simultaneously by the same model, or they are evaluated with two model where one tends to be robust and the other one tends to be more accurate (it can be so by employing different ). I would like to see how the proposed method balances those two evaluations.
-
Missing key ablations. The authors design a special initialization to make the update of A and B approximately represents the modifications of singular values and vectors of original weight matrix. I just wonder how the performance is if we don't employ such an initialization, i.e., we just simply add LoRA to the pretrained model which is initialized as zero matrix. I would like to see the comparison between plain LoRA initialization and the proposed initialization. Otherwise, the effectiveness of proposed initialization cannot be proved.
-
Lack hyper-parameter study.
问题
See the weakness part.
局限性
The authors adequately addressed the limitations.
Thank you sincerely for your thoughtful feedback on our work. Below, we have provided a detailed explanation for your concerns as follows. Please do not hesitate to let us know if you have any further questions.
Q1. Technical contributions concerning adversarial singular value and vector perturbation and robust LoRAPool
- We acknowledge the reviewer's scrutiny of our technical contributions and would like to clarify that our work is definitely not a simple summation of existing techniques. Different components of our methods are closely related and adapted to our goal: to boost the robust generalization of pre-trained models in out-of-domain few-shot learning.
- Specifically, inspired by the low-rank strategies when training transformers, we do not directly perturb the weight matrices as done in SAM, instead our perturbations are in the spectral space: we perturb the singular values and the singular vectors to boost the performance.
- In addition, we use different adversarial perturbations (either different types or different magnitudes) to generate a LoRA pool consisting of multiple low-rank structures and use adaptive merging to construct the best-adapted parameters for different out-of-distribution tasks during the test time. We acknowledge that both SAM and LoRA motivate our solution, but we need to clarify that our algorithm is significantly different from vanilla SAM or LoRA and much better adapted to the problem we aim to solve. Below are additional comparisons to validate the effectiveness of our novel solutions.
- The results, reported in Table 2 of the rebuttal PDF, demonstrate that our method outperforms SAM by 1.56% on average.
- As shown in Table 1 of the rebuttal PDF, the robust LoRAPool with perturbation-specific parameters effectively avoids interference between attacks and significantly enhances the OOD generalization without ID compromise. For uniform strategy, we adopt the average attack strength () of candidate configurations and meta-tune 4 LoRAs with different seeds. The random strategy means we randomly sample one attack budget for each training task from the same attack candidate configurations.
- We also humbly highlight that our contribution of singular value trimming and non-parametric test-time merging mechanism is also novel and effective for few-shot learning, as supported by the results in Tables 4 and 5 of the main paper.
Q2. Clarification of the top_k operation
We apologize for any potential misunderstandings. The operation top_k before softmax in Eq. (7) refers to selecting the top LoRA modules with the largest score of and the rest LoRAs are deactivated for the current task.
Q3. Clarification of the tuning-free setting and test-time fine-tuning setting
We apologize for any potential misunderstandings.
- The tuning-free setting does not involve additional training on the support set. We adaptively merge meta-tuned LoRA into pre-trained models via the formula in line 206 and perform prototype-based classification. Aside from this, the test-time fine-tuning setting allows for training on the support set according to different fine-tuning methods such as fine-tuning full parameters (PMF) or partial parameters (ATTNSCALE)
- We evaluate under both settings to demonstrate AMT's (1) effectiveness in learning a well-generalized initialization for pre-trained models even without time-consuming fine-tuning, and (2) compatibility with advanced fine-tuning techniques to further improve the few-shot learning performances.
Q4. Clarification of variant removing the adversarial perturbations on singular values and vectors
We appreciate the reviewer's feedback. When removing the adversarial perturbations on singular values and vectors, we inject the worst-case perturbations into the input only, without perturbing the singular value and vectors of LoRA. Other configurations remain the same as the final version of AMT. We would like to humbly clarify that this adjustment does not imply not using SAM.
Q5. Clarification of evaluated model
We appreciate the reviewer's invaluable feedback and we clarify that all results presented in Table 1, 2 and 3 are achieved simultaneously by the same model using a meta-tuned robust LoRAPool on the source domain ImageNet.
- We would like to highlight that thanks to the diverse design of the adversarial LoRAPool, our AMT improves the trade-offs between adversarial robustness and clean accuracy, as well as between ID and OOD generalization, as supported by the results compared to the clean meta-tuning method (PMF) and adversarial few-shot learning method (StyleAdv) in Table 1, 2 and 3.
- We conducted additional experiments, employing different . The results, in Table 3 of rebuttal PDF, demonstrate that we can use to adjust the preference of our robust LoRAPool to either clean or adversarial environments.
Q6. Ablation studies to support the improvement of perturbation on singular values and vectors
We appreciate the reviewer's insightful suggestion and conducted additional comparisons, pitting AMT against original LoRA initialization, for which we incorporated adversarial perturbations in the weight space. The results in Table 4 of rebuttal PDF demonstrate that AMT achieves superior performance, highlighting the effectiveness of our adversarial singular value and vector perturbation in boosting the model's generalization capability.
Q7. Hyper-parameter study
- We have conducted supplementary experiments. The results, as illustrated in Tables 5, 6, 7, and 8 of the rebuttal PDF, indicate that (1) our model is not very sensitive to the rank of LoRA and the number of attack steps; (2) a sufficiently diverse but large pool leads to improvements in performance. The results also justify our choice of top-2.
Thanks for the authors' detailed response. The response clarify most of my concerns. I would like to raise my score. Hope the authors can include all the necessary details and results and polish their paper to make it clearer.
The paper proposes a method for training loras for vision transformer models such that the model easily adapts to an unseen few shot classification task. The goal is to have these loras robust to adversarial noise.
优点
The experiments seem comprehensive and show impressive performance. The idea of perturbing eigenvectors of weights instead of weights seems interesting to me. Further, the idea of training and merging multiple loras that are robust to different levels of adversarial noise is interesting.
缺点
- What are the instances in medical and self-driving domain "where encountering novel and adversarial environments is common" (line 26)?
- The writing can be improved a lot. a. I struggling to understand how the start of the second sentence of the introduction (line 15) is connected to the previous sentence. b. Same with the starting two sentences of the next paragraph (line 21). I don't see why "However" is necessary on line 23. c. Third, fourth, and fifth sentences of the third paragraph (lines 29, 31, and 33) are not connected at all. d. What is "double strongest perturbation"? (line 45) e. This is a prime example of a sentence that needs to be simplified: "To robustify the learned meta-knowledge, adversarial meta-tuning adopts the worst-case optimization by injecting the adversarial perturbation δ to the input x through the minimax strategy" (lines 130-132) I consider myself a decently well educated researcher and the sentence doesn't tell me anything.
The paper seemed promising to me, but I had to parse out a lot of stuff to get to the interesting part. Hence, I couldn't spend time properly understanding the method and results. For now, I am keeping accept as the result seem promising.
问题
See above.
局限性
The limitations section in the paper seems adequate.
We sincerely thank the reviewer for providing valuable feedback. We detail our response below point by point. Please kindly let us know whether you have any further concerns.
Q1. Instances of novel and adversarial environments
We appreciate the reviewer's great feedback and agree that discussion of instances of adversarial environments in real-life applications will strengthen the motivation of this work. We will incorporate the discussions and the related literature into our revised manuscript.
- Instances of novel environments: In real-world deployments, deep learning models in both medical and self-driving domains often encounter novel environments and suffer from distribution shifts between training and deployment data, including unseen pathologies [1], variations in hospital equipment and protocols [1], and diverse urban road scenarios [2].
- Instances of vulnerability to adversarial attacks: Furthermore, deep neural networks are vulnerable to physical-world adversarial attacks, which can lead to harmful diagnoses or unsafe driving decisions. For instance, adversaries can perturb sensor signals to deceive 2D or 3D medical imaging models [3, 4], manipulate traffic signs with malicious stickers to mislead autopilot systems [5], or fool the autopilot into following fake lane lines[6] or making unsafe trajectories [7].
[1] Understanding silent failures in medical image classification. In MICAAI, 2023.
[2] Are we ready for autonomous driving? the kitti vision benchmark suite. In CVPR, 2012.
[3] Self-adaptive adversarial training for robust medical segmentation. In MICAAI, 2023.
[4] Adversarial attacks and defenses on AI in medical imaging informatics: A survey. In Expert Systems with Applications, 2022.
[5] Robust Physical-World Attacks on Deep Learning Visual Classification. In CVPR, 2018.
[6] Dirty road can attack: Security of deep learning based automated lane centering under physical-world attack. In USENIX Security, 2021
[7] On Adversarial Robustness of Trajectory Prediction for Autonomous Vehicles. In CVPR, 2022.
Q2. Clarification of writing
We apologize for the ambiguity and thank the reviewer for valuable feedback on our manuscript. We have carefully considered each of the points and propose the following revisions to enhance the coherence and readability of our paper:
a) Connection with the previous sentence (line 15): We will clarify the transition between sentences. Specifically, we will revise the first two sentences as follows: "Building upon the capabilities demonstrated by large-scale pre-trained vision transformers in zero-shot scenarios, a few annotated examples can be leveraged to further enhance generalization capability, achieving impressive performance across a broad spectrum of downstream tasks."
b) Transition in Paragraph (line 21): We acknowledge the abruptness of the transition and the unnecessary use of "However." We will revise as follows: "While few studies have explored how meta-tuning can maintain high performance under these conditions at the same time, it is crucial for real-world applications..."
c) Coherence in Third Paragraph (lines 29, 31, 33): We will revise these sentences to ensure a more cohesive flow of ideas.
d) Clarification of "double strongest perturbation" (line 45): We inject the strongest perturbations twice by initially attacking the input and subsequently perturbing the singular values and vectors with adversarial examples.
e) Simplification of the sentence (lines 130-132): We will simplify the sentence to "To robustify the learned meta-knowledge, adversarial meta-tuning injects the worst-case adversarial perturbation to the input ."
Summary of changes
We extend our sincere thanks to the reviewers for their constructive feedback. We have summarized additional experiments and clarification made during the rebuttal period as follows.
Clarification:
- Illustrated our technical contributions concerning adversarial singular value and vector perturbation, robust LoRAPool, and test-time merging mechanism. (Reviewer cLZu Q1 and Q6)
- Clarified the writing and added instances of novel and adversarial environments. (Reviewer ek5A Q1 and Q2)
- Clarified the top-k operation. (Reviewer cLZu Q2)
- Clarified the experiment details and evaluation settings. (Reviewer cLZu Q3, Q4, Q5)
- Added discussion of other parameter-efficient tuning methods and data augmentation techniques. (Reviewer VJpc Q1 and Q2)
Additional Experiments:
- Conducted a comparative analysis between our different design choices of proposed AMT. (Reviewer cLZu Q1 and Q6)
- Conducted the hyper-parameter study of LoRA rank, pool size, and loss coefficient (Reviewer cLZu Q5, Q7, Reviewer cdgf Q3)
- Analyzed the robust generalization for unseen attacks and AutoAttack under distribution shifts. (Reviewer cdgf Q1 and Q2)
- Conducted a comparative analysis with other parameter-efficient tuning methods and data augmentation techniques. (Reviewer VJpc Q1 and Q2)
Adversarial Meta-Tuning (AMT) addresses generalization in few-shot learning for image classification by mixing LoRA updates subjec tto perturbations. These perturbations are on inputs, as in preceding work, and on the singular value decompositions, which is novel. The mixture is determined at test time to customize the model to a given task during meta-testing. Results on multiple benchmarks validate the approach.
Four reviewers vote for acceptance at weak (ek5A, VJpc) and borderline levels (cLZu, cdgf). The authors provide a detailed rebuttal to each review, and half of the reviewers engage (with cLZu and cLZu raising their scores). During AC-reviewer discussion the AC provided the opportunity to highlight outstanding issues, and the remaining reviewers did argue against the submission. As such and after careful inspection of the submission and rebuttal threads, the AC finds the weaknesses to be adequately resolved and sides with acceptance.
The authors are strongly encouraged to incorporate the results on adapter pools and generalization to unseen attacks and AutoAttacks to reinforce the content and more convincingly inform the community. Clarifications 1 & 2 and the Additional Experiments 2-3 are of particular use for this purpose to communicate the contributions. As a last note, the AC underlines the importance of including and discussing adapter pools, since adapters are in wide use and well-studied (and a request for this emphasis was raised during the AC-reviewer discussion phase).
Note: the AC acknowledges the confidential comment by the authors and has resolved the matter by close inspection of the threads and the AC-reviewer discussion phase.
Miscellaneous feedback: please fix the typo of "Meta-Datset" in the caption of Table 4.