PaperHub
6.4
/10
Poster4 位审稿人
最低4最高4标准差0.0
4
4
4
4
3.8
置信度
创新性2.8
质量3.0
清晰度2.8
重要性2.5
NeurIPS 2025

Data Selection Matters: Towards Robust Instruction Tuning of Large Multimodal Models

OpenReviewPDF
提交: 2025-05-12更新: 2025-10-29
TL;DR

We design an efficient data selection approach to construct a robust training mixture for visual instruction tuning in large multimodal models.

摘要

关键词
Large Multimodal ModelsData SelectionRobustness

评审与讨论

审稿意见
4

This paper propose a gradient-free robustness-aware data selection framework, a data filtering strategy that selects the worst-case samples based on embedding distances from a lightly fine-tuned proxy model, to improve robustness and efficiency for instruction tuning. The approach is evaluated across multiple multimodal and language understanding benchmarks, and the authors report that selecting approximately 30% of the data yields comparable or even better performance than using the full training set.

优缺点分析

Strength:

  1. The proposed ARDS method is conceptually simple yet well-motivated.
  2. The method achieves consistently strong performance across a diverse set of benchmarks, highlighting the effectiveness of the proposed selection strategy.
  3. The proposed data selection strategy is especially relevant for scenarios where computational resources are limited or training on full datasets is infeasible.

Weakness:

  1. One major concern lies in the conceptual link between the data selection strategy and the robustness claims made throughout the paper. The authors heavily emphasize robustness as the central motivation, yet the core selection mechanism relies on embedding distances to identify worst cases. It is not sufficiently explained why samples that are semantically distant or harder to fit necessarily correspond to those that improve robustness. The assumption that worst-cases are aligned with robustness cases remains unsubstantiated, and the paper lacks an ablation or theoretical justification to support this connection. As a result, the method appears more as a form of relevance or difficulty-based sampling, rather than one explicitly targeting robustness.
  2. I found the results on ARC-e and BoolQ in Table 3 puzzling, the full-data training performs significantly worse than using only a random 30% subset. This seems counterintuitive unless the full training set introduces severe biases that the smaller subset helps avoid. Could the authors elaborate on this surprising behavior? It would be helpful to include further analysis or visualization, such as label distributions, training loss curves, or predictions breakdown, to confirm this effect is genuine rather than due to setup discrepancies.
  3. Figure 3 partially justifies the 30% ratio on two representative tasks, but it remains unclear whether ARDS consistently holds its advantage across tasks of varying modality complexity and reasoning depth. Given that the benchmark includes visual multiple-choice, text-based classification, and math QA, a brief discussion on whether this ratio remains effective across different cognitive requirements would strengthen the generality claim.
  4. One point that remains underexplored is the practical motivation for data selection in the current setup. Since the full training data-LLAVA665k is assumed to be available, clean, and standard across benchmarks, it is not immediately clear why discarding a large portion (70%) of it would be desirable. The authors should clarify under what practical or deployment conditions such a selection strategy is intended to be applied. For example, is the method primarily motivated by training efficiency, overfitting concerns, or robustness to spurious correlations? Making this clearer would help situate the proposed method in real-world training scenarios.

问题

See weakness above. Overall, my main confusion with this paper lies in how the authors directly associate worst-case examples and accuracy improvements with robustness. I would appreciate further clarification on this point, and I’m open to engaging in further discussion.

局限性

yes

最终评判理由

The author's supplementary experiments solved my q2 and q3. Although I still have some doubts about q1, I think it can be improved to score 4.

格式问题

NA

作者回复

We would first like to express our genuine appreciation for the time and effort the reviewer has dedicated to reviewing our manuscript. We believe these comments and questions provide valuable feedback and will greatly help us improve the overall quality of our work. Please let us know whether our response has sufficiently addressed your concerns. We are more than happy to engage in further discussion.

Q1.  Clarification of the association between worst-case selection and robustness gains

  • We appreciate the opportunity to elucidate the connection of our ARDS and robustness goal and will incorporate the following discussion into our revised manuscript.
    • To begin, we would like to humbly highlight our key idea is to customize targeted instruction selection (e.g., LESS) for improving the robustness against specific dataset biases while avoiding gradient-based calculation.
    • Targeted instruction selection improves robustness.
      • Theoretical Analysis. Built on TracIn[1], LESS[2] quantifies the influence of a training datapoint zz on the loss of a test data zz^{\prime} via (z;θt+1)(z;θt)ηt(z;θt),(z;θt)\ell\left(z^{\prime} ; \theta^{t+1}\right)-\ell\left(z^{\prime} ; \theta^t\right) \approx -\eta_t\left\langle\nabla \ell\left({z} ; {\theta}^t\right), \nabla \ell\left(z^{\prime} ; \theta^t\right)\right\rangle. Let mm denote vulnerable samples selected by our worst-case evaluation subgroups and mm^{\prime} denote perturbed test sample for robustness evaluation.
        • If the gradient inner-products (m;θt),(m;θt)\left\langle\nabla \ell\left(m ; \theta^t\right), \nabla \ell\left(m^{\prime} ; {\theta}^t\right)\right\rangle is notably non-negative, then the loss of perturbed test samples (m;θt+1)(m;θt)\ell\left(m^{\prime} ; \theta^{t+1}\right)-\ell\left(m^{\prime} ; \theta^t\right) decreases.
        • The decreased loss mirrors adversarial training for enhancing adversarial robustness [3].
      • Empirical Validation.
        • We first compute this gradient inner product using 100 training batches (batchsize=32) and 128 perturbed ScienceQA samples. The results of training batches from our robust training mixture show the average (m;θt),(m;θt)=25.76\left\langle\nabla \ell\left(m ; {\theta}^t\right), \nabla \ell\left(m^{\prime} ; {\theta}^t\right)\right\rangle=25.76, while that of training batches from random selection is 1.951.95. This strongly supports our theoretical premise.
        • We kindly refer to Table 6 for the contribution of worst-case evaluation subgroups.
    • Alignment between ARDS and targeted instruction selection. To avoid the heavy cost of gradient-based calculations and sidestep the need for access to downstream data, we leverage the rich hidden representation of LMMs and construct worst-case evaluation subgroups on the training corpus.
      • Theoretical Analysis. In Appendix M, we provide a theoretical analysis under a linear model to demonstrate that the cosine similarity discrepancy between feature embeddings rr and gradients gg of two data samples Δij=cos(ri,rj)cos(gi,gj)\left|\Delta_{i j}\right|=\left|\cos \left(r_i, r_j\right)-\cos \left(g_i, g_j\right)\right| can be upper-bounded.
      • Empirical Validation. We utilize the established gradient datastore in LESS and conversation vector datastore in our ARDS to compute the overlap between gradient-based selection and our feature-based selection for ScienceQA. The results, in the table below, demonstrate high selection overlap.
    • The transitive relationship thus substantiates why ARDS enhances robustness, supported by both theoretical insights and empirical evidence.
Selection ratio30%40%50%60%70%
Selection Overlap58.87%69.00%75.94%79.97%83.60%

[1] Estimating training data influence by tracing gradient descent. In NIPS, 2020

[2] Less: Selecting influential data for targeted instruction tuning. In ICML, 2024

[3] Towards deep learning models resistant to adversarial attacks. In ICLR, 2018

Q2. Why does full-data training perform worse than a random selection on pure-text tasks?

  • We appreciate the reviewer's great feedback, and delve into a meticulous analysis for the decreased performance of full-data trained model on ARC-e and BoolQ.
    • We hypothesize that this phenomenon arises from the catastrophic forgetting driven by imbalanced training data across modalities, which aligns prior work that identified text-only forgetting during visual instruction tuning [1, 2] For example, WINGS [1] reports up to 13.33% text-only degradation after visual modality expansion.
    • In LLaVA-665K, only around 40688/665298 \approx 6% of samples are pure-text instructions. During large‑scale multimodal tuning, the optimizer focuses overwhelmingly on vision‑centric tasks and gradually forgets language knowledge acquired from text‑only instruction tuning.
    • We have incorporated additional analysis by measuring text-only forgetting with random subsets of increasing size for LLaVA‑1.5 (7B), owing to compute limits. Larger subsets are expected to include more image-text pairs and thus amplify modality mismatch between multimodal and text-only inputs. The results, shown in table below, confirm that adding vision‑heavy data helps model excel in multimodal tasks while accentuates forgetting on text‑only tasks.
Task TypeTraining Size5%10%30%50%70%90%100%
text-only taskBoolQ Acc.64.1058.4455.9353.4355.6648.3537.77
multimodal taskA-OKVQA Acc.72.0574.3278.2579.0480.0080.7980.52
  • We would like to humbly highlight that after once data selection, all subsequent results (e.g., ScienceQA, ARC and BoolQ) were obtained with the identical checkpoint trained with the selected data under the same evaluation setup.

[1] Wings: Learning multimodal llms without text-only forgetting. In NIPS, 2024

[2] Vila: On pre-training for visual language models. In CVPR. 2024

Q3. More discussions concerning 30% ratio on other tasks

  • We appreciate this great suggestion, and in response, we have incorporated additional experiments to assess the generalization of our curated robust training mixture with 30% ratio.
    • Specifically, we have, following your advice, included a comparison with other mentioned methods on another text-only benchmark (SocialIQA[1]) and two additional math benchmarks (MathVista[2] and DynaMath[3]). These two mathematical benchmarks can be viewed as more difficult Out-of-Domain benchmarks, as the original LLaVA-665K does not explicitly involve math training data [4].
      • MathVista[2] is a mathematical reasoning benchmark introduced to analyze reasoning capabilities in visual complex scenarios (e.g., tables, function plots).
      • DynaMath[3] is a dynamic visual math benchmark designed to systematically analyze the robustness of mathematical reasoning capabilities across different topics (e.g., arithmetic, graph theory) under diverse condition changes (e.g., numerical value change). We report the average clean and robust accuracy on each benchmark.
    • The results, shown in the table below, demonstrate our curated robust training mixture with the 30% consistently boosts the generalization and robustness of visual instruction tuning, compared to full-data training and previous state-of-the-art selectors.
Methoddata percentageSocialIQA-AvgMathVista-AvgDynaMath-Avg
Full48.7023.3824.48
Random30%49.2519.8621.06
LESS30%50.9323.3825.19
COINCIDE30%52.3018.9819.98
ARDS(ours)30%58.5625.0026.14

[1] Social IQa: Commonsense reasoning about social interactions. In EMNLP, 2019.

[2] MathVista: Evaluating Mathematical Reasoning of Foundation Models in Visual Contexts. In ICLR, 2024

[3] DynaMath: A Dynamic Visual Benchmark for Evaluating Mathematical Reasoning Robustness of Vision Language Models. In ICLR, 2025

[4] Improved Baselines with Visual Instruction Tuning. In CVPR, 2024

Q4. More in-depth discussions concerning the motivation

  • We appreciate the reviewer's invaluable feedback.
    • We would like to humbly highlight our primary motivation is to improve the robustness of visual instruction tuning against specific dataset biases.
      • We agree with the reviewer that full training data-LLAVA665k are assumed to be clean, but individual clean samples do not imply they will not reinforce spurious correlations or detrimental bias [1]. Large multimodal models tend to memorize common concepts or prioritize stereotypes present in the training corpus [2]. For example, both "roses are red" and "banana is yellow" image-text pairs are clean samples, but training corpus dominated by these samples may make the model disregard the actual user input that deviates from those patterns, and thus harm generalization and robustness under distribution shift.
      • As shown both theoretically and empirically in Q1 discussion, our curated robust training mixture improves not only the robustness of visual instruction tuning but generalization capabilities as well—sometimes even exceeding the full-data baseline.
    • Beyond robustness, our smaller training mixture favorably enjoys data efficiency, which is also targeted by prior selection methods such as LESS and COINCIDE. The smaller training mixture is appealing for its lightweight storage requirements and reduced training cost. This makes our approach practical for real-world deployment.

[1] An investigation of why overparameterization exacerbates spurious correlations. In ICML, 2020

[2] Holistic analysis of hallucination in gpt-4v (ision): Bias and interference challenges. arXiv preprint arXiv:2311.03287, 2023.

评论

Thank you for the author's response, which has addressed some of my concerns. Including these additional experiments in the appendix would likely strengthen the paper further. I have updated my score to 4.

审稿意见
4

This paper designs a novel visual instruction selection approach, named ARDS. ARDS first divides clusters based on the extracted embeddings of conversations and then calculate ranking scores for each visual instruction by taking both embedding similarity and loss into account. Since ARDS only need one forward process and no backward process for each instance, ARDS is faster than gradient-based competitive methods. Extensive comparisons and ablation experiments are conducted to validate the effectiveness of ARDS.

优缺点分析

Strengths:

  1. The designed pipeline of ARDS is clear and intuitively meaningful. Authors make sufficient explorations in three key steps of ARDS: i) representation of conversation used for clustering; ii) identifying vulnerable samples as representations of clusters; iii) scoring mechinism that combines embedding similarity and average loss.
  2. According to experimental results, ARDS outperforms competing methods across multiple benchmarks and requires less time consume.

Weaknesses:

  1. The experimental section lacks an analysis of the distribution of the selected samples across the clusters.
  2. The ablation settings are to some extent hard to understand.

For a more detailed explanation, please refer to the Questions section below.

问题

  1. According to Equation 6, the value score of an instance is computed as an aggregation of the average cluster loss and the similarity between the instance and the cluster. Typically, there are some clusters with low average loss and others with significantly higher average loss. Clearly, instances that belong to or are closer to high-loss clusters tend to receive higher value scores. As a result, the selected samples are likely to be predominantly from these high-loss instances. My concern is that, in extreme cases, if the majority of selected samples come from such high-loss clusters, the model might partially or even largely forget the knowledge associated with low-loss clusters due to their underrepresentation in training. Therefore, I believe a statistical analysis of the selected samples across clusters is necessary, but the paper lacks this component.
  2. I do not understand how the ablation study in Table 6 is conducted. If no perturbation is applied, then x′ does not exist—how is Equation 4 computed in that case? And if clustering is not performed, how is the value score calculated? The paper lacks necessary clarification regarding this part of the experimental setup.

If my concerns are addressed, I would consider raising my rating.

局限性

yes

最终评判理由

The authors'rebuttal has addressed my concerns. After comprehensively assessing the paper's novelty and writing quality, I have decided to maintain the Rating of 4.

格式问题

No

作者回复

We sincerely thank the reviewer for providing valuable feedback. We detail our response below point by point. Please kindly let us know whether you have any further concerns.

W1 & Q1: More analysis of selected samples across worst-case evaluation subgroups (clusters).

  • We sincerely appreciate the reviewer's invaluable suggestion. Following your advice, we have incorporated the statistical analysis on the loss distribution of our worst-case evaluation subgroups.
    • Specifically, we measure balance through Shannon entropy given as e(S)=m=0Mp(Sm)logp(Sm)e(\ell _ {\mathcal{S}})=-\sum _ {m=0}^M p\left(\ell _ {\mathcal{S} _ m}\right) \log p\left(\ell _ {\mathcal{S} _ m}\right), where p(Sm)p\left(\ell _ {\mathcal{S} _ m}\right) is the weight of the mm-th cluster after Softmax(Sm)\text{Softmax}(\ell _ {\mathcal{S} _ m}). As entropy depends on the number of subgroups, we further normalize entropy w.r.t. the number MM of possible clusters: eˉ(S)=e(S)log(M)\bar{e}(\ell _ {\mathcal{S}})=\frac{e(\ell _ {\mathcal{S}})}{\log (M)}, where log(M)\log (M) is equal to the entropy of a uniform distribution of size MM. We obtained the normalized entropy eˉ(S)=0.92\bar{e}(\ell _ {\mathcal{S}})=0.92. The result empirically verifies that selected samples remain well dispersed under the loss-based weighted-sum aggregation strategy.
  • We would like to humbly highlight that we only perform clustering for building the worst-case subgroups and the importance score of each training sample is aggregated from all subgroups in a weighted-sum manner. To validate that this choice mitigates imbalance, we replaced it with a max operator (i.e., the importance score of each datapoint is determined only by its closest cluster). The results, in Appendix Table 10, show that the weighted-sum aggregation strategy yields a better clean and robustness trade-off and avoids the skew that the max variant exhibits. We appreciate the reviewer for prompting this analysis and will highlight these results in our main paper.

W2 & Q2: Clarification on the ablation settings in Table 6

  • We apologize for any potential confusion and misunderstandings.
    • For the first row in Table 6, we randomly sample the same number of samples MBMB as the final total size of the worst-case subgroups and the importance score of each training data is then aggregated with the sample-to-sample weighted-sum rule ( I(xi)=j=1MBexp(j)dijj=1MBexp(j)\mathcal{I}\left(x_i\right)=\frac{\sum _ {j=1}^{MB} \exp \left(\ell_{j}\right) \cdot d _ {i j}}{\sum _ {j=1}^{MB} \exp \left(\ell _ j\right)} ). For the second row, we apply Equation 4 to retrieve top-MBMB samples from the training dataset with the largest loss difference, and the importance score is aggregated in a sample-to-sample weighted-sum manner. The third row denotes our final construction strategy of worst-case subgroups, where we first perform clustering and use Equation 4 to retrieve top-BB candidates from each cluster. The importance score of training data is then aggregated with the sample-to-cluster weighted-sum scheme (Equation 6).
评论

Thank you for your reply, which has effectively addressed my concerns. I was surprised to see that the normalized entropy value is 0.92, as this indicates that the distribution of the selected samples is relatively uniform. I suggest incorporating this statistical experiment into the paper, as it would greatly enhance the intuitive understanding of the sample distribution. Additionally, I acknowledge the contributions of this work. That said, given that the explanation of the ablation settings constitutes a crucial component of the experimental section—and that the original version lacks sufficient clarity in this regard—I will maintain the rating at 4.

审稿意见
4

This paper proposes a data selection strategy for the SFT stage of MLLMs. The core idea is to utilize embeddings from a warmed-up proxy model to identify hard samples by measuring model consistency under random input variations, and then selecting data points around these identified hard samples.

优缺点分析

  • It is encouraging to see that principles from active learning and data selection, such as coreset selection and consistency/uncertainty metrics [1, 2, 3], can be straightforwardly adapted and remain effective in the context of MLLMs.
  • The manuscript would benefit from a more thorough discussion of its relationship to the broader field of active learning and data selection. Given the direct application of concepts from this area, explicitly positioning the work within this context would provide valuable perspective.
  • The set of comparative methods is insufficient. The paper would be substantially strengthened by including comparisons against more recent and relevant baselines, such as D3M, which appears to be closely related, and ICONS [5], a direct extension of LESS. Furthermore, the evaluation should be expanded to include a wider range of standard MM benchmarks.
  • A more detailed analysis of the time consumption is needed. This analysis should critically compare the proposed method's efficiency against strong baselines, particularly random sampling. And meantime take the performance gains into consideration (eg. comparison done by COINCIDE).
  • Table 2, as the primary results table, is presented in a way that could be misleading. Following the established practice in MLLM data selection literature, the main results table should focus on performance comparisons against the full data pool and other selection methods under standard, unaltered evaluation benchmarks. It is unfair to evaluate competing methods on an augmented evaluation set where the specific augmentations are leveraged by the proposed method but are unseen by the baselines. The analysis of robustness should be presented in a separate table to ensure a fair and clear comparison of core performance.

[1]Coresets for Scalable Bayesian Logistic Regression [2]Consistency-Based Semi-supervised Active Learning: Towards Minimizing Labeling Cost [3]Not All Labels Are Equal: Rationalizing the Labeling Costs for Training Object Detection [4]box-level active detection [5]ICONS: Influence Consensus for Vision-Language Data Selection

问题

  • How many augmented samples are generated for each seed instance during the selection process? What is the sensitivity of the proposed method's performance and computational cost to this number? A brief ablation study on the number of augmentations would be valuable.
  • The analysis is primarily conducted on the SEED and ScienceQA benchmarks. Could the authors provide a more detailed justification for this specific choice?
  • What is the performance of the proposed method on Vision-FLAN, the widely used dataset for selection?

局限性

The Limitation section currently reads more like a concluding summary and a brief outlook on future work. I recommend that the authors revisit the limitation section after expanding their experimental evaluation. This would allow for a much deeper and more valuable discussion.

最终评判理由

Most of my concerns have been addressed or clarified with concrete results. My remaining concerns are about the fair presentation of attack-proof and attack-unknown work in the data selection field, as well as ensuring the code is made public, which the author has also addressed. Therefore, I will maintain my positive score.

格式问题

na

作者回复

We would first like to express our genuine appreciation for the time and effort the reviewer has dedicated to providing such a detailed and in-depth review of our submission. We believe these comments and questions will contribute significantly to improving the overall quality of our work. We are more than happy to engage in further discussion.

Q1. More discussions concerning the relationship to active learning and data selection.

  • We sincerely thank the reviewer for the constructive comments and will incorporate the following discussion into our revised manuscript.
    • Both data selection and active learning aim to identify a subset of data that yields the best possible model, while the distinction lies in the focus on settings. Active learning defaults to that data is unlabled and emphasize on reducing annotation cost [1-5]. In contrast, data selection methods usually have full access to all labels when selecting [6], and the goal is to improve training efficiency [7] or robustness[8].
    • Our method ARDS is a data selection method with no extra annotation required. Our focus is to boost robustness against specific dataset biases for visual instruction tuning while avoiding computation-intensive gradient-based calculation in [7, 8].

Q2. More comparisons with other related work and on a wider range of standard benchmarks.

  • We appreciate the reviewer's great suggestion, and in response, we have conducted supplementary experiments to compare with D3M [8], which is a gradient-based data selection method designed for improving image classification robustness. Specifically, we reproduce D3M in our setting using the same warmed-up proxy model for extracting gradient vectors. We follow its standard procedure to project gradients, estimate the coefficient for each training sample and calculate the group alignment as the final score.
    • The results show that compared to this gradient-based method, the proposed ARDS achieves better robustness improvement while avoiding computation-intensive gradient calculation.
Methoddata percentageScienceQA-CleanScienceQA-PAScienceQA-SAScienceQA-SA+PAScienceQA-Avg.
Random30%69.7652.6059.4423.7551.39
D3M30%69.8656.8762.5227.1754.10
ARDS(ours)30%69.2659.4068.5747.6061.21
  • Additionally, we have included more standard benchmarks for a more comprehensive comparison. The results, shown in the table below, demonstrate that our method achieves comparable generalization capability and consistently boosts robustness across various benchmarks. This further demonstrates the effectiveness of the proposed ARDS.
Methoddata percentageTextVQAGQASocialIQA-AvgMathVista-AvgDynaMath-Avg
Full57.2361.9448.7023.3824.48
Random30%55.9059.6149.2519.8621.06
LESS30%49.4556.4950.9323.3825.19
COINCIDE30%56.0959.1552.3018.9819.98
D3M30%54.4156.7351.2823.6920.55
ARDS(ours)30%56.9760.3258.5625.0026.14

Q3. More computational analysis with baselines

  • We appreciate the reviewer's valuable suggestion, and in response, we have incorporated additional computational analysis.
    • We would first like to humbly emphasize that we use the standard LLaVA-1.5 training configurations when comparing all baseline methods, and thus the wall-clock time of visual instruction tuning on the same size of training mixture should be the same for all baseline methods.
    • While ARDS, together with prior competitor COINCIDE, extract sample features for information measure to perform data selection, which can indeed increase the computational cost compared to random selection, this trade-off is justified by our results significantly enhancing the robustness—a key motivation of our work.
MethodSelection RatioWall-clock time of data selectionWall-clock time of visual instruction tuningSciPA-Avg.SEED-Avg.
Random10%<1min3.6h46.2733.31
COINCIDE10%100 min3.6h47.3033.91
ARDS (ours)10%39 min3.6h61.4245.08
Random30%<1min6.9h51.3937.97
COINCIDE30%100 min6.9h52.2739.58
ARDS (ours)30%39 min6.9h61.2146.80

Q4. Split Table 2 into separate tables and add more results to ensure fair comparisons

  • We appreciate the reviewer’s invaluable suggestion.
    • To begin, we would like to humbly emphasize that our primary motivation is to improve the robustness of visual instruction tuning against specific dataset biases. We also humbly emphasize that our method does not require access to downstream few-shot examples used in LESS or a reference model well-trained on the full data required by COINCIDE.
    • We will follow your advice to divide clean and robust accuracies in Table 2 into separate tables to make comparisons clearer.
    • Following your advice to ensure fair comparison, we have conducted additional experiments by adding our constructed worst-case evaluation subgroup as the selection source for the previous state-of-the-art selector COINCIDE. The results, shown in the table below, demonstrate the effectiveness of our ARDS to achieve superior robustness across benchmarks.
MethodData PercentageScienceQA-AvgSEED-AvgMMBench-EN-AvgA-OKVQA-Avg
COINCIDE30%52.2739.5859.5464.97
COINCIDE+worst-case subgroup30%53.2140.9760.44567.77
ARDS (ours)30%61.2146.8065.2672.95

Q5 Clarification and more results on the worst-case perturbation.

We deeply appreciate the reviewer’s feedback. To begin, we would like to humbly highlight that perturbations are introduced solely to build worst-case evaluation sub-groups, which are not applied during training. To build the worst-case subgroups, we inject task-specific visual and textual perturbations (i.e., symbol plus permutation attacks) simultaneously, and thus only one augmented sample is generated. We further explored the effect of generating multiple samples. The results, shown in the table below, demonstrate that the performance of the proposed method is not very sensitive to the number of augmentations.

#Aug. in Worst-caseComputational Costdata selection ratioScienceQA-CleanScienceQA-PAScienceQA-SAScienceQA-SA+PAScienceQA-Avg
139min20%69.2659.4068.5747.6061.21
250min20%68.6758.8568.0748.3460.98

Q6. Justification for our analysis of more evaluation benchmarks.

  • We appreciate the reviewer's invaluable suggestion and we have incorporated additional, more detailed ablational analysis using A-OKVQA, MMMU, MMBench-EN, and MMBench-CN benchmarks. We report the average clean and robust accuracies on each benchmark. The results, in the table below, demonstrate that each component of the proposed method is consistently effective across benchmarks.
Worst-case PerturbationEvaluation Subgroup ClusteringData PercentageScienceQA-Avg.SEED-Bench-Avg.A-OKVQA-Avg.MMMU-Avg.MMBench-EN-Avg.MMBench-CN-Avg.
×\times×\times30%52.3940.4366.0320.9260.5355.42
\checkmark×\times30%55.6545.4071.5721.9964.0958.79
\checkmark\checkmark30%61.2146.8072.9522.8865.2659.80

Q7 More results on Vision-FLAN

  • We sincerely appreciate the reviewer's invaluable suggestion and have conducted additional experiments on another training dataset Vision-Flan using the proposed ARDS. The results, shown in the table below, demonstrate that the robust training mixture created by ARDS on Vision-Flan achieves superior robustness, highlighting the effectiveness of our robustness-aware data selection method in mitigating the model's biased behaviors.
MethodData PercentageSci-CleanSci-PASci-SASci-SA+PASci-Avg.
Full100%64.0639.7153.5913.7342.77
Random50%61.3834.4153.5918.8942.06
ARDS (Ours)50%62.8737.9860.5433.5648.74

Q8. More discussion concerning limitations.

  • We appreciate the reviewer's valuable feedback and will incorporate the following discussion of limitations in the revised manuscript. Although the efficacy of the proposed method has been confirmed through empirical experiments for mitigating several dataset biases, such as position bias, symbol-content spurious correlations, and visual spurious correlations, opportunities for refinement persist. One avenue for improvement involves a more diverse investigation into the potential dataset biases and their interaction effects.

[1]Coresets for Scalable Bayesian Logistic Regression. In NIPS, 2016

[2]Consistency-Based Semi-supervised Active Learning: Towards Minimizing Labeling Cost. In ECCV, 2020

[3]Not All Labels Are Equal: Rationalizing the Labeling Costs for Training Object Detection. In CVPR, 2022

[4]Box-level active detection. In CVPR, 2023

[5]A Survey on Deep Active Learning: Recent Advances and New Frontiers, In TNNLS, 2024

[6]A Survey on Data Selection for Language Models. In TMLR, 2024

[7]Less: Selecting influential data for targeted instruction tuning. In ICML, 2024.

[8]Improving subgroup robustness via data selection. In NIPS, 2024.

评论

Thank authors for their response and the effort put into the additional experiments. I have two concerns I would like to follow up on:

Regarding W5, the stated goal is to select samples that are most robust to attack. A truly practical and meaningful validation of this would be to demonstrate robustness against novel, unknown attacks. The current experimental setup, which uses an attack-aware method to compare against general selection baselines, creates an unfair comparison where the proposed method is essentially acting as both the player and the referee. In my opinion, a proper way to validate this setting would be to hold out a set of attacks or devise a new attack group for testing, rather than evaluating on a test where the challenges are known to the method beforehand.

Additionally, I would like to inquire if the authors intend to release the code for reproducibility.

评论

We appreciate the reviewer for reviewing our rebuttal and for allowing us to further clarify your concerns.

Concern 1: More in-depth discussions concerning attack awareness during data selection

  • We appreciate the reviewer's insightful suggestion to further evaluate our proposed data selection setting on novel attacks.
    • We first kindly refer to Appendix Table 13 for the robustness evaluation on ScienceQA under previously unseen attacks, where we introduced two previously unseen symbol attacks and an additional visual corruption.
    • We have conducted additional comparisons on more benchmarks to extend Appendix Table 13, pitting ARDS against Random selection, COINCIDE, and its strengthened variant—COINCIDE supplied with our worst-case subgroup to fully align the seen data during selection.
      • SA1 and SA2 denote the canonical answer labels A/B/C/D are replaced by S/N/V/F and U/I/O/P, respectively, which are injected as novel variations not encountered during data selection (recall we used Q/W/E/R to build worst-case evaluation subgroups)
      • PA means the adversarial permutation attack, adaptively generated for each new question by enumerating all possible permutations of the answer options.
      • The results, presented in table below, demonstrate that ARDS consistently achieves strong robustness and generalization even in the presence of these new, previously unseen attacks. Even when compared with the COINCIDE+worst-case subgroup, ARDS maintains superior robust accuracy across all scenarios, confirming that the robustness gains generalize well beyond attacks seen during data selection.
  • We would also like to humbly highlight that the setting of defender knowing the attack type of adversary aligns with standard practices in adversarial training and evaluation, where models are usually trained against specific adversarial perturbation budget, such as ll_{\infty} [1], l2l_{2}[2], l1l_{1}[3]-norm bounded perturbations, and later evaluated on the corresponding type of attacks to measure robustness improvement over clean training.
  • We appreciate the reviewer's invaluable feedback again, which has strengthened our work.
Selection MethodData PercentageSci.-SA1Sci.-SA1+PASci.-SA2Sci.-SA2+PASci.-Avg.SEED-SA1SEED-SA1+PASEED-SA2SEED-SA2+PASEED-Avg.A-OKVQA-SA1A-OKVQA-SA1+PAA-OKVQA-SA2A-OKVQA-SA2+PAA-OKVQA-Avg.MMBench-SA1MMBench-SA1+PAMMBench-SA2MMBench-SA2+PAMMBench-Avg.
Full100%67.7240.8066.7336.4952.9356.6225.4755.0419.8839.2577.2152.4076.0748.0363.4269.0546.2968.9136.6155.21
Random30%64.0632.9263.8131.2848.0147.8213.5747.228.2929.2260.2624.9865.1518.3442.1860.1829.3662.5121.3043.33
COINCIDE30%65.4437.0463.4129.8548.9451.6519.4048.5811.8232.8668.4737.0366.5525.6849.4364.2235.2764.2923.9846.94
COINCIDE+worst-case subgroup30%62.8733.9663.1631.0447.7550.2816.1951.0816.0333.3967.3429.2671.7935.9051.0763.7630.4965.9026.3146.62
ARDS (Ours)30%68.5748.3466.1439.7655.7057.4833.8154.9522.2742.1378.6063.6774.5049.0066.4471.7551.9870.4842.5059.18

[1] Towards deep learning models resistant to adversarial attacks. In ICLR, 2018

[2] Adversarial training and robustness for multiple perturbations. In NIPS, 2019

[3] Towards stable and efficient adversarial training against l1 bounded adversarial attacks. In ICML, 2023

Concern 2: Additionally, I would like to inquire if the authors intend to release the code for reproducibility.

Yes. To promote transparency and reproducibility, we will release the source code.

评论

Thank the authors for additional experiments and analysis, which partly addressed my concerns. I understand that dealing with known attack is a standard practice in adversarial training. However, for MLLM data selection methods, this direction of research is relatively new and unknown to the design process of previous competitors. While I recognize its value, Id like to make sure the comparison is presented as fair as possible. I will maintain my positive score.

审稿意见
4

This paper focuses on developing data selection method to alleviate the data bias and supious correlation in the visual instruction tuning data mixture. The paper is well-written and the experimental section is well-organized. The proposed method ARDS is evaluated in serveral aspects and many benchmarks.

优缺点分析

Pros:

  • The proposed method is gradient-free.
  • The paper is well-organized.
  • The paper evaluates the ARDS extensively across various benchmarks and settings.

Cons:

  • The main concern is that in a real scenario, one cannot obtain the whole training dataset at one time, if new data comes, one should recalculate the worst-case evaluation groups again, which would cost a lot. Ideally, we would like a data selection method that could directly give a decision on each candidate data before training and inference.

  • If you compress the 30% data again, how small will the final data mixture be?

Small typo errors: There is no Figure 4 in the main paper. Figure 4 in Line 271 cannot be clicked.

问题

See weakness.

局限性

Please add the limitation discussion at the end of paper.

最终评判理由

After reviewing the authors’ responses to the weaknesses I raised, I am satisfied that they have fully addressed my concerns. I continue to regard this work as highly valuable and will maintain my positive score. Moreover, I hope this discussion will inspire future research to develop methods that can be applied without access to the entire dataset for data selection, an approach better suited to real‐world scenarios, where data and knowledge are constantly evolving.

格式问题

Add limitation section.

作者回复

Thank you sincerely for your thoughtful feedback on our work. Below, we have provided a detailed explanation for your concerns as follows. Please do not hesitate to let us know if you have any further questions.

Q1. More results in scenarios where the whole training dataset is not available.

  • We appreciate the reviewer bringing up the concern in a scenario where one cannot obtain the whole training dataset at one time.
    • We would first like to highlight that our experimental setting follows existing static visual instruction selection methods, where the selectors are aware of all training data to build a gradient database [1] or perform clustering on the whole dataset [2]. The key idea of our worst-case evaluation subgroups is to find representative samples that are most vulnerable to model-biased behavior while not needing the downstream few-shot examples required in [1].
    • We totally agree with the reviewer that a selection method supporting dynamically coming data is desired. To further address the concern about the training data requirement for building our worst-case evaluation subgroups, we have conducted supplementary experiments. We randomly sample 10% training data as the obtained small subset of the original training corpus and treat the remaining 90% as new incoming training data. We denote the variant as ARDS^*. Specifically, to build worst-case evaluation subgroups, we perform clustering on these 10% data with the same number of clusters KK and subgroup budget BB and then apply task-specific perturbations (see Appendix Line 424-428 for more details). The results indicate that despite seeing only a tenth of the corpus during subgroup construction, ARDS\*^\* achieves robustness close to the full-information ARDS—and still outperforms both LESS and COINCIDE. This highlights the potential and applicability of our method to more dynamic data selection scenarios.
MethodSelection RatioSci-CleanSci-PASci-SASci-PA+SASci-Avg.
LESS [1]30%68.4255.6364.7034.9555.93
COINCIDE [2]30%67.7252.2161.0828.0652.27
ARDS30%69.2659.4068.5747.6061.21
ARDS^*30%68.8658.3067.1342.3959.17

[1] Less: Selecting influential data for targeted instruction tuning. In ICML, 2024.

[2] Concept-skill transferability-based data selection for large vision-language models. In EMNLP, 2024.

Q2. More results on further data selection.

  • We appreciate the reviewer's insightful suggestion regarding performing data selection a second time.
    • To begin, we would like to humbly emphasize our current 30% ratio is determined by the trade-off between data efficiency as well as both clean and robust performance across varying tasks (kindly refer to Appendix Figure 5)
    • Following your advice, we have incorporated additional experiments to explore the impact of performing data selection again. The total selection ratio equals to the product of the two selection ratios. The results, in the table below, demonstrate that iterative selection, which first chooses 30% and then selects that subset to 33.3%, outperforms a single 10% selection. This indicates the great potential of our ARDS method to be extended to other scenarios. We will add the results and the discussion to our revised manuscript.
MethodTotal Selection Ratio1st Selection Ratio2nd Selection RatioSci-CleanSci-PASci-SASci-PA+SAAvg.
Random10%10%-66.4845.3155.6817.6046.26
COINCIDE10%10%-66.2448.1456.5718.2447.30
ARDS (ours)10%10%-68.7255.7368.8652.3561.42
ARDS (ours)10%30%33.33%71.1060.1471.1556.4264.70

Q3. Typos correction

We appreciate the reviewer for pointing out these typos. We have revisited and corrected the typos in our manuscript.

Q4. Clarification of limitation discussion

We appreciate the reviewer's feedback, and we kindly refer to Appendix A for the limitation discussion.

评论

Thank you to the authors for their response.

Regarding Weakness 1, I totally agree with the current setting. However, I would also suggest exploring a novel method to address the data‐selection problem without full knowledge of the entire training set, as I believe this more closely reflects real-world scenarios.

For Weakness 2, I appreciate the additional experiments you’ve conducted. They address my curiosity about iterative data selection, and the results are intriguing: two-stage selection outperforms one-stage selection at the same overall ratio. It would be valuable to see an experiment in which you select 50% (just an example) of the data in the first stage, then an additional XX % in the second stage, resulting in a total of 30 % selected, and compare this to the baseline 30 % one-stage selection. Such an analysis would further strengthen the impact of your paper.

Overall, I find this work very valuable and will maintain my positive score.

评论

More results on extending our method to iterative data selection.

We appreciate the reviewer’s invaluable suggestion and have conducted additional experiments that compare one-stage and two-stage iterative selection with ARDS while keeping the total selection ratio at 30%.

  • Specifically, we vary the stage-1 ratio in {40%,50%,60%} and set the stage-2 ratio to keep the same total selection ratio.
  • We observe that performance depends on how the 30% budget is split: the 60%→50% configuration yields a comparable performance to one-stage 30%, while other ratio allocations are slightly lower. Compared to the benefits shown in our earlier iterative selection results (30%→33.3% outperforms a single 10% selection), this also suggests that iterative selection gains might be saturated when the total selection ratio is relatively large. In addition, we believe exploring improved ratio budget splits for iterative selection deserves future work.
MethodTotal Selection Ratio1st Selection Ratio2nd Selection RatioSci-CleanSci-PASci-SASci-PA+SAAvg.
Random30%30%-69.7652.6059.4423.7551.39
COINCIDE30%30%-67.7252.2161.0828.0652.27
ARDS (ours)30%30%-69.2659.4068.5747.6061.21
ARDS (ours)30%40%75%69.5157.4666.4440.1158.38
ARDS (ours)30%50%60%69.1657.2666.3941.7558.64
ARDS (ours)30%60%50%69.4658.0667.7747.1560.61
评论

We extend our sincere thanks to the reviewers for their constructive feedback. In order to provide greater clarity on the revisions made to our paper and the experiments we conducted to address the reviewers' questions, we have summarized the experiments and clarifications made during the rebuttal period as follows.

Additional Experiments

  • Extended ARDS to scenarios where the complete training set is unavailable (Reviewer 5LuU Q1)
  • Extended ARDS to a second-round selection. (Reviewer 5LuU Q2)
  • Added comparisons to more related work and on a wider range of benchmarks. (Reviewer LJnh Q2, Reviewer NiSF Q3)
  • Provided a more detailed comparative analysis of the computational overhead. (Reviewer LJnh Q3)
  • Applied our proposed worst-case evaluation subgroups to prior selection methods for a direct comparison. (Reviewer LJnh Q4)
  • Analyzed how performance varies with the number of generated samples when building worst-case subgroups. (Reviewer LJnh Q5)
  • Performed additional ablations on more benchmarks. (Reviewer LJnh Q6)
  • Extended ARDS to other training datasets. (Reviewer LJnh Q7)
  • Analyzed the loss distribution across worst-case evaluation subgroups. (Reviewer HSHf Q1)
  • Analyzed the observed catastrophic forgetting on pure-text tasks. (Reviewer NiSF Q2)

Clarifications

  • Added a high-level theoretical analysis of ARDS. (Reviewer NiSF Q1)
  • Added a deeper discussion of our core motivation (Reviewer NiSF Q4)
  • Clarified experimental settings used in ablations. (Reviewer HSHf Q2)
  • Added a deeper discussion of the ARDS's position within active learning and data selection areas. (Reviewer LJnh Q1)
  • Added a deeper limitation discussion (Reviewer LJnh Q8)
  • Corrected typos (Reviewer 5LuU Q3)
最终决定

This paper proposes a visual instruction selection approach, named ARDS, for the SFT stage of MLLMs. The core idea is to utilize embeddings from a warmed-up proxy model to identify hard samples by measuring model consistency under random input variations, and then selecting data points around these identified hard samples. As ARDS only needs one forward process and no backward process for each instance, it is faster than gradient-based methods. However, it is also somewhat costly due to its heavy pipeline, which involves hierarchical clustering, perturbation, and loss difference evaluations. Extensive comparisons and ablation experiments are conducted to validate the effectiveness of ARDS.

The paper received unanimously positive feedback from the reviewers. Most of the reviewers' concerns have been addressed during the rebuttal phase. I encourage the authors to include the supplementary experiments as well as some of the discussions during the rebuttal in their final version.