PaperHub
6.0
/10
Poster4 位审稿人
最低6最高6标准差0.0
6
6
6
6
4.3
置信度
正确性2.8
贡献度2.5
表达3.0
ICLR 2025

MIA-Bench: Towards Better Instruction Following Evaluation of Multimodal LLMs

OpenReviewPDF
提交: 2024-09-27更新: 2025-02-28
TL;DR

We introduce MIA-Bench, a benchmark designed to assess MLLMs’ ability to strictly adhere to complex instructions.

摘要

关键词
Multimodal LLM; Instruction Following; Benchmark

评审与讨论

审稿意见
6

The paper introduces MIA-Bench, a benchmark for evaluating the ability of multimodal large language models (MLLMs) to follow complex instructions. The benchmark consists of 400 image-prompt pairs that test the models' compliance with layered instructions to generate accurate responses. The evaluation of various state-of-the-art MLLMs shows significant performance variations, indicating areas for improvement in instruction adherence. The paper also discusses the creation of additional training data and supervised fine-tuning to enhance the models' instruction-following capabilities without affecting their performance on other tasks.

优点

  • The paper presents a new benchmark that rigorously tests multimodal large language models' ability to follow complex instructions, addressing a previously underexplored area.

  • This paper offers a comprehensive set of 400 diverse image-prompt pairs that challenge models with layered instructions, enhancing the assessment of their linguistic and descriptive capabilities.

  • Experiment demonstrates performance improvements in instruction adherence through supervised fine-tuning.

缺点

  • Complex instruction flowing ability might not be the main interests for MLLM users currently, especially when it comes to writting style constraints such as length and genre. Such abilities are mostly evaluated on LLMs.

  • The performance of models on MIA-Bench does not necessarily correlate with their performance on other benchmarks, suggesting that excelling in this specific task may not translate to generalized improvements across different multimodal tasks.

问题

  • In MIA-Bench, are instructions all related to the corresponding image content or randomly picked from a instruction bank?
  • Instruction following abilities highly depends on LLMs used in the MLLM models. It would be better to group the performances in Table 1 and Table 2 according to the LLMs' size.
评论

Thank you for your thoughtful review of our submission. We appreciate your feedback on both the strengths and areas for improvement in our paper. Below, we provide clarifications to your concerns.

Complex Instruction Following Relevance for MLLM Users

We understand your concern regarding the importance of complex instruction-following capabilities, especially with respect to writing style constraints like length and genre. However, we argue that instruction adherence is increasingly important for MLLMs as they are used in applications requiring precise, multimodal interaction, such as visual-based assistants, educational tools, and creative applications where adherence to complex instructions (even stylistic) enhances user experience. For example, in educational scenarios, such as automated tutors, a teacher might instruct a model to "Summarize the image content as a short story suitable for a 7-year-old, incorporating a cheerful tone." Here, the combination of age-appropriate language, a specific tone, and multimodal comprehension tests the model’s ability to follow complex instructions critical for engaging and effective learning experiences. Multimodal cooking assistants could be tasked with, "Generate a recipe based on this image of ingredients, ensuring the instructions are concise, use metric measurements, and fit within a single screen view." Adherence to format, length, and context-specific guidelines enhances usability for users in real-time cooking environments. The MIA-Bench aims to address this need by focusing on layered and multifaceted instruction adherence tasks, which evaluate MLLMs beyond basic visual recognition, for more user-aligned multimodal interactions.

Performance Correlation with Other Benchmarks

While MIA-Bench performance may not correlate directly with general-purpose benchmarks, this result aligns with our paper’s goal of highlighting instruction adherence as a distinct and specialized capability for MLLMs. This is our strength and motivation. Our findings show that excelling in MIA-Bench, which focuses on instruction adherence, serves as a unique measure of MLLMs’ capability that is not fully covered by most popular benchmarks.

Are instructions all related to the corresponding image content or randomly picked from a instruction bank?

They are all related to the corresponding image, and MLLMs should not be able to provide a completely correct response without seeing the image. Each instruction is unique, written by human, and not picked from an instruction bank.

It would be better to group the performances in Table 1 and Table 2 according to the LLMs' size.

Yes your suggestion makes a lot of sense. We have updated the paper to group open-source models into three categories: those with fewer than or equal to 8B parameters, those with 8B to 13B parameters, and those with more than 13B parameters.

We sincerely appreciate your valuable feedback, which will contribute to enhancing the quality and impact of our work.

评论

Thanks for your feedback. Writting instructions accordingly and uniquely for each image makes this dataset helpful and valuable for the community. I have raised my rating for that.

审稿意见
6

To evaluate and guide the instruction-following capabilities of multimodal models, MIA-Bench introduces a set of benchmarks for multimodal instruction. MIA-Bench primarily measures the abilities of MLLMs on following layered and compositional multimodal instructions and provides a set of multimodal instructions to enhance model performance in these areas.

优点

  1. The paper proposes a benchmark for evaluating the abilities of MLLMs on following compositional instructions, covering a variety of categories and tasks, and providing guidance for MLLMs in following complex composite instructions in the future.

  2. The paper proposes a set of instructions that can effectively enhance the ability of MLLMs to follow layered and compositional instructions.

缺点

  1. Since the images in MIA-Bench are sourced from widely used datasets such as COCO 2017, SBU, TextVQA, and Flickr, I believe you should prioritize evaluating whether the current open-source models exhibit any data leakage issues on these datasets to demonstrate that MIA-Bench does not suffer from data contamination.

  2. I notice that the experiments in Table 3 were conducted on LLaVA-Next-13b. Is this done after fine-tuning on a preference-aligned model? If so, I think this continued SFT setup leading to a performance improvement in MIA-Bench, while other benchmarks exhibit varying degrees of decline, is quite intuitive. Could you please provide additional experiments on the impact of mixing generated instructions with original instructions on model performance? Furthermore, I notice that there are only 5000 generated instructions here; how would the model performance on MIA-Bench be affected if the number of instructions were increased?

问题

  1. The 400 instructions in MIA Bench are manually annotated. It is important to ensure the reasonableness of the sub-instructions, such as whether the length limitation is appropriate. Additionally, has there been any manual verification of the scoring results from GPT-4o to confirm their accuracy and reasonableness?

  2. The instructions in MIA-Bench are composed of multiple sub-instructions. How is the number of sub-instructions per instruction determined? Is there a difficulty grading for the instructions, such that a higher number of sub-instructions indicates a greater level of difficulty?

审稿意见
6

This paper proposes MIA-Bench, which mainly focuses on the instruction-following capability of current Large Multimodal Models (LMMs) under complex and compositional instructions. Unlike previous close-ended benchmarks (e.g., multiple choice for MMBench) and open-ended benchmarks (e.g., LLaVA-Bench), MIA-Bench aims to evaluate precise adherence to complex instructions. As for evaluation, the authors leverage GPT-4o as the judge model. The authors have evaluated many representative LMMs on the proposed MIA-Bench.

优点

  1. Evaluating the capability of LMMs to adhere to complex instructions is worth studying.
  2. Table 2 demonstrates that MIA-Bench is not highly correlated with existing benchmarks.
  3. Figure 8 illustrates that MIA-Bench is vision-centric, which is not strongly correlated with the performance of the LLM backbone.

缺点

In the context of evaluating open-ended freeform responses like MIA-Bench, it is crucial to account for the inherent variability introduced by the judge model. For example, the performance of MM-Vet has been observed to fluctuate by up to 10 points when assessed with different versions of GPT.

This variability raises several important considerations:

  1. Standard Deviation Reporting: When evaluating Large Multimodal Models (LMMs) using the same judge model (e.g., GPT-4o) across multiple trials, it is essential to report the standard deviation of the performance metrics.
  2. Detailed Generation Configuration: The specific generation parameters of the judge model, such as top-p, top-k, temperature, and num_beams, should be explicitly documented.
  3. Impact of Generation Configuration: It is necessary to investigate whether the above generation configuration of the judge model has a substantial impact on the performance metrics.
  4. Performance Across Different Judge Models: Detailed performance metrics should be reported for various judge models. Specifically, the performance of LMMs should be evaluated using different versions of judge models, such as Claude-3-Opus, GPT-4o-20240806, GPT-4o-20240513, GPT-4v-20240409, GPT-4-1106-preview, or even open-sourced models (e.g., Qwen2.5 and LLaMA-3.1). This comprehensive approach will help in identifying any model-specific biases and in providing a more reliable assessment of the LMMs.

问题

I do not have any further questions. Please refer to the "weaknesses" section.

评论

Thank you for your valuable feedback on our paper. We appreciate your insights into the importance of robust evaluation metrics and variability considerations for MIA-Bench, and we are happy to address your concerns below.

Q1: Standard deviation reporting

A1: In the table below we show STD information of the evaluation scores. For each model, we run inference for three times on MIA-Bench and compute STD of total score.

ModelFuyu-8bInstructBLIP-13bKosmos-2mPLUG-Owl2InternVL-Chat-v1.5SphinxQwen-VL-ChatLLaVA-1.5-7bLLaVA-1.5-13bLLaVA-1.6-7bLLaVA-1.6-13bLLaVA-1.6-34bIdefics-2-8bGemini-1.0-ProClaude-3-OpusClaude-3-HaikuClaude-3-SonnetGPT-4v
STD0.007470.005860.008080.000960.012090.005170.003730.010150.011340.01060.002730.012950.014930.007020.017240.009470.004650.00326

Q2: Detailed generation configuration

A2: The setting of GPT-4o is set to default. Top-p and temperature default to 1. More settings can be found here in their official reference: https://platform.openai.com/docs/api-reference/chat

Q3: Impact of generation configuration

A3: The common practice of using GPT-4/GPT-4v/GPT-4o as a judge is to use default settings.

Q4: Performance across different judge models

A4: In an attempt to analyze if GPT-4o’s evaluation is reliable, we used Claude-3-opus as a second judge and reported results in section ‘Other external models as the judge‘, from line 417 to line 424.

We hope that our response addressed all your concerns. Thank you again for your feedback.

评论

I believe the authors have not catched my key points, and this response is perfunctory. Specifically:

  1. Towards Q1, more details should be reported regarding the STD.

    • Does the STD result from multiple inferences using the same model, or from evaluating the same inference results using the same judge model?
    • Why is the STD so small? For example, is the STD for Fuyu-8B 0.00747 (0.747%) or 0.00747%? The main table reports 24.52% for Fuyu-8B if my understanding is correct, so clarity on the STD value is essential.
  2. Towards Q3, it is highly recommended to conduct multiple times using the same judge model under the same inference results with different generation configurations of the judge model. This will help ensure the robustness and reliability of the evaluation results.

  3. Towards Q4, the evaluation should incorporate more judge models rather than relying solely on Claude-3-opus. Using a variety of judge models will provide a more comprehensive and balanced assessment of the performance.

If the authors continue to provide perfunctory responses and do not take my concerns seriously, I will consider lowering my score.

评论

Q3 Towards Q3, it is highly recommended to conduct multiple times using the same judge model under the same inference results with different generation configurations of the judge model. This will help ensure the robustness and reliability of the evaluation results.

A3: We set temperature to default (0) as this provides a more deterministic result; higher temperature increases the randomness which is not what we want from the judge. Top-p controls the diversity of generated text, this is by common practice of MLLM benchmarks to set to default as well, as the judge's main purpose is generating evaluation scores. Following your suggestion, we conducted additional evaluations using the gpt-4o-2024-11-20 model under different temperature settings (0, 0.1, and 0.2) and top p (0.8, 0.9, 1) to assess the robustness and reliability of the evaluation results. Below are the result tables.

ModelScore by gpt-4o-2024-11-20 (temp=0)Score by gpt-4o-2024-11-20 (temp=0.1)Score by gpt-4o-2024-11-20 (temp=0.2)
GPT-4o89.9491.0490.73
Claude-3-Opus84.8987.6687.41
Reka82.6884.1383.30
MiniCPM-Llama3-V-2.578.7579.1879.55
Gemini76.3276.2076.76
LLaVA-1.5-13b66.0567.6367.25
ShareGPT4v65.7266.9767.18
Idefics-2-8b53.6153.6653.92

Table 1: Evaluation score by gpt-4o-2024-11-20 with different temperature.

ModelScore by gpt-4o-2024-11-20 (top p=1)Score by gpt-4o-2024-11-20 (top p=0.9)Score by gpt-4o-2024-11-20 (top p=0.8)
GPT-4o89.9489.5590.80
Claude-3-Opus84.8986.8186.68
Reka82.6883.4883.60
MiniCPM-Llama3-V-2.578.7578.9579.35
Gemini76.3275.7375.69
LLaVA-1.5-13b66.0567.2667.43
ShareGPT4v65.7267.1866.88
Idefics-2-8b53.6153.7454.29

Table 2: Evaluation score by gpt-4o-2024-11-20 with different top p.

The results demonstrate consistent ranking across models, with minimal fluctuations in the scores despite changes in temperature. This consistency indicates that the judge model’s scoring is stable across varying generation configurations, further validating the robustness of our evaluation framework. The minor score variations observed are within an acceptable range.

评论

Thank you for your additional comments. Here we answer your questions one by one:

Q1 Does the STD result from multiple inferences using the same model, or from evaluating the same inference results using the same judge model?

A1: We ran inference for three times on MIA-Bench, then evaluated the three different inference results using GPT-4o.

Q2 Why is the STD so small? For example, is the STD for Fuyu-8B 0.00747 (0.747%) or 0.00747%? The main table reports 24.52% for Fuyu-8B if my understanding is correct, so clarity on the STD value is essential.

A2: 0.00747 means 0.747%. Below we update the table. The evaluation scores from multiple runs are pretty consistent, thus the STD is small.

ModelFuyu-8bInstructBLIP-13bKosmos-2mPLUG-Owl2InternVL-Chat-v1.5SphinxQwen-VL-ChatLLaVA-1.5-7bLLaVA-1.5-13bLLaVA-1.6-7bLLaVA-1.6-13bLLaVA-1.6-34bIdefics-2-8bGemini-1.0-ProClaude-3-OpusClaude-3-HaikuClaude-3-SonnetGPT-4v
STD0.747%0.586%0.808%0.096%1.209%0.517%0.373%1.015%1.134%1.06%0.273%1.295%1.493%0.702%1.724%0.947%0.465%0.326%
评论
ModelTotal ScoreDescriptionLength LimitGenresGrammarMentionMathPerspectiveOCR
GPT-4o0.9097040.9278750.9123710.9420570.8624340.9004410.8571430.9166670.882883
Claude-3-Opus0.8560770.8907210.8684900.9170700.7745900.8193860.8611110.7250000.809524
Reka0.8399050.8947520.7850880.9056430.7136610.8016670.9259260.6574070.828571
MiniCPM-Llama3-V2.50.7980230.8289160.7717950.8230870.7519440.7639760.7212640.8166670.841880
Gemini-1.0-Pro0.7735690.8174220.7354700.7889110.7978140.6830200.8660710.8703700.806373
LLaVA-1.5-7b0.6839470.7588170.7037500.6740460.6302080.6176200.4252870.8000000.602564
ShareGPT4v0.6890460.8004610.6577380.6087330.6547620.6017540.5000000.8000000.743056
Idefics-2-8b0.5417550.5602430.6193180.4892760.6468250.4553420.4055560.3750000.627193

Table 5: Details of model scores evaluated by gpt-4o-2024-05-13.

ModelTotal ScoreDescriptionLength LimitGenresGrammarMentionMathPerspectiveOCR
GPT-4o0.8994100.9093790.9162040.9693950.8548850.8612470.9202900.9074070.878378
Claude-3-Opus0.8489490.8615430.8716860.8965520.7971700.8087770.8461540.6458330.865741
Reka0.8268440.8818410.8092590.8732760.7253900.7702250.8141030.7500000.819444
MiniCPM-Llama3-V2.50.7875370.8188130.7902460.7957960.7681820.7363590.6766670.7166670.828125
Gemini-1.0-Pro0.7632400.8143790.7500000.7851590.7576820.6722550.7583330.7857140.776042
LLaVA-1.5-7b0.6604720.7518730.6618220.6498510.4985120.5727190.5166670.7500000.571429
ShareGPT4v0.6571860.7653090.6326820.5575780.5833330.5451040.4642860.6750000.717742
Idefics-2-8b0.5361340.5899640.5418870.4558820.6115820.4498210.4066670.5277780.576190

Table 6: Details of model scores evaluated by gpt-4o-2024-11-20.

Due to legal agreements the authors are obligated to adhere to, we have been requested not to use other close sourced MLLMs (eg, Clause-3-Opus, or Reka) for full rounds of evaluation. We did not use GPT-4v series because they are deprecated, and we did not use gpt-4-1106-preview, LLaMA-3.1, and Qwen2.5 as they are LLMs, but our judge needs to be a vision-language model to take in images as images are mandatory in the evaluation process. We chose not to convert the images into textual captions then use LLMs to evaluate because 1) our instructions a lot of times require MLLMs to focus on details in the images; if images are converted into captions, these details can be lost, thus the evaluation will be inaccurate, 2) MLLMs like GPT-4o and GPT-4v can serve as judges better than LLMs. As for open-source models; in our evaluation result table, we showed that their performance on MIA-Bench are not as good as proprietary models, thus not suitable as judges for this task.

We hope that we addressed your concerns. Thank you again for your review.

评论

I really appreciate the authors' sufficient empirical results, where I found that different generation configurations and judge models contribute to similar rankings.

Therefore, I will keep my rating and recommend acceptance.

评论

Q4 Towards Q4, the evaluation should incorporate more judge models rather than relying solely on Claude-3-opus. Using a variety of judge models will provide a more comprehensive and balanced assessment of the performance.

A4: The judge we use for MIA-Bench is GPT-4o. To alleviate this concern that GPT-4o may favorably score its own responses, as reported in the paper, we use Claude-3-opus, a strong performer, to evaluate responses from GPT-4o and itself, and compare their scores with each other to double check if GPT-4o is the best performing model on this benchmark. We find that even using Claude-3 Opus to score its own and GPT-4o's responses, GPT-4o still achieves a superior score. Based on this observation, we use GPT-4o for evaluation by default, to ensure the correctness of evaluation. It's common practice to use one judge model instead of multiple to evaluate inference result for efficiency. Here we list a few benchmarks that use one version of GPT as the judge: LLaVA-Bench[1] (NeurIPS 2023), MMBench[2] (ECCV 2024), MathVista[3] (ICLR 2024), HallusionBench[4] (CVPR 2024), CV-Bench[5] (NeurIPS 2024), MM-Vet[6], MMHAL-BENCH[7], MLVU[8], etc.

We conducted additional experiments following your suggestion to use different judges, namely gpt-4o-2024-11-20, gpt-4o-2024-05-13, chatgpt-4o-latest, and gpt-4o-mini-2024-07-18. The results are updated in the appendix: A.4 Comparison of Scores and Rankings across Different Judge Models. We also paste the result tables here. We evaluated eight MLLMs using these four judges. The ranking is consistent with a minor difference. (LLaVA-1.5-13b and ShareGPT4, when evaluated by gpt-4o-2024-05-13, has a different ranking order from when evaluated by other judges. This is not surprising though, as their performance is similar.)

ModelScore by chatgpt-4o-latestRanking by chatgpt-4o-latestScore by gpt-4o-2024-11-20Ranking by gpt-4o-2024-11-20Score by gpt-4o-2024-05-13Ranking by gpt-4o-2024-05-13Score by gpt-4o-mini-2024-07-18Ranking by gpt-4o-mini-2024-07-18
GPT-4o89.69189.94190.97181.361
Claude-3-Opus86.16284.89285.61278.952
Reka83.09382.68383.99377.703
MiniCPM-Llama3-V2.578.10478.75479.80473.724
Gemini75.77576.32577.36567.455
LLaVA-1.5-13b66.78666.05668.39761.546
ShareGPT4v66.61765.72768.90660.307
Idefics-2-8b53.51853.61854.18844.288

Table 3: Comparison of scores and rankings across different judge models.

ModelTotal ScoreDescriptionLength LimitGenresGrammarMentionMathPerspectiveOCR
GPT-4o0.8968930.9062880.9179960.9559520.8305080.8679490.8466670.8333330.896396
Claude-3-Opus0.8616280.8953630.8660390.9277300.8070180.8205490.8576390.6666670.800926
Reka0.8308850.8698670.8216850.8834030.7955970.7728940.8132180.6750000.848485
MiniCPM-Llama3-V2.50.7809660.8311970.7660260.7962570.7261900.7220370.6913580.6562500.768018
Gemini-1.0-Pro0.7577330.7938600.7241380.7454550.8541670.6703490.8160560.8229170.822581
LLaVA-1.5-7b0.6678260.7431370.6388890.6758270.5712120.5945050.5000000.7583330.596774
ShareGPT4v0.6660920.7739050.6612900.5739040.5625000.5707220.4583330.6388890.695238
Idefics-2-8b0.5350570.5979630.5318100.4837680.5930560.4523610.3260870.4583330.569444

Table 4: Details of model scores evaluated by chatgpt-4o-latest.

审稿意见
6

This paper introduces MIA-Bench, a benchmark designed to evaluate multimodal large language models (MLLMs) on strict adherence to complex instructions. Based on a dataset of 400 image-instruction pairs, MIA-Bench assesses the performance of 29 popular MLLMs, highlighting that current MLLMs struggle with precise instruction adherence. In addition, the paper also explored the supervised fine-tuning (SFT) method based on the LLaVA-NeXT model and achieved positive results, demonstrating the potential effectiveness of this method in improving model instruction adherence.

优点

(1) MIA-Bench fills a gap in existing benchmarks by focusing on the ability of multimodal large language models (MLLMs) to adhere to complex instructions, a crucial yet previously underexplored area of evaluation. This benchmark’s design reveals potential deficiencies in models when following strict, multi-layered instructions, supporting the practical deployment of multimodal models in complex, instruction-based tasks. (2) The MIA-Bench dataset consists of 400 image-instruction pairs covering various instruction types, such as description, length limit, genre, grammar, and OCR. With data from diverse sources, it reflects the variety found in real-world scenarios. This diverse set of instructions enhances the benchmark’s comprehensiveness and real-world applicability. (3) The paper provides a systematic evaluation of 29 popular MLLMs, analyzing their performance across different instruction categories. This large-scale comparison offers researchers a detailed performance reference and targeted insights for future model improvements.

缺点

(1) Although the SFT experiments in the paper demonstrate improved performance in adhering to complex instructions, they may lack in terms of model generalization. The results may be specific to the LLaVA-NeXT model, with no validation of applicability to other models, leaving it unproven whether SFT is equally effective for enhancing complex instruction adherence in MLLMs of different architectures and sizes. (2) The differences between MIA-Bench and other benchmarks are indeed striking, and the authors believe that this may indicate that MIA-Bench's unique design focuses more on evaluating the model's strict instruction-following ability. However, the authors' explanation does not completely rule out the possibility that MIA-Bench itself may have design biases. This explanation is based on speculative experimental results and has not been verified in depth by sufficient experiments. (3) Although the authors introduced Claude-3-Opus for comparison to mitigate scoring bias, the two models may exhibit similar biases in evaluation. Therefore, using Claude-3-Opus as the sole comparison tool may be insufficient to fully reveal potential scoring biases in GPT-4o.

问题

please refer to the weakness part.

评论

Q3. Although the authors introduced Claude-3-Opus for comparison to mitigate scoring bias, the two models may exhibit similar biases in evaluation. Therefore, using Claude-3-Opus as the sole comparison tool may be insufficient to fully reveal potential scoring biases in GPT-4o.

A3. In our evaluation, we use GPT-4o as the default judge model because of two reasons: The most widely recognized free-form evaluation benchmarks currently adopt ChatGPT-series models as their judge, as they represent the state-of-the-art MLLMs available. Examples of such benchmarks include LLaVA-Bench [1] (NeurIPS 2023), MMBench [2] (ECCV 2024), MathVista [3] (ICLR 2024), HallusionBench [4] (CVPR 2024), and CV-Bench [5] (NeurIPS 2024), etc. To align with this common practice, we have chosen GPT-4o as the default evaluation model.

To further assess the reliability of the scoring, as outlined in Section 3.2, we employ the second-highest-performing model on MIA-Bench, Claude-3 Opus, to evaluate its own responses alongside those of GPT-4o. Interestingly, Claude-3 Opus also favors the GPT-4o responses over its own on MIA-Bench. This preliminary experiment demonstrates consistency across these two judge models.

We sincerely appreciate your constructive feedback, which will help improve the quality and depth of our work. Thank you again for your time and valuable insights.

References: [1] Haotian Liu, Chunyuan Li, Qingyang Wu, and Yong Jae Lee. Visual instruction tuning, NeurIPS 2023. [2] Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, and Dahua Lin. Mmbench: Is your multi-modal model an all-around player?, ECCV 2024. [3] PanLu,HritikBansal,TonyXia,JiachengLiu,ChunyuanLi,HannanehHajishirzi,HaoCheng,Kai-Wei Chang, Michel Galley, and Jianfeng Gao. Mathvista: Evaluating mathematical reasoning of foundation models in visual contexts, 2024. [4] Tianrui Guan, Fuxiao Liu, Xiyang Wu, Ruiqi Xian, Zongxia Li, Xiaoyu Liu, Xijun Wang, Lichang Chen, Furong Huang, Yaser Yacoob, Dinesh Manocha, and Tianyi Zhou. Hallusionbench: An advanced diagnostic suite for entangled language hallucination and visual illusion in large vision-language models, 2024. [5] Shengbang Tong, Ellis Brown, Penghao Wu, Sanghyun Woo, Manoj Middepogu, Sai Charitha Akula, Jihan Yang, Shusheng Yang, Adithya Iyer, Xichen Pan, Austin Wang, Rob Fergus, Yann LeCun, and Saining Xie. Cambrian-1: A fully open, vision-centric exploration of multimodal llms, 2024.

AC 元评审

This paper introduces a new MIA-Bench benchmark specifically designed for MLLM's complex instruction following ability study. The authors' major contribution is a new dataset containing 400 image-instruction pairs, where the instructions are written by humans with high quality. Besides, the authors provided a comprehensive evaluation of the existing MLLMs and provided SFT positive results. The weaknesses and concerns are the model generalization issue, the MIA-Bench's design biases, the diverse model judgment issue, and the potential data leakage issue. After rebuttal, the authors addressed most of these concerns, and all reviewers agreed with the acceptance. Meanwhile, the value of the new benchmark with 400 high-quality image-instruction pairs is highlighted. Therefore, the final recommendation is accept.

审稿人讨论附加意见

This submission received four reviews. Reviewer drhM's major concerns were about the SFT model generalization issue, the MIA-Bench's design biases issue, and the solo Claude-3-Opus judgment issue. The authors provided one-by-one responses to these questions with new experimental results. The reviewers did not provide final comments, but the initial score is a marginal acceptance. Reviewer yNxN's concerns were mainly about the robustness of the evaluation. The authors provided comprehensive new results according to the reviewer's suggestions. The reviewer's final decision is a marginal acceptance. Reviewer PupP's major concern was the potential data leakage issue. The authors acknowledged that images may be covered by previous models, but emphasized their instructions are of high quality and written by humans, especially composed of multiple levels of instructions. The reviewer accepted the explanation and improved the score from 5 to 6. Reviewer UieD's concerns were about the value of complex instruction for MLLMs and the issue of performance. After rebuttal, the reviewer appreciated the contribution of the new dataset and improved the score from 5 to 6. Considering all reviewers' opinions and the discussion, the final recommendation is accept.

最终决定

Accept (Poster)