PaperHub
5.8
/10
Poster4 位审稿人
最低5最高6标准差0.4
6
6
6
5
4.0
置信度
正确性3.0
贡献度2.5
表达2.8
NeurIPS 2024

MindMerger: Efficiently Boosting LLM Reasoning in non-English Languages

OpenReviewPDF
提交: 2024-05-13更新: 2025-01-09
TL;DR

Merging the built-in capabilities of LLMs and external capabilaties of multilingual model to boost multilingual reasoning.

摘要

关键词
Large Language ModelMultilingualReasoning

评审与讨论

审稿意见
6

The paper introduces MergeMinds, a method that integrates LLMs with multilingual models to enhance reasoning capabilities across multiple languages. Through a two-step training scheme, MergeMinds effectively boosts performance in multilingual reasoning and language understanding tasks, particularly excelling in low-resource languages, as demonstrated by significant accuracy improvements on various datasets.

优点

  1. This work propose MergeMinds method offers a novel solution by merging LLMs with external multilingual capabilities and wo-step training process, enhancing the robustness of the model.
  2. The method shows substantial improvements in multilingual reasoning and language understanding tasks, particularly with notable gains in low-resource languages, demonstrating the effectiveness of the approach.

缺点

  1. From my perspective, the motivation of MergeMinds is extremely similar with LangBridge, both of which ultilize the mapping layer to map multilingual encoder to exsisting MLLM.
  2. In fact, the work of this paper has already been explored on multi-modal large models a long time ago, such as InstructBlip's Qformer and LLaVA's two-stage alignment training.

问题

  1. It appears that the Te language is not included in the MGSM. Could you explain the reason for this omission?
  2. I noticed that some work similar to xcot was referenced but not included in the comparative table. What was the rationale behind this decision?
  3. In my view, does this work adapt the paradigm from InstructBlip's Qformer and LLaVA's two-stage alignment training to a multilingual setting?
  4. Missing Reference [1-3]

[1] MAPO: Advancing Multilingual Reasoning through Multilingual-Alignment-as-Preference Optimization

[2] The Power of Question Translation Training in Multilingual Reasoning: Broadened Scope and Deepened Insights

[3] Multilingual large language model: A survey of resources, taxonomy and frontiers

局限性

The authors have adequately addressed the limitations.

作者回复

Thank you for your encouragement of our work. We reply to your questions as follows.

W1: From my perspective, the motivation of MergeMinds is extremely similar with LangBridge, both of which utilize the mapping layer to map multilingual encoder to existing MLLM.

  • In related works about MLLM, LLM does not have visual capabilities at all. Compared with this, the most significant difference in our research is that LLM has built-in multilingual capabilities, .
  • Based on this difference, we need to explore whether there is a more efficient training solution than MLLM to effectively activate the multilingual capabilities built-in LLM.
  • LangBridge is a solution like the works in MLLM that only considers integrating encoder to obtain capabilities that are not available in LLM. LangBridge does not consider boosting the built-in multilingual capabilities of LLM, which results in its performance being significantly lower than MindMerger. It lags behind MindMerger by 8.0% and 7.1% in low-resource languages and all languages in the MGSM dataset. Therefore, different from the existing work of MLLM, with MindMerger as a starting point, there are still many challenges to explore in the future to stimulate the multilingual as well as other bulit-in capabilities of LLM.

W2, Q3: The work of this paper has already been explored on multi-modal large models. Does this work adapt the paradigm from InstructBlip's Qformer and LLaVA's two-stage alignment training to a multilingual setting?

  • Because the focus of our work is on how to combine multilingual capabilities from external model and LLM itself, we face the challenge of avoiding over-reliance on the capabilities of the external multilingual model during training. Similarly, we also face the challenge of avoiding over-reliance on the multilingual capabilities built into LLM during training.
  • Moreover, our goal is to activate the multilingual capabilities of the LLM rather than to increase new capabilities it does not have, it is possible to explore a more efficient training solution than the MLLM. It can be observed from Table 6 in the paper that MindMerger has surpassed all baselines under the using of only 5,000 training samples per language at the second stage.
  • In addition, the research on integrating multilingual models into LLM is still under-explored and different from the existing work on MLLM. For example, what data to use for training, what structure to use for integration, etc. still need more insightful discussions.
MGSMTeBnThSwJaZhDeFrRuEsEnLrlHrl
xCoT42.840.449.248.450.050.047.249.650.048.848.445.249.1
LangBridge34.842.850.443.240.045.250.852.456.458.063.242.852.3
MindMerger52.054.054.055.651.654.860.059.657.663.265.253.958.9

Q1: It appears that the Te language is not included in the MGSM. Could you explain the reason for this omission?

  • This is not a technical issue, since the QAlign and MathOctopus work we followed did not compare Te, so we ignored Te as well.
  • We added the Te experiment in the table above. MindMerger is still the best model on Te, with a 9.2% improvement over xCoT.

Q2: I noticed that some work similar to xCoT was referenced but not included in the comparative table. What was the rationale behind this decision?

  • We selected the two strongest methods for each of the two categories of baselines: relearning-based and replacement-based.
  • xCoT is a relearning-based method, but its performance is lower than MultiReason (MultiReason-Lora) and QAlign. Considering the page limit of the paper, we did not include it in the comparison.
  • We show the performance of xCoT in the table above, and we consider adding it to the appendix.

Q4: Missing Reference.

  • Thank you for the reviewers’ kind reminder, we will cite them in the next version of paper.

Best wishes.

评论

Thanks for your great response. I have no additional problems.

审稿意见
6

This paper proposes a framework which merges LLMs with the external language understanding capabilities from multilingual models to improve multilingual reasoning performance. Specifically, the authors introduce a two-step training scheme: they (i) train the framework to embed the multilingual model into the LLM by using translation data, and (ii) further train the framework to collaboratively utilize the built-in capabilities of LLM and the embedded external language capabilities.

优点

  • This paper proposes a new method to boost the multilingual reasoning of LLMs, which embeds external language understanding capabilities from multilingual models into LLMs.
  • Extensive experimental results and analyses on several datasets demonstrate the effectiveness of the proposed method.

缺点

  • In Section 3.2, the authors mention that they use translation data and query translation task data generated from public translation models for the two-stage training; while in lines 182-192, it shows that they also use several English task datasets, such as MetaMathQA. It is unclear how they use these data.
  • Although the author mentioned that only a small number of parameters are trained during two-stage training, the paper lacks a comparison with baselines in this apsect. Particularly, the proposed method requires the external multilingual model, which is expected to affect the inference efficiency.
  • I know asking for new results might be unrealistic at this stage, but I am curious what would happen if you only used two-stage training for LLM without including an external multilingual model.

问题

See above.

局限性

See above.

作者回复

Thank you for the valuable reviews. We have added some experiments and will include them in the next version of paper.

W1. In Section 3.2, the authors mention that they use translation data and query translation task data generated from public translation models for the two-stage training; while in lines 182-192, it shows that they also use several English task datasets, such as MetaMathQA. It is unclear how they use these data.

  • For all models, we first use English task datasets to fully fine-tune LLM in order to obtain the reason capabilities, and then use other data to train.
  • Taking the mathematical reasoning task as an example, both our model and all baselines are based on MetaMath-Llama (denote as MonoReason in our paper) as the base model, and MetaMath-Llama is fully fine-tuning Llama on the MetaMathQA dataset.

W2. Although the author mentioned that only a small number of parameters are trained during two-stage training, the paper lacks a comparison with baselines in this apsect. Particularly, the proposed method requires the external multilingual model, which is expected to affect the inference efficiency.

  • Compared with the parallel encoding input query, the serial generation time is much longer.
  • As report in the follow table, compared with the single LLM model QAlign under batch size 1, the average inference speed for MindMerger even faster than QAlign, where MindMerger spends an average of 3.65s inferring a sample while QAlign uses 3.81.
Inference time per sample (s)BnThSwJaZhDeFrRuEsEnAvg.
QAlign3.774.213.574.094.133.843.613.623.793.483.81
MindMerger3.883.573.643.533.713.493.473.943.743.553.65
  • This phenomenon is caused by the length of the text generated by QAlign is slightly larger than that of MindMerger as reported as belows:
The average length of generated text (token)BnThSwJaZhDeFrRuEsEnAvg.
QAlign126.30139.94118.34135.86135.94126.52120.20119.76128.24117.60126.87
MindMerger122.32121.42121.94121.78126.74117.56116.30133.22125.84120.34122.75
  • As for the cost of the external multilingual model, it is very insignificant. As reported in the below table, due to its small number of parameters, the time used to encode the query of external model is only half of QAlign (0.02s vs 0.04s).
Encoding time per sample (s)BnThSwJaZhDeFrRuEsEnAvg.
QAlign0.050.040.040.030.030.030.040.030.030.030.04
Multilingual Encoder + Mapping Layer0.020.020.020.020.020.020.020.020.020.020.02
评论

W3. I know asking for new results might be unrealistic at this stage, but I am curious what would happen if you only used two-stage training for LLM without including an external multilingual model.

  • There are three types of training data in the paper: (a) translation data, (b) English task data, and (c) query translation task data.
  • We experimented with four dataset usage settings (b, c, b->c, c->b) to fully fine-tune Llama and compare the influence of using translation data for training in the first stage (a->b, a->c, a->b->c, a->c->b).
  • The experimental results are shown in the table below. Regardless of the way the training data is used, MindMerger has an average accuracy that is at least 6.2 higher than the fully fine-tuned method.
  • Using translation data as the first stage training data can help improve performance on 3 out of 4 fully fine-tuning settings (a->c, a->b, and a->c->b).
  • The gain of translation data for MindMerger is significantly greater than that of fully fine-tuning Llama. With the help of translation data, MindMerger improved the average accuracy rate by 5.5%, while the improvements for fully fine-tuning Llama were 0.8%, 3.8% and 1.0% respectively.
BnThSwJaZhDeFrRuEsEnLrl.Hrl.Avg.
c33.236.834.836.037.239.639.638.437.640.434.938.437.4
a->c36.436.037.231.634.442.440.440.442.441.236.539.038.2
b (MonoReason)6.87.26.836.438.455.254.452.057.268.86.951.838.3
a->b21.221.224.438.834.452.856.852.457.661.622.350.642.1
b->c (MultiRason)33.240.042.042.042.045.244.845.248.052.038.445.643.4
a->b->c33.640.440.441.237.638.441.244.441.647.638.141.740.6
c->b40.846.445.244.850.452.852.852.056.459.644.152.750.1
a->c->b40.850.050.447.652.854.854.056.050.054.847.152.951.1

Thanks again to the reviewer for the experimental suggestions. We believe that these additional experiments will help further enhance the impact of our work.

Best wishes.

评论

Dear authors,

Thanks for your response and update with the results! I decided to raise the soundness score.

审稿意见
6

The paper propose a way to improve multilingual reasoning in LLM in own native languages without relying on pivoting methods such as translating to English. The method assume a hypothesis that LLMs have built in knowledge and reasoning abilities in a lower-resource language and not just common language like English. The method plugs in LLMs with the external multilingual models as embedding layers and introduce a training scheme build those new knowledge it. Experiments on MGSM show better accuracy across languages.

优点

  • The method is pretty novel in ways of incorporating multilingual encoders as additional embedding layers to augment existing embedding layers of the LLMs.
  • The experiments are done with many languages comprehensively, including many low-resource languages, and ablation studies, which are good.

缺点

  • The method is incomplete and ambiguous on the generation side. It makes sense that injecting multilingual embeddings at the bottom for the input may help the LLM understand the context better, but that does not necessarily make the LLM generate in those native language better, especially if the LLMs are weak in producing content in native languages.

    • It is no surprise that the tasks presented in the paper are all "understanding" tasks where the generation workload is light and trivial. Other "generation" tasks, such as summarization, writing, translation... may not fair well with this method. I would highlight if the authors could show good results for the method for these tasks.
  • Missing baseline comparison: https://arxiv.org/abs/2306.11372 , https://aclanthology.org/2023.findings-emnlp.826.pdf

  • Confusing word choices, such as "mapping stage", "merging" and "augmentation stage". These words often lead to different understanding rather than what are described in the paper.

  • "Query translation task" and "query translation data" and many parts of the methodology are not explained clearly. It would be better if there are explicit visual examples or diagrams.

问题

NA

局限性

The paper discussed limitations

评论

W3. Confusing word choices, such as "mapping stage", "merging" and "augmentation stage". These words often lead to different understanding rather than what are described in the paper.

  • Thank you for your feedback, we will fine-tune the wording and description in the next version based on your comments.

W4. "Query translation task" and "query translation data" and many parts of the methodology are not explained clearly. It would be better if there are explicit visual examples or diagrams.

  • Thank you for your feedback again. We followed the settings of QAlign to translate the query from English to non-English automatically as training data. Below are three examples of query translation task data:
Augmentation stage (Query translation task data)
MathInput (Zh): 伯特每天都填报纸上的每日填字游戏。他每两周就用完一支铅笔。平均而言,他用完一支铅笔需要1050个字。每个填字游戏平均有多少个字? <br> Input (En): Bert fills out the daily crossword puzzle in the newspaper every day. He uses a pencil to fill out the puzzles every two weeks. On average, it takes him 1050 words to use up a pencil. How many words are in each crossword puzzle on average? <br> Training Target: If Bert uses up a pencil to fill out the puzzles every two weeks and it takes him 1050 words to use up a pencil ... 1050/14 = 75 words in each daily crossword puzzle on average. #### The answer is: 75
CommonsenseInput (Zh): 酒后驾车的人可能会被指控什么? (A)惩罚 (B)逮捕 (C)车祸 (D)胡言乱语 (E)酒后驾驶 <br> Input (En): What is someone operating a vehicle likely to be accused of after becoming inebriated? (A) punish (B) arrest (C) automobile accidents (D) talking nonsense (E) drunk driving <br> Training Target: E
NLIInput (Zh): Premise: 她不太明白。Hypothesis: 事实上,她没有理解 <br> Input (En): Premise: She doesn’t really understand. Hypothesis: Actually, she doesn’t get it. <br> Output: Entailment

If you have any questions about our work, we look forward to discussing with you and, if possible, hope that you can improve our score appropriately.

Best wishes.

评论

W2. Missing baseline comparison XLT.

  • We have discussed XLT in related work, lines 80-81, which is a prompt-based approach that brings limited improvements over open source models such as Llama through carefully crafted prompts.
  • Our model and the baselines we compared are all supervised fine-tuning methods. Although the work of XLT is inspiring, prompt-based methods are not in the same category works as our model.
  • We added a comparison between XLT and our model. The experimental results are shown in the following two tables. Under the same base model, MindMerger has a huge lead over XLT.
  • Due to page limitations, we consider adding the experimental results of XLT to the appendix.
BnThSwJaZhDeFrRuEsEnLrlHrlAvg.
XLT (Llama2-chat)3.25.64.09.212.816.017.614.019.619.64.315.512.2
MindMerger (Llama2)50.452.857.254.453.661.257.660.858.466.853.559.057.3
BnThSwJaZhDeFrRuEsEnLrlHrlAvg.
XLT (Llama3-Instruct)27.653.632.452.058.053.639.650.853.252.837.951.447.4
MindMerger (Llama3)64.465.666.462.068.071.672.473.272.875.265.570.769.2
作者回复

Thank you for your reviews.

The main concern is the performance of our model on the generation task. We believe that adding experiments on the generation task will help further expand the scope of our work.

We are grateful for the feedbacks on some writing improvement and we will revise them in the next version.

The specific responses are as follows:

W1. The method is incomplete and ambiguous on the generation side. It makes sense that injecting multilingual embeddings at the bottom for the input may help the LLM understand the context better, but that does not necessarily make the LLM generate in those native language better, especially if the LLMs are weak in producing content in native languages.

  • Our model can be a key step towards improving LLM multilingual generation capabilities, since language understanding is the basis of language generation.

  • To response this concern, we added an experimental comparison on the translation task based on Flores-101 dataset. We implemented four settings: X-En, En-X, X-Zh, and Zh-X. In each setting, we compare fully fine-tuned Llama2 (Llama-SFT) with MindMerger using 100 K and 200 K training samples per language, respectively.

  • The experimental results are shown in the following four tables. MindMerger consistently outperforms Llama-SFT in translation quality under all comparison settings.

  • Considering that the top of MindMerger is frozen and not trained, the improvement in generation quality comes from the enhancement of query understanding, which may be a valuable insight for future research on multilingual generation tasks.

X-EnBnThSwJaZhDeFrRuEsLrl.Hrl.Avg.
Llama-SFT (100 K)14.89.622.716.414.628.726.325.020.715.722.019.9
MinderMerger (100 K)18.616.732.923.322.237.432.631.725.022.728.726.7
Llama-SFT (200 K)16.018.827.920.223.638.139.330.428.720.930.127.0
MinderMerger (200 K)27.124.540.226.028.942.542.434.130.130.634.032.9
En-XBnThSwJaZhDeFrRuEsLrl.Hrl.Avg.
Llama-SFT17.18.220.813.16.218.919.712.616.715.414.514.8
MinderMerger20.211.930.119.212.427.632.223.223.720.723.122.3
Llama-SFT (200 K)23.915.626.618.916.725.135.621.224.322.023.623.1
MinderMerger (200 K)24.417.432.623.020.731.640.826.626.124.828.127.0
X-ZhBnThSwJaDeFrRuEsEnLrl.Hrl.Avg.
Llama-SFT (100 K)7.05.76.912.915.713.514.311.616.36.514.111.5
MinderMerger (100 K)10.18.910.316.018.718.017.415.821.29.817.915.2
Llama-SFT (200 K)8.57.58.213.214.714.413.611.113.18.113.411.6
MinderMerger (200 K)11.811.712.017.320.520.019.017.023.911.819.617.0
Zh-XBnThSwJaDeFrRuEsEnLrl.Hrl.Avg.
Llama-SFT (100 K)9.93.67.312.911.514.79.611.012.76.912.110.4
MinderMerger (100 K)14.17.813.919.418.824.817.517.725.911.920.717.8
Llama-SFT (200 K)14.27.811.316.517.720.514.115.718.611.117.215.2
MinderMerger (200 K)16.211.416.220.720.226.618.718.327.214.622.019.5
评论

Dear reviewer -- could you see whether the author response addressed your concern, especially about your comment on how their method can improve generation of LLMs? they provided new experiments, as well as a comparison with baseline you proposed. If you can modify the scores based on the response, that'd be very helpful!

评论

Thanks for the author response, I pump up the scores for the effort!

审稿意见
5

The paper suggests incorporating an embedding block using external multilingual models to improve the models' understanding. Additionally, comprehensive experiments are conducted to demonstrate its efficacy.

优点

  1. The paper proposes a straightforward and easily implementable method to enhance models' multilingual capability.
  2. Experimental results validate its effectiveness.

缺点

The paper lacks a clear understanding of why the method can work.

  1. To my understanding, the Embedding Layer of the model merely converts the query from text space to the token/embedding space. This transformation may not be interpreted as "understanding the input" since the information hasn't been operated and combined by the attention layer. However, XX after the multilingual model has been "understood" by the model. Therefore, it does not make much sense to concatenate these two embeddings.
  2. As X~\tilde{X} contains the information of query q, TT includes duplicated information. This may be why the method can surpass baselines such as QAlign and Translate-En. It would be more convincing if baselines also input duplicated information.
  3. This is further proved by Figure (4b), which illustrates that the mapping layer tends to obtain a unified representation for different languages and this is the same as QAlign. Therefore, it is confusing why this method can outperform QAlign.
  4. The paper uses Llama2 as the base model. However, its multilingual performance is far from sota. Until submission, I believe Mistral, Gemma, and Llama3 have been released, whose multilingual performance is much better.

问题

The details of the two training stages lack clarity, specifically regarding the utilization of translation and parallel data, as well as the precise training tasks involved.

局限性

The authors suggest a limitation on the integration of external multilingual models with the model's inherent understanding capabilities, a concern partly explored through the ablation study.

评论

Q1. The details of the two training stages lack clarity, specifically regarding the utilization of translation and parallel data, as well as the precise training tasks involved.

  • Thank you for your feedback, we will improve the description of training data in the next version of our paper.
  • In the first stage (mapping stage), all tasks including Math, Commonsense, and NLI are trained using non-English to English translation data. The following is an example:
Mapping stage (Translation data)
Input (Zh): 文档中没有元素属于该组时,该名称会被作为根结点显示在结构树上。
Training Target: It will be shown in the structure tree as a top node when there are no elements belonging to this group in the document.
  • In the second stage (augmentation stage), the input query of each task is translated into non-English to construct multilingual training samples, both English and non-English samples are used to train our model. Below are examples of training samples for three tasks:
Augmentation stage (Query translation task data)
MathInput (Zh): 伯特每天都填报纸上的每日填字游戏。他每两周就用完一支铅笔。平均而言,他用完一支铅笔需要1050个字。每个填字游戏平均有多少个字? <br> Input (En): Bert fills out the daily crossword puzzle in the newspaper every day. He uses a pencil to fill out the puzzles every two weeks. On average, it takes him 1050 words to use up a pencil. How many words are in each crossword puzzle on average? <br> Training Target: If Bert uses up a pencil to fill out the puzzles every two weeks and it takes him 1050 words to use up a pencil ... 1050/14 = 75 words in each daily crossword puzzle on average. #### The answer is: 75
CommonsenseInput (Zh): 酒后驾车的人可能会被指控什么? (A)惩罚 (B)逮捕 (C)车祸 (D)胡言乱语 (E)酒后驾驶 <br> Input (En): What is someone operating a vehicle likely to be accused of after becoming inebriated? (A) punish (B) arrest (C) automobile accidents (D) talking nonsense (E) drunk driving <br> Training Target: E
NLIInput (Zh): Premise: 她不太明白。Hypothesis: 事实上,她没有理解 <br> Input (En): Premise: She doesn’t really understand. Hypothesis: Actually, she doesn’t get it. <br> Output: Entailment

Thank you again for your comments. We hope our responses can resolve your concerns. We hope you can improve our score appropriately and are willing to have more discussions with you.

Best wishes.

评论

W3. This is further proved by Figure (4b), which illustrates that the mapping layer tends to obtain a unified representation for different languages and this is the same as QAlign. Therefore, it is confusing why this method can outperform QAlign.

  • This reflects that our model has better capability to obtain unified representation than QAlign.
  • In Table 9 of the paper, we calculated the cosine similarity and top-1 retrieval recall between representations of the same query expressed in different languages to quantify the model’s capability to obtain a unified representation. The results show that our model significantly improves LLM’s capability to obtain a unified representation.
  • We further compare the last layer pooling representation of QAlign and our method in the following table. It can be observed that although QAlign slightly outperforms MonoReason, the score of cosine similarity and recall@1 are still far behind our model, which proves that our model is better able to obtain unified representations.
Cosine SimilarityBn->EnTh->EnSw->EnJa->EnZh->EnDe->EnFr->EnRu->EnEs->EnLrlHrlAvg.
MonoReason0.080.07-0.080.110.130.040.120.050.090.020.090.07
QAlign0.070.04-0.060.110.140.100.180.060.150.020.120.09
MindMerger0.320.320.460.460.460.550.620.520.560.370.530.48
Recall@1 (%)Bn->EnTh->EnSw->EnJa->EnZh->EnDe->EnFr->EnRu->EnEs->EnLrlHrlAvg.
MonoReason0.41.80.25.812.611.547.344.927.00.824.916.8
QAlign2.84.90.512.326.121.261.737.341.02.733.323.1
MindMerger52.876.064.391.696.498.099.298.188.964.495.485.0

W4. The paper uses Llama2 as the base model. However, its multilingual performance is far from sota. Until submission, I believe Mistral, Gemma, and Llama3 have been released, whose multilingual performance is much better.

  • In fact, we have compared Mistral as well as the larger MetaMath-Llama-13B in Table 5. The experimental results show that our method is also consistently effective on LLM with stronger multilingual capabilities.
  • We further compare our method with baselines based on Llama3 in the following table. Our method still outperforms all baselines.
Llama3, MGSMBnThSwJaZhDeFrRuEsEnLrlHrlAvg.
MonoReason40.453.232.053.258.064.467.267.669.276.441.965.158.2
MonoReason41.255.629.250.056.862.464.466.466.074.842.063.056.7
MultiReason-Lora52.062.052.458.864.866.068.872.474.075.255.568.664.6
MultiReason-SFT39.247.648.048.046.452.848.047.252.857.244.950.348.7
QAlign50.059.656.054.058.462.463.669.670.873.255.264.661.8
LangBridge49.251.656.448.054.869.668.467.669.678.052.465.161.3
Translate-En52.041.658.054.853.263.660.455.667.676.450.561.758.3
MergeMinds64.465.666.462.068.071.672.473.272.875.265.570.769.2
作者回复

We thank the reviewer for taking the time to provide reviews. However, we would like to clarify some misunderstandings as follows:

W1. To my understanding, the Embedding Layer of the model merely converts the query from text space to the token/embedding space. This transformation may not be interpreted as "understanding the input" since the information hasn't been operated and combined by the attention layer. However, after the multilingual model has been "understood" by the model. Therefore, it does not make much sense to concatenate these two embeddings.

  • Our model is based on the finding that providing LLM with inputs from two representation spaces can improve LLM's multilingual reasoning capability, where one representation comes from the native space of LLM's embedding layer, and the other representation is the mapping of multilingual queries to unified space provided by the multilingual model.

  • The two different types of representations are the key to boost better understanding of the query in subsequent layers of LLM.

W2. As contains the information of query q, includes duplicated information. This may be why the method can surpass baselines such as QAlign and Translate-En. It would be more convincing if baselines also input duplicated information.

  • We duplicated the input for all baselines during training and inference. As shown in the below table, the performance of 4 out of 6 baselines decreased with duplicated input.
  • Note that our model outperforms all baselines by at least 6.7% and 8.0% across all languages and low-resource languages, which far exceeds the improvement of the baseline by duplicated input. Therefore, it can be considered that duplicate input is not the mechanism for improving the performance of our model.
MetaMath-Llama-7B, MGSMBnThSwJaZhDeFrRuEsEnLrlHrlAvg.
MonoReason6.87.26.836.438.455.254.452.057.268.86.951.838.3
MonoReason (duplicated input)8.47.24.833.242.853.653.653.652.063.26.850.337.2
MultiReason-Lora29.635.228.052.054.859.658.462.459.664.830.958.850.4
MultiReason-Lora (duplicated input)28.843.233.650.454.857.657.661.262.062.835.258.151.2
MultiReason-SFT33.240.042.042.042.045.244.845.248.052.038.445.643.4
MultiReason-SFT (duplicated input)31.241.239.242.047.649.652.049.648.858.846.049.846.0
QAlign32.439.640.444.048.454.856.852.459.668.037.554.949.6
QAlign (duplicated input)30.838.040.444.446.455.655.256.857.665.249.054.549.0
LangBridge42.850.443.240.045.256.450.852.458.063.245.552.350.2
LangBridge (duplicated input)34.042.842.033.235.653.652.051.257.261.239.649.146.3
Translate-En48.437.637.649.246.860.456.447.659.665.541.255.150.9
Translate-En (duplicated input)47.633.638.448.448.856.049.244.854.863.239.952.248.5
MindMerger52.053.454.059.061.764.164.063.365.067.753.163.560.4
评论

Dear reviewer, could you check at the authors new experimental results and see whether it address your concerns, and update the score accordingly if they do?

评论

Thanks so much for so detailed reply in rebuttal to address my concern. Thay addressed my concerns, and I have raised my score accordingly.

最终决定

The paper presents a method to improve multilingual capability of LLMs by providing backbone LLMs the undecided query representation from multilingual model to LLM that is more optimized for reasoning. The proposed merging recipe includes two stages (1) mapping multilingual model into the backbone LLM and (2) teaching models to perform task given mapped multilingual model’s representation and backbone LLM’s representation.

The reviewer (iF4J) have asked for more thorough evaluation against with the baselines, making input setting more equivalent and trying different base LLMs, and the proposed method consistently shows gains in these new settings. The approach is evaluated throughly on various benchmarks, including generation benchmark (suggested by reviewer q66P) and efficiency comparison (suggested by reviewer VBLh). These results strengthens the value of the proposed approach.