MPO: An Efficient Post-Processing Framework for Mixing Diverse Preference Alignment
摘要
评审与讨论
This paper proposes MPO, an efficient post-processing framework for mixing diverse preference alignment. The authors use batch stochastic mirror descent to find the optimal combination coefficients for output combination.
update after rebuttal
Most of the concerns are resolved, so the reviewer raises the score to weak accept.
给作者的问题
-
According to Algorithm 1, do the authors simply combine the output logits of different LLMs? From Figure 1, the reviewer thinks the authors conduct something like model merging.
-
The reviewer thinks the experimental evaluation of the method is not enough.
-
In Table 1, do the authors adopt provided by others or train them by yourselves? If the authors train these models by yourselves, the reviewer thinks the authors should add comparison results with models provided by others. For example, comparison with PKU-SafeRLHF [1] in terms of helpfulness and harmlessness.
-
If we consider the win rate of the reference model as 50%, the aligned model is only marginally better than the baseline. Are such results reasonable?
-
Following the above point, in Table 1, a large leads to better results, this is quite uncommon because general alignment algorithms like PPO usually use a small constraint. Can we think a large represents a similar model as the original one? If so, what is the role of the alignment method during the process?
-
-
In practice, the reviewer thinks it is quite difficult for us to serve multiple LLMs at the same time. Especially when the model is large. If the authors adopt the model merging method, this should not be a problem.
[1] https://huggingface.co/datasets/PKU-Alignment/PKU-SafeRLHF
论据与证据
Yes.
方法与评估标准
Yes.
理论论述
The reviewer assumes all proofs are correct.
实验设计与分析
Yes.
补充材料
No.
与现有文献的关系
This paper proposes a post-processing method for diverse preference optimization. From the reviewer's perspective, it is novel compared to previous scientific literature.
遗漏的重要参考文献
None.
其他优缺点
-
Strengths
- The proposed method is training-free. Theoretically, the authors can combine several existing well-aligned LLMs to get the desired LLM without any cost training process.
- The authors provide theoretical guarantees for the proposed method.
-
Weaknesses
- Lack of comprehensive evaluation of the proposed method.
- The reviewer doubts the feasibility of serving multiple LLMs simultaneously in practice.
其他意见或建议
None.
We greatly appreciate your constructive and insightful feedback! Here we provide a detailed response to address all of your concerns below.
Confusion of Algorithm 1 and Figure 1.
Thank you for the question and apologies for any confusion. As shown in Thm3.4, we have which can be seen as a logit-level aggregation of multiple LLMs. However, our main goal is to compute that maximizes the minimum reward, which balances the trade-offs among different objectives. Based on your comment, we will revise Figure 1 to better highlight the role of in shaping the final policy.
Evaluation of in Table 1.
Thank you for your feedback.
- in Table 1 were trained by our own, and the results demonstrate a significant surplus in the corresponding rewards—indicating the effectiveness of our single-objective policies.
- However, our primary goal is to balance multiple objectives rather than to optimize individual objectives. Given the differences in training data, we believe that direct comparisons of single-objective policies may not be entirely fair.
- Nonetheless, we have added further comparisons with the PKU-SafeRLHF model with MPO under in Eq.10. As shown in FIgure 4, our mostly rely on and , which aligns closely with the objectives considered in the PKU-SafeRLHF. The results below show that MPO still achieves the highest minimum win rate.
Table : Win rate(%) against the Reference Model
| Model | Helpful | Harmless | Humorous | Min |
|---|---|---|---|---|
| 53.1 | 40.8 | 56.1 | 40.8 | |
| 46.3 | 53.1 | 54.1 |
- Moreover, the normalized reward for and are listed below:
| Model | |||
|---|---|---|---|
| -0.176 | 0.564 | 0.104 | |
| 0.150 | -0.05 | 0.150 |
Here, larger normalized rewards indicate better alignment with the corresponding objective. These results indicate that the PKU-SafeRLHF model places a greater emphasis on helpfulness, which is likely due to its constrained optimization loss. We will add more discussion with PKU-SafeRLHF in the revision.
Aligned model is only marginally better than the baseline. Are such results reasonable?
Thank you for the question.
- Table 1 demonstrates MPO's optimality among all methods under the max-min setting, with the aligned model outperforming the baseline on all individual objectives.
- The improvement appears only marginal because of the inherent conflict between objectives such as helpfulness and harmlessness. Multi-objective alignment tasks involve complex trade-offs, and simultaneously improving all conflicting objectives is nearly impossible [1]. This challenge is precisely why we consider the max-min setting in our work.
References: [1] Rame, A, et al. Rewarded soups: towards pareto-optimal alignment by interpolating weights fine-tuned on diverse rewards. NeurIPS 2023.
What is the role of the during the alignment process?
Thank you for your question.
- While a larger encourages policy to stay close to the reference model, it plays a critical role in balancing conflicting objectives in our multi-objective setting. Stronger regularization helps stabilize aggregation and maintain desirable baseline behaviors while optimizing the minimum reward.
- is a tunable hyperparameter. As shown in Table 1, outperforms and . In particular, the case of collapsed to the reference policy with a 50% win rate. This highlights the importance of tuning for optimal performance.
In practice, the reviewer thinks it is quite difficult for us to serve multiple LLMs at the same time. Especially when the model is large. If the authors adopt the model merging method, this should not be a problem.
- We understand the concern about serving multiple LLMs concurrently. However, the efficiency of MPO primarily comes from the training phase. In Section 4.2 of our experiments, training policies using PPO-based approaches (e.g., MaxMin-RLHF, MORLHF) requires approximately 10 A100 GPU hours, whereas MPO only requires around 2.5 A100 GPU hours, since it avoids reinforcement learning step.
- In the inference phase, instead of running full inference on all LLMs simultaneously, we can compute single-objective policy outputs logits in parallel, then aggregate logits using the MPO framework.
- Additionally, as shown in Table 1, our method achieves better alignment than parameter merging baselines like reward soups and future work will focus on further optimizing the inference pipelines to enhance the scalability of MPO in real-world applications.
This paper studies how to align diverse objectives of human preferences. The authors propose a post-processing approach that combines the optimal policies for each objective without requiring retraining from scratch. Moreover, the author also studies the max-min RLHF, and shows that we can find the optimal policy by adjusting the weight using mirror descent. Experimental results are provided to support their theoretical findings.
给作者的问题
- In Line 139, the author states that "Balancing multiple, often competing objectives leads to training instability, while the need to train multiple reward models and perform RL updates makes them computationally expensive." However, obtaining the optimal policy for each objective also requires training multiple reward models. Therefore, I do not think that the computational cost of MPO is lower than that of previous algorithms like Reward Soups. Could the authors provide further clarification on why MPO is not computationally expensive?
论据与证据
Yes. The claims are correct, clear and easy to follow.
方法与评估标准
The authors evaluate their algorithms using a classical MORLHF dataset named HH-RLHF, which contains three objectives Helpful, Harmless, and Humorous.
理论论述
I check the proof and it is correct to me.
实验设计与分析
The experiment compares MPO with previous algorithms such as Reward Soups and Max-Min RLHF, along with baselines like the reference model and single-reward algorithm. My main concern is that the authors should include comparisons with more prior works, such as MOD (Shi et al., 2024), as well as a baseline algorithm that aggregates the reward and trains the model directly on it. Additionally, since Algorithm 1 has learned the optimal weight , the authors could utilize this weight to implement RS instead of using a uniform weight .
Shi R, Chen Y, Hu Y, et al. Decoding-time language model alignment with multiple objectives. NeurIPS 2024.
补充材料
I read the proof part and the experiment details.
与现有文献的关系
The key contribution of this paper is the proposed algorithm that enables language models to align with diverse objectives by leveraging the optimal policy for each objective, rather than requiring training from scratch. However, the novelty appears to be limited, as the main theorem (Theorem 3.4) has already been studied in Theorem 1 of (Shi et al., 2024). Could the authors clarify the differences between their main theorem and the one in (Shi et al., 2024)? Additionally, the use of mirror descent for weight adjustment closely resembles the approach in (Ramesh et al., 2024). As a result, it remains unclear whether this paper offers a novel theoretical contribution.
Shi R, Chen Y, Hu Y, et al. Decoding-time language model alignment with multiple objectives. NeurIPS 2024.
Ramesh S S, Hu Y, Chaimalas I, et al. Group robust preference optimization in reward-free rlhf. NeurIPS 2024.
遗漏的重要参考文献
Two essential references that are not discussed are (Shi et al., 2024) and (Ramesh et al., 2024). The former presents a result similar to Theorem 3.4 in this paper, while the latter introduces a similar idea of adjusting weights using mirror descent to achieve max-min goal.
其他优缺点
The strengths and weaknesses are provided above.
其他意见或建议
The author should provide more theoretical comparisons (and if possible, empirical comparisons) between MPO and previous algorithms, and the novelty of MPO.
We greatly appreciate your constructive and insightful feedback! Here we provide a detailed response to address all of your concerns.
Differences between main theorem and the one in Shi et al., 2024
Thank you for your feedback. Compared to Thm 1 in Shi et al., our approach differs in both objective and applicable setting, with some overlap in a special case:
- Our Thm 3.4 addresses the max-min setting without explicit preference weights, whereas Shi et al. uses predefined weights. In the special case of linear reward aggregation, Lem 3.9 shows that the optimization leads to a closed-form solution, which indeed reaches same conclusion in Shi et al.
- We introduce an auxiliary normalizing operator for rewards, which is crucial for transforming the optimization over into Eq12 (line 199). Without it, the reward function cannot be avoided in optimization.
- In terms of the derivation, Shi et al. uses a Legendre transform, converting the problem into where is unspecified. In contrast, our proof uses a direct reward–policy mapping, leading to the closed-form expression for directly, providing a more interpretable and transparent theoretical derivation.
The use of mirror descent for weight adjustment closely resembles the approach in Ramesh et al., 2024.
Indeed the high-level idea of Ramesh et al. is to perform robust alignment, which is similar to us. However, we highlight several differences in terms of less resources required, applicable settings, and methodology developed.
- Ramesh et al. aims to derive a group robust preference optimization objective and conduct robust alignment from scratch, which is computationally expensive and requires extensive hyperparameter tuning. In contrast, we could directly use existing single-objective policies, avoiding full retraining and significantly reducing computational cost.
- By reusing pretrained or open-source LLMs, we simplify robust training to only updating preference weights—a lightweight post-processing step that aligns better with practical academic and industry use cases. Ramesh et al. 's method is better suited for settings where LLMs must be trained from scratch.
- Methodologically, Ramesh et al. uses a gradient descent-mirror ascent method to update the policy and the weight simultaneously as its objective is to solve a min-max optimization. In addition, it has access to unbiased gradient estimators. In our case, since we do not train policies, weight optimization becomes a conditional stochastic problem without unbiased gradient estimators. To tackle this, we designed a biased mirror descent method and analyzed its convergence.
Could the authors provide further clarification on why MPO is not computationally expensive?
Thanks for your question.
- While standard RLHF requires training multiple reward models, this can be avoided by using DPO, which is mathematically equivalent and was adopted in our experiments for efficiency.
- For , training the policy via MORLHF (or maxmin-RLHF) for a fixed takes approximately 10 A100 GPU hours, whereas solving for in MPO only takes 2.5 A100 GPU hours.
- Unlike standard MORLHF, which re-runs PPO for different , MPO avoids this overhead by efficiently combining logits. While Reward Soups has similar cost with predefined weights, MPO achieves better alignment (shown Table 1), highlighting its effectiveness in balancing multiple onjectives.
Comparisons with more prior works.
Thank you for the suggestion, we have added further comparisons with prior works:
- We compare our approach with MORLHF, which employs linear aggregated rewards using PPO, as well as with reward soups that utilize learned weights. As shown in Table 1, our MPO method still achieves the highest minimum win rate, where corresponds to the Reward Soups with learned weights
| Model | Helpful | Harmless | Humorous | Min |
|---|---|---|---|---|
| 44.8 | 59.4 | 56.4 | 44.8 | |
| 45.4 | 52.2 | 51.3 | 45.4 | |
| 42.9 | 56.7 | 54.5 | 42.9 | |
| 46.3 | 53.1 | 54.1 | ||
| 51.9 | 53.7 | 50.0 | 50.0 | |
| 53.7 | 50.8 | 48.8 | 48.8 | |
| 41.7 | 54.4 | 52.9 | 41.7 | |
| 54.9 | 53.1 | 57.1 | ||
| Table 1: Win rate(%) against the Reference Model |
- While MOD employs a linear combination of logits, in Figure 2(b), we have evaluated with different . Our results show that the policy obtained via MPO achieves the best objective performance across this grid, outperforming MOD with predefined weights in the max-min setting.
Thanks for the response. However, I still think this paper is very similar to the main theorem in [Shi et al. 2024], with a similar weight adjustment in [Ramesh et al. 2024]. Hence, the novelty seems limited. I will keep my score.
Dear Reviewer FxzW,
Thank you for your follow-up comment. We appreciate your continued engagement with our work. However, we respectfully disagree with your assessment and would like to offer a clarification of the key differences between our approach and the references you mentioned.
First, the primary focus of our paper is on multi-objective alignment via a max-min formulation that does not require pre-specified preference weights. This focus is very different from that of [Shi et al. 2024], which assumes fixed weights as input. Yet how to choose such a fixed weight can be highly nontrivial in practice. On the contrary, our method is specifically designed to infer a robust solution without relying on explicit prior knowledge of user preferences.
Second, as noted in our rebuttal, we introduce an auxiliary normalization operator on the reward functions, which plays a critical role in enabling a closed-form expression for the optimization over (see Eq. 12). This normalization step changes the analytical landscape and requires a different theoretical treatment than that in [Shi et al. 2024].
Regarding your comment on "a similar weight adjustment" in [Ramesh et al., 2024], we would like to emphasize 1) from a computational perspective, our method can directly use existing policies, avoid policy retraining, and thus is much cheaper; 2) from a technical point of view, to achieve such a computational reduction, we propose to use a novel biased mirror descent to update the weights, which makes it possible to keep the policy fixed. This is fundamentally different from the approach in [Ramesh et al. 2024], which employs a joint min-max optimization framework and thus involves both updating weights and policy retraining at each step. In particular, they have access to unbiased gradient estimators of the weight at the costs of much higher computational costs. That being said, our biased optimization method is tailored to our setting to avoid updating the policy and ensure smaller computational costs. In addition, we provide a novel convergence analysis.
This work proposes Mixing Preference Optimization (MPO), a post-processing framework for aggregating single objective policies with a mixing of preference alignment. Specifically, the authors combine two multi-objective RLHF approaches, MORLHF and MaxMin-RLHF using a post-processing strategy to combine single-objective policies. The overall objective of MPO can be viewed as a max-min game between the policy and the combinational factor of MORLHF.
After rebuttal: Thank you to the authors for their feedback. However, my concerns regarding [Theoretical Claims] remain unresolved. In particular, the primary issue is that the correctness of Equation (10) has not been substantiated. This affects the overall confidence in the correctness of the proposed method. I will maintain my original score.
给作者的问题
-
How can the correctness of Eq. 10 be verified?
-
Whether the proposed method will improve the computational complexity in the inference phase?
-
Can we improve the size of in the experiments?
-
Where are the key novelty and soundness compared with MORLHF and MaxMin-RLHF?
论据与证据
The authors claim that MPO significantly reduces training costs and computational overhead. However, I can not find any theoretical or empirical analysis of the computational complexity compared with the baselines such as MORLHF and MaxMin-RLHF.
方法与评估标准
The proposed method uses a post-processing algorithm to tackle alignment with diverse human preferences. The method, itself, is easy to follow and the evaluation criteria make sense.
理论论述
The main concerns are the correctness of the derivation of the algorithm, specifically:
-
The correctness of Eq. 10 can not be verified, which is an important premise for deriving the optimization objective in Theorem 3.4. (Main Theorem).
-
is a global and certain optimality of the policy, but is obtained by stochastic descent in Alg. 2. It is unclear how the estimation of can approximate optimize the objective of Eq. 10.
-
The optimization of is with respective to the input-output . However, the optimization of is input-output agnostic. This anti-intuitive result makes me concerned about the soundness of the proposed method in comparison with MaxMin-RLHF.
实验设计与分析
I have some concerns about the experimental design and analyses as follows.
-
The size of is so small (2 or 3) in the experiments. This can make the soundness of the proposed algorithm weak in practical usage.
-
The comparison does not consider other important RLHF baselines such as PPO-based approaches. The existing work has proposed to aggregate the reward model of PPO to mix diverse preference alignment.
-
The analysis of training costs is missing in the experimental analyses. Besides, it is unclear whether the proposed method will improve the computational complexity in the inference phase.
补充材料
I have reviewed the supplementary material.
与现有文献的关系
An improvement method on the alignment of diverse human preferences upon previous work such as MORLHF and MaxMin-RLHF.
遗漏的重要参考文献
Not found.
其他优缺点
For strengths, 1. the proposed method is easy to follow for mixing diverse preference alignment. 2. The problem of alignment with diverse human preferences is novel and interesting within the area of RLHF.
For weaknesses, 1. the writing is needed to improve, e.g., section 3 is hard to follow. 2. the contribution and novelty are marginal to some extent, the proposed algorithm is a straightforward combination of two previous works MORLHF and MaxMin-RLHF.
其他意见或建议
N/A
We greatly appreciate your constructive and insightful feedback! Here we provide a detailed response to address all of your concerns.
How can the correctness of Eq. 10 be verified?
Thanks for the question. As we explained after Eq. 10 in our original submission, the result follows directly from Sion’s minimax theorem. The objective function is convex in ( fixed) and concave in ( fixed), which satisfies the conditions for the theorem and guarantees the interchange of max and min in Eq. 10.
is a global and certain optimality of the policy, but is obtained by stochastic descent in Alg. 2. It is unclear how the estimation of can approximate optimize the objective of Eq. 10.
- If refers to the reward model, our approach does not involve reward training; Instead, we use an auxiliary normalizing operator to directly derive the optimal policy
- If refers to the preference weight as solved by Alg. 2, then Thm 3.8 provides a KL-based error bound between the learned policy and . The theorem formally states that closely approximates under mild conditions.
The optimization of is with respect to the input-output (x,y). However, the optimization of is input-output agnostic.
Thank you for your question. Again, we believe you are referring to . As shown in Thm 3.4, our main task is to solve for via Eq .12. Once obtained, the optimal policy is effectively a linear combination of the logits from the single-objective policies. Additionally, when applying Alg 2, we only need the individual policies and a set of prompts.
Can we improve the size of in the experiments?
Thank you for your comment. We would like to point out that is consistent with prior works [1,2,3], which also adopts up to three objectives. This dimensionality has proven effective for capturing the scalability of multi-objective alignment, balancing soundness and computational efficiency.
References:
- Chakraborty, S, et al. MaxMin-RLHF: Alignment with diverse human preferences. ICML 2024.
- Yang, R, et al. Rewards-in-context: Multi-objective alignment of foundation models with dynamic preference adjustment. NeurIPS 2024.
- Zhou, Z, et al. Beyond one-preference-fits-all alignment: Multi-objective direct preference optimization. arXiv preprint arXiv:2310.03708 (2023).
The comparison does not consider other important RLHF baselines such as PPO-based approaches.
Thank you for the comment. Standard MORLHF approaches [1,2] optimize linearly aggregated reward functions. while Maxmin-RLHF perform PPO to obtain the policy by optimizing: In light of your comments, we have added more comparisons of such methods. Notably, our MPO method still achieves the highest minimum win rate.
| Model | Helpful | Harmless | Humorous | Min |
|---|---|---|---|---|
| 44.6 | 56.1 | 51.4 | 44.6 | |
| 42.9 | 56.7 | 54.5 | 42.9 | |
| 46.3 | 53.1 | 54.1 | ||
| 46.1 | 53.8 | 54.8 | 46.1 | |
| 41.7 | 54.4 | 52.9 | 41.7 | |
| 54.9 | 53.1 | 57.1 |
References:
- Ji, J, et al. Beavertails: Towards improved safety alignment of llm via a human-preference dataset. NeurIPS 2023.
- Wu, Z, et al. Fine-grained human feedback gives better rewards for language model training. NeurIPS 2023.
The analysis of training costs is missing in the experimental analyses.
Thanks for the suggestion. For computation cost, training the policy using MORLHF (or maxmin-RLHF) with aggregated reward models takes approximately 10 A100 GPU hours, since both methods rely on PPO for policy optimization and differ only in how they aggregate reward functions, In contrast, our approach avoids PPO entirely, solving for the preference weights via Alg 2 requires only about 2.5 A100 GPU hours, offering a significant reduction in training time while still achieving competitive performance.
Where are the key novelty and soundness compared with MORLHF and MaxMin-RLHF?
Thanks for the comment, We have stated the key novelty and soundness of MPO compared to MORLHF and MaxMin-RLHF in both the Introduction and Conclusion sections. In summary:
- MPO connects reward aggregation to policy aggregation, yielding a closed-form aggregation rule. 2. Our approach operates directly on single-objective policies, avoiding extra RL updates and reward model training, thereby significantly reducing computational overhead. 3. The method is backed by rigorous theoretical error bounds that ensure robustness relative to the optimal policy.
This paper proposes MPO, a framework designed to mix diverse single-objective policies for aligning LLMs with human preferences. Instead of training a costly multi-objective RLHF model from scratch, this paper shows how pre-trained, single-objective policies can be aggregated using a batch stochastic mirror descent (BSMD) algorithm. This paper derives a closed-form solution relating the aggregated policy to the individual policies and provide theoretical guarantees. Experiments on multiple multi-objective preferences tasks illustrate that MPO can outperform baseline methods while having lower computation cost.
给作者的问题
Please see the weakness.
论据与证据
Yes, most of the claims are supported by the theoretical guarantees and experiment results.
方法与评估标准
The evaluation mainly relies on the judgment of GPT 3.5/4, it would be better to include human validation.
理论论述
Yes, I carefully checked the key steps of the aggregated policy and convergence guarantees for BSMD, and they appear to be correct.
实验设计与分析
The experimental design is sound in that it considers multiple objectives and compares MPO with representative baselines. One potential issue is that the evaluation relies on GPT 3.5/4 models, although common in current research, may be sensitive to prompt design.
补充材料
I reviewed the supplementary sections including proof derivations, implementation details, and detailed results.
与现有文献的关系
This paper makes meaningful contribution to the preference learning and alignment of LLMs.
遗漏的重要参考文献
This paper discusses a wide range of relevant works of RLHF, preference learning and the diverse alignment objectives.
其他优缺点
Strengths
- This paper proposes an effective method MPO that utilizes existing single-objective policy to combine into a unified one for diverse preferences alignment. The proposed method is well formalized and supported by theoretical analysis.
- MPO has a clear advantage in efficiency, which avoids alignment from scratch.
- Experiment results demonstrate across a wide range of alignment tasks demonstrate the effectiveness of MPO.
Weaknesses
- It would be better to add more validations on the GPT-based evaluation.
- It would be better to include a discussion on the scalability of MPO with respect to computation resources when increasing the number of objectives.
其他意见或建议
Please see the weakness.
We greatly appreciate your constructive and insightful feedback! Here we provide a detailed response to address all of your concerns below.
It would be better to add more validations on the GPT-based evaluation.
Thanks for the suggestion. When utilizing GPT-based evaluations, we have experimented with multiple GPT versions and leveraged prompts similar to previous works such as [1] and [2]. We found that while there are some variations in the output, the overall performance trends remain stable, which gives us confidence in the reliability of this evaluation method. We are also open to further validations and comparisons to ensure that our evaluation framework is as comprehensive and robust as possible.
References:
- Rafael Rafailov, Archit Sharma, Eric Mitchell, Christopher D Manning, Stefano Ermon, and Chelsea Finn. Direct preference optimization: Your language model is secretly a reward model. NeurIPS 2023.
- Zhou, Z., Liu, J., Shao, J., Yue, X., Yang, C., Ouyang, W. and Qiao, Y., 2023. Beyond one-preference-fits-all alignment: Multi-objective direct preference optimization. arXiv preprint arXiv:2310.03708.
It would be better to include a discussion on the scalability of MPO with respect to computation resources when increasing the number of objectives.
Thanks for the question. From our observation, computational cost increases approximately linearly with the dimensionality of . Specifically, obtaining an approximate optimal (using 600 iterations of Algorithm 2) requires:
| dim() | A100 GPU hours |
|---|---|
| 2 | 1.8 |
| 3 | 2.5 |
| 4 | 3.3 |
This linear scaling indicates that the computational burden remains manageable when increasing the number of objectives.
My recommendation is to accept the paper.
The paper proposes a method for inference-time aggregation of policy models that have been tuned to optimize different preference objectives. The paper shows an equivalence between a max-min RL approach to balancing multiple objectives at tuning time and a post-processing approach that balances single-objective-optimized policies according to a normalized max-min objective at inference time. The authors derive a mirror descent algorithm for learning the parameters of this aggregation, and demonstrate its effectiveness against other multi-objective RL and model averaging alternatives.
Reviewers agreed that the paper addresses an interesting problem, and the solution seems to work. To me, the most serious questions raised were around the novelty of the work, especially wrt [Shi et al 2024], which addressed a similar inference-time alignment problem. However, I found the authors' response to be convincing, specifically that Shi et al do address a different problem formulation, where a specification for linear aggregation is given a priori, whereas this paper addresses max-min aggregation, and addresses the additional technical details (i.e., normalization) necessary for solving the problem in this different context. This being said, I am sympathetic to FxzW's concern about novelty, and hope that the authors make their contributions relative to this prior work clear. In addition to discussion in related work, remarks around the main results in the paper and potentially in the introduction would also be appropriate. It may also be useful to include an example where linear aggregation and max-min aggregation give distinct results.
Beyond these questions of novelty, the methodology seems sound, and with the experiments that were added during the rebuttal phase, adequately tested.
Given that there was some confusion among reviewers about the distinctions between training-time vs inference-time aggregation, I might also suggest that the authors consider how these distinctions could be made clearer in the writing, but this is not essential.