PaperHub
7.0
/10
Poster4 位审稿人
最低6最高8标准差1.0
8
6
8
6
3.3
置信度
正确性3.0
贡献度2.5
表达2.3
ICLR 2025

Uncertainty and Influence aware Reward Model Refinement for Reinforcement Learning from Human Feedback

OpenReviewPDF
提交: 2024-09-18更新: 2025-05-17

摘要

关键词
Reward ModelingLarge Language ModelData Augmentation

评审与讨论

审稿意见
8

This paper presents a new method for improving reward model training by prioritising on-policy samples to relable to get the highest performance with the least relabelling cost. The method uses a combination of empirical uncertainty estimated by an ensemble of reward models, and gradient-based influence calculation to select data. The paper presents empirical results showing that their method outperforms baselines across a range of metrics in the HH dataset for two different reward model sizes. Their method improves both reward model performance and policy performance, and in a series of ablations the paper shows that all components of the proposed method are important for the success of the method.

优点

The proposed method is interesting and novel, and shows strong improvements over reasonable baselines. The problem of reward model overoptimisation is an important and timely one, making the contribution reasonably significant. While the method adds additional complexity on top of the baselines, this does seem to pay off in improved performance, and ablations demonstrate that how each component contributes. The idea is intuitive which makes the contribution more compelling. The paper is mostly easy to read and understand. The contribution seems sufficiently original.

缺点

lack of clarity on RLR

The RLR baseline is not explained in the main body of the paper. Could you add a description of this algorithm so that it's clear what UGDA is being compared to?

Limited evaluation dataset

The experiments are only performed on anthropic HH. It would be beneficial to demonstrate improvements on another RLHF dataset as well. This is partially remedied by showing improvements on a range of metrics and for reward models as well as policies, but this is still a weakness of the paper.

Lack of error bars

It would be useful to have error bars or confidence intervals on any of the reported results. I acknowledge that running experiments for additional seeds is costly, but it would make the results much more robust

Summary

Overall I give this paper a 6. I would be willing to raise my score to a 7 if the clarity and error bars issues mentioned above were fixed, and I would raise my score higher if additional datasets where considered and results where still positive for UGDA.

Update

Given the improved results and clarity fixes in the rebuttal, I am raising my score to an 8, assuming all of that is in the final version of the paper.

问题

The RLR baseline is not explained in the main body of the paper. Could you add a description of this algorithm so that it's clear what UGDA is being compared to?

评论

Dear Reviewer V4Mi,

Thanks for your careful and insightful comments, we provide our responses to your questions in the following.

W1: lack of clarity on RLR

A1: The RLR uses the filtered instructions in the data selection stage, but the responses are the corresponding pairwise responses in the preference dataset; we use the samples in the preference dataset to refine the reward model ensembles.

W2: Limited evaluation dataset

A2: To answer this question, we conduct addtional experiments based on the ultra feedback dataset [1] to train the policy, the results are presented in the following.

MethodAlpaca Eval (LC_Winrate)Alpaca Eval (Winrate)Arena HardMTBench 101
PPO11.88.011.74.7
LCB13.08.610.75.2
UWO12.69.112.85.5
RLR13.38.812.45.3
UGDA14.810.913.56.1

Reward Model Evaluation Results

  • Helpful results
Reward ModelTraining MethodsAverage_RewardVar_Reward
Gemma 2BPPO0.660.12
LCB0.610.13
UWO0.650.17
RLR0.740.15
UGDA0.910.09
Gemma 7BPPO0.730.14
LCB0.730.13
UWO0.830.15
RLR0.820.12
UGDA0.940.11
  • Harmless results
Reward ModelTraining MethodsAverage_RewardVar_Reward
Gemma 2BPPO0.640.18
LCB0.630.14
UWO0.650.16
RLR0.700.17
UGDA0.840.13
Gemma 7BPPO0.740.14
LCB0.780.13
UWO0.720.13
RLR0.800.11
UGDA0.930.12

GPT Comparision Results

WinTieLose
UGDA vs. PPO49.6%24.5%25.9%
UGDA vs. LCB51.2%16.6%32.2%
UGDA vs. UWO46.7%21.8%31.5%
UGDA vs. RLR49.8%19.8%30.4%

From the above results, we can see that UGDA still outperforms than all the baselines, which further verify the effectiveness of our architecture. Additionally, we do not apply different random seeds for the new dataset experiments due to the time limitation of the rebuttal. If needed, we will add the results with the error bars.

References

[1] Cui, G., Yuan, L., Ding, N., Yao, G., He, B., Zhu, W., ... & Sun, M. (2024, June). ULTRAFEEDBACK: Boosting Language Models with Scaled AI Feedback. In Forty-first International Conference on Machine Learning.

评论

W3: Lack of error bars

A3: We conduct additional experiments with different random seeds, specially, we use five different random seeds for the policy optimization. The results are shown in the following.

Benchmark Evaluation Results

MethodAlpaca Eval (LC_Winrate)Alpaca Eval (Winrate)Arena HardMTBench 101
PPO11.7 ± 0.0158.0 ± 0.01011.7 ± 0.0214.7 ± 0.018
LCB12.8 ± 0.0098.6 ± 0.01410.7 ± 0.0145.2 ± 0.013
UWO12.4 ± 0.0129.1 ± 0.01312.8 ± 0.0155.5 ± 0.008
RLR13.1 ± 0.0208.8 ± 0.01611.7 ± 0.0095.3 ± 0.017
UGDA14.2 ± 0.01110.8 ± 0.01813.4 ± 0.0116.2 ± 0.012

Reward Model Evaluation Results

  • Helpful results
Reward ModelTraining MethodsAverage_RewardVar_Reward
Gemma 2BPPO0.61 ± 0.03720.19 ± 0.0231
LCB0.52 ± 0.04010.19 ± 0.0196
UWO0.53 ± 0.03830.20 ± 0.0228
RLR0.71 ± 0.02190.17 ± 0.0390
UGDA0.80 ± 0.02320.12 ± 0.0289
Gemma 7BPPO0.62 ± 0.03980.22 ± 0.0121
LCB0.74 ± 0.03840.18 ± 0.0212
UWO0.70 ± 0.04030.16 ± 0.0426
RLR0.80 ± 0.03540.15 ± 0.0399
UGDA0.89 ± 0.04920.13 ± 0.0276
  • Harmless results
Reward ModelTraining MethodsAverage_RewardVar_Reward
Gemma 2BPPO0.63 ± 0.02390.19 ± 0.0393
LCB0.58 ± 0.03770.22 ± 0.0411
UWO0.60 ± 0.02880.17 ± 0.0407
RLR0.62 ± 0.02970.14 ± 0.0342
UGDA0.79 ± 0.01970.13 ± 0.0296
Gemma 7BPPO0.72 ± 0.03770.21 ± 0.0223
LCB0.74 ± 0.03430.17 ± 0.0356
UWO0.71 ± 0.02680.13 ± 0.0402
RLR0.72 ± 0.03720.15 ± 0.0389
UGDA0.86 ± 0.04740.11 ± 0.0322

GPT Comparision Results

WinTieLose
UGDA vs. PPO51.3% ± 6.6221.2% ± 4.4727.5% ± 5.67
UGDA vs. LCB49.5% ± 3.3613.4% ± 3.8337.1% ± 7.39
UGDA vs. UWO48.6% ± 5.0720.3% ± 2.1531.1% ± 5.51
UGDA vs. RLR53.9% ± 4.0721.1% ± 2.7425.0% ± 4.45

From the above results, we can know that the superiority of our UGDA is not affected by random seed, and we will add the above results to our final version.

评论

Thanks for your response. Given the improved results and clarity fixes in the rebuttal, I am raising my score to an 8, assuming all of that is in the final version of the paper.

评论

Thanks for your positive feedback! We will follow your advice and add all the content in our final version.

审稿意见
6

This paper presents Uncertainty-Gradient-based Data Augmentation (UGDA), an approach to enhance reinforcement learning from human feedback for training LLMs. UGDA addresses a key issue: the limitations of an imperfect reward model that serves as a proxy for human preferences, which becomes increasingly problematic as policy optimization alters the training data distribution. UGDA aims to refine the reward model by selectively relabeling the most uncertain and impactful interaction data during policy optimization. The proposed approach follows a three-stage pipeline: Reward LoRA Ensembles, Data Selection, and Reward LoRA Refinement. Experiments demonstrate that UGDA improves policy performance without the costly requirement of additional human preference data collection.

优点

  1. The paper introduces a new method for selecting the uncertain and impactful interaction data during policy optimization to refine the reward model.
  2. The paper is well-organised and articulated with clarity, the complex concepts are explained in a clear and concise manner.
  3. The results, partly supported by some confidence/uncertainty based methods, demonstrate considerable potential for improving alignment of LLMs with human values.

缺点

The results have a few issues which make evaluating the contribution difficult:

  1. The paper lacks a comparison with some existing works, particularly methods involve iterative PPO/DPO method that train a reward model simultaneously and reward ensembles [1].
    [1] Coste T, Anwar U, Kirk R, et al. Reward model ensembles help mitigate overoptimization.
  2. The alignment of relabeled reward data with human annotator judgments remains insufficiently validated.

问题

Could the authors provide a more detailed description of the process for selecting the validation dataset? Given that the impact score depends on the validation set, what are the implications if the validation dataset distribution significantly differs from the training set? Will this also introduce high bias in the impact score.

How does the computational complexity of UGDA scale with model and dataset size, considering the use of an ensemble of reward models and the validation dataset?

What if the variance in reward signals provided by different reward models for the same questions and responses remains high after refining the reward models? The paper lacks an analysis of how refining the reward models impacts the consistency of the reward signals.

评论

Dear Reviewer Jzfj,

Thanks for your careful and insightful comments, we provide our responses to your questions in the following.

W1: Comparison with iterative PPO/DPO method and reward ensembles.

A1: We have conducted the reward ensembles method UWO [1] in the current version of our paper as the baseline. Additionally, follow your advice, we conduct the experiment based on the DPO [2] and iterative DPO [3], the results are based on the Gemma 7B policy.

Benchmark Evaluation Results

MethodAlpaca Eval (LC_Winrate)Alpaca Eval (Winrate)Arena HardMTBench 101
DPO11.17.910.84.9
iterative DPO11.98.311.75.2
UGDA14.210.713.66.1

Reward Model Evaluation Results

Test SettingTraining MethodsAverage_RewardVar_Reward
HelpfulDPO0.610.15
iterative DPO0.690.14
UGDA0.910.12
HarmlessDPO0.580.13
iterative DPO0.650.12
UGDA0.880.10

GPT Comparision Results

WinTieLose
UGDA vs. DPO51.8%23.9%24.3%
UGDA vs. Iterative DPO48.9%21.7%29.4%

From the results, we can see that iterative DPO outperforms DPO, but is less effective than UGDA. This may be because of the superior generalization ability of PPO compared with DPO.

References

[1] Coste, T., Anwar, U., Kirk, R., & Krueger, D. (2023). Reward model ensembles help mitigate overoptimization. arXiv preprint arXiv:2310.02743.

[2] Rafailov, R., Sharma, A., Mitchell, E., Manning, C. D., Ermon, S., & Finn, C. (2024). Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36.

[3] Xu, J., Lee, A., Sukhbaatar, S., & Weston, J. (2023). Some things are more cringe than others: Preference optimization with the pairwise cringe loss. arXiv preprint arXiv:2312.16682.

W2: Relabeled reward data validation.

A2: We provide the comparision of the relabeled reward data with human or GPT annotator in Table 1, which shows the similarity of the human labelers and GPT-4, which is suffecient for supporting us to use GPT annotator in the following experiments. If needed, we are willing to conduct more experiments to verify the results with your advice.

Q1: Validation dataset construction.

A1: We build the validation dataset by randomly selecting 10% samples of the perference dataset, specially each samople is constructed by the instruction with chosen response. We will add the related description in the experimental setup in our final version.

Q2: The computational complexity.

A2: The complexity of reward model training with ensembles scales with the preference data size Dp\mathcal{D}_p and the number of ensembles KK, which is O(DpK)\mathcal{O}(|\mathcal{D}_p|\cdot K).

For the gradient based data selection, we divide this into two stage to analysis the complexity. The gradient feature computation is the most expensive step, and the cost scales linearly with candidate dataset size D_train|\mathcal{D}\_{\text{train}}|, the number of checkpoints NN and the projection dimension dd, which is O(D_trainNd)\mathcal{O}(|\mathcal{D}\_{\text{train}}|\cdot N \cdot d). From the experimental results in our work, we can see that a small projection dimension is sufficient.

The data selection process requires minimal computation, and scales with the size of D_train|\mathcal{D}\_{\text {train}}| and D_val|\mathcal{D}\_{\text{val}}|, which is O(D_trainD_val )\mathcal{O}\left(|\mathcal{D}\_{\text {train}}| \cdot\left|\mathcal{D}\_{\text {val }}\right|\right).

Furthermore, we will add the above analysis to our final version.

评论

Q3: Variance problem.

A3: To answer this question, we count the variance of all training samples. Before the iteration of reward model refining, the smallest variance among the top 50% of samples is 0.24. After refining, only 21% of the samples have a variance greater than 0.24, indicating that the consistency of sample scoring by reward model ensembles has increased.

The above results also mean that, if we want to get more consistent reward model ensembles, we can further introduce more iterations for the policy optimization and reward refining. But considering the time cost and the training difficulty of PPO, combining with the results in our work, we can observe that one iteration is enough for the better results to show the effectiveness of UGDA.

评论

Dear Reviewer Jzfj,

In the above responses, we have try our best to answer your questions and solve your confusions. Due to the rebuttal ddl is coming, we are very willing to have a more in-depth discussion with you, and we welcome you to give us more suggestions. If you have additional suggestions, please let us know and we will try to respond as quickly as possible.

评论

The reply did not address all of my concerns, particularly, I still believe that selectively relabeling the most uncertain and impactful interaction data during policy optimization could be computationally costly. However, based on the detailed experiments provided, I have decided to raise my score.

评论

Thanks for your positive feedback, we will add more details in our final version, and we will further improve our method in future work.

审稿意见
8

The paper considers the reward learning problem in reinforcement learning from human feedback (RLHF). As a reward model is trained from preference data and can be inaccurate, it would be beneficial to allow the model to query the labels of some samples. The key question is: querying the labels of which samples can improve the model’s performance the most.

The paper proposes Uncertainty-Gradient based Data Augmentation (UGDA). The method selects samples for labeling based on a joint objective combining (1) the uncertainty of the sample’s prediction (based on an ensemble of reward models) and (2) the influence of the sample on a validation set. The model identifies samples that maximize this joint objective and queries their labels. Empirical results demonstrate that UGDA outperforms other sample selection methods in improving model performance.

优点

Novelty: The algorithm is novel in using gradient-based data influence for reward learning. To the best of my knowledge, this approach has not been previously applied to reward learning in RLHF.

Motivation and Empirical Performance: The proposed method is well-motivated and well explanable. Empirical results show UGDA's effectiveness.

缺点

Motivation and algorithm design. I wonder if UGDA can be over-complicated for the goal it wants to achieve. For example, can I use simpler heuristics to select samples that may lead to a larger learning gradients?

For example, does UGDA tend to select out-of-distribution samples in the training set (as these samples can lead to larger training gradients)? If we simply find samples that are outliers in the training dataset (for example, using data density distribution), would they perform similarly to UGDA?

Evaluation. The reward models are all finetuned from Gemma models, and Llama-2-13B model is used for evaluation. Would UGDA have similar results if tested with different model architectures?

Presentation. Notational Clarity: The notation could be improved for clarity. For example, RvR_v looks like a specific reward function, but it actually represents variance.

The paper would benefit from overall improvements in clarity and grammar. Specific issues include: Line 18-19, ungrammatical: “Since the policy optimization continuously shifts the human preference training dataset’s distribution.” Line 48 “Recently, there are some works [that] focus on”

问题

I have some questions in the weakness section above.

I didn’t see details on how the validation set is constructed. Does the validation set have the same distribution as the training set?

评论

Dear Reviewer wH1u,

Thanks for your careful and insightful comments, we provide our responses to your questions in the following.

W1: Motivation and algorithm design

A1: In the training procedure of RLHF, for the gradient-based data selection, our motivation is to identify the most relevant data from these extensive datasets to effectively develop specific capabilities, which can be used for the further reward model refining. Specifically, we study the problem of identifying the influence of training samples on the prediction of the validation samples. The data influence can be learned by gradient information [1,2].

Additionally, our experiments based on the complex QA tasks, thus we can not directly define the samples that are out of distribution by the data visualization. Thus, we use a heuristic way to define the out of distribution samples by the length of the instruction, specially, to achieve the fair comparision, we sort the samples by the length of the instructions, we use 25% of them to be relabeled by the GPT annotator and further refine the reward models. We show the results of the policy optimization in the following, and we use OOD (length) to represent the above method.

MethodAlpaca Eval (LC_Winrate)Alpaca Eval (Winrate)Arena HardMTBench 101
OOD (length)12.48.311.95.2
UGDA14.210.713.66.1

From the results, we can observe that, if we use the length of instructions as the criterion to select the samples, the results are not much more superior than the random selection for the policy optimizaiton. This may be because the samples that influence the gradient may not only be the OOD samples, which means, in the policy optimization process, the learned policy can generalize to some OOD scenarios. Thus, these OOD samples may not be the most important factor affecting the gradient.

References

[1] Pruthi, G., Liu, F., Kale, S., & Sundararajan, M. (2020). Estimating training data influence by tracing gradient descent. Advances in Neural Information Processing Systems, 33, 19920-19930.

[2] Han, X., Simig, D., Mihaylov, T., Tsvetkov, Y., Celikyilmaz, A., & Wang, T. (2023). Understanding in-context learning via supportive pretraining data. arXiv preprint arXiv:2306.15091.

W2: Evaluation

A2: As shown in the scaling law of the reward model [1], the bigger reward model has strong ability, thus we use it to evaluate the result. To answer this question, we additionally use the Gemma-2 27B to evaluate the results.

  • Helpful results
Reward ModelTraining MethodsAverage_RewardVar_Reward
Gemma 2BPPO0.530.17
LCB0.570.15
UWO0.440.14
RLR0.480.09
UGDA0.690.08
Gemma 7BPPO0.550.19
LCB0.630.16
UWO0.640.17
RLR0.730.13
UGDA0.840.10
  • Harmless results
Reward ModelTraining MethodsAverage_RewardVar_Reward
Gemma 2BPPO0.510.13
LCB0.480.15
UWO0.490.14
RLR0.550.12
UGDA0.700.09
Gemma 7BPPO0.670.14
LCB0.710.12
UWO0.680.11
RLR0.700.11
UGDA0.810.07

From the results, we can see that our UGDA still outperform than other baselines, the results show that our method can acheive better performance with different bigger reward model evaluation, which is independent of the specific model structure.

References

[1] Gao, L., Schulman, J., & Hilton, J. (2023, July). Scaling laws for reward model overoptimization. In International Conference on Machine Learning (pp. 10835-10866). PMLR.

W3: Presentation.

A3: We will follow your advice and revise all the misleading presentations. And we will also check the full paper again to polish the presentation in our final version.

评论

Q1: I didn’t see details on how the validation set is constructed. Does the validation set have the same distribution as the training set?

A1: We build the validation dataset by randomly selecting 10% samples of the perference dataset Dp\mathcal{D}_p, specially each sample is constructed by the instruction with chosen response. The validation set and training set have different distributions. The training dataset are the interaction data during policy optimization, but the validation dataset is formed by the preference dataset. Furthermore, we will add the related description to the experimental setup in our final version.

评论

Dear Reviewer wH1u,

In the above responses, we have try our best to answer your questions and solve your confusions. Due to the rebuttal ddl is coming, we are very willing to have a more in-depth discussion with you, and we welcome you to give us more suggestions. If you have additional suggestions, please let us know and we will try to respond as quickly as possible.

评论

Thanks for providing additional experiment results and clarifications. The authors have addressed my concerns. The proposed method has more solid empirical justifications and I have raised the score.

The paper can still benefit further from a clearer presentation in the revision.

评论

Thanks for your positive feedback! We will follow your advice to add more details and further improve the presentation in our final version.

审稿意见
6

This paper study the reward modeling using LLMs in RLHF. The authors argue that the existing reward models, which serve as proxies for human preferences, become less effective as the policy optimization process shifts the distribution of the training dataset. To address this, UGDA leverages policy interaction samples, selecting them based on the uncertainty of the reward ensembles and the gradient-based influence on policy optimization. The selected samples are then used to refine the reward model through supervised learning, which in turn improves the policy. Empirical results show that UGDA can enhance policy performance without the need for costly human preference data collection, surpassing state-of-the-art methods in experiments.

优点

This paper is well-motivated and well-written.

This paper aims to address the off-distribution reward overoptmization problem, which is a critical issue in RLHF.

缺点

Lacks comparisons to active learning and data augmentation baselines.

Considering that off-policy methods have become popular, additional analysis of the robustness of UGDA under different policy optimization scenarios, such as iterative DPO, could strengthen the paper.

问题

Could you apply the method to the full fine-tuning reward ensembles?

评论

Dear Reviewer DX3G,

Thanks for your careful and insightful comments, we provide our responses to your questions in the following.

W1: Lacks comparisons to active learning and data augmentation baselines.

A1: We conduct experiments with the ablation study which only uses uncertainty. This can be seen as an active learning-based method.

To achieve the fair comparision, we have conducted the experiments based on the random data augmentation as the baseline of our method in the experiments, which can be seen as the data augmentation baseline. The results are shown in Section 5.3.

W2: Off-policy methods.

A2: We follow your advice to conduct experimental analysis based on the DPO [1] and iterative DPO [2]. Specially, the for the iterative DPO, to achieve the fair comparison, we only apply one round iteration with 25% samples for data augmentation, we use GPT annotator as the reward model for iterative DPO. Moreover, the experimental results are based on the Gemma 7B policy.

Benchmark Evaluation Results

MethodAlpaca Eval (LC_Winrate)Alpaca Eval (Winrate)Arena HardMTBench 101
DPO11.17.910.84.9
iterative DPO11.98.311.75.2
UGDA14.210.713.66.1

Reward Model Evaluation Results

Test SettingTraining MethodsAverage_RewardVar_Reward
HelpfulDPO0.610.15
iterative DPO0.690.14
UGDA0.910.12
HarmlessDPO0.580.13
iterative DPO0.650.12
UGDA0.880.10

GPT Comparision Results

WinTieLose
UGDA vs. DPO51.8%23.9%24.3%
UGDA vs. Iterative DPO48.9%21.7%29.4%

From the results, we can see that iterative DPO outperforms DPO, but is less effective than UGDA. This may be because of the superior generalization ability of PPO compared with DPO.

References

[1] Rafailov, R., Sharma, A., Mitchell, E., Manning, C. D., Ermon, S., & Finn, C. (2024). Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36.

[2] Xu, J., Lee, A., Sukhbaatar, S., & Weston, J. (2023). Some things are more cringe than others: Preference optimization with the pairwise cringe loss. arXiv preprint arXiv:2312.16682.

Q1: Could you apply the method to the full fine-tuning reward ensembles?

A1: Our method can be directly applied to the full fine-tuning without additional modification, but due to the computational cost and the time limitation, we can not provide the fine-tuning results during the rebuttal, we are sorry for this. From the results of the LoRA fine-tuning, the effectiveness of our UGDA can be verified.

评论

Dear Reviewer DX3G,

In addition to the above response, we re-evaluated the feasibility of the full fine-tuning experiments, and try our best to conduct the experiments to answer Q1.

But due to some limitation about the computational and time cost, we use another small base model to implement the baselines and our method.

Specifically, in the full fine-tuning process, we use the Pythia 70M model as the reward model and the Pythia 1.4B model as the policy model. The reward judge model is the Pythia 6.9B. We regret that, due to limited time constraints, we are unable to train a larger model for full fine-tuning.

Benchmark Results

MethodAlpaca Eval (LC_Winrate)Alpaca Eval (Winrate)Arena HardMTBench 101
PPO0.310.170.340.16
LCB0.330.190.410.19
UWO0.420.220.370.17
RLR0.440.240.430.22
UGDA0.650.320.520.26

Reward Model Evaluation Results

Test SettingTraining MethodsAverage_RewardVar_Reward
HelpfulPPO0.480.22
LCB0.450.24
UWO0.530.21
RLR0.590.19
UGDA0.660.16
HarmlessPPO0.510.26
LCB0.490.25
UWO0.550.24
RLR0.640.26
UGDA0.680.19

GPT comparision Results

WinTieLose
UGDA vs. PPO38.5%35.3%26.2%
UGDA vs. LCB41.7%36.6%21.7%
UGDA vs. UWO39.6%31.2%29.2%
UGDA vs. RLR34.7%33.5%31.8%

From the results, it is evident that our UGDA outperforms various baselines. However, compared to the results on the Gemma model, the improvement is less significant. This could be attributed to the small size of the base model and the capabilities of the Pythia model. Nevertheless, the experimental results still support the superiority of our UGDA in the full fine-tuning scenario.

Moreover, we are very willing to have a more in-depth discussion with you, and we welcome you to give us more suggestions. If you have additional suggestions, please let us know and we will try to respond as quickly as possible.

评论

Dear Reviewer DX3G,

In the above responses, we have try our best to answer your questions and solve your confusions. Due to the rebuttal ddl is coming, we are very willing to have a more in-depth discussion with you, and we welcome you to give us more suggestions. If you have additional suggestions, please let us know and we will try to respond as quickly as possible.

评论

Thanks for your response. My concerns are addressed and I have raised my score to 6.

评论

Thanks for your positive feedback! We will follow your advice and add more details in our final version.

AC 元评审

This paper studies reward modeling in RLHF. It uses uncertainty and gradient information to select important data, has further improved the learned reward model, and shows SoTA performance empirically.

Strengths:
This paper is well-written, novel, and well-motivated. The comprehensive experiments show strong performance.

Weaknesses:
Some reviewers questioned the motivation and algorithm design, and worried that the proposed method is over-complicated. Some other reviewers were concerned about the lack of sufficient experiments. However, most of these weaknesses were addressed during the rebuttal period.

All reviewers of this paper agreed that this paper made significant contributions to reward modeling in RLHF, so I would vote to accept.

审稿人讨论附加意见

During the rebuttal period, the authors provided extensive additional experimental results, and addressed most concerns from all reviewers.

最终决定

Accept (Poster)