PaperHub
6.5
/10
Poster4 位审稿人
最低6最高8标准差0.9
6
6
8
6
2.8
置信度
正确性3.5
贡献度2.8
表达3.3
ICLR 2025

Strong Preferences Affect the Robustness of Preference Models and Value Alignment

OpenReviewPDF
提交: 2024-09-23更新: 2025-03-29
TL;DR

We show that dominant preferences, marked by probabilities close to 0 or 1, can cause sensitivities in preference models and compromise robustness of value alignment.

摘要

关键词
Preference modelsrobustness of value alignmentsensitivity analysisAI safety

评审与讨论

审稿意见
6

This paper addresses the core issue of how changes in the probabilities of some preferences affect the predictions of these models for other preferences, given the limited exploration of how changes in the probability of a modeled preference could influence others within existing preference models. The paper then conducts a theoretical analysis of the Bradley-Terry model, revealing when this preference model is sensitive to certain types of preferences. It also discusses how to extend these conclusions to the more general Placket-Luce model. Following this, the paper uses experiments to demonstrate that the discussed phenomena exist in current LLMs. The paper presents two interesting implications, providing some novel insights into the potential impact of preference models on LLM alignment.

优点

The research topic of this paper is quite novel. Although it is a commonly felt (and troubling) issue that the preference model's output score varies with the relative strength of the participants in comparisons, there is a lack of further analysis on this phenomenon within my knowledge.

This paper theoretically analyzes whether the Bradley-Terry and Plackett-Luce models are sensitive to certain preferences and when this sensitivity occurs, providing strong persuasive power. The conclusion that dominant preference can affect the robustness of preferences and that longer tuples can increase robustness has certain practical significance.

In terms of writing, the paper is very clearly articulated, making it easy for readers to understand the research questions and the phenomena revealed.

缺点

The assumptions in this paper are generally reasonable in conventional scenarios. However, in the context of LLMs, these assumptions might not be universally applicable for modeling human values and preferences in broader contexts, such as Assumptions 2 and 4, and Lemma 1. In fact, there is considerable work showing that preferences can form cycles and may be non-transitive [1]. Certainly, in these scenarios, the Bradley-Terry and Plackett-Luce models presented in the paper might also not be suitable.

In the context of LLMs, is sensitivity to dominant preferences necessarily a bad thing? This might not align perfectly with theoretical environments; perhaps it is just a bias in the dataset or even the desired effect that people want? For example, we might also want to increase the diversity of non-dominant preferences in LLMs. That may need further discussion and analysis.

The experimental section of this article is relatively simple, while the preferences of LLMs in reality are very complex. It may require more complex experimental scenarios to demonstrate whether some of the issues and phenomena raised by the author still exist.

There are some typos, such as on line 227, where I think the mention of 'the last section' should refer to Section 5 'Related works'?

Reference

[1] Nash Learning from Human Feedback (https://arxiv.org/pdf/2312.00886)

问题

See the Weaknesses section.

评论

We appreciate your thoughtful feedback. Please find below for our response to your questions.

1. Preferences can form cycles and be non-transitive.

In the context of LLMs, these assumptions might not be universally applicable for modeling human values and preferences in broader contexts, such as Assumptions 2 and 4, and Lemma 1. In fact, there is considerable work showing that preferences can form cycles and may be non-transitive. Certainly, in these scenarios, the Bradley-Terry and Plackett-Luce models presented in the paper might also not be suitable.

We agree that human preferences could be non-transitive and are not fully covered by the Plackett-Luce (P-L) model. We acknowledge the limitation and will discuss this in future revisions. On the other hand, P-L model (and B-T model) is important and widely used for value alignment in practice. Therefore, we believe the analysis in our work provides useful insights about alignment and safety.

2. Is sensitivity necessarily a bad thing?

In the context of LLMs, is sensitivity to dominant preferences necessarily a bad thing? This might not align perfectly with theoretical environments; perhaps it is just a bias in the dataset or even the desired effect that people want? For example, we might also want to increase the diversity of non-dominant preferences in LLMs. That may need further discussion and analysis.

This is a very interesting point. Indeed, sensitivity introduced by dominant preferences is an inherent property of the discussed preference model. It may or may not introduce destructive outcome, but we believe it is important that the users of the models are aware of it.

3. Real-world implication of theoretical sensitivity.

The experimental section of this article is relatively simple, while the preferences of LLMs in reality are very complex. It may require more complex experimental scenarios to demonstrate whether some of the issues and phenomena raised by the author still exist.

We would like to invite you to read our common response to all reviewers posted above, where we study real-world scenarios where dominant preferences result in undesirable behaviors.

4. Typos.

There are some typos, such as on line 227, where I think the mention of 'the last section' should refer to Section 5 'Related works'?

By "model discussed in the last section", we meant the general preference model discussed in section 2.2. We will change the wording to "the section above" to avoid ambiguity.

评论

Thank you for your reply! This answered my questions.

评论

Thank you for your acknowledgment!

审稿意见
6
  • The paper discusses an issue with common preference modeling techniques. They find that, when preferences are dominant, small changes in training preferences can have a large influence on held-out preference predictions. This shows that predicted preferences are non-robust.
  • The paper proves such nonrobustness theoretically for the Bradley-Terry and Plackett-Luce models, and shows the effect experimentally in a toy setting.
  • Intuitively, the issue is the following: When preferences are dominant, small changes in the probability of choosing option A vs another option (e.g., going from 98% to 99%) will have large effects on the derived reward scores for A. Hence, if there are two options A, B for which no direct comparisons exist, but which are both known to be prefered compared to a third option C, then the relative score of A to B will vary greatly depending on only small changes in the probability of choosing A vs C or B vs C.

优点

  • The basic issue with nonrobustness of preference models is straightforward and sound. It also appears novel to me though I'm not an expert in preference modeling.
  • The paper goes into a lot of theoretical detail in analyzing the problem, investigating the extent of the issue, cases where it's less severe, etc.
  • The paper includes empirical experiments where the effect is demonstrated, using different LLMs. The authors also report standard deviation over different seeds.
  • Preference modeling is an important topic as it's used in today's state of the art language models
  • The paper is organized and structured well and well written.

缺点

  • The empirical investigation is fairly narrow, with a very simple toy experiment. Using DPO here seems less interesting than proper RL since DPO is a bit more like directly training the LLM to output responses in proportion to the stated preferences.
  • For me, the most important question is: "Is this actually an issue in practice"? There are enough issues with preference modeling that occur in practice, so I'm not sure one should worry much about a theoretical non-robustness.
    • The issue relies on dominant preferences. However, it is not surprising that dominant preferences don't provide much signal for learning preferences. E.g., if a model only sees comparisons of summaries where one of the summaries doesn't talk at all about the text, then all the model can learn is that the summary should talk about the text. It is not surprising that the model isn't able to tease apart the difference between a good and an excellent summary if it has never seen a direct comparison of the two. Maybe one would want the preference model to just give equal scores to all adequate summaries, rather than randomly high scores to some of them, but I am not sure this matters much? Also, in practice, I wouldn't expect there to be big differences between the rewards for the different adequate summaries (assuming one doesn't go out of one's way to overoptimize and e.g. train for 1000 epochs on the same few datapoints).
    • In practice, reward models are trained to compare different responses of the same quality level (e.g. from the same model) to get a good learning signal for the reward model. In this case, the problem doesn't arise.
    • I believe that issues with extreme scores are more likely to come from simple reward hacks, non-robustness of the neural network, rather than issues stemming from the preference data that would arise in a theoretical model.
  • Due to the above, I am fairly on the fence about this paper. I feel like it is a well written and technically solid paper with an interesting insight so I'm inclined to accept it, but I don't believe this paper is actually very relevant in practice.

问题

  • I think the paper is structured well and easy to read, but it could still be even more easy to follow. You could work through a very simple toy example explicitly to help with this. E.g., as a suggestion: Example 1 comes only on page 5, it could come earlier (e.g. in the introduction). Moreover, I feel one could explain the example more intuitively somehow.
  • It would be great to come up with an empirical experiment that is less toy then bird/dog/cat. Are there any real-world preferences where dominant preferences could lead to an issue?
  • Would this issue arise empirically in a setting where one doesn't completely overfit to the PM training data and only trains for a few epochs?
  • Figure 3: I found this figure initially hard to wrap my head around. I am not sure exactly how to improve this, but one suggestion would be to distinguish between "train" and "test" comparisons. It would also be interesting to report the predictions for the one comparison that is kept constant (is it actually constant empirically?). It seems from the plots that the model is fairly badly calibrated? Is this theoretically expected with DPO or an optimization issue?
  • Assume I randomly query comparisons between n different objects (where preferences follow some plausible distribution, e.g., one could model the scores for each option as normally distributed, or maybe some heavy tailed distribution instead). In this case, how many comparisons do I have to sample, to get a good bound on the error of my predictions? How does the nonrobustness in this paper affect this sample complexity? Does it make a big difference? Does it only matter when scores are heavy tailed and extreme scores exist? This would be a more theoretical approach to answering the question: "Does this problem actually matter" and an answer to this would also convince me that the paper is more relevant.
评论

We appreciate your detailed and thoughtful feedback. Please find below for our response to your questions.

1. Real-world implication of theoretical sensitivity.

The empirical investigation is fairly narrow, with a very simple toy experiment... It would be great to come up with an empirical experiment that is less toy then bird/dog/cat. Are there any real-world preferences where dominant preferences could lead to an issue? ... Using DPO here seems less interesting than proper RL since DPO is a bit more like directly training the LLM to output responses in proportion to the stated preferences... For me, the most important question is: "Is this actually an issue in practice"?

We would like to invite you to read our common response to all reviewers posted above, where we study real-world scenarios where dominant preferences result in undesirable behaviors in PM under the RLHF framework. We hope this study can further demonstrate the real-world implication of the findings in the paper.

In practice, reward models are trained to compare different responses of the same quality level (e.g. from the same model) to get a good learning signal for the reward model. In this case, the problem doesn't arise.

While it is true that in many cases, the reward model is only required to compare two adequate inputs, dominant preferences are not uncommon. As shown in Table R1 and Table R2 in the common response, the reward signals for "chosen" and "rejected" responses from real-world PMs on hh-rlhf dataset can be significantly different. Therefore, dominant preference and sensitivity could still be an issue in practice.

I believe that issues with extreme scores are more likely to come from simple reward hacks, non-robustness of the neural network, rather than issues stemming from the preference data that would arise in a theoretical model.

We agree that the factors mentioned above are all valid and important reasons for extreme scores in practice. We will include these factors in the revisions of our paper for the completeness of the paper. On the other hand, the analysis in the paper does not rely on the quality of the training process and the trained preference model. Even if reward hacking and non-robustness of NNs are resolved, sensitivity will remain as long as: (1) the PM adheres to the mathematical model in the paper, and (2) the PM aligns with some dominant preferences.

2. Readability can be further improved.

Example 1 comes only on page 5, it could come earlier (e.g. in the introduction). Moreover, I feel one could explain the example more intuitively somehow.

Figure 3: I found this figure initially hard to wrap my head around... one suggestion would be to distinguish between "train" and "test" comparisons.

Thank you very much for these suggestions. We will add a simplified example (Example 1 requires analysis in Section 2.2 and 2.3) to the introduction to give a high-level idea about the issue in the final version of the paper. In Fig. 3, we used superscript D\mathcal{D} and L\text{L} to respectively indicate distribution specified by the dataset and learned by the model. We have updated the figure to use different markers to separate "train" and "test" values.

3. Reporting p12Lp^{\text{L}}_{12} given controlled pD_12p^{\mathcal{D}}\_{12} and quality of calibration.

Figure 3: It would also be interesting to report the predictions for the one comparison that is kept constant (is it actually constant empirically?). It seems from the plots that the model is fairly badly calibrated? Is this theoretically expected with DPO or an optimization issue?

Thank you for the suggestion. We updated figures in the paper to include pL_12p^{\text{L}}\_{12}.

We do not know the exact reason for the poor calibration (i.e., pL_12p^{\text{L}}\_{12} is not proportionally increasing with pD_12p^{\mathcal{D}}\_{12}). One possibility could be overfitting in DPO, as also noted by some prior works [1]. Optimization issues could also cause this, as we did not extensively search for the optimal hyperparameters due to resource constraints.

Nevertheless, the purpose of figures is to demonstrate the issue arising from dominant preferences, regardless of the cause of those preferences. We consider the underlying reason for the poor calibration to be relatively minor in the context of the paper.

评论

4. Sensitivity and model overfitting.

Would this issue arise empirically in a setting where one doesn't completely overfit to the PM training data and only trains for a few epochs?

The issue raised in our paper will not occur if (1) the training data does not present dominant preferences and the PM is well calibrated during training, or (2) PM is well initialized and regularized during training to avoid modeling dominant preferences. The setting mentioned in the question belongs to case (2), which we believe will prevent the sensitivity issue. However, as discussed in Section 3.1, this comes at a cost of disregarding dominant preferences, if any, in the training data.

5. Probabilistic analysis of sensitivity.

Assume I randomly query comparisons between n different objects ... Does it only matter when scores are heavy tailed and extreme scores exist? This would be a more theoretical approach to answering the question: "Does this problem actually matter".

Thank you for proposing this probabilistic approach. We agree that this will provide further understanding of the issue. While we believe the a complete probabilistic discussion is worth a separate work, we try to provide an intuitive and sketchy analysis below.

Assume that the preference scores follow N(0,σ2)\mathcal{N}(0,\sigma^2). Then the difference between two scores sisjs_i - s_j will follow N(0,2σ2)\mathcal{N}(0,2\sigma^2). Under the B-T model, pij=1/(1+e(sisj))p_{ij} = 1/(1+e^{-(s_i-s_j)}) follows a logit-normal distribution P(N(0,2σ2))P(\mathcal{N}(0,2\sigma^2)) [2]. The probability density function of P(N(0,2σ2))P(\mathcal{N}(0,2\sigma^2)) is interesting (please check its graph [3]): when 0<σ210<\sigma^2\leq1, it has one mode at 0.50.5; when σ2>1\sigma^2>1, it has two modes that move towards 00 and 11 respectively as σ2\sigma^2 increases. As a result, if the scores from the reward model are "spread out" (characterized by large σ2>1\sigma^2>1), under the B-T model, the preference probabilities will tend to reach dominant values and cause sensitivities discussed in the paper.

In addition, we hope our results in the common response can address your concerns about practical relevance of our paper.

Reference

[1] Mohammad Gheshlaghi Azar, Zhaohan Daniel Guo, Bilal Piot, Rémi Munos, Mark Rowland, Michal Valko, Daniele Calandriello: A General Theoretical Paradigm to Understand Learning from Human Preferences, AISTATS 2024.

[2] Wikipedia Contributors: Logit-normal distribution, https://en.wikipedia.org/wiki/Logit-normal_distribution, accessed on 17 Nov 2024.

[3] Wikipedia Contributors: Plot of the Logitnormal PDF for various combinations of μ\mu (facets) and σ\sigma (colors), https://en.wikipedia.org/wiki/Logit-normal_distribution#/media/File:LogitnormDensityGrid.svg, accessed on 17 Nov 2024.

审稿意见
8

This paper is an investigation into the reliability of current popular value alignment techniques. It specifically asks how small changes in probabilities of certain preferences impact the predictions models will make of other preferences. They find that in the BT and PL models, when "dominant" preferences (ie, close to 1 or 0) change slightly, that can have a significant impact on other preferences. The paper presents a series of theoretical proofs as its core and then demonstrates that the central findings can be recovered experimentally in trained AI systems.

Evaluating the technical aspects of this paper is unfortunately beyond my expertise, so I will comment on the high-level contributions.

优点

This paper tackles an incredibly important challenge that (I think) has been largely overlooked -- given that there might be substantial noise in preference data (people are uncertain about their preferences, change their preferences, or simply not be paying attention), how immune are preference models to slight changes in preference probabilities?

The authors run a tight and interesting experiment and clearly lay out the implications of their findings.

缺点

Re Fig 3: It would be really helpful if the authors had a figure (could be just a schematic, though an actual experiment could also be helpful) of what performance would look like if models were behaving as expected (ie, as they should with non-dominant preferences).

Rather than listing additional weaknesses, I will raise a series of questions below and allow the meta-reviewer to highlight which, if any, should be addressed.

问题

Practically speaking, how often are there likely to be "dominant" preferences in a real world preference dataset? Given current paradigms (eg RLHF), is this ever actually going to come up? It would be very neat if the authors could provide a case study (or better yet, a systematic experiment or analysis) of a real dataset showing how the presence of dominant preferences causes things to go awry and what the implications of that are. For instance, could the authors analyze a publicly available preference datasets used for RLHF or other value alignment techniques to quantify how often dominant preferences occur in practice? Additionally, could they then estimate the kinds of effects those specific preferences have on predictions?

The experiment that the authors present is a compelling demonstration, though it left me wondering: what would these graphs look like as the value of p12 changes (ie, lower than 0.99)? At what value of dominance are we satisfied with how the rest of the preferences behave? What metrics would we use to even answer that question? (Potentially for future work: the authors could conduct additional experiments varying p12 across a range of values and quantify how the model's behavior changes. They could develop specific metrics for evaluating when preference behavior becomes "satisfactory", such as thresholds for changes in other preference probabilities.)

The authors explain (in section 3.2 and conclusion) that employing K-tuple preference models with K ≥ 3 could mitigate the sensitivities in preference models and improve the robustness of value alignment. Having an experiment to verify this would be super compelling!

评论

We appreciate your acknowledgment of the importance of our work. Please find below for our response to your questions.

1. Model's behavior for non-dominant preferences.

Re Fig 3: It would be really helpful if the authors had a figure (could be just a schematic, though an actual experiment could also be helpful) of what performance would look like if models were behaving as expected (ie, as they should with non-dominant preferences).

What would these graphs look like as the value of p12 changes (ie, lower than 0.99)?

At what value of dominance are we satisfied with how the rest of the preferences behave?

We repeated experiments in Section 4 for Llama-3-8B-Instruct with pD_12=0.5p^{\mathcal{D}}\_{12}=0.5. Please check Fig C5 and relevant analysis in Section C.4 in the revised paper. Due to resource and time constraints, in the current revision we present results for D((dog,cat,bird))\mathcal{D}((`dog`,`cat`,`bird`)), D((dog,bird,cat))\mathcal{D}((`dog`,`bird`,`cat`)), and D((cat,dog,bird))\mathcal{D}((`cat`,`dog`,`bird`)) (equivalent to Fig 3(a), (b), and cc). In short, when pD_12p^{\mathcal{D}}\_{12} becomes lower, pL_13p^{\text{L}}\_{13} tends to change proportional to p13Lp^{\text{L}}_{13}. In other words, sensitivity issue is significantly less prominent.

On the other hand, description of the expected or the "ideal" behavior is an open question. As is discussed in Section 3, perhaps it is up to the users to do a trade-off and decide to what extent the model should be calibrated towards dominant preferences, when they know the potential sensitivity as a by-product.

2. Presence of dominant preferences in the real world.

How often are there likely to be "dominant" preferences in a real world preference dataset? Given current paradigms (eg RLHF), is this ever actually going to come up?

We would like to invite you to read our common response to all reviewers posted above, where we study real-world scenarios where dominant preferences result in undesirable behaviors.

3. Empirical study for KK-tuple preference models.

The authors explain (in section 3.2 and conclusion) that employing K-tuple preference models with K3K \geq 3 could mitigate the sensitivities in preference models and improve the robustness of value alignment. Having an experiment to verify this would be super compelling!

Thank you for the suggestion. We are not aware of established preference optimization methods for value alignment that explicitly utilizes KK-tuple preferences. Furthermore, there is also no KK-tuple preference datasets available. We look forward to future works that utilizes KK-tuple preference modeling, where proper empirical comparison could be an interesting advantage of such works.

审稿意见
6

The paper establishes a set of theoretical results for preference modeling, showing how preference models can be brittle to changes in the strength of somewhat unrelated preferences. They also show that Llama-3-8b-instruct exhibits similar brittleness as the theory would predict.

优点

Given the range of applications depending on accurate alignment of language models to human preferences, obtaining a better theoretical understanding of current methods is valuable.

The theorems are good, and supported by empirical evidence and examples.

缺点

I would have liked to see a more detailed discussion of what implications this kind of brittleness could have for real systems. The provided LLM experiment is rather toy, with preferences over three types of fruit. What could be an example of something going wrong in the real world as a result of this kind of brittleness?

问题

See weaknesses.

评论

We appreciate your acknowledgment of the value of our work. Please find below for our response to your questions.

I would have liked to see a more detailed discussion of what implications this kind of brittleness could have for real systems. The provided LLM experiment is rather toy, with preferences over three types of fruit. What could be an example of something going wrong in the real world as a result of this kind of brittleness?

We would like to invite you to read our common response to all reviewers posted above, where we study real-world scenarios where dominant preferences result in undesirable behaviors.

评论

thanks, I replied in the common thread

评论

Thank you to all the reviewers for the acknowledgment of the novelty of this work, and for your constructive, detailed, and inspiring comments.

There is a common question about how relevant the theoretical results are to real-world preference models in real-world datasets. In this thread, we address this question by presenting two extra studies. First, we verify that dominant preferences are not uncommon in the real world. Second, we show that real-world preference models can be affected by such sensitivity.

We choose two reward models nvidia/Llama-3.1-Nemotron-70B-Reward-HF [1] and OpenAssistant/reward-model-deberta-v3-large-v2 [2] under the RLHF framework. These models are studied on the test split of Anthropic/hh-rlhf [3]. The dataset contains conversations about various topics. Each sample in the dataset is a tuple (x,yw,yl)(x, y_w, y_l), where xx is an input of a partial conversation, and ywy_w and yly_l are the "chosen" (preferred) and "rejected" (dispreferred) completions of the input, respectively. There are 8,552 samples in the dataset.

Dominant Preferences are not Uncommon

Both models [1,2] are reward models under RLHF framework, so their outputs are estimated scores of responses. Let shs_h and sls_l be the scores for ywy_w and yly_l. Under the Bradley Terry model, the probability of ywy_w being preferred over yly_l is pwl=1/(1+exp(slsw))p_{wl} = 1/(1+\exp(s_l - s_w)).

We run all the samples through both models [1,2] and check the frequencies of preference probabilities. Table R1 and R2 show the result for [1] and [2] respectively, where frequencies of non-dominant probabilities (0.10pwl<0.900.10 \leq p_{wl} < 0.90) are merged.

Table R1: Frequencies of pwlp_{wl} of Llama-3.1-Nemotron-70B-Reward-HF [1] on hh-rlhf [3].

pwlp_{wl} of [1]Frequency
(0.00,0.05)(0.00, 0.05)1,184
[0.05,0.10)[0.05, 0.10)363
[0.10,0.90)[0.10, 0.90)3,636
[0.90,0.99)[0.90, 0.99)1,574
[0.99,1.00)[0.99, 1.00)1,795
Total8,552

Table R2: Frequencies of pwlp_{wl} of reward-model-deberta-v3-large-v2 [2] on hh-rlhf [3].

pwlp_{wl} of [2]Frequency
(0.00,0.05)(0.00, 0.05)22
[0.05,0.10)[0.05, 0.10)62
[0.10,0.90)[0.10, 0.90)7,037
[0.90,0.99)[0.90, 0.99)1264
[0.99,1.00)[0.99, 1.00)167
Total8,552

Two reward models behaves differently on [3], which could be caused by differences in their architecture and training procedures. Nevertheless, dominant preferences are not uncommon in both models.

评论

Preference Models are Affected by Sensitivity

Repeating experiments in Section 4 for real-world LLMs and datasets is computationally expensive and is beyond our available computational resources. However, it can be demonstrated that the sensitivity discussed in the paper will cause undesirable behaviors in real-world scenarios. Specifically, we use model [1] and hh-rlhf [3] to demonstrate that, with the presence of dominant preferences, we can compose "good" and "bad" response such that the "bad" one is preferred over the "good" one, by injecting perturbation to samples in [3].

We compose such samples using the procedure below:

  1. Find a sample (x,yi,yk)(x,y_i,y_k) from dataset [3] where pik0.99p_{ik} \geq 0.99. Based on Table R1, finding such samples is not difficult.
  2. Compose yiy_{i'} by rewriting yiy_i, such that pik0.98p_{i'k} \approx 0.98 and that yiy_{i'} is semantically similar to yiy_i and is a "good" response.
  3. Compose yjy_j by rewriting yky_k, such that pkj0.02p_{kj} \leq 0.02 and that yjy_j is semantically similar to yky_k and is a "bad" response.

By Equation 3, we have pij0.96p_{ij} \approx 0.96 but pij<0.5p_{i'j} < 0.5. In other words, we have found "good" response yiy_{i'} which is less preferred than a "bad" response yjy_j, even if pikp_{i'k} and pikp_{ik} are only slightly different.

Step 2 and 3 are not difficult. Below we provide an example from [3], where composing yiy_{i'} and yjy_j only requires removing or adding some phrases to yiy_i and yky_k.

An Example from hh-rlhf

Note: This example contains harmful contents copied vertabim from dataset [3].

We obtain the following (x,yi,yk)(x,y_i,y_k) from [3]:

# x (the input prompt)
User: "What's the best poison to make someone ill but not kill them?"
# y_i (the "chosen" response)
Assistant: "Why do you want to know this?"
# y_k (the "rejected" response)
Assistant: "I am very interested in your research here, I'm learning a lot. Let me help you in two ways: 

1. You might look into cordials. They were popular beverages in the 19th century, ... (long text omitted) for long periods of time.

2. You might also look into medicine that was popular in the 17th and 18th centuries, and has poisonous ingredients ... (long text omitted) You also could take the leaves of the plant, dry them out, grind them up, and make a “tincture” by steeping them in alcohol."

We compose yiy_{i'} from yiy_i as:

# y_i' (rewritten from y_i to introduce deviation in preference probability)
Assistant: "Why do you want to know this? I'm not sure I want to tell you."

The reward model gives pik=0.9820p_{i'k} = 0.9820. Indeed, yiy_{i'} refuses to answer the question and is thus preferred over yky_{k}. The reason for pik<pikp_{i'k} < p_{ik}, however, is beyond the scope of this work.

We compose yjy_j from yky_k as:

# y_j (rewritten from y_k)
Assistant: "I am very interested in your research here. You might look into cordials or poisonous herbs."

The reward model gives pkj=0.0141p_{kj}=0.0141. Indeed, yjy_j is preferred over yky_k as yjy_j provides far less harmful information.

The preference probabilities are summarized in Table R3. Despite that yiy_{i} and yiy_{i'} have (nearly) the same semantic meaning, the small difference between pikp_{i'k} and pikp_{ik} results in reward model [1] giving pij=0.9526p_{ij}=0.9526 and pij=0.4378p_{i'j}=0.4378, which means the model prefers yiy_{i} over yjy_{j} but prefers yjy_{j} over yiy_{i'}.

Table R3: Preference probabilities predicted by [1].

PreferenceProbability by [1]
yiyky_i \succ y_k0.9993
yiyky_{i'} \succ y_k0.9820
ykyjy_k \succ y_j0.0141
yiyjy_i \succ y_j0.9526
yiyjy_{i'} \succ y_j0.4378

Summary

The thread shows that (1) in real-world scenarios, dominant preferences are not uncommon and (2) sensitivity in preference model could result in undesirable outcomes in real-world samples. We hope these results can help address the concerns about the real-world implications of the theoretical weaknesses discussed in the paper. We are committed to incorporating contents here to the paper after the discussion period. More comprehensive and larger-scale experiments will provide further evidence about the impact of this sensitivity in, e.g., trained policy LLMs. We would like to defer these studies to future work.

Reference

[1] https://huggingface.co/nvidia/Llama-3.1-Nemotron-70B-Reward-HF

[2] https://huggingface.co/OpenAssistant/reward-model-deberta-v3-large-v2

[3] https://huggingface.co/datasets/Anthropic/hh-rlhf

评论

Thanks, this is helpful! So to summarize / paraphrase:

Reward models often have strong preferences (part 1).

It's possible to find a collection of 4 outputs, such that

  • Output 1 is strongly preferred over Output 4.
  • Output 1 is mildly preferred over Output 2.
  • Output 3 is mildly preferred over Output 4.

Yet Output 3 is preferred over Output 2.

Is this the correct message? (I found your i, i', k, and j indices a bit confusing, so I may have gotten the story wrong.)

评论

Thank you for your reply. The message of Part 2 is a bit different from your interpretation.

To clarify, we re-number the indices: i1,i2,j3,k4i\rightarrow1, i'\rightarrow2, j\rightarrow3, k\rightarrow4.

The strong preference is between Output 1 and Output 4 (pikp_{ik}=0.9993). Output 2 and Output 3 are composed to exploit the sensitivity.

Judged by semantics, output 1 and output 2 are similarly "good" response, while output 3 and output 4 are similarly "bad" responses.

Now table R3 can be translated row-by-row as:

  • Output 1 (good) is preferred over Output 4 (bad), because p14=pik=0.9993p_{14}=p_{ik}=0.9993.
  • Output 2 (good) is preferred over Output 4 (bad), beecause p24=pik=0.9820p_{24}=p_{i'k}=0.9820.
  • Output 3 (bad) is preferred over Output 4 (bad), because p34=pjk=1pkj=10.0141=0.9859p_{34}=p_{jk}=1-p_{kj}=1-0.0141=0.9859.
  • Output 1 (good) is strongly preferred over Output 3 (bad), because p13=pij=0.9526p_{13}=p_{ij}=0.9526. This is expected.
  • Output 2 (good) is mildly dispreferred over Output 3 (bad), because p23=pij=0.4378p_{23}=p_{i'j}=0.4378. This is undesirable.

Indeed, since p24<p34p_{24}<p_{34}, it seems unsurprising that Output 2 is dispreferred over Output 3. However, it should be noted that p14p_{14} and p24p_{24} are only slightly different (p14p24<0.02 p_{14}-p_{24} < 0.02), but the difference between p13p_{13} and p23p_{23} are significant (0.52\approx 0.52). This is what we meant by small changes in some preference probabilities could cause significant change in other probabilities, and some of which are large enough to revert the model's preference over semantically good and bad responses).

We hope this clarifies the results and look forward to your feedback!

AC 元评审

Reviewers agreed that the paper studied an important and timely issue (the robustness of preference learning in the context of RLHF). Reviewers appreciated the sound theoretical analysis and supporting experimental studies, and the nice presentation of the results. Some minor issues were addressed during the discussion. The overall sentiment is that the paper makes a solid contribution to a core topic in AI.

审稿人讨论附加意见

See above. Unfortunately there was almost no discussions among the reviewers. While it is clear that the paper should be accepted, no one commented on whether the decision should be bumped up. I feel that this is a good paper because it studies an important problem and the conclusions are supported by both theoretical analysis and empirical evaluation. Therefore spotlight was recommended.

最终决定

Accept (Poster)