Dueling Convex Optimization with General Preferences
convex optimization, dueling feedback, comparison, preference, transfer function, smooth convex functions, convergence rate, strongly convex functions
摘要
评审与讨论
This paper considers the problem of dueling bandits with general preferences, where the preference model between two decisions (called the transfer function) is not specified to some specific choices. The most technical challenge in this problem is the estimation of gradients since only the preference feedback (a one-bit feedback generated from a Bernoulli distribution) is accessible, which is even harder than the one-point bandit feedback. To this end, the authors proposed a novel estimation of , which does not necessarily align with the true gradient but is still sufficient for a gradient descent update. As for results, the authors achieved an oracle query complexity to obtain an -optimal decision. Furthermore, for strongly convex functions, the aforementioned bound can be further improved to using a similar idea like epoch-GD, which is theoretically optimal.
给作者的问题
Not many.
论据与证据
Yes.
方法与评估标准
Yes.
理论论述
Yes, mostly.
实验设计与分析
There are no experiments in this paper.
补充材料
Yes.
与现有文献的关系
N/A.
遗漏的重要参考文献
No.
其他优缺点
Strengths
-
Studying general preference functions is a meaningful problem, which may be further applied to real-world applications, for example, RLHF.
-
The adopted assumption is mild and reasonable. The authors have provided detailed explanations to validate its rationality.
-
The proposed algorithm is simple and easy to implement. It generally follows the update rule of gradient descent but with a carefully chosen gradient estimation. The key challenge lies in the technical analysis of this update rule.
Weaknesses
- The guarantee of Theorem 2 only holds for some unknown decision generated in the optimization process. To find it out, additional operations are needed, which will lead to extra query complexity.
其他意见或建议
I noticed that Thm 2 has an assumption of . What if the required error ? I guess this should not be a big deal since the problem would become easy if the required error is large.
“The guarantee of Theorem 2 only holds for some unknown decision generated in the optimization process. To find it out, additional operations are needed, which will lead to extra query complexity.”
You are right to notice that, due to the challenging feedback model, the nature of our convergence guarantee becomes more intricate than simply holding for the final iterate of the algorithm. We address this issue in detail in the paragraph in lines 167-177 (right column). For additional details regarding the factor see our reply to reviewer br4X.
“What if the required error ”
Indeed, as you probably realized, in this case we can simply set and obtain an only stronger guarantee, and the convergence bound will be independent of and it will only depend on the problem parameters .
Thanks to the authors for the reply. I acknowledge the contributions of this work. However, as mentioned in the review comments, the condition on seems to be an assumption, which would confuse readers. As the authors have said, when the required error is larger, a stronger guarantee can be achieved. As a result, I highly recommend that the authors could add this case and its corresponding results to the theorem to make it self-contained. I maintain my current score.
We appreciate the reviewer’s additional comments and thoughtful discussion. We agree that a comment about the condition on is in order, to clarify that it does not limit the generality of our result in any way. We will carefully incorporate this into the final version - thanks!
This work studies convex optimization with dueling feedback and general transfer functions. The main contribution is an algorithm with convergence rate for smooth and convex functions and convergence rate for smooth and strongly-convex functions, where is the minimal degree (with a non-zero coefficient) in the transfer’s series expansion about the origin.
给作者的问题
Are there any example of concrete problems in practice where ?
论据与证据
The statements are supported with proofs.
方法与评估标准
NA
理论论述
The proofs in the main paper seem to be correct.
实验设计与分析
NA
补充材料
I didn't review the supplementary material.
与现有文献的关系
NA
遗漏的重要参考文献
NA
其他优缺点
Strength: This work improves the prior art by considering a very general class of transfer functions where can be greater than . The rates in smooth and strongly convex objectives match the existing lower bounds.
Weakness: However, this setting seems not very important. This work lacks discussions of meaningful examples and concrete problems/applications in practice.
其他意见或建议
NA
“Are there any example of concrete problems in practice where p>1?”
First, we should note that even for our results constitute the first convergence bounds for convex optimization with (approximately) linear transfer functions.
Next, to your question, it is worth recalling what the transfer function abstracts. One motivation is similar to Reinforcement learning from human feedback (RLHF), where the transfer function abstracts how human preference behaves given similarly-valued alternatives. The most natural one is to assume that when the alternatives are of similar values, humans select almost randomly. The degree abstracts how small changes in the quality of the two alternatives are translated to differences in human evaluations. A higher value implies that differences are harder to detect. For this reason, is equally well-motivated as , and there is no particular fundamental reason to model preferences using a linear function.
Again, note that the transfer function transfers between the valuations (e.g. in LLM: how good is a response to a prompt) to the human response (e.g. in LLM: the likelihood that a human will prefer each alternative). There is no reason to believe the transfer should be necessarily linear.
This paper proposes an algorithm for the setting of dueling convex optimization, with a broad class of transfer functions.
给作者的问题
- Could the authors further elaborate the following statement to solve the problem of unknown time index that achieves good performance: Since errors accumulate only along the path from the root to the true minimum, both the accuracy and confidence of the entire comparison procedure will decrease by a factor.
- Can the assumption of be relaxed? Can the results be improved if is an odd function?
论据与证据
The claims are well supported.
方法与评估标准
The proposed methods make sense.
理论论述
I checked the proof for the non-strongly-convex case.
实验设计与分析
N/A
补充材料
I reviewed all of the supplementary materials.
与现有文献的关系
The two most related works are Jamieson et al. (2012) and Saha et al. (2021). This work is a natural extension of Saha et al. (2021).
遗漏的重要参考文献
N/A
其他优缺点
Strengths:
- This work follows a clear logic, and is technically solid, especially the analysis about the stopping time.
Weaknesses:
- This work could be improved by justifying the algorithm design with experiments.
- In general, the paper lacks interpretation of the results. I was not sure whether the proposed algorithm can be compared against some naive algorithms, given that the algorithm is the first one for the setting being studied. The choice of hyperparameters also need more discussions.
- The algorithm design of this paper is mostly an extension of Saha et al. (2021b). The authors could have over-claimed the contribution of the algorithm design, especially the "relative gradient".
其他意见或建议
1, The concept of "admissible transfer functions" appears abruptly in the title of Subsection 2.2, without further explanations. 2. In terms of the structure of the paper, the main text includes too much proof detail and inadequate discussions of the main theorems.
“This work could be improved by justifying the algorithm design with experiments”
This is primarily a theoretical work in optimization with partial feedback, and our main goal is to understand the fundamental achievable convergence bounds in this setting. This is also the focus of much of the prior work in this space. We feel that the provided theory gives ample justification for the algorithm.
“I was not sure whether the proposed algorithm can be compared against some naive algorithms, given that the algorithm is the first one for the setting being studied”
When we have dueling feedback, it is even challenging to design “naive” algorithms for the general convex (and smooth) case. One could view the existing algorithm of Jamieson et al (2012) as an attempt to generalize a natural noisy binary search algorithm to a high-dimensional setting, but unfortunately, their convergence requires strong assumptions such as strong convexity as their extension relies on a coordinate descent scheme. Our main contribution is to give precise convergence bounds for the general convex (and smooth) dueling setting.
“The choice of hyperparameters also need more discussions”
The hyperparameters of our algorithms are set to guarantee the best performance bound we can derive. The rationale for their choice is to optimize, in hindsight, the convergence bound we establish.
“The algorithm design of this paper is mostly an extension of Saha et al. (2021b)”
First, we should emphasize that, in our view, expanding the scope of comparison-based / dueling optimization to a variety of nonlinear transfer functions is novel and significant. In fact, even the most well-studied transfers, like the sigmoid, were not covered by the existing literature in this context. Our aim was precisely to fill in this gap.
Regarding technical novelty, specifically compared to Saha et al. (2021b): on a very high level, like Saha et al. we follow the approach of designing a gradient estimator using the dueling feedback and using it for running gradient descent. However, our gradient estimation analysis is very different from that of Saha et al., due to the generality and nonlinearity of the transfer function in the feedback. This can be seen in Lemma 3 and its proof, which analyzes the gradient estimator in the case of nonlinear transfer: roughly, we do this by inspecting the local polynomial behavior of the transfer around zero and working out the effect of the polynomial transformation on the expected value of the gradient estimator. In the estimation process, the magnitude of the gradient gets distorted (see the leading factor in the expected value) so we can no longer use the standard subgradient descent analysis, which we replace with a different approach akin to the analysis of normalized gradient descent.
We agree a more detailed explanation of the technical novelty compared to Saha et al. should be included in the paper itself - we will improve this in the final version.
Other Comments Or Suggestions:
Thank you for your suggestions, we will incorporate them in the final version of the paper.
“Since errors accumulate only along the path… will decrease by a factor”
We will have at the end of the run points, we would like to select the minimum of those corresponding function values. If we had exact comparisons, we could build a complete binary tree over the points, and compute the minimum by selecting at each node of the tree the minimum value between its descendents. With our actual noisy comparisons, we have “errors” of magnitude at each node of the tree, and if we sum these errors along the path to the minimal leaf, the error is amplified by a factor (which is the depth of the tree). Alternatively, we can start with a goal error of and end with a cumulative error of .
“Can the assumption of be relaxed?”
It is worth recalling what the transfer function abstracts. One motivation is similar to Reinforcement learning from human feedback (RLHF), where the transfer function abstracts how human preference behaves given similarly-valued alternatives. The most natural one is to assume that when the alternatives are of similar values, humans select almost randomly. In our formal feedback model, this essentially implies that .
Having said that, technically our algorithm works without modifications in the case where , since it is agnostic to additive bias in (this can be seen from the statement of Lemma 9 in the appendix).
“Can the results be improved if is an odd function?”
We are unaware of whether the bounds can be improved for odd transfer functions. In fact, our preliminary results were initially for odd functions and we were happy to see that they extend naturally to more general transfer functions, with the same rates, as we describe in the paper.
I appreciate the authors for the response. The rebuttal revolves most of my concerns, especially the technical contributions. I have raised my score, and it would be great to see the discussion about the technical novelty compared with previous works in the revised version.
We appreciate the reviewer's engagement and openness to revisit the initial concerns after reading our response. And thanks again for the thoughtful and constructive feedback---we will be sure to incorporate these discussions in the final version.
Dear Authors,
Thank you for submitting your paper to ICML and for your contribution to the theoretical foundations of optimization. This work tackles the convex optimization setting under general preference-based comparisons and introduces a novel algorithm that achieves optimal convergence rates for both smooth and strongly convex functions.
The reviewers found the paper technically solid and appreciated the generalization of prior results. Reviewer br4X, after engaging with the rebuttal, increased the score, noting the clarity of the contributions. Other reviewers highlighted the elegance of the gradient estimator and the soundness of the theoretical analysis, while noting that the lack of experiments and applied motivation. Reviewers iszU and WhdB pointed out the importance of articulating practical examples and relaxing assumptions in theorems, which was addressed in the response.
Given the strength of the theoretical contributions, the soundness of the analysis, and the reviewers’ overall support, I am pleased to recommend acceptance to ICML 2025. I encourage you to clarify the assumptions and extend the discussion in the final version to help a broader audience understand the potential applications and interpretability of your results.
Congratulations on your contribution.
Best regards, AC