PaperHub
6.6
/10
Poster4 位审稿人
最低3最高4标准差0.5
3
4
3
4
ICML 2025

Clustering Items through Bandit Feedback: Finding the Right Feature out of Many

OpenReviewPDF
提交: 2025-01-23更新: 2025-07-24
TL;DR

We consider a problem of clustering in an adaptive and sequential setting, and provide an efficient and optimal algorithm.

摘要

We study the problem of clustering a set of items based on bandit feedback. Each of the $n$ items is characterized by a feature vector, with a possibly large dimension $d$. The items are partitioned into two unknown groups, such that items within the same group share the same feature vector. We consider a sequential and adaptive setting in which, at each round, the learner selects one item and one feature, then observes a noisy evaluation of the item's feature. The learner's objective is to recover the correct partition of the items, while keeping the number of observations as small as possible. We provide an algorithm which relies on finding a relevant feature for the clustering task, leveraging the Sequential Halving algorithm. With probability at least $1-\delta$, we obtain an accurate recovery of the partition and derive an upper bound on the required budget . Furthermore, we derive an instance-dependent lower bound, which is tight in some relevant cases.
关键词
clusteringbandit theorypure explorationinformation-theoretic boundsmachine learning

评审与讨论

审稿意见
3

This paper studies a problem of clustering nn items into two groups using bandit feedback. This setting considers an n×dn \times d matrix MM, where each row represents an item's feature vector. The nn rows are partitioned into two unknown groups, such that items within the same group share the same feature vector. The learner sequentially and adaptively an item It[n]I_t \in [n] and a feature Jt[d]J_t \in [d], and gets a noisy observation drawn from an unknown distribution with mean MIt,JtM_{I_t,J_t}. The goal is to recover the correct item partition with minimal observations. The authors propose an algorithm called BanditClustering, which operates in two steps: (1) identifying two representative items from different groups, and (2) selecting a discriminative feature to classify all items. The authors provide theoretical results such as a tight upper bound on the required budget and a matching instance-dependent lower bound. Numerical experiments validate the efficiency of the method compared to non-adaptive clustering approaches.

update after rebuttal

Overall, I appreciate the technical novelty of this paper. Therefore, I decided to maintain my positive score.

给作者的问题

  1. Why do the authors restrict their analysis to only two clusters? What challenges arise when extending the approach to multiple clusters?
  2. How does the proposed algorithm perform when the data is imbalanced (i.e., θ\theta is small), and how does it compare to the baselines?

论据与证据

All claims are well-supported by clear and convincing evidence.

方法与评估标准

The proposed methods and evaluation criteria make sense for the problem.

理论论述

I have not checked all the proofs in detail but did not identify any obvious errors.

实验设计与分析

The experiments rely solely on synthetic data and consider only uniform sampling and K-means as baselines. It would be better to compare other adaptive clustering methods, such as the ones mentioned in the related work. Also, the comparison does not examine performance in highly imbalanced scenarios (when θ\theta is small).

补充材料

I reviewed the source code in the supplementary material and found no obvious issues.

与现有文献的关系

This work extends the bandit clustering problem studied in prior works, such as Thuot et al. (2024) and Yang et al. (2024). Unlike these studies, where the entire noisy feature vector of an item is observed at each step, the setting introduced in this paper allows only a single noisy feature to be observed per step. This introduces a new and inherently more challenging problem.

遗漏的重要参考文献

N/A.

其他优缺点

Strengths

  1. This paper is well-organized and easy to follow.
  2. This paper studies a new bandit clustering problem where at each step, only a single noisy feature of a given item is observed. This setting is different from previous studies where the entire noisy feature vector is observed.
  3. The authors provide rigorous upper and lower bounds on sample complexity, ensuring that their method is provably efficient.

Weaknesses

  1. The algorithms and theoretical analysis assume only two clusters, which is a strong limitation. While the authors suggest that they could "straightforwardly extend" their methodology to multiple clusters, this extension is not formally developed or analyzed.
  2. The paper primarily focuses on theoretical contributions and lacks strong motivation. Although the introduction includes a motivating example, it does not clearly articulate the significance of the problem.

其他意见或建议

Quotation marks in LaTeX should be formatted using `` and '' instead of " ".

The legend in Figure 1 is unclear. What does the algorithm "Cluster" refer to? Also, it seems that the uniform sampling algorithm is not explicitly labeled. Furthermore, what does the shaded area indicate?

作者回复

First, we would like to thank you for your time and effort in reviewing our paper. Below, we address the key remarks and questions raised in your review:

  • Extension to K>2K>2 clusters

    In the paper, we analyze the case of two clusters, as even in this simpler setting, significant challenges appeared to obtain optimality. It happens that it is possible to use our algorithms as subroutines in order to handle the extension to KK cluster. To avoid redundancy, we invite you to read the answer to reviewer RdCv where this extension is discussed in detail. When dealing with K>2K>2 groups, it is significantly more challenging to find a strategy which adapts optimally to the relative position of the centers of the KK groups, we will leave this question for future works.

  • Dependency on the balancedness θ\theta

    Thank you for the question. From a theoretical point, we do prove in the paper that the balancedness parameter θ\theta has only an impact on the first step of our procedure, i.e., the identification of representatives. Furthermore, gathering Proposition C.1 (analysis of Algorithm 2), and the lower bound in Lemma E.1, we are able to prove that our procedure exhibits the optimal dependency on the budget with respect to θ\theta. The only difficulty to deal with very unbalanced clusters is that it is hard to detect a row in the smallest cluster. We will add further discussion about the dependency on θ\theta of our method in the final paper.

    We have also run numerical experiments to investigate the influence of θ\theta. They will be included in the paper. In a first setup, we are able to see that indeed, θ\theta mainly influences the CandidateRow-step of our clustering procedure and that the influence on the budget of this step is comparable to a factor 1/θ1/\theta, which fits in with our theoretical findings. Furthermore we can see that in this setup, our algorithm clearly outperforms uniform allocation strategies for very small values of θ\theta, while being competitive for larger values of θ\theta.

  • Suggestions

    Thank you to your suggestion, we will correct all quotation marks, and improve the explanation and legend of Figure 1.

  • Motivation

    We thank the reviewer for pointing out that we should extend the discussion of our motivating example, which comes from image labeling. Recent work about distinguishing "doppelganger" animals uses expert annotators to ensure a correct labeling (see [Herde, Marek, et al. "dopanim: A Dataset of Doppelganger Animals with Noisy Annotations from Multiple Humans." Advances in Neural Information Processing Systems 37 (2024): 51085-51117.]). This works because experts know which features are relevant for a correct distinction. If no experts are available, our approach still provides a solution, because we can find the relevant features without prior knowledge and use non-expert workers for the classification task.

  • Numerical experiments

    As mentioned above, we will illustrate the dependency with θ\theta in an additional numerical experiment. Besides, we will also compare our method with the method from [Ariu et al.(2024) "Optimal clustering from noisy binary feedback"]).

审稿意见
4

This paper investigates the problem of clustering items via bandit feedback. The items can be partitioned into two unknown groups that share the same feature vector within each cluster. The authors propose a sequential and adaptive setting where the learner can only select one item-feature pair per round. The objective is to accurately recover the correct partition while minimizing the number of observations. The paper presents an algorithm that identifies a relevant feature for clustering by leveraging the Sequential Halving algorithm. With probability at least 1δ1-\delta, the algorithm achieves accurate recovery of the partition, and the authors derive an upper bound on the required budget. Additionally, they establish an instance-dependent lower bound that is tight in certain relevant cases.

给作者的问题

Please comment/discuss the weakness part. I am happy to raise my score if these questions are well discussed/solved.

-------------Post-rebuttal----------------

The reviewer thanks the authors for their detailed responses, including the discussion on extending to more than two groups (K>2K > 2), addressing heterogeneous groups, elaborating on the trade-offs between dd and Δ\Delta, and discussing the missing references. My concerns have been resolved, and I would like to raise my score. I encourage the authors to incorporate these discussions into the final version to improve clarity and completeness of the current work.

论据与证据

Yes

方法与评估标准

Yes

理论论述

Yes, I checked the main content of the paper.

实验设计与分析

Yes.

补充材料

No.

与现有文献的关系

Contributions:

  1. Introduction of a new bandit cluster model where only one item-feature pair can be selected in each round, distinguishing it from existing works that select an item with all features.
  2. Development of the first lower bound that characterizes the fundamental difficulty of the bandit clustering problem.
  3. Design of a near-optimal bandit clustering algorithm with theoretical guarantees.

遗漏的重要参考文献

There is another line of clustering of bandits’ works. For example:

  1. Gentile, Claudio, Shuai Li, and Giovanni Zappella. "Online clustering of bandits." In International conference on machine learning, pp. 757-765. PMLR, 2014.
  2. Li, Shuai, Wei Chen, and Kwong-Sak Leung. "Improved algorithm on online clustering of bandits." arXiv preprint arXiv:1902.09162 (2019).
  3. Liu, Xutong, Haoru Zhao, Tong Yu, Shuai Li, and John CS Lui. "Federated online clustering of bandits." In Uncertainty in Artificial Intelligence, pp. 1221-1231. PMLR, 2022.
  4. Li, Zhuohua, Maoli Liu, Xiangxiang Dai, and John Lui. "Demystifying Online Clustering of Bandits: Enhanced Exploration Under Stochastic and Smoothed Adversarial Contexts." arXiv preprint arXiv:2501.00891 (2025).

其他优缺点

Weaknesses:

  1. In Section 6, the authors discuss two important extensions: increasing the number of groups (K>2K>2) and handling heterogeneous groups, which would better model real-world applications. However, the discussion provides only vague ideas without specific results for each extension. More concrete analysis or preliminary results would strengthen the paper's practical relevance.
  2. Regarding the upper bound result, there appears to be a factor of dd in the regret upper bound in Equation (3). It seems possible that fixing one feature to explore could remove this d factor, potentially at the cost of a larger Δ\Delta term. The paper would benefit from a more thorough discussion of this trade-off, especially in cases where the d improvement might dominate any negative effects on the Δ\Delta term.

其他意见或建议

Minor Issues:

There are several typographical errors throughout the paper. For example, on line 230, a sentence is incomplete.

作者回复

First, we thank you for your time and effort in reviewing our paper. We will correct the typographical errors you identified. Below, we discuss about the different weaknesses that you identified. We will provide intuitions and thorough discussions on these points in the final version of the paper.

  • Extension to K>2K>2 clusters

    Our algorithm identifies a discriminative feature to separate two groups. To achieve this. We fix arbitrarily a first item, and find an item from a different cluster (Algorithm 2). Then, we identify a feature that separates the two clusters well, balancing the cost of identifying such a feature with the cost of classification (Algorithm 3). We proved that this requires a budget of H(μaμb)log(1/δ)H(\mu^a-\mu^b)\log(1/\delta) (see Thm 3.1 and Eq. (3)).

    Here is a way to extend the method to K>2K>2 clusters, assuming KK is known. We first identify KK representatives from different clusters, one at a time. If k<Kk<K items have already been identified from kk different clusters, we can run kk calls of Algorithm 1 in parallel to identify an item whose mean vector differs from all the previously identified representatives. Repeating this operation KK times yields KK representatives. Then, we use these representatives to learn K(K1)/2K(K-1)/2 discriminative features (one for each pair of representatives), ultimately classifying all items. Compared to the case where K=2K=2, the total budget then scales by a factor of K2×minkkH(μkμk)K^2 \times \min_{k\ne k'} H(\mu^k-\mu^{k'}), where μ1,,μK\mu^1,\dots,\mu^K are the mean vectors of the KK clusters. This approach would be order-optimal if KK is considered a constant.

We will add this procedure for K>2K>2 clusters in the Appendix of the final version, along with this (non-optimal) bound. However, finding a distribution-dependent optimal strategy, taking into account the relative positions of all means μ1,,μK\mu^1,\dots, \mu^K, remains highly non-trivial and is way beyond the scope of our paper.

  • Robustness to heterogeneity within the clusters

    Actually, our method can be adapted to deal with problems where there exists some heterogeneity inside the groups. We invite you to read the answer in the rebuttal to reviewer AdRS.

  • Intuition on the complexity HH for sparse vector

    We would like to provide more intuition on the trade-off HH (see Eq. (3)). In particular, we will explain why a dependency in the dimension dd is unavoidable and that HH is the optimal trade-off between dd and Δ\Delta.

    Most of our intuitions about the problem lies in the simple case (Corollary 3.2), where the gap vector is ss-sparse with a constant magnitude hh. In this case, the l2 norm of the gap vector Δ\Delta (appearing in HH) becomes Δ22=sh2||\Delta||_2^2=sh^2. Before explaining the intuition behind this rate, we emphasize that the lower bound matches Theorem 4.1 in this setting thereby showing that our budget is indeed optimal.

    To perform the clustering task, we need to select a feature that discriminates the two groups. if one item in each cluster are identified , this problem is equivalent to finding a good arm of reward hh among dd. In this case, a simple algorithm --which is at the core of our approach-- proceeds as follows.

    • First, sub-sample dsθlog(1/δ)\frac{d}{s\theta}\log(1/\delta) entries from the matrix, so that, with high probability, a nonzero entry is selected. If s=ds=d, the problem reduces to one dimension; if s=1s=1, all features must be explored.
    • Second, sample all these entries 1h2log(1/δ)\frac{1}{h^2}\log(1/\delta) times, select the entry with the largest (absolute) mean, and store the associated feature.
    • Finally, sample 1h2log(1/δ)\frac{1}{h^2}\log(1/\delta) each item on this feature and classify the items.

    In the paper, our algorithm (especially through the use of sequential halving) adapts to the unknown parameters ss and hh, and leads to a budget dθsh2log(1/δ)2+nh2log(1/δ)\frac{d}{\theta sh^2}\log(1/\delta)^2+\frac{n}{h^2}\log(1/\delta). We can argue that all these steps cannot be improved with respect to d,s,hd,s,h. We will provide more intuition on this optimal trade-off in the final version. Also, you can find a discussion on the optimality of HH for general shapes of Δ\Delta in the rebuttal of reviewer z8W5.

  • Literature on online clustering of bandits

    We appreciate your suggestion regarding references on the problem of online clustering of bandits and its extensions. This problem is indeed related to our setting, as it involves exploring a bandit environment where the items exhibit an underlying clustering structure. We will include a discussion on the similarities and differences between our approach and this interesting line of research in the literature review of our paper. Still, we would like to point out there are two major differences with our problem: (1) the learner has no control over which items are presented at each time step, and (2) the algorithms (such as CLUB and its extensions) are evaluated in the cumulative regret setting.

审稿意见
3

This paper considers the problem of classifying nn items into two categories based on bandit feedback. Each item is associated with dd features that vary depending on the class it belongs to. In each observation, an item and a feature are selected, and the corresponding feature value is observed with noise. Under this setting, the paper focuses on the fixed confidence regime, evaluating the minimum number TT of observations required to achieve a given accuracy level. It proposes an algorithm along with an upper bound and a nearly matching lower bound. Additionally, the proposed algorithm is evaluated through numerical experiments.

给作者的问题

At the end of Section 4, it is stated, "For more general Δ\Delta, we conjecture that the trade-off in HH in (3) is optimal and unavoidable." Could you provide any supporting facts or intuitions that lead to this conjecture?

论据与证据

The main contributions of this paper are theoretical, and all appear to be supported by correct proofs.

方法与评估标准

The evaluation metric used in this paper is natural and reasonable.

理论论述

I have briefly checked the proofs of Theorem 3.1 and Theorem 4.1. No particular issues were found.

实验设计与分析

I was not able to spend much time verifying the correctness of the numerical experiments.

补充材料

I have briefly checked the proofs of Theorem 3.1 and Theorem 4.1.

与现有文献的关系

The problem setting considered in this paper is a special case of the problem studied in (Ariu et al., 2024). However, to the best of my knowledge, this is the first paper to provide theoretical guarantees and analysis for this type of problem.

遗漏的重要参考文献

I am not aware of any particular relevant literature not discussed.

其他优缺点

As mentioned in Section 4, one limitation is that the matching upper and lower bounds are restricted to specific cases of Δ\Delta. Additionally, the fact that the number of groups is limited to two poses a practical constraint. While Section 6 outlines a general approach for extending the method to three or more groups, obtaining tight upper and lower bounds in such cases would likely require nontrivial further analysis. That being said, this paper serves as a strong starting point for exploring broader problem settings. Overall, the paper is written in a very clear and readable manner.

其他意见或建议

There seems to be an extraneous period at the end of Extension to a larger number of groups. in Section 6.

作者回复

First, we would like to thank you for your time and effort in reviewing our paper. We corrected the small typo in Section 6 that you pointed out. Besides this, we would like to provide more insights regarding the two limitations you mentioned.

  • Discussion on the optimal trade-off HH (Eq. (3)) for general Δ\Delta

    In Corollary 3.2, we prove that our procedure is optimal when the gap vector is ss-sparse with a constant magnitude hh. You can find a discussion on the intuitions behind HH on this sparse setting in the rebuttal to reviewer RdVc. For a general gap vector Δ\Delta, we have good reasons to think that understanding the optimality of this trade-off for this simple example allows us to understand (at least intuitively) the optimality for general vectors.

    Our upper bound is constituted of two terms. A first term dlog(1/δ)θΔ2\frac{d\log(1/\delta)}{\theta||\Delta||^2}, we proved that this is optimal (see Thm 4.1) for the sub-problem of identifying an item in the second group. The second term is defined as this minimum mins[d][dsΔ(s)2+nΔ(s)2]log(1/δ)\min_{s\in[d]} \left[\frac{d}{s\Delta_{(s)}^2} + \frac{n}{\Delta_{(s)}^2}\right] \log(1/\delta). Indeed, the term nΔ(s)2log(1/δ)\frac{n}{\Delta_{(s)}^2} \log(1/\delta) is the price for clustering if we use a feature with a gap Δ(s)|\Delta_{(s)}| while dsΔ(s)2log(1/δ)\frac{d}{s\Delta_{(s)}^2}\log(1/\delta) corresponds to the price for selecting a feature with a magnitude at least Δ(s)|\Delta_{(s)}|. Define the effective sparsity ss^* for which the minimum holds, and define the effective magnitude as Δ(s)\Delta_{(s^*)}. Intuitively, we can argue that entries significantly larger than Δ(s)\Delta_{(s^*)} are too rare to be detected (otherwise, ss^* would be smaller), and entries much smaller than Δ(s)\Delta_{(s^*)} are too weak to be used for classification. Our insight is that the problem is as hard as if the gap vector were ss^*-sparse with a constant magnitude Δ(s)\Delta_{(s^*)}, a setting where we have matching lower and upper bounds (see corollary 3.2), leading to our complexity.

    We believe that proving formally this optimality is difficult in general. We have currently a sketch of proof for which we reduce the problem into a problem of good-arm identification -- we prove that each algorithm that solves the problem should be able to identify a feature larger (up to a constant) than Δ(s)\Delta_{(s^*)} in the gap-vector. However, we would then need an instance-dependent lower bound for such problem. Unfortunately, contrary to the problem of best-arm identification, such a lower bound does not exist yet in the literature (see [Zhao et al.(2023), Chaudhuri et Kalyanakrishnan(2019),Katz-Samuels et Jamieson(2020), De Heide et al.(2021)]. We are currently working on such lower bound, we believe that this would be a more valuable contribution in a separate paper fully dedicated to good-arm identification.

  • Extension to K>2K>2 clusters

    In this manuscript, we focused on the case of two groups because it serves as an informative baseline for analyzing the optimal trade-off in clustering with bandit feedback. This problem had not been studied before, even in the binary setting. Naturally, extending the approach to K>2K>2 groups is an important and practical direction. We will add to the appendix an algorithm that extends our method to the general case where K>2K>2, together with an analysis of its (non-optimal) budget. However, as you observed, obtaining the optimal dependency with respect to KK is significantly more challenging. We invite you to read the detailed discussion on this point in our response to Reviewer RdCv.

审稿意见
4

This paper addresses the problem of clustering items based on bandit feedback in a sequential and adaptive setting. Each of nn items is characterized by a dd-dimensional feature vector, and the items are partitioned into two unknown groups where items within the same group share the same feature vector. The learner sequentially selects an item and a feature, observes a noisy evaluation, and aims to recover the correct partition with a small number of observations.

The main contribution is the BanditClustering procedure, which operates in three steps:

  • Identifying two items from distinct groups
  • Finding a discriminative feature between them
  • Using this feature to cluster all items.

The authors provide non-asymptotic bounds on the budget required and prove their algorithm's optimality through matching lower bounds in certain cases. The approach leverages techniques from best arm identification (Sequential Halving) and adaptive sensing strategies.

The paper presents a theoretically sound algorithm for clustering with bandit feedback that achieves optimal sample complexity by efficiently identifying discriminative features.

The paper includes theoretical analysis characterizing the sample complexity based on the difficulty of the clustering task, which is quantified by the difference between feature vectors of the two groups. Experimental results demonstrate the advantage of their adaptive approach over uniform sampling, particularly in sparse regimes.

给作者的问题

  • Is there a way to use the samples from row 1 more efficiently? The current approach seems to involve significant oversampling from this row. Could this sampling strategy be optimized, and what would be the impact on the theoretical guarantees?

  • How would the algorithm perform in settings with more than two clusters? Would the approach of finding discriminative features scale well, or would the sample complexity increase significantly?

  • The paper focuses on the case where items within the same group have identical feature vectors. How robust is the approach to small variations within groups? Could the analysis be extended to allow for some heterogeneity is of the form that there is a linear seperator between the two groups? There is some discussion but would like to know more.

论据与证据

The main claims of the paper are supported by convincing theoretical and empirical evidence:

  • The upper bound on sample complexity (Theorem 3.1) is rigorously proven and demonstrates how the algorithm adapts to the sparsity pattern of the gap vector.

  • The lower bounds (Theorem 4.1) establish the optimality of the approach in certain regimes, particularly when the gap vector takes only two values.

  • The experimental results in Section 5 supports the theoretical findings, showing the algorithm's superior performance compared to uniform sampling in sparse regimes.

The claims about the algorithm's adaptivity to unknown structures are well-supported by the theoretical analysis, which shows how the budget scales with problem-dependent parameters without requiring prior knowledge of these parameters.

方法与评估标准

The methods and evaluation criteria are generally sound, though there might be room for improving sample efficiency:

  • The three stage approach with primary focus on identifying a single feature goes well with the narration of sparse feature distinction vectors.

  • The budget (number of observations) is an appropriate metric for evaluating efficiency in the bandit setting.

  • The theoretical analysis properly characterizes the algorithm's performance in terms of relevant problem parameters (gap magnitude, sparsity, balancedness).

However, there appears to be potential inefficiency in the sampling strategy, particularly regarding samples from row 1. Maybe a better book-keeping of what samples are already present?

理论论述

I skimmed through the details of the proof, they did appear correct. My primary focus was on :

  • Lemma B.1 (guarantees for CSH algorithm)

  • Theorem 3.1 (upper bound)

  • Theorem 4.1 (lower bound)

The theoretical analysis is generally sound. I noticed an issue in Section 3, where there appears to be a typo: μijμijμ_{ij}-μ_{ij} should likely be μic,jμ1,jμ_{ic,j}-μ_{1,j} or similar?

实验设计与分析

Summary of Experimental setup:

  • Experiment 1 examines how sparsity affects budget requirements, comparing BanditClustering against uniform sampling with K-means.

  • Experiment 2 studies how the budget scales with problem dimension when the sparsity is fixed.

These experiments showcase support to the theoretical claims about adaptivity to sparsity and scaling with problem dimensions. The comparison with uniform sampling is informative, demonstrating clear advantages of the adaptive approach.

However, the experimental section is relatively brief compared to the theoretical analysis. As a proof-of-concept it works, though more detailed experiments wouldn't be a bad addition.

补充材料

The main sections in supplemental material correspond to :

  • Notation
  • Proof of performance statements of Algo 1,2,3
  • Lower bound proof

I have tried to go through the supplemental material and they seem correct (at-least no obvious mistake), but it could be possible that I might have missed some crucial detail.

与现有文献的关系

The paper appropriately positions its contributions within several related research areas:

  • Best arm identification literature : The paper builds on Sequential Halving algorithms and extends these techniques to the clustering context.

  • Adaptive sensing strategies : The three phase approach is definitely a nice novel addition to the sparse sensing literature

  • Dueling bandits : The cleaver isolation of 1-d dueling bandit setup as a sub-sampler is a good addition.

  • Other bandit clustering problems : Limited impact here since the setup is of a bipartite graph kind of setup.

The authors claim their approach allows for more efficient budget allocation by focusing on the most relevant features, leading to lower observation budgets than previous methods, particularly in high-dimensional settings.

遗漏的重要参考文献

The paper does a good job at comparing with other bandit clustering frameworks but could potentially elaborate more on the quantitative differences in sample complexity. I don't identify any critical omissions in the literature review.

其他优缺点

Strengths: The paper's main strengths lie in its theoretical rigor and adaptive approach, while limitations include the restricted problem setup and questions about practical applicability.

  • Clearly written paper with well structured arguments and explanation.
  • The paper provides a clear theoretical characterization of the sample complexity, showing how it depends on the sparsity pattern of the gap vector.
  • The algorithm is adaptive to unknown structures without requiring prior knowledge of the problem parameters.
  • The theoretical analysis is rigorous, with matching upper and lower bounds in certain regimes.
  • The three-step approach (finding representatives, identifying discriminative features, clustering) is conceptually clear and intuitive.
  • The experimental results convincingly demonstrate the advantage over uniform sampling approaches.

Weaknesses:

  • The setup is limited to binary partitioning (just two clusters) with identical feature vectors within clusters. The scaling up of the computational complexity and regret due to more clusters is something of importance
  • The algorithm seems to oversample from row 1, raising questions about sample efficiency. The paper doesn't explore potential downstream effects or implications for regret bounds.
  • The practical impact of the theoretical findings on real-world applications is not extensively discussed.
  • The experimental section is relatively brief compared to the theoretical analysis. It would be valuable to see how the algorithm performs on a real-world dataset where the feature sparsity pattern is naturally occurring rather than artificially constructed.

其他意见或建议

  • The paper would benefit from a more intuitive explanation of why the specific trade-offs in the algorithm design are optimal, perhaps with a simplified example.

  • The paper mentions budget comparison with other approaches in the introduction but doesn't provide detailed quantitative comparisons in the main text. (maybe I missed this.)

作者回复

We first would like to thank you for your time and effort in reviewing it, and for your insightful questions. We now overview the remarks and questions that you formulate in your review:

  • Oversampling from row 1

    First, we would like to emphasize that improving the budget compared to non-active settings requires over-sampling certain rows, which we refer to as representatives, following [Thuot et al.(2025) "Clustering with bandit feedback: breaking down the computation/information gap"]. This step is crucial for accurately estimating the means μ1\mu_1 and μ2\mu_2 and, ultimately, for accelerating the clustering task. This is not, per se, a weakness of our procedure but rather an essential aspect for achieving order-wise budget optimality (up to logarithmic factors). Notably, we chose row 1 as the first representative, but by symmetry, any randomly selected row could have served the same purpose.

    Second, it may indeed be possible to slightly reduce the budget allocated to the first row by reusing previously sampled data. For instance, if an index pair (i,j)(i,j) is identified such that μijμ1j|\mu_{ij}-\mu_{1j}| is large, we can allocate cnΔ^2log(n/δ)c\frac{n}{\hat{\Delta}^2}\log(n/\delta) samples to estimate μ1j\mu_{1j} and μij\mu_{ij} once and for all. These estimates could then be reused for classifying the remaining rows, reducing the confidence bounds required for perfect classification by a constant factor. However, applying such refinements in other parts of the algorithm would significantly reduce its transparency, both in terms of implementation and theoretical analysis. Moreover, the resulting improvement in the budget would be limited to a factor 2. For the final version, we will keep our method as it is, but we will clarify our motivations, and highlight potential ways to improve numerical constants.

  • Extension to K>2K>2 clusters

    In this manuscript, we focused on the case of two groups because it serves as an informative baseline for analyzing the optimal trade-off for the problem. We will add to the appendix an algorithm that extends our method to the general case (K>2K>2), together with an analysis of its budget which scales with a factor K2K^2. However, obtaining the optimal dependency with respect to KK is significantly more challenging. We invite you to read the detailed discussion on this point in our response to Reviewer RdCv.

  • Robustness to heterogeneity within the clusters

    As mentioned page 8, our algorithm can be adapted to handle some degree of heterogeneity by appropriately adjusting certain tuning parameters. For instance, assume that we have prior knowledge that, for any feature j[d]j\in[d], the within-group variation satisfies, maxg(i)=g(i)μijμijcming(i)g(i)μijμij\max_{g(i)=g(i')} |\mu_{ij}-\mu_{i'j}|\leqslant c\min_{g(i) \ne g(i')} |\mu_{ij}-\mu_{i'j}|. If c<1/4c<1/4, our algorithm remains correct -- searching a single discriminative feature is still meaningful and enables classification. However, if the within-group heterogeneity is comparable to the inter-group differences in some features, our method could fail, and further investigation would be required. Observe also that if c=1/2c=1/2, the problem is not identifiable anymore. We will expand on this discussion in the final version.

  • Discussion on the optimal trade-off HH

    We appreciate your suggestion and will provide more intuition about the complexity term HH (see Eq. (3)), along with a discussion on why we believe it is optimal. Also, we will base our discussion, as for the upper bound, on the simple example of sparse vectors.You can find a discussion on the intuitions about HH in the sparse setting in the rebuttal of review RdVc, and a discussion on its optimality for general values of Δ\Delta in the rebuttal of review z8W5.

  • Budget comparison with the literature

    We acknowledge that further quantitative comparisons with other approaches and existing literature are missing. We will add this discussion in the final version, we will in particular discuss about the benefice of our model in high dimension settings. Thank you for pointing this out.

  • Real-world applications

    While we described in the introduction that image labeling might be a very natural application for our algorithm, it seems challenging to process existing datasets such that we obtain data for the bandit framework. We agree that the paper would benefit from real-world experiments, but in view of the fact that this work is intended to focus on an in-depth theoretical investigation, we decided to only consider synthetical data. A possibility could be to setup experiments using crowdsourcing platforms or a group of experts. One could use image data (as done in [Herde, Marek, et al. "dopanim: A Dataset of Doppelganger Animals with Noisy Annotations from Multiple Humans." Advances in Neural Information Processing Systems 37 (2024): 51085-51117.]), but ask experts about specific features of depicted animals instead of the concrete class.

审稿人评论

I would like to thank the authors for their detailed responses, especially regarding my oversampling query. I encourage the authors to incorporate these discussions (especially oversampling and budget comparision) into the final version to improve clarity and completeness of the current work.

My concerns have been resolved.

最终决定

This paper considers the problem of clustering items under bandit feedback in a fixed-confidence setting. It is assumed that the items are divided into two unknown groups, and items within the same group share the same dd-dimensional feature vector. At each time step, the agent selects a pair consisting of an item and one feature (out of dd features), and receives noisy feedback pertaining to that feature component. The objective is to correctly cluster all items with a confidence guarantee of at least 1δ1 - \delta, while minimizing the number of samples used.

Previous research on fixed-confidence bandit clustering has typically assumed that the entire feature vector of an item can be observed at each time step. In contrast, this study introduces the challenge of observing only a single feature of the feature vector, leading to the difficulty of “Finding the Right Feature out of Many.” This paper is the first to provide analysis for adaptive algorithms under such a setting.

Several limitations are noted, such as the lower and upper bounds not being tight, optimality being achieved only under restricted values of Δ\Delta ((it would be good to have comparisons with the lower/upper bounds in existing studies such as Yang et al., Thuot et al., Ariu et al., and Yavas et al.), the number of clusters being limited to two, and heterogeneity within the same cluster not being considered. In terms of algorithm design, there appear to be many areas for improvement—for example, they apply an anytime version using the doubling trick from Successive Halving (SH), a method suitable for fixed-budget problems, to the fixed-confidence problem. The experiments were limited, and there was mention of a lack of ablation studies on θ\theta and comparisons with existing methods (Ariu et al., 2024).

The authors have addressed these concerns to some extent in their rebuttal, and all reviewers seem to have agreed to assign or maintain a positive score in response. I expect these points to be reflected in the revised version.