PaperHub
5.8
/10
Rejected4 位审稿人
最低5最高6标准差0.4
6
6
6
5
3.5
置信度
正确性2.8
贡献度3.0
表达3.0
ICLR 2025

Learning from interval targets

OpenReviewPDF
提交: 2024-09-24更新: 2025-02-05

摘要

关键词
Weak supervisionPartial-label learningLearning from side informationLearning theory

评审与讨论

审稿意见
6

This paper studies a regression model where only an interval around the true response is known to the learner. Given such weak supervision, the goal of the learner is to learn a predictor whose population risk is small. To tackle this problem, the authors propose two methods. The first modifies the standard regression loss functions to account for interval ground truths. In particular, a projection loss is defined which penalizes the the output a learned predictor when its outputs lie outside the interval for a specific instance. The second formulates a minimax problem with an objective that minimizes the original regression loss with respect to the worst-case label within the interval (available to the learner). In both settings, the authors show how the smoothness of the function class (measured via Lipschitzness) can be beneficial. The authors provide theoretical guarantees for both methods and corroborate their results with experiments.

优点

  • Paper is well written and organized
  • Paper improves upon previous works by removing assumptions and the ideas seem novel as far as I can tell
  • I think the Figures do a good job of giving intuition about the methods

缺点

  • Unclear objective. I think the authors can do a better job about making explicit the fact that although the learner observes interval targets, the goal is to still do well with respect to the original regression loss. This goal is sort of hidden in lines 101-102, and it would be nice to emphasize this more.

  • Lack of specification on how error rates decay with sample size. It is not clear how the excess risk in lines (13) and (16) decays with the sample size nn. For all I know, it may not decay at all, making the error bounds vacuous. At the very minimum, it would be nice if the authors can provide error rates for a specific class of functions. This is also an issue in line (22) - I have no intuition on how the error might decay with the sample size here. I think this is a significant limitation and is the main reason for my rating. I will raise my score if the authors can give me a sense of how the non-asymptotic error rates decay with nn, even it is for a specific function class.

  • Unclear method. I am confused about certain aspects of the methods proposed in Section 4.1 See Questions below for more details.

问题

  • Does Proposition 3.1 hold uniformly for all xXx \in X? If so, I am confused since fF0~f \in \tilde{F_{0}} only guarantees that f(x)[lx,ux]f(x) \in [l_x, u_x] almost surely. So there could still be xx's (albeit in a measure 00 set), for which this is not true.
  • What norm is being used on line 236?
  • Is the guarantee on line 267 pointwise or in expectation? If it is pointwise for every xx, how are you accounting for the fact that the functions fFη~f \in \tilde{F_{\eta}} may have projection loss bigger than η\eta for low probability xx? Is this accounted for by definition of rη(x)r_{\eta}(x) and sη(x)s_{\eta}(x)? If so, it would be good to explain how the buffer rη(x)r_{\eta}(x) and sη(x)s_{\eta}(x) account for that.
  • Is F~0\tilde{F}_0 known to the learner in Equation (23)? It seems like the learner needs to know the distribution DD over XX to compute F~0\tilde{F}_0 based on line 170. So because the learner doesn't know DD, it shouldn't be able to perform the optimization in Equation (23). What am I missing here?
  • Based on the bullet points above, how does the learner know that the f1,...,fjf_1, ..., f_j in line 409-410 belong to F~0\tilde{\mathcal{F}}_0 if F~0\tilde{\mathcal{F}}_0 is not known?
  • Do the upper bounds in Equations (13) and (16) decay with the sample size nn? If so, it would be good to comment and be explicit about how they decay with nn (i.e what is the rate of decay?).
  • In Equation (24), are the minimizations and maximizations over ff and ff^{\prime} being done both over F\mathcal{F}? Or is the maximization of ff^{\prime} being done over F~0\tilde{F}_0? If it is the latter, the same question I had above holds - how can the learner do this if F~0\tilde{F}_0 is unknown?
评论

Thank you for your feedback and suggestions. We hope the following responses address all your questions.

  1. I think the authors can do a better job about making explicit the fact that although the learner observes interval targets, the goal is to still do well with respect to the original regression loss.

We have made this setting more explicit in the revised version.

  1. Lack of specification on how error rates decay with sample size. It is not clear how the excess risk in lines (13) and (16) decays with the sample size n.

We have added an explicit generalization bound for both realizable (13) and agnostic setting (16) in terms of the number of data points nn in Appendix A. This result is applicable to any hypothesis class F\mathcal{F} where the Rademacher complexity grows as O(1/n)O(1/\sqrt{n}). For instance, the class of two-layer neural network with bounded weight satisfies this property.

  1. Does Proposition 3.1 hold uniformly for all xXx \in X? If so, I am confused since fF0f \in \mathcal{F}_0 only guarantees that f(x)[lx,ux]f(x) \in [l_x, u_x] almost surely. So there could still be xx (albeit in a measure 0 set), for which this is not true.

Our result in Proposition 3.1 holds for any xXx \in \mathcal{X} with a positive probability density function p(x)>0p(x) > 0. We have made this more clear in the revised version. We note that this is stronger than f(x)[lx,ux]f(x) \in [l_x, u_x] almost surely, and we deal with a set with measure 0 with an assumption that the distribution DID_I is a non-atomic distribution (Assumption 2 in Appendix E).

  1. What norm is being used on line 236?

We used L1L_1 norm and this is made clearer in the revision.

  1. Is the guarantee on line 267 pointwise or in expectation? If it is pointwise for every xx, how are you accounting for the fact that the functions fFηf \in F_\eta may have projection loss bigger than η\eta for low probability x? Is this accounted for by definition of rη(x)r_{\eta}(x) and sη(x)s_{\eta}(x)? If so, it would be good to explain how the buffer rη(x)r_{\eta}(x) and sη(x)s_{\eta}(x) account for that.

The guarantee in line 267, is pointwise and holds for any xXx \in \mathcal{X} with a positive probability density p(x)>0p(x) > 0. We are able to deal with any xx with a large projection loss (although with low probability) by relating xx with its neighbor xx'. The main idea is that, with smoothness, if f(x)f(x) is too small (too large) then for any neighbor xx', f(x)f(x') has to be close to f(x)f(x) and thus, would be too small (too large) as well. But this can't happen too much since the expected projection loss is smaller than η\eta and we bound how much f(x)f(x) can be far away from the given interval with rη(x)r_{\eta}(x) and sη(x)s_{\eta}(x). The full proof in Appendix C. We will make this connection clearer in the revised version.

  1. Is F0F_0 known to the learner in Equation (23)? It seems like the learner needs to know the distribution D over X to compute F0F_0 based on line 170. So because the learner doesn't know D\mathcal{D}, it shouldn't be able to perform the optimization in Equation (23). What am I missing here?

Yes, we would not have access to F0F_0 in practice. Therefore, we propose to approximate the objective in equation (23) with the regularization approach and with pseudo labels. In general, we could learn fjFηf_j \in F_\eta for some small η\eta and we could treat a set of such functions as an approximation of F0F_0 (pseudo labels idea). We have made this connection more explicit in the revised version.

  1. Based on the bullet points above, how does the learner know that the f1,...,fjf_1,...,f_j in line 409-410 belong to F0F_0 if F0F_0 is not known?

In practice, f1,...,fjf_1,...,f_j would belong to FηF_{\eta} for some small η\eta. We can achieve this by training with a different random seed. FηF_\eta is a superset of F0F_0 where we use it to approximate F0F_0. We have made this approximation aspect is more clear in the revised version.

  1. In Equation (24), are the minimizations and maximizations over ff and ff' being done both over F\mathcal{F}? Or is the maximization of ff' being done over F0\mathcal{F}_0? If it is the latter, the same question I had above holds - how can the learner do this if F0\mathcal{F}_0 is unknown?

The minimizations and maximizations over ff and ff' are being done both over F\mathcal{F}. However, we have a regularization on ff' to encourage ff' to be in F0\mathcal{F}_0.

评论

I thank the authors for their response. As promised, I have increased my score to 6 given their explicit generalization bounds.

审稿意见
6

This paper considers a regression analysis problem where the outcome of each training sample is given by not single value but an interval. An existing method assumes the "realizability" and "ambiguity degree < 1" in the existing method, and it employs the loss function called the "projection loss" and proved the theoretical performances. However, since these assumptions are too restrictive, the paper relaxes the assumptions, and accordingly constructed another learning setup called "minmax objective".

优点

  • The formulations of the problem and the method looks clear.
  • The method introduced less restrictive constraint than existing work: the class of hypotheses F{\cal F} must contain only mm-Lipschitz functions. Also, the experiment in Section 5.2 examines the effect of the choice of mm in the algorithm. (As discussed with equation (16), mm should not be too small or too large.)

缺点

  • It looks that a part of experimental setups are lacked. Please see Section "Question" below.
  • Just notation problems, but it is difficult to follow the effect of the Lipschitz constant mm. It may be better to add "mm" to all operations affected by mm (e.g., replacing lxxl_{x^\prime\to x} with lxx(m)l^{(m)}_{x^\prime\to x}).

问题

  • Section 2, line 133: What is the "\ell1 loss as the surrogate" for the 0-1 loss? The l1-loss surrogate for (1) is not so obvious, although we can understand it after reading its mathematical expression in (Cheng et al., 2023a).
  • Section 2, equation (2): What are the capital letters LL and UU? Perhaps the lower and upper bounds as random variables? (If so, what are their distributions?)
  • Section 3, equation (6) and the succeeding sentences: Are π\pi_{\ell} and πl\pi_l the same? (Perhaps just non-unified notations?)
  • Section 3.2, Proposition 3.2: What situation the induced lower bound and upper bounds of f(x)f(x) is effective? As far as my understanding, if the size of the bounds is small at xx^\prime, the size at another point xx must be accordingly small under the constraint of Lipschitz continuity.
  • Section 3.3.1, equation (13): What are the definitions of I0I_0 and IηI_\eta? (Although I could guess them,) their specific definitions seems not to be given.
  • Section 3.3.1, equation (14): Is the equation correct? The rightmost expression of the equation looks like independent from xx.
  • Section 4: Are the notations "ff^\prime"'s in equations (20) and (21) are related? If not, it is better to introduce different variables.
  • Section 5 (Appendix D): What is the obstacle (e.g., computational cost) to run the proposed method for a larger dataset? The datasets used in the experiments have up to 10000 samples.
  • Section 5: The proposed method assumes that the class of hypotheses is mm-Lipschitz, but in the experiment how can we limit the prediction model (MLP in this case) as mm-Lipschitz? (The paper refers the existing work (Cheng et al., 2023a), but in the existing work it looks that mm-Lipschitz is not employed.)
  • Section 5: What loss function does the experiment employed? (More specifically, ψ\psi in Assumption 1)
评论

Thank you for your feedback and questions. We hope the following responses address all your questions.

  1. Just notation problems, but it is difficult to follow the effect of the Lipschitz constant m. It may be better to add m to all operations affected by m (e.g., replacing lxxl_{x' \to x} with lxx(m)l_{x' \to x}^{(m)}).

We have incorporated this in the revised version.

  1. Section 2, line 133: What is the "1\ell_1 loss as the surrogate" for the 0-1 loss? The l1-loss surrogate for (1) is not so obvious, although we can understand it after reading its mathematical expression in (Cheng et al., 2023a).

Yes, by 1\ell_1 surrogate loss for 0-1 loss, we refer to equation 12) in Cheng et al. 2023a. We made this more explicit in the revised version.

  1. Section 2, equation (2): What are the capital letters L and U ? Perhaps the lower and upper bounds as random variables? (If so, what are their distributions?)

Yes, L and U are the lower and upper bounds as random variables. We assume that the tuple (xi,li,ui)(x_i, l_i, u_i) is sampled i.i.d. from a distribution DI\mathcal{D}_I (line 106).

  1. Section 3, equation (6) and the succeeding sentences: Are π\pi_\ell and πl\pi_l the same? (Perhaps just non-unified notations?)

Yes, we have fixed this in the revised version.

  1. Section 3.2, Proposition 3.2: What situation the induced lower bound and upper bounds of f(x)f(x) is effective? As far as my understanding, if the size of the bounds is small at XX' ,the size at another point xx must be accordingly small under the constraint of Lipschitz continuity.

First, we note that under a small ambiguity degree assumption, our bound would be very effective where we would recover the true label, [lDx,uDx]={y}[l_{\mathcal{D} \to x}, u_{\mathcal{D} \to x}] = \{y\}. Our bound is also applicable when the ambiguity degree is large and yes, you have the right intuition. Our bound is more effective when the hypothesis class is smoother, and when the neighbor intervals are "far away'' from the given interval, so that it would force the hypothesis to be near the edge of the given interval (see Figure 1b for an illustration).

  1. Section 3.3.1, equation (13): What are the definitions of I0I_0 and IηI_\eta ? (Although I could guess them,) their specific definitions seems not to be given.

We had defined IηI_\eta in Theorem 3.3 in the previous version. In the revision, we define this explicitly in Section 3.3 so that it is easier for the reader.

  1. Section 3.3.1, equation (14): Is the equation correct? The rightmost expression of the equation looks like independent from xx.

Yes, this is correct. Here we consider and example when our error bound is tight when the hypothesis class only contains a constant function. Therefore, the right-hand side is independent from xx.

  1. Section 4: Are the notations ff in equations (20) and (21) are related? If not, it is better to introduce different variables.

Yes, they are the same hypothesis ff that we learn from the hypothesis class F\mathcal{F}. However, they are not related and we can incoporate this in the revised version.

  1. Section 5 (Appendix D): What is the obstacle (e.g., computational cost) to run the proposed method for a larger dataset? The datasets used in the experiments have up to 10000 samples.

There is no bottleneck for running this method on a larger dataset. In particular, the complexity of the projection method is the same as the complexity of traditional regression (with the same function class). For the pseudo-label methods, the complexity is kk times of the complexity of the traditional regression method when kk is the number of pseudo labels for each instance. We expect our proposed methods to scale well with the size of the datasets.

  1. Section 5: The proposed method assumes that the class of hypotheses is m-Lipschitz, but in the experiment how can we limit the prediction model (MLP in this case) as m-Lipschitz? (The paper refers the existing work (Cheng et al., 2023a), but in the existing work it looks that m-Lipschitz is not employed.)

We use an MLP with a spectral normalization layer (Miyato et al. 2018)(see Section 5.2). The normalization ensures that the Lipschitz constant of the MLP is less than 1. We then scale the output of the MLP by a constant factor m to ensure that the Lipschitz constant of the hypothesis is less than mm.

  1. Section 5: What loss function does the experiment employed? (More specifically, ψ\psi in Assumption 1)

We use mean absolute error or l1l_1 loss in our experiment. This is equivalent to ψ(x)=x\psi(x) = x in Assumption 1.

评论

Dear reviewer, please let us know if you have any other questions. If we have resolved all your concerns, would you be willing to reconsider your score?

评论

I was convinced with author's responses, however, I will retain my score after reading others' reviews.

审稿意见
6

This paper investigates a weakly supervised regression setting called “learning from interval targets," where the exact real-valued targets are not directly available; instead, supervision is provided in the form of intervals around the targets. The authors suggest that previous research assumptions for theoretical analysis were overly simplistic (SMALL AMBIGUITY DEGREE ASSUMPTIONS). Therefore, they propose new theoretical analyses to address more robust real-world scenarios. To tackle this problem, the authors propose two methods: (1) using surrogate loss functions to align targets with weak supervision labels (PROJECTION APPROACH); and (2) considering the worst-case scenario within the intervals and minimizing that risk (MINMAX OBJECTIVE). Extensive experiments validate the effectiveness of the authors' methods.

优点

  • The authors discuss an intriguing weakly supervised regression problem: learning from interval targets. Most research has only considered classification problems, but they overlook that regression problems are equally important.
  • The authors provide (reasonable) theoretical analyses for their methods, demonstrating the effectiveness of the proposed approaches.
  • The methods proposed by the authors are both natural and reasonable, and the paper is well-organized and easy to understand.

缺点

  • The experimental setup considered by the authors is too simplistic, as it only involves a small number of UCI datasets, and the selected datasets are simple (small and low-dimensional). Previous research seems to have also considered experimental setups in CV and NLP.

  • The authors' analysis is based on the rejection of the SMALL AMBIGUITY DEGREE ASSUMPTIONS, but I believe their rejection of this assumption is not convincing enough, as it would be impossible to learn if noise is always present.

问题

  • The authors believe that the SMALL AMBIGUITY DEGREE ASSUMPTIONS are not applicable in regression scenarios primarily because regression tasks differ from classification tasks in having exact labels. Therefore, directly applying the previous SMALL AMBIGUITY DEGREE ASSUMPTIONS is unreasonable. Should the AMBIGUITY DEGREE ASSUMPTIONS be redefined instead of being directly rejected? A larger AMBIGUITY DEGREE might result in intervals that are always mixed with noise, making it impossible to discern the true label distribution. Alternatively, could the size of intervals be used to constrain the AMBIGUITY DEGREE to satisfy a new relationship?
  • In the PL (max) and PL (mean) methods, the authors utilize pseudo labels, but it seems there is no mention of how these pseudo labels are generated.
  • The projection loss proposed by the authors (Equation 5) appears to be consistent with the surrogate loss function in [1]. Does this imply that the method in [1] can still be applicable under the new setting (SMALL AMBIGUITY DEGREE ASSUMPTIONS does not apply) ?
评论

Thank you for your feedback and suggestions. We hope that the following address your main questions regarding the ambiguity degree assumption.

  1. The experimental setup considered by the authors is too simplistic, as it only involves a small number of UCI datasets, and the selected datasets are simple (small and low-dimensional). Previous research seems to have also considered experimental setups in CV and NLP.

Our focus on a small set of UCI datasets aims to complement the theoretical contributions of this work with clear, reproducible results. While it would be very valuable, our focus was not on a large-scale and thorough evaluation on the experimental front.

  1. The authors' analysis is based on the rejection of the SMALL AMBIGUITY DEGREE ASSUMPTIONS, but I believe their rejection of this assumption is not convincing enough, as it would be impossible to learn if noise is always present.

In regression, predictions close to the target (within an error tolerance ϵ\epsilon) are often sufficient, making learning with noise more tolerable. For example, when for each label yy the given interval is a ball with radius ϵ>0\epsilon > 0 around yy, [l,u]=[yϵ,y+ϵ][l,u] = [y - \epsilon, y+\epsilon]. When ϵ\epsilon is small enough, then we would expect to be able to learn a good function, albeit with some irreducible (small) error that depends on ϵ\epsilon. However, since the ambiguity degree is large in this setting, the previous work would not apply.

  1. In the regression setting,) should the AMBIGUITY DEGREE ASSUMPTIONS be redefined instead of being directly rejected? A larger AMBIGUITY DEGREE might result in intervals that are always mixed with noise, making it impossible to discern the true label distribution. Alternatively, could the size of intervals be used to constrain the AMBIGUITY DEGREE to satisfy a new relationship?

We have now added a note in the paper outlining an alternate definition in line with your proposal, and how it relates to our definition, in Appendix B. Here's a preview:

Recall that the small ambiguity degree assumption implies that for each instance xx, the intersection of all intervals of xx is the true label yy; [lx,ux]=y\bigcap [l_x, u_x] = y. One possibility along the lines of your suggestion is to redefine it so that the intersection falls within a radius ϵ\epsilon around the true label: [lx,ux]B(y,ϵ)\bigcap [l_x, u_x] \subseteq B(y, \epsilon). In this case, our analysis already captures the essence of this interval intersection for each xx. If we have this assumption then our bound would already imply that f(x)I0(x)B(y,ϵ)f(x) \in I_0(x) \subseteq B(y, \epsilon). Further, I0(x)I_0(x) might even be a proper subset of the ball B(y,ϵ)B(y, \epsilon) based on the smoothness property.

  1. In the PL (max) and PL (mean) methods, the authors utilize pseudo labels, but it seems there is no mention of how these pseudo labels are generated.

The pseudo labels are generated from a hypothesis fjf_j which is a function from F~η\widetilde{\mathcal{F}}_\eta (see equation (25), (26)) for some small η\eta. We have updated the paper to make this more clear now.

  1. The projection loss proposed by the authors (Equation 5) appears to be consistent with the surrogate loss function in [1]. Does this imply that the method in [1] can still be applicable under the new setting (SMALL AMBIGUITY DEGREE ASSUMPTIONS does not apply) ?

We suppose that [1] here refers to a prior work "Weakly supervised regression with interval targets" by Cheng et al., 2023a. The surrogate loss function in [1] is a special case of our LpL_p loss when p=1p=1. In our paper, we have a theoretical result for LpL_p loss for all p1p \geq 1 and also for a general loss function that satisfies Assumption 1. Our general result implies that [1] is still applicable in the setting when the small ambiguity degree assumption does not apply. On the other hand, previous work does not use the smoothness assumption and do not tune the Lipschitz constant of the hypothesis class. Our analysis found that this is an important variable and demonstrate that tuning the Lipschitz constant leads to a better downstream performance (Figure 3).

评论

I thank the authors for their response, which has addressed some of my concerns. However, I still have a question regarding Question 4.

While I understand that the pseudo labels are generated by the hypothesis fjf_{j}, it remains unclear how the hypothesis fjf_{j} is obtained. The additional section mentions that fjf_{j} is derived by minimizing the empirical projection loss, but this explanation is still insufficiently detailed. For example, what exactly constitutes the empirical projection loss? What data is used in this process? Specifically, the setup used in the experiments to determine fjf_{j} is not clearly explained, which should be an important component of PL (max) and PL (mean) methods.

评论

We are happy to provide an additional clarification. The empirical projection is given by Equation (5). That is, if we have training samples (xi,li,ui)(x_i, l_i, u_i) for i=1,,ni = 1,\dots, n , then

fjargminfF1ni=1nπ(f(xi),li,ui).f_j \in \arg\min_{f \in \mathcal{F}} \frac{1}{n} \sum_{i=1}^n \pi_\ell(f(x_i), l_i, u_i).

In practice, training a model using stochastic gradient descent with different random initializations and random seeds ensures that fjf_j will be different. This process can be repeated kk times to obtain {f1,,fk}\{f_1, \dots, f_k\}. We will revise this to be more explicit in the updated version.

评论

Dear reviewer, please let us know if you have any other questions. If we have resolved all your concerns, would you be willing to reconsider your score?

评论

Thank you to the authors for the additional clarification, which has resolved my confusion regarding hypothesis fjf_{j}. I will maintain my current score 6, which is a positive rating.

审稿意见
5

This paper investigates the regression problem in a situation where the learner has access only to an interval around the true response. It proposes two approaches:

  1. Modifying the regression loss function with a projection and learning a hypothesis lying within the interval
  2. Minimizing a typical regression loss with respect to the worst-case label

优点

This work investigates an interesting situation where the learner has access only to an interval around the true response, rather than the exact response as in the standard supervised learning setting. Due to the high cost of getting exact responses, learning from intervals has significant practical meanings.

It proposes two approaches and provides both theoretical analysis and experiments on real-world datasets. For the theoretical results, the proofs are clear and easy to follow, accompanied by a sufficient explanation of the derivations.

缺点

  1. This paper highlights a generalization bound in Section 3.3 as its main result. However, this bound seems too straightforward and can be quite vacuous - e.g., Theorem 3.5 is only taking an expectation of Equation (12), which is the worst-case analysis on a single instance.
  2. The multiple parameters appearing in the generalization bound are difficult, if impossible at all, to obtain. For example, in order to get the bound in Theorem 3.5, one needs to know IηI_\eta. IηI_\eta consists rηr_\eta, which involves the calculation of an expectation with respect to XX (Equation 9). Usually, one is not able to compute the expectation efficiently.
  3. To obtain an estimate of η\eta, one needs to know the Rademacher complexity of Π(F)\Pi(F), which can be quite difficult to obtain.

问题

  1. It's mentioned that Theorem 3.5 can be tight for certain hypothesis classes, for instance, constant hypotheses. However, the constant hypothesis seems to be too restrictive. Could you provide another instance of hypothesis class, where this theorem can be relatively tight?
  2. Is it possible to obtain the parameters in the generalization bounds in an efficient way?
  3. In lines 179-182, it's mentioned that a value of η\eta is obtained from the RHS of Equation (6); how would one calculate the Rademacher complexity of Π(F)\Pi(F)? How does the tightness of an estimation of η\eta affect the later generalization bounds?
  4. In lines 201-202, why ηn=O(1/n)\eta_n = O(1/\sqrt n)? How does the Rademacher complexity of Π(F)\Pi(F) affect the order of ηn\eta_n?
评论

Thank you for your feedback and suggestions. We hope that the following address your questions regarding our theoretical results.

  1. This paper highlights a generalization bound in Section 3.3 as its main result. However, this bound seems too straightforward and can be quite vacuous.

We want to highlight that one of the technical difficulties in proving Theorem 3.3 is that we allow for ff that only approximately lies inside the interval. That is, for certain xx, f(x)f(x) could be arbitrarily far away from the given interval [lx,ux][l_x, u_x]. Nevertheless, our bound provides a pointwise reduced interval for each individual xx. This is non-trivial as we have to work with f(x)f(x) that could be arbitrarily far away from the given interval [lx,ux][l_x, u_x].

  1. The multiple parameters appearing in the generalization bound are difficult, if impossible at all, to obtain. \ldots To obtain an estimate of η\eta one needs to know the Rademacher complexity of Π(F)\Pi(F) which can be quite difficult to obtain.

Our bounds emphasize learnability in the agnostic setting, which was not known in previous work. That is, as the sample size increases, we can bring error down to the error rate for the best hypothesis in class. Our bounds also reveal that the Lipschitz constant affects generalization. In practice, we would not estimate IηI_\eta or rηr_\eta but instead tune relevant parameters, and our analysis shows that the smoothness parameter is an important one to tune. Our experiments confirm the insights of the analysis hold in practice.

  1. It's mentioned that Theorem 3.5 can be tight for certain hypothesis classes, for instance, constant hypotheses \dots Could you provide another instance of hypothesis class, where this theorem can be relatively tight?

The bound would also be relatively tight on both extreme when the hypothesis class is smoother (Lipschitz constant mm is close to 00) and when the hypothesis class is very flexible (mm is very large). The first setting follows from the fact that the bound holds for a constant hypothesis class when m=0m = 0. For the second example, when the hypothesis class is very flexible so that the label of any xx, is independent of other xx', the only information we have is that the true label lies inside a given interval then the best thing we could hope for is that the error for each instance xx is at most the size of the interval of each xx. Since the hypothesis class is too flexible, we can't derive additional information from it.

  1. In lines 179-182, it's mentioned that a value of η\eta is obtained from the RHS of Equation (6); how would one calculate the Rademacher complexity of Π(F)\Pi(F) ? How does the tightness of an estimation of η\eta affect the later generalization bounds? \dots In lines 201-202, why ηn=O(1/n)\eta_n=O(1/\sqrt{n})? How does the Rademacher complexity of Π(F)\Pi(F) affect the order of ηn\eta_n?

The result in the line 179-182 is an upper bound of the true value of η\eta of the learned hypothesis ff (denoted as ηn\eta_n). This bound is for a theoretical purpose where we hope to communicate that as we have more training data points nn, η\eta would converge to zero as ηn\eta_n is of O(1/n)O(1/\sqrt{n}). As a result, we would be able to learn a hypothesis that lies inside the given intervals more often with more training data.

To see that ηn=O(1/n)\eta_n = O(1/\sqrt{n}), we recall the definition that ηn=2Rn(Π(F))+Mln(1/δ)n\eta_n = 2R_n(\Pi(\mathcal{F})) + M\sqrt{\frac{\ln(1/\delta)}{n}} when Rn(Π(F))R_n(\Pi(\mathcal{F})) is the Rademacher complexity. We can show that Rn(Π(F))R_n(\Pi(\mathcal{F})) can be reduced to Rn(F)R_n(\mathcal{F}) and it is known that for many hypothesis class such as a class of two-layer neural networks with bounded weights or linear models, Rn(F)R_n(\mathcal{F}) is O(1/n)O(1/\sqrt{n}). Therefore, we can conclude that ηn=O(1/n)\eta_n = O(1/\sqrt{n}). We have provided a rigorous derivation of this in Appendix A.

On the effect of the η\eta on the final generalization bound, we have provided an additional Theorem in Appendix A that explicitly how the final error bound decay with the number of training sample nn (which also affect how large η\eta is).

评论

Dear reviewer,

In the revision, we have comprehensively addressed your main concern regarding explicit bounds. A new Appendix section was added that shows ηn=O(1/n)\eta_n = O(1/\sqrt{n}) for classes with the Rademacher complexity Rn=O(1/n)R_n = O(1/\sqrt{n}). In the rebuttal comment before this one, we elaborate on this bound, as well as your other concerns.

We request you to reconsider your score.

Authors

评论

I thank the authors for the responses and clarifications on explicit bounds. I am increasing my score to 5.

评论

We thank all reviewers for detailed and constructive feedback, and suggestions to improve the paper. Reviewers found our problem setting well-motivated. They appreciated the exposition, finding the paper "well-organized", "clear", and "easy to follow". The key improvement to previous work––a finite-sample generalization analysis under less restrictive assumptions––was clear to all reviewers.

A recurring concern among reviewers was the opacity of our main results, particularly the lack of clarity on how generalization depends on the number of samples nn. A major update in the revised version is explicit sample-complexity bounds for classes whose Rademacher complexity is O(1/n)O(1/\sqrt{n}). We show that apart from irreducible additive terms, the part of the excess risk that can be brought down by learning also goes down at the same rate O(1/n)O(1/\sqrt{n}). Thus, in the setting we analyze, the statistical behavior of learning with interval feedback is similar to behavior when learning with point feedback. This result and its proof are currently in Appendix A of the main paper, and will also be incorporated into the main paper for the camera-ready version.

We have incorporated additional reviewer suggestions in this revision. These include an alternative definition of the ambiguity degree for a regression setting, clarification of proposed methods, problem setting and the theoretical analysis as well as correction of typos. Updates are highlighted in blue in the revised version. Reviewers also had a number of specific questions that we will answer in individual responses. We believe this substantial revision significantly strengthens the paper and hope the reviewers agree.

AC 元评审

This paper addresses the challenge of regression with interval targets, where the exact target value is replaced by an interval around the true response. The authors propose two methods: the projection approach, which modifies the regression loss to penalize predictions falling outside the interval, and the minimax objective, which minimizes the worst-case loss within the interval. The paper provides theoretical guarantees for both methods and supports them with empirical experiments on real-world datasets.

Reviewers acknowledged the novelty of the paper and appreciated its theoretical contributions, particularly in tackling the regression problem with interval targets. However, the significance of the theoretical results was questioned, with some reviewers finding them less impactful than expected. For instance, the parameter η\eta, which depends on Rademacher complexity, is challenging to compute accurately. Additionally, the experimental setup was considered too simplistic, relying on small datasets. As a result, the paper was viewed as borderline, with no reviewer offering strong endorsement.

Given the high competition among ICLR submissions, I lean towards rejection. The authors are encouraged to take into accounts the reviewers' comments and resubmit to another venue.

审稿人讨论附加意见

After discussions, the authors have addressed some of the reviewers' concerns. However, the paper has not received strong support overall. There remain questions regarding the significance of the theoretical results and ongoing concerns about the experimental setup.

最终决定

Reject