PaperHub
7.3
/10
Poster4 位审稿人
最低4最高5标准差0.5
5
4
4
5
3.0
置信度
创新性2.8
质量3.0
清晰度3.0
重要性3.0
NeurIPS 2025

Agnostic Continuous-Time Online Learning

OpenReviewPDF
提交: 2025-05-06更新: 2025-10-29
TL;DR

We develop a general theory of agnostic online learning from continuous-time data streams under limited queries, providing tight regret bounds for both oblivious and adaptive settings.

摘要

We study agnostic online learning from continuous-time data streams, a setting that naturally arises in applications such as environmental monitoring, personalized recommendation, and high-frequency trading. Unlike classical discrete-time models, learners in this setting must interact with a continually evolving data stream while making queries and updating models only at sparse, strategically selected times. We develop a general theoretical framework for learning from both *oblivious* and *adaptive* data streams, which may be noisy and non-stationary. For oblivious streams, we present a black-box reduction to classical online learning that yields a regret bound of $T \cdot R(S)/S$ for any class with discrete-time regret $R(S)$, where $T$ is the time horizon and $S$ is the *query budget*. For adaptive streams, which can evolve in response to learner actions, we design a dynamic query strategy in conjunction with a novel importance weighting scheme that enables unbiased loss estimation. In particular, for hypothesis class $\mathcal{H}$ with a finite Littlestone dimension, we establish a tight regret bound of $\tilde{\Theta}(T \cdot \sqrt{\mathsf{Ldim}(\mathcal{H})/S})$ that holds in both settings. Our results provide the first *quantitative* characterization of agnostic learning in continuous-time online environments with limited interaction.
关键词
continuous-time online learningquery efficiencyadaptive adversariesimportance weightingLittlestone dimension

评审与讨论

审稿意见
5

This paper investigates agnostic continuous-time online learning. In contrast to the classical formulation of online learning, in which the input is a discrete sequence, the input here is a continuous data stream. While previous research has addressed the realizable setting, the more realistic and challenging agnostic setting has remained largely unexplored. The authors demonstrate that, given a hypothesis class with Littlestone dimension dd, the regret of online learning this class is exactly Θ~(Td/S)\tilde{\Theta}(T\sqrt{d / S}) for both oblivious and adaptive adversarial settings, where TT represents the time-span and SS denotes the number of queries the algorithm can make.

优缺点分析

Strengths:

The authors establish a nearly tight Θ~(Td/S)\tilde{\Theta}(T\sqrt{d / S}) regret bound for the fundamental problem of agnostic continuous-time online learning. For an oblivious adversary, they propose a simple black-box reduction to the discrete case. In the case of an adaptive adversary, they present a novel re-weighting scheme attaining the same bound. Additionally, they provide a matching lower bound that holds even for the most challenging scenario (an adaptive query strategy against an oblivious nature).

Furthermore, the organization and presentation of the paper are quite commendable, making it easy to read and understand.

Weakness:

I did not identify any significant weakness. However, I have concerns regarding the correctness of Lemma 6. Please refer to "Questions" for further details.

In the preliminaries, the authors define the protocol (Protocol 1) in a way that it chooses a predictor h^sYX\hat{h}_s\in\mathcal{Y}^{\mathcal{X}}. However, it appears that sometimes the output is indeed a distribution over predictors (particularly in the adaptive setting, where this is necessary), and at any time step the actual predictor is drawn from this distribution (rather than keep outputting h^s\hat{h}_s for a period). I recommend the authors clarify this part to make the definition consistent.

问题

  1. I am concerned that Lemma 6 may not hold as I1,,IkI_1,\dots,I_k could overlap. For example, let m=k=2m = k = 2. Given that Y1=Y2=0Y_1 = Y_2 = 0, it holds that E[BJY1=Y2=0]>0E[B_J \mid Y_1=Y_2=0]>0 since it is possible (with positive probability) that JI1=I2J\neq I_1 = I_2 and BJ=1B_J = 1. However, the Lemma gives 1/2+k/m(Yˉ1/2)=01/2 + k/m(\bar{Y} - 1/2) = 0. As this lemma serves as a key inequality in the proof of the lower bound, can you provide a more detailed explanation regarding its validity?
  2. It seems the lower bound only holds when SS is not excessively large: in the formula before line 413, the inequality requires T/S=Ω(1)T / \sqrt{S} = \Omega(1) (i.e., S=O(T2)S = O(T^2)). In contrast, the upper bound seems to hold even as SS \to \infty. Therefore, the bounds do not match in the extreme regime. Is my understanding correct?

局限性

yes

最终评判理由

This paper investigates agnostic continuous-time online learning and provides tight upper and lower bounds. Though there were some issues in the original proof, they were fully addressed in the rebuttal. I recommend acceptance.

格式问题

N/A

作者回复

We thank the reviewer for the thoughtful feedback and helpful comments. We address the questions raised below:

Q1: This is indeed a bug, and we are grateful to the reviewer for identifying it. The correct expression should be

12+k~m(Yˉdistinct12),\frac{1}{2} + \frac{\tilde{k}}{m} \left( \bar{Y}_{\mathrm{distinct}} - \frac{1}{2} \right),

where k~k\tilde{k} \le k is the number of distinct ItI_t's, and Yˉdistinct\bar{Y}_{\mathrm{distinct}} is the average of all YtY_t's corresponding to distinct indices ItI_t. The same replacement should also be made in Lemmas 7 and 8.
This correction does not affect the subsequent argument, as the upper bound in Lemma 8 still holds since k~k\tilde{k} \le k. We can apply Khintchine's inequality (upper bound) to obtain

k~mYˉdistinct12k~2mk2m,\frac{\tilde{k}}{m} \left| \bar{Y}_{\mathrm{distinct}} - \frac{1}{2} \right| \le \frac{\sqrt{\tilde{k}}}{2m} \le \frac{\sqrt{k}}{2m},

by conditioning on the set of distinct indices. This provides an even better constant 12\frac{1}{2} compared to 22\frac{\sqrt{2}}{2}, as the counters NjN_j are simply indicators. We will make this correction in the revised version and again thank the reviewer for the careful reading.

Q2: Indeed, our current lower bound proof is only non-vacuous when S=O(T2)S = O(T^2), due to the constant factor in (8). For general ST2S \ge T^2, the Ω(T/S)\Omega(T/\sqrt{S}) lower bound still holds. This can be remedied by selecting the block length to be 1/S21/S^2 (instead of 1/S1/S as in the paper), so that the constant factor in (8) becomes O(1/S)O(1/S), which is negligible compared to the main term T/ST/\sqrt{S}. We will clarify this in the revision.

Randomness of Predictors: Indeed, for the 0–1 loss, it is necessary for the predictor to be randomized in order to achieve sublinear regret in the adaptive case. We will clarify in Protocol 1 that the predictor h^s\hat{h}_s is allowed to use internal randomness.

评论

Thank you for your response. After reviewing the rebuttal, I believe the issues I raised in my review have been adequately addressed. I will maintain my positive score and recommend acceptance.

审稿意见
4

This paper considers a continuous online learning framework in which a learner selects discrete query times at which the data stream is observed (to evaluate the loss) and the learning strategy is updated. This framework was recently studied in [7], but only in the realizable setting—where the data is assumed to be generated by a function within the hypothesis class. In contrast, the present work focuses on the more challenging agnostic setting. The authors generalize the notion of expected regret to this continuous-time framework by integrating the loss over time, rather than summing it at discrete points.

Two settings are considered: the oblivious data stream, where the data is generated arbitrarily in advance and independently of the learner’s randomness; and the adaptive data stream, where the data stream may depend on any past information. For both settings, the authors propose a black-box reduction to an expert algorithm, using the observed losses at the random query times as unbiased estimators of the integrated continuous loss. The regret bound then follows by plugging these estimators into the standard regret analysis of the expert algorithm.

In addition to upper bounds, the authors provide a nearly matching lower bound (up to logarithmic factors) in the oblivious setting, showing that the regret guarantees are essentially optimal.

优缺点分析

Strenghts:

  • The paper is well-written and easy to follow.
  • The problem considered is interesting and the authors provide a clean and complete analysis with tight upper and lower bounds.

Weaknesses:

  • The results are restricted to finite hypothesis classes. Is it not possible to extend the black-box reduction to continuous hypothesis classes, and more generally to reduce to any discrete online convex optimization algorithm, rather than limiting to expert algorithms?
  • The paper would benefit from synthetic numerical experiments to illustrate the practical implications of the continuous-time framework on regret. In particular, when setting T=ST = S, we would expect to recover similar performance—does this hold in practice? Additionally, how stable are the results, given that they rely on an estimation technique?
  • The proofs appear to be fairly standard and rely on combining known results: the agnostic lower bound follows techniques from [1]; the sampling process for query points is borrowed from [7]; and the use of unbiased loss estimators within an expert algorithm mirrors approaches commonly used in adversarial bandits and online learning with partial feedback.

问题

  • Is it possible to derive high-probability guarantees, rather than bounds only in expectation? Alternatively, can one obtain upper bounds on the actual expected regret (i.e., using the infimum over hh in Equation (5)) instead of pseudo-regret bounds (i.e., the expectation as in Equation (4))?

  • The results are presented differently for the oblivious and adaptive settings. Is it not possible to unify the presentation, perhaps by giving a generic analogue of the black-box reduction in Theorem 1 for the adaptive setting—similar to what is done for the oblivious case—rather than relying on the more case-specific Lemma 4 and Theorem 2? The motivation for diverging in presentation is not entirely clear.

局限性

Yes

最终评判理由

I have read other reviews and the rebuttal. I think the authors address well my concerns. In particular, they added synthetic experiments that seem to well illustrate their theoretical results. I lean toward acceptance.

格式问题

NA

作者回复

We thank the reviewer for the thoughtful feedback and helpful comments. We address the questions raised below:

W1: In fact, our results for the adaptive case also apply to hypothesis classes with finite Littlestone dimension, not just finite classes, as shown in Theorem 2. For general OCO algorithms, our reduction could still apply by running the algorithm (e.g., OGD or FTRL) on the weighted loss function. We only require a regret guarantee of the form stated in Lemma 3. We thank the reviewer for pointing this out and will be happy to include this clarification in the revised version.

W2 & Q1: Our regret guarantee relies on estimating the loss, which is bounded by ΔT/S\Delta \approx T/S. For S=TS = T, the variance is constant; in this case, an application of Freedman's inequality for martingales (with filtration induced by tst_s) yields a high-probability bound on the learner's risk (or any fixed comparator risk) with deviation O(S)=O(T)O(\sqrt{S}) = O(\sqrt{T}). For general SS, the deviation scales as O(ΔS)=O(T/S)O(\Delta \sqrt{S}) = O(T/\sqrt{S}), which matches the order of our expected regret O~(TLdim(H)/S)\tilde{O}(T\sqrt{\mathsf{Ldim}(\mathcal{H})/S}).

In the oblivious case, the learner's high-probability risk can be translated into a high-probability bound for the actual (pointwise) regret. However, for the adaptive case, it is unclear whether such a high-probability pointwise regret bound is attainable, since the optimal comparator may depend on the learner’s internal randomness. To our knowledge, no such adaptive high-probability pointwise regret bound exists even in the adversarial bandit setting. Nevertheless, the above argument still yields the following weaker form of adaptive high-probability regret: for any given expert hHh \in \mathcal{H}, with high probability, the (pointwise) regret of the learner against hh is O~(TLdim(H)/S)\tilde{O}(T\sqrt{\mathsf{Ldim}(\mathcal{H})/S}).

W3: We agree that many of our technical ingredients are inspired by classical ideas. However, synthesizing these ideas in our (novel) setting is far from straightforward. For example, the selection of the weighting factor is derived from the randomness of the query times tst_s and is not arbitrarily chosen. The fact that it may appear natural in hindsight does not imply that it is easy to derive or even to recognize it's existence. Moreover, our lower bound proof in Appendix D significantly extends the construction in Ben-David et al. [1] by designing an oblivious process (see Remark 4 and the accompanying discussion).

Q2: We thank the reviewer for the helpful suggestion. As noted in our response to W1, a general reduction is indeed achievable in the OCO framework. We will include this discussion in the main text following Theorem 2.

Synthetic Experiments:
We conducted synthetic experiments to demonstrate the robustness of our algorithm. We used the expert class

H=hθ:=1xθ:θ0,0.001,,1.000,\mathcal{H} = \\{ h_{\theta} := \mathbf{1}\\{x \le \theta\\} : \theta \in \\{0, 0.001, \ldots, 1.000\\} \\},

so that H=1000|\mathcal{H}| = 1000. We set the time horizon to T=5T = 5 and discretize it into 1000 small blocks to approximate the integral regret without normalization. So, the simulated regret values reported below are scaled by a factor of 200×200\times compared to the actual continuous-time regret used in the paper. We simulate our Algorithm 1 under different query budgets SS, using the Exponential Weights Algorithm (EWA) as the expert algorithm B\mathcal{B}, on the following data streams (all features are generated iid uniform from [0,1][0,1]):

  1. alt_blocks: The stream is divided into SS equal chunks where the label for the chunks follows an (deterministic) alternating 0/1 pattern.
  2. iid: Each label is drawn independently from a Bernolli source with parameter p=0.3p=0.3.
  3. realizable: The labels are generated by a ground-truth function from H\mathcal{H}.
  4. upper bnd.: Our theoretical upper bound.
  5. lower bnd.: Our theoretical lower bound.

The considered data processes span a broad spectrum, ranging from stationary to highly non-stationary regimes. Results are shown below (each experiment is repeated 100 times, and the average regret is reported):

SSalt_blocksiidrealizableupper bnd.
5161.98172.93197.131175.9
10102.30138.96144.92831.4
2065.9199.07120.65588.0
4057.1679.3486.10415.6
8047.6570.9663.50293.3
16035.9761.7648.97207.6
32041.8635.4736.01147.0
64042.1827.7322.81103.9

We also conducted simulations on the hard data stream lower_stream used in our lower bound proof (cf. Appendix D):

SSlower bnd.lower_streamupper bnd.
571.6132.771175.9
1050.642.80831.4
2035.889.83588.0
4025.339.83415.6
8017.929.51293.3
16012.627.75207.6
3208.928.40147.0
6406.333.08103.9

Observation 1: For all four data streams, the regret is upper bounded by our theoretical upper bound. In fact, the curves have very similar slopes when plotted on a log-log scale, demonstrating that the 1/S1/\sqrt{S} order is genuine.

Observation 2: The regret for the lower_stream data is tightly sandwiched between our theoretical lower and upper bounds, demonstrating the optimality of our theoretical analysis.

We expect similar results to hold for Algorithm 2 as well, since the weighting factor lies in [0,1][0,1] and does not increase the variance of the loss estimation. We are happy to incorporate the experimental results reported above in our revision, along with additional discussion.

评论

I thank the authors for their rebuttal, which clarified several points. After reading the other reviews and corresponding rebuttals, I will keep my score for now and wait for the discussion with the other reviewers.

审稿意见
4

This paper studies agnostic online learning from continueous-time data streams. learners have to interact with continually envolving data stream while making queries and updating model at strategically selected times. This paper develop theoretical framework for learning from both oblivious and adaptive data streams.

优缺点分析

Strengths: This paper develops generic algorithmic frameworks for agnostic continuous-time online learning under both oblivious and adaptive data streams with optimal regret guarantees, giving both upper and lower bounds.

Weakness: No experimental results on either synthetic or real world data are conducted in the paper, it remains unclear whether the algorithms proposed can be used in practice.

问题

  1. The paper builds on and extends the update-and-deploy framework of Devulapalli and Hanneke [7], which focused on the realizable setting. Could the authors clearly delineate the conceptual and technical differences and challenges introduced by the agnostic setting?
  2. The paper proposes a novel importance weighting scheme (Algorithm 3) to obtain unbiased loss estimates under adaptive query times with dynamic epochs. Could the authors clarify how sensitive is the performance and regret bound to the exact choice of weighting function?
  3. Practical data streams can be complex and non-stationary beyond Littlestone dimension and scale characterization. Could the authors discuss or exemplify scenarios or hypothesis classes where these bounds translate to meaningful learning guarantees in real-world continuous data and elaborate on how these theoretical guarantees relate to computational or sample complexity in applications like high-frequency trading or personalized recommendations?

局限性

Yes, the authors have acknowledged and discussed the limitations in Appendix A,

格式问题

no

作者回复

We thank the reviewer for the thoughtful feedback and helpful comments. We address the questions raised below:

Q1: The primary conceptual difference compared to the framework of Devulapalli and Hanneke [7] is that our setting does not assume the existence of a perfect labeling rule from the hypothesis class. We also explicitly treat the query budget as a resource, which allows us to derive a precise quantitative trade-off between regret and the query budget.

From a technical standpoint, the work in [7] applies dynamic querying without weighting, leveraging the fact that the plain loss serves as an overestimate that is sufficient to bound mistakes in the realizable setting. However, in the agnostic setting, such an overestimate is insufficient to lower bound the competitor's loss. Our main technical innovation is the introduction of a weighting factor that produces exactly unbiased loss estimates, along with novel reductions to classic expert algorithms. Additionally, the lower-bounding techniques presented in Section 3 and Appendix D are entirely novel compared to those in [7].

Q2: The weighting factor was not arbitrarily chosen, but rather derived from the randomness in the query times tst_s, as shown in Lemma 2. In fact, other methods of selecting tst_s could also work, potentially leading to different weighting factors using the same derivation technique as in the proof of Lemma 2. Note that the weighting factor introduced in our Algorithm 3 lies in [0,1][0,1], so it does not increase the variance of the estimator used in Algorithm 1 (which we simulate below).

Q3: In fact, we consider the worst-case scenario; that is, our results apply to any non-stationary data-generating process. The "complexity" our framework addresses is the query (or update) budget. For example, in high-frequency trading, one must deploy a model that operates in real time, and updating the model can be costly—whether due to retraining overhead or concerns about system stability. As a result, updates must be sparse and strategic. Our work provides a worst-case performance baseline for such applications.

Synthetic Experiments:
We conducted synthetic experiments to demonstrate the robustness of our algorithm. We used the expert class

H=hθ:=1xθ:θ0,0.001,,1.000,\mathcal{H} = \\{ h_{\theta} := \mathbf{1}\\{x \le \theta\\} : \theta \in \\{0, 0.001, \ldots, 1.000\\} \\},

so that H=1000|\mathcal{H}| = 1000. We set the time horizon to T=5T = 5 and discretize it into 1000 small blocks to approximate the integral regret without normalization. So, the simulated regret values reported below are scaled by a factor of 200×200\times compared to the actual continuous-time regret used in the paper. We simulate our Algorithm 1 under different query budgets SS, using the Exponential Weights Algorithm (EWA) as the expert algorithm B\mathcal{B}, on the following data streams (all features are generated iid uniform from [0,1][0,1]):

  1. alt_blocks: The stream is divided into SS equal chunks where the label for the chunks follows an (deterministic) alternating 0/1 pattern.
  2. iid: Each label is drawn independently from a Bernolli source with parameter p=0.3p=0.3.
  3. realizable: The labels are generated by a ground-truth function from H\mathcal{H}.
  4. upper bnd.: Our theoretical upper bound.
  5. lower bnd.: Our theoretical lower bound.

The considered data processes span a broad spectrum, ranging from stationary to highly non-stationary regimes. Results are shown below (each experiment is repeated 100 times, and the average regret is reported):

SSalt_blocksiidrealizableupper bnd.
5161.98172.93197.131175.9
10102.30138.96144.92831.4
2065.9199.07120.65588.0
4057.1679.3486.10415.6
8047.6570.9663.50293.3
16035.9761.7648.97207.6
32041.8635.4736.01147.0
64042.1827.7322.81103.9

We also conducted simulations on the hard data stream lower_stream used in our lower bound proof (cf. Appendix D):

SSlower bnd.lower_streamupper bnd.
571.6132.771175.9
1050.642.80831.4
2035.889.83588.0
4025.339.83415.6
8017.929.51293.3
16012.627.75207.6
3208.928.40147.0
6406.333.08103.9

Observation 1: For all four data streams, the regret is upper bounded by our theoretical upper bound. In fact, the curves have very similar slopes when plotted on a log-log scale, demonstrating that the 1/S1/\sqrt{S} order is genuine.

Observation 2: The regret for the lower_stream data is tightly sandwiched between our theoretical lower and upper bounds, demonstrating the optimality of our theoretical analysis.

We are happy to incorporate the experimental results reported above in our revision, along with additional discussion.

审稿意见
5

This paper extends te setting of (realizable) continuous-time online learning from Devulapalli and Hanneke to the agnostic case. In this framework, there is a stream in continuous time of queries XtX_t, and the algorithm needs to make online predictions on the stream by picking a hypothesis to use with a few discrete observation of the true labels YtY_t to the stream. The original assumed that the labels in the stream were generated by some hypothesis in the hypothesis class. This paper drops this assumption, and show algorithms (and some matching lower bounds) for both oblivious and adaptive adversaries. In the oblivious case the algorithm splits the time-horizon in equal length epochs and samples a uniform time for each epoch to see the label and use as a black-box an OL algorithm. In the adaptive case this may not work anymore, and they show how to use dynamic epochs, also used by Devulapalli and Hanneke, and dynamic step sizes on EWA to circumvent this issue.

优缺点分析

This paper studies a very natural (and more challenging) extension of the continuous-time online learning framework proposed by Devulapalli and Hanneke, which by itself it already a somewhat natural problem. Many of the techniques in the algorithms seem to drawn upon the ideas of DH24, but are certainly not trivial extensions in any way. The algorithm (or, better saying, reduction) for the oblivious case is clean and serves as a nice way to "warm-up" to the adaptive case, although the oblivious case is interesting in itself. Also, the slight modification of EWA leads to a (relatively) elegant analysis. Unfortunately I did not have the time to check the proofs as I would like to have done, but the few derivations that I had time to check were relatively clearly exposed and seemed to be correct.

Overall, although there does not seem to be major novel ideas in the framework nor in the algorithms, this is a quite solid contribution, and shows optimal rates for both the oblivious and adaptive cases, which makes this a even more complete and solid work. While I have some questions, none of them seem to be crucial for clarity/significance/originality/quality of the paper.

问题

Small technical question, but DH24 assume the process ZtZ_t to be countable. I don't think this matters too much except on some of the technical conditions required later on (such as measurability of the sample paths). Do the authors think there are some technical kinks that might need to be ironed out because of this?

局限性

The limitations of the setup seem to be well discussed in the paper.

最终评判理由

I did not have many concerns, and it seems like the other reviewers only raised minor points. I still lean towards accepting this submission.

格式问题

No formatting concerns

作者回复

We thank the reviewer for the thoughtful feedback and helpful comments. Indeed, in our paper, we explicitly assume that the function induced by the loss is measurable for any sample path, so that the risk defined via integration is well-defined. The measurability and boundedness of f(t)f(t) ensure that the function is absolutely integrable; thus, the exchange of the order of integration in Lemma 2 follows from Fubini’s theorem, which does not require assumption on the countability of the ZtZ_t's.

From a practical point of view, one can also consider a sufficiently small partition of the time horizon to approximate the integral and thereby avoid measurability concerns—in which case, our result still holds.

评论

I would like to thank the authors for their replies to my review and also to the other reviewers. I did not have many concerns, and it seems like the other reviewers only raised minor points. I still lean towards accepting this submission.

最终决定

This paper provides the first guarantees for continuous-time online learning in the agnostic setting for hypothesis classes with a finite Littlestone dimension. All of the reviewers were positive about this submission and overall in favor of acceptance. For the camera-ready version, the authors need to include the fix to the bug in Lemma 6 that was discovered during the review process. Additionally, they are highly recommended to add the experimental results that they shared in the rebuttal, and provide context for the computational complexity and generality of the procedure that some reviewers believed was missing/unclear in the initial submission.