PaperHub
6.4
/10
Poster4 位审稿人
最低4最高4标准差0.0
4
4
4
4
3.5
置信度
创新性2.8
质量2.5
清晰度2.3
重要性2.8
NeurIPS 2025

Self-Training with Dynamic Weighting for Robust Gradual Domain Adaptation

OpenReviewPDF
提交: 2025-05-09更新: 2025-10-29
TL;DR

The paper introduces a dynamic self-training method for gradual domain adaptation that effectively transfers knowledge across domains.

摘要

In this paper, we propose a new method called Self-Training with Dynamic Weighting (STDW), which aims to enhance robustness in Gradual Domain Adaptation (GDA) by addressing the challenge of smooth knowledge migration from the source to the target domain. Traditional GDA methods mitigate domain shift through intermediate domains and self-training but often suffer from inefficient knowledge migration or incomplete intermediate data. Our approach introduces a dynamic weighting mechanism that adaptively balances the loss contributions of the source and target domains during training. Specifically, we design an optimization framework governed by a time-varying hyperparameter $\varrho$ (progressing from 0 to 1), which controls the strength of domain-specific learning and ensures stable adaptation. The method leverages self-training to generate pseudo-labels and optimizes a weighted objective function for iterative model updates, maintaining robustness across intermediate domains. Experiments on rotated MNIST, color-shifted MNIST, portrait datasets, and the Cover Type dataset demonstrate that STDW outperforms existing baselines. Ablation studies further validate the critical role of $\varrho$'s dynamic scheduling in achieving progressive adaptation, confirming its effectiveness in reducing domain bias and improving generalization. This work provides both theoretical insights and a practical framework for robust gradual domain adaptation, with potential applications in dynamic real-world scenarios.
关键词
Gradual Domain AdaptationAdversarial Domain AdaptationDistribution ShiftOut-of-distribution Generalization

评审与讨论

审稿意见
4

In this paper, the authors propose a novel gradual self-training method named Generalized Dynamic Osmosis (STDW), which largely improves the performance of the model on the gradual domain adaptation problem.

The method involves three innovations. First, instead of assigning labels for the next whole domain and keeping them unchanged when fine-tuning, STDW uses a batchwise strategy of predicting pseudo-labels for a mini-batch of data so the pseudo-labels are consistently updated by a new model. Second, they use cyclic batch matching to facilitate robust inter-domain adaptation. Third, a weighted loss with a dynamic weight is adopted to stabilize the training process.

The authors conduct some experiments to validate the efficacy of the method and the effect of some hyperparameters.

优缺点分析

Pros:

  • The method is interesting and effective. First, the pseudo-labeling dynamic learning can make more efficient use of the current model since they predict the pseudo-labels using the up-to-date model. Second, the cyclic batch matching with a dynamic weight can stabilize the training process and decrease the bad effect of the incorrect pseudo-labels. The method is novel to me, and the effectiveness is verified by the experiments.

Cons:

  • The writing needs to be polished, especially the math notations. Those conflicting notations largely hinder a smooth reading for the readers and raise some misunderstandings. There are so many conflicting math notations. In Line 107, you use kk to denote the class number, while it also represents the batch index later in Line 169. In Line 113, the model is a mapping from XX to Y={1,2,...,k}Y=\{1,2,...,k\}, while the expressions later in Eq. (2), (3) indicate it is a mapping from XX to [0,1]k[0,1]^k. In Line 127, you add the signsign to denote the final prediction function f^\hat{f} but this is only for binary classification---you use a multi-classification setting in Line 107. In Line 146, you use DD to denote a domain set. Since it does not have a subscript, I was wondering if this indicates a dataset combined from all domains or if it indicates one specific domain DtD_t? In Eq. (11), xix_i or xx? In Line 176, the domain is denoted as DiD_i, while the index in the previous section is tt for the domain (Line 128). I totally understand that the gradual domain adaptation problem, which involves many domains, increases the complexity of the notation system, but those conflicting notations make it difficult for readers to understand the method of this work.
  • The motivations of the method should be explained more clearly. In section 4.2, the motivation of combining two batches from two adjacent domains is not clear. To my understanding, I think the idea is a little bit similar to Momentum. When the weight ρ\rho is small at the early stage, the model is trained based on the loss on the previous domain rather than the current domain, where the accuracy of the pseudo-label could be high. So this strategy could prevent the model from collapsing due to the incorrect pseudo-labels of the current domain.
  • The framework illustration in Fig. 2 is not clear. What is the meaning of "sync" in the figure? To my understanding, you use only one classifier to learn the batches BtB_t and Bt+1B_{t+1}. I suggest replacing "sync" with "shared weights," which is more widely used in papers. Besides, what is HtH_t? In the method, you never use the feature extractor and classifier in a separate way. So is it really necessary to separate the network into three parts—classifier, HtH_t, and feature extractor—in the illustration figure?
  • More ablation studies are needed. I suggest the authors validate the effects of the three components in Section 4 via an ablation study. For example, you could remove the cyclic batch matching and compare the results to show the important role of cyclic batch matching.
  • As the authors note in Line 141, the mislabeling would hurt performance. So it would be more convincing if the average accuracy of the pseudo-labels on intermediate domains can be shown and compared to further validate the effectiveness of your method.
  • Another baseline method. The work of [1] proposes an adversarial self-training method, which significantly improves the performance of GDA. Please consider comparing your method with [1].

[1]: Adversarial Self-Training Improves Robustness and Generalization for Gradual Domain Adaptation. Lianghe Shi, Weiwei Liu. NeurIPS 2023.

问题

  • In the GST method, a new model is trained from scratch on each domain, while this work trains only one model and fine-tunes the model in each iteration. I am a little curious about whether keeping fine-tuning only one model would also improve the performance of the GST method.

局限性

Yes, the authors discuss some limitations of their method in the appendix.

最终评判理由

The authors have resolved the notation conflicts and added additional experiments in their rebuttal. However, I remain concerned about their reported results for the baseline method AST. I have independently implemented the AST method and found that it can achieve approximately 97% accuracy on the Rotated MNIST dataset, which is very close to the performance reported for their proposed method (97.6%). In contrast, the authors report an accuracy of only 92.5% for AST.

Overall, I remain neutral regarding the acceptance of this work.

格式问题

No

评论

Thanks for the detailed response. Most of my concerns are addressed well and I will increase the rate to 4.

But I am still concerned about the baseline method AST you compared. Did you use the implementation from their github and adopt the optimal hyper parameters used in their paper? They set two important hyper parameters τ=9,ζ=0.05\tau=9, \zeta=0.05 to get the optimal accuracy of 97.15 on rotating mnist.

评论

Thank you for your comments! We have confirmed that we have reproduced the relevant code and used the same parameters. Our experimental parameters are based on GOAT [1], and we have optimized the gst-related code in GitHub. At this point, gst performs better than goat in the paper during reproduction. If the article is accepted, we will release our code, including the code for other reproduction methods. Once again, we appreciate your affirmation of our method!

[1] https://arxiv.org/pdf/2310.13852

评论

Futhermore, we provide the Lyapunov Stability Proof for Cyclic Batch Matching.

Let θt\theta_t be the parameter vector after step tt and define the mixed loss

L(θt)=ϱt(θt,Bt(d))+(1ϱt)(θt,Bt(d1)),L(\theta_t)=\varrho_t\,\ell(\theta_t,\mathcal{B}_t^{(d)})+\bigl(1-\varrho_t\bigr)\,\ell(\theta_t,\mathcal{B}_t^{(d-1)}),

where \ell is cross-entropy,

Bt(d)Dd\mathcal{B}_t^{(d)}\subset D_d

and

Bt(d1)Dd1\mathcal{B}_t^{(d-1)}\subset D_{d-1}

form a cyclically matched pair, and ϱt[0,1]\varrho_t\in[0,1] increases monotonically. Suppose \ell is LL-smooth and μ\mu-strongly convex in a neighbourhood of the joint optimum θ\theta^{\ast}. Consider the Lyapunov candidate V(θt)=12θtθ2V(\theta_t)=\tfrac12\|\theta_t-\theta^{\ast}\|^2. With a gradient-descent update θt+1=θtηθL(θt)\theta_{t+1}=\theta_t-\eta\nabla_{\theta}L(\theta_t) and any step size η(0,2/L)\eta\in(0,2/L), we have

V(θt+1)=V(θt)η(μLη2)θtθ2.V(\theta_{t+1}) = V(\theta_t) - \eta\Bigl(\mu-\tfrac{L\eta}{2}\Bigr)\|\theta_t-\theta^{\ast}\|^2.

Choosing η<2μ/L2\eta<2\mu/L^2 makes the bracket positive, hence V(θt+1)<V(θt)V(\theta_{t+1})<V(\theta_t) unless θt=θ\theta_t=\theta^{\ast}. Because the cyclic schedule enumerates every cross-domain batch exactly once per epoch, the loss remains LL-smooth and μ\mu-convex throughout training, ensuring that VV is strictly decreasing. Therefore VV qualifies as a Lyapunov function establishing global asymptotic stability of the optimisation trajectory under cyclic batch matching.

评论

Response to Reviewer #KmnV

Thank you for your valuable suggestions. As your suggestions involve many experiments, we apologize for the late reply. Please let us know if you have any further comments.

Q1 (Notation).

A: Thank you for flagging this issue. In the revision we introduce a single, page-two symbol table that defines every recurring variable. The number of classes is now KK, batch indices are bb, and domain indices are dd. The model is consistently written fθ ⁣:X ⁣ ⁣Δ(K)f_{\theta}\!:\mathcal{X}\!\to\!\Delta^{(K)} and pseudo-labels are y^(x)=argmaxkfθ(x)k\hat{y}(x)=\arg\max_k f_{\theta}(x)_k. A domain set is D={D0,,Dn}\mathcal{D}=\{D_0,\dots,D_n\} and an individual domain is DdD_d. These changes eradicate the earlier symbol collisions between class count and batch index, binary and multi-class mappings, and domain collections versus single domains.

Q2 (Motivation of two-domain batch composition).

“In Section 4.2, the motivation of combining two batches from two adjacent domains is not clear. … I think the idea is a little bit similar to Momentum …”

A: First, when the dynamic weight ϱ\varrho is small at the start of adaptation, gradients are dominated by the safer pseudo-labels from the previous domain Dd1D_{d-1}, thereby damping noise from the current domain DdD_d. Second, in an optimal-transport perspective, adjacent domains occupy neighbouring regions of feature space; pairing their mini-batches traces a geodesic of minimal 1-Wasserstein cost, yielding a smoother optimisation trajectory. A proof based on Lyapunov stability—included below—shows the mixed-batch update is globally convergent under standard smoothness and strong-convexity assumptions.

Q3 (Framework illustration in Fig. 2).

“What is the meaning of ‘sync’ in the figure? … is it really necessary to separate the network into three parts …?”

A: The revised figure now labels the shared parameters explicitly as “shared weights’’ rather than “sync’’ and depicts only two actual modules: a feature extractor gg and a classifier hh. The dynamic-weight loss is drawn directly on the joint head so that no artificial third block appears.

Q4 (More ablation studies).

“I suggest the authors validate the effects of the three components …”

We have added another ablation study. Removing each component in turn demonstrates its independent contribution.

ConfigurationRot-MNIST Acc (%)
Base model (no dynamic labels, no cycle, no ww)83.8
+ Dynamic pseudo-label refresh89.7
+ Cycle batch matching92.1
Full STDW (adds dynamic weight ϱ\varrho)97.6

Q5 (Pseudo-label accuracy).

“It would be more convincing if the average accuracy of the pseudo-labels on intermediate domains can be shown …”

We add an experiments for pseudo-label accuracies across nine Portrait domains. STDW maintains 96.2%96.2\% on D0D_0 and never dips below 72.8%72.8\% at D4D_4, whereas GST bottoms out at 65.7%65.7\%. We also compute a Pearson correlation of r=0.86r=0.86 between intermediate pseudo-label fidelity and final-domain accuracy, confirming the robustness of our STDW.

Q6 (Additional baseline AST 2023).

“Please consider comparing your method with [Lianghe Shi & Weiwei Liu, NeurIPS 2023].”

A: We have reproduced Adversarial Self-Training (AST) under identical backbones and hyper-parameters. The table below summarises the results.

MethodRot-MNISTColour-ShiftPortraitsCoverType
AST (NeurIPS ’23)92.594.784.972.8
STDW (Ours)97.698.387.174.2

STDW outperforms AST by two to five percentage points on every benchmark, which we attribute to our cyclic matching and dynamic weighting mechanisms.

Q7 (GST with fine-tuning).

“I am curious about whether keeping fine-tuning only one model would also improve the performance of the GST method.”

We implemented “GST-fine-tune,” keeping a single model throughout training. While this variant does improve GST, it remains behind STDW.

DatasetGSTGST-fine-tuneSTDW
Rot-MNIST83.886.997.6
Portraits82.683.887.1

The gap confirms that cyclic matching and adaptive weighting supply gains beyond continual fine-tuning.

评论

Dear AC and authors,

Since I did not receive the authors’ rebuttal in my system, I will keep the score unchanged.

Reviewer KmnV

审稿意见
4

The paper introduces Self-Training with Dynamic Weighting (STDW) to Gradual Domain Adaptation (GDA). It addresses the challenge of smooth knowledge migration from source to target domains by introducing a dynamic weighting mechanism controlled by a time-varying hyper-parameter ϱ, which balances the loss contributions of source and target domains during training. The method leverages self-training to generate pseudo-labels and optimizes a weighted objective function for iterative model updates. The authors provide theoretical insights and validate the method through extensive experiments on multiple benchmark datasets, demonstrating that STDW outperforms existing baselines and achieves SOTA performance.

优缺点分析

Strengths:

  1. Formal analysis (optimal transport, Lyapunov stability) provides error bounds for pseudo-label consistency and convergence (Appx. B).
  2. The paper introduces a novel dynamic weighting mechanism that adaptively balances the loss contributions of source and target domains during training. This innovation effectively addresses the challenges of GDA.
  3. Extensive validation across 6 datasets, varying intermediate domains (0–6), and 15 corruption types. Ablation studies confirm criticality of ϱ scheduling and intermediate domains.
  4. The proposed STDW method has demonstrated significant improvements in handling continuous distribution shifts compared to existing methods.

Weaknesses:

  1. The performance of STDW relies heavily on the availability and quality of intermediate domains, which may not always be accessible or well-defined in real-world scenarios. This limitation could restrict the applicability of the method in certain practical situations.
  2. The dynamic weighting mechanism introduces additional hyperparameters (e.g., the scheduling of ϱ) that require careful tuning, potentially increasing the computational overhead during training.
  3. The method assumes a smooth transition between domains, which might not hold for highly discontinuous or abrupt domain shifts. The authors acknowledge this limitation and suggest that future work could explore ways to handle more diverse and challenging domain shifts.

问题

  1. The paper assumes a smooth transition between domains. How would the method perform in scenarios with highly discontinuous or abrupt domain shifts? Could the authors provide some insights or potential solutions for such cases?
  2. The linear ϱ schedule (0→1) is heuristic. Are there theoretical or data-driven guidelines (e.g., based on domain discrepancy)? Could ϱ be learned?
  3. The paper mentions that the performance of STDW relies on the availability and quality of intermediate domains. What strategies could be employed to generate or select high-quality intermediate domains in cases where they are not readily available?
  4. Cyclic batch matching and iterative pseudo-labeling may increase training time. How does STDW’s complexity compare to baselines (e.g., GST)?

局限性

Yes, The authors have adequately addressed the limitations of their work in the paper.

格式问题

No

评论

Q4. Cyclic batch matching and iterative pseudo-labeling may increase training time. How does STDW’s complexity compare to baselines (e.g., GST)?

A: Analytically, the per-step cost of STDW is dominated by the forward/backward pass exactly as in conventional training. Iterative pseudo-labeling does not require an extra pass because we reuse the logits already computed for the current mini-batch; thresholding or argmax\arg\max is computed on-the-fly. Cyclic batch matching pairs each batch from Dd1D_{d-1} with one from DdD_d by round-robin indices, which is an O(1)O(1) operation at iteration time and O(B)O(B) bookkeeping per epoch where BB is the number of batches; no global assignment is solved. In contrast, GST trains a fresh model per domain stage, which multiplies optimization warm-up and increases total epochs. Therefore, under equal batch sizes and epochs per stage, STDW’s wall-clock is typically lower than GST’s despite dynamic weighting and cyclic pairing.

Lyapunov Stability Under Cyclic Matching and Switching Anchors in Q2

For completeness, we extend stability argument to the piecewise-smooth path constructed for discontinuous shifts. Let θt\theta_t denote the parameters at step tt and consider the mixed loss on a matched pair of mini-batches

Ld(θt)=ϱt(θt,Bt(d))+(1ϱt)(θt,Bt(d1)),L_d(\theta_t)=\varrho_t\,\ell\big(\theta_t,\mathcal{B}^{(d)}_t\big)+\bigl(1-\varrho_t\bigr)\,\ell\big(\theta_t,\mathcal{B}^{(d-1)}_t\big),

where \ell is cross-entropy and ϱt[0,1]\varrho_t\in[0,1]. Assume \ell is LL-smooth and μ\mu-strongly convex in a neighbourhood of the local joint minimizer θd\theta_d^{\ast} for the pair (Dd1,Dd)(D_{d-1},D_d). With gradient descent θt+1=θtηLd(θt)\theta_{t+1}=\theta_t-\eta\nabla L_d(\theta_t) and step size η(0,2/L)\eta\in(0,2/L), define the Lyapunov function Vd(θ)=12θθd2V_d(\theta)=\tfrac12\|\theta-\theta_d^{\ast}\|^2 . Standard smooth strongly-convex analysis yields

Vd(θt+1)Vd(θt)η(μLη2)θtθd2,V_d(\theta_{t+1}) \le V_d(\theta_t) - \eta\Bigl(\mu-\tfrac{L\eta}{2}\Bigr)\|\theta_t-\theta_d^{\ast}\|^2,

so VdV_d strictly decreases whenever θtθd\theta_t\neq\theta_d^{\ast} and η<2μ/L2\eta<2\mu/L^2. Cyclic matching enumerates all batch pairs once per epoch without changing smoothness or curvature, so the decrease persists across iterations. For a discontinuous path approximated by anchors {A0,,Aq}\{A_0,\dots,A_q\} we obtain a switched system with modes d{1,,q}d\in\{1,\dots,q\}. If there exists a common quadratic Lyapunov function V(θ)=12θθˉ2V(\theta)=\tfrac12\|\theta-\bar{\theta}^{\ast}\|^2 valid in the union of neighbourhoods of {θd}\{\theta_d^{\ast}\}—for instance when θdθˉ\|\theta_d^{\ast}-\bar{\theta}^{\ast}\| is bounded and the Hessians share eigenvalue bounds (μ,L)(\mu,L)—then the same inequality holds for all modes with the same η\eta, ensuring global asymptotic stability under arbitrary switching. When a common Lyapunov function is too conservative, we adopt a dwell-time condition implemented by the discrepancy-aware controller: the mode can switch (i.e., we advance ϱ\varrho or move to the next anchor) only after the gradient norm on the current mode falls below a tolerance and the consistency gates are met. In both cases, the energy decreases monotonically, preventing divergence even under abrupt transitions.

评论

Response to Reviewer #BwCA

Q1. The paper assumes a smooth transition between domains. How would the method perform in scenarios with highly discontinuous or abrupt domain shifts? Could the authors provide some insights or potential solutions for such cases?

A: We agree that abrupt, non-smooth shifts stress any gradual adaptation procedure. In our framework, two practical modifications improve robustness without changing the core objective. First, we replace the single linear schedule by a discrepancy-aware controller that reacts to a jump. Concretely, when the measured discrepancy between the current and previous batches exceeds a threshold—using, for example, an empirical 1-Wasserstein distance or an MMD estimate computed in feature space—we temporarily reduce ϱ\varrho so that the loss is dominated by the better-calibrated pseudo-labels from Dd1D_{d-1}. This behaves like a safety brake and prevents large, noisy gradients caused by unreliable pseudo-labels on the new domain. Second, we construct a piecewise-smooth path by clustering the unlabeled target pool and ordering clusters along a minimum-spanning tree defined by an OT metric, effectively creating “virtual” intermediate anchors. STDW then runs over this induced path; cyclic matching operates between successive anchors even when the original shift is discontinuous. From a stability viewpoint, this leads to a switched-system interpretation in which each anchor pair admits the same Lyapunov argument as in the smooth case. If a common quadratic Lyapunov function exists in a neighbourhood of each local joint optimum, and if we enforce a minimal “dwell time” before advancing ϱ\varrho, then the overall trajectory remains stable.

Q2. The linear ϱ\varrho schedule (0 ⁣ ⁣1)(0\!\to\!1) is heuristic. Are there theoretical or data-driven guidelines (e.g., based on domain discrepancy)? Could ϱ\varrho be learned?

A: We now provide both a principled guideline and learnable variants. A data-driven guideline sets ϱ\varrho proportional to a normalized discrepancy proxy and to the estimated reliability of current-domain pseudo-labels. Let δ^t\widehat{\delta}_t be a bounded discrepancy score (e.g.,

δ^t=MMD(Dd1,Dd)MMD(Dd1,Dd1)+ϵ\widehat{\delta}_t=\frac{\mathrm{MMD}(D_{d-1},D_d)}{\mathrm{MMD}(D_{d-1},D_{d-1})+\epsilon}

mapped to [0,1][0,1]) and let q^t[0,1]\widehat{q}_t\in[0,1] estimate pseudo-label accuracy via agreement between two perturbations or via a small held-out split with delayed labels when available. A simple controller is

ϱt+1=clip(ϱt+κ(q^tγδ^t),0,1)\varrho_{t+1}=\mathrm{clip}\big(\varrho_t+\kappa\big(\widehat{q}_t-\gamma\,\widehat{\delta}_t\big),0,1\big)

with small gain κ\kappa and trade-off γ\gamma, which increases ϱ\varrho only when label reliability dominates discrepancy. Beyond such rules, ϱ\varrho can be learned end-to-end. A differentiable parametric schedule uses a sigmoid

ϱt=σ(αt+β)\varrho_t=\sigma(\alpha t+\beta)

where (α,β)(\alpha,\beta) are trained by implicit differentiation to minimize a held-out loss on DdD_d. Alternatively, we model

ϱt=σ(wst)\varrho_t=\sigma(\mathbf{w}^\top s_t)

where sts_t concatenates discrepancy features, confidence statistics, and training time; w\mathbf{w} is learned by bilevel optimization in which the inner loop updates θ\theta and the outer loop updates w\mathbf{w} to minimize the final-domain validation loss.

Q3. The paper mentions that the performance of STDW relies on the availability and quality of intermediate domains. What strategies could be employed to generate or select high-quality intermediate domains in cases where they are not readily available?

A: When explicit intermediate domains are missing, we find that feature-space interpolation and OT-barycentric synthesis provide stable substitutes with low engineering cost. We embed samples using the current encoder gg, compute an OT map TT from Dd1D_{d-1} to DdD_d using entropic regularization, and generate intermediates by barycenters zλ=(1λ)z+λT(z)z_\lambda=(1-\lambda)z + \lambda T(z) for λ(0,1)\lambda\in(0,1); the corresponding inputs are reconstructed either by a lightweight decoder or used directly in feature space by training the classifier head hh on (zλ,y^)(z_\lambda,\hat{y}). A simpler alternative uses distribution mixup, drawing pairs (xd1,xd)(x_{d-1},x_d) and training on x~=ϕ1((1λ)ϕ(xd1)+λϕ(xd))\tilde{x}=\phi^{-1}\big((1-\lambda)\phi(x_{d-1})+\lambda\phi(x_d)\big) where ϕ\phi is a fixed feature extractor (e.g., a self-supervised backbone). For selection rather than synthesis, we retrieve “bridge” samples from a large unlabeled pool by minimizing MMD({x},Dd1)+MMD({x},Dd)\mathrm{MMD}(\{x\},D_{d-1})+\mathrm{MMD}(\{x\},D_d), thereby constructing a narrow corridor of real samples that approximates the geodesic path. These strategies can be combined with weak style transfer or Fourier-domain amplitude mixing to further densify the path while keeping semantics intact.

评论

Dear reviewer, we sincerely appreciate your taking the time and effort to review our paper. As the rebuttal phase is drawing to a close, we would greatly appreciate it if you could take the time to review our response and engage in discussion. We believe our response addresses your main concerns, and we look forward to further discussion to ensure a fair and comprehensive evaluation of our work. Thank you again for your valuable contribution.

Sincerely,

Authors in NeurIPS 2025 Conference Submission10853.

审稿意见
4

This paper introduces Self-Training with Dynamic Weighting (STDW), a novel method designed for Gradual Domain Adaptation (GDA) that improves robustness by smoothly migrating knowledge from the source domain to the target domain. STDW generates dynamic pseudo-labels and uses a time-dependent hyperparameter to progressively shift learning emphasis from source to target domains. Empirical results on various benchmarks (demonstrate that STDW significantly surpasses existing GDA approaches.

优缺点分析

Strengths

  • Paper is clearly written.
  • Performance is strong.
  • Setting is practical.

Weaknesses Overall, I don't think the contribution from this paper is very novel.

  1. Generating dynamic pseudo-labels is already common in self-training [1]. This is very similar to the approach Section 4.1. FixMatch even has an additional step of selecting high-confidence pseudo-labels.
  2. The paper proposes cyclic batch matching in Section 4.2. But I am confused why data sampling is designed in such a manner. One can simply randomly sample a batch from the left and right domain each time. This simplifies the training pipeline, and I feel it wouldn't impact the performance too much.
  3. While the stepwise dynamic osmosis makes a lot of sense, it feels like an engineering technique (or hyperparameter tweaking) and offers little technical contribution.

[1] FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence

问题

  1. Does random sampling perform about the same with cyclic batch matching? If not, can you please offer an explanation why cyclic batch matching work?
  2. While I think gradual domain adaptation is a more practical assumption than traditional UDA, it can be difficult to collect data from gradual domains in practical situations. What is your view of this?

局限性

Yes.

最终评判理由

My concerns are resolved after reading authors' response. Hence I will increase my score.

I encourage authors to include the discussions about W1-W2 in the revision, which help clarify the contribution of the paper.

格式问题

No concern.

评论

Author Response to Reviewer pTMZ

1. “Generating dynamic pseudo-labels is already common in self-training. This is very similar to the approach in Section 4.1. FixMatch even has an additional step of selecting high-confidence pseudo-labels.”

We agree that refreshing pseudo-labels is a familiar ingredient in semi-supervised learning, including FixMatch. What distinguishes STDW is not merely that labels are refreshed, but when and how they are trusted. Our time-varying weight ϱ(t)\varrho(t) begins near zero, anchoring optimisation on the source domain until target pseudo-labels reach a reliability threshold; only then does ϱ\varrho rise toward one. In FixMatch, by contrast, the target loss dominates from the outset and collapses under distribution shift if early labels are wrong.

2. “The paper proposes cyclic batch matching in Section 4.2. But I am confused why data sampling is designed in such a manner. One can simply randomly sample a batch from the left and right domain each time. This simplifies the training pipeline, and I feel it wouldn’t impact the performance too much.”

Cyclic pairing gives every mini-batch from the preceding domain Dd1D_{d-1} exactly one partner from the current domain DdD_d per epoch, ensuring balanced gradients and low variance in the ϱ\varrho update. Random sampling may over-represent easy or hard sub-modes of DdD_d early on, causing ϱ\varrho to fluctuate and amplifying noise in the pseudo-labels. Because the Lyapunov proof assumes bounded gradient variance, the deterministic schedule aligns neatly with the theory, whereas random pairing weakens that guarantee.

3. “While the stepwise dynamic osmosis makes a lot of sense, it feels like an engineering technique (or hyper-parameter tweaking) and offers little technical contribution.”

The stepwise schedule is best viewed as an automatic pacing mechanism rather than a manual hyper-parameter. Its purpose is to align the model’s level of trust in target pseudo-labels with the stage of adaptation: low trust when the model is still source-biased, higher trust as it becomes more competent on the target. By embedding this pacing directly in the loss—through ϱ(t)\varrho(t)—we avoid per-dataset knob turning; the same rising profile works across all benchmarks we tested. Conceptually, the schedule fulfils the same role as a curriculum in curriculum learning: it lets the model tackle easier, better-labelled data first and only later rely on noisier labels. This controlled progression is the key technical idea behind “dynamic osmosis’’ and is what enables the method to migrate knowledge smoothly.

4. “Does random sampling perform about the same with cyclic batch matching? If not, can you please offer an explanation why cyclic batch matching work?”

We performed an ablation that keeps every other STDW component intact but replaces cyclic pairing by independent random draws. On Rot-MNIST accuracy drops from 97.6%97.6\% to 95.8%95.8\%, and on Colour-Shift from 98.3%98.3\% to 96.7%96.7\% (three seeds each). The variance reduction afforded by deterministic pairing explains the gap: every batch participates once, gradients are evenly weighted, and ϱ\varrho evolves smoothly. The table summarises the numbers.

Sampling strategyRot-MNIST (%)Colour-Shift (%)
Random pairing95.8 ± 0.396.7 ± 0.2
Cyclic pairing97.6 ± 0.298.3 ± 0.2

5. “While I think gradual domain adaptation is a more practical assumption than traditional UDA, it can be difficult to collect data from gradual domains in practical situations. What is your view of this?”

The stepwise schedule is best viewed as an automatic pacing mechanism rather than a manual hyper-parameter. Its purpose is to align the model’s level of trust in target pseudo-labels with the stage of adaptation: low trust when the model is still source-biased, higher trust as it becomes more competent on the target. By embedding this pacing directly in the loss—through ϱ(t)\varrho(t)—we avoid per-dataset knob turning; the same rising profile works across all benchmarks we tested. Conceptually, the schedule fulfils the same role as a curriculum in curriculum learning: it lets the model tackle easier, better-labelled data first and only later rely on noisier labels. This controlled progression is the key technical idea behind “dynamic osmosis’’ and is what enables the method to migrate knowledge smoothly.

评论

Dear reviewer, we sincerely appreciate you taking the time and effort to review our paper. As the rebuttal phase is drawing to a close, we would greatly appreciate it if you could take the time to review our response and engage in discussion. We believe our response addresses your main concerns, and we look forward to further discussion to ensure a fair and comprehensive evaluation of our work. Thank you again for your valuable contribution.

Sincerely,

Authors in NeurIPS 2025 Conference Submission10853

审稿意见
4

The authors propose a novel method, Self-Training with Dynamic Weighting (STDW), for the problem of Gradual Domain Adaptation (GDA). This method employs a self-training mechanism with pseudo-labels and introduces a dynamic hyperparameter ϱ, which evolves from 0 to 1 to balance the contributions of the source and target domains in the loss function. This dynamic weighting ensures a smooth and controlled knowledge transfer across intermediate domains during adaptation. This mechanism further incorporates a dynamic, mini-batch-wise pseudo-label generation strategy to mitigate the accumulation of pseudo-labeling errors during continual adaptation. The proposed method is evaluated on multiple image datasets and consistently outperforms both traditional domain adaptation approaches and other state-of-the-art GDA methods. Comprehensive ablation studies further confirm the essential role of the dynamic parameter ϱ and the use of intermediate adaptation steps, which together enhance stability and reduce domain bias.

优缺点分析

Strengths:

  1. Introduces a novel optimization framework employing a time-varying hyperparameter ϱ, which flexibly adjusts the relative contributions of the source and target domains during adaptation, thereby ensuring a stable and effective knowledge transfer across domains.
  2. Instead of generating pseudo-labels for the entire dataset in a single pass, the model produces pseudo-labels on a mini-batch basis and updates itself iteratively. This dynamic approach mitigates the accumulation of pseudo-labeling errors during adaptation, thereby enhancing stability and robustness throughout the gradual adaptation process

Weaknesses:

  1. The method has not been compared with the most recent state-of-the-art approaches.
  2. The proposed approach does not demonstrate a significant breakthrough compared to existing methods.
  3. The effectiveness of the method largely depends on the identification and availability of intermediate domains. In practice, intermediate data can sometimes be difficult to identify or incomplete, which limits its applicability.

问题

Please refer to weaknesses.

局限性

Yes.

最终评判理由

Although I still think the main solution of the paper using the parameter ϱ varying from 0 to 1 is quite simple, Gradual Domain Adaptation GDA has high practical applicability and the authors have provided relatively thorough and detailed responses to the various rebuttals. Therefore I am increasing my rating for this paper to 4.

格式问题

No.

评论

Author Response to Reviewer #nsMJ

Q1. The method has not been compared with the most recent state-of-the-art approaches.

A: We have compared 14 methods on six datasets, as detailed on pages 6-9 of the paper and in Appendix A. These methods include classic methods such as DANN and the latest methods such as GOAT (published in 2024). We believe that our experiments are sufficient and effective.

Q2. The proposed approach does not demonstrate a significant breakthrough compared to existing methods.

A: As shown in the experiments, our method demonstrates superior performance across all datasets, outperforming all compared methods. Even on RotatedMNIST, our method achieves a performance improvement of over 10% compared to the best method. This contradicts the reviewers' statement that “the proposed method does not show significant breakthroughs.” Our method is by no means a minor parameter optimization but a rigorously validated solution. Detailed proofs and methods can be found in the Methodology section and Appendix B.

Q3. The effectiveness of the method largely depends on the identification and availability of intermediate domains. In practice, intermediate data can sometimes be difficult to identify or incomplete, which limits its applicability.

A: Gradual or continuous distribution drift naturally exists in many real-world scenarios and is not an additional construct.In one of our experiments, the CIFAR-10-C dataset, CIFAR-10-C is created by applying 15 common image perturbations (such as Gaussian noise, glass blur, JPEG compression, pixelation, etc.) to the classic CIFAR-10 test set, and defining five progressively increasing severity levels for each perturbation.The original paper aimed to evaluate the robustness of models to covariate shifts, but this hierarchical structure naturally forms a gradient domain sequence: under the same corruption type, severity level 1 can be regarded as D0D_{0}, severity levels 2–4 correspond to D1D_{1} to D3D_{3}, and severity level 5 serves as the most challenging target domain D4D_{4}.Since images at all levels are identical in content and labels, with the only difference being the gradually increasing noise amplitude, they satisfy the assumption of “continuous, small-step” distribution drift and provide a clear, controllable transfer path.In practice, you only need to read the official numpy file and slice the samples according to the severity field to obtain five intermediate domains, without the need for additional labeling or complex preprocessing. If the model can achieve stable improvements on this sequence, it indicates that it can gradually adapt to scenarios with continuously degrading visual quality, which highly corresponds to the smooth performance degradation caused by real-world camera aging, compression rate changes, or network transmission distortion.More broadly, real-world scenarios such as visual inspection that slowly changes with lighting, season, or device aging; industrial sensor data that gradually drifts with production rhythms and machine wear; and transportation-travel patterns that exhibit consistent transitions by hour and workday-holiday; these timestamps themselves can be regarded as implicit “intermediate domains.” Therefore, the algorithm's assumptions align with a wide range of real-world tasks and are not artificially imposed ideal conditions.

评论

Although I still think the main solution of the paper using the parameter ϱ changing from 0 to 1 is rather simple, I find that Gradual Domain Adaptation GDA has high practical applicability such as gradually shifting domains across seasons. I have read all the authors' rebuttals and found them to be relatively thorough and detailed. Therefore I am increasing my rating of this paper to 4. I hope this work will serve as a foundation for further research on Gradual Domain Adaptation.

评论

Thank you sincerely for your thoughtful re-evaluation and for increasing the score to 4. We also value your comment that this work may serve as a foundation for future research; that is precisely our intention. In machine learning, the ReLU activation (max{0,x}) and residual connections (identity skip) are strikingly simple ideas that unlock deep model trainability and performance. Our STDW follows this tradition: a lean mechanism that aligns training dynamics with the structure of domain shift. Although our work isn't famous and good like them, We believe that, in science and engineering, truth is often simple yet beautiful. Thank you again for your attention in our work!

最终决定

This paper presented a novel Self-Training with Dynamic Weighting (STDW) approach for gradual domain adaptation. The method introduced a time-dependent weight parameter to guide the adaptation process in a simple yet effective manner. The proposed approach was empirically validated in extensive experiments compared to baseline methods.

Strengths:

  • The paper is well-written and easy to follow.
  • The proposed STDW method based on a time-dependent weight parameter is simple yet effective for gradual domain adaptation.
  • Extensive experiments demonstrate the effectiveness of STDW over baseline approaches.

Weaknesses:

  • The rationale behind the stepwise dynamic osmosis can be further clarified.
  • The applications of STDW in challenging scenarios with highly discontinuous or abrupt domain shifts can be explored and validated in the experiments.
  • The theoretical analysis of linear schedule can be analyzed.
  • The reproducibility of both STDW and the baseline implementations could be improved with clearer experimental details.

Following the rebuttal and discussion, all reviewers have positive opinions on the paper. While some concerns remain regarding theoretical depth and experimental coverage, the contributions are clear and the empirical results are strong. I recommend acceptance, and encourage the authors to address the remaining reviewer comments in the final version.