PaperHub
4.8
/10
Poster3 位审稿人
最低2最高3标准差0.5
3
3
2
ICML 2025

Multidimensional Adaptive Coefficient for Inference Trajectory Optimization in Flow and Diffusion

OpenReviewPDF
提交: 2025-01-23更新: 2025-08-12
TL;DR

We introduce MAC, a plug-in for flow and diffusion models that enables multidimensionality and adaptability in inference trajectories, improving generation quality and efficiency.

摘要

关键词
Multidimensional Adaptive CoefficientInference Trajectory OptimizationFlowDiffusionAdversarial Optimization

评审与讨论

审稿意见
3

This paper proposes to optimize model parameters and forward process interpolation coefficients with respect to a simulation-based adversarial loss. The training process consists of two stages. In the first stage, the flow model is trained w.r.t. randomly sampled multi-dimensional interpolation schemes via flow matching. To constrain the hypothesis space of interpolation schemes, interpolation coefficients are parameterized as a sum of weighted sinusoidals. In the second stage, the flow model and interpolation coefficients are optimized to minimize a simulation-based adversarial loss. Here, the weights for sinusoidals are parameterized as neural net outputs conditioned on xTx_T, providing an additional level of adaptivity. The authors demonstrate that the proposed method achieves competitive FID scores on CIFAR-10, ImageNet 32×3232 \times 32, FFHQ 64×6464 \times 64, and AFHQ 64×6464 \times 64.

给作者的问题

I am willing to raise the score to Weak Accept if the authors can address the following concerns.

  • [Q1] Can the authors provide any clarifications regarding the comment in Claims And Evidence?
  • [Q2] Can the authors provide additional experiments on ImageNet 64×6464 \times 64 as a demonstration of scalability?

论据与证据

  • (Left column, lines 260-262) "Given that Hθ(t,x(t))x0H_\theta(t, x(t)) \approx x_0, HθH_\theta can be adversarially refined by optimizing it with the discriminator DψD_\psi."

This claim holds only when generative paths are sufficiently straight. However, as cited below, the authors also claim that the benefit of using MAC arises from its ability to discover better non-linear trajectories.

(Left column, lines 300-305) "This suggests that a straight trajectory is not always optimal, even in OT-trained models, and MAC can adaptively discover better trajectories to correct errors that arise during transportation. Figure 3 further illustrates how MAC adjusts the trajectory direction to optimize transportation, resulting in a path that is not straight."

Hence, the authors' assumption Hθ(t,x(t))x0H_\theta(t, x(t)) \approx x_0 that justifies the usage of adversarial learning on velocity HθH_\theta output instead of generator Gθ,ϕG_{\theta,\phi} output seems to be at odds with the benefits of using MAC.

方法与评估标准

CIFAR-10, FFHQ-64, AFHQ-64 are standard datasets / FID and IS are standard metrics for evaluating the performance of generative models. However, based on previous works such as [1,2,3], I believe results on at least ImageNet 64×6464 \times 64 are necessary to demonstrate the practicality and scalability of the proposed method.

[1] Consistency Trajectory Models: Learning Probability Flow ODE Trajectory of Diffusion, ICLR, 2024

[2] Constant Acceleration Flow, NeurIPS, 2024

[3] Simple ReFlow: Improved Techniques for Fast Flow Models, ICLR, 2025

理论论述

There are no new theoretical claims.

实验设计与分析

I did not find any issues with experimental designs or analysis.

补充材料

I reviewed the supplementary material for experimental details.

与现有文献的关系

This paper proposes to adversarially optimize forward process interpolation coefficients for better generative modeling.

遗漏的重要参考文献

This paper is missing a discussion of [1,2,3]. [1] also optimizes the forward process to learn faster flows. [2,3] also incorporates adversarial learning into flow matching for improved generative modeling.

[1] Minimizing Trajectory Curvature of ODE-based Generative Models, ICML, 2023

[2] Constant Acceleration Flow, NeurIPS, 2024

[3] Simple ReFlow: Improved Techniques for Fast Flow Models, ICLR, 2025

其他优缺点

  • originality and significance : while the usage of adversarial learning to improve the perceptual quality of neural ODEs have been already introduced in works such as [1] or [2], the idea of optimizing forward process parameters via simulation-based training is novel and interesting. Still, I find the experimental setup rather limited, and as discussed in Methods And Evaluation Criteria, additional results on higher resolution or more complex datasets will significantly strengthen this submission.

[1] Consistency Trajectory Models: Learning Probability Flow ODE Trajectory of Diffusion, ICLR, 2024

[2] Constant Acceleration Flow, NeurIPS, 2024

其他意见或建议

  • Right column, lines 43-48 : x_est,θ,ϕx\_{est,\theta,\phi} is undefined.
  • Right column, Eq. (2) : x^_0,θ\hat{x}\_{0,\theta} and x^_1,θ\hat{x}\_{1,\theta} are undefined.
  • Table 4 : please clarify the difference between zρTz \sim \rho_T and xTρTx_T \sim \rho_T.
作者回复

URL for Additional Figures: https://imgur.com/a/3UiYDVF

[A1]**[A1] H_\thetaestimatesthevectorfieldandestimates the vector field and\gamma_\phiperformsinferencetimeplanning performs inference-time planning**

We clarify the misunderstanding here. As shown in Additional Figure 2, the roles of θ\theta and ϕ\phi differ clearly. HθH_{\theta} estimates the vector field by predicting the endpoint x0x_0 regardless of trajectory curvature. Adaptive coefficients (γϕ\gamma_{\phi}) determine nonlinear trajectory planning by adjusting velocity directions and step sizes based on predictions from HθH_{\theta}. Thus, the capability for nonlinear trajectory planning depends entirely on γϕ\gamma_{\phi}, which adaptively guides the inference until endpoints are reached.

For clarity, we added Additional Figure 1 to illustrate MAC enabling distinct, adaptively curved trajectories for inference planning. Our core contribution, MAC, provides inference-time planning, meaning optimal inference plans are computed and optimized offline via simulation and then directly deployed during inference without additional computations.

[A2]ScalabilityExperimentsonImageNet**[A2] Scalability Experiments on ImageNet**

We agree with the reviewer that experiments on ImageNet are essential for demonstrating scalability and practical value. However, due to substantial resource requirements (32 A100 GPUs for 2 weeks needed for pre-training EDMγEDM_\gamma on ImageNet-64), conducting large-scale experiments within the rebuttal period is currently impractical.

Nevertheless, we emphasize our method’s scalability, supported by the following points:

As in current SoTA methods on ImageNet-256 ([5], [6], [7], [8]), diffusion models’ scalability primarily relies on latent diffusion paradigms [4]. Our ongoing research explicitly integrates MAC with LDM, discrete diffusion frameworks [7], and IMM [8]. These frameworks substantially reduce computational demands, making ImageNet-scale experiments feasible by leveraging available pre-trained models.

Specifically, given that discrete diffusion frameworks naturally use highly multidimensional interpolated values for pre-training, optimizing MAC on discrete diffusion frameworks enables large-scale experiments using existing pre-trained models. Additionally, combining MAC with advanced models like IMM [8] can achieve high performance on large-scale datasets by using existing pre-trained models (explained in our response to Reviewer TFVT [A3]).

MAC can be readily integrated with these frameworks, and we are currently preparing dedicated large-scale follow-up studies.

For revision, we propose including experiments on ImageNet-64 using our existing setup with available EDMαEDM_\alpha models. We believe this sufficiently demonstrates our method’s scalability within the scope of the original submission.

[A3]EssentialReferencesDiscussion**[A3] Essential References Discussion**

Thank you for identifying the missing literature ([1], [2], [3]). Briefly, the major conceptual difference is that prior works ([1], [2], [3]) optimize vector fields by predefined trajectory properties (e.g., straightness or minimal curvature) guided by optimal transport theory to reduce numerical errors. In contrast, our method does not enforce predefined inference trajectory optimality; instead, we adaptively discover optimal nonlinear trajectories by final transportation quality through simulation-based inference-time planning. This allows greater flexibility and potentially superior performance.

[A4]NotationClarifications**[A4] Notation Clarifications**

Gθ,ϕ(τ,xT)=xest,θ,ϕG_{\theta, \phi}(\tau, x_T) = x_{\text{est}, \theta, \phi} (Details in Appendix C).

x^\hat{x} values are predictions from Hθ(t,x(t))H_{\theta}(t, x(t)) for terminal points x0ρ0x_0 \sim \rho_0 and x1ρ1x_1 \sim \rho_1.

xTx_T denotes the actual initial sample point, while zz is sampled from the same distribution but unused during inference.

We sincerely appreciate your thoughtful consideration and hope these clarifications effectively address your concerns.

[1] Minimizing Trajectory Curvature of ODE-based Generative Models, ICML, 2023

[2] Constant Acceleration Flow, NeurIPS, 2024

[3] Simple ReFlow: Improved Techniques for Fast Flow Models, ICLR, 2025

[4] High-resolution Image Synthesis with Latent Diffusion Models, CVPR, 2021

[5] Scalable Diffusion Models with Transformers, ICCV, 2023

[6] SiT: Exploring Flow and Diffusion-based Generative Models with Scalable Interpolant Transformers, ECCV, 2024

[7] [Mask] is All You Need, arXiv, 2025

[8] Inductive Moment Matching, arXiv, 2025

审稿人评论

Thank you for the reply. Given the authors rebuttal and additional experiments on ImageNet64, I have raised the score from Weak Reject to Weak Accept.

审稿意见
3

This paper introduces Multidimensional Adaptive Coefficients (MAC) for flow and diffusion models, allowing coefficients to vary across dimensions and adapt to different starting points. The two-stage approach combines pre-training with multidimensional coefficients and adversarial refinement. Experiments across multiple frameworks and datasets show improved performance, including state-of-the-art results on CIFAR-10 conditional generation.

update after rebuttal

The author resolved my confusion. However, considering that reviewer MR8r's review was not addressed, I'm unsure what happened. Finally, considering the contribution of the current work, my current rating remains unchanged.

给作者的问题

When considering the total computation (pre-training + adversarial finetuning), how does your method compare to these alternatives?

论据与证据

The performance improvement claims are generally supported by comprehensive experiments across multiple frameworks and datasets.

However, the "training efficiency" claim is problematic since the method requires training models from scratch rather than fine-tuning existing pre-trained models, which is a significant practical limitation.

方法与评估标准

The evaluation metrics and benchmark datasets are appropriate for the task.

However, there is a concern with the methodology. The hypothesis space design for MAC introduces many additional hyperparameters (M, s, q, LPF configuration). The paper lacks analysis of how baseline methods might perform if augmented with similar parameterization advantages. For example, it would be valuable to see how CTM would perform if pre-trained with comparable additional parameters before distillation.

理论论述

There is no theoretical claims in this paper.

实验设计与分析

The experimental design is overall sound, with comprehensive testing across multiple frameworks and datasets.

The ablation study in Table 6 shows that adversarial training provides most of the performance gain. Since adversarial training is not a novel contribution of this paper, this suggests that MAC's specific contribution to performance improvement is relatively limited.

补充材料

I reviewed all the supplementary materials.

与现有文献的关系

The paper positions itself as novel by optimizing trajectory quality through simulation, but this approach is conceptually similar to:

  1. Kim et al. (2024) (CTM) which also uses adversarial refinement on diffusion trajectories
  2. Lu et al. (2024) (CD + GAN) which likewise employs adversarial refinement

The primary difference, multidimensional coefficients, appears to provide only incremental benefits over standard adversarial refinement of model parameters.

遗漏的重要参考文献

I find that the existing citations sufficiently cover the related literature, though the novelty of the proposed approach remains limited.

其他优缺点

Strengths:

  1. Comprehensive experiments across multiple frameworks demonstrate the method's versatility

  2. Good visualizations of learned trajectories help explain the method's operation

  3. The approach achieves state-of-the-art results on CIFAR-10 conditional generation

Weaknesses:

  1. The requirement to retrain models from scratch makes the method impractical for large-scale applications

  2. The performance gains attributable specifically to MAC (versus general adversarial refinement) appear modest

  3. The method introduces significant additional complexity through MAC hyperparameters

其他意见或建议

The paper would benefit from providing a more detailed analysis of computational overhead. Additionally, testing on higher-resolution images would help demonstrate the method's scalability to more complex generation tasks.

作者回复

URL for Additional Figures: https://imgur.com/a/3UiYDVF

[A1]MACscorevalueliesininferencetimeplanning,orthogonaltovectorfieldtuningmethodslikeCTMandCD+GAN**[A1] MAC’s core value lies in inference-time planning, orthogonal to vector field tuning methods like CTM and CD+GAN**

CTM and CD+GAN optimize trajectories by adjusting the vector field parameter θ\theta. In contrast, our method (MAC) performs inference-time planning, meaning the optimal inference plans are computed and optimized offline via simulation and then directly deployed during inference without additional computations.

As shown in Additional Figures 1 and 2, without MAC, trajectory planning’s search scope is limited to linear paths, restricting performance gains. MAC enables adaptive, nonlinear, dimension-wise trajectory planning, significantly expanding optimization flexibility.

CTM and CD+GAN do not employ actual inference-time simulation feedback. Our method dynamically optimizes trajectories and timestep plans using simulations identical to inference, making it uniquely dynamic.

MAC acts as a final-stage inference planning strategy, enhancing performance beyond what is achievable by optimizing the vector field alone. This means that MAC can be integrated with existing frameworks such as CTM and other diffusion or distillation methods, as their methodologies attempt to solve different problems.

[A2]MACReducesHyperparameterEngineeringCosts**[A2] MAC Reduces Hyperparameter Engineering Costs**

The modest gains observed with EDM result from EDM’s already highly optimized configuration, leaving limited margin for further improvement. (This is connected to the reason why we use EDMγEDM_\gamma and coefficient labeling, explained below in [A3].) However, Sections 4.1 and 4.2 demonstrate significant improvements (~10x NFE efficiency) in less optimized frameworks (DDPM, FM, SI) by using MAC.

MAC alleviates the extensive manual tuning burden associated with framework-specific hyperparameters, like in EDM. Although MAC requires tuning of hypothesis space parameters (M,s,qM, s, q, LPF configs), these parameters are framework-independent. Hence, once identified, these parameters generalize well across various flow and diffusion models, providing considerable performance benefits without framework-specific engineering, as shown in our experiments.

Regarding baseline comparison with similar parameterization: As demonstrated in Table 6, directly incorporating pre-training parameterization into EDM (EDMγEDM_\gamma) negatively impacts performance. Using this as a teacher network will degrade CTM performance unless combined with MAC for inference-time planning. MAC provides clear advantages specifically when used for adaptive inference-time planning, not as a general pre-training strategy.

[A3]PracticalityviaExistingPretrainedModels**[A3] Practicality via Existing Pre-trained Models**

Our method indeed requires higher computational costs compared to CTM (approximately 201 Mimg vs. 25 Mimg) when using the specific EDMγEDM_\gamma pre-training setup, chosen deliberately to push the limits and achieve SoTA performance on CIFAR-10. However, we emphasize that using EDMγEDM_\gamma and coefficient labeling is optional. MAC can practically and effectively leverage existing pre-trained models (such as EDMαEDM_\alpha), significantly reducing computational overhead without substantially compromising performance.

Using γ\gamma increases the probability of multidimensional interpolated values x(t)x(t), but even using α\alpha inherently yields multidimensional interpolations, as Gaussian noise is independently sampled across dimensions. Specifically, given x1ρ1=N(0,Id)Rdx_1 \sim \rho_1 = \mathcal{N}(0, I_d) \in \mathbb{R}^d, each dimension x1,ix_{1,i} is independently drawn from N(0,1)\mathcal{N}(0,1), and thus even a linear interpolation x(t)=tx1+(1t)x0x(t) = t \cdot x_1 + (1-t) \cdot x_0 naturally results in dimension-wise distinct noise contributions.

Empirical evidence from our experiments supports this claim. Section 4.1 demonstrates effective nonlinear trajectory planning in OTSIMACOT-SI_{MAC} using models pre-trained with α\alpha. Additionally, Figure 6 shows the trained ϕ\phi‘s distribution significantly differs from the pre-training distribution, indicating MAC’s effectiveness is not critically dependent on pre-training specifics.

Thus, for large-scale practical scenarios where training from scratch is costly, MAC can effectively utilize existing models (e.g., EDMαEDM_\alpha), ensuring practicality and computational efficiency.

[A4]Scalability(ImageNet)**[A4] Scalability (ImageNet)**

Please refer to our response to Reviewer jRK3’s comment [A2].

We sincerely appreciate your thoughtful consideration and hope these clarifications effectively address your concerns.

审稿人评论

Thanks for the authors' responses. My final concern lies in complete experimental comparisons, just like all the other reviewers, i.e., experiments on ImageNet64. While the authors note that extensive GPU resources would be needed for these experiments, there are strategies available to reduce computational demands, such as gradient accumulation. Although a complete comparison would be time-intensive, the preliminary results produced by a not fully converged model, could provide meaningful insights into the method's scalability. Given this, I maintain my current score.

作者评论

[A]ImageNet64ResultDemonstratingScalabilityandPracticalityUsingOnlyExistingPretrainedModels**[A] ImageNet-64 Result Demonstrating Scalability and Practicality Using Only Existing Pre-trained Models**

We would like to inform the reviewers that we have conducted an additional experiment on ImageNet-64, as requested.

We achieved an FID of 1.47**1.47** with 5(+) NFE by applying inference trajectory optimization usinganexistingpretrained**using an existing pre-trained** EDMαEDM_\alpha model, with only 30k**30k** training images for MAC. This result outperforms CTM with NFE = 2 (FID = 1.73), which requires 61.4M**61.4M** training images. Except for the batch size, which was set to 32 for γϕ\gamma_\phi, and 512 for DψD_\psi, all other configurations—including the model size for wϕw_\phi—were kept identical to those used in the CIFAR-10, FFHQ, and AFHQ experiments.

This demonstrates a substantial gain in training efficiency and scalability—achieving better performance with 2048×**2048**\times fewer training samples than CTM. Furthermore, this result empirically supports our earlier rebuttal point [A3] to Reviewer TFVT, as it was obtained withoutanypretrainingstage**without any pre-training stage**, directly proceeding to the adversarial trajectory optimization stage using an existing pre-trained model without any modification.

We note that further training may improve FID scores, but we report the result as early as possible within the rebuttal period to provide a timely response. Additional ablation studies on ImageNet-64 will be included in the revision.

We hope this new result effectively addresses the concerns regarding the practicality and scalability of our method.

审稿意见
2

This paper introduces a new way to handle the interpolation between noise and data in diffusion and flow-based generative models. Unlike standard approaches that use an interpolation with a uniform scale across the entire image (like Rectified Flows, DDPM, and IDDPM), the authors propose extending this interpolation to a 2D space, matching the dimensions of the image and noise. This 2D interpolation is learned using a coefficient model, offering potentially more flexibility in the interpolation process. The training procedure involves three main components: a standard diffusion model trained with a 2D sampling of the time variable, a discriminator, and the coefficient model, all optimized jointly in the second stage to find better interpolation weights.

update after rebuttal

As I didn't receive any rebuttal from the author I'll remain my current rating unchanged.

After reviewing the author's rebuttal with other reviewers, I believe most of my concerns remain unaddressed, and the paper, especially the experiment section, still appears incomplete and not yet ready from my perspective.

给作者的问题

  1. In the introduction, you mention a connection to Neural ODEs. Could you please elaborate on the nature of this relationship and how your work builds upon or relates to concepts from Neural ODEs?
  2. Why was the specific parameterization for the coefficient model chosen as described in equation 8? Were other parameterizations explored, and if so, what were the results? An ablation study on this design choice would be beneficial.
  3. The experimental evaluation is currently limited to CIFAR-10. Do you have plans to extend the evaluation to higher-resolution datasets like Imagenet 64x64 or include more comprehensive comparisons on datasets like FFHQ or AFHQ? Given the 2D time sampling, are there potential scalability challenges with increasing image size?
  4. In equation 5, x_T is used but not defined. Similarly, in Algorithm 1, ρ_T is used without definition. What are these variables, and what is the difference between ρ_1 and ρ_T?
  5. From Section 3.2.2 it seems that the method is currently applied to diffusion models predicting noise. Could you clarify if this approach can be extended to other types of diffusion and flow models, such as those predicting velocities?
  6. Tables 2, 3, and 4 present results with NFE=10, which is unusually low for standard DDPM and Flow Matching methods. The performance also differs from Table 5. Were different models or training settings used in these tables? Please clarify these discrepancies.
  7. Table 5 is missing many baseline results, including 1-Rectified Flow (Flow Matching). Could you provide these missing results for a more comprehensive comparison?
  8. What is the primary motivation for using adversarial training in your method? Could you compare the performance of your method with and without the adversarial loss to isolate the impact of the proposed interpolation technique? It's unclear whether the performance gains are solely due to the proposed interpolation method or if the adversarial training plays a significant role, especially in terms of FID improvement.
  9. In Table 6, the FID scores for the first two rows (presumably vanilla EDM) are significantly higher than the EDM results in Table 5. Could you explain this difference? I assume the first row is vanilla EDM so it should match table 5 row 3 in diffusion models?

论据与证据

The central claim is that the proposed 2D learned interpolation offers more flexibility and can lead to improved generative models. The paper presents experimental results on CIFAR-10 to support this claim. However, the evidence might be considered somewhat limited due to the restriction to a single, relatively low-resolution dataset. The lack of thorough ablation studies on the design choices of the coefficient parameterization also weakens the evidence for the specific design being optimal.

方法与评估标准

The proposed method, involving the joint training of a diffusion/flow model, a discriminator, and a coefficient model with a 2D time sampling, presents an interesting alternative to standard interpolation techniques. The evaluation seems to primarily rely on FID scores obtained on the CIFAR-10 dataset. While FID is a common metric for generative model evaluation, the limited scope of the experiments raises questions about the generalizability of the findings.

理论论述

This paper doesn't include new proofs or theoretical claims

实验设计与分析

The experimental design and analyses raise several concerns:

  • Limited Dataset: The evaluation is limited to CIFAR-10 (32x32 images), which is insufficient to demonstrate the effectiveness of the proposed method on more complex and higher-resolution datasets. I suspect that the increased dimensionality of the 2D time sampling might lead to optimization difficulties as the image size grows.
  • Lack of Ablation: The absence of ablation studies on the design choices for the coefficient parameterization is a significant oversight. It's quite hand-crafted in terms of the specific parameterization in equation 8, and whether other designs might yield better results.
  • Inconsistent Results: There are discrepancies in the reported performance across different tables (e.g., Table 2, 3, and 4 compared to Table 5), and it's unclear if different models or settings were used. The comparison in Tables 2-4 is limited to NFE=10, which is unusually low for standard diffusion and flow models to achieve good performance.
  • Missing Baselines: Table 5 lacks many baseline results, making it difficult to properly assess the improvement offered by the proposed method. Furthermore, the comparison only includes distillation methods and inexplicably omits 1-Rectified Flow (Flow Matching), which seems like a relevant baseline.

补充材料

Yes the supplementary mainly includes the experiment details and visualizations.

与现有文献的关系

Previous work like rectified flow (flow matching), DDPM or other variants mainly focus on using a predefined and fixed intepolation. This work extend it using a learned intepolation and expand it to 2D

遗漏的重要参考文献

The paper include the related references.

其他优缺点

Strengths: The paper proposes a novel idea of using a learned 2D interpolation for diffusion and flow models, which could potentially offer more flexibility than standard interpolation.

Weaknesses: Please see the questions below

其他意见或建议

Please see the questions part.

最终决定

Standard diffusion and flow-based models interpolate between noise and data, and the interpolating coefficients are usually shared across dimensions. This paper proposes a simple method for having dimension-dependent interpolation coefficients.

The paper received three reviews, yet one was flagged as being AI-generated. Upon inspecting the review, I believe the flag to be very plausible since the review explicitly mentioned the authors conducting experiments only on CIFAR-10 despite this not being the case. I thus ignored this review in my decision making process, and instead read the paper and engaged with the authors.

Reviewers found the method novel and sensible, but criticized the lack of experiments on ImageNet. The authors provided these experiments in their rebuttal in a way that I believe satisfactorily addressed the main criticisms raised. Overall I believe this paper should be accepted: it presents a simple yet well-motivated and novel method which achieves good empirical results.