UniDB: A Unified Diffusion Bridge Framework via Stochastic Optimal Control
We present UniDB, a unified diffusion bridge framework using stochastic optimal control, significantly improving detail preservation and image quality in generative tasks with minimal code modifications.
摘要
评审与讨论
This paper proposes a unified framework, UniDB, of diffusion bridge models based on Stochastic Optimal Control. This framework enhances the quality and detail of generated images by balancing control cost and terminal penalty.
给作者的问题
See the above comments.
论据与证据
Claim 1: UniDB helps to understand and generalize Doob’s -transform. The theoretical derivations in the paper demonstrate that Doob’s -transform is a special case of UniDB when the penalty coefficient approaches infinity. This provides a solid mathematical foundation for the claim.
Claim 2: UniDB improves image quality by allowing the design of different controllers . The claim is supported by a lot of qualitative and quantified experimental results, demonstrating that controller design enhances image quality.
方法与评估标准
The use of PSNR, SSIM, LPIPS, and FID as evaluation metrics in experiments on three high-resolution datasets (CelebA-HQ, Rain100H, and DIV2K) is reasonable. These metrics comprehensively assess image quality, covering aspects of pixel-wise similarity (PSNR, SSIM), perceptual quality (LPIPS), and generative diversity (FID). The chosen datasets are also well-suited for high-resolution image generation and restoration tasks, ensuring a robust evaluation.
理论论述
The mathematical derivations in the paper are complete and generally reliable. However, two aspects require further clarification from the authors:
-
Choice of norm in the training objective (Equation 19): The paper does not provide a clear justification for using the norm instead of other alternatives like norm. The authors should explain whether this choice is based on empirical performance, theoretical considerations, or robustness to outliers.
-
Introduction of the state vector term in linear SDE: One of the paper's key novelties is introducing the term in the linear SDE form, but it is not explicitly explained how it is computed or designed in the main context. Further elaboration on the motivation, computation, and impact of would improve the clarity of the contribution.
实验设计与分析
The experiments are comprehensive and well-designed, covering three different tasks to ensure robustness and generalizability. Additionally, the paper conducts an ablation study on the key penalty coefficient , which helps evaluate its impact on model performance.
However, I suggest that the authors include DDBM as a benchmark for comparison. Since DDBM is also a Doob’s -transform-based model and is mentioned in the Preliminaries, it would be beneficial to compare the proposed method against it. This would provide a clearer assessment of the advantages and potential improvements offered by the proposed framework.
补充材料
The supplementary material is complete and provides sufficient additional details to support the main paper.
与现有文献的关系
-
UniDB as a Generalization of Doob’s h-Transform: Doob’s h-transform has been widely studied in stochastic processes and has been applied in bridge modeling [1][2]. The paper demonstrates that Doob’s -transform is a special case of the proposed UniDB framework when the penalty coefficient tends to infinity, providing a broader theoretical foundation for diffusion bridge models.
-
Controller Design for Improved Image Generation: Stochastic Optimal Control has been explored for diffusion bridge models [3], but existing approaches often lead to artifacts such as blurred or distorted details. By allowing the design of different controllers , UniDB provides greater control over generation quality, leading to improved image fidelity and diversity across multiple datasets.
[1] Zhou, Linqi, et al. "Denoising diffusion bridge models.", 2023.
[2] Yue, C., et al. "Image restoration through generalized ornstein-uhlenbeck bridge", 2024.
[3] Park, B., et al. Stochastic optimal control for diffusion bridges in function spaces, 2024.
遗漏的重要参考文献
The key contribution of this paper is the introduction of Stochastic Optimal Control into the DDBM theoretical framework. However, a similar approach was explored last year in:
[1] Zhang, Shaorong, et al. "Exploring the Design Space of Diffusion Bridge Models via Stochasticity Control." arXiv preprint arXiv:2410.21553 (2024).
To differentiate from prior work, the paper should explicitly highlight the key distinctions between this work and Zhang et al. (2024), particularly in how Stochastic Optimal Control is formulated and applied.
其他优缺点
The paper is well-structured and clearly organized, making it easy for readers to follow. However, the novelty is sceptical, as a similar approach was explored last year.
其他意见或建议
See the above comments.
Thank you for your feedbacks and comments.
Claim 1: Choice of norm in the training objective (Equation 19): The paper does not provide a clear justification for using the norm instead of other alternatives like norm. The authors should explain whether this choice is based on empirical performance, theoretical considerations, or robustness to outliers.
Thanks for your feedback. Using norm is a common trick in the field of image restoration with many prior works [G1-G4]. We just follow to use the same trick as we mentioned in line 270 of our paper. The reason that choosing norm over other alternatives like norm in image restoration tasks have been analyzed in [G2-G4]. In [G2], "when the network is trained using as a cost function, instead of the traditional , the average quality of the output images is superior for all the quality metrics considered." [G3, G4] establish the loss's superiority over through improved convergence and outlier robustness.
[G1] Yue et al. "Image restoration through generalized ornstein-uhlenbeck bridge", ICML 2024.
[G2] Zhao et al. "Loss Functions for Image Restoration with Neural Networks", IEEE Transactions on Computational Imaging 2017.
[G3] Lim et al. "Enhanced Deep Residual Networks for Single Image Super-Resolution", CVPR 2017 workshop.
[G4] Mu et al. "Riemannian Loss for Image Restoration", CVPR 2019 workshop.
Claim 2: Introduction of the state vector term in linear SDE: One of the paper's key novelties is introducing the term in the linear SDE form, but it is not explicitly explained how it is computed or designed in the main context. Further elaboration on the motivation, computation, and impact of would improve the clarity of the contribution.
We want to clarify that we've never claimed the introduction of the state vector term in linear SDE is the key novelty of our framework. Instead, our main novelty lies in constructing the diffusion bridge in the form of stochastic optimal control and revealing that Doob's h-transform is a special case of ours. The introduction is just a simple reformulation referenced by prior works IR-SDE [G5] and GOUB [G1]: we reformulated the drift term into . Particularly in our experiments of UniDB-GOU, for a fair comparison, we set which is identical to GOUB [G1], ensuring consistency with baselines.
[G5] Luo et al. "Image Restoration with Mean-Reverting Stochastic Differential Equations", ICML 2023.
Experimental Designs Or Analyses: The reviewer suggests that the authors should include DDBMs as a benchmark for comparison.
Actually, we have compared DDBMs as a benchmark in Appendix E "Additional Experimental Results" of our paper, theoretically analyzing the application of UniDB to DDBMs (Appendix: A.8. Examples of UniDB-VE and UniDB-VP) and conducting extensive experiments (Tables 3, 4, and 5). Specifically, it outperforms DDBMs on both LPIPS and FID metrics, with gains reaching up to ~20% in some cases.
Essential References Not Discussed: a similar approach was explored last year in: Zhang, Shaorong, et al. "Exploring the Design Space of Diffusion Bridge Models via Stochasticity Control". To differentiate from prior work, the paper should explicitly highlight the key distinctions between this work and Zhang et al. (2024), particularly in how Stochastic Optimal Control is formulated and applied.
Our work takes a completely different approach from the prior work Zhang et al. (2024) (following simply denoted as SDB), specifically:
- Different purposes. SDB mainly focuses on solving singularities and accelerations for training and sampling caused by the neglection of the impact of noise in sampling SDEs. In contrast, our UniDB aims to address the issues resulting from Doob's h-transform (e.g. artifacts along edges and unnatural patterns) that occur in the existing diffusion bridge models.
- Different methods. Although both that paper and our work mention "Stochastic Control" in titles, the two "Stochastic Control" are totally different. SDB is more inclined towards stochastic process, adding noise into the base distribution and stochasticity to the reverse process which is still based on Doob's h-transform. While our UniDB focuses on stochastic optimal control, modeling the forward process as an optimization problem to analyze the drawbacks of Doob's h-transform.
Therefore, the two articles are fundamentally different in nature.
Thanks for the author's clarification on my confusion. I've raised up my rating.
This paper introduces UniDB, a diffusion bridge model framework that utilizes Stochastic Optimal Control (SOC) for process optimization, providing an analytical solution for the optimal controller. UniDB generalizes existing diffusion bridge models by showing that Doob’s h-transform is a special case where the penalty coefficient γ tends to infinity. By adjusting the trade-off between control costs and terminal penalties, UniDB improves image detail and quality while maintaining compatibility with existing models. Experimental results demonstrate its effectiveness in image restoration tasks.
update after rebuttal
I have carefully read the rebuttal and would like to maintain my original score.
给作者的问题
Please refer to the questions and comments provided above.
论据与证据
UniDB claims to generate higher-quality images than Doob’s h-transform and supports this claim with experimental evidence on tasks such as super-resolution, inpainting, and deraining. However, additional explanation is needed regarding whether the optimal controller in LQ SOC directly contributes to producing sharper and more detailed images. Furthermore, through Proposition 4.3, it is shown from an LQ SOC perspective that the optimal controller obtained with a finite γ is preferable to that of the infinite case. However, it remains unclear whether there is a systematic analysis regarding the choice of γ. In Proposition 4.5 and Figure 2, a sufficiently large γ is selected to minimize differences in terminal point positions. Nevertheless, based on the results in Table 2, it appears that using an optimal controller with finite γ does not always lead to improvements in actual evaluation metrics. I would appreciate further clarification on this point.
方法与评估标准
The authors demonstrate their approach’s strength through evaluation criteria widely used in prior works
理论论述
The detailed proofs of the Theorems and Propositions in the main paper are clearly described in the Appendix. However, I noticed some minor issues when cross-referencing the statements with their proofs.
-
There are minor typos regarding the connection between the statements and the proofs. The authors are encouraged to check the appendix number references carefully.
-
In equation (53) of Appendix A.3, could the authors provide more details on the derivation of this part? The intermediate steps would help improve clarity.
实验设计与分析
This paper follows a standard experimental design and evaluation process, so I believe there are no issues in this regard.
补充材料
I examined the validity of each section in the Appendix as well as their connections to the main paper, and the related concerns have been raised above.
与现有文献的关系
No.
遗漏的重要参考文献
No.
其他优缺点
Please refer to the questions and comments provided above.
其他意见或建议
Please refer to the questions and comments provided above.
We greatly appreciate your feedback and inquiries.
Claims And Evidence 1: Additional explanation is needed regarding whether the optimal controller in LQ SOC directly contributes to producing sharper and more detailed images.
The over-control in Doob's h-transform violates the natural statistical properties of images, prioritizing mathematical precision (pixel-perfect endpoints) over visual authenticity (realistic SDE trajectories). UniDB leads to better overall performance considering both realistic SDE trajectories and target endpoint matching. A comprehensive analysis can be found in response for Reviewer 33jh's Claim and response for Reviewer AzN7's Claim and Question 2.
Claims And Evidence 2: It remains unclear whether there is a systematic analysis regarding the choice of . Based on the results in Table 2, it appears that using an optimal controller with finite does not always lead to improvements in actual evaluation metrics.
In Figure 2 of our paper, we illustrate that a range of to for is pretty good. What we want to emphasize is not that there is a single best value of for each dataset, but rather that most value of chosen within this interval can yield a better result than Doob's h-transform (). While suboptimal values may occasionally degrade performance, our approach provides a simple, efficient, and nearly cost-free way to improve model performance, which works well in most cases.
It is worth noting that when is set to , our model outperforms all baselines across multiple tasks, including super-resolution (on DIV2K, CelebA, and FFHQ), deraining on (Rain100H), and inpainting on (CelebA).
Question: Could the authors provide more details on the derivation of this part: ?
Yes, below is the detailed proof: At line 770 of our paper, we defined for simple denotation. The first equation of Eq. 51 is a simple simplification by substituting the expression from Eq. 50 into Eq. 51. A more detailed proof is as follows:
Then, just mutiplying on both sides of the equation, we can learn that . We will add the detailed proof in the revised version.
Suggestion: There are minor typos regarding the connection between the statements and the proofs.
Yes, you are right. Thanks for pointing out them and we will correct them in the revised version.
This paper proposed a diffusion-based method for image restoration problems, e.g., super-resolution, deraining, and inpainting. Given a dataset of corrupted and clean image pairs, the goal is to construct a diffusion model that at inference generates clean images given corrupted images. The proposed method is based on stochastic optimal control (SOC), which re-frames the problem as an optimization over the drift of diffusion models. Such an SOC reformulation introduces a tunable hyper-parameter () and recovers prior Doob-h-transform's based diffusion models as approaches infinity. It is shown empirically that a rather large but finite (~e7) improves performance.
给作者的问题
-
Eq 8 and 16 need more explanation. Why is it okay to drop the Brownian motion? This seems more like an empirical trick to get better PSNR and SSIM.
-
There seems to be some typos in Eq 15,16 \nabla p --> \log \nabla p
论据与证据
I'm not convinced by the implication of Prop 4.3 (the paragraph below Prop 4.3, before Sec 4.3). The SOC objective is (artificially) constructed as a surrogate for searching the drift of SDE. That is smaller for finite does not imply suboptimality in empirical performance. Intuitively, it does make sense to set to infinity since we'd like xT^u to converge exactly to the given xT. Any finite would fail to achieve that, as shown in Prop 4.5. What's presented in this paper is somewhat counter-intuitive but also interesting, yet it's rather empirically observed than theoretically justified.
方法与评估标准
Yes
理论论述
Thm 4.1 is expected but still quite nice to see the actual analytic form! Though, I did not check proof carefully.
实验设计与分析
Comparison to baselines is insufficient. GOUB is a special case of UniDB (these values re-appear in Table 2 last column) and DDRM is not quite comparable as a non-learning-based method and requiring knowing additionally corruption type. The authors should compare their method to I2SB and its follow-up works (e.g., CDDB https://arxiv.org/abs/2305.19809). These works are also SOC-inspired method for solving image restoration.
补充材料
Yes
与现有文献的关系
N/A
遗漏的重要参考文献
N/A
其他优缺点
N/A
其他意见或建议
- What are x0 and xT in practice? To my best guess, x0 would be the clean image and xT would be the corrupted image. So in Eq 12 we learn a forward process that bring x0 close to xT for every (x0, xT) pair, and then reverse it with Eq 15. Is this correct? It'd be better to have clarification in Sec 3.
We sincerely thank you for your comments and questions. We will provide a detailed response to these concerns.
Claim: The reviewer is not convinced by the implication of Prop 4.3: is smaller for finite does not imply suboptimality in empirical performance. Intuitively, it does make sense to set to infinity since we'd like to converge exactly to the given .
Thank you for raising this important point. We agree that the intuition behind Proposition 4.3 deserves clarification. Below, we address why strict terminal matching () is not ideal for practical performance, even if it seems mathematically appealing.
1. Exact terminal matching () harms image quality.
While Doob’s h-transform achieves , it requires large control () to force exact endpoint matching. The large in SDE trajectory()may disrupt the inherent continuity and smoothness of images. Prioritizing pixel-perfect endpoints over smooth trajectories leads to "mathematically correct but visually unrealistic" outputs. Our experiments (Figure 1) confirm that Doob's h-transform can lead to artifacts along edges and unnatural patterns in smooth regions.
2. Why finite works better?
By keeping finite, UniDB explicitly balances two goals: target matching (terminal penalty) and trajectory smoothness (controller). Proposition 4.3 shows that finite achieves a lower total cost not by sacrificing performance, but by optimally trading minor terminal mismatches for significantly smoother, more natural diffusion paths.
Thus, Proposition 4.3 reflects a key insight: real-world image generation benefits more from stable trajectories than rigid mathematical constraints. We will clarify this intuition in the revised manuscript.
Experiment: Comparison to baselines is insufficient: the authors should compare their method to I2SB and its follow-up works (e.g. CDDB).
Thank you for your feedback. Below, we clarify why our current comparisons are sufficient and address your concerns about I2SB and CDDB:
-
Direct Comparison to State-of-the-Art (DDBM > I2SB):
- DDBM (a SOTA baseline) is strictly superior to I2SB (see Table 2 in [DDBM paper], where DDBM achieves FID 4.43 vs. I2SB’s 9.34 on DIODE-256×256, and 1.83 vs. I2SB’s 7.43 on Edges→Handbags-64×64).
- UniDB outperforms DDBM across multiple tasks (Tables 3–5 in Appendix E): Super-resolution (DIV2K, CelebA-HQ), Deraining (Rain100H).
- Critically, DDBM is a special case of UniDB (UniDB with ), meaning our method inherently subsumes and improves upon this stronger baseline.
-
I2SB's Computational Burden.
- Training I2SB is prohibitively expensive. It requires 16×V100 GPUs with more than 1 week (per communication with I2SB authors).
- The inference time of I2SB is 90 seconds per 256×256 image (vs. our method’s less than 5 seconds).
- Given hardware/time constraints, reproducing I2SB (and its follow-ups) for fair comparison is practically infeasible during rebuttal. Since DDBM already outperforms I2SB, and UniDB outperforms DDBM, we argue that direct I2SB comparisons are redundant.
-
CDDB is Orthogonal to Our Contribution.
- CDDB is a training-free, plug-and-play refinement module for I2SB, not a standalone method.
- Such post-hoc techniques could also integrate with UniDB (e.g., applied during inference), but this is beyond our paper’s scope, which focuses on developing a unified training framework for diffusion bridge.
We appreciate your suggestion and will explicitly discuss the relationship between UniDB, DDBM, I2SB, CDDB in the revised manuscript.
Comment: What are x0 and xT in practice? x0 would be the clean image and xT would be the corrupted image, is this correct?
Yes, you're right. It's standard in diffusion-based image restoration research using x0 as the clean image and xT as the corrupted image. We'll make it clearer in the revised version.
Q1: Eq 8 and 16 need more explanation. Why is it okay to drop the Brownian motion? This seems more like an empirical trick to get better PSNR and SSIM.
Yes, Dropping Brownian motion is an empirical trick that we followed from GOUB as stated in line 165 of our paper. It can obtain better results on image restoration tasks, capturing more pixel details and structural perceptions of images (which contributes to better PSNR and SSIM). This phenomenon has been verified by GOUB's three ablation experiments.
Q2: There seems some typos in Eq 15,16 \nabla p -> \log \nabla p.
Yes, you're right. We will correct it in the revised version.
This paper proposes a framework that unifies and extends various diffusion bridge methods by way of stochastic optimal control. In the case of linear dynamics, they derive a computationally tractable method, which can be thought of as a regularization of previous methods by the introduction of a new hyperparameter. Implementing this change requires only a minimal modification to existing code. They show that, by tuning this new hyperparameter, improved performance can be achieved in a number of benchmark examples.
update after rebuttal
My assessment has not changed substantially. I have maintained my score at 4: Accept
In addition, earlier I posted a followup discussion regarding the proof of Prop. 4.3 under an incorrect heading, so the authors were unable to see it. I believe it would further improve the paper if the following was addressed, though it would not change my score either way:
I do still believe that Proposition 4.3 is much more elementary than it is made out to be. Below is a more detailed version of the comment in my original review, on which the authors could base a revised elementary proof.
Starting with the case, the optimal control must lead to , else the cost is infinite, and therefore
Now looking at the finite case, we can bound the minimum by the value at the control :
The SDE determining doesn't depend on , hence in the above expression we still have . Therefore we arrive at
给作者的问题
-
Regarding Proposition 4.3, perhaps I misunderstand, but I believe this proposition is trivial from the definitions. Under the hard constraint () the terminal cost is enforced to be exactly zero, hence so both costs agree on . Therefore the minimal cost when can't be larger than the cost of the control . I'm not sure this fact is deserving of its own proposition. I am also not convinced that comparing the minimal cost values in this way says anything about why one method performs better than another in any operational sense. I think the answer probably lies more in the direction of the soft constraint being a numerically better behaved regularization than the hard constraint. I think the authors need to provide a better intuitive discussion about why (17) should be expected to correspond to better performance in the experiments in section 5 or else remove this proposition and the surrounding paragraph from the main text, as well as alter the discussion in Section 4.5.
-
Could you comment further on the performance of UniDB (SDE) vs UniDB (ODE) in table 1. Specifically, is there a reason why the former performs well for LPIPS and FID while the latter performs well for PSNR and SSIM?
论据与证据
The majority of their claims are supported by clear and convincing evidence. Generalization through the methods of of stochastic control is well-grounded theoretically, and they show improved performance empirically on a number of convincing benchmarks
The one claim that I don’t find substantiated revolves around their Proposition 4.3. They claim that this proposition lends theoretical support to the observation that introducing their new hyperparameter leads to improvements in practice. I find that result to be mathematically trivial and also to be disconnected from saying anything about how well the method performs in practice. However, I don’t think this proposition is integral to their work, and it could be removed (or de-emphasized) without any negative effects.
方法与评估标准
I find the methods and evaluation criterial to make sense for the problem.
理论论述
I checked the proof of theorem 4.1 and didn’t find any substantial issues.
实验设计与分析
I reviewed the experimental designs in Section 5 and did not observe any issues.
补充材料
There were no attached supplementary materials. I did review the appendices.
与现有文献的关系
The key contribution is a reformulation of diffusion bridge methods in the language of stochastic control, which motivates a natural one-parameter family of methods extending previous work, especially GOUB (Yue et al., ICML, 2024). I find this to be a natural and interesting extension.
遗漏的重要参考文献
I am not aware of any missing essential references.
其他优缺点
I found the contribution to be original and well-motivated theoretically, the writing to be relatively clear, and the performance improvements to be nontrivial.
其他意见或建议
- Equations 9 and 11 require expected values, yes? They are still stochastic control problems.
- Above equation 19, the formula for the score should have a log.
- Computations in Eq 36 and 37 have extraneous commas at the end of lines.
- Line 705 appears to be repeated.
We sincerely thank you for your comment.
Claim & Question 2: The reviewer finds that result to be mathematically trivial and also to be disconnected from saying anything about how well the method performs in practice. However, The reviewer doesn’t think this proposition is integral to their work, and it could be removed (or de-emphasized) without any negative effects. The authors need to provide a better intuitive discussion about why (17) should be expected to correspond to better performance in the experiments in section 5 or else remove this proposition and the surrounding paragraph from the main text.
We appreciate this feedback and can move Proposition 4.3 to Appendix.
Here we add more analysis to help understand the practical implications of our UniDB. When approaches , our UniDB reduces to Doob's h-transform, and the control term in Eq (12) becomes ineffective, leading to . Although Doob's h-transform ensures reaches the target endpoint exactly, i.e., , it may force the model to preserve even harmful noise/artifacts in the target. This is becuase Doob's h-transform will apply disproportionately large control inputs to achieve such exact matching. The large in SDE trajectory may disrupt the inherent continuity and smoothness of images. The over-control in Doob's h-transform violates the natural statistical properties of images, prioritizing mathematical precision (pixel-perfect endpoints) over visual authenticity (realistic SDE trajectories). As shown in our Figure 1, Doob's h-transform can lead to artifacts along edges and unnatural patterns in smooth regions.
Moreover, we want to emphasize that the discovery of Proposition 4.3 is non-trivial. Proposition 4.3 and related mathematical derivations in Appendix A.3 show that . Generally, one cannot determine the magnitude between and because although but . We use strict mathematical derivations in Appendix A.3 to show that is true. Though Proposition 4.3 does not directly mean better image quality by UniDB, it can show that UniDB leads to better overall performance considering both realistic SDE trajectories and target endpoint matching.
Question 1: Is there a reason why the UniDB (SDE) performs well for LPIPS and FID while the UniDB (ODE) performs well for PSNR and SSIM?
This is a common phenomenon in various Diffusion models [R1-R3]. As analyzed in [R1], "although solvers for the probability flow ODE allow fast sampling, their samples typically have higher (worse) FID scores than those from SDE solvers if no corrector is used". In the paper SDE-Drag [R2] and GOUB [R3], they compare the performance of the SDE and ODE models and get the similar phenomenon: SDE model performs well for LPIPS and FID. Particularly in GOUB [R3], the experimental results also demonstrate the better performance of ODE model for PSNR and SSIM.
[R1]: Song et al. "Score-Based Generative Modeling through Stochastic Differential Equations.", ICLR 2021.
[R2]: Nie et al. "The Blessing of Randomness: SDE Beats ODE in General Diffusion-based Image Editing", ICLR 2024.
[R3]: Yue et al. "Image restoration through generalized ornstein-uhlenbeck bridge", ICML 2024.
Comments Or Suggestions: Equations 9 and 11 require expected values, yes? They are still stochastic control problems; Above equation 19, the formula for the score should have a log; Computations in Eq 36 and 37 have extraneous commas at the end of lines; Line 705 appears to be repeated.
Yes, you're right. Thanks for pointing out these typos and we will correct them in the revised version.
The authors reformulated a diffusion bridge matching problem as a stochastic control problem and developed a UniDB framework.
The proposed UniDB approach focuses on stochastic optimal control, modeling the forward process as an optimization problem to analyze the drawbacks of Doob's h-transform.
As a result they arrived to a one-dimensional parametric family of solutions. Thanks to an additional parameter it is possible to improve quality of the distribution matching results.
Proofs are correct. Experimental validation is solid. The approach demonstrated the best accuracy on a number of relatively complex benchmarks.
The reviewers did come to a consensus here, so the decision is relatively easy.