PaperHub
7.3
/10
Poster4 位审稿人
最低4最高5标准差0.5
4
4
5
5
3.8
置信度
创新性2.3
质量3.0
清晰度2.8
重要性2.8
NeurIPS 2025

Variational Regularized Unbalanced Optimal Transport: Single Network, Least Action

OpenReviewPDF
提交: 2025-05-06更新: 2025-10-29
TL;DR

Var-RUOT is a new framework improving RUOT for recovering dynamics from high-dimensional snapshots. It incorporates first-order optimality conditions, achieving better performance on single-cell datasets.

摘要

关键词
unbalanced optimal transportSchrödinger bridgetrajectory inferencesingle-cell

评审与讨论

审稿意见
4

The authors of the paper introduce Var-RUOT, a method to perform deep regularised unbalanced optimal transport. Var-RUOT parameterizes a scalar potential and derives the necessary conditions required by the associated field and growth penalty to minimise the isotropic time invariant RUOT problem. The potential is trained by minimising a loss function computed over simulated particles from an empirical measure. The loss has multiple components. First, a reconstruction term ensures that the model recapitulates the marginal distributions over time and preserves the total mass of individual snapshots. The second component is a Hamilton-Jacobi-Beltrami loss enforcing field optimality under the ITI-RUOT problem. Furthermore, the least action loss is added to regularize the dynamic optimal transport field. Finally, the paper proposes a new biological interpretation of the growth penalty function, showing that it varies in the direction of the transporting field. By controlling the second derivative of the penalty function, one can encode biological processes, such as the loss of stemness. In their experiments, the authors compare their model with existing methods on the task of reconstructing time marginals of real and simulated single-cell data. Together with distribution matching metrics the paper additionally shows that the model achieves lower path action and faster convergence compared to its counterparts, offering an efficient approach to study cellular development in cellular biology.

优缺点分析

Strengths

I generally find the paper well-written, for which I commend the authors. The math is introduced gradually and is easy to digest, which is a very positive aspect in my opinion. I also think that the problem of enforcing stronger optimality constraints to dynamical OT is relevant and well-justified, specifically in the unbalanced setting, which is very useful in computational biology. I particularly enjoyed the growth penalty explanation part, as a better interpretation of its analytical form could provide end users with no expertise in OT with a biological rationale on how to set its value to optimally model the system they are considering.

Weaknesses

At the moment, there are some concerns on my side regarding the presentation of the paper that induced me to a borderline rejection score. I would like to know more about the authors' opinions during the rebuttal phase to be able to produce a final informed assessment.

  • Selection of the growth penalty: Though I appreciate the formalisation of the growth penalty in biological terms, the results proposed by the authors show that using more sophisticated terms than the WFR standard approach is suboptimal on the snapshot reconstruction task. Therefore, I am not sure this specific aspect of the paper is ripe enough to be used as an effective contribution to the paper. I still think this aspect is interesting, and maybe additional options for the penalty functions could be considered in further iterations of the paper, or at least be shown in a setting where it is actually useful.
  • Line 103 Here, I recommend that the authors define α\alpha with a short sentence.
  • Time index: Overall, I would personally find it easier to understand if all the time-resolved variables and quantities in the paper had an index indicating time. Some of them already have this (for example, the density ρ\rho), but some others miss it (as, for instance, the empirical measure μ\mu or the weights ww).
  • Benchmark: One thing that I find sub-optimal in the current presentation is the way benchmarks are conducted. First, some models are presented in Tab. 1 and 2, while many of them are dropped later in Tab. 4. I personally believe that the baselines should be the same throughout to ensure transparency.
  • Missing baselines: Moreover, nowadays, there are very powerful OT-based methods for trajectory inference in low dimensions, like OT-CFM and derivatives thereof. I think I would have given some priority to them as baselines to test whether stochasticity and unbalancedness help in trajectory reconstruction. Also, combinations of OT Flow Matching and unbalancedness were tackled in some existing works [1], which I think is worth comparing against. Along the same line, comparison with [2], which is presented as a special case of the current model, is beneficial. What I am trying to say is that in a research setting where simulation-based methods are overtaken by their simulation-free counterpart, comparison with the latter models is meaningful to show that the proposed approach is a valuable option.
  • Dataset choice. There are widespread benchmarks of methods used to match temporal snapshots in scRNA-seq. They are based on the Embryoid body dataset from [3] and the NeurIPS benchmark in [4]. They also test slightly higher dimensionalities of the input space, which I think is relevant here in a simulation-based setting.
  • Figure presentation. In my opinion, Fig. 3 is not very clear. I can see what the colors mean, but the paths are not clear to me. I think the presentation could be improved by making the source and the target more understandable. For example, I don't understand where one can see higher curvature (line 256) in the results from DeepRUOT. Moreover, I would add a legend explaining the difference between crosses and full dots. Moreover, in Fig. 2, Var-RUOT produces different growth patterns compared to DeepRUOT, where the growth rate first increases and then goes down again in the last time point. Is it expected? I believe this figure could benefit from more explanations. In other words, in most of the provided qualitative evidence, I cannot tell why Var-RUOT is better than DeepRUOT.
  • Wall clock time: Very minor point. I would add the unit of measure of wall clock time somewhere in the table caption.
  • Description of performance: Sometimes, the authors seem to interpret the scores in absolute terms. For example, in lines 248-250 or in lines 286-288. I think these scores are mostly comparative, so I would phrase it differently and say that the results demonstrate that Var-RUOT better models trajectories than the baselines, rather than the accurate recovery claim.

I wish I could be more positive in my assessment at this stage. I am looking forward to hearing more from the authors in their responses.

[1] Eyring, L., Klein, D., Uscidda, T., Palla, G., Kilbertus, N., Akata, Z. and Theis, F., 2023. Unbalancedness in neural monge maps improves unpaired domain translation. arXiv preprint arXiv:2311.15100.

[2] Neklyudov, K., Brekelmans, R., Tong, A., Atanackovic, L., Liu, Q. and Makhzani, A., 2023. A computational framework for solving wasserstein lagrangian flows. arXiv preprint arXiv:2310.10649.

[3] Moon, K.R., Van Dijk, D., Wang, Z., Gigante, S., Burkhardt, D.B., Chen, W.S., Yim, K., Elzen, A.V.D., Hirn, M.J., Coifman, R.R. and Ivanova, N.B., 2019. Visualizing structure and transitions in high-dimensional biological data. Nature biotechnology, 37(12), pp.1482-1492.

[4] Luecken, M.D., Burkhardt, D.B., Cannoodt, R., Lance, C., Agrawal, A., Aliee, H., Chen, A.T., Deconinck, L., Detweiler, A.M., Granados, A.A. and Huynh, S., 2021, August. A sandbox for prediction and integration of DNA, RNA, and proteins in single cells. In Thirty-fifth conference on neural information processing systems datasets and benchmarks track (Round 2).

问题

  • Why are the weights initialised as 1? Is there a way to instead learn them?
  • What is the ground truth for the mass objective? In other words, how do you compute M(Tk) if you only have the weights for the solution of the ODE?
  • Why did the authors pick 215\frac{2}{15} as an exponent of ϕ2\phi_2? How does the penalty function behave for different values of the exponent?
  • Line 71: isn’t the Schrödinger bridge always dynamic as it operates via regularisation to a reference probability path?

局限性

The limitations of the work have been sufficiently addressed in the Appendix.

最终评判理由

Dear AC,

Thank you very much for handling this paper.

In my final assessment, I increased the score assigned to the present submission from 3 to 4. The authors have addressed most of my concerns with substantial experiments, explanations, and analyses. I believe the paper is above the acceptance bar, as it introduces new theory while providing value for downstream applications.

The reason I did not score a 5 is that I believe some aspects are not fully polished, such as the exploration of the relevance of the new formulation of the non-canonical growth penalty and the lack of presentation quality in the qualitative results.

I remain available to provide further justification for my score.

Best regards,

The Reviewer

格式问题

I did not come across any formatting constraints during my review.

作者回复

Thank you for carefully reviewing our paper and for providing valuable comments. Below please find our point-by-point response:

Summary & Strengths

We appreciate your recognition of our work's strengths. Our paper derives the first-order optimality conditions for the RUOT problem and solves it by fitting a single scalar field, thereby modeling unbalanced distributions and stochastic dynamics. Compared to previous methods, our Var-RUOT finds paths with lower action, achieves stable training, and converges faster. We also discuss the biological priors associated with the RUOT penalty function. With each action guiding its corresponding dynamics, we hope our framework offers a more physical approach to modeling single-cell omics data.

[W1, Q3] On the Selection of g215g^\frac{2}{15}

Thank you for your question regarding ψ(g)\psi(g). To clarify, we use g2/15g^{2/15} as an example to show that when ψ(g)>0\psi'(g)>0 and ψ(g)<0\psi''(g)<0, g(x,t)g(\boldsymbol{x},t) increases along u(x,t)\boldsymbol{u}(\boldsymbol{x},t) (Theorem 4.2). We chose ψ2(g)=g2/15\psi_2(g)=g^{2/15} because:

  1. It is the simplest form that meets dψ(g)dg>0\frac{d\psi(g)}{d|g|}>0, ψ(g)=ψ(g)\psi(g)=\psi(-g), and ψ(g)<0\psi''(g)<0.
  2. From the optimality condition we obtain
g(λ)=(λα2q+12p)2q+12p(2q+1)(1λ)2q+1(2q+1)2p.g(\lambda) = \left(\frac{\lambda}{\alpha}\frac{2q+1}{2p}\right)^{\frac{2q+1}{2p-(2q+1)}} \propto \left(\frac{1}{\lambda}\right)^{\frac{2q+1}{(2q+1)-2p}}.

For a more stable training process, we keep g(λ)g(\lambda) from being too steep near λ=0\lambda=0. So we set p=1p=1 and choose a large qq (i.e., q=7q=7).

Note that g(λ)g(\lambda) is undefined at λ=0\lambda=0 due to the necessary discontinuity in ψ(g)\psi'(g) at g=0g=0. To address this, we linearly interpolate between g(δ)g(-\delta) and g(δ)g(\delta) (Appendix A.2). This singularity causes slowing convergence and degrading distribution reconstruction when using ψ2(g)\psi_2(g). A more reasonable way to deal with this problem is to fit the reciprocal of λ\lambda, which is our future work.

Empirically, on the 2D Mouse Blood Hematopoiesis dataset (results are shown below), ψ(g)=g2/15\psi(g)=g^{2/15} performs similarly to ψ(g)=g2/21\psi(g)=g^{2/21}, while ψ(g)=g2/9\psi(g)=g^{2/9} (with a steeper g(λ)g(\lambda)) leads to unstable training and poorer reconstruction. Our experiments demonstrate that choosing the exponent on gg within a certain range is feasible. We will update the paper to include the above analysis in the sensitivity analysis section of the Appendix.

ψ(g)\psi(g)/W1\mathcal{W}_1 Distancet=1t=2
ψ(g)=g(2/9)\psi(g)=g^{(2/9)}0.94761.2197
ψ(g)=g(2/15)\psi(g)=g^{(2/15)}0.29530.1917
ψ(g)=g(2/21)\psi(g)=g^{(2/21)}0.28580.2565

As noted in our Limitations, choosing ψ(g)\psi(g) remains open. Future work may derive it from physical principles (e.g., Branching SDE, Stochastic Thermal Dynamics) or learn it with a neural network. We will include these discussions in the Appendix.

[Q1] On Weight Initialization

Using particle methods, we approximate the true measure μ\mu with the empirical measure μN\mu^N. We initialize by placing a particle at every t=0t=0 data point with the uniform weight 1. This yields

μN(0)=1Ni=1Nwiδ(xXi)μ(0),\mu^{N}(0) = \frac{1}{N}\sum_{i=1}^{N} w_i \delta(x - X_{i} ) \rightarrow \mu(0),

which justifies the rationale to set wi(0)=1w_i(0)=1.

[Q2] On the Ground Truth for Mass Matching

In RUOT modeling, because cell counts vary (due to growth or death), the number of data points differs over time. If we set the total mass at t1=0t_1=0 as the reference number 1, then the mass at time tit_i as NiN1\frac{N_i}{N_1} (with NiN_i as the count at tit_i), which serves as the ground truth for mass matching loss. We will clarify this in our paper.

[Q4]On the Schrodinger Bridge Problem

We apologize for the confusion. The original Schrodinger Bridge is dynamic—minimizing the KL divergence over path space between a stochastic process and a reference one. In our paper, "static" means some method[(Tong, A.,et al) S2FM] first compute a static OT solution via entropy-regularized OT, then fit the velocity field using Conditional Flow Matching. We will clarify this in the paper.

[W4,5,6] Additional Baselines for Existing Datasets

Thanks for the great suggestion. Follow your advice, we supplemented three baselines on existing datasets; in addition, we provided results for nine baselines and Var-RUOT on two new datasets. We first supplemented the additional baselines on existing datasets with:

  • Wasserstein Gradient Flow (WGF in the table) [(Neklyudov, K., et al.) Wasserstein Lagrangian Flows]
  • OTCFM [(Lipman, Y., et al.) Flow Matching]
  • UOTCFM [(Eyring, L., et al.)Unbalancedness in Neural Monge Maps]

Specifically, for the 2D Mouse Blood Hematopoiesis dataset, we have added three previously unevaluated baselines: MIOFLOW, S2FM, and PISDE.

Simulation Dataset

Method/W1\mathcal{W}_1 Distancet1t_1t2t_2t3t_3t4t_4
OTCFM0.10350.20780.28980.3107
UOTCFM0.10020.16530.17110.4129
WGF0.49830.83460.80460.4493
Var-RUOT0.04520.03850.04450.0572

EMT Dataset

Method/W1\mathcal{W}_1 Distancet1t_1t2t_2t3t_3
OTCFM0.25570.27010.2799
UOTCFM0.25380.26960.2771
WGF0.39010.43810.2848
Var-RUOT0.25400.26700.2683

2D Mouse Dataset

Method/W1\mathcal{W}_1 Distancet1t_1t2t_2
MIOFLOW0.35460.2772
S2FM0.17060.1602
PISDE0.31240.2531
OTCFM0.36740.3222
UOTCFM0.33010.2051
WGF0.33020.2051
Var-RUOT0.12000.1917

Next, we show experiments on additional datasets. We use a 100-dimensional, 5-time-point (t=0,1,2,3,4t=0,1,2,3,4) EB dataset [(Moon, K.R., et al) Visualizing structure and transitions...] and the NeurIPS challenge dataset [(Luecken, M.D., et al.) A sandbox for prediction and integration...]. For NeurIPS dataset, we extract gene expression data from donor 28483, cluster them via K-Means by pseudo time, and order the clusters (by centroid) into t=0,1,2,3,4t=0,1,2,3,4. This new dataset is then used for further tests. The results are as follows:

EB Dataset

Method/W1\mathcal{W}_1 Distancet1t_1t2t_2t3t_3t4t_4
S2FM9.673211.071112.543614.7100
PISDE8.83179.42839.787211.1102
MIOFLOW8.71829.41399.754711.0080
TIGON8.79929.649710.013011.3284
RUOT8.80299.651810.036511.3555
OTCFM9.445911.378614.052919.7489
UOTCFM9.479411.606014.671521.7653
ActionMatching12.714516.780913.466318.1165
WGF9.984711.056211.971913.2417
Var-RUOT8.91589.795510.264711.4927

NeurIPS Dataset

Method/W1\mathcal{W}_1 Distancet1t_1t2t_2t3t_3t4t_4
S2FM6.187612.61709.183012.5831
PISDE5.66425.23586.27938.9086
MIOFLOW6.42676.27157.10798.9035
TIGON5.94815.75996.55398.9573
RUOT6.01665.69726.50998.8047
OTCFM5.50825.44767.093510.1416
UOTCFM5.55265.43496.74829.6782
ActionMatching7.37018.51569.28429.9665
WGF7.17857.71358.25218.6629
Var-RUOT4.79614.18166.23129.1058

And we compared the path action learned by TIGON, DeepRUOT and VarRUOT:

Method/Path ActionTIGONDeepRUOTVarRUOT
EB Dataset4.49275.05714.1536
NeurIPS Dataset10.425010.98983.6982

In summary, the experiments show that:

  • For low-dimensional data: On the EMT dataset, Var-RUOT achieves distribution reconstruction comparable to OTCFM and UOTCFM, and on the three-gene simulation and 2D Mouse Blood Hematopoiesis datasets, it outperforms the other baselines.
  • For high-dimensional data: On the EB dataset, Var-RUOT yields results similar to TIGON and DeepRUOT, and outperforms simulation-free methods (OTCFM, UOTCFM) as well as Wasserstein Gradient Flow; on the NeurIPS dataset, it outperforms existing methods. Notably, compared to DeepRUOT and TIGON, Var-RUOT finds a path with significantly lower action.

Thus, the experiments confirm that Var-RUOT remains competitive on high-dimensional data, achieving good distribution reconstruction while finding lower-action paths. We will include these discussions on new datasets & baselines in the Appendix.

[W7] On the Figures

Thank you for your questions regarding our figures. For Figures 2 and 3:

  • Crosses represent real dataset points, and circles represent data generated by our algorithm.
  • Trajectory colors indicate particle mass, with brighter areas showing higher mass.

We will add legends and annotations for clarity.

Regarding the EMT results, Figure 3 (a) shows DeepRUOT produces curved trajectories (even passing through sparse top regions), while Figure 3 (b) shows nearly straight lines connecting the start and end points. Since straight lines minimize action, Table 2 compares the Path Action of DeepRUOT and Var-RUOT, with Var-RUOT achieving significantly lower action.

The growth pattern in Var-RUOT emerges naturally from using the WFR Metric (Theorem 4.2), where the growth rate increases along the trajectory. Without a "minimum action" objective, many valid velocity fields and growth patterns could reconstruct the marginals. Both methods correctly reconstruct the marginals, while Var-RUOT finds a lower-action solution, which aligns with our objective.

We clarify that our goal in presenting these figures is not to show that Var-RUOT has a significantly lower reconstruction loss, but to demonstrate that both methods perform similarly in marginal reconstruction. Var-RUOT’s advantages lie in it obtains a lower-action solution, simpler scalar field fitting, faster convergence, and the elimination of the pretraining phase in DeepRUOT.

[W2,3,8,9]Other Writing Issues

Thank you for your valuable suggestions. We will revise the paper to:

  • Define α\alpha at line 103 as the penalty coefficient for growth/death.
  • Clarify in lines 248–250 and 286–288 that the W1\mathcal{W}_1, W2\mathcal{W}_2 distances and Path Action are relative indicators, with Var-RUOT performing better relative to the baseline.
  • Add the unit (seconds) for Wall Clock time in Table 3.
  • Add time index to time-dependent variables.
评论

Dear authors, 

Thank you very much for your rebuttal. I am impressed and commend the amount of experimental effort you put into it. Below, I will provide some additional comments on the paper after rebuttal. 

Value of the exponent. I still believe that this theoretical analysis could benefit from a case where changing the exponent of gg provides biological advantages compared to the standard WRF approach, cause as of now it sounds a little bit like an underexplored avenue. However, I do appreciate the experiments on the variation of the parameter and the explanation. Thank you for those. 

Experiments. The experiments you provided are comprehensive. The only thing I wonder is whether you used the same pipeline as the compared publications to process and reconstruct the data, since the Wasserstein distance values on Cite and EB differ from the original publications, even just in terms of range. 

Figures. Thank you for your explanation. I now see what you mean. I would recommend making the paths more distinguished from the colours of the points, because now the qualitative analysis is a bit hard to grasp. I am still not sure that I understand the behaviour of Var-RUOT in Fig.2. The growth coefficient increases, decreases again and then increases again along the trajectory. Is this behaviour expected?

All in all, I think the paper passes the acceptance bar for the conference, so I decided to increase my 3 to a 4. All the best to the authors for the remainder of the discussion period.

评论

Thank you again for taking the time to review our paper and read our responses. We are especially glad to hear your positive evaluation of our work. We are also very happy to further discuss some questions here.

Regarding the exponent of gg

Thank you for your suggestion. In the revised paper, we will especially highlight the biological implications of various exponent choices as meaningful future work directions, and as a on-going work, we are indeed cooperating with wet-lab biologists now to incorporate lineage-tracing information to provide more biological insights into this problem.

Regarding the Experiments

Thank you for your reminders, and we have taken the chance to carefully check all the details of implementation. In our experiments, we standardized the data so that the mean of each principal component is 0 and the standard deviation is 1. Moreover, in our analysis, the EB Dataset was reduced to 50 dimensions using PCA, which may differ from the dimensional settings in other works (e.g. 5 dimensions). These data pre-processing differences explain why our results vary from those reported in the original literature. We will detail the data processing workflow in the revised version of the paper, especially highlighting the differences. For a more comprehensive evaluation, we will also provide another table using the processing pipelines in the original publications. More importantly, we will open-source all the data processing code and the processed datasets to ensure fully reproducible results of all the experiments.

Regarding the Figures

Thank you for your nice suggestion. We will adjust the trajectory colors so that they are clearly distinguishable from the points in the figure.

In addition, the outcome shown in Figure 2 is indeed the implication of Theorem 4.2 and one of the main messages conveyed by our work: if the WFR Metric is employed, the growth rate increases along the direction of the velocity field for the data points at each time step, which might bring such biologically "weird" phenonmenon. Because the assumed dynamical system is non-autonomous (with tt explicitly present in λ\lambda), this theorem does not impose constraints on the growth rate across data points from different time steps. As a result, the figure shows the growth rate first rising and then falling in order to fit the overall distribution change. Overall, this indeed motivates us to discuss the choices of ψ\psi, and implies that WFR metric needs to be carefully used in terms of biological interpretation. We will include more detailed discussion in the revised paper about this result.

Once again, thank you for all your insightful and constructive comments. Your feedback has greatly helped us improve the paper, and we will incorporate your suggestions into the final version.

Sincerely,

The Authors

评论

Dear authors,

Thank you again for your final remarks. Everything sounds reasonable. I confirm my final score of 4.

All the best!

审稿意见
4

The authors present a new method for learning high dimensional stochastic dynamics for unnormalized distributions. The authors build on the recent RUOT framework modifying the loss to ensure the learned solution satisfies the optimality condition of least action. This in turns allows them to use a single network and parametrize only a scalar field. Additionally they propose a growth penalty function which better aligns with biological priors.

优缺点分析

Strengths

  • The motivation of trying to satisfy the condition of least action is well motivated.
  • There is a substantial theory section.
  • The numerical results are quite strong. The method really does seem to be an improvement. The faster convergence speed is promising.

Weaknesses

  • The overall presentation is not great with regards to what the key methodological innovations are and how they fit within the existing literature. It is quite difficult to tease out the exact differences between this paper other work such as the original RUOT paper [1] or action matching style papers [2]. This is quite important because the latter works also explicitly enforce optimality conditions WRG to least action and the loss terms look very similar to what the paper calls "Action Loss".
  • In particular one of the key approaches of the proposed method is to paramzeraize with a single scalar field where by the velocity is the gradient of this field and the growth function is equal to it. The title of paper might suggest that this is a new innovation, but it is in fact exactly what is done in the original action matching paper [2]. While the authors cite this paper and state "these methods typically employ separate neural networks to parameterize the velocity and growth functions, without leveraging their optimality conditions or the inherent relationship between them". To my knowledge this is incorrect with respect to [2].

[1] Learning stochastic dynamics from snapshots through regularized unbalanced optimal transport. Zhang. 2024. [2] Action Matching: Learning Stochastic Dynamics from Samples. Neklyudov. 2023.

问题

  • It would be great to see a table (or something like this) clearly showing the differentiate factors and or advantages of their method compared to the many other (see above).
  • Recent works show that estimating action losses can be quite difficult and they proposed quadrature of these types of losses [3]. Do the authors encounter similar issues?
  • How and why are the authors able to eliminate the pertaining steps of the original RUOT algorithm [1] ?
  • How exactly was the g^2/15 value obtained? Was this a tuned hyper-parameter?

[1] Learning stochastic dynamics from snapshots through regularized unbalanced optimal transport. Zhang. 2024. [3] Parametric model reduction of mean-field and stochastic systems via higher order action matching. Berman. 2024.

局限性

  • The Limitations section is pushed to the appendix, which in my opinion is not in the spirit of the guidelines and neurips checklist.

最终评判理由

The response was very thorough and addressed my concerns.

格式问题

  • Large emojis in the first figure make the paper feel unserious. I recommend removing these.
作者回复

Thank you for carefully reviewing our paper and raising valuable questions. Below we address your concerns in a point-to-point response:

Summary & Strengths

We appreciate your recognition of our work's strengths. Our paper derives the first-order optimality conditions for the RUOT problem and solves it by fitting a single scalar field, thereby modeling unbalanced distributions and stochastic dynamics. Compared to previous methods, our Var-RUOT finds paths with lower action, achieves stable trainin, and converges faster. We also discuss the biological priors associated with the RUOT penalty function. With each action guiding its corresponding dynamics, we hope our framework offers a more physical approach to modeling single-cell omics data.

[W1, W2, Q1] Differences from Previous Methods

Thank you for your question and sorry for the confusion. First, we sincerely apologize for not thoroughly checking our citation positions and misunderstandings caused by it. We fully agree that Action Matching is a very innovative work uses a single network to parameterize sθ(x,t)s_{\theta}(\boldsymbol{x}, t) that inspires our research; we will review the manuscript to correct any improper citations and clarify the significance of previous methods.

Following your great suggestion, we also made a table to clearly highlight the contribution of Var-RUOT that will later be included in the revised Appendix:

MethodUnbanlanced DistributionStochastic DynamicsUnbalanced + StochasticLeast Action
SF2M-+--
PISDE-+-+
MIOFlow----
Action Matching++-+
TIGON+---
DeepRUOT+++-
Var-RUOT++++
  • Unbalanced Distribution indicates that the method can handle non-normalized distributions.
  • Stochastic Dynamics means the method supports stochastic dynamics.
  • Unbalanced + Stochastic shows the method can handle both simultaneously (e.g., DeepRUOT).
  • Least Action signifies that the method employs a first-order condition to minimize the action, rather than simply adding the action as a weighted loss.

The key difference between Var-RUOT and DeepRUOT is that Var-RUOT leverages first-order optimality conditions in its parameterization and loss design (e.g., using only a scalar field rather than three neural networks as in DeepRUOT). Compared to DeepRUOT, our approach finds paths with lower action and addresses unstable training and slow convergence, thus eliminating the need for a pre-training phase.

Var-RUOT and Action Matching differ in three key ways:

Firstly, Var-RUOT can simultaneously handle imbalanced data distributions and stochastic dynamics, both crucial for single-cell omics.

Secondly, Action Matching's training algorithm needs to sample from intermediate distribution qt(x)q_{t}(x), while in single-cell omics data, we only have snapshot data at a few time points as boundary sample. So acquiring intermediate samples from qt(x)q_{t}(x) may be challenging due to missing evolution data in-between, potentially hindering the ability to capture complex dynamics. Empirically, our experimental results (Tables 1, 2, 4) show that Var-RUOT achieves more satisfactory results over existing results including Action Matching.

Lastly, Action Matching defines the Action Gap loss as:

LAM=Eq0(x)[s0(x)]Eq1(x)[s1(x)]+01Eqt(x)[12st(x)2+st(x)]dtL_{\text{AM}} = E_{q_{0}(x)} [s_{0}(x)] - E_{q_{1}(x)} [s_{1}(x)] + \int_{0}^{1} E_{q_{t}(x)} \left[ \dfrac{1}{2} \| \nabla s_{t}(x) \|^{2} + \dfrac{\partial s}{\partial t}(x) \right] \mathrm{d} t

While Var-RUOT directly adopts the HJB Loss(Line 204-209) to ensure that the computed solution λ(x,t)\lambda(\boldsymbol{x},t) satisfies the HJB equation. In other words, Var-RUOT incorporates the first-order condition into the loss, enabling it to find a solution with lower action. We will add these discussions in the revised manuscript.

[Q2] On the Estimation of the Action Loss

Thank you for your question. In our experiments, we employed two methods to compute the integral losses for both the action and HJB terms. When using the Euler-Maruyama method, setting the step size too large (e.g., Δt=0.2\Delta t = 0.2) results in high variance estimates for the action and HJB losses, leading to unstable training. Therefore, we use Δt=0.1\Delta t = 0.1 in all experiments. The Stochastic Runge-Kutta method offers better convergence, allowing an integration step of Δt=0.2\Delta t = 0.2. Since our approach is simulation-based, a longer step size significantly reduces training time. In practice, we thus choose the Stochastic Runge-Kutta method and select a larger step size as long as training remains stable. We will discuss this in our final version and cite [(Berman.) High Order Action Matching] in our paper.

[Q3] Why the Related Steps in the Original RUOT Algorithm Can Be Omitted

In the original DeepRUOT algorithm, three neural networks are employed to approximate v(x,t)\boldsymbol{v}(\boldsymbol{x},t), g(x,t)g(\boldsymbol{x},t), and logρ(x,t)\log \rho(\boldsymbol{x},t), with all constraints imposed as soft constraints via loss functions. For instance, to ensure that v\boldsymbol{v}, gg, and ρ\rho satisfy the Fokker–Planck equation, DeepRUOT introduces the Fokker–Planck loss

LFP=tpθ+x(pθv)gθpθ+λwpθ(x,0)p0L_{\text{FP}} = \|\partial_{t} p_{\theta} + \nabla_{x} \cdot (p_{\theta}\boldsymbol{v})- g_{\theta} p_{\theta}\| + \lambda_{w} \| p_{\theta}(\boldsymbol{x},0) - p_{0} \|

Training these three interdependent networks simultaneously is challenging and often causes convergence issues. To accelerate training by obtaining a good initialization, DeepRUOT incorporates a pretraining phase: (1) training v\boldsymbol{v} and gg using a reconstruction loss LRecons{L}_{\text{Recons}}, and (2) training pp with Conditional Flow Matching.

In Var-RUOT, we use a single neural network λθ(x,t)\lambda_\theta(\boldsymbol{x},t) to approximate a scalar field. The velocity field u(x,t)\boldsymbol{u}(\boldsymbol{x},t) and growth rate g(x,t)g(\boldsymbol{x},t) are derived analytically from λ(x,t)\lambda(\boldsymbol{x},t) using first-order optimality conditions (Theorem 4.1). As a result, Var-RUOT avoids handling multiple interdependent networks and eliminates the need for the pretraining phase used in DeepRUOT. This is indeed one of the main contribution of our work; ablation experiments in DeepRUOT demonstrate that these pretraining steps are indispensable (DeepRUOT[(Zhang.), DeepRUOT], Appendix B.5), and we have removed them in Var-RUOT. We will add these discussions in the revised manuscript.

[Q4] Why we Choose g2/15g^{2/15}

Thanks for your question regarding ψ(g)\psi(g). we choose g2/15g^{2/15} primarily to serve as an example showing that when ψ(g)>0\psi'(g) > 0 and ψ(g)<0\psi''(g) < 0, at each time step g(x,t)g(\boldsymbol{x},t) increases along the direction of the velocity field u(x,t)\boldsymbol{u}(\boldsymbol{x},t) (Theorem 4.2). As for why we adopt ψ2(g)=g2/15\psi_{2}(g) = g^{2/15} as the example, it is because: (1) It is the simplest choice that satisfies the conditions given in our paper (lines 224–229): dψ(g)dg>0\dfrac{\mathrm{d}\psi(g)}{\mathrm{d}|g|} > 0, ψ(g)=ψ(g)\psi(g) = \psi(-g), and ψ(g)<0\psi''(g) < 0. (2) For training considerations: when choosing ψ(g)=g(2p)/(2q+1)\psi(g)=g^{(2p)/(2q+1)}, the first-order optimality condition allows us to derive

g(λ)=(λα2q+12p)2q+12p(2q+1)(1λ)2q+1(2q+1)2p.g(\lambda) = \left(\dfrac{\lambda}{\alpha} \dfrac{2q+1}{2p}\right)^{\frac{2q+1}{2p - (2q+1)}} \propto \left(\dfrac{1}{\lambda}\right)^{\frac{2q+1}{(2q+1)-2p}}.

We aim for more stable training, and since gg is computed directly from λ\lambda, we do not want the g(λ)g(\lambda) to be overly steep (especially near λ=0\lambda=0), which requires the exponent 2q+1(2q+1)2p\dfrac{2q+1}{(2q+1)-2p} to be as small as possible. This means choosing pp as small as possible (hence p=1p=1) and choosing qq as large as possible (we select q=7q=7). Therefore, we ultimately choose the penalty function ψ(g)=g2/15\psi(g)=g^{2/15}.

*Note: It is worth noting that g(λ)g(\lambda) is undefined at λ=0\lambda=0. In fact, for any function ψ(g)\psi(g) that meets the conditions outlined in lines 224–229 along with ψ(g)<0\psi''(g)<0, its first derivative must be discontinuous at g=0g=0, so this singularity is inevitable. To handle this issue, we select a small value δ\delta and linearly interpolate between g(δ)g(-\delta) and g(δ)g(\delta) by redefining

g(λ)=12δ(g(δ)g(δ))(λ+δ)+g(δ),λ[ϵ,ϵ),g(\lambda) = \dfrac{1}{2\delta}\left(g(\delta) - g(-\delta)\right)(\lambda + \delta) + g(-\delta), \quad \lambda \in [-\epsilon, \epsilon),

to bypass the singularity (Appendix A.2). Another approach to deal with the singularity is to fit the reciprocal of λ\lambda, denoted as μ\mu, thereby moving the singularity at λ=0\lambda=0 to μ\mu \rightarrow \infty. In practice, μ\mu \rightarrow \infty does not occur, just as the singularity where g|g| \rightarrow \infty is not reached. This method avoids artificially modifying the g(λ)g(\lambda) relationship and is more reasonable. This will be our future work.*

To further explain our training considerations, we experimented with two additional parameter sets on the 2D Mouse Blood Hematopoiesis dataset. The experimental results are as follows:

ψ(g)\psi(g) / W1\mathcal{W}_1 Distancet=1t=2
ψ(g)=g(2/9)(p=1,q=4)\psi(g)=g^{(2/9)}(p=1, q=4)0.94761.2197
ψ(g)=g(2/15)(p=1,q=7)\psi(g)=g^{(2/15)}(p=1, q=7)0.29530.1917
ψ(g)=g(2/21)(p=1,q=10)\psi(g)=g^{(2/21)}(p=1, q=10)0.28580.2565

The experimental results indicate that choosing ψ(g)=g2/15\psi(g)=g^{2/15} produces performance comparable to ψ(g)=g2/21\psi(g)=g^{2/21}, while selecting ψ(g)=g2/9\psi(g)=g^{2/9} leads to unstable training and poorer reconstruction of the distribution. Our experiments demonstrate that choosing the exponent on gg within a certain range is feasible. We will update the paper to include the above analysis in the sensitivity analysis section of the Appendix.

On Writing Issues

Thank you for your suggestions regarding our writing. We will review and revise our paper to:

  • Move the Limitations section into the main text.
  • Remove the emoji from Figure 1.
评论

Thank you these responses were helpful. I feel my concerns have been addressed. Conditional on the authors making these changes, I think this work is a good edition to the conference.

评论

Thank you again for taking the time to review our paper and to read our responses. Your insightful and constructive feedback has helped us further improve our work. We will incorporate your suggestions into the revised version of the paper.

Sincerely,

The Authors

审稿意见
5

This paper introduces Var-RUOT, a new method for learning the dynamics of evolving systems using just a few snapshot observations. Traditional approaches rely on multiple neural networks and often struggle to find the most efficient, biologically realistic paths. Var-RUOT simplifies the problem by learning only a single scalar function, from which all relevant dynamics can be derived. This leads to more stable training, faster convergence, and solutions that follow the principle of least action (i.e., the most efficient evolution). The method also allows for customizing the behavior of growth or decay in the system to better reflect biological processes. Through experiments on synthetic and real biological datasets, the authors show that Var-RUOT achieves more accurate and efficient results than existing approaches.

优缺点分析

Strengths

  1. By reducing the RUOT problem to learning a single scalar field \lambda(x,t), the method avoids redundant parameterization of velocity and growth fields, leading to a compact, interpretable, and theoretically grounded framework.

  2. Var-RUOT achieves faster convergence and more stable training compared to prior methods, as shown through detailed empirical comparisons across multiple datasets and metrics (e.g., action, Wasserstein distance, wall time).

  3. The paper connects the choice of the growth penalty function \psi(g) to biological priors, enabling control over whether growth increases or decreases along inferred trajectories, offering practical relevance in single-cell dynamics applications.

Weaknesses

  1. The paper enforces only first-order optimality conditions through the HJB residual loss, which are necessary but not sufficient for achieving true minimal action. There is no guarantee that the learned solution corresponds to a global or even local minimum of the action functional.

  2. While the method relies on a weighted particle system to approximate density evolution, it lacks any quantitative analysis of convergence rates or variance bounds for the empirical measure, especially under high-dimensional stochasticity.

  3. The method assumes that the mapping \lambda = \alpha \psi'(g) is invertible across all relevant values of \lambda, but this is not always true for non-convex or saturating forms of \psi. No mechanism is introduced to ensure this invertibility is respected during training.

问题

  1. Your method relies on enforcing first-order necessary conditions for action minimization via the HJB loss. However, these conditions are not sufficient for global or even local optimality. Can you provide any theoretical or empirical evidence that minimizing the HJB residual leads to low-action trajectories beyond satisfying a PDE constraint pointwise?

  2. The derivation assumes that the mapping λ=αψ(g)\lambda = \alpha \psi'(g) can be inverted to recover g(x,t)g(x,t). What guarantees do you have that λθ(x,t)\lambda_\theta(x,t) stays within the image of ψ\psi' during training? How is the domain of gg enforced or regularized when ψ\psi is non-convex or has flat regions?

  3. In lower-dimensional or isotropic settings, optimal transport paths with minimal action (e.g., in the WFR or Benamou-Brenier case) are analytically or numerically solvable. Why did you not include comparisons against such ground-truth geodesics to rigorously evaluate whether Var-RUOT truly finds lower-action trajectories?

  4. What is the quantitative effect of violating the HJB equation on the evolution of ρ(x,t)\rho(x,t)? Can small residual errors compound over time and distort the learned transport path, particularly in extrapolation tasks?

  5. The biological interpretation of ψ(g)\psi''(g) controlling the direction of growth is intuitive but lacks formal backing. Why is ψ(g)=g2/15\psi(g) = g^{2/15} considered a principled choice beyond its curvature sign? Have you validated this with any empirical or theoretical biological models?

  6. You claim that learning a single scalar field simplifies the problem. However, could this choice overly constrain the model class, especially in settings where velocity and growth cannot be jointly derived from a single potential? Are there situations where this coupling leads to underfitting?

局限性

Yes

最终评判理由

I believe this is a valuable paper and authors addressed my concerns thoroughly, therefore I increased my score to 5.

格式问题

N/A

作者回复

Thank you for carefully reviewing our paper and for providing valuable comments. Below please find our point-by-point response:

Summary & Strengths

We appreciate your recognition of our work's strengths. Our paper derives the first-order optimality conditions for the RUOT problem and solves it by fitting a single scalar field, thereby modeling unbalanced distributions and stochastic dynamics. Compared to previous methods, our Var-RUOT finds paths with lower action, achieves stable training, and converges faster. We also discuss the biological priors associated with the RUOT penalty function.

[W1] Regarding Var-RUOT’s Guarantee for the Minimal Action Solution

Thank you for your comment. Since Var-RUOT is a deep learning–based RUOT solver, it guarantees only a local minimum of the action. Empirically, experiments (Table 1,2,4) along with all additional results in the rebuttal show that Var-RUOT achieves a lower action than DeepRUOT, and its Path Action is comparable to that of traditional OT solvers.

We consider guaranteeing a global minimum an interesting direction for future work and will discuss it in the revised manuscript. We plan to start with analytical solutions for simple cases (e.g. Dirac-to-Dirac transport) and then extend to real data by treating complex distributions as a superposition of Dirac deltas and solving the associated velocity field via a Flow Matching–like approach.

[W2] Convergence Rate

Thanks for your insightful question. From the perspective of numerical integration, we use the particle method to estimate LHJBL_{\text{HJB}} and LActionL_{\text{Action}}, which is equivalent to performing integration via Monte Carlo methods. Let the number of particles be NN. Then, for any smooth function f(x)f(\boldsymbol{x}) (In the Action Loss and HJB Loss, the integrand.), the central limit theorem tells us that the error of the integral f(x)dμN\int f(\boldsymbol{x})\,\mathrm{d}\mu^N is of order O(1N)\mathcal{O}\left(\dfrac{1}{\sqrt{N}}\right). We will revise the paper and include this discussion on the convergence rate. In practical applications of the Var-RUOT algorithm, our default setting is N=512N=512, which ensures that the estimates of both loss have very small variance.

[Q1, Q4] Regarding the effect of the HJB Loss.

Thanks for the insightful question. Var-RUOT controls the penalty for λ(x,t)\lambda(\boldsymbol{x},t) violating the HJB equation by adjusting the weight γHJB\gamma_{\text{HJB}}: a higher weight increases the importance of the HJB loss, reducing the violation after training. Sensitivity analysis (Section C.2, Figure 7) shows that as γHJB\gamma_{\text{HJB}} increases, the Path Action decreases, proving that a lower HJB loss leads to a smaller action solution. To address the concerns, we further provide sensitivity results for the EMT dataset and additional results for the 2D Mouse dataset:

  • EMT Dataset
γHJB\gamma_{\text{HJB}}006.25×1036.25 \times 10^{-3}3.125×1023.125 \times 10^{-2}6.25×1026.25 \times 10^{-2}6.25×1016.25 \times 10^{-1}3.1253.125
LHJB\mathcal{L}_{\text{HJB}}1.90161.30910.62050.36850.02230.0006
Path Action0.46120.45840.38550.35440.29750.2808
  • 2D Mouse Dataset
γHJB\gamma_{\text{HJB}}006.25×1036.25 \times 10^{-3}3.125×1023.125 \times 10^{-2}6.25×1026.25 \times 10^{-2}6.25×1016.25 \times 10^{-1}3.1253.125
LHJB\mathcal{L}_{\text{HJB}}1.91071.46051.07450.68190.07250.0074
Path Action3.30023.21653.23973.14913.11823.1200

These results confirm that incorporating the HJB loss effectively finds lower action solutions. Based on our findings, we recommend setting γHJB\gamma_{\text{HJB}} between 10210^{-2} and 10110^{-1} to balance minimal action with data reconstruction quality. We will include these discussion in our final version.

[W3,Q2] Reversibility Issue of ψ(g)\psi'(g)

Thank you for your valuable comment and apologize for the insufficient discussions in previous paper. Here we discuss the two scenarios of ψ\psi selection in more details:

  • ψ1(g)=g2p\psi_{1}(g)=g^{2p}, ψ1(g)=2pg2p1\psi_{1}'(g)=2p\cdot g^{2p-1}, pZ+p\in\mathbb{Z}^{+}. Here, ψ1(g)\psi_{1}'(g) is a bijection from R\mathbb{R} to R\mathbb{R}, so gg is uniquely determined from λ=αψ1(g)\lambda=\alpha\psi_{1}'(g).

  • ψ2(g)=g2p2q+1\psi_{2}(g)=g^{\frac{2p}{2q+1}}, ψ2(g)=2p2q+1g2p(2q+1)2q+1\psi_{2}'(g)=\frac{2p}{2q+1}g^{\frac{2p-(2q+1)}{2q+1}}, p,qZ+p,q\in\mathbb{Z}^{+} with 2q+1>2p2q+1>2p. Here, ψ2(g)\psi_{2}(g) is a bijection from R{0}\mathbb{R}\setminus\{0\} to R{0}\mathbb{R}\setminus\{0\} (undefined at 0). Directly solvingλ=αψ2(g)\lambda=\alpha\psi_{2}'(g) yields

g=(αλ2q+12p)2q+1(2q+1)2p.g=\left(\frac{\alpha}{\lambda}\frac{2q+1}{2p}\right)^{\frac{2q+1}{(2q+1)-2p}}.

At λ=0\lambda=0, g(λ)g(\lambda) is undefined. To handle this inevitable singularity, we set a small δ\delta and linearly interpolate between g(δ)g(-\delta) and g(δ)g(\delta) for λ[ϵ,ϵ)\lambda\in[-\epsilon,\epsilon),i.e., g(λ)=12δ(g(δ)g(δ))(λ+δ)+g(δ)g(\lambda) = \dfrac{1}{2\delta}\left(g(\delta) - g(-\delta)\right)(\lambda + \delta) + g(-\delta) (Appendix A.2)

Another approach is to fit the reciprocal of λ\lambda, denoted as μ\mu, thereby moving the singularity at λ=0\lambda=0 to μ\mu \rightarrow \infty. In practice, μ\mu \rightarrow \infty does not occur, just as the singularity where g|g| \rightarrow \infty is not reached. This method avoids artificially modifying the g(λ)g(\lambda) relationship and is more reasonable. This will be our future work. We will update the paper for more thorough disucssions on ψ(g)\psi'(g) choice and include these detailed analysis in the appendix.

[Q3] Comparison with Traditional OT Solvers

Thank you for your valuable comment. In additional experiment, here we built two 2D normalized and unbalanced datasets to test Var-RUOT’s ability to find minimal action paths. For normalized data, we used the pot library to obtain ground truth; for unbalanced data, we applied a Sinkhorn-based static WFR solver [(Wang Z., et al.) Wasserstein-fisher-rao document distance]. With σ=0\sigma=0, Var-RUOT achieves similar—or even lower—Path Action values than traditional solvers(results are shown below), demonstrating its effectiveness. We will include these analysis in the appendix.

DatasetNormalizedNormalizedUnbalancedUnbalanced
-Simple GaussianGaussian MixtureSimple GaussianGaussian Mixture
Path Action (Traditional Solver)0.50461.90621.37762.4464
Path Action (Var-RUOT)0.49981.87541.01022.3909

[Q4] Extrapolation Performance

We evaluated Var-RUOT’s extrapolation capability and error propagation by extending the three-gene simulation dataset to nine time points. The first five were used for training, while the remaining four tested extrapolation. We compared TIGON, DeepRUOT, and Var-RUOT. The results are shown below:

Method/W1\mathcal{W}_{1} Distancet=5t=6t=7t=8
TIGON0.19320.34370.52090.6913
DeepRUOT0.10270.19400.33590.5047
VarRUOT0.10570.18060.28900.4260

Var-RUOT outperforms TIGON and performs similarly to DeepRUOT in long-term extrapolation. However, as time increases, error grows because the model’s non-autonomous system learns λ(x,t)\lambda(\boldsymbol{x}, t) only within the training window. Beyond this range, the tt-dependency is less accurate, leading to larger errors. Nonetheless, this non-autonomous system improves flexibility and helps prevent underfitting (see Q6). We will include these analysis in the appendix.

[Q5] Choice of ψ(g)=g2/15\psi(g) = g^{2/15}

Sorry for the confusion. We clarify that choosing ψ(g)=g2/15\psi(g)=g^{2/15} primarily to showcase an example with ψ(g)<0\psi''(g)<0. There are two reasons.

Firstly, the simplest function family satisfying our conditions (lines 224–229)—that is, dψ(g)dg>0\frac{d\psi(g)}{d|g|}>0, ψ(g)=ψ(g)\psi(g)=\psi(-g), and ψ(g)<0\psi''(g)<0—is given by

ψ(g)=g2p2q+1,p,qZ+,  2p<2q+1.\psi(g)=g^{\frac{2p}{2q+1}},\quad p,q\in\mathbb{Z}^+,\; 2p<2q+1.

Secondly, for training stability, the first-order condition yields

g(λ)=(λα2q+12p)2q+12p(2q+1)(1λ)2q+1(2q+1)2p.g(\lambda)=\left(\frac{\lambda}{\alpha}\,\frac{2q+1}{2p}\right)^{\frac{2q+1}{2p-(2q+1)}} \propto \left(\frac{1}{\lambda}\right)^{\frac{2q+1}{(2q+1)-2p}}.

To avoid a steep g(λ)g(\lambda) curve near λ=0\lambda=0, we choose the smallest pp (i.e., p=1p=1) and a large qq (here, q=7q=7), which leads to ψ(g)=g2/15\psi(g)=g^{2/15}.

Here we conduct additional experiments on the 2D Mouse Blood Hematopoiesis dataset (results are shown below) to show that while ψ(g)=g2/15\psi(g)=g^{2/15} and ψ(g)=g2/21\psi(g)=g^{2/21} perform similarly, however ψ(g)=g2/9\psi(g)=g^{2/9} yields unstable training and poorer reconstruction.

ψ(g)\psi(g) / W1\mathcal{W}_1 Distancet=1t=2
ψ(g)=g(2/9)(p=1,q=4)\psi(g)=g^{(2/9)}(p=1, q=4)0.94761.2197
ψ(g)=g(2/15)(p=1,q=7)\psi(g)=g^{(2/15)}(p=1, q=7)0.29530.1917
ψ(g)=g(2/21)(p=1,q=10)\psi(g)=g^{(2/21)}(p=1, q=10)0.28580.2565

Overall, our experiments confirm that, within a reasonable range, different choices for the exponent on gg are feasible. We will include this sensitivity analysis in the appendix.

[Q6] On Whether a Single Scalar Field Causes Underfitting

Thanks for the insightful question. Here we clarify that:

  • Experiments show that Var-RUOT reconstructs the marginal distributions at all time points well—both on low-dimensional data (Sections 6.1 and 6.3) and high-dimensional data (Appendix C.4), the results are comparable to DeepRUOT (which uses three networks) without underfitting. Please also refer to the responses to Reviewer kRW9 [Q1] and Reviewer YDV3 [W4,5,6], which include quantitative results of Var-RUOT on high-dimensional datasets.
  • Allowing λ\lambda to depend on both x\boldsymbol{x} and tt enables capturing both trajectory evolution and total mass fluctuations; if λ\lambda depended solely on x\boldsymbol{x}, it could not fit distributions showing some specific mass changing pattern (e.g., mass that first rises then falls).
  • Moreover, the DeepRUOT method employs three networks, while they are interdependent (via a Fokker–Planck loss), which makes training unstable and requires pretraining, while Var-RUOT is simpler and more stable.
评论

Dear Reviewer s4d8,

Thank you once again for your valuable comments and suggestions. We have addressed each of your points in detail and hope that our responses have adequately resolved your concerns. As the discussion period is nearing its end, please do not hesitate to let us know if you have any further questions or concerns. We are more than willing to provide additional clarifications if needed.

Best regards,
The Authors

评论

I thank the authors for their comprehensive rebuttal. All of my questions have been answered satisfactorily, and I am satisfied with the responses. Therefore, I am increasing my score to 5.

评论

We sincerely appreciate the time and effort you dedicated to reviewing our paper and considering our responses. Your insightful and constructive feedback has been invaluable, and we are committed to incorporating your suggestions into the revised version of the manuscript.

Sincerely,
The Authors

审稿意见
5

This paper proposes Var-RUOT, a new framework for solving the Regularized Unbalanced Optimal Transport (RUOT) problem (Def. 4.1). The authors reformulate the RUOT problem as a weighted particle system (Theorem 5.1). By applying the Hamilton–Jacobi–Bellman (HJB) equation (Theorem 4.1), they derive an action-matching loss based on the value function. In addition, they introduce an HJB loss term to enforce necessary conditions and a reconstruction loss to ensure that the distribution generated by the model remains consistent with the real data distribution. The method is evaluated on synthetic datasets (Sec. 6.1) and applied to biological problems (Sec. 6.2, 6.3).

优缺点分析

  • The paper is well-organized and mathematically sound, with clear examples and intuitive explanations.

  • To the best of my knowledge, this is the first work to apply a weighted particle method in this context. This formulation provides valuable intuition and could inspire extensions to related problems. I believe this is an innovative contribution that merits attention from the NeurIPS community.

  • The method demonstrates an efficient application of RUOT compared to previous works. The experimental results consistently outperform prior approaches. Moreover, the combination of toy experiments for validation and real-world biological applications highlights both the technical soundness and practical potential of the approach.

  • However, there are some limitations. Estimating the value function solely through a PINN-like loss (via the HJB equation) may lead to inaccuracies, as solving PDEs is challenging and in this case requires high precision—since the control is derived from differentiating the value function. Even small errors in estimating the value surface could cause significant deviations. Additionally, while the action-matching loss is novel, it may have limited scalability, and the need to estimate second-order derivatives with respect to the state space adds further computational burden. Despite these limitations, I still believe the paper’s contributions are valuable.

问题

  • Does the method scale to high-dimensional settings (e.g., more than 100 dimensions) where the marginals are Gaussian Mixture Models with many modes? It seems challenging to generalize this approach to such complex, high-dimensional cases.

  • What hyperparameters require careful tuning? I imagine that the ratio between the loss function might be crucial. Could the authors elaborate on this aspect or provide additional ablation studies? Also, since estimating λ\lambda relies on accurately minimizing the HJB loss, could the authors include a training curve for the HJB loss? Furthermore, have the authors considered using an L1 loss instead, as it might yield sharper estimates of the value surface? Any comments on this would be appreciated.

局限性

Please refer to Strength and Weakness section.

最终评判理由

I initially thought this paper should be clearly accepted, and the authors addressed all of my concerns.

格式问题

.

作者回复

Thank you for carefully reviewing our paper and for providing valuable comments. Below please find our point-by-point response:

Summary & Strengths

We appreciate your recognition of our work's strengths. Our paper derives the first-order optimality conditions for the RUOT problem and solves it by fitting a single scalar field, thereby modeling unbalanced distributions and stochastic dynamics. Compared to previous methods, our Var-RUOT finds paths with lower action, achieves stable training, and converges faster. We also discuss the biological priors associated with the RUOT penalty function. With each action guiding its corresponding dynamics, we hope our framework offers a more physical approach to modeling single-cell omics data.

[W1] Relying solely on a PINN-like HJB loss to estimate the value function can produce significant control deviations, as even small errors cause large inaccuracies due to sensitive differentiation.

Thank you for your insightful question. We fully agree that since deep learning solver for the RUOT problem converts hard constraints into soft ones via the loss function, so there's no guarantee that the solution exactly satisfies the marginal constraints nor minimizes the action at every time point. Instead, the HJB loss serves as a regularizer that promotes a lower action value without necessarily reducing to 0. On the other hand, we point out that empirically, sensitivity analysis (lines 656–670) shows that increasing the loss weight weight γHJB\gamma_{\text{HJB}} would decrease the action value, indicating the effectiveness of such term to achieve the decrease of action for the comnputed path. We will add the relevant discussions in the revised manuscript.

[W2] Compute Second-order derivatives causes computational burden.

Thank you for your comments and suggestion. We fully agree that compute second-order derivatives using PyTorch autograd by default will requires retaining a larger graph and thus more memory and time than first-order derivatives. Meanwhile we clarify that

  • Unlike previous RUOT methods such as DeepRUOT that train three networks for v\boldsymbol{v}, ρ\rho, and gg, VAR-RUOT fits a single scalar field λ\lambda, reducing the search space, improving stability, and converging faster (see Section 6.2).

  • Since Var-RUOT only need second-order derivatives for the LHJB\mathcal{L}_{\text{HJB}} loss (to compute x2λ\nabla_x^2 \lambda), meaning we only require the trace of the Hessian matrix HH. We can estimate it using the Hutchinson method:

    trace(H)=Ev[vTHv],vN(0,I)\text{trace}(H)=\mathbb{E}_{\boldsymbol{v}}[\boldsymbol{v}^{T} H \boldsymbol{v}], \quad \boldsymbol{v}\sim N(0,I)

    which computes the derivative along one direction, reducing cost (error O(1/m)O(1/\sqrt{m}), with mm samples). We can choose to use Hutchinson early in training for speed and later switch to full autograd for accuracy. This feature will be made as an option in our open-source code to speedup the computation.

[Q1] Can Var-RUOT be Generalized to high-dimensional settings?

Thank you for your question on Var-RUOT's scalability to high-dimensional datasets. In previous manuscript, we validated its scalability on several datasets: 50 and 100 dimensional Gaussian Mixture datasets (lines 709–715), a 50-dimensional Mouse Blood Hematopoiesis dataset, and a 30-dimensional β-cell Differentiation dataset (lines 716–723). For the high-dimensional Gaussian Mixture dataset, Figure 13 qualitatively shows the learned particle trajectories and growth rates on gaussian datasets; Figures 14 and 15 display the learned velocity fields and growth rates for the real 50D and 30D datasets, demonstrating Var-RUOT's applicability.

To quantitatively evaluate scalability, we further designed simulated data at three time points (t = 0, 1, 2) where each point is generated from a mixture of three Gaussians in 150 dimensions. We then compared various baselines on this dataset, with results as follows:

Method/W1\mathcal{W}_1 Distancet=1t=2PathAction
SF2M7.2869.457-
PISDE6.0517.546-
MIOFlow6.4947.812-
TIGON6.1577.61513.4913
DeepRUOT6.2027.64813.5992
VarRUOT6.0727.54410.1033

Var-RUOT achieves comparable reconstruction accuracy to PISDE, slightly outperforms TIGON and DeepRUOT, and significantly outperforms other methods on the 150-dimensional Gaussian Mixture dataset. It also yields a significantly lower path action than TIGON/DeepRUOT, confirming its scalability in high-dimensional settings. We will include these experiments in our revised manuscript.

[Q2]Questions on hyperparameter tuning and ablation studies.

Thank you for your question on hyperparameter tuning. Our method requires tuning three hyperparameters: the WFR metric parameter α\alpha (controlling the penalty for mass growth/decay) and two loss parameters, γHJB\gamma_{\text{HJB}} and γAction\gamma_{\text{Action}}. In Appendix C.2, we show that our method is robust to variations in α\alpha (using the 2D Mouse Blood Hematopoiesis dataset, lines 648–655, Table 7). Similarly, sensitivity experiments on γHJB\gamma_{\text{HJB}} and γAction\gamma_{\text{Action}} (lines 656–670, Figures 7–8) demonstrate that increasing either parameter reduces the Path Action. However, very high values diminish the impact of the reconstruction loss that enforces marginal constraints, which degrades performance.

To provide further insights, here we have conducted additional sensitivity tests on γHJB\gamma_{\text{HJB}} and γAction\gamma_{\text{Action}} using the Mouse Blood Hematopoiesis dataset.

  • Case with fixed γAction=6.25×102\gamma_{\text{Action}} = 6.25\times {10}^{-2}:
γHJB\gamma_{\text{HJB}}006.25×1036.25\times10^{-3}3.125×1023.125\times10^{-2}6.25×1026.25\times10^{-2}6.25×1016.25\times10^{-1}3.1253.125
W1\mathcal{W}_{1}0.15730.16090.15500.13160.21090.2539
PathAction3.30023.21653.23973.14913.11823.1200
  • Case with fixed γHJB=6.25×102\gamma_{\text{HJB}}= 6.25\times {10}^{-2}:
γAction\gamma_{\text{Action}}006.25×1036.25\times10^{-3}3.125×1023.125\times10^{-2}6.25×1026.25\times10^{-2}6.25×1016.25\times10^{-1}3.1253.125
W1\mathcal{W}_{1}0.14540.13440.16740.13160.40010.7914
PathAction3.40433.44973.27373.14912.47161.2059

The results from these additional sensitivity experiments are consistent with those presented in the original manuscript: higher values of γHJB\gamma_{\text{HJB}} and γAction\gamma_{\text{Action}} lead to a lower Path Action, yet if they are set too high, the reconstruction quality deteriorates significantly. In fact, all experiments in the main text use the same parameters, γHJB=γAction=6.25×102\gamma_{\text{HJB}} = \gamma_{\text{Action}} = 6.25 \times 10^{-2} (see Table 6). In practice, we recommend setting both parameters in the range of 10210^{-2} to 10110^{-1} and we will make it clear in the open-source package.

[Q3]Training curve for the HJB loss.

Thank you for your question. Due to rebuttal restrictions, we cannot provide an image of the HJB loss curve. The HJB loss initially rises for 10–20 epochs (since LReconL_{\text{Recon}} dominates while parameters are near zero) and then gradually declines as LReconL_{\text{Recon}} reaches a similar scale as LHJBL_{\text{HJB}} and LActionL_{\text{Action}}. We will include this curve in the appendix of revised manuscript.

[Q4]Can we use L1 loss instead?

Thank you for your valuable suggestion. In additional experiment, we have modified the HJB loss per your recommendation by replacing the quadratic penalty with a linear penalty. Then we re-ran experiments on three datasets presented in the main text for the Var-RUOT method, comparing the performance when using the original L2L_{2} loss versus the L1L_{1} loss.

  • Three Gene Simulation Dataset:
Method/W1\mathcal{W}_1 Distancet=1t=2t=3t=4PathAct
Var-RUOT(L2L_2Loss)0.04520.03850.04450.05721.1105
Var-RUOT(L1L_1Loss)0.04890.05560.05890.06641.1249
  • EMT Dataset:
Method/W1\mathcal{W}_1 Distancet=1t=2t=3Path Action
Var-RUOT(L2L_{2}Loss)0.25400.26700.26830.3544
Var-RUOT(L1L_{1}Loss)0.26310.27960.28090.3528
  • 2D Mouse Blood Hematopoiesis Dataset:
Method/W1\mathcal{W}_1 Distancet=1t=2Path Action
Var-RUOT(L2L_{2}Loss)0.12000.14313.1491
Var-RUOT(L1L_{1}Loss)0.17470.16903.1889

The experimental results indicate that with our chosen hyperparameter values (γHJB=6.25×102\gamma_{\text{HJB}} = 6.25 \times 10^{-2} and γAction=6.25×102\gamma_{\text{Action}} = 6.25 \times 10^{-2}), using either the L1L_{1} loss or the L2L_{2} loss for the HJB loss yields similar reconstruction performance and Path Action. We will add the above analysis to the appendix during the revision of the paper.

评论

Thank you for your response. My concerns are fully addressed.

评论

Thank you once again for dedicating your time to review our paper and to read our responses. Your insightful and constructive comments have significantly enhanced our work, and we will certainly integrate your suggestions into the revised version of the paper.

Sincerely,

The Authors

最终决定

This paper presents Var-RUOT, a new approach to recover the dynamics from snapshot data, aproblem central in the computational biology literature. The core innovation is to directly incorporate the first-order optimality conditions (HJB) of the RUOT problem into the models parameterization and loss. This leads to significant simplifications over the previous methods.

The work received unanimous and positive recommendation from all four reviewers who were particularly impressed by the method's theoretically grounding, the performance, and a strong rebuttal that addressed all concerns from all reviewers. For these reasons I recommend accepting this work.