OPTIMAL TRANSPORT BARYCENTER VIA NONCONVEX CONCAVE MINIMAX OPTIMIZATION
摘要
评审与讨论
This paper uses a primal-dual formulation for the Wasserstein barycenter problem without any regularization. The method is performed by using Wasserstein gradinet descent for the optimization on the density, and for the Kantarovich potential, a H^1 metric optimization is used. The algorithm is a first order method which seems to quickly converge to the stationary point.
优点
-
The presentation is clear, especially on the optimization for the Wasserstein gradient.
-
The numerical experiments seem promising.
缺点
The reviewer first wishes to comment that the methodology seems an exciting and promising direction in OT and an illuminating extension of the back-and-forth OT algorithm. However, it is hard to take the writing at face value to know the validity of the result. While the reviewer mainly asks for clarification in the remainder of this part, the unclarity of the implementation details in itself is a major weakness.
The major drawback of this work is that the methodology is not clear from the paper, which is why currently it does not seem possible to assess the quality of this work. Specifically, here are some questions that need to be answered:
- How is the differentiation implemented in this work? Terms such as are not continuous and admits discrete proxies. Is it done by finite difference? And, if so, what is the scheme? Or, alternatively, is it done by performing FFT, and is some smoothing filter performed?
- How is the pushforward implemented in this work? Assuming is computed, the term (\nabla \phi)_{#} might exceed the grid specified. How is this going to be resolved in the numerical example? These are questions that need to be addressed within this work. As of now, the reviewer's understanding is that this work only quotes the Jacobs & Leger paper, but that would be insufficient.
- It is quite confusing as to why the authors use a 10241024 grid for Sinkhorn in Figure 1. From the plot, it is clear that the distributions occupy a small region of the grid. Is a smaller size used? This should a priori be mentioned in the numerical experiment section. If somehow the full 10241024 space is used, then it would be quite unfair to the Sinkhorn algorithm. Alternatively, does the code for WDHA use the sparsity of the source and target distribution?
- The comparison between DSB and the proposed method is still inconclusive. It may happen that numerical optimization in addition to a smaller regularization number may lead to better performance in the barycenter functional value and better wall-clock time. The reviewer is not sure if the POT code package result can be taken at face value.
Other major issues:
- The experiments primarily focus on Algorithm 4, but the analysis is done for Algorithm 3. The authors did not clearly show this intricacy in the contribution subsection.
- There does not seem to be an implementation for Algorithm 3.
Minor issues:
- This work does not sufficiently mention the limitation that it is only efficient for low-dimensional grids. Sinkhorn, in contrast, can work by discretization of continuous densities by empirical distributions. The omission to stress this limitation needs to be resolved.
问题
The questions are listed in the weakness section.
We just want to add the following points.
(iii) doesn't have a ground truth that can convince the readers of the validity of a plot that doesn't look good and has bad Sobolev norm
Ans: we added a new simulation in the appendix subsection C.1 which has gournd truth. The ground truth of this example is the uniform distribution on the disk of radius 0.15 centered at (0.5, 0.5), which is not smooth and has -norm (negative Sobolev norm) bounded by 1. Please kindly see our response to point 4.
(i) doesn't look good
Ans: In our newly added simulation example in the appendix of subsection C.1, the Wasserstein distance between the barycenter of our method and the truth is smaller than that of both CWB and DSB. The -distance between the density of barycenter of our method and the truth is also the smallest. We think that our barycenter look better visually.
In practice, one would pass the WSB result over a sharpening filter or pass the DSB result over a smoothing filter.
Ans: we indeed conducted an additional thresholding step for WSB and CSB in exmaple 1. Interested readers can check Figure 1 of the paper for the results.
We sincerely appreciate your time, thoughtful consideration, and constructive guidance on our paper. Following your and the other referees' valuable comments, we added three additional simulation studies in the appendix of the revised version to address your questions. If we have successfully addressed some of your concerns, we would kindly ask you to consider reflecting this in your rating. Thank you again for your invaluable feedback.
Below we provide point-by-point response to your comments in italics.
1. How is the differentiation implemented in this work? Terms such as are not continuous and admit discrete proxies. Is it done by finite difference? And, if so, what is the scheme? Or, alternatively, is it done by performing FFT, and is some smoothing filter performed?
Ans: you are right. We used finite difference approach. In particular, let and be the equally spaced grid points, we have and for . Given the evaluations of function on these grid points, we compute the gradient at point as
$
\nabla \varphi (x_i, y_j) = \left( \begin{array}{c} \frac{\varphi_{i,j} - \varphi_{i-1, j}}{h} \ \frac{\varphi_{i,j} - \varphi_{i, j-1}}{h} \end{array} \right),
$
where . No smoothing or filtering is performed for computing the gradient. FFT is used to solve the Poisson equation () with zero Neumann boundary condition.
2. How is the pushforward implemented in this work? Assuming is computed, the term (\nabla \phi)\_{\\#} might exceed the grid specified. How is this going to be resolved in the numerical example? These are questions that need to be addressed within this work. As of now, the reviewer's understanding is that this work only quotes the Jacobs & Leger paper, but that would be insufficient.
Ans: To discuss the implementation of (\nabla \varphi)\_{\\#} \nu, we describe how the mass (density value of at a point ) is splitted and mapped \citep{jacobs2020fast} as follows. Since is convex, we observe that and . Let be the quadrilateral formed by 4 points and pick the mesh grids , where
with . The mass of is first uniformly distributed to the meshed grids . Then, the mass at each point is distributed to 4 nearest grid points in , inversely proportional to their distances. If (\nabla \varphi)\_{\\#} \nu exceeds the grid specified, the mass will be distributed to the boundary points instead. The discussions about computing , and (\nabla \phi)_{\\#} \nu are added to appendix subsection D.2 of the revised version.
3. It is quite confusing as to why the authors use a 1024 1024 grid for Sinkhorn in Figure 1. From the plot, it is clear that the distributions occupy a small region of the grid. Is a smaller size used? This should a priori be mentioned in the numerical experiment section. If somehow the full 1024 1024 space is used, then it would be quite unfair to the Sinkhorn algorithm. Alternatively, does the code for WDHA use the sparsity of the source and target distribution?
Ans: our method is advantageous for high-resolution densities. The is the standard pixel size for images and has nothing in particular. Our method wins on images (demonstrated by example 2) as well. All the methods are compared on the full grid and are fed with the same densities, and thus the comparison is fair. In example 1, we put four shapes in the corners solely for visualization purpose, we can enlarge and center the shapes, the results would be the same. Example 2 would correspond to the case that the distributions occupy large regions around the center.
4. The comparison between DSB and the proposed method is still inconclusive. It may happen that numerical optimization in addition to a smaller regularization number may lead to better performance in the barycenter functional value and better wall-clock time. The reviewer is not sure if the POT code package result can be taken at face value.
Ans: We added a new simulation in the revised version for which the ground truth is known. We emphasize that regularization parameters reg= 0.003 and reg=0.005 are the optimal choices for CWB and DSB respectively. Smaller regularization parameters lead nonconvergent and worse results for both CWB and DSB, please see the table below and Figure 3 of the revised version. With known ground truth, we can compare directly the Wasserstein distance between computed barycenter and the truth, and -distance between the computed barycenter density and the true density. This simulation demonstrates that both CWB and DSB may output blurry barycenters, while our result is visually no different from the truth.
In this new simulation, the goal is to compute the barycenter of four uniform distributions supported on round disks of radius 0.15, centered at , , , respectively. It's clear that the true barycenter is uniform on the disk of radius 0.15 centered at . The computed barycenter densities by WDHA, CWB and DSB are shown in Figure 3 of the revised version. We report in the following table the Wasserstein distance between computed barycenter distribution to the truth, -distance between computed barycenter densities and the true density, and the barycenter functional value. Our method is uniformly the best, and in particular, the improvement in the Wasserstein distance is of orders of magnitude.
| reg | Wasserstein distance | -distance | ||
|---|---|---|---|---|
| WDHA | 0.4869 | |||
| CWB | 0.003 | 1.321 | ||
| CWB | 0.002 | 3.041 | 0.1154 | |
| DSB | 0.005 | 0.8219 | ||
| DSB | 0.004. | . | 2.281 | 0.1016 |
We mark that the barycenter functional value of the truth is 0.09. This simulation is included in the appendix subsection C.1 of the revised version. Our code is attached for reproducibility check.
5. The experiments primarily focus on Algorithm 4, but the analysis is done for Algorithm 3. The authors did not clearly show this intricacy in the contribution subsection. There does not seem to be an implementation for Algorithm 3.
Ans: Thank you. Empirically, the projection can be computed by the approach in [Simonetto, 2021], which is implemented as the python function reg4opt.regression.convex\_regression in library reg4opt. Unfortunately, this requires solving a large convex quadratically constrained quadratic problem, and the code could take weeks to run.
Instead, we implement the algorithm with projection (Algorithm 3) and test the method on 1D distributions. For each repetition and , we let be the truncated version of on the domain , where and . Let , , be the true barycenter, computed barycernter from Algorithm 3 (with projection onto ), computed barycernter from Algorithm 4 (with double convex conjugates) respectively. We repeat the experiment 300 times and report two types of average distances between the groundtruth and each computed barycenters : average -distance and average -distance between densities . Algorithm 3 performs slightly better with -distance and -distance 0.608, while Algorithm 4 has -distance and -distance 0.6434. We have included this study in subsection C.2 in the appendix of the revised version.
We remark that the same issue also exists in other works that utilize the -gradient if any convergence theories are to be developed; see the assumptions in proposition 1 [Jacobs and L´eger, 2020] and Theorem 2.8 [Jacobs et al., 2021]. We don't think this should be the reason to deny its satisfying empirical performance either. Please check our additional simulation with ground truth. Without additional regularities, generalization of the Theorem can be challenging, the main obstacle is that the map is not Lipschitz, see [Berman, 2021] for more details regarding this.
The reviewer thanks the author for the helpful clarification. The reviewer has decided to maintain the score on the basis of two significant weaknesses: (A) The structure of the manuscript focuses too much on the less practical Algorithm 3, and the Algorithm 4 is less developed as it should be, and (B) The numerical is not as strong as that of Jacob & Leger. For (A), the detailed comment is that this factor is a major issue in the presentation, as the readers would read into Section 4 thinking that the Algorithm has a convergence guarantee, but in fact the presented algorithm lacks so. For (B), the main issue is that the WDHA Wasserstein barycenter does look worse than the WSB, and it looks worse in the sense that the barycenter looks quite nonsmooth, which begs the question of if the wasserstein gradient outer loop proposed in this work is sound, or if better algorithms can be used. In practice, one would pass the WSB result over a sharpening filter or pass the DSB result over a smoothing filter. The issue is that the WDHA result (i) doesn't look good, (ii) the will have a Soblev norm that is excessively large, and (iii) doesn't have a ground truth that can convince the readers of the validity of a plot that doesn't look good and has bad Sobolev norm.
The reviewer would have given the work a score of 4 were it an option, while the reviewer thinks the omission of technical details were probably minor omissions and giving a score of 3 would be quite off in the reviewer's assessment, which was the reasoning for choosing 5. After the clarification, the reviewer thinks the work warrants a 5, but higher than that grade would not be possible due to issues (A) and (B). Fixing either (A) or (B) in the reviewer's opinion could very well lead to a 6, and fixing both will make the work very good. The reviewer thinks the work is nevertheless quite promising and wishes the authors the best of luck.
This paper introduces a new algorithm called the "Wasserstein-Descent H-Ascent (WDHA) algorithm" for computing the optimal transport barycenter. Considering the dual form of the original problem, this paper transforms the barycenter problem as a nonconvex-concave minimax optimization problem, and then propose a gradient algorithm for optimization. One advantage of the proposed WDHA algorithm is its runtime complexity, which grows linearly with the grid size , while the naive method computing the optimal transport map requires time complexity of . Moreover, the under certain assumptions, the paper also shows the convergence rate of the algorithm to its stationary point. Finally, numerical studies are provided to demonstrate the effectiveness of the method.
优点
-
Algorithm: The paper introduces the WDHA algorithm by considering the dual form of the problem of the optimal transport barycenter and using a nonconvex-concave minimax optimization procedure, which is novel and interesting. Moreover, The proposed WDHA algorithm achieves nearly linear runtime complexity, improving the complexity of the naive approach.
-
Theory: Under certain assumptions, the authors provide a convergence rate of for the WDHA algorithm, which matches the convergence rate as in the Euclidean nonconvex-concave optimization problems.
-
Numerical study: The paper includes numerical studies that demonstrate the effectiveness of the WDHA algorithm over existing methods, showing that the proposed algorithm runs faster than the existing methods.
缺点
-
Theory: One weakness is that the theoretical analysis is performed for Algorithm 3, while the practical implementation uses Algorithm 4. Also, the theorem assumes lower bounded densities of the distributions, which may not be true in practice (for example, in the case of Figure 1). Even in this ideal setting, there is also an extra assumption on the uniform boundedness of . Therefore, the convergence guarantee of the algorithm in practice remains unclear.
-
Numerical Studies: While the numerical studies show faster computation times for the proposed method, I think the current results are not sufficiently convincing to demonstrate its superiority. For example, the "barycenter functional values" reported in line 515 differ very slightly across differet methods, and it seems to me that in the figures the DSB method performs generally well. More quantitative and comprehensive experiments should be conducted to validate the claims.
Moreover, I would like to ask the authors to provide the codes for reproducibility.
-
Comparison with Literature: A detailed discussion and comparison regarding the settings, runtime complexity, and theoretical guarantees with existing methods for optimal transport barycenter computation should be included for better demonstration.
问题
-
Can you provide theoretical/empirical insights for the statement in Line 317 "By embedding all densities functions into L2(Ω), we can alternatively use the L2-gradient to update ν. However, in simulations, the Wasserstein gradient update significantly outperforms the L2-gradient update."
-
Can you discuss the impact of the additional regularity by considering on the resulting barycenter accuracy, the convergence rate and computational cost? How these quantities are chosen in practice?
-
I would like the authors to discuss more on the effect of tuning the parameters in the numerical studies.
We sincerely appreciate your time, thoughtful consideration, and constructive guidance on our paper. Following your and the other referees' valuable comments, we added three additional simulation studies in the appendix of the revised version to address your questions. If we have successfully addressed some of your concerns, we would kindly ask you to consider reflecting this in your rating. Thank you again for your invaluable feedback.
Below we provide point-by-point response to your comments in italics.
1. Theory: One weakness is that the theoretical analysis is performed for Algorithm 3, while the practical implementation uses Algorithm 4. Also, the theorem assumes lower bounded densities of the distributions, which may not be true in practice (for example, in the case of Figure 1). Even in this ideal setting, there is also an extra assumption on the uniform boundedness of . Therefore, the convergence guarantee of the algorithm in practice remains unclear.
Ans: Thank you. The same issue also exists in other works that utilize the -gradient if any convergence theories are to be developed; see the assumptions in proposition 1 [Jacobs and L´eger, 2020] and Theorem 2.8 [Jacobs et al., 2021]. We don't think this should be the reason to deny its satisfying empirical performance either. Please check our additional simulation with ground truth. Without additional regularities, generalization of the Theorem can be challenging, the main obstacle is that the map is not Lipschitz, see [Berman, 2021] for more details regarding this.
Empirically, the projection can be computed by the approach in [Simonetto, 2021], which is implemented as the Python function reg4opt.regression.convex\_regression in library reg4opt. Unfortunately, this requires solving a large convex quadratically constrained quadratic problem, and the code could take weeks to run. Instead, we implement the algorithm with projection (Algorithm 3) and test the method on 1D distributions. For each repetition and , we let be the truncated version of on the domain , where and . Let , ,
be the true barycenter, computed barycernter from Algorithm 3 (with projection onto ), computed barycernter from Algorithm 4 (with double convex conjugates) respectively. We repeat the experiment 300 times and report two types of average distances between the groundtruth and each computed barycenters : average -distance and average -distance between densities . Algorithm 3 performs slightly better with -distance and -distance 0.608, while Algorithm 4 has -distance and -distance 0.6434. We have included this study in subsection C.2 in the appendix of the revised version.
In example 1, we put four shapes in the corners solely for visualization purpose, we can enlarge and center the shapes, the results would be the same. In addition, it's possible to separate the domains of the measures by following precedures in Chartrand et al. [2009] and considering the following formulation
where the support of is in , the suport of is in and . The convex conjugate of is defined to be for any . It's then possible to extend the results by using Proposition 2.1 and Lemma 2.2 in Chartrand et al. [2009]. Then, will be bounded only on .
We note that the boundedness assumption of is minor, because this can be easily checked in practice. If this assumption is violated, a warning message can be sent to the user.
2. Numerical Studies: While the numerical studies show faster computation times for the proposed method, I think the current results are not sufficiently convincing to demonstrate its superiority. For example, the ``barycenter functional values" reported in line 515 differ very slightly across different methods, and it seems to me that in the figures the DSB method performs generally well. More quantitative and comprehensive experiments should be conducted to validate the claims. Moreover, I would like to ask the authors to provide the codes for reproducibility.
Ans: that's a good point. We added a new simulation in the revised version for which the ground truth is known. So, we can compare directly the Wasserstein distance between the computed barycenter and the truth, and -distance between the computed barycenter density and the true density. This simulation demonstrates that both CWB and DSB may output blurry barycenters, while our result is visually no different from the truth.
In this new simulation, the goal is to compute the barycenter of four uniform distributions supported on round disks of radius 0.15, centered at , , , respectively. It's clear that the true barycenter is uniform on the disk of radius 0.15 centered at . The computed barycenter densities by WDHA, CWB and DSB are shown in Figure 3 of the revised version. We note that regularization parameters reg= 0.003 and reg=0.005 are the optimal choices for CWB and DSB respectively. Smaller regularization parameters lead nonconvergent and worse results for both CWB and DSB. We report in the following table the Wasserstein distance between computed barycenter distribution to the truth, -distance between computed barycenter densities and the true density, and the barycenter functional value. Our method is uniformly the best, and in particular, the improvement in the Wasserstein distance is of orders of magnitude.
| reg | Wasserstein distance | -distance | ||
|---|---|---|---|---|
| WDHA | 0.4869 | |||
| CWB | 0.003 | 1.321 | ||
| CWB | 0.002 | 3.041 | 0.1154 | |
| DSB | 0.005 | 0.8219 | ||
| DSB | 0.004. | . | 2.281 | 0.1016 |
We mark that the barycenter functional value of the truth is 0.09. This simulation is included in the appendix subsection c1 of the revised version. Our code is attached for reproducibility check.
3. Comparison with Literature: A detailed discussion and comparison regarding the settings, runtime complexity, and theoretical guarantees with existing methods for optimal transport barycenter computation should be included for better demonstration.
Ans: our method is advantageous for high-resolution densities, i.e., densities supported on extensive grid points. To our best knowledge, CWB and DSB are the only two methods are applicable to this setting, i.e., can have an output within reasonable time, and have available implementations. The runtime for example 1 and example 1 are reported in our paper. In addition, we have added another example with gourd truth. Please see our response for point 2. We are the first to use -gradient in the barycenter problem. The only theoretical guarantees of optimization algorithms that use -gradient are the ones developed in [Jacobs and L´eger, 2020, Jacobs et al., 2021], which are also based on similar regularities (smoothness and strong convexity) of potentials.
4. Can you provide theoretical/empirical insights for the statement in Line 317 ``By embedding all densities functions into , we can alternatively use the L2-gradient to update . However, in simulations, the Wasserstein gradient update significantly outperforms the L2-gradient update."
Ans: Certainly, we implemented this algorithm and tested it on uniform distributions supported on round disks for which the ground truth is known. Let , be density functions. We can write
Let be the space of functions that are -integrable with respect to the Lebesgue measure and be the set of density functions. Notice that the -gradient of with respect to is , we may consider the -descent -ascent algorithm: for each iteration , do
- For , compute
- Compute
Here, is the projection of onto , which can be computed using python function pyproximal.Simplex in library 1PyProximal. We apply this algorithm to the uniform distributions supported on round disks, and observe that it actually diverges leading to a wrong result. Nevertheless, we plot the output in Figure 4 of the revised version. This simulation is included in appendix subsection C.3 of the revised version.
5. Can you discuss the impact of the additional regularity by considering on the resulting barycenter accuracy, the convergence rate and computational cost? How these quantities are chosen in practice?
Ans: for the class of problems that the potentials between true barycenter and each are strongly convex and smooth, we expect the additional regularities to be redundant. Without this additional regularity, the inner maximization problem is not strongly concave and one may only show convergence up to -precision. This can be reflected by its Euclidean counterpart for nonconvex-concave problem, where we only have concavity, but not strong concavity, see section 4.2 of Lin et al. [2020]. Projecting the potential function to is expensive having time complexities [Simonetto, 2021], and we thus recommend replacing the projection step by double convex conjugates having time complexity .
6. I would like the authors to discuss more on the effect of tuning the parameters in the numerical studies.
Ans: we revised our discussion of parameters on page 420 in the old version to ``Theorem 1 suggests that the parameters , should be bounded above. Empirically, we find that the above algorithm works better with diminishing step sizes, potentially due to two reasons: (1) diminishing step sizes may reduce the effect of discrete approximations; and (2) diminishing step sizes may be more effective for nonsmooth convex functions. Since the second convex conjugate does not enforce strong convexity and smoothness, Lemma 1 no longer holds, and is only guaranteed to be concave. In addition, diminishing step size may speed up the learning process in the early stages and inverse time decay for , i.e., , works equally well in the simulation studies." in the revised version.
I would like to thank the authors for their detailed response and new experiments. I will raise the scores accordingly.
We sincerely appreciate your constructive comments and time!
In this paper, the authors propose a novel method to compute the unregularized Wasserstein barycenter of probability distributions discretized on compact sets. This setting is typically encountered in image settings, where one image corresponds to a probability distribution over a grid of pixels. In essence, their method consists in reformulating the Wasserstein barycenter problem as a min-max optimization problem (by relying on the dual formulation of the Kantorovitch transport problem), where the 'max'-based problem is concave while the 'min'-based problem is not convex. Then, they propose to solve this new problem via gradient ascent-descent approach, an iterative algorithm which alternates between a gradient descent step operating on the measure space (with Wasserstein gradient) and gradient ascent steps operating on the space of Kantorovitch dual potentials. Under reasonable assumptions on the input distributions, the authors provide convergence rate to a stationary point and iteration complexity of their algorithm named WDHA. Then, they apply their method to image settings (synthetic data + binarized MNIST) and compare WDHA to previous methods, which rely on regularization.
优点
- The paper is well written, the authors paying attention to define their mathematical notation and justifying carefully their computations (especially on gradients). A valuable effort is made on introducing elements in the introduction that will serve later for defining the theoretical framework of the method. In particular, the algorithm is clear to understand.
- The numerical experiments provide the intuition that the current method provide good quality of the barycenter.
缺点
- The choice of numerical setting of the main hyperparameters (step-sizes for ascent and descent) is not discussed.
- The burden of dimension is not discussed: in the current paper, only 2-dimensional settings are considered while Ot problems are generally of interest in larger scale. It is hard to see if the current method is scalable.
- The authors compare their methods to regularized approaches that enable to compute Wasserstein barycenters, but did not consider certain regularized methods that provide the exact same type of numerics (barycenter over images), see [1,2], without proper justification.
- Numerical results do not include uncertainty quantifications, which hurts the statistical significance of the results.
- The current paper lacks experiments where the actual barycenter is known (for example Gaussian experiment), which would enable to have a clear comparison with ground truth samples.
[1] Fast Computation of Wasserstein Barycenters. Cuturi and Doucet. 2014.
[2] Continuous Regularized Wasserstein Barycenters. Li et al. 2020.
问题
- How much critical is the compactness assumption on the input probability distributions for your method ?
- Assumptions in Theorem 1 suggest that the input distributions should have the same support (although it is not verified in the very first experiment). Could you improve this assumption to consider a case of distinct supports ?
- Could you explain what prevents the term to be arbitrarily large in the upper bound of Theorem 1, Which may hurt the convergence rate ?
- Could you detail how you efficiently implement the computation of the convex conjugate and the negative inverse Laplacian in the case of probability density functions discretized over a grid ?
- How do your method compare with methods mentioned in the section 'Weaknesses' ?
- What motivated to set the hyperparameters of your algorithm (in particular ) as you did ? Do you have guidelines for general use ?
- How do you compute the value function for competing methods ?
Suggestions
- The authors should mention in the introduction/related work the other common setting for Wasserstein barycenters related to generative modeling and deep learning, where the input probability distributions are not tractable and are represented by subset of samples. In the image setting, this would correspond to take one image as a sample from a high-dimensional distribution (not a distribution itself). A lot of recent work has been done in this direction recently.
- I think there is a typo at the end of page 5: it should be a max not a min.
We sincerely appreciate your time, thoughtful consideration, and constructive guidance on our paper. Following your and the other referees' valuable comments, we added three additional simulation studies in the appendix of the revised version to address your questions. If we have successfully addressed some of your concerns, we would kindly ask you to consider reflecting this in your rating. Thank you again for your invaluable feedback.
Below we provide point-by-point response to your comments in italics.
1. The choice of numerical setting of the main hyperparameters (step-sizes for ascent and descent) is not discussed. What motivated to set the hyperparameters of your algorithm (in particular ) as you did? Do you have guidelines for general use?
Ans: We revised our discussion of parameters on line 420 in the old version to ``Theorem 1 suggests that the parameters , should be bounded above. Empirically, we find that the above algorithm works better with diminishing step sizes, potentially due to two reasons: (1) diminishing step sizes may reduce the effect of discrete approximations; and (2) diminishing step sizes may be more effective for nonsmooth convex functions. Since the second convex conjugate does not enforce strong convexity and smoothness, Lemma 1 no longer holds, and is only guaranteed to be concave. In addition, diminishing step size may speed up the learning process in the early stages and a inverse time decay for , i.e., , works equally well in the simulation studies." in the revised version.
2. The burden of dimension is not discussed: in the current paper, only 2-dimensional settings are considered while Ot problems are generally of interest in larger scale. It is hard to see if the current method is scalable. The authors should mention in the introduction/related work the other common setting for Wasserstein barycenters related to generative modeling and deep learning, where the input probability distributions are not tractable and are represented by subset of samples. In the image setting, this would correspond to take one image as a sample from a high-dimensional distribution (not a distribution itself). A lot of recent work has been done in this direction recently.
Ans: Thank you for the comment. Using deep neural networks and generative models may be the only way to break the curse of dimensionality, we have added the following discussion in the revised paper on line 58 ``To break the curse of dimensionality, generative models have been investigated for the Wasserstein barycenter problem [Korotin et al., 2022]”". Feel free to let us know if we missed any relevant references.
3. The authors compare their methods to regularized approaches that enable to compute Wasserstein barycenters, but did not consider certain regularized methods that provide the exact same type of numerics (barycenter over images), see [1,2], without proper justification. How do your method compare with methods mentioned in the section ’Weaknesses’?
Ans: We note that [1] is an earlier work of Marco Cuturi who also co-authored CWB and DSB, which are both compared in the paper. DSB is an fixed support method that is more efficient than free support method; see [1] for this conclusion. We tried to add [2] as a comparison method. Unfortunatly, the code in (https://github.com/lingxiaoli94/CWB/tree/master) is not directly applicable to our examples, we thus provide the following justification on line 55 in the revised version ``Also, Li et al. [2020] propose a new dual formulation for the regularized Wasserstein barycenter problem such that discretizing the support is not needed.". Since both methods rely on entropic regularizations, the issue of having blurry output is still expected to exist.
[1] Fast Computation of Wasserstein Barycenters. Cuturi and Doucet. 2014. [2] Continuous Regularized Wasserstein Barycenters. Li et al. 2020.
4. Numerical results do not include uncertainty quantifications, which hurts the statistical significance of the results.
Ans: We added a new exmaple for 1D distributions to compare the algorithm with projection (Algorithm 3) and the algorithm with double convex conjugates (Algorithm 4). In this simulation, we repeat the experiments 300 times and report the average distance to address uncertainty quantification. Please see subsection C.2 in the appendix of the revised version for more details.
5. The current paper lacks experiments where the actual barycenter is known (for example Gaussian experiment), which would enable to have a clear comparison with ground truth samples.
Ans: That's a good point. We added a new simulation in the revised version for which the ground truth is known. This simulation demonstrates that both CWB and DSB may output blurry barycenters, while our result is visually no different from the truth.
In this new simulation, the goal is to compute the barycenter of four uniform distributions supported on round disks of radius 0.15, centered at , , , respectively. It's clear that the true barycenter is uniform on the disk of radius 0.15 centered at . The computed barycenter densities by WDHA, CWB and DSB are shown in Figure 3 of the revised version. We note that regularization parameters reg= 0.003 and reg=0.005 are the optimal choices for CWB and DSB respectively. Smaller regularization parameters lead to nonconvergent and worse results for both CWB and DSB. We report in the following table the Wasserstein distance between computed barycenter distribution to the truth, -distance between computed barycenter densities and the true density, and the barycenter functional value. Our method is uniformly the best, and in particular, the improvement in the Wasserstein distance is of orders of magnitude.
| reg | Wasserstein distance | -distance | ||
|---|---|---|---|---|
| WDHA | 0.4869 | |||
| CWB | 0.003 | 1.321 | ||
| CWB | 0.002 | 3.041 | 0.1154 | |
| DSB | 0.005 | 0.8219 | ||
| DSB | 0.004. | . | 2.281 | 0.1016 |
We mark that the barycenter functional value of the truth is 0.09. This simulation is included in the appendix subsection C.1 of the revised version.
6. How much critical is the compactness assumption on the input probability distributions for your method?
Ans: in practice, we can use finite grid points to approximate unbounded support, but this will incur some truncation losses.
7. Assumptions in Theorem 1 suggest that the input distributions should have the same support (although it is not verified in the very first experiment). Could you improve this assumption to consider a case of distinct supports?
Ans: in example 1, we put four shapes in the corners solely for visualization purpose, we can enlarge and center the shapes, the results would be the same. In addition, it's possible to separate the domains of the measures by following procedures in \cite{chartrand2009gradient} and considering the following formulation
where the support of is in , the suport of is in and . The convex conjugate of is defined to be for any . It's then possible to extend the results by using Proposition 2.1 and Lemma 2.2 in Chartrand et al. [2009].
8. Could you explain what prevents the term to be arbitrarily large in the upper bound of Theorem 1, Which may hurt the convergence rate?
Ans: this is true because all measures are assumed to be bounded above and have compact supports. Thus, the Wasserstein distance between and is uniformly bounded for all .
9. Could you detail how you efficiently implement the computation of the convex conjugate and the negative inverse Laplacian in the case of probability density functions discretized over a grid?
Ans: we note that for the 1D case, convex conjugate can be computed efficiently using the method in Corrias [1996]. For the 2D case, notice that
Thus, the convex conjugate in the 2D case can be computed by iteratively applying the 1D solver to each row and column. Poisson's equation () with zero Neumann boundary conditions is classical and can be solved efficiently using fast Fourier transform, and we refer to chapter 3 of [Taylor, 1996] for detailed derivations of using Fourier transform to solve Poisson's equations. We have added more details on the computation of , and (\nabla \varphi)\_{\\#} \nu in the appendix subsection D.2 of the revised version
10. How do you compute the value function for competing methods?
Ans: once we have an estimated barycenter , we apply the back-and-forth approach Jacobs and L´eger [2020] to compute the Wasserstein distance between and each .
References
- Robert J Berman. Convergence rates for discretized Monge–Amp`ere equations and quantitative stability of optimal transport. Foundations of Computational Mathematics, 21(4):1099–1140, 2021.
- Rick Chartrand, Brendt Wohlberg, Kevin Vixie, and Erik Bollt. A gradient descent solution to the monge- kantorovich problem. Applied Mathematical Sciences, 3(22):1071–1080, 2009.
- Lucilla Corrias. Fast legendre–fenchel transform and applications to hamilton–jacobi equations and conser- vation laws. SIAM journal on numerical analysis, 33(4):1534–1558, 1996.
- Matt Jacobs and Flavien L´eger. A fast approach to optimal transport: The back-and-forth method. Nu- merische Mathematik, 146(3):513–544, 2020.
- Matt Jacobs, Wonjun Lee, and Flavien L´eger. The back-and-forth method for wasserstein gradient flows. ESAIM: Control, Optimisation and Calculus of Variations, 27:28, 2021.
- Alexander Korotin, Vage Egiazarian, Lingxiao Li, and Evgeny Burnaev. Wasserstein iterative networks for barycenter estimation. In Advances in Neural Information Processing Systems, volume 35, pages 15672–15686. Curran Associates, Inc., 2022.
- Lingxiao Li, Aude Genevay, Mikhail Yurochkin, and Justin M Solomon. Continuous regularized wasserstein barycenters. Advances in Neural Information Processing Systems, 33:17755–17765, 2020.
- Tianyi Lin, Chi Jin, and Michael Jordan. On gradient descent ascent for nonconvex-concave minimax problems. In International Conference on Machine Learning, pages 6083–6093, 2020.
- Andrea Simonetto. Smooth strongly convex regression. In 2020 28th European Signal Processing Conference, pages 2130–2134, 2021.
- Michael Eugene Taylor. Partial differential equations. 1, Basic theory. Springer, 1996.
Thank you for the precise answers. I have some remarks regarding the response:
-
About the burden of dimension: I would suggest the authors to insist that their method is only applicable to 2D settings with compact support (unless numerical evidence of the contrary), I don't think this is highlighted enough. Since the direction of the OT community tends to consider high-dimensional problems (with neural networks as mentioned by the authors), this has to be clearly stated either in the abstract/setting or as a limitation of the curent approach. In particular, it is not precised by the authors how the computation of the convex conjugate may be done in dimension ...
-
About the related work: I would suggest to also include [1] in the discussion.
-
About the uncertainty quantification: sorry if I was unclear, but displaying the average performance of the competing methods does not sufficiently reflect this uncertainty; the standard deviations should at least appear as it commonly expected in ML papers, see [2] for instance. This should be done each time a metric is displayed, ie, including the experiments of the main paper (for instance, the value functions are really close to each other in the first experiment) . Do you have these values ? Moreover, why did you not include the other methods for this 1D experiment ? Could you do more than 1d while still having access to the ground truth ?
-
About the additional experiment: thank you for this additional result, where a ground truth is available and which demonstrates the superiority of the current method. Nonetheless, I wonder why the authors chose this uncommon setting (where each input measure has the same structure, and therefore may be too 'toyish'), while most of papers consider a Gaussian setting, also proposed by Reviewer eGnH, which may bring more complexity, since the covariances may be different for each input measure, and may also enable to scale in dimension ?
[1] Continuous Wasserstein-2 Barycenter Estimation without Minimax Optimization. Korotin et al. 2021
[2] Continuous Regularized Wasserstein Barycenters. Li et al. 2020.
We want to thank reviewer yaTH for your consideration and quick responses.
(a) About the burden of dimension: I would suggest the authors to insist that their method is only applicable to 2D settings with compact support (unless numerical evidence of the contrary), I don't think this is highlighted enough. Since the direction of the OT community tends to consider high-dimensional problems (with neural networks as mentioned by the authors), this has to be clearly stated either in the abstract/setting or as a limitation of the curent approach.
Ans: thank you, our methods porvide more accurate computation for the low-dimensional case. The computation burden, to a large extent, depends on the total number of grid points. (Jacobs & L´eger,2020) used spatial grids as large as (2D) and (3D). We added right after stating the contributions on line 95 of the revised version: ``For limitations, the current approach is mainly limited to computing the Wasserstein barycenter of 2D or 3D distributions supported on a compact domain".
(b) In particular, it is not precised by the authors how the computation of the convex conjugate may be done in dimension .
Ans: we note that the procedure we described in response to point 9 is generalizable to arbitraty dimensions as long as we have a seperable cost function. We illustrate using the 3D case. The cost function for our case is seperable as . Then,
= \sup\_{x\_1, x\_2} \left( (x\_1-y\_1)^2/2 + (x\_2-y\_2)^2/2 + \sup\_{x\_3} \left\\{ (x\_3-y\_3)^2/2 - \varphi(x\_1, x\_2, x\_3) \right\\} \right) = \sup\_{x\_1} \left( (x\_1-y\_1)^2/2 + \sup\_{x\_2} \left\\{ (x\_2-y\_2)^2/2+ [\varphi(x\_1, x\_2, \cdot )]^{*}(y\_3) \right\\} \right)which can be computed by applying the 1D solver to each dimension iteratively.
(c) About the related work: I would suggest to also include [1] in the discussion.
Ans: we apologize for missing this reference of scalable algorihtms, and have added ``To break the curse of dimensionality, scalable algorithms using input convex neural networks \citep{korotin2021continuous} and generative models \citep{NEURIPS2022_6489f2c6} have been investigated for the Wasserstein barycenter problem" on line 59 of the revised version.
We want to thank reviewer yaTH for your consideration and quick responses.
(d) About the uncertainty quantification: sorry if I was unclear, but displaying the average performance of the competing methods does not sufficiently reflect this uncertainty; the standard deviations should at least appear as it commonly expected in ML papers, see [2] for instance. This should be done each time a metric is displayed, ie, including the experiments of the main paper (for instance, the value functions are really close to each other in the first experiment) . Do you have these values ? Moreover, why did you not include the other methods for this 1D experiment ? Could you do more than 1d while still having access to the ground truth?
Ans: Thank you. We added the standard deviations and included the results of CWB and DSB. We revised the writing in appendix subsection C.1 to ``Algorithm 3 (with ) performs slightly better with -distance (standard deviation: ) and -distance 0.608 (standard deviation: ), while Algorithm 4 has -distance (standard deviation: ) and -distance 0.6434 (standard deviation: ). In addition, CWB has -distance (standard deviation: ) and -distance 0.082 (standard deviation: ). DSB achieves -distance (standard deviation: ) and -distance 0.0699 (standard deviation: ). DSB performs the best in Wasserstein distance for this example". We didn't include CWB and DSB initially, because our primary goal is to demonstrate that Algorithms 3 and 4 perform similarly. For 1D distributions, the true barycenter can be easily computed by averaging the quantile functions, while the barycenter of 2D distributions with compact support are generally unknown, please also refer our response to point (e) below for the 2D Gaussian case.
(e) About the additional experiment: thank you for this additional result, where a ground truth is available and which demonstrates the superiority of the current method. Nonetheless, I wonder why the authors chose this uncommon setting (where each input measure has the same structure, and therefore may be too 'toyish'), while most of papers consider a Gaussian setting, also proposed by Reviewer eGnH, which may bring more complexity, since the covariances may be different for each input measure, and may also enable to scale in dimension ?
Ans: Our method works on distributions with compact support, because it needs to discretize the support and operates on the grid points. Technically, we first need to truncate Gaussian distributions to have compact support. The true barycenter of truncated Gaussian distributions are actually unknown.
Thank you for pursuing the discussion and adding the missing details. I am willing to increase my score regarding the changes that have been made.
We sincerely appreciate your constructive comments and time!
This paper considers the problem of finding the unregularized barycenter between discrete probability measures. The authors reformulate this problem as a saddle-point optimization problem and propose a kind of gradient descent-ascent algorithm to solve it. The established algorithm is supported by theoretical investigation of its convergence rates. The approach is tested in two experimental setups where it is claimed to achieve better performance than two other baselines in terms of accuracy and time complexity.
优点
The authors propose an interesting approach to solving the unregularized barycenter problem using the Wasserstein and gradients. The proposed approach outperforms the chosen competitors in time and space complexity.
缺点
My major concern is related to some discrepancy between the theoretical investigations of the proposed approach and its use in practical tasks. Namely, in Section 3.4, the authors prove several theoretical results regarding their method, including the important Theorem 1, which shows that the proposed algorithm allows finding stationary points of the objective functional. From the proofs in the Appendix, one can see that all theoretical results are logically connected, i.e. Lemma 2 uses the results of Lemma 1, Lemma 3 uses Lemma 2, and so on. Thus, the fulfillment of Theorem 1 also depends on the fulfillment of Lemma 1. At the same time, in Section 3.5 the authors propose to modify the algorithm for practical use in order to reduce the time complexity. This modification consists in replacing the projecting the potentials on with computing the second convex conjugate . Such a replacement entails non-fulfilment of Lemma 1, as the authors themselves write in lines 422-423. Accordingly, Theorem 1 is also no longer true, i.e. the convergence of the algorithm's solutions to ground-truth solutions is no longer theoretically justified. I will appreciate if the authors could comment on the generalization of Theorem 1 to the case where Lemma 1 is not satisfied or conduct experiments with the original version of the algorithm.
The practical evaluation also lacks an experiment showing that the proposed algorithm (modified version) actually gives the ground-truth solutions for the considered barycenter problem. The authors only perform a comparison of some functional values computed using a back-and-forth approach (Jacobs and Le ́geret, 2020) with the two baseline approaches WB (Solomon et al, 2015) and DSB (Janati et al., 2020). However, I have doubts regarding this metric for comparison, i.e., the computed functional values. It is not clear from the text how these values are calculated - more details about the back-and-forth approach (Jacobs and Le ́geret, 2020) should be included. This is especially important since the approaches the authors compare with are entropy-regularized. Besides, I recommend the authors to conduct a comparison with ground truth solutions of the barycenter problem which can be constructed for the case of gaussian measures using, e.g., the iterative algorithm from (Álvarez-Esteban, Pedro C., et al., 2016), and perform the comparison with the baselines in this setup.
Moreover, I have some doubts regarding the significance of the developed algorithm since it borrows ideas which are already present in the field. -gradient ascent algorithm for finding the maximizers of dual OT problem was proposed by (Jacobs & Le ́ger,2020), while the Wasserstein gradient for solving the barycenter problem was initially proposed in (Zemel & Panaretos, 2019; Chewi et al., 2020). The algorithm in the current work combines these two ideas.
In summary, I think that the contribution of the paper is questionable since the algorithm combines the ideas which are already present in the field and, thus, to represent interest for the community it should have some theoretically and practically justified advantages over the previous approaches. However, in the current version of the paper, it is not clear whether the modified algorithm which was used in the experimental part actually learns the ground-truth barycenter after the considered number of epochs or not. The theoretical investigations (convergence analysis) are not valid for this modified algorithm, while the experimental section also lacks the setup with known ground-truth solutions which can be used to mitigate this concern.
Minor:
- line 132: descent ascent;
- mistake in line 269 - there should be instead of over potentials.
问题
-
The idea in lines 290-291 is not clear: “the definitions in subsection 2.3 imply that the Wasserstein gradient of J is given by , where …”. Please explain how this Wasserstein gradient is derived?
-
Why do you consider only two baseline models, i.e., CWB (Solomon et al, 2015) and DSB (Janati et al., 2020) approaches? These models use the entropy regularization which makes the comparison not entirely fair (yet, I agree that it can be done in the case of small regularization parameter ). Meanwhile, you have mentioned many other approaches which compute the barycenter in an unregularized setting, e.g., (Zemel & Panaretos, 2019; Chewi et al., 2020). Why do you not compare with these algorithms? I will highly appreciate it if you perform the comparison with some of these approaches during the rebuttal phase.
-
What do you mean by in line 211?
-
In lines 315-318 you write: “we can alternatively use the L2 -gradient to update $\nu . However, in simulations, the Wasserstein gradient update significantly outperforms the L2-gradient update.” Can you provide an experiment justifying this claim?
References.
Álvarez-Esteban, P. C., Del Barrio, E., Cuesta-Albertos, J. A., & Matrán, C. (2016). A fixed-point approach to barycenters in Wasserstein space. Journal of Mathematical Analysis and Applications, 441(2), 744-762.
Matt Jacobs and Flavien Le ́ger. A fast approach to optimal transport: The back-and-forth method. Numerische Mathematik, 146(3):513–544, 2020.
Hicham Janati, Marco Cuturi, and Alexandre Gramfort. Debiased Sinkhorn barycenters. In Pro- ceedings of the 37th International Conference on Machine Learning, pp. 4692–4701, 2020.
Sinho Chewi, Tyler Maunu, Philippe Rigollet, and Austin J Stromme. Gradient descent algorithms for Bures-Wasserstein barycenters. In Conference on Learning Theory, pp. 1276–1304, 2020.
Yoav Zemel and Victor M. Panaretos. Fre ́chet means and Procrustes analysis in Wasserstein space. Bernoulli, 25(2):932 – 976, 2019.
Justin Solomon, Fernando de Goes, Gabriel Peyre ́, Marco Cuturi, Adrian Butscher, Andy Nguyen, Tao Du, and Leonidas Guibas. Convolutional Wasserstein distances: Efficient optimal transporta- tion on geometric domains. ACM Transactions on Graphs, 34(4), 2015.
Thank you for your detailed answers and new experiments. The newly conducted experiments indeed clarify some of my concerns. However, I agree with the reviewer yaTH that the fact that your approach is tested only in the experiments with dimension , raises questions about its applicability to high dimensions.
Still, I appreciate that the authors conduct the experiments which I suggested and increase my score to 6.
We sincerely appreciate your constructive comments and time!
We sincerely appreciate your time, thoughtful consideration, and constructive guidance on our paper. Following your and the other referees' valuable comments, we added three additional simulation studies in the appendix of the revised version to address your questions. If we have successfully addressed some of your concerns, we would kindly ask you to consider reflecting this in your rating. Thank you again for your invaluable feedback.
Below we provide point-by-point response to your comments in italics.
1, My major concern is related to some discrepancy between the theoretical investigations of the proposed approach and its use in practical tasks. Namely, in Section 3.4, the authors prove several theoretical results regarding their method, including the important Theorem 1, which shows that the proposed algorithm allows finding stationary points of the objective functional. From the proofs in the Appendix, one can see that all theoretical results are logically connected, i.e. Lemma 2 uses the results of Lemma 1, Lemma 3 uses Lemma 2, and so on. Thus, the fulfillment of Theorem 1 also depends on the fulfillment of Lemma 1. At the same time, in Section 3.5 the authors propose to modify the algorithm for practical use in order to reduce the time complexity. This modification consists in replacing the projecting the potentials on with computing the second convex conjugate. Such a replacement entails non-fulfilment of Lemma 1, as the authors themselves write in lines 422-423. Accordingly, Theorem 1 is also no longer true, i.e. the convergence of the algorithm's solutions to ground-truth solutions is no longer theoretically justified. I will appreciate if the authors could comment on the generalization of Theorem 1 to the case where Lemma 1 is not satisfied or conduct experiments with the original version of the algorithm.
Ans: Thank you for your comments. The same discrepancy issue also exists in other works that utilize the -gradient if any convergence theory is to be developed; see the assumptions in Proposition 1 [Jacobs and Leger, 2020] and Theorem 2.8 [Jacobs et al., 2021. We believe that the satisfying empirical performance of our algorithm should not be discounted merely because of this. Please check our additional simulation result with ground truth. Without additional regularities, the generalization of the Theorem can be challenging. The main obstacle is that the map is not Lipschitz; see [Berman, 2021] for more details regarding this.
In practice, the projection can be computed following the approach developed by Simonetto [2021], which is implemented as the Python function reg4opt.regression.convex\_regression in library reg4opt. Unfortunately, this projection algorithm requires solving a large convex quadratically constrained quadratic problem so that it takes weeks to run an experiment in 2D.
Instead, we implement the algorithm with projection (Algorithm 3) and test the method on 1D distributions. More precisely, for each repetition and , we let be the truncated distribution of on the domain , where and . Let , , be the true barycenter, the barycernter computed from Algorithm 3 (with projection onto ), and the barycernter computed from Algorithm 4 (with double convex conjugates) respectively. We repeat the experiment times and report two types of average distances between the groundtruth and each computed barycenters : average -distance and average -distance between densities . Algorithm 3 performs slightly better with -distance and -distance 0.608, while Algorithm 4 has -distance and -distance 0.6434. We have included this study in subsection C.2 in the appendix of the revised version.
2. The practical evaluation also lacks an experiment showing that the proposed algorithm (modified version) actually gives the ground-truth solutions for the considered barycenter problem. The authors only perform a comparison of some functional values-computed using a back-and-forth approach (Jacobs and L'{e}geret, 2020) with the two baseline approaches WB (Solomon et al, 2015) and DSB (Janati et al., 2020). However, I have doubts regarding this metric for comparison, i.e., the computed functional values. It is not clear from the text how these values-are calculated - more details about the back-and-forth approach (Jacobs and L'{e}geret, 2020) should be included. This is especially important since the approaches the authors compare with are entropy-regularized. Besides, I recommend the authors to conduct a comparison with ground truth solutions of the barycenter problem which can be constructed for the case of gaussian measures using, e.g., the iterative algorithm from ('{A}lvarez-Esteban, Pedro C., et al., 2016), and perform the comparison with the baselines in this setup.
Ans: that's a good point. We added a new simulation in the revised version for which the ground truth is known. This simulation demonstrates that both CWB and DSB may output blurry barycenters, while our result is visually no different from the truth. We, in addition, report the -distance between computed barycenter densities and the true density.
In this new simulation, the goal is to compute the barycenter of four uniform distributions supported on round disks of radius 0.15, centered at , , , respectively. It's clear that the true barycenter is uniform on the disk of radius 0.15 centered at . The computed barycenter densities by WDHA, CWB and DSB are shown in Figure 3 of the revised version. We note that regularization parameters reg= 0.003 and reg=0.005 are the optimal choices for CWB and DSB respectively. Smaller regularization parameters lead nonconvergent and worse results for both CWB and DSB. We report in the following table the Wasserstein distance between computed barycenter distribution to the truth, -distance between computed barycenter densities and the true density, and the barycenter functional value. Our method is uniformly the best, and in particular, the improvement in the Wasserstein distance is of orders of magnitude.
| reg | Wasserstein distance | -distance | ||
|---|---|---|---|---|
| WDHA | 0.4869 | |||
| CWB | 0.003 | 1.321 | ||
| CWB | 0.002 | 3.041 | 0.1154 | |
| DSB | 0.005 | 0.8219 | ||
| DSB | 0.004. | . | 2.281 | 0.1016 |
We mark that the barycenter functional value of the truth is 0.09. This simulation is included in the appendix subsection C.1 of the revised version.
Thank you for pointing out ('{A}lvarez-Esteban, Pedro C., et al., 2016). We have added the following on line 47 of the revised version ``Alvarez-Esteban et al. (2016) propose a fixed point approach that is effective for any location-scatter family."
3. Moreover, I have some doubts regarding the significance of the developed algorithm since it borrows ideas which are already present in the field. -gradient ascent algorithm for finding the maximizers of dual OT problem was proposed by (Jacobs & L'{e}ger,2020), while the Wasserstein gradient for solving the barycenter problem was initially proposed in (Zemel & Panaretos, 2019; Chewi et al., 2020). The algorithm in the current work combines these two ideas.
Ans: We proposed a gradient-ascent type algorithm for solving the nonconvex-concave minimax formulation of the Wasserstein barycenter problem. Our method works beyond Gaussian distributions. The work of (Chewi et al., 2020) focus on Gaussian distributions, as the optimal transport maps have closed forms in this case. A work that can be considered as combination of the two ideas is to solve the inner maximization problem completely by applying -gradient descent until it converges, while our method updates the current potential with just one round of -gradient descent.
4. The idea in lines 290-291 is not clear: “the definitions in subsection 2.3 imply that the Wasserstein gradient of is given by , where ...". Please explain how this Wasserstein gradient is derived?
Ans:To compute Wasserstein gradient of , we apply the definition in Subsection 2.3. Note that
which implies that . By the definition of Wasserstein gradient, Wasserstein gradient of . We have added this part to the appendix subsection D.1 of the revised version.
5. Why do you consider only two baseline models, i.e., CWB (Solomon et al, 2015) and DSB (Janati et al., 2020) approaches? These models use the entropy regularization which makes the comparison not entirely fair (yet, I agree that it can be done in the case of small regularization parameter ). Meanwhile, you have mentioned many other approaches which compute the barycenter in an unregularized setting, e.g., (Zemel & Panaretos, 2019; Chewi et al., 2020). Why do you not compare with these algorithms? I will highly appreciate it if you perform the comparison with some of these approaches during the rebuttal phase.
Ans: We note that the algorithms in (Zemel & Panaretos, 2019; Chewi et al., 2020) are not applicable to the examples in our paper, as they implicitly assume that in each iteration, the optimal transport between and each are known, which is only true for Guassian distributions. We fill this gap by proposing the the *Wasserstein-Descent \dot{\mathbb{H*}^1-Ascent} (WDHA) algorithm.
6. What do you mean by in line 211?
Ans: We have changed to indicating gradient w.r.t in the revised version.
7. In lines 315-318 you write: ``we can alternatively use the L2 -gradient to update . However, in simulations, the Wasserstein gradient update significantly outperforms the L2-gradient update." Can you provide an experiment justifying this claim?
Ans: Certainly, we implemented this algorithm and tested it on uniform distributions supported on round disks for which the ground truth is known. Let , be density functions. We can write
Let be the space of functions that are -integrable with respect to the Lebesgue measure and be the set of density functions. Notice that the -gradient of with respect to is , we may consider the -descent -ascent algorithm: for each iteration , do
- For , compute
- Compute
Here, is the projection of onto , which can be computed using python function pyproximal.Simplex in library 1PyProximal. We apply this algorithm to the uniform distributions supported on round disks, and observe that it actually diverges leading to a wrong result. Nevertheless, we plot the output in Figure 4 of the revised version. This simulation is included in appendix subsection C.3 of the revised version.
References
- Robert J Berman. Convergence rates for discretized Monge–Amp`ere equations and quantitative stability of optimal transport. Foundations of Computational Mathematics, 21(4):1099–1140, 2021.
- Rick Chartrand, Brendt Wohlberg, Kevin Vixie, and Erik Bollt. A gradient descent solution to the monge- kantorovich problem. Applied Mathematical Sciences, 3(22):1071–1080, 2009.
- Lucilla Corrias. Fast legendre–fenchel transform and applications to hamilton–jacobi equations and conser- vation laws. SIAM journal on numerical analysis, 33(4):1534–1558, 1996.
- Matt Jacobs and Flavien L´eger. A fast approach to optimal transport: The back-and-forth method. Nu- merische Mathematik, 146(3):513–544, 2020.
- Matt Jacobs, Wonjun Lee, and Flavien L´eger. The back-and-forth method for wasserstein gradient flows. ESAIM: Control, Optimisation and Calculus of Variations, 27:28, 2021.
- Alexander Korotin, Vage Egiazarian, Lingxiao Li, and Evgeny Burnaev. Wasserstein iterative networks for barycenter estimation. In Advances in Neural Information Processing Systems, volume 35, pages 15672–15686. Curran Associates, Inc., 2022.
- Lingxiao Li, Aude Genevay, Mikhail Yurochkin, and Justin M Solomon. Continuous regularized wasserstein barycenters. Advances in Neural Information Processing Systems, 33:17755–17765, 2020.
- Tianyi Lin, Chi Jin, and Michael Jordan. On gradient descent ascent for nonconvex-concave minimax problems. In International Conference on Machine Learning, pages 6083–6093, 2020.
- Andrea Simonetto. Smooth strongly convex regression. In 2020 28th European Signal Processing Conference, pages 2130–2134, 2021.
- Michael Eugene Taylor. Partial differential equations. 1, Basic theory. Springer, 1996.
In this paper, the authors present a new, computationally efficient approach for the Wasserstein barycenter problem. Without resorting to entropic-regularization, which degrades the quality of the result, they propose to formulate the problem as a nonconvex-concave minimax optimization problem, solved with a mixed gradient descent in the Wasserstein space and gradient ascent in Sobolev space. The algorithm is shown to converge to a fixed point with the classic rate for this class of problems. The final complexity is in if one projection step is approximated by a double conjugate (which violates the theory but performs well in practice), to be compared with in most previous approaches. Experiments on synthetic data and handwritten digits are performed.
优点
I liked this paper. The idea is natural but novel and efficient, and the presentation is clear and well-written despite the technicity of the work. Experiments are convincing. The authors are also honest in giving credit to inspirations from previous works (esp. Jacobs and Léger, and Lin et al.).
缺点
One point that would deserve more discussion is the property of the nonconvex-concave problem. Unless I am mistaken (I could very well miss something), by reformulating the problem as a nonconvex problem, one loses convergence to the true solution, but only towards a stationary point. This is a bit obfuscated in the current formulation, and would deserve a more honest discussion: efficient computation is obtained by "sacrificing" some theoretical guarantees, even with good practical performance. (If I missed something and convergence to the true barycenter is guaranteed, then it also deserved to be written!)
Minor typo: "dicuss", "Sythentic"
问题
See above.
We sincerely appreciate your time, thoughtful consideration, and constructive guidance on our paper. Following your and the other referees' valuable comments, we added three additional simulation studies in the appendix of the revised version to address your questions. If we have successfully addressed some of your concerns, we would kindly ask you to consider reflecting this in your rating. Thank you again for your invaluable feedback.
Below we provide point-by-point response to your comments in italics.
1, One point that would deserve more discussion is the property of the nonconvex-concave problem. Unless I am mistaken (I could very well miss something), by reformulating the problem as a nonconvex problem, one loses convergence to the true solution, but only towards a stationary point. This is a bit obfuscated in the current formulation, and would deserve a more honest discussion: efficient computation is obtained by ``sacrificing" some theoretical guarantees, even with good practical performance. (If I missed something and convergence to the true barycenter is guaranteed, then it also deserved to be written!)
Ans: Thank you. In Euclidean space, the goal of minimax problem is to find the minimizer of , where . If is nonconvex in , then is nonconvex, and thus a feasible goal is to find such that is small.
For the Wasserstein barycenter problem, by definition, is a Wasserstein barycenter if , where is the Kantorovich potential between and . If for all , then is a stationary point of , i.e., Wasserstein gradient of is 0. Reversely, if we assume that the Kantorovich potential between the true barycenter and each is in , then Wasserstein gradient of is 0 would mean that is a Wasserstein barycenter. We have included this discussion in Remark 1 of the revised version.
Thank you for answering my questions. I keep my positive score.
We sincerely appreciate your constructive comments and time!
This paper presents the Wasserstein-Descent–Ascent (WDHA) algorithm, a novel approach for computing unregularized Wasserstein barycenters via a nonconvex-concave minimax optimization framework. The key contribution lies in its primal-dual formulation that alternates between Wasserstein and Sobolev geometries, achieving computational efficiency for low-dimensional settings. While the method shows potential, several critical weaknesses lead to the recommendation of rejection. The numerical results are underwhelming, with artifacts indicating potential instability. The experiments are restricted to low-dimensional settings, raising concerns about scalability to higher-dimensional problems, which are central to optimal transport applications. Furthermore, the theoretical contributions are limited, with some theorems criticized as lacking explanatory power or practical relevance. Reviewers also noted insufficient acknowledgment of the method's limitations, such as its reliance on approximations that weaken theoretical guarantees. While the authors provided additional clarifications during the rebuttal, the core issues remain unresolved, making the work feel premature for acceptance at this stage. Addressing these gaps is crucial to solidify the contribution and ensure its robustness and applicability.
审稿人讨论附加意见
During the rebuttal period, reviewers raised concerns about numerical instability, limited scalability to high-dimensional settings, and the superficial nature of the theoretical contributions. The authors responded with additional simulations and clarifications, addressing some questions about their algorithm's behavior. However, critical issues, such as the robustness of the numerical results, scalability, and the practical relevance of the theoretical insights, were not sufficiently resolved. While the new experiments added some value, they did not mitigate the overarching concerns. After considering the discussion, I weighed the lack of stability and scalability as decisive factors in my recommendation for rejection, as these are fundamental for advancing research in this area.
Reject