Asymptotics of Alpha-Divergence Variational Inference Algorithms with Exponential Families
We provide asymptotic results for algorithms optimizing the alpha-divergence criterion in the context of Variational Inference, using an exponential variational family.
摘要
评审与讨论
This paper focuses on the theoretical study of variational inference using alpha-divergences using exponential family. This is an important problem in the field of variational inference. Specifically, this work proves a geometric convergence rate for the algorithm proposed in [9]. Moreover, this paper also proposes an alternative optimization algorithm and its convergence guarantees are provided. The proposed algorithm finds applications in VAE.
优点
-
Asymptotic convergence rate of the algorithm proposed in [9] is derived in this work. Moreover, this paper proposes an alternative unbiased algorithm with convergence guarantees.
-
Assumptions (H1)-(H3) follow by reasonable explanations.
-
Applications of the proposed algorithm to VAE are provided.
缺点
-
It only considers the variational family to be exponential models.
-
The convergence is proved for the proposed algorithm in the asymptotic sense.
问题
-
Typo at line 81, ''negative'' ''positive'' since you drop the negative term.
-
What is the technical difficulty to derive a convergence result similar to Theorem 1 for the proposed algorithm?
局限性
Yes.
Thank you for your thorough review and honest feedback. Below, we provide detailed responses to your second question. The two weaknesses you pointed out are adressed in our global rebuttal. We hope that our responses will clarify the issues you raised.
What is the technical difficulty to derive a convergence result similar to Theorem 1 for the proposed algorithm?
It does not present any major difficulty and can be done by following the exact same steps as in the proof of Theorem 1. We would get a similar result, with a slightly slower rate however: the in the convergence speed would be replaced by , and we have by Jensen’s inequality. It is easy to gain some intuition for why this is the case: under the mean parameterization, the update is changed from to , so we are still performing a convex combination of and but with a less aggressive coefficient, thus yielding a slower rate of convergence. This also provides a heuristic as to why UNB moves slower at the start of the experiments: when the variational parameter is poorly chosen, is closer to , which caps the progress that can be made at each step when the only values allowed for are between and .
We are grateful for your careful review, we hope our responses have adequately addressed your questions and provided further insights.
Thank you for the response. I would like to keep my score.
This procedure studies the asymptotic properties of a variational algorithm that ensures a monotonic decrease in the alpha-divergence. Specifically the authors investigate behavior in the setting where the variational distribution belongs to the exponential family of distributions. In this setting, and when a key integral can be analytically calculated, the the authors show monotonic convergence at a geometric rate. For scenarios where a key integral cannot be calculated explicitly, the authors show an unbiased empirical approach. Furthermore the authors show that this empirical approach enjoys almost sure convergence to a local minimizer of the alpha-divergence. Finally, the paper presents, both, simulated and real-data experiments to support the theoretical claims.
优点
This is an very well-written paper and clear despite its high level of mathematical rigor. The authors consider an important problem of alpha-divergence minimization, which has broad impacts in the ML community. The convergence analysis and unbiased empirical minimization algorithm appear novel and interesting.
缺点
The paper does not contain any major weaknesses that I could find, beyond limitations of the analysis. These limitations include the specific assumption of an exponential family variational distribution, as well as the stated assumptions H1-H4 and C0-C3.
Below are some detailed comments:
- L25 : The inclusive KL divergence should be .
- L69 : Shouldn't the posterior be proportional to the joint rather than the marginal ?
- L80 : "Assuming the argmax is uniquely defined at each iteration" it is unclear whether this is a reasonable assumption
- Eq. (5) : Notation undefined
- L108 : Is it obvious that ?
- L196 : Notation undefined (is this a composition?)
Perhaps the biggest weakness is Sec. 5. This section falls a little flat as it concludes "we both obtain a biased and unbiased algorithm" without making those algorithms explicit or obvious (at least to me). Also, it is unclear how strong an assumption it is that the exponential family does not depend on as in . In what reasonable scenario would the encoder be independent of the input?
问题
See "Weaknesses" section.
局限性
The authors do not explicitly state limitations of the proposed methodology. Nevertheless the authors do make explicit assumptions of the analysis.
Thank you for your meticulous review and kind feedback. Below, we provide detailed responses to each of your questions, aiming to further clarify our work and address any remaining uncertainties.
Diverse remarks on notation + L196 : Notation is undefined (is this a composition?)
Yes, denotes a composition. We thank you for carefully reading our paper and will make sure to correct these mistakes.
L69 : Shouldn't the posterior be proportional to the joint rather than the marginal ?
Indeed, we can take . The dependency in is dropped for notational ease.
L80 : "Assuming the argmax is uniquely defined at each iteration" it is unclear whether this is a reasonable assumption.
As you point out, this assumption can be challenging in the general case, but when the variational family is an exponential family as in (H1), the argmax in (1) is uniquely defined by if and only if this quantity is in . This is why the choice of is so critical: if is not in , we must take small enough to ensure that the next iterate will be valid. Under (H1), the set is open and convex, which implies two things. First, by convexity, if is a valid choice, then all are also valid choices. Second, by openness, it is always possible to find such a .
L108 : Is it obvious that ?
Note that is the normalizing constant that makes a probability density function. By linearity of the integral, is the expectation of the measurable function under a distribution with density , i.e. . This explains why .
[Sec. 5.] falls a little flat as it concludes "we both obtain a biased and unbiased algorithm" without making those algorithms explicit or obvious. Also, it is unclear how strong an assumption it is that the exponential family does not depend on as in .
The assumption in Section 5 was used only to illustrate a specific point: when is fixed and the decoder does not depend on (i.e., and we only update ), the exact gradient of the VR bound equals (up to a negative factor) . This was meant to show that, in this particular scenario, an iteration of the algorithm discussed in Section 3 corresponds to a gradient ascent step on the VR bound. This assumption is not representative of the VAEs' training process, which uses the gradient estimators (17) and (18) with an Adam optimizer. Specifically, estimator (18) is used for VR and estimator (17) for UB. We will reorganize this section to clarify these points.
Thank you once again for your detailed review and supportive comments. We hope that our responses have satisfactorily addressed your questions and concerns.
Thanks for the thorough rebuttal. I think this is a strong piece of work and will keep my score. Unfortunately I can't increase my confidence score as I am unfamiliar with the relevant work in this area and I have not had a chance to thoroughly validate the analysis beyond a standard detailed read through.
As a side note, I personally find notation to be very confusing / misleading. Especially when you also refer to the prior (L57). You may want to consider revising this. At the very least you should explicitly state that x is assumed implicit for brevity.
Thank you again for your positive feedback. Regarding the notation, we understand your concern about potential confusion. We will emphasize the fact that the data is fixed throughout the optimization process, and will rename the prior to avoid any ambiguity and improve the accessibility of the paper.
The paper studies the converge properties of an optimization algorithm, used to minimize the alpha divergence when the variational approximation belongs to the family of exponential distributions. The optimizer is a modification of an existing method and its performance is competitive with existing methods (without being stronger). However, the new approach enjoys theoretical guarantees.
优点
This paper is line with recent publications on proving the convergence of VI when minimizing the KL-divergence. Obtaining similar results for the alpha-divergence is a natural next step. The main contribution of the paper is theoretical, and I find the results convincing (although I did not read the appendix in detail).
The theory motivates a revision to existing methods, which seems to be competitive. Minimizing alpha-divergence is notoriously difficult and I appreciate the paper's discussion of algorithms to do so, and demonstration on some examples.
缺点
The paper lacks clarity and is difficult to read.
First, the notation is heavy. Two ways to address this might be to introduce a table of notation and to clearly delineate definitions as formal statements. I'm guessing this wasn't done because of space constraints, but this could be a good use of the additional page for accepted papers.
In Section 5, I was puzzled by the assumption that , that is the factor does not depend on . Does this mean that the encoder maps each image to the same latent representation? It seems like the VAE would then be useless, at least for the task of data compression. This also makes me question the results of Section 6. I'm guessing that while the VAE fails to do any meaningful compression of existing images, it may still generate convincing new images. I'd like to see such images (potentially in the appendix), as a supplement to Table 1, and a simple check that the model is trained in a useful manner.
Once the authors clarify this point, I can adjust my score.
问题
I'll use this for minor comments and clarification questions.
- instead of (or in addition to) exclusive and include KL, write KL(q||p) and KL(p||q)
- Figure 1 and Figure 2 are hard to read. In Figure 1, indicate the optimum. Could the colors for the objective function be more distinct than the colors of the trajectories?
- Figure 2: font size should font size of text. Could the box plots be replaced with trajectories and shaded intervals?
- line 314: fix the reference to eq:vrbound
- Define the IWAE and VAE baselines. Does this simply correspond to minimizing KL(q||p) and KL(p||q)? Is this using the variational family as prescribed in line 275?
局限性
Yes.
We are grateful for your careful reading and constructive comments on our submission. Below, we respond to each of your questions, hoping to clarify any uncertainties.
The paper lacks clarity and is difficult to read.
We are sorry to hear that despite our best efforts to make the paper and the notation as clear and readable as possible, you encountered some difficulty in reading our work. We acknowledge that the notation in the paper is dense and will do our best to improve the overall readability by following your advice.
In Section 5, I was puzzled by the assumption that , that is the factor does not depend on . Does this mean that the encoder maps each image to the same latent representation?
The assumption in Section 5 was used only to illustrate a specific point: when is fixed and the decoder does not depend on (i.e., and we only update ), the exact gradient of the VR bound equals (up to a negative factor) . This was meant to show that, in this particular scenario, an iteration of the algorithm discussed in Section 3 corresponds to a gradient ascent step on the VR bound. This assumption is not representative of the VAEs' training process, which uses the gradient estimators (17) and (18) with an Adam optimizer. Specifically, estimator (18) is used for VR and estimator (17) for UB. We will reorganize this section to clarify these points.
Figure 2: font size should be font size of text. Could the box plots be replaced with trajectories and shaded intervals?
Like you suggest, we originally intended for Figure 2 to show full trajectories with shaded intervals, but the result was cluttered and difficult to interpret due to the overlapping of five lines and their respective intervals. Additionally, the trajectories were quite noisy, although Polyak-Ruppert averaging could mitigate this specific issue. We will make further efforts to improve the clarity of the plot.
Define the IWAE and VAE baselines. Does this simply correspond to minimizing KL(q||p) and KL(p||q)?
Yes, IWAE is obtained when and VAE corresponds to .
We extend our gratitude for your careful review and insightful feedback, and we hope our responses have effectively addressed your questions and concerns. We acknowledge the issues you noticed in the writing of our paper and will make sure to fix them.
I've read the authors' rebuttals and I thank them for addressing my comments.
I'd like to ask for a clarification regarding section 5 and the special case . The authors write "we can exploit this link to derive a VAE training procedure for the unbiased algorithm (10)." After re-reading the paragraph, I still do not understand how this link is exploited in order to generate a training algorithm for the case where .
Regarding the figure, I understand that plotting trajectories can sometimes hurt readability. I hope the authors will consider adjusting the font size.
Yes, IWAE is obtained when and VAE corresponds to .
Thank you for the clarification.
We acknowledge that Section 5 lacks clarity and appreciate the opportunity to resolve this issue. In what follows, we expand on the reasoning behind this part of the paper.
Consider an algorithm whose updates are of the form , where is a positive definite matrix and . Usually, is the gradient of some function that we want to maximize.
In our case, it is equivalent to consider and , or and , where is the Fisher Information Matrix of the model at ,. This arises directly from the chain rule, along with the identities and recalled in Appendix A.1. Indeed, they imply .
Since the gradients of interest are easier to compute w.r.t. , we will choose the first option. Specifically, we have , which leads to .
For the unbiased algorithm , we set , which we recognize as the gradient of .
Multiplying by the positive constant and writing it in integral form, we find that (remember that and are related by the one-to-one mapping ).
In the more general case where we have instead of and instead of , this is the Variational Rényi (VR) bound: .
In other words, the exact biased algorithm is a particular case of gradient ascent on the VR bound.
By doing a similar work for the unbiased algorithm, where , we get . Hence, the unbiased algorithm is a particular case of gradient ascent on a new bound given by .
Using the REINFORCE gradient and backpropagation, we train VAEs to maximize these bounds. The respective methods are referred to as VR and UB in the paper.
We hope this reformulation of the proof will help clarify any ambiguities. Should you have any further questions or require additional explanations, please let us know. Thank you again for your time and constructive comments.
Thank you for engaging. However these additional details do not provide the clarification I'm looking for. Namely, why do we need to study the case where in order to derive a learning algorithm?
In the above derivation, I do not see where in the calculation of we exploit the fact that and . And once we derive in this particular case, why can we backtrack to the more general case?
I believe the writing should make it clear why this detour is necessary, how is the simpler case exploited, and why can we then go back to the more general case. Would it not be possible to directly work in the more general case?
Using the REINFORCE gradient and backpropagation, we train VAEs to maximize these bounds.
The authors have not defined "reinforce" gradient, could you define it?
Thank you for your feedback. We appreciate the opportunity to clarify our approach.
In the paper, we distinguish between two contexts of variational inference: the traditional framework discussed in sections 3 and 4, and the more complex VAE training setup.
-
In the traditional setting, we want to approximate the true posterior by learning the parameters of a single distribution . Since the data remains fixed throughout the entire training process, we drop all dependencies in : we suppose that is known up to a constant (through the function ), and simply we denote . This does not mean that the learnt parameter isn't dependent on the data. Rather, we are optimizing a parametric distribution over the entire dataset, e.g., coefficients in Bayesian logistic regression.
-
In contrast, VAEs involve optimizing both an encoder and a decoder simultaneously. One of the main goals with VAEs is to learn a meaningful latent representation of the data, so each data point must be effectively mapped to the latent space, hence the particular conditioning on . Unlike the traditional setup, where we learn a single distribution, VAEs require joint optimization of two networks. This complexity necessitates adjusting the approach.
The derivation in our paper shows that the algorithms MAX and UNB perform gradient ascent on specific variational bounds, namely the VR bound and .
The VR bound, for example, is defined as .
If we are in setup 1, the only thing we can act on is the parameter . Differentiating the previous formula w.r.t. shows that MAX is a form of gradient ascent on .
For VAEs, we need to optimize both the encoder and decoder networks, and we are not changing and but the weights of the networks, of which the parameters are functions (say and , where denotes the weights of both networks). To handle this new task, we use the reparameterization trick for differentiable optimization and the REINFORCE gradient, i.e., we write with and denotes the derivative with respect to the networks' weights. If further details are needed, we can provide a more comprehensive derivation.
What ties the algorithms from sections 3 and 4, and the VAE optimization process we propose together is the fact that they minimize the alpha-divergence by maximizing the same variational bounds. The procedures that lead to maximizing the bounds, however, are very different.
We hope this clarifies the fact that these two frameworks are different and thus need be treated adequately. Thank you for your consideration.
Thank you for providing additional details.
I'm well familiar with how training a VAE works and this is not my point of confusion. Moreover, while everything that the authors state in their response is correct (distinction between Bayesian inference and learning for VAEs), this still does not answer my question. I'll restate it once last time: why do you need to first derive the case where does not take in an input , i.e. you're not learning an encoder and doing amortized variational inference, in order to derive the training procedure for the VAE? The sentence that I find confusing is
We can exploit this link to derive a VAE training procedure for the unbiased algorithm.
I still don't see how the link is exploited.
Perhaps you can clarify the following: when you write "" and " does not depend on ", what setting are you examining here? As far as I can tell, this is not what you call "the traditional setting" where you do posterior learning, since you're still learning (or the weights depends on). It is simply an encoder which doesn't take in or some specific input .
As is, I still believe this is a valuable contribution to the NeurIPS community and I'll maintain my score, which a weak accept.
We appreciate your patience and apologize for misunderstanding your point of confusion, which is absolutely valid. This specific sentence in the paper is indeed wrong, due to a poor formulation of the underlying idea.
Throughout the entire paper, the function always depends on and can be written . For simplicity and to reduce notation clutter, we omit this dependency until Section 5.
In Section 5, the function denoted is not the density of a distribution in an exponential family parameterized by . Instead, it could be expressed , where is a neural network. When we wrote "the parameter is not updated, belongs to an exponential family and and does not depend on ", we meant that we were temporarily reverting to the "traditional VI" setting. The goal there was to explain why MAX performs gradient ascent on the VR bound when we are in the traditional setting and is the density of an exponential distribution.
The exact reasoning that led to the VAE training algorithms is the one described in the comments above. To summarize, we start by noticing that in the setting of Sections 3 and 4, the algorithms MAX and UNB are gradient ascent procedures, respectively on the VR bound and . Training VAEs to maximize these bounds leads to the methods referred to as VR and UB.
We hope this explanation will satisfactorily address your concerns about this part of the paper. We will delete the problematic sentence and reorganize section 5 so that it better reflects the train of thought behind the VAE training methods we use.
Thank you for your thorough review and for helping us ensure the clarity of our paper.
Thank you for getting back to me. My intention is not to be stubborn, and I'm quite keen to better understand your work.
The content of Section 5 is still not entirely clear to me, but at this point, I believe I need to have another close look at the paper in light of the comments provided by the author, and I'll discuss the matter with the other reviewers. Ideally, I would like to read a revised Section 5. Since this section is fairly short, I invite the authors to rewrite Section 5 to include their clarifying comments. That said, I recognize this is an unusual request and even if the authors did not do this, I'd still consider changing my score after re-reading the paper and our exchange during the rebuttal period.
Thank you for your interest in our work and for giving us the opportunity to improve it. Below is a draft of the revised Section 5 you asked for. We hope that this revision will clarify the points you raised.
In this section, we explain how to transpose the algorithms presented in Section 4.1 to the training of Variational Auto-Encoders (VAEs) [22]. We start by showing that the exact versions of the biased and unbiased algorithms correspond to gradient ascent procedures on two different variational bounds. Let us first address the case of the biased algorithm. Recall that it writes . For and , we define the Variational Rényi (VR) bound [26] by
Noticing that , we can thus express an iteration of the biased algorithm as Under (H1), the Fisher Information Matrix is positive definite, hence the biased algorithm is a gradient ascent procedure on the VR bound.
The unbiased algorithm, which writes , similarly amounts to performing gradient ascent on the variational bound defined by
In the context of VAEs [22], we learn both a probabilistic encoder and a probabilistic decoder , where and are densities from families parameterized respectively by and , while and are neural networks. For simplicity and to align with the usual notation for VAEs, we will denote and . We will also use the shorthand .
Since the biased and unbiased algorithms studied in the previous sections minimize the alpha-divergence by maximizing the variational bounds and , we propose to train VAEs to maximize those same bounds. In this new setting, they write
To update both the encoder and decoder simultaneously, we differentiate them with respect to using the reparameterization trick [22]. If and there exists a mapping such that has the same distribution as when , then we have
where and .
To train VAEs, we simply plug batch estimates of these gradients into an optimizer like Adam. Notably, can be estimated unbiasedly, while estimators of are subject to bias. We will study the practical implications of this fact in Section 6.
This paper proposes the asymptotic analysis for both exact and empirical alpha-divergence minimization algorithms, especially in the case of infinite number of iterations. The paper mainly focuses on the exponential family setting, and provides geometric convergence analysis of exact minimization algorithms using fixed point theories. To bypass the difficulty on studying the empirical minimization algorithms due to the bias of previous algorithms, a novel unbiased algorithm is proposed and analyzed, including almost sure convergence to a local minimizer and a law of the iterated logarithm. Finally, the paper experiments on toy gaussians and on variational auto-encoders to show the effectiveness of the proposed algorithm.
优点
- The paper is clearly organized, including both exact and empirical analysis.
- The paper evaluates the asymptotic properties as the number of iterations goes to infinity, which are seldom discussed in previous works.
- Extensive discussions on the convergence of empirical alpha-divergence minimization algorithms are provided in the paper.
缺点
-
The novelty of the paper appears to be limited. The basis of the theoretical analysis is mostly covered by [1], and the main theoretical analysis focuses only on exponential families.
-
Some assumptions in the theoretical analysis could be further evaluated.
- For Assumption (H3) in section 3, it remains unclear when the mapping is a contraction.
- It would be beneficial to further verify this for and in the case of the exponential family.
- For Assumption (C2) in section 4, the paper states that for “ suppose specific behaviors on the relative tails”. It would be illuminating to verify this assumption in experiments, especially in VAE experiments, where the choice of can make a difference to the empirical results.
-
The experiments of this paper could be extended. It would be nice to add experiments like Bayesian logistic regression as in [1]. Furthermore, the empirical results lack significance. In the toy gaussian experiments, the proposed UNB method achieves more accurate but slower convergence results compared to the biased MAX method, and is surpassed by the NAT method with proper step size. In VAE experiments, the UB approach is not significantly better than the VR approach in some cases.
-
The relation between the unbiasedness and the computational intensiveness of the proposed algorithm could be further justified. The bias issue is not analyzed theoretically because “biased gradient estimators hinder any theoretical study”, and there seems to be a trade-off between unbiasedness and computational intensiveness regarding the comparison of MAX and UNB methods in toy experiments.
[1] K. Daudel, R. Douc, and F. Roueff. Monotonic Alpha-Divergence Minimisation for Variational Inference. Journal of Machine Learning Research, 24(62):1–76, 2023.
问题
- The paper asserts that “biased gradient estimators hinder any theoretical study”. Is it possible to compare the biased and unbiased algorithms on a relatively simple example like gaussian distributions or two-mixture of gaussians?
- If assumption (H3) is not satisfied, do we still have geometric convergence? If not, what would be the convergence rate?
- In VAE experiments, different result in different empirical results for both VR and UB methods. Moreover, UB achieved relatively worse results for , which is the interval where assumption C2 could be satisfied. Could the authors provide a heuristic understanding of this phenomenon?
- In line 208 of section 4, the paper states “on top of that, we lose the monotonicity property”. The monotonicity property is essential for the geometric convergence in the exact alpha-divergence minimization algorithm analysis. Does it also contribute to the empirical algorithm analysis? Is it beneficial to solving the instability problems for large step sizes?
局限性
- The paper states the limitation of only considering the exponential family. Still, there are additional limitations when focusing on this setting. Please refer to the second point of weaknesses.
- Although the paper claims to focus mostly on asymptotic theory, the quality of experiments could still improve to better support the theoretical analysis. Please refer to the third point of weaknesses.
Thank you for your thorough and constructive review of our paper. We greatly appreciate the time and effort you have dedicated to providing feedback. We are committed to addressing the concerns raised and improving our work. Below, we respond to each of your points in detail.
The basis of the theoretical analysis is mostly covered by [1].
While our work indeed builds on the algorithm proposed in [1], which had already established the monotonicity property and explicit update form for the exponential family, our contributions lie in proving the algorithm’s convergence at an asymptotically geometric rate and towards a minimizer of the alpha-divergence. We further consider this algorithm in the empirical setting and propose an unbiased version, for which we provide two convergence theorems.
For Assumption (C2), the paper states that “ supposes specific behaviors on the relative tails”. It would be illuminating to verify this assumption in experiments.
We have found that convergence can still occur even if assumption (C2) is not met. For example, when is the density of a Cauchy distribution and the variational family is Gaussian, we still observe convergence (Appendix C). So far, we have not encountered a simple case where convergence entirely fails. This suggests that assumption (C2) could be relaxed or even replaced in future studies.
It would be nice to add experiments like Bayesian logistic regression.
Due to space constraints, we had to make choices regarding which experiments to present in the paper. We opted for the toy Gaussian examples as we found them to be more insightful than Bayesian logistic regression or other classical applications of variational inference algorithms.
The empirical results lack significance.
We believe that the main takeaway from the toy gaussian experiments is that the MAX approach is highly robust against aggressive hyperparameter tuning strategies, and generally converges faster than other methods with theoretical guarantees.
Although the NAT method performs well when it converges, it has shown occasional instability and is costlier per-iteration.
In VAE experiments, the UB approach is not significantly better than the VR approach in some cases.
It is true that the UB approach does not always show a significant advantage over the VR approach in our VAE experiments. However, we believe that this study offers useful insights. As you pointed out, the choice of makes a difference in the empirical results, which is quite exciting as the idea behind the use of alpha-divergences in variational inference is to overcome some limitations of the traditionally used KL divergence which can be problematic on certain datasets. We observe that a proper tuning of can greatly improve the model’s performance, though the optimal value seems to depend on both the gradient estimator and the dataset. A thorough analysis of the phenomena at play is beyond the scope of this paper, but is certainly a compelling question for future research.
The bias issue is not analyzed theoretically because “biased gradient estimators hinder any theoretical study”.
Analyzing the MAX approach theoretically presents two main difficulties. First, if it converges, it converges to a minimizer of an approximation of the alpha-divergence, rather than the true alpha-divergence itself. Second, this approximation is hard to characterize, since it amounts to finding and analyzing a function whose gradient is .
Is it possible to compare the biased and unbiased algorithms on a relatively simple example like gaussian distributions or two-mixture of gaussians?
In the experiments section, both the biased and unbiased algorithms are compared in the case of a Gaussian variational family with Gaussian, mixture of Gaussian and Cauchy targets (see also Appendix C). The biased algorithm is referred to as MAX, while the unbiased algorithm is UNB.
If assumption (H3) is not satisfied, do we still have geometric convergence? If not, what would be the convergence rate?
We need assumption (H3) to obtain the geometric rate of convergence. Lemma 2 (found at the beginning of Appendix A.1) establishes that, as long as the sequence remains bounded, at least one fixed point of the mapping will be a limit point of this sequence. However, evaluating whether this limit point is the definitive limit and determining the convergence rate without (H3) is beyond the scope of this study.
UB achieved relatively worse results for , which is the interval where assumption C2 could be satisfied. Could the authors provide a heuristic understanding of this phenomenon?
We believe that the phenomenon at play is quite complex and depends mostly on the dataset, since opposite behaviors are observed between CIFAR10 and CelebA. Moreover, as we already discussed, assumption (C2) might be subject to relaxation, thus it may not fully explain the deeper reasons behind the observed convergence behavior.
Does [the monotonicity property] contribute to the empirical algorithm analysis? Is it beneficial to solving the instability problems for large step sizes?
Empirically, we can not guarantee the monotonicity property unless an accept-reject step is integrated into the procedure. The main idea behind the analysis of the sample-based algorithm is to use the mean parameterization of the exponential family to write it as a Robbins-Monro procedure, hence we do not leverage the monotonicity property seen in the exact setting. However, the fact that the exact versions of MAX and UNB enjoy this property may explain why these two approaches appear to be quite stable and follow direct paths to minimizers. This behavior is illustrated in Figure 1, and similarly observed in the mixture of Gaussian and Cauchy cases. We can include additional figures in Appendix C to further exemplify this matter.
This paper explores alpha-divergence Variational Inference (VI) from a theoretical perspective, and in particular the monotonic alpha-divergence minimization algorithm. It includes an asymptotic analysis of the algorithm applicable to exponential families, establishing conditions that ensure convergence to a local minimizer at a geometric rate. The theoretical examination of the sample-based counterpart of the algorithm leads to a modified unbiased version, for which both almost sure convergence and a law of the iterated logarithm are provided. Experimental validation using synthetic and real-world datasets illustrates and supports the theoretical findings of the paper.
优点
This is a relevant analysis of alpha-divergence VI algorithms, delivering an in-depth study of their behavior. It is well-written, with technically intricate proofs that are articulated with clarity. Overall, it is a very nice contribution to the field.
缺点
I mainly have minor comments:
- Some assumptions would benefit from more detailed discussion. Specifically, I am unsure why conditions (H4) and (C0') are considered realistic and sensible.
- The impact of the number of samples K used in the sample-based algorithm is underexplored: what influence does it have? Additionally, is it important to maintain the same number of samples as the number of iterations increases?
问题
局限性
Thank you for your thorough review and encouraging feedback on our submission. We provide detailed responses to each of your questions below, hoping to clarify any uncertainties you may have had reading our paper.
I am unsure why conditions (H4) and (C0') are considered realistic and sensible.
The rationale behind assumption (H4) is that as we approach the optimal solution, we can afford to be progressively less conservative in our choice of , since should always remain sufficiently close to the set , or even belong to that set. More specifically, one can show that (H4) holds if the set is compact, meaning that the set of 's that verify is bounded. Consequently, each decrease in the alpha-divergence allows for an increase in the highest permissible value of . We are open to including a concise discussion on this matter in the Appendix.
Regarding (C0’), it is essentially a stricter version of (C0) and involves a design choice left to the practitioner's discretion. It's important to note that (C0) and (H4) are incompatible, which partly explains why we do not get a geometric convergence rate in empirical scenarios.
What influence does the number of samples have in the sample-based algorithm?
Increasing the number of samples reduces the bias of the estimator used in the first sample-based algorithm (MAX). Notably, the estimator becomes asymptotically unbiased, meaning that as , it converges to . This is illustrated in the toy Gaussian experiment, where a small sample size () causes the MAX algorithm to converge to suboptimal parameters with higher alpha-divergence compared to unbiased algorithms. To highlight the impact of sample size on bias, we can include additional plots in Appendix C. Aside from this, a smaller sample size accelerates per-iteration computation but results in noisier estimates of the intractable integrals. This creates a tradeoff, a detailed exploration of which is beyond this paper's scope.
Is it important to maintain the same number of samples as the number of iterations increases?
Theoretically, maintaining the same number of samples across all iterations is useful to satisfy condition (iv) in the proof of Theorem 3, as it guarantees the continuity of the function . However, in practice, one might opt to use fewer samples in the early iterations and increase the number later on to improve the accuracy of the final outcome.
Thank you again for your attentive review and encouraging feedback on our submission, we hope our responses have helped resolve any ambiguities.
We deeply thank the reviewers for their careful and detailed reviews of our manuscript. We are grateful for their constructive feedback and for offering us an opportunity to improve our work. Below, we provide responses to some points that have been raised by multiple reviewers. We hope that this discussion will satisfactorily address any concerns the reviewers may have.
(Reviewer pDqJ) The main theoretical analysis focuses only on exponential families.
(Reviewer FQDG) It only considers the variational family to be exponential models.
We chose to focus on exponential families as they offer a convenient setting and enjoy interesting theoretical properties. First, they ensure the existence of a unique solution to the argmax problem in (3), given that is chosen appropriately (this is not a restrictive assumption). They also allow us to state (H3) in an understandable way, rather than through obscure integral conditions. Finally, we believe that the principles established in this restricted setting offer insights and intuition about the problem at hand, potentially serving as a foundation for future analyses in more general contexts.
(Reviewer pDqJ) For Assumption (H3) in section 3, it remains unclear when the mapping is a contraction. It would be beneficial to further verify this for and in the case of the exponential family.
(Reviewer FQDG) The convergence is proved for the proposed algorithm in the asymptotic sense.
In Proposition 1, we explore the contraction property of the mapping . The analysis reveals that, under assumption (H3), this mapping is a contraction in a neighborhood of the parameter . It can be noticed that when is an exponential family as in (H1) and the target belongs to , i.e. , assumption (H3) is always verified. Indeed, we can show that in this setting, the only fixed point (c.f. Lemma 1) is the target parameter , and that we have .
Determining the size of the neighborhood in which is a contraction is challenging in the general case. However, in the simpler case where is a real Gaussian family with fixed variance and is Gaussian (not necessarily with the same variance), this condition is verified for any mean in .
Again, we thank all reviewers for their comprehensive and constructive feedback. We hope that our responses have adequately addressed your questions and concerns, and that our work has been significantly improved as a result.
This paper provides convergence results for two algorithms for minimizing the alpha-divergence between an exponential family and a target distribution. While reviewers were overall positive and I think this paper has some worthwhile ideas, the paper also has some weaknesses.
The first algorithm analyzed (essentially that in Equation 5) is an algorithm that was previously proposed and proven to decrease monotonically. The algorithm is also in practice unimplementable, as it relies on expectations that are not available for practical target distributions. The contribution of this paper is to give a rate of convergence (theorem 1).
The second algorithm analyzed (essentially that in equation 10) seems to be novel. (And may be valuable). The analysis for this algorithm (theorems 2 and 3) give something similar to a \sqrt{\log(t)/t} rate. This algorithm is implementable. Convergence is proven to some unknown parameter vector μ_* which is in a set of critical points of the alpha divergence. This weakness is exhibited in the fact that the results in Theorem 2 and Theorem 3 make no reference to the minibatch size K, despite the fact that we hope that algorithm to become more accurate when K is larger. The experimental evidence for the algorithm is quite limited.
A weakness of the paper is the form of the assumptions being made. Ultimately, the only sources of variability in the problem are the exponential family chosen, the target distribution, and algorithmic choices (e.g. step sizes). So ultimately, any assumption must be about one of these things. But the form in which the assumptions are given makes it extremely difficult to understand for which target distributions the algorithms are claimed to work. For example, where does the assumption on the target distribution lie in Theorem 1? I asked some reviewers, but could not obtain an answer. I believe the answer is that H3 makes reference to φ^α_η which is defined in equation 2 to be a geometric mixture of the variational family and the target distribution. But what does this mean? For what distributions would this be satisfied? This is not discussed.
Another weakness of the paper is a rather "severe" use of notation, with various symbols seemingly either undefined or at least requiring the reader to hunt through the previous pages to find the definition. If the paper is technical fine, but it would be a great service to the reader in the main assumptions and results to make this clear in the statement, e.g. add a few lines of setup to the theorem clarifying what algorithm is being used, and the previous lines where symbols are defined. Sometimes the notation is quite obscure, e.g. in Eq. 13, Cov_η is covariance with respect to the variational family defined by parameters η, whereas Cov_{φ^α_η} is a covariance with respect the distribution φ^α_η.
In conclusion, while algorithms for alpha divergences are of fundamental interest and this paper contributes valuable new ideas, I believe that the relative lack of clarity of the paper made some of its weaknesses obscure to the readers. After consultation with reviewers and the SAC, I am recommending that the paper be accepted, but I urge that the paper be revised to increase clarity to maximize the potential impact.