Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis
摘要
评审与讨论
This work develops an algorithm for score-based diffusion models over a discrete state space. Their new algorithm relies on the uniformization of the CTMC developed by Chen and Ying and discretization in time. The main innovation of the work is theoretical. Using this new algorithm, the authors can derive bounds on the generated samples under less restrictive assumptions than in previous work. The main assumption is an approximate score function, where analogous to the results on continuous denoising diffusion models, the authors are able to develop a bound with early stopping. Under the further assumption of bounded scores for the data distribution, the authors show that early stopping is no longer needed and can derive bounds at .
优点
The paper is well-written and represents a strong contribution to the generative model literature. The assumptions are weaker than those for existing algorithms. There is a strong case for accepting this paper due to its novel theoretical analysis of an important problem. The theoretical framework given may be useful in analyzing more general settings and motivating further algorithmic improvements.
缺点
First, It is unclear how tight these bounds are. Can the authors compare these to bounds in the continuous setting? Second, I don’t understand everything in the table, and it would be better to flesh out the comparison with Chen and Ying. In particular, their assumptions are different, can you tell us what they are? Finally, there are no tests of the algorithm in practice. While the theoretical contribution is nice, it remains to be seen if it will have an impact on practice.
问题
Can the authors provide more explanation comparing their result to the Chen and Ying result?
Also, can they provide a comparison between their guarantees and the guarantees for continuous models? I am curious if they are analogous or what the fundamental differences are.
Q3. There are no tests of the algorithm in practice. While the theoretical contribution is nice, it remains to be seen if it will have an impact on practice
A3. We appreciate your strong support of our theoretical contribution. While our focus is theoretical, we build upon the CTMC framework that has been empirically validated to be effective in previous works (Campbell et al., 2022; Lou et al., 2024). Our paper belongs to the theoretical line of research in diffusion models, similar to the analysis of continuous diffusion models (Chen et al., 2023b; Benton et al., 2024a) and discrete diffusion models (Chen & Ying, 2024)
References
[1] Sitan Chen, Sinho Chewi, Jerry Li, Yuanzhi Li, Adil Salim, and Anru Zhang. Sampling is as easy as learning the score: theory for diffusion models with minimal data assumptions. In The Eleventh International Conference on Learning Representations, 2023b.
[2] Joe Benton, Valentin De Bortoli, Arnaud Doucet, and George Deligiannidis. Nearly d-linear convergence bounds for diffusion models via stochastic localization. In The Twelfth International Conference on Learning Representations, 2024a.
[3] Hongrui Chen and Lexing Ying. Convergence analysis of discrete diffusion model: Exact implementation through uniformization. arXiv preprint arXiv:2402.08095, 2024.
[4] Campbell, Joe Benton, Valentin De Bortoli, Thomas Rainforth, George Deligiannidis, and Arnaud Doucet. A continuous time framework for discrete denoising models. Advances in Neural Information Processing Systems, 35:28266–28279, 2022.
[5] Aaron Lou, Chenlin Meng, and Stefano Ermon. Discrete diffusion language modeling by estimating the ratios of the data distribution. In International Conference on Machine Learning, 2024.
I appreciate the authors detailed response to my questions, and I maintain that the paper is worthy of acceptance to ICLR.
Thank you for your support. We're glad that our rebuttal has addressed your questions. Your thoughtful comments and feedback have been invaluable to us.
Thank you for the strong support and valuable suggestion!
Q1. Can the authors provide more explanation comparing their result to the Chen and Ying result?
A1. We present a detailed comparison between our results and Chen and Ying (2024) as follows (see Table 1 in our paper):
-
Time-discretized: By utilizing the uniformization technique of CTMC, Chen and Ying (2024) proposed a sampling algorithm that exactly simulates the reverse CTMC. Hence, their convergence bounds get rid of the discretization error term. However, our algorithm discretizes the time and utilizes score estimators on the pre-defined time steps, which introduces a discretization error in the convergence bound that scales with the step size.
-
Score error assumption: Chen and Ying (2024) made the score error assumption: . Since this assumption is made over the time interval with continuous integral on the time which contains the true score functions and the score estimators across the time interval , we, therefore, refer to this assumption as the "continuous score error" in Table 1. In contrast, since we discretize the time in our sampling algorithm in our paper, we make the following score error assumption: , which contains the true score functions and the score estimators at the predefined time steps. We refer to this assumption as the "discretized score error" in Table 1, which can be viewed as a time-discretization of score entropy.
-
Chen and Ying (2024) assumed the bounded score estimator to implement their sampling algorithm. Though we did not make this assumption in the maintext, for the practical implementation of our Algorithm 1, the application of the score clipping discussed in the Appendix essentially applies this bounded score estimator assumption. Namely, since we have known the true bound of the score function, we can ensure that the score estimators are also bounded by score clipping.
-
Both Chen and Ying (2024) and our paper derived the results with and without early stopping. Our algorithm discretizes the time to simulate the reverse process, introducing an additional discretization error term in our bounds. Furthermore, our convergence analysis is conducted in a more general setting, as opposed to the hypercube framework used in their work.
Q2. Also, can they provide a comparison between their guarantees and the guarantees for continuous models? I am curious if they are analogous or what the fundamental differences are.
A2. We provide a comparison between our convergence results for discrete diffusion models and the continuous diffusion models as follows. We choose two typical convergence bounds achieved for continuous diffusion models.
Continuous state:
-
Chen et al. (2023b): , where is the Lipschitz constant of the score function, is the second moment bound of the data distribution, and is the step size.
-
Benton et al. (2024a): , where is related to the time steps, and is the number of time steps.
Discrete state:
- Chen and Ying (2024):
- Ours: (with early stopping) (without early stopping)
These convergence bounds are analogous, for the main KL divergence bounds for convergence comprise three components: score estimation error, discretization error, and the error due to insufficient mixing of the forward process. This is because the exponential integrator sampling procedure in continuous diffusion models discretized the true reverse SDE, and our sampling algorithms discretized the true reverse CTMC, both leading to a discretization error term. We remark that the appearance of the quantities and in our bound, which is absent in the continuous SDE counterpart, stems from the lack of triangle inequality for the Bregman divergence versus distance in the score-matching objective. Hence, we note that a significant difference between the discrete CMTC framework and the continuous SDE framework is the score-matching methods. In addition, we note that our main bounds are nearly linear in the dimension , matching the best result found in continuous diffusion models (Benton et al., 2024a).
This paper provides the convergence rate of the discrete diffusion models introduced by Lou and Ermon. The authors consider the finite state space , extending early analysis by Chen and Ying in the continuous time.
优点
The paper provides a timely convergence analysis of the discrete diffusion models, which is important in understanding this new class of models (with applications to LMM). I checked most proofs, and they are scientifically correct.
缺点
The weaknesses are:
(1) The main idea of proofs are very similar to Chen and Yin, with the key to the proof the representation given by Proposition 1. The differences are: (1) the paper extends the previous results from to ; (2) the paper considers the issue of discretization. I agree that (2) is important, while (1) seems to me incremental.
(2) The authors made Assumption 2, which requires the ratio constant is independent of . I am dubious on this assumption, since the main application of the discrete diffusion models is on the LLM. I would encourage the authors to provide a numerical analysis to show indeed this assumption is valid.
(3) Regarding early stopping: Chen and Yin considers early stopping since they are concerned with the continuous dynamics. It is known that the score matching has large errors near time , and even the continuous dynamics may not be well-defined at . Since the paper studies the discrete scheme, I wonder if there is a particular reason why the authors also consider early stopping.
(4) Except the error for the initialization, which uses log-Sobolev inequality (more or less expected), the other ingredients in the proof are very similar to the existing approaches to the SDE-based diffusion models. Of course, the authors dealt with CDMC, which is slightly different.
问题
Please see the weakness section.
We appreciate the reviewer's support and valuable comments.
Q1. The main idea of proofs are very similar to Chen and Ying, with the key to the proof the representation given by Proposition 1. The differences are: (1) the paper extends the previous results from to ; (2) the paper considers the issue of discretization. I agree that (2) is important, while (1) seems to me incremental.
A1. As detailed in Section 6, our proof techniques include three key components: (A) establishing exponential convergence of the forward process with uniform mixing in the general state space , (B) employing a Girsanov-based method to analyze the discretization error, and (C) deriving novel score movement bounds to control the time-variation of score functions.
Regarding point (1): In practice, the discrete state generally factorizes into , which fits many generative modeling tasks including text generation, image generation, and protein design. This makes our state space setting and sampling algorithm more broadly applicable, as opposed to the hypercube framework used by Chen and Ying (2024).
Regarding point (2): The analysis of discretization error in our discrete-time sampling algorithm is both the most challenging aspect and an important contribution of our work. In particular, our proof leverages Girsanov's Theorem to bound the path measure KL divergence between the true reverse process and the sampling process, while deriving score movement bounds represents a key and novel component of our convergence analysis.
Q2. The authors made Assumption 2, which requires the ratio constant L is independent of d. I am dubious on this assumption, since the main application of the discrete diffusion models is on the LLM. I would encourage the authors to provide a numerical analysis to show indeed this assumption is valid.
A2. We appreciate the reviewer's insights. For the validation of Assumption 2, we provide the following explanation. Assumption 2 is satisfied in many cases, e.g., when the data distribution is the product of independent and identically distributed (i.i.d.) components, each with a marginal distribution fully supported on . In this case, the uniform upper bound in Assumption 2 is thus the uniform upper bound of the score function for . Importantly, depends only on and remains independent of . We have added this supplementary explanation for Assumption 2 to the paper in Lines 378-379 (highlighted in blue).
In addition, we would like to note that Campbell et al. (2022) also made a similar assumption that requires the score uniform bound is dependent on but not on . As they pointed out, the uniform boundness for the data distribution follows trivially from the strict positiveness of if we allow to depend on , and it also makes explicit the dependence of the error bound on the dimension . We adopt this assumption which enables us to derive a bound that can be nearly linear in without early stopping, as stated in Theorem 2.
Q3. Regarding early stopping: Chen and Yin considers early stopping since they are concerned with the continuous dynamics. It is known that the score matching has large errors near time 0, and even the continuous dynamics may not be well-defined at 0. Since the paper studies the discrete scheme, I wonder if there is a particular reason why the authors also consider early stopping.
A3. Indeed, the large errors in score matching near time present a significant challenge. As explained in Lines 299 and 371–372 of our paper, the discrete score function can diverge as for data distributions lacking full support on . The unbounded score at prevents us from establishing uniform upper bounds for the score and its estimators, which are critical for the validity of our key Lemmas 2, 3, 4, and 5. To address this issue, we avoid the zero time point by performing early stopping. However, if we assume that the data distribution has full support on , the early stopping can be removed and we can derive a convergence bound without early stopping.
Q4. Except the error for the initialization, which uses log-Sobolev inequality (more or less expected), the other ingredients in the proof are very similar to the existing approaches to the SDE-based diffusion models. Of course, the authors dealt with CDMC, which is slightly different.
A4. The discretization error terms in our main bounds are novel in the literature on score-based discrete diffusion models. Specifically, in the CTMC setting, we are the first to establish a score movement bound for high-dimensional discrete score functions. This result forms a critical component of our analysis of the discretization error.
In addition, our discrete-time sampling algorithm and methodology are closely related to the previous works on the theory of continuous diffusion models within the SDE framework (Chen et al., 2023b; Benton et al., 2024b), as pointed out in Lines 74-76 of our paper. Inspired by the exponential integrator approach in the SDE framework, our sampling algorithm adopts a similar procedure, and the proof of its convergence leverages techniques analogous to those used for SDE-based diffusion models including Girsanov's Theorem and theoretical properties of score functions such as score movement bound.
References
[1] Hongrui Chen and Lexing Ying. Convergence analysis of discrete diffusion model: Exact implementation through uniformization. arXiv preprint arXiv:2402.08095, 2024.
[2] Sitan Chen, Sinho Chewi, Jerry Li, Yuanzhi Li, Adil Salim, and Anru Zhang. Sampling is as easy as learning the score: theory for diffusion models with minimal data assumptions. In The Eleventh International Conference on Learning Representations, 2023b.
[3] Joe Benton, Yuyang Shi, Valentin De Bortoli, George Deligiannidis, and Arnaud Doucet. From denoising diffusions to denoising markov models. Journal of the Royal Statistical Society Series B: Statistical Methodology, 86(2):286–301, 2024b.
The paper examines theoretical aspects of score-based discrete diffusion models within a Continuous Time Markov Chain (CTMC) framework. The authors aim to address the underexplored convergence properties of these models, especially in discrete state spaces, compared to their continuous counterparts which have been widely studied. Key contributions of the paper include: 1. The authors propose a discrete-time sampling algorithm designed for high-dimensional discrete diffusion tasks. This algorithm leverages score estimators at specific time points to approximate the reverse process of the diffusion model. 2. They provide convergence bounds for the KL divergence and TV distance between the generated sample distribution and the target data distribution. The bounds are derived for cases both with and without early stopping, depending on assumptions about the data distribution's properties. 3.The convergence analysis is performed using a Girsanov-based method, which enables the authors to assess the score estimation error, discretization error, and mixing properties of the forward process. This method is adapted from techniques in continuous diffusion models and tailored to the discrete setting. 4. The paper establishes that their convergence bounds scale nearly linearly with the data dimension, aligning with the best results for continuous models. Additionally, they discuss practical considerations, such as the need for score clipping and early stopping under certain conditions to handle potential score function divergences. 5. The authors compare their method with prior approaches, particularly highlighting differences in sampling efficiency and the elimination of certain assumptions on score estimators, thus broadening the algorithm's applicability in discrete settings.
优点
Overrall I think this work is clear and studies a clean problem which is relevant to the theoretical understanding of diffusion models.. Theorems 1 and 2 derive rigorous convergence bounds for score-based discrete diffusion models using KL divergence (Theorem 1 with early stopping, Theorem 2 without early stopping). This provides a critical theoretical foundation for discrete diffusion modeling, aligning nearly linearly with the dimension , a promising result compared to continuous models. And the use of a Girsanov-based approach to analyze the sampling algorithm in a discrete setting is particularly noteworthy (Sections 5 and 6). This method adapts well-established techniques from continuous diffusion models and is a creative application in the discrete domain, enabling a novel convergence analysis. The work also proposes a time-discretized approach (Section 4 and Algorithm 1), which involves sampling at discrete time steps rather than simulating the continuous CTMC path. This approach allows for more efficient sampling, as it does not require continuous access to the reverse CTMC. Additionally, to ensure bounded score estimates, the authors introduce score clipping as a practical solution for handling extreme values (discussed in Section 5).
缺点
I don't see major weaknesses of this work, but I do have some comments:
- Since the authors claim improved sampling efficiency (e.g., fewer function evaluations), would compare actual runtime or convergence speed with other discrete sampling methods (such as the uniformization technique in Chen & Ying) substantiate these claims?
- Assumption 2 requires the data distribution to have full support and uniform bounds. This may restrict the model.
- Although the work mentions similarities with continuous diffusion models, there is limited discussion on when to use their discrete CTMC approach versus discretized continuous models for discrete data.
- Generalizability to Other State Spaces: The works' algorithm assumes a general state space , but would it be insightful to discuss how it might extend to more complex or structured discrete spaces, such as graph-based or hierarchical state spaces, which are common in discrete data applications?
问题
- Have the authors considered adaptive step sizes as a way to minimize discretization error? If so, could author's share insights on how this might affect the overall convergence bound?
- The paper proposes a CTMC-based discrete diffusion model, but certain continuous diffusion approaches are also adaptable to discrete data. Could the authors elaborate on specific scenarios or types of tasks where their CTMC-based model would be preferred over these alternative approaches?
- Also see weaknesses part above
Thank you for your strong support and valuable feedback!
Q1. Since the authors claim improved sampling efficiency (e.g., fewer function evaluations), would compare actual runtime or convergence speed with other discrete sampling methods (such as the uniformization technique in Chen & Ying) substantiate these claims?
A1. We appreciate the reviewer's comment and acknowledge that our claim about improved efficiency requires revision. Although we derived an approximate sampling complexity of for our algorithm, i.e., running our sampling algorithm requires steps through uniformization (see Discussion on in Appendix D in our paper), this does not quantitatively demonstrate that our algorithm is more efficient than other discrete sampling methods, such as the one in Chen and Ying (2024), which has a sampling complexity of with . We have highlighted the following revision in blue on Lines 423-424: our algorithm discretizes the time to simulate the reverse process and calls the score estimator at fixed discretization points instead of randomly sampled times, as done in Chen and Ying (2024).
Q2. Assumption 2 requires the data distribution to have full support and uniform bounds. This may restrict the model.
A2. Assumption 2 is an assumption for the data distribution to give a further convergence result that is nearly linear in the dimension without early stopping. However, Assumption 2 holds in many cases, e.g., if the data distribution is the product of independent and identically distributed (i.i.d.) components, each with a marginal distribution fully supported on . We have added this supplementary explanation for Assumption 2 in our paper on Lines 378-379 (highlighted in blue). Additionally, we note that similar assumptions are also made by Chen and Ying (2024) (aimed to remove early stopping) and Campbell et al. (2022) for theoretical analysis.
Q3. Generalizability to Other State Spaces: The works' algorithm assumes a general state space , but would it be insightful to discuss how it might extend to more complex or structured discrete spaces, such as graph-based or hierarchical state spaces, which are common in discrete data applications?
A3. We appreciate the reviewer’s insightful question. Extending our methodology and analysis to more complex or structured discrete spaces is promising and interesting. In each specific setting, exploring the corresponding discrete score function and score-matching methods would be essential to design sampling algorithms. For graph-based state spaces, our framework shows direct applicability. Building on Vignac et al. (2022), we can perform the diffusion processes separately on each node and edge feature within a graph setting. In this setting, the state space comprises the discrete node types and edge types , which can be addressed as instances of our one-dimensional state space framework where . For any given node or edge, we can define distinct CTMC diffusion processes and reverse sampling processes. Investigating and validating this extension would be an exciting direction for future research.
Q4. Have the authors considered adaptive step sizes as a way to minimize discretization error? If so, could author's share insights on how this might affect the overall convergence bound?
A4. As an initial exploration of discrete-time sampling algorithms and corresponding convergence analyses for score-based discrete diffusion models, we did not incorporate adaptive step sizes, as proposed by Chen et al. (2023a) and Benton et al. (2024a) for continuous diffusion models. Instead, we followed earlier works on the convergence of continuous diffusion models and adopted a constant step size for analytical simplicity, as done in Chen et al. (2023b) and Yang and Wibisono (2022). Scheduling the step sizes indeed has the potential to decrease the discretization error. According to Lemma 3 in our paper, when is small, the score movement bound can be very large, particularly for unbalanced data distributions, leading to a significant discretization error term when is small. Hence, analogous to the exponential decay step size schedules suggested by Chen et al. (2023a) and Benton et al. (2024a), applying a decaying step size schedule to slow down the CTMC as may help decrease the overall discretization error. The critical challenge in determining an optimal step size schedule lies in deriving the score movement bound for general step sizes, which may involve a complex mathematical formulation and requires further investigation. We would leave this as a direction for future research.
Q5. The paper proposes a CTMC-based discrete diffusion model, but certain continuous diffusion approaches are also adaptable to discrete data. Could the authors elaborate on specific scenarios or types of tasks where their CTMC-based model would be preferred over these alternative approaches?
A5. Text generation is a typical task where CTMC-based discrete diffusion models are particularly well-suited due to the inherently discrete nature of the text. Zheng et al. (2023) empirically showed that the Discrete Denoising Diffusion Probabilistic Model (D3PM) (Austin et al., 2021) outperforms continuous diffusion models for text generation tasks (see Table 1 in their paper). Furthermore, Campbell et al. (2022) demonstrated that the CTMC framework offers significantly greater flexibility in defining reverse sampling schemes compared to discrete-time diffusion model approaches such as D3PM. This implies the advantages of using the CTMC framework over continuous diffusion methods for text generation. Additionally, Lou et al. (2024) showed that CTMC-based discrete diffusion models are competitive with autoregressive models, achieving superior performance over GPT-2 in particular. These empirical studies suggest that CTMC-based models would be preferred over these alternative approaches for discrete data generation tasks.
References
[1] Jacob Austin, Daniel D Johnson, Jonathan Ho, Daniel Tarlow, and Rianne Van Den Berg. Structured denoising diffusion models in discrete state-spaces. Advances in Neural Information Processing Systems, 34:17981–17993, 2021.
[2] Joe Benton, Valentin De Bortoli, Arnaud Doucet, and George Deligiannidis. Nearly d-linear convergence bounds for diffusion models via stochastic localization. In The Twelfth International Conference on Learning Representations, 2024a.
[3] Andrew Campbell, Joe Benton, Valentin De Bortoli, Thomas Rainforth, George Deligiannidis, and Arnaud Doucet. A continuous time framework for discrete denoising models. Advances in Neural Information Processing Systems, 35:28266–28279, 2022.
[4] Lin Zheng, Jianbo Yuan, Lei Yu, and Lingpeng Kong. A reparameterized discrete diffusion model for text generation. arXiv preprint arXiv:2302.05737,2023.
[5] Clement Vignac, Igor Krawczuk, Antoine Siraudin, Bohan Wang, Volkan Cevher, and Pascal Frossard. Digress: Discrete denoising diffusion for graph generation. arXiv preprint arXiv:2209.14734, 2022
[6] Aaron Lou, Chenlin Meng, and Stefano Ermon. Discrete diffusion language modeling by estimating the ratios of the data distribution. In International Conference on Machine Learning, 2024.
[7] Hongrui Chen and Lexing Ying. Convergence analysis of discrete diffusion model: Exact implementation through uniformization. arXiv preprint arXiv:2402.08095, 2024.
[8] Hongrui Chen, Holden Lee, and Jianfeng Lu. Improved analysis of score-based generative modeling: User-friendly bounds under minimal smoothness assumptions. In International Conference on Machine Learning, pages 4735–4763. PMLR, 2023a.
[9] Sitan Chen, Sinho Chewi, Jerry Li, Yuanzhi Li, Adil Salim, and Anru Zhang. Sampling is as easy as learning the score: theory for diffusion models with minimal data assumptions. In The Eleventh International Conference on Learning Representations, 2023b.
[10] Kaylee Yingxi Yang and Andre Wibisono. Convergence of the inexact langevin algorithm and score-based generative models in kl divergence. arXiv preprint arXiv:2211.01512,2022.
The paper presents a discrete-time sampling method using score estimators at fixed points in a multidimensional state space. It provides bounds on how closely the sample matches the data distribution, with KL divergence bounds nearly linear in dimension, comparable to top diffusion models. The analysis uses a Girsanov-based approach to define key properties of the discrete score function for effective sampling.
优点
The paper tackles an interesting theoretical problem. Theoretical analysis of diffusion models with finite state spaces with discrete time space is of great importance. The authors have made the effort to push further the existing methods for the current setting.
缺点
The notation is confusing and important details are often omitted. The assumptions seem to be rather stringent. The paper lacks mathematical clarity. See the below sections for details.
Mathematical comments
- Proposition 1. The formula for depends on the initial data distribution . When , the term on the right side . The latter is a fixed matrix and thus . This is not necessarily the uniform distribution.
- Proposition 1 What does approaching mean here? In what sense do you have the convergence?
- line 280 . The sentence
With rate $Q^{\leftarrow}_t$, it holds that...is not clear. Is there a reference or proof for this statement? - Assumption 2 It does not seem to be easy to verify this assumption. Is there a general class of distributions that are known to satisfy it?
- Theorem 2 The number is not always well-defined, as we deal with discrete distributions, where some of the probabilities may be equal to zero.
typos
- line 187 if setting β(t) as a time-dependent scala
- line 206 Nota -> Note
- line 323 For completeness, We -> For completeness, we
问题
- Equation on line 723. Why is equal to the Kronecker sum ?
- line 296 Why is the reverse process time-inhomogeneous?
- line 3 of the Algorithm: How easy is it to find the maximum of the estimated score function? Is there any concavity assumption to make this problem solvable in theory/real-time? Is this assumption satisfied for practically relevant examples?
- line 361 Why is there a uniform upper bound on all the scores and their estimators? On line 371 the authors mention that the score function may be as large as infinity for some data points. This means, that when is small, this uniform upper bound becomes larger, thus, it is dependent on . Can this dependence be quantified?
- Equation (4). Why is this true? Is there a reference or a proof for this statement?
Q5. Theorem 2 The number is not always well-defined, as we deal with discrete distributions, where some of the probabilities may be equal to zero.
A5. We would like to clarify this misunderstanding. Theorem 2 holds under Assumption 2 which ensures that the data distribution is strictly positive. The strict positiveness of naturally leads to the strict positiveness of marginal distribution as for any and . Thus, is well-defined and would not blow up or equal to zero. We have clarified this point on Lines 392-393 in our revised paper.
Q6. Typos
A6. We appreciate the reviewer for identifying these typos. We have corrected them accordingly on Lines 211 and 322.
Q7. Equation on line 723. Why is equal to the Kronecker sum ?
A7. The Kronecker product structure of was inspired by Chen and Ying (2024, Proposition 3) where is considered. This structure can be verified by direct calculation from the definition of and :
First, the transition rate matrix for each dimension satisfies that for and for .
We next calculate the Kronecker sum . By the definition of the Kronecker product, we know that for the Kronecker product of square matrices , its elements can be expressed using the following indexing relationship: , where and are multi-dimensional indices representing the specific row and column positions in the space. Thus, by the structure of and the identity matrix , we can see that for any , , and , it holds that
Similarly, we have that for any ,
Hence, we conclude that
We have added these mathematical details in our revised paper on Lines 720-745.
Q8. line 296 Why is the reverse process time-inhomogeneous?
A8. The reverse rate matrix is time-dependent, which indicates that the reverse process is time-inhomogeneous, according to the definition of a time-inhomogeneous CTMC (see, e.g., Holmes-Cerfon (2022)).
We appreciate your detailed review. Our responses are provided below, and all revisions have been highlighted in blue in the revised manuscript.
Q1. Proposition 1. The formula for depends on the initial data distribution . When , the term on the right side . The latter is a fixed matrix and thus . This is not necessarily uniform distribution.
A1. This is a misunderstanding and we would like to clarify this mathematical detail here. When , . This yields that , where is the uniform distribution on and the equality holds because is a discrete probability distribution whose entries sum to 1. This means that .
We have added this clarification to the revised version of our paper on Line 755.
Q2. Proposition 1 What does approaching mean here? In what sense do you have convergence?
A2. The marginal distribution has each component defined as a function of . Here, "approaching" refers to taking the limit as . As , the marginal distribution converges, in the sense of a function limit, to , the uniform distribution over .
Q3. line 280 . The sentence 'With rate , it holds that...' is not clear. Is there a reference or proof for this statement?
A3. This result directly follows from Campbell et al. (2022, Proposition 3), as cited in Line 269. Proposition 3 in Campbell et al. (2022) explicitly provides the expression for the reverse rate matrix corresponding to a given forward rate matrix. The CTMC with this reverse rate is the exact time reversal of the forward CTMC. For clarity, we have revised the statement in our paper on Lines 269-274: According to Campbell et al. (2022), the reverse process can be achieved by a CTMC starting from with , and the reverse rate is of the form
Q4. Assumption 2 It does not seem to be easy to verify this assumption. Is there a general class of distributions that are known to satisfy it?
A4. Assumption 2 is satisfied by a broad class of practical distributions. A particularly important example is when the data distribution consists of independent and identically distributed (i.i.d.) components, where each marginal distribution has full support on . In this case, has the form of Kronecker product decomposition as where is the marginal distribution of the -th token of the data on the state space and By independence, we have that (, , and ), which is exactly the score function of the marginal distribution . Thus, we know that the uniform upper bound in Assumption 2 is exactly the uniform upper bound of the score of , which is only dependent on but not on .
We have added this supplementary explanation for Assumption 2 to the paper on Lines 378-379.
Q9. line 3 of the Algorithm: How easy is it to find the maximum of the estimated score function? Is there any concavity assumption to make this problem solvable in theory/real-time? Is this assumption satisfied for practically relevant examples?
A9. We appreciate the reviewer's insightful observation. In Line 3 of Algorithm 1, we have set , where the maximum is obtained over the discrete set . When is relatively small, taking the maximum is practically reasonable. When is large, it may be impractical to get the exact maximum. By the uniformization of CTMC (Chen and Ying, 2024), we can instead get an upper bound for to implement Algorithm 1. By applying score clipping as discussed in Appendix A.4, we can ensure that for all . Therefore, we can set which is tractable given , and thus circumvent obtaining the maximum over the set .
In the revised version of our paper, we have added the discussion on on Lines 1330-1338 to clarify this point.
Q10. Line 361 Why is there a uniform upper bound on all the scores and their estimators?
A10. The fact that and are the uniform upper bounds for all the scores and their estimators has been deferred to the Appendix. By Lemmas 2 and 4 in Appendix A, we know that and are uniform bounds for all the scores and their estimators with and without early stopping respectively ( and are defined in Theorem 1 and Theorem 2 respectively). See also discussions on and in Appendix A.4 (Lines 961-971 in our paper).
We have added the explanation for this statement on Line 362 of the revised paper.
Q11. On line 371 the authors mention that the score function may be as large as infinity for some data points. This means, that when is small, this uniform upper bound becomes larger, thus, it is dependent on . Can this dependence be quantified?
A11. Yes, the dependence on of is quantified in Theorem 1 (Line 349): C_1=\max\left\\{1+\frac{S}{\delta},\max\_{\mathbf{x}\in\mathcal{X}, k\in\\{0,\cdots,K-1\\}}\Vert\hat{s}\_{(k+1)h+\delta}(\mathbf{x})\Vert_\infty\right\\} . By applying score clipping, we can ensure that (see discussions on in Appendix A.4 (Lines 961-971).
Q12. Equation (4). Why is this true? Is there a reference or a proof for this statement?
A12. By Campbell et al. (2022, Proposition 3), is the marginal along the reverse process : . With the rate matrix of the reverse CTMC, we can write out the marginal satisfies the Kolmogorov equation (see, e.g., Campbell et al. (2022); Chewi (2023), as cited in Line 186 of the revised paper):
References
[1] Andrew Campbell, Joe Benton, Valentin De Bortoli, Thomas Rainforth, George Deligiannidis, and Arnaud Doucet. A continuous time framework for discrete denoising models. Advances in Neural Information Processing Systems, 35:28266–28279, 2022.
[2] Hongrui Chen and Lexing Ying. Convergence analysis of discrete diffusion model: Exact implementation through uniformization. arXiv preprint arXiv:2402.08095, 2024.
[3] Sinho Chewi. Log-concave sampling, 2023. Book draft available at https://chewisinho.github.io.
[4] Miranda Holmes-Cerfon. Lecture 4: Continuous-time Markov Chains, 2022. URL https://personal.math.ubc.ca/~holmescerfon/teaching/asa22/handout-Lecture4_2022.pdf Lecture notes, University of British Columbia.
My concerns were properly addressed. I raised my score.
Thank you for raising your score. We appreciate your detailed review and are glad that our rebuttal has addressed your concerns.
Dear Reviewer Gwu3,
Thank you for your detailed review. We have carefully addressed each point in our response and revised the manuscript accordingly. We would be grateful for any additional feedback or questions you may have.
Best,
Authors
Dear Reviewer Gwu3,
Thank you for your detailed review. We have addressed your concerns point by point in our rebuttal and made corresponding revisions. Given the approaching PDF revision deadline, your feedback would be greatly appreciated, particularly if our explanation of the distribution convergence results in Proposition 1 addresses your concerns. We would be happy to address any additional questions you may have.
Best regards,
Authors
This paper offers a convergence analysis of score-based discrete diffusion model when the state space is , generalizing the prior work of Chen and Ying that handles the case of . The work leverages heavily the techniques from Chen and Ying, but also offers new ingredients to handle the generalization. Given the importance of discrete diffusion, the work provides a useful contribution to the theoretical understanding of such models.
审稿人讨论附加意见
The rebuttal has helped clarify the contribution of this work in relation to existing works, as well as some technical details, leading to a consensus score of acceptance.
Accept (Poster)