PaperHub
6.5
/10
Poster4 位审稿人
最低6最高8标准差0.9
8
6
6
6
3.0
置信度
正确性2.8
贡献度2.5
表达2.8
ICLR 2025

LANTERN: Accelerating Visual Autoregressive Models with Relaxed Speculative Decoding

OpenReviewPDF
提交: 2024-09-28更新: 2025-03-02

摘要

Auto-Regressive (AR) models have recently gained prominence in image generation, often matching or even surpassing the performance of diffusion models. However, one major limitation of AR models is their sequential nature, which processes tokens one at a time, slowing down generation compared to models like GANs or diffusion-based methods that operate more efficiently. While speculative decoding has proven effective for accelerating LLMs by generating multiple tokens in a single forward, its application in visual AR models remains largely unexplored. In this work, we identify a challenge in this setting, which we term token selection ambiguity, wherein visual AR models frequently assign uniformly low probabilities to tokens, hampering the performance of speculative decoding. To overcome this challenge, we propose a relaxed acceptance condition referred to as LANTERN that leverages the interchangeability of tokens in latent space. This relaxation restores the effectiveness of speculative decoding in visual AR models by enabling more flexible use of candidate tokens that would otherwise be prematurely rejected. Furthermore, by incorporating a total variation distance bound, we ensure that these speed gains are achieved without significantly compromising image quality or semantic coherence. Experimental results demonstrate the efficacy of our method in providing a substantial speed-up over speculative decoding. In specific, compared to a na\"ive application of the state-of-the-art speculative decoding, LANTERN increases speed-ups by $\mathbf{1.75}\times$ and $\mathbf{1.82}\times$, as compared to greedy decoding and random sampling, respectively, when applied to LlamaGen, a contemporary visual AR model. The code is publicly available at https://github.com/jadohu/LANTERN.
关键词
Speculative decodingVisual Autoregressive Models

评审与讨论

审稿意见
8

This paper tackles the challenge of transferring speculative decoding from LLMs to autoregressive image generation models. The authors observe that the nature of image data causes image tokens to exhibit selection ambiguity, resulting in high rejection rates and poor acceleration effects in speculative decoding. Hence they propose using the proximity set of image tokens as proxies to speculate on accepting a predicted image token, and introduce a combinatorial optimization strategy for selecting the set to ensure image quality remains intact. Extensive experiments confirm significant acceleration with minimal quality loss.

优点

The paper's structure and presentation are well-crafted. It identifies the phenomenon of "token ambiguity" and provides a viable solution. Extensive experiments, particularly the ablation studies, demonstrate the effectiveness and scalability of the method. It offers valuable insights for research on inference acceleration in AR text-to-image generation. Although the idea of using a set to proxy a single point has been extensively researched, the authors have elegantly applied it to the sampling decision process and achieved excellent results. Simplicity yet effectiveness is crucial.

缺点

1.For a specific image token, the algorithm for finding its proximity set , specifically "Find the neighborhood (i.e., Appendix B, Line 9)," lacks detailed discussion on efficiency. For example, for each quantized image token, are the corresponding sets A and B precomputed or dynamically calculated during inference? How much time does this part take? I recommend the authors supplement this section with analysis and discussion.

2.In Line 105, there might be a misstatement. EAGLE-2 is an acceleration decoding method, not a drafter.

3.To demonstrate the generalizability of the observation of "token ambiguity," i.e., that this phenomenon is present in most AR image generation models, the paper should include experiments on a broader range of AR image generation models beyond just LLaMAGen. This would enhance the experimental completeness and provide stronger evidence for the generality of the observed issue.

问题

1.There's a suggestion that the authors could provide the details of "Find the neighborhood" in Appendix A.

2.A suggestion is (since the experimental section of the current paper is already sufficient, this is just a gentle suggestion) that the authors could discuss more about the differences in token ambiguity between text generation and image generation. For example, in LLaMAGen, what common properties do tokens within the same set have? What are the patterns in the sizes of sets A and B corresponding to different image tokens? These discussions could help in designing further algorithms.

3.Could validation experiments be conducted on other AR models? This is also a optional suggestion aimed at enhancing the generalization and robustness of the method.

伦理问题详情

There are no ethical concerns in this paper.

评论

First of all, we sincerely appreciate the detailed review of our paper and the constructive feedback provided. Thank you for highlighting the clarity of our presentation and recognizing the novelty and elegance of our approach to tackle token selection ambiguity. We believe your feedback has significantly strengthened our work. To address the concerns and questions, we have conducted additional experiments and revised the manuscript as follows:


Weakness 1 & Question 1: Finding Neighborhood

To identify the proximity sets A and B of quantized image tokens, we employed both a dynamic approach and precomputation. Specifically:

  • Precomputation of BB: The set BB comprises the kk nearest tokens in the latent space with 2\ell_2 distance for each quantized image token. Since BB is independent of token probabilities (and preceding tokens), it is precomputed and stored as tensors before inference begins.
  • Dynamic Calculation of AA: The set AA is a subset of BB, determined dynamically based on token probabilities and an upper bound δ\delta. This ensures AA adapts to the current decoding context.

We have summarized the overall procedure as pseudocode and included it as Appendix D.2 in the revised manuscript.

To evaluate the impact of these computations on overall latency, we measured the time required to compute AA during a single inference step and compared it with the forward times for the target model and drafter. Target model is LlamaGen Stage I and randomly selected 1000 caption from MS-COCO 2014 validation set are used. The results are summarized below:

Target forwardDrafter forwardProximity set AA calculation
LANTERN (δ=0.40,k=1000)(\delta=0.40, k=1000)3.80×1023.80\times10^{-2}s1.08×1021.08\times10^{-2}s1.19×1031.19\times10^{-3}s

Our analysis shows that the time required to compute AA is 32 times smaller than drafter model forward and 9.1 times smaller than target model forward for k=1000,δ=0.4k=1000, \delta=0.4, which is negligible in entire process. Furthermore, the significant speedup enabled by AA also confirms the efficiency of our method.

These findings have been added to Appendix F.2, along with further discussion on implementation details. We appreciate your suggestion, which has enhanced the clarity and comprehensiveness of our work.


Weakness 2: Misstatement in Line 105

We acknowledge the incorrect description of EAGLE-2 as a drafter in Line 105. This has been corrected to EAGLE-2 as a base speculative decoding method at Line 105 in the revised manuscript. Thank you for pointing this out, allowing us to rectify this misstatement.

Weakness 3 & Question 3: Generalizability of the Token Selection Ambiguity and LANTERN

Thank you for your valuable suggestion on the generalizability of token selection ambiguity and our method LANTERN to other visual AR models. To demonstrate the generalizability of the token selection ambiguity and our method, we conducted additional experiments on other visual AR models, including LlamaGen Stage II model and Anole [1] with MS-COCO 2017 Validation set.

ModelsAverage Top-1 probabilitiesAverage Top-10 probabilitiesDrafter Test Accuracy
Vicuna-13B0.7870.7870.9890.98984.66%84.66\%
LlamaGen-XL0.2060.2060.5200.52038.80%38.80\%
Anole0.0640.0640.2040.20427.60%27.60\%

As shown in the table, like LlamaGen, Anole exhibits low average top-1 and top-10 probabilities, with values of 0.0640.064 and 0.2040.204, respectively. However, these probabilities are even lower than those of LlamaGen (0.2060.206 and 0.5200.520), suggesting that Anole faces a more severe degree of token selection ambiguity. This increased ambiguity also affects drafter performance, with Anole’s drafter achieving a test accuracy of only 27.60%27.60\%, compared to 38.80%38.80\% for LlamaGen.

By extending the analysis to Anole, we confirm that the token selection ambiguity is a recurring challenge across different models, with varying degrees of severity. This further substantiates our claim that token selection ambiguity is a critical factor influencing the performance of speculative decoding. The results have been incorporated into the Figure 2(b) and (c) in the revised manuscript.

评论
LlamaGen Stage II, τ=1\tau=1SpeedupMean Accepted LengthFIDCLIPScorePrecision / RecallHPS v2
Vanilla AR1.00×1.00\times1.00×1.00\times47.6047.600.29390.29390.41380.4138 / 0.56480.564823.8423.84
EAGLE-20.96×0.96\times1.22×1.22\times----
LANTERN (δ=0.40,k=1000)\delta=0.40, k=1000)1.64×1.64\times2.24×2.24\times46.1046.100.29250.29250.47040.4704 / 0.52220.522223.0623.06
Anole, τ=1\tau=1SpeedupMean Accepted LengthFIDCLIPScorePrecision / RecallHPS v2
Vanilla AR1.00×1.00\times1.00×1.00\times20.2720.270.32150.32150.65520.6552 / 0.63980.639823.5223.52
EAGLE-20.73×0.73\times1.10×1.10\times----
LANTERN (δ=0.50,k=100)\delta=0.50, k=100)1.17×1.17\times1.83×1.83\times23.4023.400.31860.31860.60260.6026 / 0.61780.617822.9222.92

The results show that LANTERN consistently achieves substantial acceleration (with speed-ups of 1.71×1.71\times and 1.60×1.60\times over EAGLE-2 on LlamaGen Stage II and Anole, respectively) while maintaining competitive image quality. However, its performance, in both acceleration and image quality, varies depending on the degree of token selection ambiguity inherent to the underlying model. In particular, Anole's results are somewhat less favorable compared to those of LlamaGen, a difference attributable to the distinct characteristics of the two models.

As discussed earlier, Anole exhibits a higher degree of token selection ambiguity compared to LlamaGen, resulting in lower drafter performance. This discrepancy impacts the outcomes for both EAGLE-2 and LANTERN on Anole. Nevertheless, it is essential to note that this variation does not represent a limitation of LANTERN itself but rather reflects the impact of model-specific factors, such as token selection ambiguity. Despite these challenges, LANTERN continues to outperform EAGLE-2 on Anole in terms of acceleration and maintains reasonable image quality, highlighting its robustness across different models.

Furthermore, these findings open avenues for future research. Improving drafter performance through enhanced architectures or training approaches that address token selection ambiguity could further optimize results for models with higher levels of ambiguity. While the current results affirm LANTERN’s robustness and versatility, addressing these factors could lead to even greater performance improvements in the future. We sincerely thank you for raising these important questions, which have allowed us to explore these new research directions and better understand the role of token selection ambiguity in model performance.

We have included the these results to appendix E in the revised manuscript, as we aimed to minimize significant changes to the main content of the manuscript. However, if you feel that incorporating these results into the main table would enhance clarity and comprehensiveness, we would be happy to expand the main table in the next revision. We sincerely appreciate all of your valuable feedback and remain committed to improving the manuscript in line with your suggestions.

评论

Question 2: Further Analysis on Token Ambiguity

We sincerely thank you for your thoughtful suggestion and for recognizing the potential value of further discussions on token ambiguity. We conducted additional analysis to better understand the nature of token ambiguity:

  1. Visualization of Proximity Tokens for Visual AR Models: We visualized how tokens in BB decode into image patches with LlamaGen Stage II model. The visualizations reveal that tokens in BB correspond to visually similar patches, validating the hypothesis of latent proximity in image generation. The results have been incorporated into Appendix C.2 and the same figures can be found in this anonymous links: Sample 1 Sample 2.

  2. Proximity Token Sets in Language AR Models: We also conducted an investigation into proximity token sets for language models (Vicuna-7B). Since language models do not utilize encoders like VQVAE in visual AR models, we examined the proximity of tokens based on their input embeddings (representations obtained through the embedding layer). As expected, we observed that, unlike visual AR models, tokens in language models do not display clear semantic similarity among those considered "close" in the embedding space. While this finding may not be surprising, it underscores a fundamental difference in how token relationships are structured across the two domains.

    Token: hiTop-1Top-2Top-3Top-4Top-5
    Nearest_Portály_Mediabestanden<0x4C><0x6B><0x49>
    Farthest[Line Seperator][Object Replacement Character]_infinitely_firewall_sooner
    Token: actTop-1Top-2Top-3Top-4Top-5
    Nearest_Mediabestandenoreferrer<0x24><0x71><0x54>
    Farthest[Object Replacement Character]_Bruno_Ernst_Santos_firewall
  3. Patterns in AA Sizes: Since AA is determined dynamically, we analyzed how its size varies across different positions in the generated image. With LlamaGen Stage I model and randomly sampled 100 captions in MS-COCO 2014 validation captions, we calculate avrage size of AA with respect to its position in image. First, at high δ\delta, the size of AA is generally large. Additionally, it was observed that the average size of AA tend to be larger at the left end of image. We hypotheize that this is due to the higher uncertainty at the left side caused by line change of image. Since the probabilities assigned to individual tokens are relatively smaller at the left end, allowing more tokens to be included for the same δ\delta, which increases the average size of AA. You can find actual heatmap for k=1000k=1000 and δ=0.1,0.4\delta=0.1, 0.4 in following anonymous links: delta=0.1 delta=0.4.

These insights have been incorporated into the revised manuscript, as Appendix C and G.

We hope these revisions and additional experiments address your concerns comprehensively. We deeply value your thoughtful and constructive input, which has played a pivotal role in enhancing the rigor and quality of our work. It has been a privilege to engage with your insightful questions and critiques. Thank you once again for your careful consideration of our manuscript and for giving us the opportunity to further refine our contribution.

References

[1] Chern, Ethan, et al. "Anole: An open, autoregressive, native large multimodal models for interleaved image-text generation." arXiv preprint arXiv:2407.06135 (2024).

评论

Thank you for the detailed feedback and experiments. I think most of my concerns have been addressed, so I will raise my score to an Accept.

评论

Thank you for taking the time to provide such thoughtful feedback and for your consideration of our work. We’re glad to hear that your concerns have been addressed, and we truly appreciate your support and the raised score.

审稿意见
6

This paper proposed speculative decoding for AR Image generation, AR models tend to assign uniformly low probabilities across a wide range of tokens, making it difficult to select the most appropriate token during decoding.

To solve this problem, the author introduce LANETERN (Latent Neighbor Token Acceptance Relaxation), leveraging the interchangeability of the tokens and relaxing the acceptance for decoding.

The benefits of LANETREN can be found at generating speed with comparable quality.

优点

  1. The motivation behind the author’s proposed approach is straightforward: in standard sampling processes, an excessively high token rejection rate can significantly slow down generation. The author suggests a more relaxed policy to mitigate this issue, thereby improving generation speed.

  2. To mitigate the distributional shift introduced by the lenient sampling approach, the author proposes a constraint rule for sampling based on total variation.

  3. Through extensive ablation experiments, the author demonstrates how the lenient sampling scheme influences sampling speed and elucidates the specific roles of various hyperparameters within this scheme.

缺点

My main concern lies in whether the metrics employed in the experimental section sufficiently capture the effectiveness of the proposed approach, particularly in Table 3 and Section 4.2.3. For evaluating basic generation quality, would it be possible to provide precision and recall (P/R) instead of FID? This change could allow for a clearer view of how LANTERN and limited distribution divergence affect and restore distribution. Additionally, since LlamaGen is used as the base model, could the authors include some modern image quality scores, such as PickScore or HPS, to better quantify aesthetic quality loss?

问题

see weakness.

In summary, the author presents a promising sampling scheme for AR models, which improves sampling speed at the cost of some image quality.

The primary concern is that the author’s experiments lack sufficient evidence to demonstrate that the loss in image quality is within an acceptable range. Given this limitation in quality, I believe the impact of the work may be constrained. Therefore, I am inclined to recommend rejection.

评论

We sincerely thank you for your thoughtful feedback and constructive suggestions. We appreciate your recognition of the motivation and soundness of our approach, as well as your valuable insights on evaluation metrics and image quality. Your comments provide important guidance, and we address them in detail below.


Metrics for Evaluating Basic Generation Quality (Precision and Recall)

We appreciate your suggestion to include precision and recall (P/R) metrics to better understand the quality and diversity trade-offs in our approach. To this end, we have evaluated LANTERN with varying values of the hyperparameters δ\delta and kk, and compared its performance to the baseline. The results are as follows:

  • Precision / Recall for standard AR: 0.47810.4781 / 0.56330.5633

  • LANTERN (τ=1.0\tau=1.0)

    Precision / Recallδ=0.05\delta=0.05δ=0.1\delta=0.1δ=0.2\delta=0.2δ=0.4\delta=0.4
    k=100k=1000.48670.4867 / 0.53890.53890.47960.4796 / 0.53030.53030.47890.4789 / 0.51400.51400.48250.4825 / 0.49460.4946
    k=300k=3000.48560.4856 / 0.53670.53670.48340.4834 / 0.52310.52310.48940.4894 / 0.49010.49010.48950.4895 / 0.47190.4719
    k=1000k=10000.48650.4865 / 0.53340.53340.48690.4869 / 0.51720.51720.48800.4880 / 0.48880.48880.49090.4909 / 0.44970.4497

The results demonstrate that LANTERN achieves comparable or slightly improved precision relative to the baseline across various settings, highlighting its ability to maintain the quality of generated images. While recall decreases slightly with increasing δ\delta, this indicates that the quality of individual images is preserved, though there is a modest reduction in diversity.

The slight decrease in recall can be attributed to token selection ambiguity in the drafter. When the drafter is not optimally trained, it may struggle to produce sufficiently diverse or accurate predictions. Consequently, increasing the acceptance probability enhances acceleration but can lead to reduced image diversity. Nevertheless, this trade-off does not significantly impact overall image quality, as evidenced by consistent FID, Precision, and HPS v2 scores (in the next section). These precision and recall results illustrate that our method effectively balances quality and diversity across different hyperparameter configurations.

Overall, these evaluations provide further evidence that LANTERN maintains generation quality within acceptable bounds while achieving substantial speed improvements.


Incorporation of Modern Image Quality Metrics (PickScore and HPS v2)

To further quantify the aesthetic quality of generated images, we have evaluated our approach using HPS v2 [1], as suggested. Evaluating PickScore [2] was challenging due to its high time requirements, which involve measuring Elo ratings or win rates for each individual samples to properly evaluate it. Both PickScore and HPSv2 are metrics for assessing human preference; however, HPSv2 reportedly aligns more closely with human preference scores. Therefore, we chose to use only HPSv2 in our evaluation. This modern image quality metrics provide a more holistic measure of perceptual quality and aesthetic appeal. The updated findings are as follows:

| HPS v2

Vanilla AR : 24.11δ=0.05\delta=0.05δ=0.1\delta=0.1δ=0.2\delta=0.2δ=0.4\delta=0.4
k=100k=10024.0124.0123.9423.9423.8623.8623.7523.75
k=300k=30023.9723.9723.8523.8523.7023.7023.5523.55
k=1000k=100023.9123.9123.7523.7523.4723.4723.2223.22

These evaluations confirm that while there is a slight reduction in aesthetic quality compared to the baseline, the trade-off is well-justified by the significant improvements in generation speed. This aligns with the intended design of LANTERN, which emphasizes efficiency while preserving acceptable quality.

The results have been incorporated to our main table (Table 2 in the revised manuscript) and full results have been appended to Appendix H. The added P/R analysis complements our FID results, offering a more comprehensive view of the method’s performance.

评论

Impact of Quality Loss and Robustness Across Diverse Model

To ensure that the observed quality loss remains within an acceptable range, we conducted additional experiments across diverse AR model including LlamaGen stage II and Anole [3] with MS-COCO 2017 Validation set. The results validate the robustness of our method across different settings and highlight that the trade-off between speed and quality is consistent and favorable. These are the tables of additional experiment results :

LlamaGen Stage II, τ=1\tau=1SpeedupMean Accepted LengthFIDCLIPScorePrecision / RecallHPS v2
Vanilla AR1.00×1.00\times1.00×1.00\times47.6047.600.29390.29390.41380.4138 / 0.56480.564823.8423.84
EAGLE-20.96×0.96\times1.22×1.22\times----
LANTERN (δ=0.40,k=1000)\delta=0.40, k=1000)1.64×1.64\times2.24×2.24\times46.1046.100.29250.29250.47040.4704 / 0.52220.522223.0623.06
Anole, τ=1\tau=1SpeedupMean Accepted LengthFIDCLIPScorePrecision / RecallHPS v2
Vanilla AR1.00×1.00\times1.00×1.00\times20.2720.270.32150.32150.65520.6552 / 0.63980.639823.5223.52
EAGLE-20.73×0.73\times1.10×1.10\times----
LANTERN (δ=0.50,k=100)\delta=0.50, k=100)1.17×1.17\times1.83×1.83\times23.4023.400.31860.31860.60260.6026 / 0.61780.617822.9222.92

The results demonstrate that LANTERN consistently delivers significant acceleration (achieving 1.71×1.71\times and 1.60×1.60\times speed-ups compared to EAGLE-2 on LlamaGen Stage II and Anole, respectively) while maintaining competitive image quality. However, its performance, both in acceleration and image quality, can vary based on the degree of token selection ambiguity inherent to the model. In particular, the results on Anole are slightly less favorable compared to LlamaGen, which can be explained by the distinct characteristics of these models.

Anole exhibits much lower next-token prediction probabilities (average top-11 probability of 0.0640.064 and average top-1010 probability of 0.2040.204) compared to LlamaGen (0.2060.206 and 0.5200.520, respectively), as shown in Figure 2(c) of the revised manuscript. This highlights that Anole faces a more severe degree of token selection ambiguity, which directly affects the drafter’s ability to provide accurate predictions. Consequently, the drafter trained for Anole achieves a test accuracy of only 27.60%27.60\%, significantly lower than the 38.80%38.80\% test accuracy achieved by the drafter for LlamaGen, as presented in Figure 2(b). This gap in drafter performance impacts the results for both EAGLE-2 and LANTERN on Anole.

It is important to clarify that this variation is not a limitation of LANTERN itself but rather a reflection of model-specific factors, such as token selection ambiguity. Even under these challenging conditions, LANTERN still outperforms EAGLE-2 on Anole in terms of acceleration and maintains acceptable image quality, demonstrating its robustness across different models.

These findings also suggest opportunities for further research. Enhancing drafter performance through better architectural designs or incorporating training methods that account for token selection ambiguity could improve results across models with significant ambiguity issues. While the current results validate LANTERN’s robustness and broad applicability, addressing these aspects could unlock even greater performance improvements in the future.

We have included the these results to appendix E in the revised manuscript, as we aimed to minimize significant changes to the main content of the manuscript. However, if you feel that incorporating these results into the main table would enhance clarity and comprehensiveness, we would be happy to expand the main table in the next revision. We sincerely appreciate all of your valuable feedback and remain committed to improving the manuscript in line with your suggestions.

Once again, we would like to express our heartfelt gratitude for your detailed review and constructive suggestions. Your feedback has not only helped us identify areas for improvement but has also motivated us to refine our work further. We hope that the additional experiments, analyses, and revisions we have outlined above adequately address your concerns. Please do not hesitate to let us know if there are any other aspects we should clarify or improve. We sincerely value your time and effort in reviewing our submission and are deeply appreciative of the insights you have provided.

评论

References

[1] Wu, Xiaoshi, et al. "Human preference score v2: A solid benchmark for evaluating human preferences of text-to-image synthesis." arXiv preprint arXiv:2306.09341 (2023).

[2] Kirstain, Yuval, et al. "Pick-a-pic: An open dataset of user preferences for text-to-image generation." Advances in Neural Information Processing Systems 36 (2023): 36652-36663.

[3] Chern, Ethan, et al. "Anole: An open, autoregressive, native large multimodal models for interleaved image-text generation." arXiv preprint arXiv:2407.06135 (2024).

审稿意见
6

The work presents a novel approach to enhance the efficiency of image generation using Auto-Regressive (AR) models, which traditionally suffer from slow sequential processing. The authors introduce LANTERN, a method that leverages speculative decoding with a relaxed acceptance condition to significantly speed up inference while maintaining image quality. By utilizing a smaller drafter model to predict multiple tokens simultaneously, LANTERN addresses the challenges of token selection ambiguity inherent in visual AR models. The results demonstrate notable improvements in speed and efficiency compared to baseline methods, highlighting the potential of LANTERN to advance the capabilities of AR models in generating high-quality images.

优点

The methodology of this paper is clear and comprehensible, and the research question it addresses is interesting. Although the overall approach is relatively straightforward, I believe the findings are still valuable.

缺点

  1. My primary concern is that the paper lacks an evaluation of the image generation quality. The authors present several generated images, but these images are stylistically very similar and exhibit poor consistency with the accompanying generated text. Some images even contain clear visual errors. I recommend that the authors conduct a more comprehensive assessment of image generation quality to better demonstrate the effectiveness of the proposed method.

  2. The problem the paper aims to address is the low Mean Accepted Length that occurs when applying speculative decoding to AR image generation models. Specifically, although speculative decoding is used, a large number of predictions generated by the lightweight model are rejected, forcing the larger model to regenerate many tokens. As a result, the expected efficiency gains from speculative decoding are not realized. While the authors present a complete exploration from two perspectives, I still feel that, given the inherent ambiguity in token selection, improving the acceptance rate alone might suffice. A large set of tokens could be considered acceptable. This raises the question: Is speculative decoding still necessary in such a case? What would the quality and efficiency be if tokens were instead randomly generated within certain constraints?

问题

Please see the weaknesses.

评论

Weakness 2 : Is speculative decoding still necessary in such a case?

We sincerely thank you for raising this important question. To explore the necessity of speculative decoding, we conducted an experiment using a randomly initialized drafter model to evaluate its ability to accelerate decoding. For this experiment, we utilized LlamaGen Stage I model as the target model and randomly sampled 1000 captions from the MS-COCO 2014 validation set to measure each model's performance. The results are summarized in the tables below:

τ=0\tau=0Mean Accepted Length
Random init drafter1.001.00
trained drafter (4 epochs)1.541.54
trained drafter (20 epochs)1.601.60
τ=1\tau=1Mean Accepted Length
Random init drafter1.001.00
trained drafter (4 epochs)1.181.18
trained drafter (20 epochs)1.201.20

The results clearly demonstrate that a randomly initialized drafter model fails to improve the mean accepted length, achieving a value of 1.00 for both τ=0\tau=0 and τ=1\tau=1, equivalent to standard decoding without speculative decoding. In contrast, the trained drafter significantly enhances performance, achieving a mean accepted length of 1.60 and 1.20 for τ=0\tau=0 and τ=1\tau=1, respectively, after 20 epochs of training. These findings highlight the necessity of speculative decoding and the importance of having a well-trained drafter to achieve meaningful acceleration. Without proper training, the drafter cannot generate effective predictions, rendering speculative decoding ineffective. We are grateful for your insightful suggestion, as it allowed us to further validate the critical role of the drafter in speculative decoding and to provide a clearer understanding of its importance in our approach. This experiment reinforces the robustness and continued relevance of speculative decoding in achieving both speed and quality improvements.

Once again, we thank you for your thoughtful comments and suggestions, which have significantly contributed to refining our work. We hope the additional experiments and explanations provided above adequately address your concerns. Please do not hesitate to reach out if further clarification or additional results are needed.

评论

We deeply appreciate your positive evaluation of our work and your thoughtful feedback. Your recognition of the strengths in our methodology and the potential contributions of our approach is highly encouraging. Additionally, your detailed comments and constructive suggestions have been invaluable in guiding us to improve the quality and clarity of our work further. Below, we address each of the points you raised and provide additional experiments and explanations.

Weakness 1 : Evaluation of the image quality

We have conducted additional experiments using a broader set of evaluation metrics. The updated results are presented below (which is the same as in the general response), expanding on the main table in our original submission with same setting:

τ=1\tau=1SpeedupMean Accepted LengthFIDCLIPScorePrecision / RecallHPS v2
Vanilla AR×1.00\times1.00×1.00\times1.0015.2215.220.32030.32030.4781/0.56330.4781 / 0.563324.1124.11
EAGLE-2×0.93\times0.93×1.20\times1.20----
LANTERN (δ=0.10,k=1000)\delta=0.10, k=1000)×1.13\times1.13×1.75\times1.7516.1716.170.32080.32080.4869/0.51720.4869 / 0.517223.7523.75
LANTERN (δ=0.40,k=1000)\delta=0.40, k=1000)×1.69\times1.69×2.40\times2.4018.7618.760.32060.32060.4909/0.44970.4909 / 0.449723.2223.22

As shown in the table above, LANTERN achieves comparable or slightly higher precision while exhibiting a modest reduction in recall. This suggests that the quality of individual images is preserved, albeit with a slight decrease in diversity.

This modest decline in recall is likely due to the token selection ambiguity faced by the drafter. When the drafter is not optimally trained, it may struggle to generate predictions that are both diverse and accurate. As a result, increasing the acceptance probability to enhance acceleration could inadvertently reduce the diversity of the generated images. Nevertheless, this trade-off between speed and diversity has minimal impact on overall image quality, as demonstrated by stable FID, Precision and HPS v2 scores. Further improvements in drafter training could mitigate this effect, enhancing recall while maintaining the efficiency gains achieved by LANTERN. Additionally, the HPS v2 score, which is derived from a preference model trained on a human preference dataset, does not exhibit any significant degradation, further supporting the robustness of LANTERN's performance.

We also understand your concern regarding the stylistic similarity and lack of text-image consistency in the generated samples. While we acknowledge this issue, we believe it primarily stems from the base visual AR model (LlamaGen Stage II) rather than being specific to LANTERN. To address this, we revised the text prompts used for generation of qualitative samples and have conducted additional experiments focusing on cases where the base model generates images that meet the following criteria: (1) stylistically dissimilar, (2) consistent with the text prompt, and (3) free of visual errors. For these cases, we provide samples generated by LANTERN to demonstrate its performance under such conditions. In addition, we provide samples generated by random replacement decoding as well. The qualitative samples are available at here : LANTERN samples, random replacement decoding, and qualitative samples. Each image have been incorporated into the Figure 1, 3, and 4 in the revised manuscript.

We hope these additional results provide a more comprehensive perspective on the image quality generated by LANTERN.

评论

Thank you for your kind words and for maintaining a positive recommendation for our research. We believe that your feedback has helped make our work even more robust.

审稿意见
6

In this paper, the authors propose LANTERN, a sampling strategy to speed up the image generation without losing too much quality. The method is implemented by accumulating probabilities from nearby tokens of the current sampling token. The authors further propose a thresholding technique to prevent the accumulated distribution deviating too much from the original one. Experiment results seem to be effective, however, the analysis is intuitive and empirical without deep insights.

优点

  • The proposed method is intuitive and easy to follow.

  • Experimental results seem to verify the proposed method in terms of speeding up image generation without significant quality loss.

缺点

  • There is a gap between the introduction and methodologies in writing. The related works are deferred to the appendix and problem definition and preliminaries such as Speculative Decoding are missing, resulting in difficulties in understanding the problem and challenge for readers who are not exactly working on this domain. On the other hand, Section 2.1 and 2.2 have some overlaps about the experiments and observation, which can be more concise.

  • The evaluation seems to be insufficient. For the testing data, it is ideal to use the same setting as the vanilla baseline (i.e., LlamaGen) rather than just 100 captions from MSCOCO. The statement “Since measuring speedup with more than 100 samples shows no significant difference, we use 100 captions for efficiency” is not very convincing to me. In addition, other models are suggested for evaluation as well including other variations of LlamaGen.

  • Some analyses in experiments are not sufficient. In sampling, the authors only mentioned that when δ\delta is small, using larger k results in speeding up without significant degradation in performance, yet the reason for it is less explored. Section 4.3.2 also lacks further analysis why using TVD is better than JSD except some empirical results.

问题

  • See the weakness.
  • The claim of interchangeability in Section 3.1 is only empirically explored via a few examples. Are there some theoretical insights or statistics supporting this claim?
评论

We sincerely thank you for your thoughtful and constructive feedback. Your comments have highlighted critical areas for improvement and provided valuable insights that have strengthened our work. We are especially grateful for your positive remarks regarding the clarity and intuitiveness of our proposed method and its effectiveness in accelerating image generation. Below, we address each of the points raised with detailed responses.

Weakness 1: Gap between the introduction and methodologies in writing

We agree that the gap between the introduction and methodologies could have been improved. To address this, we revised the paper to include a concise preliminary section (Section 2 in the revised manuscript) in the main body, bridging the introduction and methodologies sections. Due to ICLR's page limit policy, we kindly ask for your understanding as we were unable to include the entire related work section. Additionally, Sections 2.1 and 2.2 have been merged into a single section (Section 3 in the revised manuscript) and refined to eliminate overlaps and improve conciseness. We believe these updates enhance the coherence of the paper, making it more accessible to readers and better aligning the introduction with the subsequent sections.

Weakness 2: Insufficiency in evaluating efficiency

First and foremost, we sincerely apologize for the misstatement in the original manuscript regarding the measurement of mean accepted length. While we had stated that the mean accepted length was measured on 100 images, this was incorrect. In fact, mean accepted lengths were evaluated on the full MS-COCO dataset, and this has been corrected in the revised manuscript.

Additionally, we acknowledge that using only 100 captions to measure actual speedup (wall-clock time) may be insufficient for providing convincing evidence. To address this, we have re-measured the actual speedup using a larger sample size of 1000 captions in the revised version. We deeply regret any confusion caused by this oversight and thank the reviewers for bringing it to our attention, allowing us to improve the clarity and accuracy of the manuscript.

τ=0\tau=0SpeedupMean Accepted Length
Vanilla AR×1.00\times1.00×1.00\times1.00
EAGLE-2×1.29\times1.29×1.60\times1.60
LANTERN (δ=0.05,k=1000)\delta=0.05, k=1000)×1.56\times1.56×2.02\times2.02
LANTERN (δ=0.20,k=1000)\delta=0.20, k=1000)×2.26\times2.26×2.89\times2.89
τ=1\tau=1SpeedupMean Accepted Length
Vanilla AR×1.00\times1.00×1.00\times1.00
EAGLE-2×0.93\times0.93×1.20\times1.20
LANTERN (δ=0.10,k=1000)\delta=0.10, k=1000)×1.13\times1.13×1.75\times1.75
LANTERN (δ=0.40,k=1000)\delta=0.40, k=1000)×1.69\times1.69×2.40\times2.40

In addition, we performed an analysis to confirm that speedup results stabilize as the number of captions increases. The table below illustrates this stability, showing that measurements with 1000 captions are consistent with those using larger sets:

Num CaptionsActual Speedup (τ=0\tau=0, LANTERN, k=1000,0.05k=1000,0.05)Actual Speedup (τ=0\tau=0, LANTERN, k=1000,0.2k=1000,0.2)Actual Speedup (τ=1\tau=1, LANTERN, k=1000,0.1k=1000,0.1)Actual Speedup (τ=1\tau=1, LANTERN, k=1000,0.4k=1000,0.4)
1001.56×1.56\times2.33×2.33\times1.13×1.13\times1.73×1.73\times
10001.56×1.56\times2.26×2.26\times1.13×1.13\times1.69×1.69\times
20001.57×1.57\times2.27×2.27\times1.13×1.13\times1.69×1.69\times
50001.56×1.56\times2.26×2.26\times1.13×1.13\times1.69×1.69\times

Please note that since the captions were randomly sampled, the results for the 100 captions may differ slightly from the acceleration reported in Table 3 of the original paper (Table 2 in the revised version). We updated our main result in Table 2 (in the revised version) with this re-measured speedup, and we hope these additional evaluations provide a clearer understanding of the robustness of our efficiency claims. The analysis of a number of captions has been added to Appendix F.1 in the revised manuscript.

Weakness 3: Further Analyses

  • Further analysis on δ\delta and kk

    As part of an extended evaluation prompted by another reviewer's question, we conducted additional experiments with image quality metrics (e.g., precision/recall and HPS v2) under the same settings as our main results. Through this process, we observed that increasing kk values does not consistently improve performance for small δ\delta, contrary to our initial claim. The table below illustrates this behavior:

    ConfigurationPrecision / RecallHPS v2
    k=100,δ=0.05k=100, \delta=0.050.48670.4867 / 0.53890.538924.0124.01
    k=300,δ=0.05k=300, \delta=0.050.48560.4856 / 0.53670.536723.9723.97
    k=1000,δ=0.05k=1000, \delta=0.050.48650.4865 / 0.53340.533423.9123.91
评论

As shown in the table above, a decreasing trend in performance evaluation metrics can be observed as kk increases for the two new metrics. We sincerely apologize for the earlier claim that larger kk always improves speedup without significant quality degradation, as this does not hold consistently under these metrics. To address this, we have revised the manuscript to exclude this claim and provide a more accurate representation of the results. The updated findings have been appended to Appendix H in the revised manuscript. We thank the reviewers for highlighting this aspect, allowing us to refine our analysis and improve the clarity of the work.

  • Further analysis on TVD and JSD

    To further address your suggestion, we compared TVD and JSD with respect to their impact on image quality and latency. The experiment setup is identical to our initial experiments on distance metrics. Actual computation time is evaluated in average on randomly selected 1000 samples. The results are as follows:

    Distance MetricMean Accepted LengthFIDCLIP Score
    TVD (δ=0.3\delta=0.3)2.29×2.29\times18.2718.270.32060.3206
    JSD (δ=0.2\delta=0.2)2.29×2.29\times18.2118.210.32060.3206
    TVD (δ=0.2\delta=0.2)2.09×2.09\times17.4317.430.32080.3208
    JSD (δ=0.13)\delta=0.13)2.09×2.09\times17.4817.480.32060.3206
    Distance MetricComputation Time for Distance MetricTotal Computation Time of Single Decoding Step
    TVD1.19×1031.19\times 10^{-3} s4.89×1024.89\times 10^{-2}s
    JSD4.03×1034.03\times 10^{-3} s4.92×1024.92 \times 10^{-2}s

    In the table, it can be observed that when TVD and JSD yield similar mean accepted lengths, there’s no significant differences in terms of image quality. While either metric can be used without notable impact on image quality, TVD proves to be a more practical choice when considering the computation time. Since JSD requires more computation than TVD, selecting TVD is more beneficial for achieving speedup in practical applications. To validate this, we measured the computation time for TVD and JSD within a single decoding step of LANTERN. As shown in the table above, JSD requires more than three times the computation time of TVD. Although the time difference between these distance metrics is relatively small compared to the total time of whole decoding step, this difference accumulates over multiple decoding steps and can result in a significant impact on overall efficiency. The results discussed above have been incorporated into Table 3 and Table 9 in Appendix F.3 in the revised manuscript.

We hope that these additional analyses adequately address your concerns. Thank you for your insightful comments, which have guided us in refining our work.

Question 1 : Theoretical insights or statistical supports to the claim of interchangeability

Thank you for raising this critical question. We acknowledge that empirical qualitative evidence based on a few examples is insufficient to fully support the claim of interchangeability. To address this, we have conducted additional experiments to provide statistical evidence. Specifically, we evaluated the quality of generated images using random replacement decoding and compared them to standard decoding. We used LlamaGen Stage I model and evaluate FID and CLIP Score on MS-COCO 2017 validation set.

Randomly Replaced by one of kk-th nearest tokenFIDCLIP Score
Vanilla AR25.0625.060.32140.3214
k=50k=5026.8826.880.31200.3120
k=100k=10030.7630.760.30910.3091
k=1000k=100088.0388.030.27150.2715

The results show that as kk increases, the replaced token is selected from a broader set of latent space neighbors, but the image quality remains well-preserved up to a certain threshold. For k=50k=50, the FID increases slightly from 25.0625.06 (Vanilla AR) to 26.8826.88, and the CLIP Score decreases marginally from 0.32140.3214 to 0.31200.3120, indicating minimal degradation. Similarly, for k=100k=100, the FID rises moderately to 30.7630.76, and the CLIP Score drops slightly to 0.30910.3091, demonstrating that even with k=100k=100, the image quality remains stable and acceptable.

It is only at k=1000k=1000 that a significant decline becomes apparent, with the FID increasing sharply to 88.0388.03 and the CLIP Score dropping to 0.27150.2715, highlighting the negative impact of selecting tokens from more distant neighbors. These results confirm our earlier qualitative observations in Figure 3 in the revised manuscript that increasing kk up to 100 maintains reasonable image quality, making it a viable strategy for generative tasks. This highlights the robustness of the model in preserving image fidelity under controlled token replacement within this range.

评论

The result has been included in the revised manuscript, at Appendix C.1. We believe these findings strengthen the validity of our claim and provide a more robust foundation for this observation. Additionally, we aim to position the development of theoretical insights into interchangeability as a promising direction for future research. Your feedback has been invaluable in highlighting the need for a deeper exploration of this aspect, and we are sincerely grateful for your thoughtful comments.

Once again, we would like to express our heartfelt gratitude for your detailed review and constructive suggestions. Your feedback has not only helped us identify areas for improvement but has also motivated us to refine our work further. We hope that the additional experiments, analyses, and revisions we have outlined above adequately address your concerns. Please do not hesitate to let us know if there are any other aspects we should clarify or improve. We sincerely value your time and effort in reviewing our submission and are deeply appreciative of the insights you have provided.

评论

We sincerely thank the reviewers for their thoughtful feedback and valuable suggestions. Your insights have been instrumental in improving the clarity and quality of our work, and we greatly appreciate the opportunity to address your concerns. In this response, we have included a General Response for questions raised by two or more reviewers, ensuring consistency and transparency. For individual responses, we have provided tailored responses, with relevant excerpts from the general response included where appropriate.


Broader range of image quality evaluation

To further validate LANTERN's ability to maintain image quality, we conducted additional evaluations using two text-to-image quality metrics: (1) Precision and Recall [1], and (2) Human Preference Score (HPS) v2 [2], with same setting as our main experiment. The results below extend our original main table (Table 3 in original paper, Table 2 in the revised version) to include these metrics. We are currently working on the evaluation for greedy decoding (τ=0\tau=0), and the results will be updated as soon as available.

τ=1\tau=1SpeedupMean Accepted LengthFIDCLIPScorePrecision / RecallHPS v2
Vanilla AR×1.00\times1.00×1.00\times1.0015.2215.220.32030.32030.47810.4781 / 0.56330.563324.1124.11
EAGLE-2×0.93\times0.93×1.20\times1.20----
LANTERN (δ=0.10,k=1000)\delta=0.10, k=1000)×1.13\times1.13×1.75\times1.7516.1716.170.32080.32080.48690.4869 / 0.51720.517223.7523.75
LANTERN (δ=0.40,k=1000)\delta=0.40, k=1000)×1.69\times1.69×2.40\times2.4018.7618.760.32060.32060.49090.4909 / 0.44970.449723.2223.22

As shown in the table above, LANTERN demonstrates comparable or slightly improved precision while showing a slight decrease in recall. This can be interpreted as maintaining the quality of individual images while slightly reducing diversity. This slight reduction in recall can be attributed to the challenges associated with token selection ambiguity in the drafter. When the drafter is suboptimally trained, it may struggle to provide sufficiently diverse predictions or precise alternatives. Consequently, the increased acceptance probability, while improving acceleration, can lead to a modest decline in the diversity of generated images. However, it is important to note that this trade-off between acceleration and diversity does not significantly impact the overall image quality, as evidenced by the consistent FID, Precision and HPS v2 scores. Future work could address this by improving the drafter’s training process to better handle token selection ambiguity, thereby further enhancing recall without compromising LANTERN’s efficiency. Additionally, the HPS v2 score, which is derived from a preference model trained on a human preference dataset, does not exhibit any significant degradation, further supporting the robustness of LANTERN's performance. As a result, LANTERN prove its effectiveness over various metric, by maintaining its score in acceptable range. The results have been updated to the Table 2 in the revised version.


Evaluation on other visual AR models

To evaluate LANTERN on additional visual AR models, we extended our experiments to include the LlamaGen Stage II model and Anole [3]. These experiments focused on random sampling (τ=1\tau=1) for the following reasons: (1) Random sampling generally produces higher-quality images than greedy decoding (τ=0\tau=0), (2) Our prior results (Table 3 in the paper) indicate random sampling is a more challenging case for acceleration, and (3) Resource constraints prevented comprehensive experiments for both settings. We use MS-COCO 2017 Validation set to evaluate LlamaGen stage II and Anole’s image generation performance. In addition, we measure actual speedup of LlamaGen model with RTX3090 and A100 for Anole. The results are summarized below.

LlamaGen Stage II, τ=1\tau=1SpeedupMean Accepted LengthFIDCLIPScorePrecision / RecallHPS v2
Vanilla AR1.00×1.00\times1.00×1.00\times47.6047.600.29390.29390.41380.4138 / 0.56480.564823.8423.84
EAGLE-20.96×0.96\times1.22×1.22\times----
LANTERN (δ=0.40,k=1000)\delta=0.40, k=1000)1.64×1.64\times2.24×2.24\times46.1046.100.29250.29250.47040.4704 / 0.52220.522223.0623.06
Anole, τ=1\tau=1SpeedupMean Accepted LengthFIDCLIPScorePrecision / RecallHPS v2
Vanilla AR1.00×1.00\times1.00×1.00\times20.2720.270.32150.32150.65520.6552 / 0.63980.639823.5223.52
EAGLE-20.73×0.73\times1.10×1.10\times----
LANTERN (δ=0.50,k=100)\delta=0.50, k=100)1.17×1.17\times1.83×1.83\times23.4023.400.31860.31860.60260.6026 / 0.61780.617822.9222.92
评论

Anole exhibits significantly lower next-token prediction probabilities (average top-11 probability of 0.0640.064 and average top-1010 probability of 0.2040.204) compared to LlamaGen (0.2060.206 and 0.5200.520, respectively), as shown in Figure 2 (c) of the revised manuscript. These results indicate that Anole suffers from more severe token selection ambiguity, which inherently affects the drafter’s training process. Consequently, Anole’s drafter achieves a test accuracy of only 27.60%27.60\%, markedly lower than the 38.80%38.80\% test accuracy observed for LlamaGen’s drafter, as shown in Figure 2 (b) in the revised manuscript. This discrepancy in drafter performance directly impacts the effectiveness of both EAGLE-2 and LANTERN on Anole.

It is important to emphasize that this variation does not reflect a limitation of LANTERN itself but rather underscores the influence of model-specific characteristics, such as token selection ambiguity, on overall performance. Despite these challenges, LANTERN continues to outperform EAGLE-2 on Anole in terms of acceleration while maintaining acceptable image quality, highlighting its robustness across diverse models.

Furthermore, these findings suggest a promising direction for future research. Addressing token selection ambiguity through improved drafter architectures or incorporating ambiguity-aware training techniques could enhance the drafter’s performance, further bolstering LANTERN’s effectiveness across models with severe token selection ambiguity. This indicates that while the current results demonstrate LANTERN's robustness and general applicability, there is potential for even greater improvements with advancements in drafter training and model design.

We have included the these results to appendix E in the revised manuscript, as we aimed to minimize significant changes to the main content of the manuscript. However, if the reviewers collectively feel that incorporating these results into the main table would enhance clarity and comprehensiveness, we would be happy to expand the main table in the next revision. We sincerely appreciate all of your valuable feedback and remain committed to improving the manuscript in line with your suggestions.


Major changes in the revised manuscript

Again, we sincerely thank the reviewers for their valuable suggestions and thoughtful feedback, which have greatly contributed to improving the quality of our work. Based on these insights, we have carefully revised the manuscript to address the concerns raised and incorporate additional analyses.

We have made several key revisions to enhance the clarity, organization, and comprehensiveness of the manuscript:

  1. Preliminaries Section (Section 2): Added a new section to bridge the Introduction and Methodologies, providing essential background on visual AR models and speculative decoding to improve reader understanding.
  2. Consolidation of Sections 2.1 and 2.2: These sections were merged and rewritten to eliminate overlapping content and improve the flow of information. While the text was significantly revised to achieve a clearer and more concise presentation, the underlying content and key messages remain unchanged.
  3. Expanded Experimental Section: Incorporated additional results using new evaluation metrics into the main table (Table 2) and corresponding explanations in the Experimental Setup (Section 5.1) and Results sections.
    • Updates in ablation study : To provide a clear comparison between TVD and JSD, we compared the differences in image quality between TVD and JSD when both achieved the same level of acceleration. This analysis has been reflected in Table 3 and Section 5.3.2 of the revised manuscript.
  4. Updated Qualitative Samples: Replaced the samples in Figures 1, 3, and 4 with more representative examples to better align with text prompts, address visual errors, and showcase stylistic diversity. The updated figures also can be found in these anonymous links: LANTERN samples, random replacement decoding, and qualitative samples.

We hope these additional experiments further support the contributions of our work and provide clarity on its robustness. Once again, thank you for your time and effort in reviewing our submission. We are grateful for your constructive feedback and look forward to your continued insights.

References

[1] Kynkäänniemi, Tuomas, et al. "Improved precision and recall metric for assessing generative models." Advances in neural information processing systems 32 (2019).

[2] Wu, Xiaoshi, et al. "Human preference score v2: A solid benchmark for evaluating human preferences of text-to-image synthesis." arXiv preprint arXiv:2306.09341 (2023).

[3] Chern, Ethan, et al. "Anole: An open, autoregressive, native large multimodal models for interleaved image-text generation." arXiv preprint arXiv:2407.06135 (2024).

评论

As previously mentioned, we further evaluate the performance of LANTERN under greedy decoding (τ=0\tau=0) and summarize the results in the following table. Similar to the case of sampling (τ=1\tau=1), we report both acceleration and image quality across various metrics. Similar to the sampling, the findings demonstrate that LANTERN achieves substantial speedup while maintaining competitive image quality compared to standard AR decoding in the greedy decoding as well. These results reaffirm LANTERN's capability to balance efficiency and quality across different decoding strategies. This additional analysis has also been incorporated into the main table (Table 2) of the revised manuscript.

τ=0\tau=0SpeedupMean Accepted LengthFIDCLIPScorePrecision / RecallHPS v2
Vanilla AR×1.00\times1.00×1.00\times1.0028.6328.630.31690.31690.42320.4232 / 0.35170.351723.1823.18
EAGLE-2×1.29\times1.29×1.60\times1.60----
LANTERN (δ=0.05,k=1000)\delta=0.05, k=1000)×1.56\times1.56×2.02\times2.0229.7729.770.31640.31640.44840.4484 / 0.31580.315822.6222.62
LANTERN (δ=0.20,k=1000)\delta=0.20, k=1000)×2.26\times2.26×2.89\times2.8930.7830.780.31540.31540.47710.4771 / 0.27730.277321.6921.69
评论

Dear Reviewers, AC, and SAC of paper #12935,

As we approach the conclusion of the author-reviewer discussion phase, we would like to highlight few points:

  1. We are sincerely grateful to the reviewers for their constructive feedback, which has significantly contributed to improving the clarity and depth of the manuscript. The thoughtful comments and suggestions have been invaluable in refining our work.

  2. During the discussion period, we made every effort to address all concerns raised by the reviewers, including conducting additional experiments, collecting data, and providing thorough responses. While we are glad that this process helped garner more favorable assessments from some reviewers, we regret that two reviewers were unable to engage further despite our polite reminders. Out of respect for their time, we chose not to send repeated reminders, trusting that our detailed rebuttal and the positive feedback from other reviewers sufficiently highlight the merits of our work. Nevertheless, if circumstances allow, we would still greatly appreciate their feedback, which could further enrich the evaluation process.

  3. Understanding the limitations and potential for speculative decoding (SD) for vision AR models has been an open problem. Our paper "LANTERN" not only takes the first attempt to understand the limitations of SD for vision AR but also identifies the associated potential causes and presents a novel solution to mitigate that. We demonstrate LANTERN can yield inference speed up by up to 1.82×1.82\times compared to any existing alternative. We believe this work would inspire the community to adopt SD for the emerging vision AR use cases for latency-critical applications.

We remain hopeful and trustful in the reviewing system and grateful for the reviews and discussions (though only two out of four reviewers provided them). We hope our efforts are acknowledged and taken into account during the decision.

Best regards,

The Authors

Paper ID #12935

AC 元评审

The paper introduces LANTERN, a method that accelerates visual autoregressive models by adapting speculative decoding -- a mechanism originally proposed for large language models -- to the domain of visual autoregressive generation. The proposed approach demonstrates the ability to achieve speed gains while maintaining image quality. This work opens up new possibilities in enhancing the efficiency of autoregressive visual generative models and highlights how mechanisms from large language models can be effectively tailored for visual-specific applications. Considering the positive feedback from all reviewers, I recommend the acceptance of this paper.

审稿人讨论附加意见

The authors effectively addressed the concerns raised by the reviewers by incorporating additional metrics (e.g., Precision/Recall and HPS v2) to comprehensively evaluate image quality, conducting analyses to validate theoretical claims such as token interchangeability and the advantages of TVD over JSD, and making clarifications and improvements to the paper’s structure and coherence.

最终决定

Accept (Poster)