PaperHub
6.4
/10
Poster4 位审稿人
最低3最高5标准差0.7
5
3
4
4
3.8
置信度
创新性2.5
质量3.0
清晰度2.5
重要性2.5
NeurIPS 2025

Switchable Token-Specific Codebook Quantization For Face Image Compression

OpenReviewPDF
提交: 2025-05-09更新: 2025-10-29

摘要

关键词
image compressionimage tokenization

评审与讨论

审稿意见
5

This paper proposes a novel method for compressing facial images using a Switchable Token-Specific Codebook Quantization (STSCQ) framework. Traditional VQ-based image compression methods degrade under ultra-low bitrate (bpp) due to reliance on a single global codebook. To address this, the authors introduce a two-level codebook structure: (i) image-level routing to select a specialized codebook and (ii) token-level codebook decomposition to assign each token its own quantizer. The architecture is trained in a three-stage progressive manner, incorporating ArcFace loss for identity preservation. Extensive experiments across five face recognition benchmarks show improved rate-distortion trade-offs and recognition accuracy over prior methods like TiTok and MaskGIT. The method demonstrates superior compression efficiency and recognition fidelity, particularly at low bpp (e.g., 66.02% accuracy at 0.004 bpp).

优缺点分析

The paper offers a compelling, though not flawless, contribution. Its chief strength lies in the originality of a two-level quantization design that marries switchable image-level routing with token-specific sub-codebooks, an architectural twist that demonstrably boosts face-recognition accuracy at ultra-low bitrates and thus promises real utility for storage-constrained biometric pipelines.

Methodologically, the authors conduct sensible ablations, test the idea on both 1-D and 2-D latent tokenizers, and document a clear three-stage training schedule. Well-designed figures also aid reader comprehension.

However, important weaknesses remain: all quantitative results lack error bars or confidence intervals, so the statistical robustness of the reported gains is uncertain. Furthermore, the novelty of image-level routing is somewhat incremental given prior work on mixture-of-experts and product-quantization, and the paper omits a discussion of computational overhead, which is crucial for edge deployment.

If the authors add statistical significance reporting and quantify the computational cost, the work would merit acceptance; without these corrections, the contribution falls just short of the bar for a premier venue.

问题

  1. Can you rerun the full evaluation over ≥ 3 random seeds and report mean ± standard deviation (or 95 % confidence intervals) for all metrics in Tables 1 and 2?
  2. What are the training-time and inference-time costs, FLOPs, and latency, relative to TiTok-s and MaskGIT?

局限性

No. The paper lists technical limitations (dependency on the underlying auto-encoder) and briefly claims positive societal value, but it omits two key areas:

  1. Supplementary Fig. 2 shows that routing groups correlate with ethnicity; however, the authors provide no demographic parity analysis or mitigation strategy. A brief experiment on RFW or a similar dataset, along with a discussion of safeguards (e.g., bias-aware training or post-compression calibration), is necessary.

  2. Ultra-low-bpp face compression can lower the cost of large-scale surveillance. The paper should acknowledge this and outline responsible-release measures (e.g., gating access to pretrained weights, adding watermarking, or providing usage guidelines).

最终评判理由

The authors’ rebuttal offers a generally adequate and structured response to the raised concerns, addressing each key issue with new experimental evidence and clarifications. For statistical robustness, they reran experiments over three random seeds and reported mean ± standard deviation for the TiTok-S baseline. While this partially satisfies the request, the omission of multi-seed evaluations for other backbones limits completeness, though they reasonably justify this with resource constraints.

The concern regarding computational overhead is fully addressed: detailed tables report inference time, GPU memory, storage, and FLOPs across configurations, comparing their method with TiTok and MaskGIT and demonstrating significant improvements, particularly under the Codebook Routing strategy.

On demographic fairness, the authors provide per-ethnicity accuracy evaluations using the RFW dataset and argue that the high number of codebooks dilutes any strong alignment with specific groups. While no direct mitigation is implemented, the authors propose future work involving group-specific decoder fine-tuning, which reasonably addresses the concern.

Ethical risks around ultra-low-bpp face compression are now acknowledged, with promised countermeasures such as watermarking and usage guidelines to be included in the final version. Although the authors do not explicitly expand on the originality concern, they reiterate the architectural innovations of their token-specific quantization and routing strategy.

Overall, the rebuttal resolves the primary issues of evaluation rigor, computational efficiency, and ethical responsibility. The rebuttal justifies an upgrade to an accept.

格式问题

No

作者回复

Answer 1:

We re-evaluated our method using three different random seeds, and the results are presented in the table below. Due to time and resource constraints, we are only able to report results based on the TiTok-S baseline. As shown in the table, the impact of random seed variation is not significant, which can be attributed to the fact that the core of our method lies in training the codebook, while the encoder and decoder are initialized with pre-trained weights. We plan to update the re-evaluation results for other baselines in future versions.

Table: Switchable Token-Specific Codebook Quantization on different baselines.

Backbone# Tokens# CodebooksCodebook SizeMeanAcc(%)IDSStorage(bits)bpp
TiTok-s1281409691.65±0.020.6130±0.002015360.0234
TiTok-s12816102489.75±0.140.5681±0.005912840.0196
TiTok-s12825625687.63±0.030.5179±0.000210320.0157

Answer 2:

We have reported the FLOPs of the model, the required storage overhead and the inference time per image on a V100 GPU in the table below. As shown, our method does introduce additional storage and inference latency, and this effect becomes more pronounced as the number of codebook groups and the size of the codebook latent space increase. We attribute this to the need to maintain all codebook groups during inference when selecting the nearest codebook based on the minimum error criterion. To address this, we experimented with directly using the routing results during inference, so that the decoder only needs to access the codebook corresponding to the routing index. The benefits of this approach are clearly reflected in the table: for example, with the MaskGIT backbone, storage overhead is reduced by approximately 60% and inference latency by about 40%, while remaining the performance of MeanAcc and IDS.

Table: Inference latency and memory overhead statistics. For ours method, 'NN' and 'CR' refer to Nearest-Neighbor search and Codebook Routing search respectively.

BackbonebppToken-specificCodebooksInfer Time (s / img)GPU Mem (MB)Storage (MB)FLOPs
Maskgit-VQGAN0.0469-10.1495474.1494.69158.14 G
Ours (NN)0.046910.15543283.391122.70158.14 G
Ours (NN)0.0391160.177112508.174194.70158.14 G
Ours (CR)0.0391160.15445116.15354.70158.14 G
TiTok-S0.0234-10.1433586.0398.68113.57 G
Ours (NN)0.023410.1437609.84122.68113.57 G
Ours (NN)0.01572560.14961894.54482.71113.57 G
Ours (CR)0.01572560.14501157.62100.21113.57 G
TiTok-L0.0059-10.17282709.241157.61269.62 G
Ours (NN)0.005910.17442716.801163.47269.62 G
Ours (NN)0.00402560.17502865.431253.50269.62 G
Ours (CR)0.00402560.17412804.021157.87269.62 G

Table: Quantitative comparison between different codebook selection policies during inference.

BackboneCodebook SelectionbppMeanAcc(%)IDS
Maskgit-VQGANNN0.039192.230.6264
Maskgit-VQGANCR0.039192.220.6253
TiTok-SNN0.015787.600.5180
TiTok-SCR0.015787.540.5125
TiTok-LNN0.004066.020.1885
TiTok-LCR0.004065.650.1864

Answer for limitations 1:

We conducted experiments on the four ethnic groups in the RFW dataset, and the results are shown in the table below. Our findings indicate that the introduction of multiple codebooks does not lead to significant changes in ethnic bias. Although our supplementary materials reveal that the codebook groupings are related to ethnicity, the number of codebook groups is much larger than the number of ethnic categories, so each group is not strongly associated with any single ethnicity. We suggest that future work could explore fine-tuning the decoder for each ethnic group in our stage 3 to better balance recognition performance across different groups. We will include a discussion of these limitations in future versions.

Table: Quantitative comparison between different codebook selection policies during inference.

Backbonebpp# CodebooksAfricanAsianCaucasianIndianMeanAcc(%)IDS
MaskGIT0.0469183.3083.1290.2285.1385.44±2.870.6932
MaskGIT0.03911680.3579.2787.7284.7083.01±3.390.6592
TiTok-S0.0234180.1380.0087.8887.6582.42±3.220.6428
TiTok-S0.015725671.9776.5279.2575.7875.88±2.600.5438

Answer for limitations 2:

We acknowledge that ultra-low-bpp face compression carries the risk of facilitating large-scale surveillance and potential misuse. To mitigate such concerns, we will provide clear usage guidelines and ethical instructions upon the open-sourcing of our project, and incorporate watermarking into the compressed images to help prevent improper use. We will include a statement addressing these issues in the limitations section of a future version.

评论

Overall, the rebuttal resolves the primary issues of evaluation rigor, computational efficiency, and ethical responsibility. The rebuttal justifies an upgrade to an accept.

审稿意见
3

This paper proposes a switchable token-specific codebook quantization method which dynamically specifies a codebook for each image using router and uses specified sub-codebooks for each token in the codebook. The hierarchically dynamic structure significantly reduces the required bpp while achieving high accuracy on face recognition datasets.

优缺点分析

Paper Strengths: 1.The paper addresses an important challenge in the low bpp facial image compression by proposing a dynamic method for specifying codebooks. 2.The paper is generally easy to follow, and the mathematical formulations are logically consistent. 3.Quantitative experiments demonstrate the advantages of the proposed method over existing baselines.

Major Weaknesses: Weakness on paper writing:

1.In Equation (12), the subscripts of the first two codebooks are both 1, which should be modified to 1 and 2. 2.In line 284-285, the original text in the paper is: “our approach achieves a 16.6% bitrate reduction (0.0469 → 0.0391 bpp) coupled with a 1.53% absolute accuracy improvement (90.70% → 92.23%).” The accuracy “90.70%” is inconsistent with the data “93.51%” provided in Table 3. Please check the validity of the data.

Weakness on method: 1.In order to achieve dynamic routing of the codebooks, the paper introduces a router to calculate the probability distribution of each codebook. However, the paper points out that the routing strategy is inconsistent during training and inference due to the exploration-exploitation dilemma inherent in entropy-regularized optimization, which is equivalent to the router only working during training. The paper lacks theoretical analysis of the dilemma. It would be beneficial to discuss the reason of the dilemma. 2.In Equation (8) and Equation (9), two losses designed for training routers are introduced, which are calculated based on the probability distribution output by the router. However, due to the lack of description of in the text, based on the known information, these two losses are actually equivalent to the information entropy of selection distributions with different weights.

Weakness on experiments: 1.For evaluating compression performance, the paper only relies on Mean Accuracy and IDS, ignoring commonly used evaluation metrics in facial image compression, such as PSNR and LPIPS. It is recommended to complete experimental results to obtain a more comprehensive evaluation. 2.The paper mentions that five facial datasets were used for evaluation in the baseline model comparison experiment. However, the paper did not provide specific results on each dataset. At the same time, it is recommended to expand the number of baseline models, which helps to more accurately evaluate the effectiveness of the method 3.The paper should compare with some facial compression methods such as [1][2][3].

Reference: [1] Qi Mao et al. Scalable Face Image Coding via StyleGAN Prior: Towards Compression for Human-Machine Collaborative Vision. TIP.2023 [2] Yuefeng Zhang et al. Machine Perception-Driven Facial Image Compression: A Layered Generative Approach. TCSVT.2024. [3] Shuai Yang et al. Towards Coding for Human and Machine Vision:: Scalable Face Image Coding.TMM.

问题

1.From the title of the paper, it appears to be a paper in the field of facial image encoding. But from the content of the article, it seems more like a downstream task in the field of facial image encoding. I recommend supplementing the experimental results according to the weakness on experiment mentioned in the previous text. 2.As the point mentioned in the weakness on method, in Equation (8) and in Equation (9) are essentially the same. The effectiveness of these two losses should be clearly discussed in the paper. Moreover, what are the consequences of using different routing strategies during training and inference due to the optimization dilemmas? It would be more convincing if the authors could address the issues related to their methods

局限性

Yes

最终评判理由

I maintain my original score. The image quality achieved by the proposed method does not appear to show significant improvement compared to the baseline. I understand that the scope of this work is focused on face recognition; however, I suggest revising the paper’s title to be more specific, as “face image compression” implies a substantially different concept in the compression domain. Given that the title explicitly claims “face image compression,” a thorough comparison with a broad range of existing methods in this area is expected. Currently, such comparisons are too limited—I have listed a few examples, but ideally, many more should be included. Moreover, the experiments primarily emphasize the low-bpp setting, without addressing whether the method can effectively compress high-resolution images. The papers cited mostly operate at very low bpp, and some are specifically designed for ID feature preservation, which can lead to particularly strong results in those conditions. As it stands, the compression application scenario targeted by this work appears very narrow. If this is positioned as a compression paper, from the perspective of the compression community, I do not consider it to fully meet that standard.

格式问题

None

作者回复

Answer for paper writing:

Thank you for pointing out the editorial error; we will correct it in future revisions. Specifically, regarding lines 284-285, our intention was to convey that our method is able to maintain—or even surpass—the baseline performance as the bpp decreases. For example, when bpp = 0.0469, the baseline achieves a MeanAcc of 90.70%, whereas our method achieves a MeanAcc of 92.23% at a lower bpp of 0.0391. We hope this clarification addresses your concerns.

Answer for method 1:

Selecting the most appropriate codebook for the model poses a significant challenge. During training, we experimented with directly using a nearest neighbor strategy for codebook selection. However, we observed that this approach resulted in a notably low codebook utilization rate of only 19.89%, indicating that the multiple codebooks were not being fully exploited. We attribute this to the nearest neighbor strategy’s tendency to favor codebooks that have been more thoroughly optimized, leading to a collapse where multiple codebooks degenerate into a single codebook.

Table: Comparisons of different codebook selection strategies used while training.

MethodCodebook Selection# Tokens# CodebooksCodebook SizeMeanAcc(%)IDSCodebook Mean Utilization(%)
Ours(TiTok-S)NN12816102489.280.570119.89
Ours(TiTok-S)Routing12816102489.890.574089.34

To address this issue, we designed a routing strategy based on entropy-regularized optimization during training, which encourages the model to select a more diverse set of codebooks and mitigates codebook overfitting, thereby promoting more comprehensive codebook training. During inference, we adopt the nearest neighbor strategy based on the minimum loss principle, as the codebook with the smallest loss corresponds to the upper bound of our model’s performance. We also experimented with using the routing strategy during inference and found that its performance was comparable to that of the nearest neighbor strategy, while offering a more efficient inference process. Specifically, it significantly reduced both inference latency and storage overhead. For example, under the TiTok-S architecture, the inference latency per image using the routing strategy was 0.0103 seconds, while remaining the MeanAcc of 87.54%, compared to 0.0154 seconds with the nearest neighbor strategy for the MeanAcc of 87.60%.

Table: Quantitative comparison between different codebook selection policies during inference.

BackboneCodebook SelectionbppMeanAcc(%)IDS
Maskgit-VQGANNN0.039192.230.6264
Maskgit-VQGANCR0.039192.220.6253
TiTok-SNN0.015787.600.5180
TiTok-SCR0.015787.540.5125
TiTok-LNN0.004066.020.1885
TiTok-LCR0.004065.650.1864

Answer for method 2:

We apologize for the oversight—there is an error in Equation 8 in the manuscript. The correct form is L_ent=_i=1Mg_θˉ(z_e)logg_θˉ(z_e)\mathcal{L}\_{\text{ent}} = \sum\_{i=1}^M \bar{g\_{\theta}}(\mathbf{z}\_e) \log \bar{g\_{\theta}}(\mathbf{z}\_e), where g_θˉ(z_e)\bar{g\_{\theta}}(\mathbf{z}\_e) denotes the average probability of the current class across samples in the batch. Equation 8 is designed to encourage the router to select diverse codebook groups for different samples within the same batch, thereby fully exploring and utilizing the expanded codebook latent space. In contrast, Equation 9 measures the entropy of the predicted distribution for each individual sample, aiming to suppress excessive uncertainty in single-sample predictions. Together with Equation 10, these three objectives not only promote comprehensive utilization of the codebook space, but also effectively balance inter-sample diversity with the certainty of individual predictions.

Answer for experiment 1:

You are absolutely right that including image quality metrics is essential for a comprehensive evaluation. In response, we have additionally assessed both 1D and 2D baselines, using three widely adopted metrics: SSIM, PSNR, and LPIPS. The results are summarized in the table below, where IBQ refers to the state-of-the-art 2D baseline published at ICCV 2025. As shown, our method achieves significant improvements in image quality, further demonstrating its effectiveness. Moreover, we observe that the LPIPS metric fluctuates across different baselines. We hypothesize that this is because the LPIPS score is computed using a network pre-trained on ImageNet to assess perceptual similarity between images, while there are substantial differences between facial data and ImageNet data. Therefore, we prefer to use MeanAcc and IDS as more reliable indicators of machine perception for facial compression tasks. We will incorporate these latest results in the revised version of our paper.

Table: Quantitative comparison about image quality with baseline methods.

MethodModel Type# TokensbppSSIM ↑PSNR ↑LPIPS ↓
IBQ2D2560.05080.852029.790.1880
MaskGIT2D2560.04690.790827.220.2204
TiTok-S1D1280.02340.779927.210.2323
TiTok-L1D320.00590.686823.250.2959
Ours(IBQ)2D2560.05080.860030.210.1765
Ours(MaskGIT)2D2560.04690.780227.270.2395
Ours(TiTok-S)1D1280.02340.782327.390.2354
Ours(TiTok-L)1D320.00590.693123.330.2968

Answer for experiment 2:

We have provided the detailed recognition accuracy and IDS metrics on five face recognition datasets, as shown in the table below.

Table: Recognition accuracy(%) on five facial datasets.

MethodModel TypebppLFWCFP_FPAgeDBCPLFWCALFWMean
MaskGIT2D0.046998.3389.1188.0388.3389.6790.70
TiTok-S1D0.023496.5185.2983.4186.1586.4587.56
TiTok-L1D0.005977.8463.0757.2962.1964.9765.07
Ours(MaskGIT)2D0.046999.4692.4091.0991.8292.7993.51
Ours(TiTok-S)1D0.023498.6890.2188.7789.1291.5291.66
Ours(TiTok-L)1D0.005987.3467.0765.7471.8273.6673.13

Table: IDS on five facial datasets.

MethodModel TypebppLFWCFP_FPAgeDBCPLFWCALFWMean
MaskGIT2D0.04690.64810.59890.61980.66430.62590.6314
TiTok-S1D0.02340.58830.55580.55700.60760.57330.5764
TiTok-L1D0.00590.17190.16820.17430.22150.16960.1812
Ours(MaskGIT)2D0.04690.67510.64280.67650.66540.66980.6659
Ours(TiTok-S)1D0.02340.61870.59670.61800.60720.61950.6120
Ours(TiTok-L)1D0.00590.25630.23810.24050.29630.26040.2583

Regarding your suggestion to compare with face compression methods, we have carefully reviewed the three representative papers you mentioned. Unfortunately, none of these works have released their code, and their experimental settings for the face compression task differ from ours. Specifically, these studies focus on preserving image details in high-resolution data. For example, [1] trains on the high-resolution FFHQ dataset and evaluates image quality on CelebA-HQ, which is not directly comparable to our setting, where the goal is to compress images from face recognition datasets while preserving identity features.

Nevertheless, we attempted to use screenshots of the visual results in Fig.6 from [2] and computed the IDS value between the original images and compressed images to assess the effectiveness in preserving facial identity. As shown in the table, our method demonstrates superior performance in identity preservation. In future versions, we will also include visual comparisons with these face compression methods.

Table: Comparison of IDS between other facial compression methods and ours.

MethodbppIDS ↑
MPDFIC[2]0.2460.4788
MPDFIC[2]0.0920.3571
MPDFIC[2]0.0450.1866
Ours(TiTok-S)0.02340.6357
Ours(TiTok-S)0.01570.5626
审稿意见
4

The authors propose a new vector quantization approach for low-bitrate image compression named Switchable Token-Specific Codebook Quantization (STSCQ). The approach addresses a key limitation of existing VQ-based compression models, where a globally shared codebook suffers at low bitrates due to reduced size. STSCQ introduces both image-level switchable codebooks (trained via routing) and token-specific sub-codebooks to better adapt to spatial and identity-specific variations in facial images. The method is evaluated on top of TiTok (Yu et al., 2024) and VQGAN (Esser et al., 2021), demonstrating improved performance on facial image reconstruction and recognition tasks at very low bpp.

优缺点分析

Strengths:

  • The idea of introducing switchable codebooks at the image-level and token-level to improve quantization capacity is well motivated.
  • The proposed training strategy appears compatible with standard latent-space models, making the approach plug-and-play.

Weaknesses:

  • While STSCQ maintains a similar bit allocation to global-codebook methods (with negligible overhead for routing), the decoder must store and access a large number of codebooks at inference time (up to M×TM \times T). This additional memory/storage overhead is not reported or discussed.
  • The ablation study in Section A.4 shows performance degradation when increasing the number of routing codebooks, suggesting potential limitations of the method on more complex or diverse datasets.
  • The paper’ presentation is not clear enough. See detailed questions below.

问题

  • In eq. 8, what is pi()p_i( \cdot )? The entropy loss seems incorrectly formulated, should it be ipilogpi\sum_i p_i \log p_i?
  • In eq. 12, the second C1C_1 seems a typo, should it be C2C_2? Also, does this imply that each of the MM image-level codebooks contains TT token-specific codebooks, for a total of M×TM \times T sub-codebooks, each being the same as CglobalC_{global}?
  • While the motivation for token-specific codebooks is reasonable, would assigning a separate codebook to each token lead to redundancy? Have you measured codebook utilization for each CijC^j_i? Are they all being used effectively?
  • For the main results (Table 1), how are the codebook size KK and the number of image-level codebooks MM selected?
  • In token-specific quantization, is only nearest-neighbor search (eq. 6) used? Why not extend the routing mechanism to this level as well to improve diversity?
  • The authors mention that the router is used only during training due to exploration-exploitation tradeoffs. Have you compared the posterior codebook utilization between learned routing and nearest-neighbor selection? This could further clarify the value and limitations of using a learned router.

局限性

The authors have discussed limitations.

最终评判理由

The rebuttal has addressed most of my concerns. While the proposed method introduces additional overhead, the authors have presented an alternative inference approach that maintains similar task performance with reduced overhead. Thus I decide to increase my scores. Additionally, I hope the authors can further improve the presentation and the clarity of their method in the revised manuscript.

格式问题

N/A

作者回复

Answer 1:

We have provided a detailed analysis of the required storage overhead and inference latency on a V100 GPU, as shown in the table below. Admittedly, selecting the codebook group based on the minimum error during inference does introduce additional overhead, which is particularly noticeable when using MaskGIT-VQGAN with 256 tokens as the backbone. To address this, we also experimented with directly utilizing the routing results during inference. In this approach, the model only needs to load the codebook group indicated by the routing decision, rather than maintaining all codebook groups on the decoder side. This strategy offers clear advantages, significantly reducing both storage overhead and inference latency. In the future, we also plan to further mitigate this issue through optimization techniques such as model quantization or pruning.

Table: Inference latency and memory overhead statistics. For ours method, 'NN' and 'CR' refer to Nearest-Neighbor search and Codebook Routing search respectively.

BackbonebppToken-specificCodebooksInfer Time (s / img)GPU Mem (MB)Storage (MB)FLOPs
Maskgit-VQGAN0.0469-10.1495474.1494.69158.14 G
Ours (NN)0.046910.15543283.391122.70158.14 G
Ours (NN)0.0391160.177112508.174194.70158.14 G
Ours (CR)0.0391160.15445116.15354.70158.14 G
TiTok-S0.0234-10.1433586.0398.68113.57 G
Ours (NN)0.023410.1437609.84122.68113.57 G
Ours (NN)0.01572560.14961894.54482.71113.57 G
Ours (CR)0.01572560.14501157.62100.21113.57 G
TiTok-L0.0059-10.17282709.241157.61269.62 G
Ours (NN)0.005910.17442716.801163.47269.62 G
Ours (NN)0.00402560.17502865.431253.50269.62 G
Ours (CR)0.00402560.17412804.021157.87269.62 G

Answer 2:

Our method can be readily transferred to images from other domains and is equally applicable to more complex and diverse datasets. For example, we report experimental results on ImageNet-1K, which contains more than 1M images from different categories. As shown in the table below, our approach achieves significant improvements over the original baseline across various metrics. Specifically, the FID score is reduced from 3.91 to 2.27, and the LPIPS score is improved from 0.3889 to 0.3564.

Table: Evaluation of reconstruction performance on the ImageNet-1K 256 × 256.

MethodbppFID ↓LPIPS ↓SSIM ↑PSNR ↑
TiTok-S0.02343.910.388917.480.4034
Ours0.02342.270.356417.670.4139

As discussed in Section A.4 of the supplementary material, we did observe that when increasing the number of routing groups while keeping the size of each codebook fixed, the evaluation metrics first improved and then declined. We attribute this to the limited size of our training dataset CASIA-WebFace, which is insufficient to fully utilize the large codebook space. Additionally, increasing the number of groups raises the routing complexity, resulting in under-trained embeddings within each codebook group. These factors together negatively impact the evaluation metrics. We plan to address this issue in our future work.

Answer for Q1:

We apologize for the oversight—there is an error in Eq. 8 in the manuscript. The correct form is L_ent=_i=1Mg_θˉ(z_e)logg_θˉ(z_e)\mathcal{L}\_{\text{ent}} = \sum\_{i=1}^M \bar{g\_{\theta}}(\mathbf{z}\_e) \log \bar{g\_{\theta}}(\mathbf{z}\_e), where g_θˉ(z_e)\bar{g\_{\theta}}(\mathbf{z}\_e) denotes the average probability of the current codebook group across samples in the batch. This loss is to encourage the router to assign different codebook groups to samples within the same batch, thereby making full use of the expanded codebook latent space.

Answer for Q2:

We appreciate you highlighting this oversight. You are correct—this was an editorial error, and the second C1C_1 should indeed be C2C_2. Regarding the codebook quantization in our method, when both switchable codebook quantization and token-specific codebook quantization are used in combination, the total number of sub-codebooks is indeed M×TM \times T. Furthermore, CglobalC_{global} refers to the globally shared codebook in previous methods, where all TT tokens share a common set of KK embeddings. We hope this clarification addresses your concern.

Answer for Q3:

We investigated the utilization of sub-codebooks when assigning separate codebooks to individual tokens. Our findings reveal that these sub-codebooks exhibit no sign of redundancy. Conversely, the token-specific codebook strategy not only further enhances the codebook utilization per token but also mitigates the imbalanced utilization allocation observed when using a single, shared codebook across all tokens, as shown in the table below. For those tokens with low utilization, we believe this may be due to their limited information content, which is related to the performance of the encoder. In the future, it would be reasonable to assign codebooks of varying sizes to different tokens to further reduce redundancy.

Table: Codebook utilization rates (%) per token on LFW dataset.

MethodbppMinMaxMeanSTD
Global-shared0.02343.4978.1254.1714.71
Ours0.023417.9083.8974.029.14

Answer for Q4:

We have now supplemented the main results in Table 1 by adding the corresponding number of codebook groups (MM) and the size of individual codebooks (KK). The updated results are shown below.

It is important to note that the main results presented here solely utilize the token-specific strategy. Consequently, the switchable codebook strategy—specifically designed to reduce the bits per pixel (bpp)—is not included in this main results table. Instead, it can be found in Table 2 within the main body of the paper.

Table: Detailed quantitative comparison with state-of-the-art methods.

MethodModel Type# Tokens# CodebooksCodebook SizeMeanAcc(%)IDSbpp
MaskGit-VQGAN2D2561409690.700.63140.0469
TiTok-S1D1281409687.560.57640.0234
TiTok-L1D321409665.070.18120.0059
Ours(MaskGit-VQGAN)2D2561409693.510.66590.0469
Ours(TiTok-s)1D1281409691.660.61200.0234
Ours(TiTok-l)1D321409673.130.25830.0059

Answer for Q5:

When quantizing each token, we employ the nearest neighbor search strategy, not the routing mechanism used at the image level. The reason for this design choice is that, if we were to continue using the routing selection mechanism for each token, it would indeed enhance the diversity of token representations. However, this approach would require storing an additional group index for every token, introducing extra overhead. This contradicts our goal of reducing bpp while increasing the overall codebook capacity. Specifically, assuming that image-level codebook routing is not considered, an image is represented by 256 tokens. If each token-specific codebook is divided into MM groups, each containing KK entries, then the number of bits required to represent an image is 256×(logM+logK)256 \times (\log M + \log K). For example, given an original codebook size of 4096, the required storage is 256×log4096=3072256 \times \log 4096 = 3072 bits. However, if we partition the codebook into 256 groups, each of size 256, the storage increases to 256×(log256+log256)=4096256 \times (\log 256 + \log 256) = 4096 bits.

Answer for Q6:

Yes, we have compared the codebook utilization between nearest neighbor search and our routing selection mechanism, and the results are shown in the table below. When using nearest neighbor search, both codebook utilization and evaluation metrics drop significantly—especially codebook utilization, which falls to only 19.89%. This clearly contradicts our original motivation for introducing multiple codebook groups. We believe this is because, with nearest neighbor search, the model tends to favor codebooks that have been more extensively optimized, leading to redundancy in other codebooks. Therefore, we introduce the routing selection mechanism to encourage the model to make full use of as many codebooks as possible, thereby leveraging the diversity of the codebook latent space.

Table: Comparisons of different codebook selection strategies used while training.

MethodCodebook Selection# Tokens# CodebooksCodebook SizeMeanAcc(%)IDSCodebook Mean Utilization(%)
Ours(TiTok-S)NN12816102489.280.570119.89
Ours(TiTok-S)Routing12816102489.890.574089.34
评论

Thank you for the rebuttal, which has addressed most of my concerns. While the proposed method introduces additional overhead, the authors have presented an alternative inference approach that maintains similar task performance with reduced overhead. I encourage the authors to further improve the presentation in the revised manuscript. I am increasing my scores accordingly.

审稿意见
4

This paper proposes a switchable token-specific codebook quantization framework for face image compression. To address the capacity limitations of conventional codebook-based image compression methods, it reorganizes the codebook at multiple levels to decompose the problem into smaller subproblems by integrating both an image-level switchable codebook and a token-level sub-codebook quantization mechanism. To enable effective training of this hierarchical structure, the authors further design a novel three-stage training strategy. The idea has practical significance and, as one of the first attempts to realize such a concept, demonstrates a clear degree of originality.

优缺点分析

The concept of using both image-level and token-level codebook quantization mechanisms is novel and timely, and the joint optimization approach is new in the context of codebook compression. During training, the integration of the MoE strategy and the loss functions is well designed. Moreover, with extensive benchmarks and detailed ablation studies, the experimental results are comprehensive and convincing. The paper is well written with clear figures and formulas. Finally, the proposed approach provides a novel framework for ultra-low bitrate face image compression, which can serve as a valuable resource for future research.

问题

a) However, the paper has several weaknesses. It does not include visual examples of the reconstructed compressed images, so the perceptual quality of the images is hard to assess. Furthermore, it remains unclear whether decreasing the single codebook size might damage the ability to preserve token diversity, thus potentially leading to information loss.

b) Also, the MoE-based routing design may lead to multiple codebooks learning redundant representations, but the paper does not provide explicit evidence to ensure codebook diversity. Additionally, the inference latency and memory overhead introduced by the multiple routing paths and token-specific codebooks are not analyzed, nor is the training resource requirement compared to other baselines.

c) The method has only been validated on face image datasets, so it remains unclear whether the token-specific strategy would generalize well to other domains such as natural scenes or anime images, where the data distribution could be more complex. In addition, the approach heavily relies on the quality of the pretrained autoencoder, and the impact of different encoder-decoder backbones has not been explored. It would also be helpful to clarify how well this design could transfer to real-world edge devices with limited computational or memory resources.

局限性

yes

最终评判理由

The reviewer went through the discussion and made the final remarks. This is the final decision.

格式问题

n/a

作者回复

Answer (a):

Thank you for your valuable suggestion. To better assess perceptual quality, we have included visualization examples in the supplementary material. However, due to policy limitations, we regret that we are unable to incorporate additional visual examples at this stage. To further validate the image generation quality of our method, we have conducted quantitative evaluations of image quality metrics, as presented in the table below. We are committed to providing more comprehensive visualizations in future iterations to further demonstrate the efficacy of our approach.

Table: Quantitative comparison about image quality with baseline methods.

MethodModel TypeTokensbppSSIM ↑PSNR ↑
TiTok-S1D1280.02340.779927.21
TiTok-L1D320.00590.686823.25
Ours(TiTok-S)1D1280.02340.782327.39
Ours(TiTok-L)1D320.00590.693123.33

Regarding "decreasing the single codebook size might damage the ability to preserve token diversity", in fact, directly reducing the size of a single codebook can indeed weaken its representational capacity. However, our method addresses this issue in two perspectives.

First, we argue that a more diverse or larger codebook feature space is not always optimistic solution; rather, it should be well-matched to the latent space of the encoder outputs. If the codebook covers a very broad space but the tokens cannot effectively utilize it, this will lead to low codebook utilization and wasted resources.

Second, We observed that the original strategy of using a global-shared codebook across all tokens leads to an uneven distribution of codebook utilization. As shown in the table below, different tokens exploit the shared codebook but with varying degrees. To address this, we propose a token-specific codebook strategy, where each token learns its own codebook to maximize the preservation of its visual characteristics. With our approach, the issue of uneven codebook utilization among tokens is significantly alleviated. This demonstrates that our codebook is better able to fit the feature space of the encoder outputs. As a result, the feature information of each token can be more effectively quantized by the current codebook, thereby reducing quantization errors caused by codebook utilization imbalance.

Table: Codebook utilization rates (%) per token on LFW dataset.

MethodbppMinMaxMeanSTD
Global-shared0.02343.4978.1254.1714.71
Ours0.023417.9083.8974.029.14

Finally, by introducing multiple groups at the image level and employing a routing strategy to assign the most suitable group to each image, our approach effectively increases the overall codebook capacity, even if the size of each individual codebook is reduced. Moreover, each codebook becomes more specialized in capturing the common features of a particular category of images, which helps maintain the diversity of the codebook representation space.

Table: Total codebook utilization rates (%) under five face recognition datasets.

bppTokensCodebooksCodebook SizeLFWCPLFWCFPFPAgeDBCALFWMean Utilization
0.02341281409674.0174.2460.5761.4467.0767.47
0.019612816102490.3994.6887.5386.1987.9389.34
0.015712825625693.1596.4594.4691.8992.0093.59

Answer (b):

To maintain the representational capacity of the codebook latent space as individual codebook size decreases, our method employs multiple groups of codebooks and effectively utilizes them. To better evaluate the utilization of these multiple codebook groups, we used the TiTok-S baseline as an example. We tested their codebook utilization under three different bpps, and the results are presented in the table below.

As shown in the table, the overall codebook utilization increases as the number of codebook groups grows. This result supports our hypothesis that, by organizing codebooks into groups, we can use smaller codebooks to capture more representative features. Consequently, this approach allows us to maintain the diversity of the codebook representation space while reducing the bpp during compression.

We have also evaluated the inference latency and storage overhead introduced by our method on a V100 GPU, as summarized in the table below. The table reports the inference latency and required storage for three baselines under different bpp (bits per pixel) settings. As shown, the expansion of the overall codebook size does indeed incur additional storage costs, along with a slight increase in inference latency. Since the number and dimensionality of tokens in the MaskGIT baseline are both significantly higher than those in the TiTok baseline (256 vs. 128 tokens, 256 vs. 12 dimensions), the storage overhead introduced by MaskGIT is substantially greater than that of TiTok. However, we would like to emphasize that in practical applications, our primary concern is the storage and transmission overhead of the images themselves. Our goal is to minimize image storage costs while preserving image quality as much as possible. Under the same bpp, our method achieves better recognition performance and image quality. Additionally, we have also explored approaches to reduce the model storage overhead. We explored a routing-based inference: only the codebook group selected by the router needs to be loaded, rather than searching with the minimum error across all codebooks. With this improvement, both inference latency and storage overhead are substantially alleviated, while the performance remains competitive with the baselines. There are also several engineering approaches to mitigate storage overhead. For example, we suggest performing quantization on the encoding side so that both the features and the codebook are matched within the int8 range. After obtaining the index sequence, the original fp32 codebook can be used on the cloud side for reconstruction and decoding.

Table: Inference latency and memory overhead statistics. For ours method, 'NN' and 'CR' refer to Nearest-Neighbor search and Codebook Routing search respectively.

BackbonebppToken-specificCodebooksInfer Time (s / img)GPU Mem (MB)Storage (MB)FLOPs
Maskgit-VQGAN0.0469-10.1495474.1494.69158.14 G
Ours (NN)0.046910.15543283.391122.70158.14 G
Ours (NN)0.0391160.177112508.174194.70158.14 G
Ours (CR)0.0391160.15445116.15354.70158.14 G
TiTok-S0.0234-10.1433586.0398.68113.57 G
Ours (NN)0.023410.1437609.84122.68113.57 G
Ours (NN)0.01572560.14961894.54482.71113.57 G
Ours (CR)0.01572560.14501157.62100.21113.57 G
TiTok-L0.0059-10.17282709.241157.61269.62 G
Ours (NN)0.005910.17442716.801163.47269.62 G
Ours (NN)0.00402560.17502865.431253.50269.62 G
Ours (CR)0.00402560.17412804.021157.87269.62 G

Table: Quantitative comparison between different codebook selection policies during inference.

BackboneCodebook SelectionbppMeanAcc(%)IDS
Maskgit-VQGANNN0.039192.230.6264
Maskgit-VQGANCR0.039192.220.6253
TiTok-SNN0.015787.600.5180
TiTok-SCR0.015787.540.5125
TiTok-LNN0.004066.020.1885
TiTok-LCR0.004065.650.1864

Regarding the comparison of training resources, the original paper reports that training TiTok-S requires 32 A100-40G GPUs for 50 hours. In contrast, our method primarily focuses on training the codebook, while the encoder is directly adopted from pretrained weights and the decoder is correspondingly fine-tuned. As a result, our training cost is 8 V100 GPUs for nearly 40 hours. We hope this clarification helps you better understand the differences in training resource requirements.

Answer (c)

Our proposed token-specific strategy can be easily adapted to images from other domains as well. To better evaluate the generalizability of our method, we also conducted experiments on the larger and more diverse ImageNet-1K dataset, with the results presented in the table below. As shown, our approach achieves significant improvements over the original baseline across various metrics. For example, the FID score is reduced from 3.91 to 2.27, and the LPIPS score is improved from 0.3889 to 0.3564. These results demonstrate the necessity and effectiveness of learning a separate codebook for each token.

Table: Evaluation of reconstruction performance on the ImageNet-1K 256 × 256.

MethodbppFID ↓LPIPS ↓SSIM ↑PSNR ↑
TiTok-S0.02343.910.388917.480.4034
Ours0.02342.270.356417.670.4139

As you mentioned, our method relies on the performance of the pre-trained autoencoder baseline, but it can be conveniently integrated into any existing codebook-based image compression framework. Here, we present updated results based on the IBQ baseline published at ICCV 2025 [1], where our approach also demonstrates its effectiveness, as shown in the table below.

Table: Evaluation on state-of-the-art IBQ.

MethodModel TypeTokensMeanAcc(%)IDSbpp
IBQ2D25695.730.84030.0508
Ours(IBQ)2D25696.130.85450.0508

The goal of image compression is to balance the trade-off between compression rate and distortion, aiming to minimize storage costs while preserving as much information from the original image as possible. In practical applications, we recommend deploying a lightweight encoder on edge devices to obtain high-quality compressed representations at lower bpp using our method, which can then be decoded or further processed by a decoder deployed in the cloud. Since our approach remains effective even at extremely low bpp (<0.01), it can significantly reduce the cost of image transmission and storage.

Reference: [1] Shi F, Luo Z, Ge Y, et al. Taming scalable visual tokenizer for autoregressive image generation. ICCV 2025.

最终决定

In this paper the aurhors introduce a switchable, token-specific codebook quantization framework for face image compression. Unlike conventional codebook-based methods, which are often constrained by limited capacity, the proposed approach restructures the codebook at multiple levels. It combines an image-level switchable codebook with a token-level sub-codebook quantization mechanism. This breaks down the overall task into smaller, more manageable subproblems. To effectively train this hierarchical design, the authors devise a novel three-stage training strategy. As one of the first implementations of this idea, the method shows both practical significance and a notable degree of originality. Strenghts include (i) novel idea of combining image-level and token-level codebook, (ii) novel idea of joint optimization approach for this problem, (iii) solid experimental evaluation, (iv) sensible ablation study, (v) good results, (vi) well written. Weaknesses include (i) lack of visual illustrations, (ii) only tested on face images, (iii) risk of redundant representations, (iv) high memory requirements for the decoder. The paper got mixed reviews before the rebuttal, but after the rebuttal and the following thorough discussions, the reviewers where predominantly positive with scores 4,4,3,5.