PaperHub
6.8
/10
Poster4 位审稿人
最低4最高5标准差0.4
5
4
4
4
3.8
置信度
创新性3.0
质量3.3
清晰度3.0
重要性3.0
NeurIPS 2025

FIPER: Factorized Features for Robust Image Super-Resolution and Compression

OpenReviewPDF
提交: 2025-05-10更新: 2025-10-29
TL;DR

An unified image representation to reconstruct the fine details.

摘要

关键词
Super-resolutionImage compression

评审与讨论

审稿意见
5
  1. The paper proposed a general feature decomposition method that can handle both image super resolution and compression.
  2. Instead of using fixed basis as fourier transform and wavelet transform, this paper is proposed to generate image specific basis, which help to improve the high-frequency repeating patterns recovery.
  3. Experiments show that the proposed method achieves SOTA performance on multiple benchmarks.

优缺点分析

  1. The experiments are solid. Toy example in Figure 2 shows the advantage of learnable basis.
  2. The pipeline is novel, which combines the advantage of both deep network and fixed basis.
  3. The description is a bit incoherent, and sometimes hard to follow. Figures (5) lack captions, which need some captions for better understanding.
  4. Experiments show that repeating patterns can be better recovered.

问题

  1. What is z in equation (9)? And what is the R here?
  2. The learned coefficients and basis are pixel-wise, but Fourier transform is global. So how to extend equation (2) to (3) is not clear to me.
  3. In Fig. 2(b), what is the meaning of "trainable basis settings" here, I thought x and y are fixed.
  4. Can authors explain more about the alpha in Figure. (4)?
  5. Any visualization of the extracted high-frequency features?
  6. How to learn the k? Is it a pure scalar or the output of some conv layers.
  7. Figure 5 looks like learning the coefficients and basis in parallel. I do not fully understand why authors claimed "our method derives the coefficient features first"
  8. Any Coordinate Transformation tried other than Sawtooth?

局限性

Can authors explain more about the following limitation? Although effective, our method requires further optimization for computation-limited scenarios such as real-time decoding.

最终评判理由

Based on the feedback from the authors, all of my concerns are solved and I have updated my scores.

格式问题

N/A

作者回复

We sincerely thank the reviewer for recognizing our solid experiments and novel pipeline. We apologize for the clarity issues and address each question comprehensively below:

Q1. Notation in Equation (9)

We apologize for any confusion caused by our notation. We mainly follow the notations and definitions in [1]. In Equation (9): L=R(y^)+R(z^)+λD(x,x^)L = R(\hat{y}) + R(\hat{z}) + \lambda\,D(x, \hat{x})

  • z^\hat{z} refers to the hyperprior, an auxiliary latent used to estimate the scale parameters for entropy coding, following the standard variational hyperprior framework.
  • R()R(\cdot) denotes the bit-rate cost of encoding a quantized variable, representing the rate term in the rate–distortion objective.

Thus, R(y^)+R(z^)R(\hat{y}) + R(\hat{z}) measures the total bit cost for both the primary latents and the hyperprior, and λD(x,x^)\lambda\, D(x, \hat{x}) balances this cost against reconstruction quality. We will clarify these definitions in the revised manuscript.

[1] Ballé, Johannes, David Minnen, Saurabh Singh, Sung Jin Hwang, and Nick Johnston. "Variational image compression with a scale hyperprior." ICLR 2018.

Q2. Pixel-wise vs Global Transform

Thank you for raising this point. In the classical Fourier representation, each frequency component is applied uniformly across the whole image: a single scalar coefficient multiplies a fixed sinusoidal pattern, so every pixel “sees” the same globally. This is powerful for capturing overall periodicity, but it cannot adapt to local variations in texture or detail.

In our Factorized Features framework, we generalize this in two key ways, trading pure global consistency for content-aware flexibility while retaining frequency modeling:

  1. Learnable, pixel-wise basis maps Instead of a fixed sinusoid, each basis becomes a full-resolution map that can vary from one pixel to the next, so that regions with different patterns (e.g., smooth sky vs. brick wall) get different frequency accommodations.

  2. Spatially-varying coefficient maps Rather than a single scalar per frequency, we produce a coefficient map for each basis. This allows each location to choose how strongly it uses each basis, adapting dynamically to local edge strength, texture complexity, or noise.

Together, these two changes turn the rigid global transform into a hybrid: - It still captures multi-frequency structure (low and high frequencies) across the image, - But it now automatically adapts to fine details and local content via learnable maps.

Why this matters

  • Local Detail Preservation: Regions that need more high-frequency emphasis (e.g., sharp edges) can boost those basis maps without affecting smooth regions.
  • Task Generality: This flexible decomposition works not only for super-resolution but also for compression (and, in principle, deblurring or denoising), since all these tasks benefit from recovering both broad structures and subtle textures.

In short, we move from “one size fits all” sinusoidal bases and scalar coefficients to a rich, pixel-wise basis–coefficient decomposition, which preserves the spirit of Fourier analysis while unlocking powerful local adaptation.

Q3. Figure 2(b) Clarification

We thank the reviewer for this question. In Fig. 2(b), “trainable basis settings” means that we enable gradient updates on the coordinate basis by setting require_grad=True for the (x,y) tensors. In other words, although initialized with fixed grid values, these coordinates become learnable parameters during training (see lines 170–171). We will clarify this point in the revised manuscript.

Q4. Alpha (α) in Figure 4

We can see in Figure 4 with sawtooth transformation γ\gamma the feature is prompted to learn pattern from different scale in a patch-like manner, e.g. γ2\gamma_2 makes the same basis repeat four times in the grid. However, only γ\gamma is not enough for very fine details; therefore, we introduce α\alpha and ψ\psi. Each basis is forced to accommodate both high-frequency (large α\alpha) and low-frequency (small α\alpha) components together with periodic functions ψ\psi, i.e., larger α\alpha will make ψ\psi oscillate sharper. To be more general, sawtooth transformation is to form a patch-like pattern, and α\alpha is to model pixel-wise high-frequency details.

Q5. High-frequency Feature Visualization

Note that due to rebuttal guidelines, we can not include images or external links here. We will include visual examples of the extracted high-frequency feature maps in the revised version to illustrate what the network learns.

Q6. Learning k

  • kk: Pure scalar hyperparameter in sawtooth transformation in line 162, where we use {256,128,64,32,16,8}\{256, 128, 64, 32,16,8\}. It is not the output of convolutional layers, found empirically using a validation set and fixed throughout training.
  • KK: The number of α\alpha. We conduct experiments to validate the setting.
N (the number of frequency components)111122666666661010
α\alpha (modulation frequencies){1}\{1\}{1,4,16,64}\{1,4,16,64\}{1,4,16,64}\{1,4,16,64\}{1}\{1\}{1,4}\{1,4\}{1,4,16,64}\{1,4,16,64\}{1,4,16,64,256,1024}\{1,4,16,64,256,1024\}{1,4,16,64}\{1,4,16,64\}
Set5 PSNR32.6232.6432.7032.6232.7132.7432.6932.73
Set5 SSIM0.90180.90200.90290.90200.90270.90320.90240.9033

We will make this explicit in the revised manuscript.

Q7. Sequential vs Parallel Processing

Although the coefficient and basis branches operate in parallel, the main feature XcoeffX_{coeff} is extracted via the backbone network. The basis transformer then downsamples those features to construct the basis. We will rephrase the caption and accompanying text to clearly separate these two stages.

Q8. Alternative Coordinate Transformations

Set5 Performance / γ\gammaSawtoothSinTriangular
PSNR32.7132.6932.71
SSIM0.90270.90260.9023

L1. Computational Limitations

Thank you for asking us to clarify this point. By “computation-limited scenarios,” we refer to environments such as on-device real-time decoding where both latency and memory are severely constrained. Although our Factorized Features design already reduces parameters, FLOPs, and peak memory compared to prior methods, the extra steps of basis generation and coefficient fusion still introduce non-negligible overhead under real-time constraints.

However, to demonstrate our method’s effectiveness even under strict computational constraints, we integrated the Factorized Features framework into state-of-the-art lightweight super-resolution models and observed consistent performance gains.

MethodLatency_ms#FLOPs_G#Params_KSet5_PSNR/SSIMSet14_PSNR/SSIMB100_PSNR/SSIMUrban100_PSNR/SSIMManga109_PSNR/SSIM
SwinIR-lt222.963.693032.44/0.897628.77/0.785827.69/0.740626.47/0.798030.92/0.9151
ELAN-lt18.054.164032.43/0.897528.78/0.785827.69/0.740626.54/0.798230.92/0.9150
OmniSR22.550.979232.49/0.898828.78/0.785927.71/0.741526.64/0.801831.02/0.9150
SRFormer-lt287.262.887332.51/0.898828.82/0.787227.73/0.742226.67/0.803231.17/0.9165
ATD-lt189.7100.176932.63/0.899828.89/0.788627.79/0.744026.97/0.810731.48/0.9198
HiT-SRF82.158.086632.55/0.899928.87/0.788027.75/0.743226.80/0.806931.26/0.9171
ASID-D861.849.6†74832.57/0.899028.89/0.789827.78/0.744926.89/0.8096-
MambaIR-lt55.884.692432.42/0.897728.74/0.784727.68/0.740026.52/0.798330.94/0.9135
MambaIRV2-lt153.475.679032.51/0.899228.84/0.787827.75/0.742626.82/0.807931.24/0.9182
RDN66.01309.22227132.47/0.899028.81/0.787127.72/0.741926.61/0.802831.00/0.9151
RCAN52.2917.61559232.63/0.900228.87/0.788927.77/0.743626.82/0.808731.22/0.9173
CATANet_CVPR25102.449.353532.58/0.899828.90/0.788027.75/0.742726.87/0.808131.31/0.9183
ESC_ICCV2521.9149.296832.68/0.901128.93/0.790227.80/0.744727.07/0.814431.54/0.9207
CATANet-F_Ours110.356.794132.69/0.901728.93/0.790627.82/0.744427.05/0.813331.62/0.9214
ESC-F_Ours24.673.495232.74/0.903228.99/0.793127.86/0.748327.11/0.815831.78/0.9233

S&W3. Figure 5 Caption

Thank you for highlighting this point. We agree that the presentation around Figure 5 can be made more coherent, and that explicit captions will help readers follow the workflow. In the revised manuscript, we will:

  • Clarify the narrative: Reorganize the text in Sections 4.2 and 4.3 to present the coefficient and basis extraction steps in a clear, sequential manner.
  • Add detailed captions:

This figure illustrates how Factorized Features are used in Super-Resolution and Image Compression. (a) Given a low-resolution input image, we first extract XcoeffX_{coeff} feature with Coefficient Backbone. Next, we generate the basis and coefficient with Basis Swin Transformer and convolution layers, respectively, from the same XcoeffX_{coeff}. Finally, the prediction is reconstructed by Factorized Features Reconstruction. (b) To decrease distortion in image compression, we replace the synthesis transform of the traditional learned image compression pipeline with (a) by aligning spatial resolution and latent channels.

These changes will ensure the description is coherent and that Figure 5 stands on its own as a self-contained illustration of our method.

评论

Thanks for the feedback from the authors. I think all my concerns are solved.

评论

Dear Reviewer evQ6,

Thank you for your effort in the review process, including the response to authors' feedback. If you believe the discussion with authors can be concluded, please submit the Mandatory Acknowledgement as required by the program of the conference.

Thank you,

AC

评论

Thank you again for your detailed feedback and for confirming that all concerns have been addressed. We appreciate your time and judgment.

审稿意见
4

This paper proposes "Factorized Features," a unified representation for low-level vision tasks, specifically demonstrating its effectiveness in Single Image Super-Resolution (SISR) and Image Compression. Unlike prior work focusing on network architecture, their approach uses a learned basis-coefficient decomposition with explicit frequency formulation to capture structural and multi-scale features. This representation is designed to generalize broadly, replacing simple feature maps in existing models. The authors further optimize image compression by leveraging a mergeable-basis property for multi-frame scenarios. Extensive experiments show state-of-the-art performance, with significant PSNR gains in SR and BD-rate reduction in image compression.

优缺点分析

Strength

  1. The concept of "Factorized Features" as a unified representation for diverse low-level vision tasks like SR and compression is a significant and novel contribution. This approach moves beyond purely architectural improvements to focus on fundamental image content representation.
  2. By replacing simple feature maps with Factorized Features, the paper successfully demonstrates the potential for broad generalizability across different low-level vision tasks and existing backbones.

Weaknesses:

  1. While "Learned Non-uniform Basis" is introduced, the mechanism for generating the entire basis map for better fitting and local reconstruction needs more in-depth explanation beyond just "networks to generate."

  2. While a "mergeable-basis property" is mentioned for multi-image compression, the explanation (Eq. 10 and related text) is brief and could be expanded to clarify the practical benefits and challenges in this setup.

问题

  1. For "broad generalizability," the paper only evaluates on two low-level vision tasks (SISR and image compression). Expanding to other tasks like denoising, deblurring, or inpainting would further strengthen the claim.

  2. The Lack of ablation: 2.1 Could a more comprehensive ablation study be provided to quantify the individual contribution of each innovative component (e.g., presence/absence of spatially variant coefficients, learned non-uniform bases, coordinate transformation, and multi-frequency modulation) to the final performance? This would be crucial for understanding which elements are the key drivers behind the observed performance improvements. 2.2 Could an ablation study on the choice of N (the number of frequency components) be provided? How do different values of N affect the model's performance, complexity, and its ability to capture fine image details? Is there an optimal value of N, or does it vary depending on the specific task?

局限性

See Questions.

最终评判理由

This paper presents a novel and effective framework for content-aware basis learning, with demonstrated benefits in low-level vision tasks. The authors have provided a thorough and convincing rebuttal, addressing all raised concerns with additional explanations, ablation studies, and extended evaluations on diverse tasks such as motion deblurring and dehazing.

格式问题

No

作者回复

We sincerely thank the reviewer for recognizing our work as "excellent" in quality and clarity, and for acknowledging our "significant and novel contribution." We address all concerns comprehensively below:

W1. Learned Non-uniform Basis Mechanism

In lines 89-100 of the Supplement, we describe implementation details of how the basis is generated. To be more clear, given an image, we extract a feature XcoeffX_{coeff} by Coefficient Backbone, and then the Basis Transformer is applied to refine the feature with downsampling for final basis generation. After the reconstruction with the coefficient and basis, the prediction is used to compute the loss and subsequently optimize the network. With gradient descent, when the Basis Transformer is optimized, the basis is optimized simultaneously, i.e., the models are trained to adapt to content-aware basis, so the basis is not calculated by heuristic rules. This design enables content-aware basis generation: smooth regions get low-frequency bases while textured areas receive high-frequency ones. We will expand this explanation with architectural diagrams in the revision.

W2. Mergeable-basis Property Explanation

Thank you for this valuable suggestion. We will expand the discussion around Eq. 10 in the main text to clarify both the benefits and limitations of our mergeable-basis approach. Specifically:

As detailed in lines 133–141 of the Supplementary, our merged basis exploits the observation that different images often share common frequency components. By learning a single, shared set of basis functions, we can leverage the frequently appearing similar or repetitive patterns across images to greatly reduce coding redundancy in multi-image compression. In practice, this means fewer total parameters and lower computational cost required when compressing an image collection. We also note the trade-off: when images are highly diverse, forcing them into one shared basis can increase reconstruction error (as seen by the quality drop for larger M in Fig. 7). Still, this content-aware information redundancy exploitation enables a significant reduction in resource usage, and we will make these points explicit in the revised manuscript.

Q1. Evaluation on More Tasks

We acknowledge this limitation and have extended our evaluation:

GoPro Dataset (Motion Deblurring)

Metric / MethodDeblurGAN-v2SRNDMPHNSDWNetMPRNetMIMO-UNet+DeepRFT+MAXIM-3SStripformerMSDI-netRestormerNAFNetFFTformerGRL-BMLWNet (CVPR'24)MLWNet-F
PSNR29.5530.2631.2031.2632.6632.4533.2332.8633.0833.2833.5733.6934.2133.9333.8334.4
SSIM0.9340.9340.9450.9660.9590.9570.9630.9610.9620.9640.9660.9670.9690.9680.9680.971

Dense-Haze Dataset (Dehazing)

Metric / MethodDehazeNetAOD-NetMSBDNFFA-NetAECR-NetDeHamerPMNetDehazeFormerTaylorFormerLH-NetMITNetPGH^2Net (AAAI'24)ConvIR (TPAMI'24)ConvIR-F
PSNR13.8413.1415.3714.3915.8016.6216.7916.2916.6618.8716.9717.0217.4518.12
SSIM0.430.410.490.450.470.560.510.510.560.5610.6060.610.8020.823

Our framework clearly generalizes across diverse low-level vision tasks. Due to time constraints in the rebuttal phase, we have focused on these two additional benchmarks, but we will include further evaluations on denoising, deraining, and other tasks in the revised manuscript.

Futhermore, to demonstrate our method’s broader effectiveness, we integrated the Factorized Features framework into state-of-the-art lightweight super-resolution models and observed consistent performance gains.

MethodLatency_ms#FLOPs_G#Params_KSet5_PSNR/SSIMSet14_PSNR/SSIMB100_PSNR/SSIMUrban100_PSNR/SSIMManga109_PSNR/SSIM
SwinIR-lt222.963.693032.44/0.897628.77/0.785827.69/0.740626.47/0.798030.92/0.9151
ELAN-lt18.054.164032.43/0.897528.78/0.785827.69/0.740626.54/0.798230.92/0.9150
OmniSR22.550.979232.49/0.898828.78/0.785927.71/0.741526.64/0.801831.02/0.9150
SRFormer-lt287.262.887332.51/0.898828.82/0.787227.73/0.742226.67/0.803231.17/0.9165
ATD-lt189.7100.176932.63/0.899828.89/0.788627.79/0.744026.97/0.810731.48/0.9198
HiT-SRF82.158.086632.55/0.899928.87/0.788027.75/0.743226.80/0.806931.26/0.9171
ASID-D861.849.6†74832.57/0.899028.89/0.789827.78/0.744926.89/0.8096-
MambaIR-lt55.884.692432.42/0.897728.74/0.784727.68/0.740026.52/0.798330.94/0.9135
MambaIRV2-lt153.475.679032.51/0.899228.84/0.787827.75/0.742626.82/0.807931.24/0.9182
RDN66.01309.22227132.47/0.899028.81/0.787127.72/0.741926.61/0.802831.00/0.9151
RCAN52.2917.61559232.63/0.900228.87/0.788927.77/0.743626.82/0.808731.22/0.9173
CATANet[1] (CVPR'25)102.449.353532.58/0.899828.90/0.788027.75/0.742726.87/0.808131.31/0.9183
ESC[2] (ICCV'25)21.9149.296832.68/0.901128.93/0.790227.80/0.744727.07/0.814431.54/0.9207
CATANet-F (Ours)110.356.794132.69/0.901728.93/0.790627.82/0.744427.05/0.813331.62/0.9214
ESC-F (Ours)24.673.495232.74/0.903228.99/0.793127.86/0.748327.11/0.815831.78/0.9233

[1] Liu, X., Liu, J., Tang, J., & Wu, G. (2025). CATANet: Efficient Content-Aware Token Aggregation for Lightweight Image Super-Resolution. In Proceedings of the Computer Vision and Pattern Recognition Conference (pp. 17902-17912).

[2] Lee, D., Yun, S., & Ro, Y. (2025). Emulating Self-attention with Convolution for Efficient Image Super-Resolution. arXiv preprint arXiv:2503.06671.

Q2. Comprehensive Component Ablation

The experiment is conducted with ESC-F as in Q1.

Comparison of multi-frequency modulation and the number of frequency components

N (the number of frequency components)111122666666661010
α\alpha (modulation frequencies){1}\{1\}{1,4,16,64}\{1,4,16,64\}{1,4,16,64}\{1,4,16,64\}{1}\{1\}{1,4}\{1,4\}{1,4,16,64}\{1,4,16,64\}{1,4,16,64,256,1024}\{1,4,16,64,256,1024\}{1,4,16,64}\{1,4,16,64\}
Set5 PSNR32.6232.6432.7032.6232.7132.7432.6932.73
Set5 SSIM0.90180.90200.90290.90200.90270.90320.90240.9033

Comparison of Coordinate Transformation

Set5 Performance / γ\gammaSawtoothSinTriangular
PSNR32.7132.6932.71
SSIM0.90270.90260.9023

We choose N=6N=6 for better quality and training efficiency (N=10N=10 converges slower). Note that due to time and resource constraint we omit spatial variant coefficients since it is quite intuitive and also learned non-uniform bases since it is discussed in Figure 2.

评论

Thanks for your detailed response! Overall, these results have mostly addressed my concerns. I will increase my score.

评论

Thank you for your kind feedback! We are happy to hear that our responses addressed most of your concerns. Please do not hesitate to point out any remaining questions we might have missed.

审稿意见
4

FIPER introduces factorized features, which can be used for low-level vision tasks such as super-resolution and image compression. Unlike previous works, FIPER decomposes features into basis and coefficients while explicitly modelling frequency components. It replaces standard feature maps and demonstrates strong generalization across both tasks. Experiments show significant performance improvement for both super-resolution and image compression tasks.

优缺点分析

Strengths: The core contribution, leveraging a frequency-aware decomposition (basis × coefficient) across both super-resolution and image compression, is well motivated and novel.

The paper reports significant improvements over prior SOTA on multiple SR benchmarks (e.g., Set5, Urban100) and achieves a 21% BD-Rate gain on image compression compared to VTM, outperforming several recent methods like TCM and ELIC.

Weaknesses:

The method requires pre-trained strong backbones (e.g., HAT, DAT). A comparison with lightweight SR/IC methods would improve practical relevance.

The paper presents Factorized Features as a “unified representation for low-level vision,” yet experiments cover only 4× SR and image compression; no evidence is given for de-blurring, denoising, and other low-level vision tasks.

问题

Could you report results (accuracy and runtime) with a lightweight CNN or Mobile-ViT backbone to demonstrate wider applicability?

Why choose N = 6 coefficient–basis pairs and alpha ∈ {1, 4, 16, 64}?  An ablation on N and frequency counts would help

Figure 3 lacks clarity, as the results for the toy example appear identical across all cases. Can you shed more light on it?

What would be the effect if, instead of applying the coordinate transformation \gamma to x in Equation 6, we use a sine or cosine transformation?

局限性

Please see weaknesses and questions

最终评判理由

The authors have addressed my questions in the rebuttal, particularly regarding the extension of their approach to other low-level vision tasks and the demonstration of its effectiveness with lightweight models. Therefore, I am inclined to recommend acceptance.

格式问题

Kindly change fig. and sec to Fig. and Sec. respectively Improve the readability of Fig.2 and Fig. 3

作者回复

We sincerely thank the reviewer for their positive assessment and constructive feedback. We are encouraged that they find our frequency-aware decomposition "well motivated and novel" and acknowledge our significant improvements (21% BD-Rate gain). We address each concern below:

W1. Lightweight Backbone Evaluation

Thank you for this important suggestion. We have conducted additional experiments with lightweight backbones and observed that Factorized Features provides consistent gains even in resource-constrained settings: lightweight models can also benefit from our Factorized Features.

MethodLatency (ms)#FLOPs (G)#Params (K)Set5 PSNR / SSIMSet14 PSNR / SSIMB100 PSNR / SSIMUrban100 PSNR / SSIMManga109 PSNR / SSIM
SwinIR-lt222.963.693032.44 / 0.897628.77 / 0.785827.69 / 0.740626.47 / 0.798030.92 / 0.9151
ELAN-lt18.054.164032.43 / 0.897528.78 / 0.785827.69 / 0.740626.54 / 0.798230.92 / 0.9150
OmniSR22.550.979232.49 / 0.898828.78 / 0.785927.71 / 0.741526.64 / 0.801831.02 / 0.9150
SRFormer-lt287.262.887332.51 / 0.898828.82 / 0.787227.73 / 0.742226.67 / 0.803231.17 / 0.9165
ATD-lt189.7100.176932.63 / 0.899828.89 / 0.788627.79 / 0.744026.97 / 0.810731.48 / 0.9198
HiT-SRF82.158.086632.55 / 0.899928.87 / 0.788027.75 / 0.743226.80 / 0.806931.26 / 0.9171
ASID-D861.849.6†74832.57 / 0.899028.89 / 0.789827.78 / 0.744926.89 / 0.8096
MambaIR-lt55.884.692432.42 / 0.897728.74 / 0.784727.68 / 0.740026.52 / 0.798330.94 / 0.9135
MambaIRV2-lt153.475.679032.51 / 0.899228.84 / 0.787827.75 / 0.742626.82 / 0.807931.24 / 0.9182
RDN66.01309.22227132.47 / 0.899028.81 / 0.787127.72 / 0.741926.61 / 0.802831.00 / 0.9151
RCAN52.2917.61559232.63 / 0.900228.87 / 0.788927.77 / 0.743626.82 / 0.808731.22 / 0.9173
CATANet[1] (CVPR'25)102.449.353532.58 / 0.899828.90 / 0.788027.75 / 0.742726.87 / 0.808131.31 / 0.9183
ESC[2] (ICCV'25)21.9149.296832.68 / 0.901128.93 / 0.790227.80 / 0.744727.07 / 0.814431.54 / 0.9207
CATANet-F (Ours)110.356.794132.69 / 0.901728.93 / 0.790627.82 / 0.744427.05 / 0.813331.62 / 0.9214
ESC-F (Ours)24.673.495232.74 / 0.903228.99 / 0.793127.86 / 0.748327.11 / 0.815831.78 / 0.9233

Due to time and resource constraints, we have focused on these initial evaluations on super-resolution; comparisons with lightweight image compression methods[3][4] will be included in the final version. We will add this analysis to the paper.

[1] Liu, X., Liu, J., Tang, J., & Wu, G. (2025). CATANet: Efficient Content-Aware Token Aggregation for Lightweight Image Super-Resolution. In Proceedings of the Computer Vision and Pattern Recognition Conference (pp. 17902-17912).

[2] Lee, D., Yun, S., & Ro, Y. (2025). Emulating Self-attention with Convolution for Efficient Image Super-Resolution. arXiv preprint arXiv:2503.06671.

[3] Bao, Y., Tan, W., Jia, C., Li, M., Liang, Y., & Tian, Y. (2025). ShiftLIC: Lightweight Learned Image Compression with Spatial-Channel Shift Operations. IEEE Transactions on Circuits and Systems for Video Technology.

[4] Wang, S., Cheng, Z., Feng, D., Lu, G., Song, L., & Zhang, W. (2024, December). Asymllic: Asymmetric lightweight learned image compression. In 2024 IEEE International Conference on Visual Communications and Image Processing (VCIP) (pp. 1-5). IEEE.

W2. Limited Task Coverage

We acknowledge this limitation and have extended our evaluation:

GoPro Dataset (Motion Deblurring)

Metric / MethodDeblurGAN-v2SRNDMPHNSDWNetMPRNetMIMO-UNet+DeepRFT+MAXIM-3SStripformerMSDI-netRestormerNAFNetFFTformerGRL-BMLWNet (CVPR'24)MLWNet-F
PSNR29.5530.2631.2031.2632.6632.4533.2332.8633.0833.2833.5733.6934.2133.9333.8334.4
SSIM0.9340.9340.9450.9660.9590.9570.9630.9610.9620.9640.9660.9670.9690.9680.9680.971

Dense-Haze Dataset (Dehazing)

Metric / MethodDehazeNetAOD-NetMSBDNFFA-NetAECR-NetDeHamerPMNetDehazeFormerTaylorFormerLH-NetMITNetPGH^2Net (AAAI'24)ConvIR (TPAMI'24)ConvIR-F
PSNR13.8413.1415.3714.3915.8016.6216.7916.2916.6618.8716.9717.0217.4518.12
SSIM0.430.410.490.450.470.560.510.510.560.5610.6060.610.8020.823

Our framework clearly generalizes across diverse low-level vision tasks. Due to time constraints in the rebuttal phase, we have focused on these two additional benchmarks, but we will include further evaluations on denoising, deraining, and other tasks in the revised manuscript.

Q1. Runtime with Lightweight Backbones

Please see W1. Note that ESC is convolution-based while CATANet is a hybrid transformer.

Q2. Hyperparameter Choices (N=6, α∈{1,4,16,64})

N111122666666661010
α\alpha{1}\{1\}{1,4,16,64}\{1,4,16,64\}{1,4,16,64}\{1,4,16,64\}{1}\{1\}{1,4}\{1,4\}{1,4,16,64}\{1,4,16,64\}{1,4,16,64,256,1024}\{1,4,16,64,256,1024\}{1,4,16,64}\{1,4,16,64\}
Set5 PSNR32.6232.6432.7032.6232.7132.7432.6932.73
Set5 SSIM0.90180.90200.90290.90200.90270.90320.90240.9033

We choose N=6N=6 for better quality and training efficiency (N=10N=10 converges slower).

Q3. Figure 3 Clarity

Figure 3 shows the difference between vanilla fields, coordinate transformation, and multi-frequency modulation. Specifically, 3.a is a vanilla basis-coefficient field, and 3.b adds sawtooth transformation. We can see in 3.b that the coordinate transformation explicitly models a patch-like pattern, e.g., the second plot from the top is divided into two periods. By enforcing such frequency components, the models can decompose the signal effectively.

Next, 3.d is our full formulation, while 3.c has no α\alpha. Intuitively, from 3.b to 3.c, applying ψ\psi (we use coscos here) should not have much effect on the performance since we do not have any constraint on these learnable bases. With the introduction of α\alpha, the models are forced to attend to signal components of different frequencies. Note that due to rebuttal guidelines, we can not include images or external links here; we will adjust the graph layout and font for better clarity.

Q4. Alternative Transformations

The experiment is conducted with ESC-F as in W1.

Set5 Performance / γ\gammaSawtoothSinTriangular
PSNR32.7132.6932.71
SSIM0.90270.90260.9023
评论

I thank the authors for their well-structured response. All of my questions have been addressed, and I will maintain my original rating.

评论

Thank you for your thorough review and positive feedback. We’re pleased that our revisions have addressed all of your questions, and we appreciate your decision to maintain the original rating. Thank you again for your valuable time and suggestions.

评论

Dear Reviewer Wt2E,

This is a reminder that the author-reviewer discussion period is ending soon on Aug. 6 (AOE), and you have not yet responded to authors' rebuttal. Please read authors' rebuttal as soon as possible, and engage in any necessary discussions, and consider if you would like to update your review and score. Please at least submit the Mandatory Acknowledgement as a sign that you have completed this task.

Thank you for your service in the review process.

AC

审稿意见
4

This paper is interesting. The authors proposes a 100M network to estimate the best bases and coefficients for feature representation for SR and compression tasks. Experiemental results show obvious improvment in terms of repetitive textures and fine details.

优缺点分析

A new network design for base and coefficient estimation. Can this network be applied to other tasks, such as image /video generations?

Strengths

  1. Unified Representation for Multiple Tasks: Factorized Features generalize across SISR and Image Compression, addressing core challenges (fine detail recovery, structural preservation) with a single framework, unlike task-specific prior methods.
  2. Explicit Frequency and Structural Modeling: By combining basis-coefficient decomposition, spatially variant coefficients, learned non-uniform bases, and multi-frequency modulation, the method effectively captures both low-frequency global structures and high-frequency local details, overcoming limitations of fixed sinusoidal bases (e.g., poor flexibility) and pure network architecture optimizations (e.g., neglect of frequency patterns).
  3. Optimized Compression Pipeline: The "mergeable-basis" property enables consolidating shared structures in multi-frame compression, reducing redundancy and improving efficiency.
  4. Strong Empirical Performance: Outperforms SOTA methods on benchmarks: In SR, achieves higher PSNR/SSIM across datasets (e.g., Set5, Urban100). In compression, delivers significant BD-rate reductions (e.g., -21.09% vs. VTM for single-image, -20.97% for multi-image with M=24). Cross-Task Prior Integration: Leverages pre-trained SR models as priors for image compression, enhancing information recovery.

Weaknesses

  1. Computational Efficiency: The method requires optimization for computation-constrained scenarios (e.g., real-time decoding) due to higher model complexity (e.g., larger parameters and MACs compared to some baselines, as shown in Table 1 and 2).
  2. Lack of Semantic Information: The framework does not incorporate semantic understanding, which could limit performance in tasks requiring context-aware detail preservation.
  3. Scalability Challenges in Multi-Frame Compression: While merging bases improves efficiency, distortion increases slightly as the number of merged frames (M) grows, stretching basis capacity.
  4. Dependence on Backbone Architecture: Performance gains rely on integrating with strong backbones (e.g., HAT, DAT), which may limit applicability to lightweight models.

问题

Failure cases can be given in the paper.

局限性

Can the Factorized Features be overfitted for certain tasks? i.e. task specific.

格式问题

No

作者回复

We sincerely thank the reviewer for their thorough and constructive feedback. We are encouraged that they find our paper "interesting" and acknowledge our contributions, including unified representation, explicit frequency modeling, optimized compression pipeline, and strong empirical performance. Below, we address each concern:

W1. Computational Efficiency

We acknowledge this important practical concern. In Table 1, only the parameter count of our HAT-F is slightly more than its counterpart, HAT-L; the MAC and Memory footprint of all our models are lower than their counterparts. In Table 2, our TCM-HAT-F-multi, though with slightly more parameters, actually achieves competitive encoding time and even the lowest decoding time of all models. Furthermore, to demonstrate the effectiveness of our method even under computation-constrained scenarios, we integrate our Factorized Features framework to state-of-the-art lightweight super-resolution models and observe consistent improvement.

MethodLatency (ms)#FLOPs (G)#Params (K)Set5 PSNR / SSIMSet14 PSNR / SSIMB100 PSNR / SSIMUrban100 PSNR / SSIMManga109 PSNR / SSIM
SwinIR-lt222.963.693032.44 / 0.897628.77 / 0.785827.69 / 0.740626.47 / 0.798030.92 / 0.9151
ELAN-lt18.054.164032.43 / 0.897528.78 / 0.785827.69 / 0.740626.54 / 0.798230.92 / 0.9150
OmniSR22.550.979232.49 / 0.898828.78 / 0.785927.71 / 0.741526.64 / 0.801831.02 / 0.9150
SRFormer-lt287.262.887332.51 / 0.898828.82 / 0.787227.73 / 0.742226.67 / 0.803231.17 / 0.9165
ATD-lt189.7100.176932.63 / 0.899828.89 / 0.788627.79 / 0.744026.97 / 0.810731.48 / 0.9198
HiT-SRF82.158.086632.55 / 0.899928.87 / 0.788027.75 / 0.743226.80 / 0.806931.26 / 0.9171
ASID-D861.849.6†74832.57 / 0.899028.89 / 0.789827.78 / 0.744926.89 / 0.8096
MambaIR-lt55.884.692432.42 / 0.897728.74 / 0.784727.68 / 0.740026.52 / 0.798330.94 / 0.9135
MambaIRV2-lt153.475.679032.51 / 0.899228.84 / 0.787827.75 / 0.742626.82 / 0.807931.24 / 0.9182
RDN66.01309.22227132.47 / 0.899028.81 / 0.787127.72 / 0.741926.61 / 0.802831.00 / 0.9151
RCAN52.2917.61559232.63 / 0.900228.87 / 0.788927.77 / 0.743626.82 / 0.808731.22 / 0.9173
CATANet[1] (CVPR'25)102.449.353532.58 / 0.899828.90 / 0.788027.75 / 0.742726.87 / 0.808131.31 / 0.9183
ESC[2] (ICCV'25)21.9149.296832.68 / 0.901128.93 / 0.790227.80 / 0.744727.07 / 0.814431.54 / 0.9207
CATANet-F (Ours)110.356.794132.69 / 0.901728.93 / 0.790627.82 / 0.744427.05 / 0.813331.62 / 0.9214
ESC-F (Ours)24.673.495232.74 / 0.903228.99 / 0.793127.86 / 0.748327.11 / 0.815831.78 / 0.9233

[1] Liu, X., Liu, J., Tang, J., & Wu, G. (2025). CATANet: Efficient Content-Aware Token Aggregation for Lightweight Image Super-Resolution. In Proceedings of the Computer Vision and Pattern Recognition Conference (pp. 17902-17912).

[2] Lee, D., Yun, S., & Ro, Y. (2025). Emulating Self-attention with Convolution for Efficient Image Super-Resolution. arXiv preprint arXiv:2503.06671.

W2. Lack of Semantic Information

Good catch! Our current focus is on frequency-domain structural modeling rather than semantic understanding. However, our framework is extensible - for example, we can insert the CLIP feature into the Coefficient Backbone [3] or condition attention on human instructions [4] in the Basis Transformer for explicit guidance. We will leave this to future work.

[3] Lu, Z., Xia, Q., Wang, W., & Wang, F. (2025). CLIP-aware domain-adaptive super-resolution. Multimedia Systems, 31(3), 257.

[4] Conde, M. V., Geigle, G., & Timofte, R. (2024, September). Instructir: High-quality image restoration following human instructions. In European Conference on Computer Vision (pp. 1-21). Cham: Springer Nature Switzerland.

W3. Scalability in Multi-Frame Compression

As shown in Figure 7, our method demonstrates robust scalability. The PSNR remains high even at M=24, with only a minor degradation of approximately 0.1 dB when increasing M from 8 to 24. This slight quality loss, resulting from more frames sharing common bases, is significantly outweighed by the twofold improvement in compression efficiency. This trade-off is more favorable than those offered by existing methods.

Regarding Table 2, the high BD-Rate observed at small values of M (e.g., M=1) is attributed to the overhead from the additional basis transmission branch, as detailed in Figure 10 of the Supplement.

W4. Backbone Architecture Dependence

We integrate our Factorized Features framework into state-of-the-art lightweight super-resolution models and witness consistent improvement as shown in the table of W1.

Q1. Failure Cases

Failure cases typically involve extremely fine textures—details that are absent in the low-resolution (LR) input but appear in the super-resolved (SR) output. However, this has been a shared inherent limitation of super-resolution methods to date. For example, in the top row of Figure 6, some white structures are present in HR but are completely absent in LR.

Note that due to rebuttal guidelines, we can not include images or external links here; visual examples of these failure cases will be provided in the revised manuscript.

L1. Task-Specific Overfitting

We thank the reviewer for this question. We understand it to ask whether our Factorized Features overfit to a single task or dataset. In fact, they exhibit strong cross-task and cross-domain generalization. For example, the same Factorized Features learned for super-resolution transfer seamlessly to image compression (as described in Section 4.3) and other low-level vision tasks such as image restoration (see below).

Even within super-resolution, a model trained exclusively on the DF2K dataset achieves excellent performance on held-out benchmarks (the lightweight SR comparison Table in Question 1 is trained with DF2K). If this does not address your concern, could you please clarify further?

GoPro Dataset (Motion Deblurring)

Metric / MethodDeblurGAN-v2SRNDMPHNSDWNetMPRNetMIMO-UNet+DeepRFT+MAXIM-3SStripformerMSDI-netRestormerNAFNetFFTformerGRL-BMLWNet (CVPR'24)MLWNet-F
PSNR29.5530.2631.2031.2632.6632.4533.2332.8633.0833.2833.5733.6934.2133.9333.8334.4
SSIM0.9340.9340.9450.9660.9590.9570.9630.9610.9620.9640.9660.9670.9690.9680.9680.971

Dense-Haze Dataset (Dehazing)

Metric / MethodDehazeNetAOD-NetMSBDNFFA-NetAECR-NetDeHamerPMNetDehazeFormerTaylorFormerLH-NetMITNetPGH^2Net (AAAI'24)ConvIR (TPAMI'24)ConvIR-F
PSNR13.8413.1415.3714.3915.8016.6216.7916.2916.6618.8716.9717.0217.4518.12
SSIM0.430.410.490.450.470.560.510.510.560.5610.6060.610.8020.823

Our framework clearly generalizes across diverse low-level vision tasks. Due to time constraints in the rebuttal phase, we have focused on these two additional benchmarks, but we will include further evaluations on denoising, deraining, and other tasks in the revised manuscript.

S&W1. Generalization

We can indeed extend our network to other tasks. For image generation, for example, we can integrate our Factorized Features into the VAE decoder of a diffusion model. For video generation, we can employ our mergeable basis and leverage its properties to enhance structural coherence. We will discuss these extensions in the Future Work / Directions section.

We hope these clarifications address the reviewer's concerns. We believe the strong empirical results, unified framework, and extensibility of our approach constitute significant contributions that merit acceptance. We will incorporate all suggested improvements in the final version.

评论

Dear Reviewer yir9,

This is a reminder that the author-reviewer discussion period is ending soon on Aug. 6 (AOE), and you have not yet responded to authors' rebuttal. Please read authors' rebuttal as soon as possible, and engage in any necessary discussions, and consider if you would like to update your review and score. Please at least submit the Mandatory Acknowledgement as a sign that you have completed this task.

Thank you for your service in the review process.

AC

最终决定

This submission presents Factorized Features, which represents an image by learning to produce both the expansion coefficient features and basis functions. With this representation, the work developed techniques for low-level vision tasks, particularly single image super-resolution and image compression, and empirical performance is shown to surpass existing approaches.

Reviewers and I appreciate the novelty of this representation framework, and the impressive empirical strength. Reviewers mentioned a few insufficiencies of the methods, primarily in the limited low-level vision task demonstration and efficiency issues, for which the authors provided more experimental results on deblurring and dehazing tasks and with light weighted models, with supportive performance. The authors also provided further clarifications of technical choices and result explanations. Overall, this paper constitutes a nice contribution to the community. I hope the authors to include the additional results and further clarifications into the final paper.