PaperHub
8.7
/10
Oral4 位审稿人
最低5最高6标准差0.4
6
5
5
5
5.0
置信度
创新性3.8
质量3.3
清晰度3.5
重要性3.3
NeurIPS 2025

Pan-LUT: Efficient Pan-sharpening via Learnable Look-Up Tables

OpenReviewPDF
提交: 2025-05-10更新: 2025-10-29

摘要

Recently, deep learning-based pan-sharpening algorithms have achieved notable advancements over traditional methods. However, deep learning-based methods incur substantial computational overhead during inference, especially with large images. This excessive computational demand limits the applicability of these methods in real-world scenarios, particularly in the absence of dedicated computing devices such as GPUs and TPUs. To address these challenges, we propose Pan-LUT, a novel learnable look-up table (LUT) framework for pan-sharpening that strikes a balance between performance and computational efficiency for large remote sensing images. Our method makes it possible to process 15K$\times$15K remote sensing images on a 24GB GPU. To finely control the spectral transformation, we devise the PAN-guided look-up table (PGLUT) for channel-wise spectral mapping. To effectively capture fine-grained spatial details, we introduce the spatial details look-up table (SDLUT). Furthermore, to adaptively aggregate channel information for generating high-resolution multispectral images, we design an adaptive output look-up table (AOLUT). Our model contains fewer than 700K parameters and processes a 9K$\times$9K image in under 1 ms using one RTX 2080 Ti GPU, demonstrating significantly faster performance compared to other methods. Experiments reveal that Pan-LUT efficiently processes large remote sensing images in a lightweight manner, bridging the gap to real-world applications. Furthermore, our model surpasses SOTA methods in full-resolution scenes under real-world conditions, highlighting its effectiveness and efficiency.
关键词
Look-Up TablePan-sharpening

评审与讨论

审稿意见
6

This paper introduces Pan-LUT, a novel learnable Look-Up Table (LUT) framework designed for pan-sharpening. The core contribution lies in the development of a computationally efficient approach that balances performance with resource constraints, particularly for deployment in scenarios where real-time processing is crucial. The experimental results demonstrate that Pan-LUT achieves competitive performance, particularly in terms of computational efficiency, when compared to existing pan-sharpening methods. Specifically, the proposed method exhibits significantly faster inference times than deep learning-based approaches, while maintaining a reasonable level of accuracy.

优缺点分析

Strengths
1、The primary strength of this paper lies in its introduction of a novel learnable LUT framework for pan-sharpening.
2、The proposed framework achieve a practical balance between performance and efficiency, making it a potentially valuable contribution to the field of remote sensing image processing.
3、Quantitative results demonstrate that the proposed framework substantially outperforms classical approaches. Moreover, it obtains acceptable results with respect to the deep-learning based models, but with a considerably lower processing time.
Weaknesses
1、The effectiveness of the proposed method, while reasonable, does not consistently achieve state-of-the-art performance when compared to some deep learning-based methods.
2、The paper lacks illustrative figures that clearly demonstrate the operations of the proposed LUT modules.

问题

1、The authors should provide a more detailed analysis of the memory complexity. This should include a breakdown of the memory requirement of each component of the proposed method, such as the PGLUT, SDLUT, and AOLUT modules. 2、Further evaluation of additional Pan-LUT variants (e.g., Pan-LUT-4 and Pan-LUT-5) is necessary to better assess the trade-off between performance and computational efficiency. 3、How are the hyperparameters λs and λm​ in Eq (11) set? Here, λs controls the smoothness regularization term, while λm​ controls the monotonicity regularization term in the loss function.

局限性

N

最终评判理由

The responses have solved all my concerns,so I decide to raise the final score.

格式问题

none

作者回复

Thank you for your review.

Weakness 1

Our Pan-LUT is specifically designed to efficiently process high-resolution remote sensing images under resource-constrained conditions. While it does not consistently achieve state-of-the-art performance when compared to some deep learning-based methods, it consistently outperforms other approaches in real-time inference and high-resolution image processing scenarios. Its lightweight architecture enables practical deployment, effectively bridging the gap to real-world applications.

Weakness 2

Since it is challenging to visualize operations in 4D or 5D space, we instead provide illustrative figures based on 3D LUTs to clearly demonstrate the structure and operations of the proposed LUT modules. For more details, please refer to the Technical Details section in the supplementary material.

Question 1

ModuleM=3M=6M=9M=17M=33
Param (K)Param (K)Param (K)Param (K)Param (K)
AOLUT0.9731.10236.205679.43156541.57
SDLUT0.081.306.5683.521185.92
PGLUT1.2138.88295.247099.28195676.96

The memory complexities of PGLUT, SDLUT, and AOLUT are O(5M5)O(5M^5), O(4M4)O(4M^4), and O(4M5)O(4M^5), respectively, where MM denotes the size of the LUT. As shown in Table above and Figure 5 of the manuscript, increasing the LUT size beyond 9 yields only minor performance gains but significantly consumes excessive storage resources. In all experiments, we use M=9M=9 as the default setting.

Question 2

We conduct additional experiments to evaluate more Pan-LUT-N variants, where NN denotes the number of stacked PGLUT and SDLUT modules.

MethodWV2GF2WV3Param (M)Inference (ms)Inference (ms)
2048times20482048\\times20484096times40964096\\times4096
Pan-LUT-140.636243.655929.62130.95790.38120.5445
Pan-LUT-240.841944.554529.79291.38430.94341.0824
Pan-LUT-341.136244.855930.12131.81081.48211.9648
Pan-LUT-441.162444.975630.19132.23732.06212.7518
Pan-LUT-541.285445.173330.24712.66372.37122.9915
Pan-LUT-641.402545.336430.30873.09022.79153.1024

As shown, deeper models (e.g., Pan-LUT-4 to Pan-LUT-6) offer only marginal improvements in performance compared to shallower variants (Pan-LUT-1 to Pan-LUT-3), while significantly increasing parameter count. The Pan-LUT-4 already contains over 2 million parameters. From a practical application perspective, we consider Pan-LUT-1, 2, and 3 as the default architectural choices.

Question 3

Metricslambdam\\lambda_m=0lambdam\\lambda_m=0.1lambdam\\lambda_m=1lambdam\\lambda_m=10lambdam\\lambda_m=100
PSNR29.428729.490229.527529.621329.3785
Metricslambdas\\lambda_s=0lambdas\\lambda_s=0.00001lambdas\\lambda_s=0.0001lambdas\\lambda_s=0.001lambdas\\lambda_s=0.01
PSNR29.416329.417629.621329.413729.4044

We tune the hyperparameters lambdas\\lambda_s and lambdam\\lambda_m to identify the optimal settings, and find that lambdam=10\\lambda_m = 10 and lambdas=0.0001\\lambda_s = 0.0001 yield the best performance. A large λs\lambda_s (e.g., >0.0001>0.0001) results in worse PSNR, as the smooth regularization limits the flexibility of LUT transformations. In contrast, the PSNR is insensitive to the choice of lambdam\\lambda_m, since monotonicity is an inherent property of LUTs.

评论

I thank the authors for providing such a detailed response in the rebuttal. After reading it carefully, I found that the authors addressed most of the issues that I raised in the initial phase. For instance, the technical details of method and extral experiments. To this end, I choose to keep my score.

评论

Thank you for taking your valuable time to carefully read our rebuttal. We are honored that our efforts to address the issues you raised. We sincerely thank you for your time and professional review.

审稿意见
5

The paper introduces Pan-LUT, a look-up table (LUT) based method designed for the pan-sharpening task. Unlike contemporary deep learning-based approaches that require significant computational resources due to their complex network architectures, the proposed method is computationally efficient. The proposed framework incorporates three types of learnable LUTs that map values from low-resolution multispectral inputs to high-resolution multispectral outputs.

优缺点分析

Strengths:

  1. The idea of using look-up table for efficient pan-sharpening is new. To the best of our knowledge, this work is the first study on LUT-based pan-sharpening.
  2. The proposed method is fast, and it can process 4K-resolution images in under 1 ms using a single NVIDIA GeForce RTX 2080 Ti GPU.
  3. The approach is simple, and its computational efficiency is outstanding.

Weaknesses:

  1. The proposed method focuses on computational efficiency, which may come at the cost of achieving the highest performance.
  2. Some parts of the paper lack clarity. For example, the meaning of "OOM" in Table 1 is not explained.

问题

  1. The authors need to analyze the impact of different LUT sizes on inference time.
  2. What is the maximum image resolution that the proposed method can handle on the specified GPU?.
  3. What contributions does the rotation ensemble strategy outlined in Section 3.3 make to the performance or robustness of the approach?

局限性

yes

最终评判理由

The author has addressed all the issues I was concerned about, so I have decided to increase the score.

格式问题

None

作者回复

Thank you for your review.

Weaknesses1

Our Pan-LUT is specifically designed to efficiently process high-resolution remote sensing images under resource-constrained conditions, even if this focus on computational efficiency may come at the cost of achieving peak performance. Its lightweight architecture enables practical deployment, effectively bridging the gap to real-world applications.

Weaknesses2

"OOM" in Table 1 refers to “Out of Memory”. We will revise the table caption and related descriptions for better clarity.

Question1

We independently evaluate the inference time of each LUT module with different sizes MM to assess their computational efficiency, where MM denotes the number of bins per dimension. Additionally, we evaluate the SDLUT module equipped with the Rotation Ensemble Strategy (RES) for comparison. All experiments are conducted using input images with a resolution of 4096times40964096 \\times 4096.

ModuleM=3M=6M=9M=17M=33
Inference (ms)Inference (ms)Inference (ms)Inference (ms)Inference (ms)
AOLUT0.06770.10780.12470.24510.2988
SDLUT + RES0.21270.25690.32750.71420.9564
SDLUT0.04690.06850.10150.14740.2063
PGLUT0.10270.11790.13450.20460.2779

From this table, we observe that the RES operation contributes to the majority of the inference time, as any operation on high-resolution images tends to be computationally expensive.

Question 2

We evaluate the maximum image resolution that our model can handle on different GPUs.

(e.g. '1K' denotes an image size of 1000times10001000 \\times 1000 pixels.)

Models2080Ti (11G)3080Ti (12G)3090 (24G)A40 (46G)A800 (80G)
resolutionresolutionresolutionresolutionresolution
Pan-LUT-18.2K8.5K11.1K16.2K20.2K
Pan-LUT-27.5K7.7K9.4K12.4K14.2K
Pan-LUT-35.6K5.7K7.3K10.1K11.6K

Our method is sufficiently lightweight to process 4K to 8K resolution images on a single NVIDIA 2080Ti GPU.

Question 3

We demonstrate the advantages of the Rotation Ensemble Strategy (RES) from both theoretical and empirical perspectives.

Reducing Memory Usage. The RES effectively increases the receptive field of SDLUT from 2times22 \\times 2 to 3times33 \\times 3. Without RES, achieving the same receptive field would require increasing the LUT dimensionality from 4 to 9, leading to a memory complexity of O(N9)O(N^9). In contrast, RES enables a larger contextual receptive field without increasing LUT size, thereby reducing memory usage.

Local Texture Enhancement. Let (w,h)(w, h) denote the pixel located at the w-th column and h-th row of the image. We provide a proof that RES extends the receptive field from 2times22 \\times 2 to 3times33 \\times 3. Ideally, we aim to incorporate all pixels within the local 3times33 \\times 3 region centered at (w,h)(w, h), corresponding to the input index set:

Gtgt=(w1,h1),(w,h1),(w+1,h1),(w1,h),(w,h),(w+1,h),(w1,h+1),(w,h+1),(w+1,h+1).G_{tgt} = \\{(w{-}1,h{-}1), (w,h{-}1), (w{+}1,h{-}1), (w{-}1,h), (w,h), (w{+}1,h), (w{-}1,h{+}1), (w,h{+}1), (w{+}1,h{+}1)\\}.

By default, SDLUT focuses on the bottom-right neighborhood of each pixel. The input index group is defined as

Gr0=(w,h),(w+1,h),(w,h+1),(w+1,h+1).G_{r0} = \\{(w,h), (w+1,h), (w,h+1), (w+1,h+1)\\}.

With the proposed RES, we apply SDLUT across three rotated versions of the input, effectively capturing additional pixel neighborhoods: Gr90=(w,h),(w1,h),(w,h+1),(w1,h+1)G_{r90} = \\{(w,h), (w{-}1,h), (w,h{+}1), (w{-}1,h{+}1)\\},

Gr180=(w,h),(w1,h),(w,h1),(w1,h1)G_{r180} = \\{(w,h), (w{-}1,h), (w,h{-}1), (w{-}1,h{-}1)\\},

Gr270=(w,h),(w+1,h),(w,h1),(w+1,h1)G_{r270} = \\{(w,h), (w{+}1,h), (w,h{-}1), (w{+}1,h{-}1)\\}.

Then, we can derive the following equation:

Gr0cupGr90cupGr180cupGr270=Gtgt.G_{r0} \\cup G_{r90} \\cup G_{r180} \\cup G_{r270} = G_{tgt}.

Definition: w/o RES' means the framework without the Rotation Ensemble Strategy.

MethodWV2GF2WV3
Pan-LUT-140.636243.655929.6213
Pan-LUT-1 (w/o RES)40.499343.214729.3354
Pan-LUT-240.841944.554529.7929
Pan-LUT-2 (w/o RES)40.724744.375429.6786
Pan-LUT-341.136244.855930.1213
Pan-LUT-3 (w/o RES)40.998844.667230.0455

We find that RES not only contributes to performance improvement but also significantly reduces the number of parameters.

评论

The author's rebuttal dispelled my doubts, so I decided to increase the score.

评论

Thank you very much for your positive feedback. We are delighted to hear that our rebuttal has successfully addressed your concerns. We sincerely appreciate you increasing the score for our manuscript. Your kind consideration is a great encouragement to us. Thank you once again for your valuable time and constructive comments.

审稿意见
5

Pan-LUT bridges the gap between computational efficiency and performance in high-resolution pan-sharpening. By replacing complex DNN operations with learnable LUTs, it achieves speedup over SOTA methods while maintaining competitive accuracy. This work pioneers hardware-friendly LUT designs for remote sensing, with implications for real-time Earth observation systems.

优缺点分析

Strengths:

  1. This paper proposed a learnable LUT framework for pan-sharpening, breaking computational bottlenecks of traditional DNNs.

  2. Pan LUT consists of three core modules: PGLUT for controlling the spectral transformation, SDLUT for capturing spatial details, and AOLUT for adaptively Aggregating channel information.

  3. Pan LUT is capable of processing ultra-high resolution (such as 4k) images, which many methods cannot achieve.

Weaknesses:

  1. There is still a gap compared to SOTA methods in terms of spectral fidelity and spatial detail preservation. In addition, Lightweight is one of the characteristics of the proposed method, but it has not been compared with recent lightweight methods in experiments.

  2. There is no specific explanation on how SDLUT processes boundary pixels in section 3.

  3. The dataset used in the experiment has a limited number of spectra, and the proposed method did not consider how to cope with the increase in the number of spectra, especially in hyperspectral scenarios.

问题

  1. This paper aims for "real-world applicability" but ignores key satellite-specific resolution variations, undermining claims like bridging the gap to real-world applications. This method has been experimentally tested with a fixed downsampling factor of 4. How to deal with situations where the downsampling factor is not fixed?

  2. This method has only been validated in scenarios with a small number of spectra. How to deal with a large number of input data spectra?

  3. In this paper, it is mentioned multiple times that the application of existing DL methods is limited in scenarios without GPUs and CPUs, overlooking critical edge device validation despite claiming “lightweight” and “hardware-friendly” attributes. Can comparative data be provided on edge devices?

局限性

Yes

最终评判理由

The authors have addressed my concerns. I lean to keep the score.

格式问题

There are no major formatting issues.

作者回复

Thank you for your review.

Weakness 1

[1] LGPConv-Learnable-Gaussian-Perturbation-Convolution-for-Lightweight-Pansharpening. IJCAI 2023

[2] Deep Adaptive Pansharpening via Uncertainty-aware Image Fusion. TGRS 2023

[3] SSUN-Net: Spatial-Spectral Prior-Aware Unfolding Network for Pan-Sharpening. AAAI 2025

UAPN-S is the most lightweight model variant proposed in [2]. SSUN-Net-1stg is refers to the version with a single reconstruction stage in [3].

MethodWV2GF2WV3Param(M)Inference(ms)Inference(ms)Inference(ms)Inference(ms)
2048times20482048\\times20484096times40964096\\times40962048times20482048\\times2048 (CPU)4096times40964096\\times4096 (CPU)
Pan-LUT-140.636243.655929.62130.95790.38120.54451012.155338.36
Pan-LUT-240.841944.554529.79291.38430.94341.08241945.789664.84
Pan-LUT-341.136244.855930.12131.81081.48211.96482844.3212523.24
LGPConv41.674246.664530.42320.026227.15OOM25113.29126547.26
UAPN-S41.566346.821230.30650.0345OOMOOM78775.02463461.26
SSUN-Net-1stg41.666346.634330.33710.0596163.69OOM34446.80202476.98

We show that current lightweight methods are only lightweight in the context of low-resolution images. These approaches are highly sensitive to input image size and rely heavily on specialized hardware such as GPUs or TPUs. As shown in Table, in the absence of GPU acceleration, these methods require a considerable amount of time to process images. UAPN-S is particularly sensitive to input resolution and fails to process 2048times20482048 \\times 2048 images even on a single 2080Ti GPU.

Weakness 2

To handle boundary pixels, we adopt a reflect padding operation, which pads the image by mirroring adjacent pixels at the borders.

Weakness 3 and Question 2

To handle a larger number of spectral channels, one can simply increase the dimensionality of PGLUT. However, this leads to a rapid growth in the number of parameters. For example, in the case of PGLUT with 12 input spectral channels and a LUT size of 3, the number of parameters reaches 12times312=6,377,29212 \\times 3^{12} = 6{,}377{,}292. Since LUTs store precomputed output values to avoid redundant computation, the resulting parameter explosion is inevitable. A practical solution is to incorporate smart architectural designs—such as the proposed Rotation Ensemble Strategy (RES)—to mitigate the memory and parameter overhead.

Rank Factorization. In multispectral and hyperspectral scenarios, rank factorization offers a promising direction to reduce the parameter count and memory consumption of the model. We plan to explore this strategy in future research.

Question 1

Typically, the LRMS images are upsampled to match the PAN resolution using bicubic interpolation before being fed into model, which is a widely adopted practice in pansharpening literature. Our method follows this approach and is inherently compatible with varying resolution ratios. Regardless of the original downsampling factor, once the LRMS image is interpolated to the PAN resolution, Pan-LUT can be applied without requiring any architectural modifications. This ensures the robustness and applicability of our method in real-world scenarios with diverse resolution settings.

Question 3

Our method focuses on scenarios involving high-resolution image processing, particularly in environments where GPUs or TPUs are unavailable, as stated in Lines 6 and 50 of the manuscript. As shown in the table provided in our response to Weakness 1, our method can be efficiently deployed on CPUs while maintaining computational efficiency. In practice, many edge devices, such as the NVIDIA Jetson and Rockchip RK series, are equipped with CPUs. And our method can be deployed on any edge device equipped with a CPU.

评论

I appreciate the authors' detailed responses to my concerns regarding the upsampling ratio, the lack of comparison with recent lightweight SOTA methods, and the absence of the testing results on edge devices. The clarifications have adequately addressed these points and I will keep my original score.

评论

Thank you very much for your continued attention and thoughtful review of our work. Your professional insights have been invaluable in enhancing the quality of our paper. We sincerely appreciate your time and constructive comments.

审稿意见
5

In this paper, to achieve high computational efficiency and a lightweight design, the authors propose a fully LUT-based framework. Specifically, they design three types of LUT modules: PGLUT for spectral information exchange across channels, SDLUT for capturing fine-grained local spatial details, and AOLUT for aggregating multimodal information to produce adaptive high-resolution results. Experimental results demonstrate that the proposed method achieves superior computational efficiency while delivering performance comparable to state-of-the-art deep learning-based approaches.

优缺点分析

Strengths:

  1. This method is capable of processing high-resolution images even in resource-constrained environments. As demonstrated in the paper, it can handle images with resolutions up to 8K and 16K.

  2. The proposed method is novel, fast, and computationally efficient, with a simple yet effective design that makes it practical for real-world applications.

Weaknesses:

  1. As mentioned in the paper, the storage requirements increase substantially with the size and number of LUTs, which is an inherent limitation of LUT. This reflects a trade-off between performance and computational efficiency, as well as between memory requirements and processing speed.

  2. Some details in the writing are not clearly presented. The paper uses two different forms of resolution description. In Section 1, the terms "2K" and "4K" are used, while in Section 4.3 (line 269), the resolutions are specified as 2048 × 2048 and 4096 × 4096. It is unclear whether "2K" specifically refers to 2048 × 2048. This inconsistency may cause confusion and should be clarified.

问题

  1. Pan-LUT-N denotes the model variant with N stacked PGLUT and SDLUT modules, analogous to increasing network depth. It would be valuable to evaluate its generalization ability across different resolutions by testing on full-resolution images and assessing the potential risk of overfitting.

  2. The rotation ensemble strategy is typically used as a Test-Time Augmentation (TTA) technique. However, its effectiveness during training remains unclear and should be validated through ablation studies or comparisons with appropriate baselines.

  3. Can quadrilinear and pentalinear interpolation be regarded as higher-dimensional extensions of bilinear interpolation?

局限性

yes

最终评判理由

The author added additional experiments to address the issues I was concerned about and provided detailed formula derivations, so I decided to increase my score.

格式问题

no

作者回复

Thank you for your review.

Weakness 1

There is a trade-off between performance and computational efficiency associated with LUT-based methods. However, this overhead is acceptable given the performance benefits achieved, especially in real-time inference and high-resolution image processing scenarios.

Weakness 2

"2K" specifically refers to a resolution of 2048times20482048\\times2048, while "4K" refers to 4096times40964096\\times4096, we will revise the manuscript to clarify the actual image sizes used in our experiments to avoid ambiguity.

Question 1

We conduct additional experiments to evaluate more Pan-LUT-N variants, where NN denotes the number of stacked PGLUT and SDLUT modules.

MethodWV2GF2WV3Param(M)Inference(ms)Inference(ms)
2048times20482048\\times20484096times40964096\\times4096
Pan-LUT-140.636243.655929.62130.95790.38120.5445
Pan-LUT-240.841944.554529.79291.38430.94341.0824
Pan-LUT-341.136244.855930.12131.81081.48211.9648
Pan-LUT-441.162444.975630.19132.23732.06212.7518
Pan-LUT-541.285445.173330.24712.66372.37122.9915
Pan-LUT-641.402545.336430.30873.09022.79153.1024

As shown, deeper models (e.g., Pan-LUT-4 to Pan-LUT-6) offer only marginal improvements in performance compared to shallower variants (Pan-LUT-1 to Pan-LUT-3), while significantly increasing parameter count. For example, Pan-LUT-4 already exceeds 2 million parameters, which limits its practicality. Therefore, we adopt Pan-LUT-1, 2, and 3 as the default architectural choices.

MetricsPan-LUT-1Pan-LUT-2Pan-LUT-3Pan-LUT-4Pan-LUT-5Pan-LUT-6
DlambdadownarrowD_{\\lambda} \\downarrow0.09560.09620.09780.09630.09870.0944
DSdownarrowD_{S} \\downarrow0.11450.12530.12760.12690.12120.1233
QNRuparrow\\uparrow0.80240.79850.79480.79340.79590.7986
Dataset: Full-WV-II

Although deeper architectures often risk overfitting, the consistent performance of Pan-LUT models on unseen datasets indicates that our method maintains strong generalization without overfitting the training data.

Question 2

We demonstrate the advantages of the Rotation Ensemble Strategy (RES) from both theoretical and empirical perspectives.

Reducing Memory Usage. The RES effectively expands the receptive field of the SDLUT module from 2times22 \\times 2 to 3times33 \\times 3. Without RES, achieving the same receptive field would require increasing the LUT dimensionality from 4 to 9, leading to a memory complexity of O(N9)O(N^9). In contrast, RES enables a larger receptive field without increasing LUT size, thereby reducing memory usage.

Local Texture Enhancement. Let (w,h)(w, h) denote the pixel located at the w-th column and h-th row of the image. We provide a theoretical proof that RES extends the receptive field from 2times22 \\times 2 to 3times33 \\times 3. Ideally, we aim to incorporate all pixels within the local 3times33 \\times 3 region centered at (w,h)(w, h), corresponding to the input index set:

Gtgt=(w1,h1),(w,h1),(w+1,h1),(w1,h),(w,h),(w+1,h),(w1,h+1),(w,h+1),(w+1,h+1).G_{tgt} = \\{(w{-}1,h{-}1), (w,h{-}1), (w{+}1,h{-}1), (w{-}1,h), (w,h), (w{+}1,h), (w{-}1,h{+}1), (w,h{+}1), (w{+}1,h{+}1)\\}.

By default, SDLUT captures the bottom-right neighborhood of each pixel, with the corresponding input index group defined as:

Gr0=(w,h),(w+1,h),(w,h+1),(w+1,h+1).G_{r0} = \\{(w,h), (w+1,h), (w,h+1), (w+1,h+1)\\}.

With the proposed RES, SDLUT is applied to three rotated versions of the input, thereby effectively capturing a broader set of pixel neighborhoods.

Gr90=(w,h),(w1,h),(w,h+1),(w1,h+1)G_{r90} = \\{(w,h), (w{-}1,h), (w,h{+}1), (w{-}1,h{+}1)\\},

Gr180=(w,h),(w1,h),(w,h1),(w1,h1)G_{r180} = \\{(w,h), (w{-}1,h), (w,h{-}1), (w{-}1,h{-}1)\\},

Gr270=(w,h),(w+1,h),(w,h1),(w+1,h1)G_{r270} = \\{(w,h), (w{+}1,h), (w,h{-}1), (w{+}1,h{-}1)\\}.

Then, we can derive the following equation:

Gr0cupGr90cupGr180cupGr270=Gtgt.G_{r0} \\cup G_{r90} \\cup G_{r180} \\cup G_{r270} = G_{tgt}.

MethodWV2GF2WV3
Pan-LUT-140.636243.655929.6213
Pan-LUT-1 (w/o RES)40.499343.214729.3354
Pan-LUT-240.841944.554529.7929
Pan-LUT-2 (w/o RES)40.724744.375429.6786
Pan-LUT-341.136244.855930.1213
Pan-LUT-3 (w/o RES)40.998844.667230.0455

Definition: w/o RES' means the framework without the Rotation Ensemble Strategy.

We find that RES not only contributes to performance improvement but also significantly reduces the number of parameters.

Question 3

The proposed quadrilinear and pentalinear interpolation methods are conceptually similar to linear and bilinear interpolation, serving as their extensions to higher dimensions. Below, we provide a theoretical analysis to illustrate their similarities.

Linear Interpolation. Given two known data points P0=(x0,y0)P_0=(x_0,y_0) and P1=(x1,y1)P_1=(x_1,y_1), we assume the function changes linearly between these two points. We then estimate the corresponding yy value for a target point xx based on its proportional position between x0x_0 and x1x_1. Specifically, if a point xx is between x0x_0 and x1x_1, and its distance to x0x_0 is tt times the total distance x1x0|x_1−x_0| (where 0t10≤t≤1), then x=x0+t(x1x0)x=x_0+t(x_1−x_0). Similarly, its corresponding yy value, yy, can be expressed as y=(1t)y0+ty1y=(1-t)y_0+ty_1.

Bilinear Interpolation. Suppose we are given function values ff at four corner points on a 2D plane: Q11=(x1,y1)Q_{11}=(x_1,y_1), Q12=(x1,y2)Q_{12}=(x_1,y_2), Q21=(x2,y1)Q_{21}=(x_2,y_1), and Q22=(x2,y2)Q_{22}=(x_2,y_2). We aim to estimate the function value f(x,y)f(x,y) for any point P=(x,y)P=(x,y) on the plane.

Step 1: Linear interpolation along the x-direction.

First, along the row where y1y_1 is constant, linearly interpolate the point (x,y1)(x, y_1). Using the values of Q11=(x1,y1)Q_{11}=(x_1, y_1) and Q21=(x2,y1)Q_{21}=(x_2, y_1), calculate the value for R1=(x,y1)R_1=(x, y_1):

f(R1)=fracx2xx2x1f(Q11)+fracxx1x2x1f(Q21)=dxf(Q11)+(1dx)f(Q21)f(R_1) = \\frac{x_2 - x}{x_2 - x_1} f(Q_{11}) + \\frac{x - x_1}{x_2 - x_1} f(Q_{21})=d_xf(Q_{11})+(1-d_x)f(Q_{21}),

where dx=fracx2xx2x1d_x = \\frac{x_2 - x}{x_2 - x_1} denotes the relative distance to the known points.

Then, along the row where y2y_2 is constant, linearly interpolate the point (x,y2)(x, y_2). Using the values of Q12=(x1,y2)Q_{12}=(x_1, y_2) and Q22=(x2,y2)Q_{22}=(x_2, y_2), calculate the value for R2=(x,y2)R_2=(x, y_2):

f(R2)=fracx2xx2x1f(Q12)+fracxx1x2x1f(Q22)=dxf(Q12)+(1dx)f(Q22).f(R_2) = \\frac{x_2 - x}{x_2 - x_1} f(Q_{12}) + \\frac{x - x_1}{x_2 - x_1} f(Q_{22})=d_xf(Q_{12})+(1-d_x)f(Q_{22}).

Step 2: Linear interpolation along the y-direction.

Finally, using the values of R1=(x,y1)R_1=(x, y_1) and R2=(x,y2)R_2=(x, y_2), linearly interpolate for the target point P=(x,y)P=(x, y):

f(P)=fracy2yy2y1f(R1)+fracyy1y2y1f(R2)=dxdyf(Q11)+(1dx)dyf(Q21)+dx(1dy)f(Q12)+(1dx)(1dy)f(Q22),f(P) = \\frac{y_2 - y}{y_2 - y_1} f(R_1) + \\frac{y - y_1}{y_2 - y_1} f(R_2)=d_xd_yf(Q_{11})+(1-d_x)d_yf(Q_{21})+d_x(1-d_y)f(Q_{12})+(1-d_x)(1-d_y)f(Q_{22}),

where dy=fracy2yy2y1d_y = \\frac{y_2 - y}{y_2 - y_1} denotes the relative distance to the known points.

Quadrilinear Interpolation in Look up table. All the interpolation methods above exhibit a structurally similar form in estimating the target value:

y=(1t)y0+ty1y=(1-t)y_0+ty_1,

f(P)=dxdyf(Q11)+(1dx)dyf(Q21)+dx(1dy)f(Q12)+(1dx)(1dy)f(Q22)f(P) = d_xd_yf(Q_{11})+(1-d_x)d_yf(Q_{21})+d_x(1-d_y)f(Q_{12})+(1-d_x)(1-d_y)f(Q_{22}).

We show that the estimation formula of quadrilinear interpolation follows the same functional form.Since quadrilinear interpolation involves 16 known points, the full expression is too lengthy to present here. For more details, please refer to the technical details section in the supplementary material.

评论

The author's careful explanation resolved my doubts about the technical details, and I decided to adjust the score I gave.

评论

Thank you for your response and feedback. We are glad that our explanation could resolve your questions. This is a great encouragement to us. Thank you again for your valuable time and professional review.

最终决定

This paper studies the problem of efficient pan-sharpening in remote sensing. By replacing complex DNN operations with learnable LUTs, Pan-LUT bridges the gap between computational efficiency and performance in high-resolution pan-sharpening.

All reviewers recognize the novelty of using a learnable LUT framework and its impact to remote sensing. Compared to previous approaches it enables processing of high-res images in practice, which is a valuable contribution to the field of remote sensing image processing.