PaperHub
7.2
/10
Poster4 位审稿人
最低3最高5标准差0.8
3
5
3
4
ICML 2025

RUN: Reversible Unfolding Network for Concealed Object Segmentation

OpenReviewPDF
提交: 2025-01-14更新: 2025-07-24
TL;DR

The first unfolding-based concealed object segmentation algorithms.

摘要

关键词
Concealed object segmentationDeep unfolding networkReversible modeling strategyCamouflaged object detection

评审与讨论

审稿意见
3

This paper tackles the challenging Concealed Object Segmentation, which aims to segment objects that are visually blended with their surroundings. It introduces the Reversible Unfolding Network (RUN), a iteratively method to refine segmentation results and minimize uncertainties. RUN includes two main modules: the Segmentation-Oriented Foreground Separation (SOFS) module, which captures non-local context and applies the reversible strategy at the mask level, and the Reconstruction-Oriented Background Extraction (ROBE) module, which addresses conflicting foreground and background regions. The method was evaluated on multiple datasets to valid its effectiveness.

给作者的问题

Please refer to the comments in the above reviews.

论据与证据

The paper claims that previous works use reversible strategies but are generally limited to the mask domain. In contrast, this work introduces RUN, which extends reversible strategies to both the mask and RGB domains.

In P3L142, the paper states that integrating model-based and learning-based approaches remains underexplored due to the lack of intrinsic models. The paper should provide discussion to clarify the specific challenges involved and explain why addressing this gap is important.

方法与评估标准

The methods is evaluated on multiple benchmarks using the standard metrics.

According to Fig. 3, the architecture appears somewhat complex. The paper should provide a comparison of FPS and parameter count to better demonstrate the efficiency of this method.

理论论述

N/A

实验设计与分析

The methods is evaluated on multiple benchmarks using the standard metrics.

Although previous works have not fully explored the RGB domain for mask refinement, this method appears to underperform compared to BiRefNet [1] and ZoomNeXt [2]. For instance, the S-measure on CAMO (higher is better) is 0.806 for this work, whereas BiRefNet achieves 0.904. Additionally, the paper does not provide the comparison results with these methods.

The comparisons are insufficient. Please include a comparison with previous segmentation refinement methods, such as [3].

[1] Bilateral Reference for High-Resolution Dichotomous Image Segmentation, CAAI AIR, 2024.

[2] ZoomNeXt: A Unified Collaborative Pyramid Network for Camouflaged Object Detection, TPAMI, 2024.

[3] SegRefiner: Towards Model-Agnostic Segmentation Refinement with Discrete Diffusion Process, NeurIPS, 2023.

补充材料

This work does not include any supplementary material.

与现有文献的关系

This work proposes an iterative method to refine segmentation results, which may benefit related segmentation tasks.

遗漏的重要参考文献

[1] Bilateral Reference for High-Resolution Dichotomous Image Segmentation, CAAI AIR, 2024.

[2] ZoomNeXt: A Unified Collaborative Pyramid Network for Camouflaged Object Detection, TPAMI, 2024.

[3] SegRefiner: Towards Model-Agnostic Segmentation Refinement with Discrete Diffusion Process, NeurIPS, 2023.

其他优缺点

N/A

其他意见或建议

N/A

作者回复

Thanks for the valuable comments.

W1. Challenges in introducing DUNs and importance of modeling high-level vision tasks

In P3L142, we state that deep unfolding networks (DUNs) are underexplored in high-level vision tasks for lacking intrinsic models.

Challenges: DUNs rely on model-based optimization, which is effective in low-level vision tasks due to well-defined physical models (e.g., Retinex model for low-light enhancement, Atmospheric Scattering model for dehazing). However, high-level vision tasks lack an explicit physical model, limiting DUN applications.

Importance:

  • Explicit constraints: Modeling high-level vision tasks enables task-specific constraints to guide optimization explicitly. For example, Eq. (4) introduces a residual sparsity constraint to refine segmentation and reduce uncertainties (P3L130, right part). This is more direct than implicit backpropagation. Table 5 shows a performance drop when removing this constraint (w/o prior M^k\hat{\mathbf{M}}_k).
  • Introducing DUNs into high-level vision: With an explicit model, DUNs integrate optimization with deep networks, balancing interpretability and generalizability (P2L58, right part). RUN benefits in effectiveness (Table 1), efficiency (see W2), and degradation resistance (Fig. 5) in COS.

W2. Efficiency analysis

The designs of M^()\hat{\mathcal{M}}(\cdot) and B^()\hat{\mathcal{B}}(\cdot) follow mathematical principles and may appear complex, but we only make fixed hyperparameters learnable and enforce weight sharing for SOFS and ROBE across stages. In the table, our method is more efficient than SOTAs.

Param. (M)/FLOPS (G)/FPS
ResNet50FocusDiff166.18/7618.49/0.23
RUN30.41/43.36/22.75
Res2Net50FEDER45.92/50.03/14.02
RUN30.57/45.73/20.26
PVT V2CamoFocus68.85/91.35/9.63
RUN65.17/61.83/15.82

W3. Comparison with more methods

BiRefNet [1] targets high-resolution segmentation (HRS); ZoomNeXt [2] focuses on multiple input scales (MIS), differing from our setting.

Higher resolution and multiscale inputs enhance subtle discriminative cues but add computational overhead. We also report our results in these settings.

For HRS, we retrain our method following BiRefNet [4], replacing our encoder with SwinL and training at 1024 x 1024 resolution. We evaluate our results on the three datasets reported in BiRefNet and find that RUN outperforms BiRefNet in most metrics. This is primarily due to our reversible modeling strategy.

HRS (MM/FβF_\beta/EϕE_\phi/SαS_\alpha)CAMOCOD10KNC4K
BiRefNet0.030/0.895/0.954/0.9040.014/0.881/0.960/0.9130.023/0.925/0.953/0.914
RUN0.032/0.898/0.951/0.9100.012/0.878/0.967/0.9200.022/0.946/0.961/0.927

For MIS, we compare with ZoomNeXt (352 × 352). In the table, RUN outperforms ZoomNeXt in 8 out of 12 metrics. Unlike ZoomNeXt’s specialized extraction-fusion strategy, RUN simply concatenates multiscale features. Adding ZoomNeXt’s strategy to RUN yields RUN++, improving results.

MIS (MM/FβF_\beta/EϕE_\phi/SαS_\alpha)CHAMELEONCAMOCOD10KNC4K
ZoomNext0.020/0.872/0.963/0.9120.069/0.782/0.883/0.8220.026/0.765/0.918/0.8550.038/0.827/0.919/0.869
RUN0.022/0.878/0.967/0.9110.064/0.807/0.902/0.8320.027/0.772/0.920/0.8430.040/0.836/0.922/0.868
RUN++0.019/0.885/0.971/0.9160.063/0.815/0.906/0.8330.025/0.781/0.930/0.8520.038/0.847/0.932/0.875

W4. Performance in segmentation refinement

RUN, an end-to-end network for COS, can also serve as a segmentation refiner by initializing M0\mathbf{M}_0 as the coarse mask.

We compare with several segmentation refiners: traditional methods (dense crf (DCRF) [3] and biliteral solver (BS) [4]) and learning-based methods (SegRefiner [5] and SAMRefiner [6]). To ensure fairness, we retrain the learning-based ones for COS, SegRefiner+ and SAMRefiner+, following their training rules.

Traditional methods fail to refine FEDER’s segmentation in COD10K due to the challenges of COS (objects blending with surroundings). SegRefiner and SAMRefiner also perform suboptimally with their provided models, as coarse COS masks contain segmentation errors and uncertain regions, complicating refinement.

Retraining improves SegRefiner+ and SAMRefiner+, with SAMRefiner+ getting results comparable to RUN. However, RUN (FPS: 22.75) outperforms SAMRefiner (FPS: 0.92) and SegRefiner (FPS: 0.89) in efficiency.

Refinement results for FGANet’s masks are discussed in W3 (reviewer 45Vr), with similar conclusions.

FEDER+DCRF+BS+SegRefiner+SegRefiner++SAMRefiner+SAMRefiner+FEDER-R (Ours)
MM0.0320.0390.0430.0370.0330.0380.0330.031
FβF_\beta0.7150.6830.6660.6910.7180.6860.7230.721
EϕE_\phi0.8920.8530.8620.8630.8890.8570.9060.897
SαS_\alpha0.8100.7970.7700.7810.8030.7830.8050.812

[1] BiRefNet, CAAI AIR24

[2] ZoomNext, TPAMI24

[3] Dense CRF, NIPS21

[4] Bilateral Solver, ECCV16

[5] SegRefiner, NIPS23

[6] SAMRefiner, ICLR25

审稿意见
5

This paper proposes the first deep unfolding network, RUN, for the COS task, aiming to cope with one of the intrinsic challenges in the COS task, which is neglecting the importance of applying reversible strategies in the RGB domain. To achieve this goal, two reversible modules, SOFS and ROBE, are further proposed. By treating the foreground-background separation problem as the noise elimination problem, ROBE serves as a valuable assistant for SOFS in promoting accurate segmentation. Abundant experiments verify the potential of this method.

给作者的问题

See weaknesses.

论据与证据

Yes

方法与评估标准

Yes

理论论述

The theoretical analysis in this paper is accurate.

实验设计与分析

The abundant experiments are both convincing and well-discussed.

补充材料

N/A

与现有文献的关系

This is the first application of deep unfolding network, a widely-used strategy in low-level vision tasks that is favored in achieving a balance in interpretability and generalizability, in high-level vision tasks. This brings positive effects to the whole high-level vision field.

遗漏的重要参考文献

N/A

其他优缺点

Strengths: The proposed RUN framework is not only a novel DUN-based application for the high-level vision task but also allows for the combination of high-level and low-level vision, ensuring robust performance in high-level vision tasks even in degraded scenarios.

Weaknesses: 1 The related work is very short. Please consider providing a more detailed description in the revised version. 2 Why use a very simple reconstruction network in ROBE, given its reconstruction capacity is highly limited in this condition? 3 Also, the reviewer wants to know if changing the simple network in ROBE to a more complex framework brings an evident performance gain. 4 The reviewer is curious about if the single SOFS module achieves comparable performance with existing methods. 5 More experiments are preferred to analyze the performance in complex degradation. The reviewer thinks this is a major challenge in COS. Please fully prove whether RUN can address this challenge and make comparison with existing strategies, such as bi-level optimization.

其他意见或建议

I think this paper is very interesting. Please consider addressing the weaknesses. I will consider modifying the score based on the response.

作者回复

Thanks for the valuable comments.

W1. More related works

In related works, the COS component focuses on the development of the reversible modeling strategy, while our coverage of DUN is limited due to the underexploration of this technique in high-level vision tasks. To provide a more comprehensive overview, we will add a section that systematically reviews the advancements in the COS field.

W2&W3. Why use a simple network in ROBE? How about replacing it with a more complex one?

(1) The reconstruction task in ROBE is relatively easy, compared to most low-level vision tasks, as the input images are of high quality. Thus, we select a lightweight network to balance performance and computational efficiency. Experimental results validate the effectiveness of this design choice.

Parameters (M)FLOPS (G)FPS
ResNet50FocusDiff166.187618.490.23
FSEL29.1551.032.19
RUN30.4143.3622.75
Res2Net50FEDER45.9250.0314.02
RUN30.5745.7320.26
PVT V2CamoFocus68.8591.359.63
RUN65.1761.8315.82

(2) As shown in Table 5, replacing the simple network with more complex networks brings limited performance gains. However, the results in Fig. 5 indicate that a more complex network can better resist degradation conditions.

(3) Furthermore, as shown in the table, increasing the stage number can also improve the network’s degradation resistance capacity. These findings can help us further exploit the potential of this DUN-based framework.

FβF_\beta00.10.20.30.40.5
RUN-4 stages0.7470.7400.7300.7180.6980.676
RUN-6 stages0.7490.7420.7330.7230.7080.694
RUN-8 stages0.7510.7430.7350.7260.7140.701
EϕE_\phi00.10.20.30.40.5
RUN-4 stages0.9030.8980.8920.8850.8760.865
RUN-6 stages0.9050.8980.8930.8860.8780.869
RUN-8 stages0.9050.9000.8960.8910.8850.877

W4. Performance of SOFS

As shown in the table, removing ROBE leads to a noticeable performance decline. The resulting model underperforms compared to the state-of-the-art method, FocusDiff, further highlighting the effect of our approach. This can also be verified by Table 11, where replacing SOFS with the core structures of existing methods results in clear performance improvements.

w/o ROBERUN (Ours)
MM0.0320.030
FβF_\beta0.7130.747
EϕE_\phi0.8920.903
SαS_\alpha0.8160.827

W5. Compare with bi-level optimization on more degradation scenes

(1) As shown in Fig. 5, replacing the original simple network with a more complex dehazing network leads to RUN+, which improves robustness against haze degradation. To further evaluate our approach, we test its performance within a bi-level optimization (BLO) framework following the setup of HQG-Net [1], where ROBE and SOFS are treated as low-level and high-level vision blocks, as illustrated in Fig. S2. Evaluating performance on COD10K using FβF_\beta and EϕE_\phi, we find that RUN+ consistently outperforms the BLO framework across different haze concentrations.

FβF_\beta0.10.20.30.40.5
BLO0.7410.7320.7200.7050.697
RUN+0.7430.7350.7260.7180.708
EϕE_\phi0.10.20.30.40.5
BLO0.8970.8900.8830.8720.860
RUN+0.8990.8940.8890.8830.876

(2) Additionally, we further analyze the performance in two degradation scenarios that are extremely challenging for the COS task, i.e., the low-light and low-resolution scenarios.

Low-light scenarios deepen the concealment of objects by lowering color contrast, while low-resolution scenes directly reduce the discriminative cues by reducing the number of valid pixels.

As with the above content, we also simulate degradation on concealed images in COD10K.

For the low-light scenarios, we applied image darkening [2] and degraded the dataset into three levels: slight low-light, medium low-light, and severe low-light. We employ Reti-Diff [3] as the reconstruction network in the ROBE module. Experiments verify that our RUN framework achieves better performance than the BLO version.

FβF_\beta / EϕE_\phislightmediumsevere
BLO0.719/0.8700.691/0.8530.662/0.813
RUN+0.735/0.8890.716/0.8780.683/0.855

For the low-resolution scenes, we use bicubic downsampling and select x2, x4, x8 for reconstruction and segmentation. DiffIR [4] is selected as the reconstruction network. As shown in the table, RUN still outperforms BLO in the low-resolution challenges.

FβF_\beta/EϕE_\phix2x4x8
BLO0.706/0.8650.662/0.8470.580/0.786
RUN+0.729/0.8810.707/0.8660.625/0.820

To sum up, our RUN framework is not only a novel unfolding-based application for high-level vision tasks but also effectively integrates high-level and low-level vision, ensuring robust performance even in degraded scenarios.

[1] HQGNet, TNNLS23

[2] StableLLVE, CVPR21

[3] Reti-Diff, ICLR25

[4] DiffIR, ICCV23

审稿人评论

The authors have resolved all my concerns, especially that the proposed RUN effectively bridges low-level and high-level tasks, ensuring robustness in degradation scenarios.

It is indeed a strong piece of work. Given the inspired contribution and the comments from the other three reviewers, I am inclined to raise my score and recommend strong acceptance.

作者评论

We sincerely appreciate the reviewer’s recognition of the significance of our contribution. The RUN framework is not only a novel unfolding-based approach but also effectively integrates high- and low-level vision, ensuring robustness even in degraded scenarios.

Your acknowledgment is highly valuable to us and reinforces our commitment to advancing research in this field!

审稿意见
3

The authors propose Reversible Unfolding Network (RUN) for Concealed Object Segmentation. RUN integrates optimization-based solutions with deep learning, enabling reversible modeling across both mask and RGB domains. It also introduces the Segmentation-Oriented Foreground Separation (SOFS) module and the Reconstruction-Oriented Background Extraction (ROBE) module. Extensive experiments validate its superiority.

给作者的问题

N/A

论据与证据

N/A

方法与评估标准

N/A

理论论述

N/A

实验设计与分析

N/A

补充材料

N/A

与现有文献的关系

N/A

遗漏的重要参考文献

N/A

其他优缺点

Strengths

  • RUN develops the first deep unfolding network for COS, revealing the potential of deep unfolding network in COS.
  • RUN introduces SOFS and ROBE modules to direct attention to uncertain regions through iterative optimization
  • RUN achieves excellent performances across five different tasks against other task-specific methods.
  • RUN can integrate with existing methods to further boost performance, showcasing its flexibility.

Weaknesses

  • This paper does not have any discussion of computational efficiency. It can be found that the performance is worse than most SOTA methods when the stage number K is small by comparing the results in Table 7 and Table 1. However, a larger K requires longer inference time. The authors should compare the computational cost between RUN and other methods.
  • Ablation studies are unclear. What is the baseline for this method? How does the model perform when the proposed SOFS and ROBE are removed?
  • The proposed method can be regarded as a post-processing method. Therefore, it should be compared with common post-processing methods in segmentation, such as dense crf and bilateral solver.

其他意见或建议

N/A

作者回复

Thanks for the valuable comments.

W1. Efficiency analysis

We compare the parameters, FLOPS, and FPS between our RUN against cutting-edge methods on three backbones. Our stage number is set to 4. As shown in the table, our method is more efficient across all three backbones with the input size of 352×352.

Parameters (M)FLOPS (G)FPS
ResNet50FocusDiff166.187618.490.23
FSEL29.1551.032.19
RUN30.4143.3622.75
Res2Net50FEDER45.9250.0314.02
RUN30.5745.7320.26
PVT V2CamoFocus68.8591.359.63
RUN65.1761.8315.82

Besides, we analyze the computational cost of our RUN framework with varying stage numbers, along with their performance using ResNet50 as the backbone on COD10K. Since the SOFS and ROBE modules at different stages share the same weights, all versions of RUN maintain a consistent parameter count. As shown in the table, our method outperforms existing state-of-the-art methods once the stage number reaches 3. For an optimal balance between efficiency and performance, we set the stage number to 4.

FLOPS (G)FPSMMFβF_\betaEϕE_\phiSαS_\alpha
RUN-1 stage29.9528.610.0330.7150.8850.803
RUN-2 stages34.1225.270.0310.7270.8930.812
RUN-3 stages38.7323.630.0300.7350.8980.822
RUN-4 stages43.3622.750.0300.7470.9030.827
RUN-6 stages52.6519.860.0300.7490.9050.826
RUN-8 stages62.0216.530.0300.7510.9050.830

W2. Ablation study: baseline and more results

As shown in Table 5, we conduct ablation studies on the ResNet50-based RUN framework.

To ensure the integrity of our reversible modeling strategy, we retain both SOFS and ROBE rather than performing breakdown ablations. Instead, in Table 5, we evaluate the contribution of each content within SOFS and ROBE by either replacing them with alternative designs or directly removing them, both of which lead to performance decreases.

Here we report the results of removing SOFS and ROBE. To remove the segmentation module SOFS, we add an extra segmentation head, i.e., the convolution blocks, after ROBE, with the segmentation loss unchanged. For the case w/o ROBE, we simply remove ROBE and keep SOFS. Besides, we include a baseline where both SOFS and ROBE are removed, leaving only the segmentation head in each stage. As shown in the table, adding SOFS and ROBE brings a clear performance gain.

Baselinew/o SOFSw/o ROBERUN (Ours)
MM0.0530.0380.0320.030
FβF_\beta0.5520.6830.7130.747
EϕE_\phi0.7180.8550.8920.903
SαS_\alpha0.6820.7970.8160.827

W3. Comparison with segmentation refinement methods

Our RUN is an end-to-end network designed for COS. For its structural characteristics, it can also be applied for refining coarse masks by initializing M0\mathbf{M}_0 as the coarse mask.

We compare our method with segmentation refinement methods, including traditional methods (dense crf (DCRF) [1] and biliteral solver (BS) [2]) and learning-based methods (SegRefiner [3] and SAMRefiner [4]). We use segmentation results from FEDER and FGANet as coarse masks. For a fair comparison, we also provide retrained versions of the learning-based refiners for the COS task, SegRefiner+ and SAMRefiner+, following their original training rules.

As shown in the table, traditional methods fail to improve the segmentation results of FEDER and FGANet. SegRefiner and SAMRefiner also achieve suboptimal performance with their provided models.

This is because concealed object segmentation is an inherently difficult task, where objects are visually blended with their surroundings. Hence, unlike common segmentation tasks, this task provides far fewer discriminative cues for feature extraction. Besides, the coarse segmentation masks often suffer from significant quality issues, such as mis-segmentation, blurred edges, and high uncertainty in certain pixel regions, making mask refinement particularly challenging.

After retraining, the learning-based refiners show noticeable performance gains, with SAMRefiner+ achieving results comparable to our RUN. However, our RUN framework achieves an FPS of 22.75, outperforming SAMRefiner with only 0.92 FPS.

FEDER+DCRF+BS+SegRefiner+SegRefiner++SAMRefiner+SAMRefiner+FEDER-R (Ours)
MM0.0320.0390.0430.0370.0330.0380.0330.031
FβF_\beta0.7150.6830.6660.6910.7180.6860.7230.721
EϕE_\phi0.8920.8530.8620.8630.8890.8570.9060.897
SαS_\alpha0.8100.7970.7700.7810.8030.7830.8050.812
FGANet+DCRF+BS+SegRefiner+SegRefiner++SAMRefiner+SAMRefiner+FGANet-R (Ours)
MM0.0320.0410.0420.0430.0340.0370.0340.032
FβF_\beta0.7080.6650.6490.6620.6980.6820.7250.716
EϕE_\phi0.8940.8460.8280.8470.8870.8660.9000.897
SαS_\alpha0.8030.7720.7640.7650.7870.7740.8030.805

[1] Dense CRF, NeurIPS21

[2] The fast bilateral solver, ECCV16

[3] SegRefiner, NeurIPS23

[4] SAMRefiner, ICLR25

审稿人评论

The rebuttal responded to my questions accordingly, and I decide to raise the final rating to weak accept.

作者评论

We sincerely appreciate the reviewer’s recognition of the significance of our contribution. The RUN framework is not only a novel unfolding-based approach but also effectively integrates high- and low-level vision, ensuring robustness even in degraded scenarios.

Your acknowledgment is highly valuable to us and reinforces our commitment to advancing research in this field!

审稿意见
4

The paper introduces a Reversible Unfolding Network (RUN), a novel deep unfolding network that integrates object segmentation and distortion restoration tasks. RUN combines a Segmentation-Oriented Foreground Separation (SOFS) module and a Reconstruction-Oriented Background Extraction (ROBE) module to achieve more accurate segmentation by reducing uncertainties. The framework allows reversible modeling of foreground and background in both the mask and RGB domains, focusing on uncertain regions to improve accuracy. Experimental results across various COS tasks demonstrate that RUN outperforms existing methods and provides flexibility for integration with other models.

给作者的问题

  1. What is the processing speed of the propose method compared to other state of the art methods? What is the trade-off and what is the best scenario for application.

论据与证据

This paper claims to be the first deep unfolding network for Concealed Object Segmentation (COS) problem. Based on the literature review of the past research works done for COS, this claim is clear and convincing. Other than this claim, this paper is mainly positioned as an application paper, which optimize the objective function with new parameters and new architecture designs to tackle COS tasks. Therefore, the claim is specified for solving the particular COS problems among the given benchmarking datasets. Based on the experiments in the main paper and the supplementary materials, the proposed optimization techniques and networks are well supported.

方法与评估标准

The theoretical part of the proposed method is inspired by the observation that segmentation mask gradients should be sparse. It introduces an extra term in the objective function to enforce the certainty. The optimization of the objective function is achieved by implementing corresponding network and train the network based on the objective function loss. The methods are evaluated in 4 widely used benchmarking datasets : CHAMELEON, CAMO, COD10K, and NC4K. The baseline methods are selected from the recent state of the art methods in this task. The proposed method consistently outperforms the baseline methods.

理论论述

I checked and verified the correctness of the equations from Eq. (4) to Eq. (17) for theoretical derivation of the proposed objective function.

实验设计与分析

The experiments have been conducted to compare the proposed method with a series of state of the art methods on the aforementioned 4 datasets. The evaluation metrics are selected properly. For the ablation study, I can see only COD10K dataset is selected, is there any particular reason for not running it over all the datasets?

补充材料

I have reviewed the entire supplementary material and the provided anonymous code.

与现有文献的关系

The proposed method has contribution to the restricted area of image segmentation. The proposed optimization techniques can be applied to a broader scientific literature where the objective function has similar content in nature. The sparsity of gradients can be found in many fields including optical flow, denoising, etc.

遗漏的重要参考文献

N.A.

其他优缺点

Strengths

  1. This paper is well written with thorough theoretical derivation, implementation details and experiment verification.
  2. The supplementary material has provided extra complement information. The code is available and valid for generating the proper results claimed in the paper.

其他意见或建议

The authors are suggested to address the weakness mentioned above in the rebuttal period.

作者回复

Thanks for the valuable comments.

W1. Why only conducting ablation studies on COD10K

Given that COD10K is a representative and high-quality dataset, we follow existing methods [1,2,3] to conduct ablation studies on COD10K. We have now included the results of ablation studies on three extra datasets. As shown in the table, the outcomes remain consistent with our initial conclusions. We will add this content to the revised manuscript.

CE(C)\mathbf{C}\rightarrow E(\mathbf{C})w/o RSSw/o VSSw/o prior M^k\hat{\mathbf{M}}_kw/o prior Bk1\mathbf{B}_{k-1}w/oEk\mathbf{E}_kB1()B()\mathcal{B}_1(\cdot)\rightarrow\mathcal{B}(\cdot)B2()B()\mathcal{B}_2(\cdot)\rightarrow\mathcal{B}(\cdot)w/o C^k\hat{\mathbf{C}}_kFixed\rightarrowLearnableRUN (Ours)
CHAMELEONMM0.0530.0300.0280.0290.0270.0280.0270.0260.0290.0290.027
FβF_\beta0.7640.8370.8420.8320.8450.8490.8590.8610.8450.8370.855
EϕE_\phi0.8850.9380.9430.9360.9450.9480.9530.9560.9440.9400.952
SαS_\alpha0.8020.8830.8890.8860.8890.8910.8930.9000.8930.8890.895
CAMOMM0.0970.0750.0710.0720.0700.0710.0690.0680.0720.0720.070
FβF_\beta0.6930.7520.7730.7660.7750.7760.7830.7800.7700.7640.781
EϕE_\phi0.7880.8360.8580.8470.8630.8610.8700.8660.8520.8450.868
SαS_\alpha0.7230.7850.7990.7900.7960.8010.8050.8080.8020.7980.806
NC4KMM0.0720.0490.0430.0460.0430.0430.0430.0420.0450.0460.042
FβF_\beta0.7350.8010.8190.8000.8200.8170.8260.8250.8160.8070.824
EϕE_\phi0.8230.8890.9030.8980.9000.9020.9040.9070.9020.8970.908
SαS_\alpha0.7920.8350.8450.8380.8490.8460.8530.8520.8450.8420.851

W2. Efficiency, trade-off, and the best scenario of RUN

(1) Efficiency: We compare parameters, FLOPS, and FPS in three backbones and find that our RUN surpasses existing cutting-edge methods across all three backbones.

Parameters (M)FLOPS (G)FPS
ResNet50FocusDiff166.187618.490.23
RUN30.4143.3622.75
Res2Net50FEDER45.9250.0314.02
RUN30.5745.7320.26
PVT V2CamoFocus68.8591.359.63
RUN65.1761.8315.82

(2) Structure-level trade-off: Although different stages in RUN share the same weights, the stage number still affects the overall computational cost. To analyze the trade-off, we investigate the impact of the stage number on our ResNet50-based framework. In the table, our method surpasses existing methods when the stage number reaches 3. For an optimal balance, we set the stage number to 4.

FLOPS (G)FPSMMFβF_\betaEϕE_\phiSαS_\alpha
RUN-2 stages34.1225.270.0310.7270.8930.812
RUN-3 stages38.7323.630.0300.7350.8980.822
RUN-4 stages43.3622.750.0300.7470.9030.827
RUN-6 stages52.6519.860.0300.7490.9050.826

(3) Framework-level trade-off: RUN achieves cutting-edge performance for the theoretical coupling of the segmentation module (SOFS) and the reconstruction module (ROBE). As shown in the table, ROBE brings an evident gain with limited costs. This can also be verified by Table 11, where replacing SOFS with the core structures of existing methods brings clear profits. Ablations in Table 5 indicate that replacing the simple reconstruction network with more complex ones yields marginal gains. Hence, we opt for a simple network here.

w/o ROBERUN (Ours)
MM0.0320.030
FβF_\beta0.7130.747
EϕE_\phi0.8920.903
SαS_\alpha0.8160.827
Parameters (M)26.9730.41
FLOPS (G)37.6443.36

(4) The best application scenarios: RUN is not only a novel unfolding-based application for COS in segmenting clean concealed objects, but also has the potential to ensure robust performance in degraded scenes. As shown in Fig. 5, when replacing the reconstruction network with a more complex one, we get RUN+ and observe better hazy resistance capacity.

To further verify this, we compare with bi-level optimization (BLO), a degradation-resistant framework. The differences between the two frameworks can be seen in Fig. S2, where our framework is superior in the theoretical coupling of the two models. We retrain our network (RUN+) in the BLO framework and report the results on COD10K with FβF_\beta and EϕE_\phi. We find that our RUN+ surpasses BLO across different hazy concentrations.

FβF_\beta/EϕE_\phi0.10.20.30.40.5
BLO0.741/0.8970.732/0.8900.720/0.8830.705/0.8720.697/0.860
RUN+0.743/0.8990.735/0.8940.726/0.8890.718/0.8830.708/0.876

Besides, we analyze performance in two other degradation scenes, i.e., the low-light and low-resolution scenes, and observe that our RUN framework consistently surpasses BLO. For space limitations, we put the results in the response of W5 of the reviewer MZkc. The reviewer can refer to the tables for more details.

[1] PFNet, CVPR21

[2] ZoomNet, CVPR22

[3] FEDER, CVPR23

最终决定

After careful consideration of the reviewers' comments and the authors' rebuttal, I believe that the paper makes a meaningful contribution to the field of Concealed Object Segmentation. The novelty of the deep unfolding network approach, combined with the strong experimental results and robustness in degraded scenarios, justifies acceptance. While the framework's complexity and lack of baseline comparison are valid concerns, they do not outweigh the paper's strengths.