PaperHub
3.8
/10
withdrawn4 位审稿人
最低1最高6标准差1.9
3
6
1
5
3.0
置信度
正确性2.3
贡献度2.0
表达2.0
ICLR 2025

Towards General Certified Robustness of Combinatorial Optimization Solvers

OpenReviewPDF
提交: 2024-09-27更新: 2024-11-26
TL;DR

We develop a mathematically sound method to certify the robustness of CO solvers and introduce a general robustness enhancement method based on a smoothing concept.

摘要

关键词
Certified Robustness; Combinatorial Optimization

评审与讨论

审稿意见
3

This paper describes an approach to certifying the robustness of parametrizable combinatorial optimization solvers with respect to certain kinds of perturbations of the solver's input. The perturbations are divided into two categories: binary, where a constraint is add/deleted, and tightness, where the constraint's bound is changed.

The paper states a definition of robustness certification for combinatorial optimization solvers, quantifying robustness in terms of the closeness of input instances. A randomized smoothing technique is used to robustify solvers. The output of these robust solvers is guaranteed to be close for any inputs that are close.

The paper presents a "robustness enhancement method" that works by searching for easy instances that are similar to the given instance.

Finally, experiments are performed demonstrating aspects of the framework.

优点

While previous works have focused on the empirical evaluation of robustness for combinatorial optimization solvers, this paper provides a theoretically sound robustness certification guarantee, which is helpful when, for example, a priori guarantees are required.

The paper also describes a way of using this guarantee to enhance the robustness of solvers using a randomized smoothing technique. This is arguable the main contribution of the paper, as it gives a way of producing useable, provably robust solvers.

The paper is well-placed within the existing literature: randomized smoothing applied to make robustness guarantees for combinatorial optimization solvers.

缺点

  • The definition of "certified radius" does not look like a definition to me. It is just an inequality. Is the certified radius the largest δ\delta that satisfies this inequality?

  • Line 186. "finding the optimal solution xx for a given problem instance QQ is often NP-hard". As stated, this is not correct. A solver that completely ignores its input and just always outputs xx returns the optimal solution for QQ in constant time, but of course this does not mean P=NPP = NP... The problem that is NP-hard is to find the optimal solution for any possible input.

  • Line 187: "indicating that no polynomial-time algorithm exists to solve it." Assuming PNPP \not= NP...

  • Equation 4 defines the CDF for the cost c(fθ(Q),Q)c(f_\theta(Q), Q). Where is the randomness in the cost coming from? I would have thought it was over the choice of inputs QQ, but in Theorem 4.1 the input QQ is taken as given and the cost still has a CDF? The source of the randomness needs to be made more clear.

  • Equation 5: Should vv be uu?

  • The matrix HbH_b. I wonder if the authors could describe more generally how this matrix is built from the set of constraints. It is stated that "1 signifies the presence of a constraint", since this set is inherently linear. In the DAG example, the constraints are between two jobs, so a matrix is implied, but I cannot see how this would be so in general.

  • Equation 6: What part of this is the definition? It seems like GHbG_{H_b} is being defined in terms of gθg_\theta and/or FHbzF_{H_b \oplus z}, but what are these?

  • Algorithm 1 and subsequent: FHb+z(j)F_{H_b + z^{(j)}} should be an \oplus?

  • Algorithm 1 is called "Randomized Smoothing..." but it outputs a number that is a distance. What has been smoothed? Shouldn't it output a smoothed/robust solver?

  • Robustness Enhancement Method (Algorithm 2 and Section 4.2): Why are we trying to find an easy instance? How does this enhance the robustness of the solver? This seems important and should be explained better.

  • Theorem 4.1: Is GHbG_{H_b} the smoothed CDF using a vector zz of Bernoulli(β\beta) RVs? Why are we interested in bounding the distance between this smoothed distribution and the perturbed distribution GHbδbG_{H_b \oplus \delta_b}? Don't we want to bound the distance with the actual cost distribution FF? The preamble defines the cost CDF FF, but then does not use it, so I suspect I am misunderstanding something...

  • Both Theorem 4.1 and 4.2 define F(r)F(r), but then never use it. Is this intentional?

  • Why are absolute values taken of CDFs? Shouldn't they be non-negative anyway?

  • Experimental details are pushed to the Appendix, in particular results for the enhancement method for certified robustness.

  • In general, I am not sure I would describe the experiments as "Extensive" (line 105)...

  • The experiments vary mm and β\beta. If I were to use this robustness certification method in practice, how would I choose values for these parameters?

  • Figure 1: It is not clear to me what the denominator of the ratios is. What is SFT (line 431) and why is this the right normalizing baseline?

Some minor issues not affecting my recommendation:

  • In text citation style is used where parenthesis citation style should be used. E.g., "Previous studies Varma & Yoshida (2021); Geisler et al. (2021); Lu et al. (2023) indicate..." -> "Previous studies (Varma & Yoshida, 2021; Geisler et al., 2021; Lu et al., 2023) indicate..."

  • Line 205 missing word? "two distinct perturbation types that can..."

  • Equation 5: I would use a naming convention that makes the connection between μ\mu and C1C_1, and between ν\nu and C2C_2, more clear.

  • Figure 1, y-axis label: radio -> ratio

问题

See Weaknesses above.

伦理问题详情

N/A

审稿意见
6

This paper studies the robustness of combinatorial optimization (CO) solvers. As claimed by the authors, previous work has not explored the certified robustness of CO solvers. In this work, they propose a definition of robustness certification for CO solvers, along with a method to improve both certified and empirical robustness. The authors conduct experiments to validate the effectiveness of their certification approach and enhancement method. They claim that their proposed framework addresses gaps in the existing literature regarding the certified robustness of combinatorial optimization solvers.

优点

  1. The paper is generally well-written, with clear explanations.

  2. This paper focus on framework studying, which is a foundational work in the robustness of combinatorial optimization solvers.

  3. Experiments are conducted on the representative problems.

缺点

  1. In Section 5.1, the solver selection criteria for the two representative problems are not consistent. The solver for DAG scheduling does not include SOTA solvers or learning methods.

  2. In the work of ROCO, an evaluation was conducted on the MC problem. Why this problem was not selected in this paper?

问题

Please reply to my comments in "Weaknesses".

伦理问题详情

Not applicable.

审稿意见
1

The paper seeks to assess the robustness of combinatorial optimization solvers. While this is a valid direction, the paper fails to articulate an appropriate theoretical framework, and exhibits several limitations in its presentation, methodology, and experiments.

优点

There has been a lot of interest recently in providing formal guarantees on the performance of trained ML models. Investigating such a direction in the context of combinatorial optimization is relevant, however the paper suffers from multiple limitations, outlined below.

缺点

It is hard to understand the logic behind the paper, and several notations / symbols / problems are used without being clearly defined. In addition to these limitations, the paper lacks a principled methodological framework. Finally, numerical experiments are incompletely stated and only small instances are considered.

Overall paper

  • How robustness of classifier relevant to robustness of combinatorial optimization setting?
  • The paper discusses the robustness of combinatorial solvers, but makes no mention of the (vast) literature on robust optimization
  • Throughout the paper, multiple concepts/notations are ill-defined or not defined at all. This includes, for instance:
    • the "strictness" of a constraint
    • the underlying randomness & probability distribution considered in Eq (4)
    • variable vv in Eq (5)
    • Matrix HbH_{b} used throughout the paper
    • gθg_{\theta} in Eq. (6)

Methodology

The paper suffers from several fundamental methodological limitations, the first of which is the absence of a clear, relevant definition of "robustness" for combinatorial optimization solvers. This severe limitation makes it impossible to follow the rest of the paper, and to evaluate the paper's contribution and relevance to the field.

  • The paper lacks a principled definition of robustness for CO solvers. Authors are invited to review the paper Compact Optimality Verification for Optimization Proxies and the references therein for relevant literature on performance verification for ML models that output solutions to optimization problems.
  • The paper states that "hard" instances have higher optimal value, which is not correct. For instance, in TSP, multiplying all the distances by γ>1\gamma > 1 increases the optimal value by the same factor γ\gamma, but the optimization problem is equivalent.
  • The function CDF F(r) in Eq. (4) seems to be defined w.r.t a probability distribution over Q... which is never defined.
  • In Theorem 4.1, the right-hand-side integral (used as upper bound on a Wasserstein distance) may be infinite. Note that this term is the integral of a CDF function, which makes it likely that it is not finite.

Experiments

The numerical experiments are hard to understand. The paper does not state the mathematical formulation of either problem, nor how instances are generated. This makes it impossible to evaluate the soundedness of the experiments and the validity of the results.

TSP instances

  • TSP instances with 100 cities are very small compared to state-of-the-art. Instances of that size can be solved exactly to global optimality on a phone using the free Concorde app. The authors are encouraged to consider (much larger) instances from the TSPLib, aiming for at least tens of thousands of cities for hard-to-solve instances.
  • It is extremely hard to understand how TSP instances are generated, and what the parameters introduced in the paper translate to in the context of TSP instances.

问题

  • The paper should explain how it differs from the (large) literature on robust optimization
  • Please provide a valid, self-contained definition of robustness for combinatorial optimization solvers. The change in objective value from instance to instance is not an appropriate metric.
  • Numerical experiments should include a clear mathematical formulation of the problem at hand, as well as a complete description of the data distribution used to generate instances.
审稿意见
5

The paper presents an approach to addressing robustness in combinatorial optimization solvers by introducing a randomized smoothing framework for certified robustness. The theoretical results are supported by experiments on DAG scheduling and TSP tasks and lend credibility to the proposed method. However, several assumptions in the robustness definitions, such as the strict reliance on CDF stability and the Wasserstein distance as robustness metrics, could limit the model's applicability to broader CO contexts.

优点

  • Novelty in the exploration of certified robustness within CO solvers.
  • Adaptation of randomized smoothing to combinatorial problems (this is nice as CO have constraints and non-categorical outputs).
  • Theoretically grounded method that could apply to several optimization tasks in dynamic environments.
  • The enhancement method for leveraging "easy instances" shows a promising direction for robustness without retraining.

缺点

  • The paper’s approach primarily targets DAG scheduling and TSP tasks, but it is unclear how generalizable the method is to structurally richer CO problems.
  • The use of Wasserstein distance as the primary robustness metric may be too restrictive, especially in CO tasks where cost functions or constraints may differ significantly.
  • The empirical robustness results, while promising, rely on specific perturbation strategies (e.g., random search, simulated annealing). A more comprehensive assessment involving adaptive or adversarial perturbation techniques would provide a more robust evaluation here.
  • Some sections of the theoretical framework (particularly the robustness guarantees) are challenging to follow. Consider to at least paraphrase the results in English or adding some illustrative examples.

问题

  1. Does the model handle adversarial perturbations, or is it only applicable to random perturbations?

  2. How does the proposed enhancement method scale with increased constraint complexity in CO tasks e.g., as the constraint structures and dependencies intensify?

  3. Could the robustness guarantees extend to CO problems with multi-objective cost functions?

撤稿通知

I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors.