PaperHub
6.3
/10
Spotlight3 位审稿人
最低5最高7标准差0.9
5
7
7
2.7
置信度
正确性2.7
贡献度2.7
表达3.3
NeurIPS 2024

Learning to Solve Quadratic Unconstrained Binary Optimization in a Classification Way

OpenReviewPDF
提交: 2024-05-14更新: 2024-11-06
TL;DR

We propose a neural solver to solve quadratic unconstrained binary optimization in a classification way, which can achieve near-optimal solutions in milliseconds, even for instances comprising thousands of variables.

摘要

The quadratic unconstrained binary optimization (QUBO) is a well-known NP-hard problem that takes an $n\times n$ matrix $Q$ as input and decides an $n$-dimensional 0-1 vector $x$, to optimize a quadratic function. Existing learning-based models that always formulate the solution process as sequential decisions suffer from high computational overload. To overcome this issue, we propose a neural solver called the Value Classification Model (VCM) that formulates the solution process from a classification perspective. It applies a Depth Value Network (DVN) based on graph convolution that exploits the symmetry property in $Q$ to auto-grasp value features. These features are then fed into a Value Classification Network (VCN) which directly generates classification solutions. Trained by a highly efficient model-tailored Greedy-guided Self Trainer (GST) which does not require any priori optimal labels, VCM significantly outperforms competitors in both computational efficiency and solution quality with a remarkable generalization ability. It can achieve near-optimal solutions in milliseconds with an average optimality gap of just 0.362% on benchmarks with up to 2500 variables. Notably, a VCM trained at a specific DVN depth can steadily find better solutions by simply extending the testing depth, which narrows the gap to 0.034% on benchmarks. To our knowledge, this is the first learning-based model to reach such a performance.
关键词
quadratic unconstrained binary optimizationcombinational optimizationmachine learningclassificationneural solver

评审与讨论

审稿意见
5

This article introduces the Value Classification Model (VCM), a neural solver for the quadratic unconstrained binary optimization (QUBO) problem. VCM utilizes a Depth Value Network (DVN) and a Value Classification Network (VCN) to efficiently generate solutions without optimal labels. It outperforms existing models in computational efficiency and solution quality, achieving near-optimal solutions in milli-seconds.

优点

This article focuses on employing learning-based approaches to solve the Quadratic Unconstrained Binary Optimization (QUBO) problem.

  • An innovative graph feature extractor, DVN, is proposed. Compared to the commonly used GCN, the DVN model presented in this work overcomes the degradation issue that occurs when the number of GCN layers increases. Experimental results demonstrate that the model's performance improves as the number of DVN layers increases.

  • Additionally, the authors introduce the VCN model, which directly generates solutions and achieves higher efficiency compared to other deep reinforcement learning-based methods. Furthermore, this article proposes the GST training approach, which uses solutions generated by BGF as labels for supervised model training, thereby avoiding the need to obtain the optimal solutions of the problem in advance.

缺点

In recent year, there are lots of research about solving QUBO problems. My major concerns are about whether the proposed approach outperforms the SOTA algorithms.

  • Besides GNN-based DRL models, there are also works that directly solve the QUBO problem using GNNs, particularly [1]. That study also employs GNNs to address the QUBO problem, handling problem sizes up to tens of thousands of nodes. I think such a relevant paper should be cited. The unsupervised method in [1] directly uses the optimization objective function as the loss function and relaxes the 0-1 variables for optimization. Intuitively, this training method, which is directly guided by the loss function, might be more effective than using local optima generated by GST as labels for training. The article should provide further explanation on this matter. Additionally, it is crucial to perform ablation studies comparing this unsupervised method with the method that uses local search solutions as labels. This is important because it pertains to the key point of the article, viewing the QUBO combinatorial problem from a classification perspective.

  • The authors should consider more advanced methods as baselines, and these comparisons should be conducted on practical problems such as MaxCut or Maximum Independent Set. The reason for this is that in practical problems like MaxCut, some methods even have performance guarantees, such as the Goemans-Williamson Max-cut algorithm, which provides a bound of 0.878.

-- From https://plato.asu.edu/ftp/qubo.html, the SOTA exact algorithm for QUBO problems is QuBowl. The authors should also consider this one as a baseline. Also, Gurobi is an exact solver for QUBO, one can also run Gurobi for a few seconds (no need to run it for hours) and obtain near-optimal solutions. I think the authors should present a more fair comparison.

[1] Schuetz, M.J., Brubaker, J.K. and Katzgraber, H.G., 2022. Combinatorial optimization with physics-inspired graph neural networks. Nature Machine Intelligence, 4(4), pp.367-377.

问题

  • In the DVN section, does the model suffer from the performance degradation seen in GCNs as the number of layers increases

  • I'm wondering why using a greedy approach to obtain local optima as labels for supervised learning is superior to unsupervised methods.

局限性

The authors discussed the limitation explicitly.

作者回复

We sincerely appreciate your thorough review and constructive feedback on our manuscript. We are grateful for your insights and will address your concerns comprehensively in our revision.

To better illustrate our work and to improve the manuscript based on your comments and suggestions, we have added the following improvements:

1. Comparison with Exact Algorithm

We have added comparisons with QuBowl from [1], and the results are as follows, where VCM and VCMG outperform two exact solvers and can achieve almost all optimal solutions (19/20) within 10ms.

SolvedGurobiQwulVCMVCMG
B100(10)10(0.1s)10(0.1s)10(4.5ms)10(5.6ms)
B250(10)0(3600s)7(610.6s)9(7.3ms)9(9.9ms)

We have also added comparisons with Gurobi within 1s, and the results are as follows.

MethodB2500P3000P4000P5000P6000P7000
GAP(%)T(ms)GAP(%)T(ms)GAP(%)T(ms)GAP(%)T(ms)GAP(%)T(ms)GAP(%)T(ms)
Gurobi-1s0.0341E30.0651E30.1081E30.1041E30.1231E30.1481E3
VCM50-dd3000.034520.066530.115520.099530.144760.144116
VCMG50-dd3000.027640.040790.0881080.0781450.1092490.108331

Clearly, compared to QuBowl and Gurobi, our VCM and VCMG are more efficient QUBO solvers.

2. Validation of practical problem

We extended our evaluation to include several MaxCut problem benchmarks. Data from [2], using the objective function as the indicator:

InstanceOPTS2V-DQNVCM50-dd100
G54100-G5410000 (10 intances)110.6108.2109.6(5ms)

Data from [3], using the average approximation ratios as the indicator:

InstanceTabuSoftTabuS2V-DQNECO-DQNVCM-dd100VCMG-dd100
G32-G34 (2000 nodes)0.9150.9830.9230.9690.990(16ms)0.991(23ms)

The results show that VCM provides impressive performance in MaxCut benchmarks, which validates its applicability to QUBO-related problems.

3. Comparison with Unsupervised Method

We appreciate your bringing this significant related work [4] to our attention, and we will include a citation and discussion of this paper in our revision. The unsupervised method, which directly uses the optimization objective function as the loss function, is intuitively appealing. We have trained VCM by the well-known supervised learning method [4] on size 50 under the same dataset and neural settings, which we call VCM-UnS. Its batch loss function in VCM can be concluded as follows. pi(θVCM)=(statei(θ_VCM)+1)/2.{p_i}( \theta_{VCM}) = ({state}_i (\theta\_{VCM})+1)/2.

L(θVCM)=1Bb=1Bk=1Npi(θ_VCM)Q_ijpj(θ_VCM)L(\theta_{VCM})=\frac{1}{B} \sum_{b=1}^B \sum_{k=1}^N{p}_i (\theta\_{VCM}) Q\_{ij}{p}_j(\theta\_{VCM})

Details on the training process are shown in Figure 1 in the submitted PDF file and the optimal training GAP (%) of VCM and VCM-UnS are obtained as follows.

VCM-UnSVCM-GST
Optimal Training Gap (%)0.2310.113

It is evident that GST provides a more efficient training process compared to UnS. UnS experiences significant fluctuations. In contrast, the GST training process is remarkably smooth. Compared to the unsupervised trainer UnS, which relies on self-driven model training, GST offers the following advantages in QUBO:

  • Integration of both self-driven and heuristic-guided training processes.
  • Current labels based on VCM’s performance.
  • Provision of a clear learning target guided by heuristic, enabling the creation of more suitable labels for training and facilitating quicker and more stable model convergence.
  • The ability to train the VCM without requiring global optimal solutions as labels.

4. The degradation in GCN

We used the GCN [5] to replace the DVN of VCM. In GCN, we used residual connections between hidden layers to facilitate the training of deeper models. The ll-th layer in GCN is calculated as follows.

Hl+1=σ(D~1/2(Q)D~1/2H(l)W(l))+H(l)H^{l+1}=\sigma\left(\tilde{D}^{-1/2}(Q)\tilde{D}^{-1/2}H^{(l)}W^{(l)}\right)+H^{(l)}

After the LL layer, the solution can be generated as follows.

x=VCN(HL)x=VCN(H^{L})

We have trained GCN and VCM by GST on size 50 under the same settings. Details and optimal training gap trends are shown in Figure 2 and Figure 3 in the submitted PDF file and the best training GAPs (%) are obtained as follows.

GCN-L1GCN-L2GCN-L3GCN-L4GCN-L5GCN-L10
Optimal Training Gap (%)37.8524.1516.2315.9131.3346.29
VCM-dd1VCM-dd2VCM-dd3VCM-dd4VCM-dd5VCM-dd10
18.187.023.542.081.360.35

Obviously, the GCN suffered from performance degradation, which is consistent with the conclusion in [5]. However, the performance of VCM steadily improves with increasing depth. Besides, the neural parameters in GCN layers are independent, whereas neural units in VCM depth are consistent, resulting in lower training costs under the same neural settings.

Overall

We are grateful for your in-depth review and valuable suggestions and hope these improvements will significantly enhance the quality and impact of our work. We remain open to addressing any additional questions or concerns you may have and sincerely invite you to reassess the contributions of our paper.

(Character limit, omitting part of ref information)

[1] Rehfeldt, D., et al. Faster exact solution of sparse MaxCut and QUBO problems.

[2] Khalil, E., et al. Learning combinatorial optimization algorithms over graphs.

[3] Nath, A., et al. A Benchmark for Maximum Cut: Towards Standardization of the Evaluation of Learned Heuristics for Combinatorial Optimization.

[4] Schuetz, M.J., et al. Combinatorial optimization with physics-inspired graph neural networks.

[5] Kipf, T. N., et al. Semi-supervised classification with graph convolutional networks.

评论

Thanks for addressing my comments. However, I have listed in my previous comment a very important and relevant baseline algorithm from [1]. The authors seemed to miss this one. If the authors would not show a comparison study on this, then they fail to convince me that the proposed algorithm outperforms the SOTA algorithms.

[1] Schuetz, M.J., Brubaker, J.K. and Katzgraber, H.G., 2022. Combinatorial optimization with physics-inspired graph neural networks. Nature Machine Intelligence, 4(4), pp.367-377.

评论

Thank you for reminding us of the omission of the comparison with the SOTA PI-GNN model from Schuetz et al. [1]. In your first round of review comments, you showed great interest in the unsupervised training method used by Schuetz et al. [1]. We thus focused on comparing the unsupervised training method with our proposed trainer GST. We fully agree with you that including the PI-GNN model, which is the most recent cutting-edge method published for solving combinatorial optimization problems, for comparative study will greatly enhance the quality of our paper. Therefore, we have conducted relevant comparative experiments. To ensure a fair comparison, we used the provided source code from Schuetz et al. [1] and compared it under the same environment using the same test data. We tested the PI-GNN with layers 2, 3, and 5 on the applied 31 benchmark instances with 2500 to 7000 variables in our work. The results are shown below.

MethodB2500(10)P3000(5)P4000(5)P5000(5)P6000(3)P7000(3)
GAP(%)T(ms)GAP(%)T(ms)GAP(%)T(ms)GAP(%)T(ms)GAP(%)T(ms)GAP(%)T(ms)
Gurobi-1s0.0341E30.0651E30.1081E30.1041E30.1231E30.1481E3
PI-GNN(2 Layers)1.68944E32.13062E31.63671E31.41886E31.986109E31.437159E3
PI-GNN(3 Layers)1.90957E32.52372E32.09280E31.945100E32.180133E32.076186E3
PI-GNN(5 Layers)3.280117E33.047103E32.761113E32.463130E32.584179E32.289265E3
VCM500.36280.86180.66980.78380.80690.7029
VCMG500.136440.2771000.2141700.2623260.2675230.224770
VCM50-dd3000.034520.066530.115520.099530.144760.144116
VCMG50-dd3000.027640.040790.0881080.0781450.1092490.108331

The results show that our proposed VCM competes very well with PI-GNN. This further demonstrates the outstanding performance of our VCM in solving QUBO problems. We have included the above results into our revised manuscript, to enhance the paper’s quality and its impact.

Thank you once again for your constructive input. We remain open to addressing any additional questions or concerns you may have.

[1] Schuetz, M.J., Brubaker, J.K. and Katzgraber, H.G., 2022. Combinatorial optimization with physics-inspired graph neural networks. Nature Machine Intelligence, 4(4), pp.367-377.

评论

Thanks for addressing my comment. I'm now satisfied now and raise my score. In the paper checklist, the authors said If the paper is accepted, the data and code will be considered for public release. I strongly recommend the authors make the data and source code available to public if this paper is accepted, since this will definitely help the community.

评论

Thank you for your positive feedback and recognition. We fully agree that making these resources available would greatly benefit the community. We are diligently organizing the data and code to ensure they are readily accessible for researchers. Once the paper is accepted, we will make the data and code publicly available to facilitate further research and comparison.

审稿意见
7

The article introduces a novel neural solver named Value Classification Model (VCM) for solving the Quadratic Unconstrained Binary Optimization (QUBO) problem. Leveraging a Depth Value Network (DVN) that exploits the symmetry of the problem's matrix, VCM captures value features effectively. The solver uses these features in a Value Classification Network (VCN) for direct solution classification, bypassing the inefficiencies of sequential decision-making models. This approach significantly reduces computational overhead and achieves near-optimal solutions rapidly.

优点

  1. The introduction of the Depth Value Network that utilizes the symmetry property of the Q matrix is innovative and effectively captures valuable features without the performance degradation seen in traditional GCN models due to increased convolution layers.
  2. The VCM demonstrates impressive computational efficiency and quality of solutions, achieving near-optimal results in milliseconds, which is commendable.
  3. The model's ability to generalize across various instance sizes without retraining is particularly notable, showing robustness and adaptability.

缺点

  1. While the model performs exceptionally well on the datasets tested, the scalability and adaptability to even larger datasets or different types of QUBO instances remain to be fully validated.
  2. The paper could benefit from a more detailed comparison with existing state-of-the-art models specifically designed for hypergraph networks, which might provide a clearer picture of the model's relative performance.

问题

  1. Have the authors considered comparing the proposed VCM with models designed for hypergraph neural networks, especially in the context of solving Quadratic Optimization problems[1]? Recent works in this area could provide a valuable benchmark.
  2. The paper focuses solely on unconstrained quadratic binary optimization problems. Have the authors considered how the framework might be adapted to address constrained optimization problems, which are more prevalent in practical applications?

[1] Xiong Z, Ye H, Zong F, et al. NeuralQP: A General Hypergraph-based Optimization Framework for Large-scale Quadratically Constrained Quadratic Programs[J].

局限性

  1. The article only discusses the application of VCM to unconstrained quadratic binary optimization problems. Many practical optimization scenarios involve constraints that might affect the solution space significantly.
  2. While the model reduces computational overhead, the dependency on specific architectural choices and hyperparameters may affect its performance across diverse scenarios not covered in the experiments.
作者回复

Thank you for your thorough review and valuable feedback on our paper. Your comments are crucial for improving the quality of our work. We would like to address each of your points and suggestions.

1. Comparison

We appreciate your suggestion to discuss NeuralQP [1], a nice work targeting Quadratically Constrained Quadratic Programs (QCQPs). NeuralQP significantly enhances solution efficiency and convergence speed for large-scale QCQPs by utilizing a hypergraph-based neural prediction and iterative neighborhood optimization. It provides a new insight into the solver horizon which is very interesting, and we will discuss it in our revised manuscript.

Specifically, NeuralQP addresses a class of constrained QCQPs with continuous variables, while our work focuses more on providing a fast near-optimal solver for QUBOs, which is a novel attempt to improve computational efficiency and solution quality in unconstrained binary settings. Consequently, it is currently challenging to reformulate NeuralQP’s hypergraph problems into the QUBO formulation, making them difficult to solve with our VCM. However your remarkable suggestion has intrigued us a lot, and NeuralQP provides valuable insights for potential future applications of our VCM, which we intend to explore in future research.

To provide as valuable a benchmark as possible, we have added additional validations with Gurobi within 1s.

MethodB2500P3000P4000P5000P6000P7000
GAP(%)T(ms)
Gurobi-1s0.0341E30.0651E30.1081E30.1041E30.1231E30.1481E3
VCM50-dd3000.034520.066530.115520.099530.144760.144116
VCMG50-dd3000.027640.040790.0881080.0781450.1092490.108331

For larger datasets, we generated G instances with 10,000 and 20,000 nodes under two distributions for validation.

MethodG10000-R0.1G10000-R0.3G20000-R0.1G20000-R0.3
GAP(%)T(ms)
Gurobi-1s0.1831E30.0451E30.0911E30.1081E3
VCM50-dd3000169017206370676

The results demonstrate that our VCM and VCMG represent more efficient QUBO solvers under the same time situation.

Furthermore, we tested VCM50 under test depth 100 with the SOTA exact algorithm QuBowl [2].

SolvedGurobiQwulVCMVCMG
B100(10)10(0.1s)10(0.1s)10(4.5ms)10(5.6ms)
B250(10)0(3600s)7(610.6s)9(7.3ms)9(9.9ms)
The results show that VCM and VCMG outperform the two exact solvers.

We have also supplemented the validation test results of GCN and we can see its performance degrades in QUBO problems as depth increases. Besides, we compared our proposed GST trainer with the unsupervised trainer from [3], highlighting the need to use self-driven and heuristic guidance in VCM training. These results are presented in our overall response, with figures included in the submitted PDF file.

2. Scalability

The test instances adopted in this study cover various variable sizes (20 to 7000) and data distributions (5 distributions in G instances and 31 benchmark instances represent diverse task distributions). VCM demonstrates consistent advantages across these instances, showcasing its strong scalability.

3. Hyperparameters

Two main hyperparameters include hidden size (hh) and depth (dd). Our experiments in Appendix F, show that increasing hh has a limited impact on model performance (less than 0.1% difference for sizes ranging from 32 to 256). On the contrary, the impact of dd is quite significant. VCM’s performance steadily improves as depth increases.

4. Applicability

The QUBO problem is a classical nonlinear optimization problem that is fundamental to many combinatorial optimization challenges. The proposed VCM aims to provide a new insight to solve this fundamental formulation. However, the way to recast other combinatorial optimization problems into a QUBO formulation is a significant yet challenging work. Over the past decades, researchers have actively advanced this area of study. Some linear or quadratic problems with linear constraints and bounded integer variables can be re-formulated as QUBO using quadratic penalties P [4]. For several simple constraint examples [4], appropriate quadratic penalties can be directly applied.

Therefore, a feasible approach to applying the VCM to solve various problems is to explore methods for reformulating these problems into QUBO.

In response to your feedback and that of other reviewers, we have expanded our evaluation to MaxCut benchmarks. It is clear that both VCM and VCMG outperform state-of-the-art methods on these benchmarks, validating their practical applicability.

Data from [5], using the average objective function as the indicator:

InstanceOPTS2V-DQNVCM50-dd100
G54100-G5410000(10)110.6108.2109.6(5ms)

Data from [6], using the average approximation ratios as the indicator:

InstanceTabuSoftTabuS2V-DQNECO-DQNVCM-dd100VCMG-dd100
G32-G34 (2000 nodes)0.9150.9830.9230.9690.990(16ms)0.991(23ms)

Overall

We appreciate your thorough review and remain open to addressing any additional questions or concerns you may have.

(Character limit, omitting part of ref information)

[1] Xiong Z, et al. NeuralQP: A General Hypergraph-based Optimization Framework for Large-scale Quadratically Constrained Quadratic Programs.

[2] Rehfeldt, D., et al. Faster exact solution of sparse MaxCut and QUBO problems.

[3] Schuetz, M.J., et al. Combinatorial optimization with physics-inspired graph neural networks.

[4] Kochenberger, G., et al. The unconstrained binary quadratic programming problem: a survey.

[5] Khalil, E., et al. Learning combinatorial optimization algorithms over graphs.

[6] Nath, A., et al. A Benchmark for Maximum Cut: Towards Standardization of the Evaluation of Learned Heuristics for Combinatorial Optimization.

评论

Thank you for your thorough and comprehensive response to my review comments. I appreciate the detailed explanations and additional validations you provided, which have addressed my concerns effectively.

I am particularly impressed with the additional benchmark comparisons and the insights you shared regarding the scalability and applicability of the VCM. Your efforts to clarify the distinctions between your work and other models, such as NeuralQP, as well as the potential for future research directions, are very much appreciated.

Given the quality of your responses and the robustness of your model's performance across various tests, I would like to raise my score to 7.

评论

Thank you for your constructive feedback. We are truly grateful for your recognition and the opportunity to address your concerns. Your comments have been invaluable in improving our paper, and we are very pleased that our additional explanations and validations were able to clarify our contributions. We sincerely appreciate your support.

审稿意见
7

This paper presents a novel approach, the Value Classification Model (VCM), for tackling the challenging Quadratic Unconstrained Binary Optimization (QUBO) problem. VCM improves by offering a classification-based solution, addressing limitations faced by existing deep reinforcement learning (DRL) methods. It directly outputs the binary solution without the need for complex policy learning steps in DRL.

The paper proposes a novel self-training strategy using a greedy flip algorithm. This approach eliminates the requirement for pre-labeled optimal solutions, which can be scarce for QUBO problems.

The paper demonstrates significant performance gains by VCM compared to existing methods.

Overall, the VCM approach presents a compelling new direction for solving QUBO problems. The paper effectively addresses limitations of existing methods and showcases promising results.

优点

Overcomes DRL Limitations: The paper effectively highlights the computational burdens of DRL for QUBO and overcomes through classification based method.

Direct Solution Output: VCM directly outputs the binary solution, eliminating the need for complex policy learning steps often required in DRL.

Greedy-Guided Self-Training: The training strategy using a greedy flip algorithm and past solutions and bypasses the need for pre-labeled optimal solutions, which can be scarce for QUBO problems.

VCM shows near-optimality, high efficiency, and generalization ability in problem-solving.

缺点

  1. Could the authors describe the datasets? what problems exactly are getting solved?

问题

see weakness

局限性

Yes

作者回复

We sincerely appreciate your thorough review and the opportunity to clarify our work regarding the datasets and specific problems solved. We are grateful for the chance to provide a more comprehensive description.

1. Datasets

Our study utilized datasets described in the format "dataset+instance size+(number of instances)". The instance size here represents the nodes in the graph.

These datasets include:

1) Generated Dataset (G):

  • Q matrix elements are integers uniformly randomized within [-100, 100], following the benchmark data format [1].
  • Used for training (512,000 instances), validation (1,000 instances), and test (1,000 instances) for various instance sizes (10, 20, 50, 100 for training and validation, and 20, 50, 100, 200, 500, 1000 for test).
  • To further evaluate VCM's distribution generalization ability, we regenerated 1,000 test instances by invalidating matrix elements in G instances to zero with probabilities of 10% (R-0.9), 40% (R-0.6), 70% (R-0.3), and 90% (R-0.1). Additionally, we generated G-RN-1 following a standard normal random distribution (RN-1).

2) Well-known Datasets (31 instances):

  • B2500(10) from ORLIB [1].
  • P3000(5), P4000(5), P5000(5), P6000(3) and P7000(3) from [2].
  • Q matrix elements are integers within [-100, 100], with varying percentages of zero elements across instances. These 31 instances from well-known datasets represent diverse task distributions and we have included a visualization of distributions in Figure 4 the submitted PDF file.

The test instances adopted in this study cover various variable sizes and data distributions, posing significant challenges to model performance.

2. Applicability

The QUBO problem is a classical nonlinear optimization problem fundamental to many combinatorial optimizations. Numerous combinatorial optimization problems can be recast into standard QUBO formulations, which can be solved using QUBO methods. In recent years, researchers have actively promoted related research [3], [4]. It can introduce the penalty function P to liberate constraints and establish QUBO formulations [5]. For several simple constraint examples, appropriate quadratic penalties can be used directly.

f(x)=xTQx+P(Axb)T(Axb)=xTQx+xDx+c=xTQ^x+c\displaystyle f(x)= {x^T}Qx + P(Ax-b)^T(Ax-b)={x^T}Qx + xDx+c={x^T}\hat{Q} x +c

Following your comment and those of other reviewers, we extended our evaluation to include several MaxCut problem benchmarks. We applied the VCM50 under testing depth 100 for validation and the results are shown below. The results obtained demonstrate that VCM's solving performance is consistent with the advantages shown in our test instances, further validating its applicability.

Results from [6], using the average objective function as the indicator:

InstanceOPTS2V-DQNVCM50-dd100
G54100110108108(5ms)
G54200112108108(5ms)
G54300106104104(5ms)
G54400114108114(5ms)
G54500112112112(5ms)
G54600110110110(5ms)
G54700112108112(5ms)
G54800108108106(5ms)
G54900110108110(5ms)
G5410000112108112(5ms)
Avg.110.6108.2109.6(5ms)

Results from [7], using the average approximation ratios as the indicator:

InstanceTabuSoftTabuS2V-DQNECO-DQNVCM-dd100VCMG-dd100
G32-G34 (2000 nodes)0.9150.9830.9230.9690.990(16ms)0.991(23ms)

Additionally, we provide comparison results with GCN to verify that GCN’s performance degrades as the layer increases in QUBO problems. We also demonstrate that our trainer GST outperforms the unsupervised trainer (UnS) from [8], emphasizing the effectiveness of self-driven and heuristic guidance for VCM training. These results are presented in our overall response, with figures included in the submitted PDF file.

Overall

We believe these additions significantly improve the clarity and impact of our work, providing readers with a more comprehensive understanding of VCM's performance across various QUBO problem instances. Thank you again for your valuable feedback. We look forward to your further comments and are prepared to address any additional questions or concerns you may have.

[1] Beasley, J. E.,1996. Obtaining test problems via internet. Journal of Global Optimization, 8, 429-433.

[2] Palubeckis, G., 2004. Multistart tabu search strategies for the unconstrained binary quadratic optimization problem. Annals of Operations Research, 131, 259-282.

[3] Glover, F., Kochenberger, G., Hennig, R., & Du, Y., 2022. Quantum bridge analytics I: a tutorial on formulating and using QUBO models. Annals of Operations Research, 314(1), 141-183.

[4] Glover, F., Kochenberger, G., Ma, M., & Du, Y., 2022. Quantum Bridge Analytics II: QUBO-Plus, network optimization and combinatorial chaining for asset exchange. Annals of Operations Research, 314(1), 185-212.

[5] Kochenberger, G., Hao, J. K., Glover, F., Lewis, M., Lü, Z., Wang, H., & Wang, Y., 2014. The unconstrained binary quadratic programming problem: a survey. Journal of combinatorial optimization, 28, 58-81.

[6] Khalil, E., Dai, H., Zhang, Y., Dilkina, B., & Song, L., 2017. Learning combinatorial optimization algorithms over graphs. Advances in neural information processing systems, 30.

[7] Nath, A., & Kuhnle, A., 2024. A Benchmark for Maximum Cut: Towards Standardization of the Evaluation of Learned Heuristics for Combinatorial Optimization. arXiv preprint arXiv:2406.11897.

[8] Schuetz, M.J., Brubaker, J.K. and Katzgraber, H.G., 2022. Combinatorial optimization with physics-inspired graph neural networks. Nature Machine Intelligence, 4(4), pp.367-377.

评论

I thank the authors for the detailed answer.. I increase my score

评论

We are so delighted that you have raised the score for our paper. We sincerely appreciate your positive feedback and support, which have greatly helped us improve the quality of the paper.

作者回复

Overall Response

Dear Reviewers,

We sincerely appreciate your thorough reviews and constructive feedback on our manuscript. Your insights have been invaluable in improving the quality and impact of our work. Below, we address your comments and outline the improvements made in response to your suggestions.

1. Comparison with Exact Algorithm

For a fair comparison, we have added validations with Gurobi within 1 second and the results are shown below.

MethodB2500P3000P4000P5000P6000P7000
GAP(%)T(ms)
Gurobi-1s0.0341E30.0651E30.1081E30.1041E30.1231E30.1481E3
VCM50-dd3000.034520.066530.115520.099530.144760.144116
VCMG50-dd3000.027640.040790.0881080.0781450.1092490.108331

For larger datasets, we generated G instances with 10,000 and 20,000 nodes under two distributions for validation.

MethodG10000-R0.1G10000-R0.3G20000-R0.1G20000-R0.3
GAP(%)T(ms)
Gurobi-1s0.1831E30.0451E30.0911E30.1081E3
VCM50-dd3000169017206370676

Besides, we have added validations with the SOTA exact algorithm QuBowl [1] and the results are shown below.

SolvedGurobiQwulVCMVCMG
B100(10)10(0.1s)10(0.1s)10(4.5ms)10(5.6ms)
B250(10)0(3600s)7(610.6s)9(7.3ms)9(9.9ms)

Clearly, the above results further verify the high efficiency of our VCM, which can provide optimal or near-optimal solutions in milliseconds.

2. Validation of Practical Problem

We extended our evaluation to MaxCut benchmarks and the results are shown below.

Data from [2], using the average objective function as the indicator:

InstanceOPTS2V-DQNVCM50-dd100
G54100-G5410000 (10 instances)110.6108.2109.6(5ms)

Data from [3], using the average approximation ratios as the indicator:

InstanceTabuSoftTabuS2V-DQNECO-DQNVCM-dd100VCMG-dd100
G32-G34 (2000 nodes)0.9150.9830.9230.9690.990(16ms)0.991(23ms)

The results confirm that VCM outperforms other methods, consistent with the advantages demonstrated in our test instances. This validates the applicability and robustness of our model in problems that can be recasted as QUBO across different scales and distributions.

3. Scalability and Performance

Our study covers various variable sizes (20 to 7000) and data distributions, demonstrating VCM's consistent performance advantages. We also investigated the impact of hyperparameters in Appendix F, particularly the depth of the DVN, on model performance. Our findings indicate that deeper models yield better performance, whereas variations in hidden size have a negligible impact.

4. GST vs. Unsupervised Method

We have compared our trainer GST with an unsupervised trainer UnS [4] under the same VCM settings. The results of optimal training gaps are shown below (while details can be found in Fig 1 in the submitted PDF file).

VCM-UnSVCM-GST
Optimal Training Gap (%)0.2310.113

We found that the GST training process offers significant advantages over the unsupervised method. GST integrates both self-driven and heuristic-guided training, enabling quicker and more stable model convergence without requiring global optimal solutions. This results in more efficient training and better performance for VCM in solving QUBO problems.

5. Addressing Degradation in GCN

We applied GCN [5] to replace the DVN in VCM and confirmed performance degradation with increased LL layers, consistent with existing literature [5]. The additional results of optimal training gaps are shown below, with details and optimal training gap trend shown in Figure 2 and Figure 3 in the submitted PDF file.

GCN-L1GCN-L2GCN-L3GCN-L4GCN-L5GCN-L10
Optimal Training Gap (%)37.8524.1516.2315.9131.3346.29
VCM-dd1VCM-dd2VCM-dd3VCM-dd4VCM-dd5VCM-dd10
Optimal Training Gap (%)18.187.023.542.081.360.35

VCM's performance steadily improves with increasing depth, highlighting the robustness of our approach. The consistent neural units across every DVN depth result in lower training costs and better scalability.

Conclusion We have meticulously addressed your comments and made significant improvements to our manuscript. We believe these enhancements will greatly improve the quality and impact of our work and we remain open to any additional questions or concerns. While QUBO has been studied for many years, it is the first time that a learning-based model can stably achieve near-optimal solutions on large instances in milliseconds, and we do believe that our work is innovative and incremental. We sincerely invite you to reassess our contributions.

Reference

[1] Rehfeldt, D., Koch, T., & Shinano, Y., 2023. Faster exact solution of sparse MaxCut and QUBO problems. Mathematical Programming Computation, 15(3), 445-470.

[2] Khalil, E., Dai, H., Zhang, Y., Dilkina, B., & Song, L., 2017. Learning combinatorial optimization algorithms over graphs. Advances in neural information processing systems, 30.

[3] Nath, A., & Kuhnle, A., 2024. A Benchmark for Maximum Cut: Towards Standardization of the Evaluation of Learned Heuristics for Combinatorial Optimization. arXiv preprint arXiv:2406.11897.

[4] Schuetz, M.J., Brubaker, J.K. and Katzgraber, H.G., 2022. Combinatorial optimization with physics-inspired graph neural networks. Nature Machine Intelligence, 4(4), pp.367-377.

[5] Kipf, T. N., & Welling, M., 2016. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907.

最终决定

This paper focuses on the use of machine learning to solve the widely applicable combinatorial QUBO problem. In this regard, the submissions makes various advances:

  • Different from the current ML based schemes utilizing a RL based approach to generating candidate solutions sequentially, the proposed approach uses a classifier to directly output candidate solutions, making the candidate solution generation step significantly more efficient.
  • Identifying the limitations of GNN based variable representations of existing schemes, the paper proposes a representation that (i) leverages the symmetry of the QUBO problem, (ii) can be applied to problems of any size at training time, and (iii) is recursive in nature, and has been shown to perform better with more recursions at test time.
  • The proposed model is trained without the need for high-quality solutions for the training problems by utilizing a GPU-accelerated greedy flip algorithm that efficiently generates progressively improving solutions that the model is evaluated against during training.

Overall the empirical results, highlighting the strong performance of the proposed scheme against various ML based schemes and off-the-shelf solvers (Gurobi). During the author-reviewer discussion, additional baselines (such as Gurobi-1, PI-GNN, QuBowl) and benchmarks (MaxCut) were brought up, and the new results continue to highlight the strong performance of the proposed scheme. Furthermore, the authors clearly evaluate the ability of the model to be trained on small instances, and tested on larger instances, highlighting the (size) generalization capability of the proposed scheme. The submission also contains thoughtful ablations, highlighting the effect of the different choices made in the overall pipeline.