Improved Theoretically-Grounded Evolutionary Algorithms for Subset Selection with a Linear Cost Constraint
摘要
评审与讨论
The manuscript presents an advanced study on subset selection problems, which are prevalent in various fields such as machine learning, operations research, and economics. The authors focus on subset selection under a linear cost constraint, a problem characterized by its NP-hardness and practical importance. The paper introduces an improved theoretically-grounded evolutionary algorithm (EA), enhancing the performance of the Pareto Optimized Multi-Criteria (POMC) algorithm and proposing a novel multi-objective EA, named EPOL.
给作者的问题
How does the proposed EPOL algorithm scale with the size of the problem (e.g., larger ground sets or higher-dimensional feature spaces)? Could the authors provide some insights or preliminary results on scalability?
Are there any parameters in the EPOL algorithm that require fine-tuning? If so, how sensitive are the results to these parameters?
The empirical study focuses on maximum coverage and influence maximization. Would the authors consider extending the study to other relevant problems such as budgeted submodular optimization?
论据与证据
yes, this paper has made complete proofs for all Lemmas.
方法与评估标准
yes, this work deepens the theoretical understanding of EAs.
理论论述
yes, i have checked the proofs of Lemma 3.2 and 3.3
实验设计与分析
yes, the experimental designs and analyses are abundant.
补充材料
yes, i have read the supplementary material with more experiments.
与现有文献的关系
yes, there remains a gap in the approximation bounds of EAs compared to greedy algorithms, and their full theoretical potential is yet to be realized.
遗漏的重要参考文献
this paper provides a theoretical understanding of EAs, which has not been discussed in previous literatures.
其他优缺点
Strengths:
The paper makes a significant theoretical contribution by improving the approximation guarantee of the POMC algorithm to 1/2, which is a substantial advancement in the field.
The proposed EPOL algorithm is innovative and achieves the best-known practical approximation guarantee of 0.6174, which is a notable empirical performance.
This work not only advances the subset selection problem but also deepens the theoretical understanding of EAs.
The writing is clear and the mathematical formulation is precise, making the paper accessible to readers familiar with evolutionary algorithms and optimization.
Weaknesses:
The manuscript would benefit from a more detailed comparison with other state-of-the-art algorithms in the literature, especially in terms of computational complexity and scalability.
The discussion on the limitations of the proposed methods is somewhat brief. A more in-depth analysis could provide insights into potential improvements or alternative applications.
I understand that this work is mainly a theoretical contribution, but if we can compare more SOTA algorithms, we can further demonstrate the superiority of the algorithms.
其他意见或建议
None
Thank you very much for your positive feedback and for recognizing the strengths of our work. We appreciate your kind words regarding our contributions and writing. Please find our detailed responses below.
I understand that this work is mainly a theoretical contribution, but if we can compare more SOTA algorithms, we can further demonstrate the superiority of the algorithms.
Thank you for your valuable suggestion. To the best of our knowledge, we have already included all state-of-the-art algorithms relevant to our problem setting in our comparisons. Nonetheless, we appreciate your perspective and will explore further comparisons in future work as more methods become available. Thank you.
How does the proposed EPOL algorithm scale with the size of the problem (e.g., larger ground sets or higher-dimensional feature spaces)? Could the authors provide some insights or preliminary results on scalability?
Thank you for your insightful comment. The proposed algorithm is designed to be highly parallelizable, which makes it well-suited for larger-scale problems. Thanks to your suggestion, we compare EPOL against three comparison algorithms—namely, the baseline GGA, the high-performance greedy algorithm 1-guess-Greedy, and POMC—using a larger network dataset email-Eu-core (1005 vertices, 25,571 edges) on the maximum coverage problem with . For each budget , we calculated the objective values (average std) and highlighted the best performance in bold. Additionally, a ‘’ indicates that EPOL significantly outperforms the corresponding algorithm, as confirmed by a Wilcoxon signed-rank test (confidence level 0.05). Consistent with the results reported in the paper, EPOL significantly outperforms other algorithms. Thank you again for your valuable feedback.
| Budget | ||
|---|---|---|
| GGA | 525 | 575 |
| 1-guess-Greedy | 528 | 578 |
| POMC | 531.3 1.3 | 581.3 0.9 |
| EPOL | 532.8 0.4 | 582.5 0.5 |
Are there any parameters in the EPOL algorithm that require fine-tuning? If so, how sensitive are the results to these parameters?
We would like to clarify that the design of the EPOL algorithm inherently avoids the need for fine-tuning of parameters. This feature contributes to its robust and consistent performance across various settings. Thank you.
The empirical study focuses on maximum coverage and influence maximization. Would the authors consider extending the study to other relevant problems such as budgeted submodular optimization?
Thank you for your insightful comment. While our empirical study has primarily focused on maximum coverage and influence maximization, our methodology is inherently adaptable to budgeted submodular optimization problems in scenarios where cost-aware resource allocation and the diminishing returns property of objective functions are central concerns. For instance: sensor placement, which balances information gain with installation costs (Krause et al., 2006); recommendation systems, which promote products within advertising budgets and respect user preferences (Ashkan et al., 2015); data summarization, which maximizes information retention under computational resource constraints (Lin & Bilmes, 2011); active learning, which selects maximally informative data samples under limited annotation budgets (Golovin & Krause, 2011); and human-assisted learning, which optimizes machine learning models with limited expert resources (De et al., 2020, 2021). As suggested by Reviewer vZX8, we will revise the paper to include more discussions on the applications of the studied problem, thereby improving its visibility. Thank you very much.
Reference:
Krause, A., et al. "Near-optimal sensor placements: Maximizing information while minimizing communication cost." Proc. IPSN, 2006.
Ashkan, A., et al. "Optimal Greedy Diversity for Recommendation." Proc. IJCAI, 2015.
Lin, H., and Bilmes, J. "A class of submodular functions for document summarization." Proc. ACL, 2011.
Golovin, D., and Krause, A. "Adaptive submodularity: Theory and applications in active learning and stochastic optimization." J. AI Research, 2011.
De, A., et al. "Regression under human assistance." Proc. AAAI, 2020.
De, A., et al. "Classification under human assistance." Proc. AAAI, 2021.
This paper contributed to 1) analyzing the existing approach for submodular optimization with a linear inequality constraint, and 2) proposing a novel approach with a better approximation guarantee and better practical performance. This paper first analyzed the existing approach, POMC, an evolutionary algorithm that maintains a population of non-dominated solutions where the constraint is treated as the second objective. The analysis of this paper improves the existing theoretical guarantee by a constant factor (from 0.5(1 - 1/e) to 0.5). The rigorous proof is provided in the main text, followed by a description of the difference in the proof from the proof of the existing result. Then, the authors propose a novel approach, EPOL, which simply repeats running POMC for subproblems that are constructed by removing each element. The theoretical analysis revealed the improved theoretical guarantee (from 0.5 to 0.61...). The empirical study shows that the proposed approach is not only theoretically improved, but also performs better than existing approaches over different test problems.
update after rebuttal
All my concerns have been addressed during the rebuttal phase. I keep my initial score.
给作者的问题
Please read the “Other Strengths And Weaknesses” part.
论据与证据
Their claims, improving theoretical guarantee of EAs for subset selection problem with a monotone and submodular objective function under a linear cost constraint as well as improving their practical performance, are supported by rigorous theoretical results and thorough experimental results.
方法与评估标准
Methodologies and Experimentation looks valid. However, honestly, as I am not familiar with this specific problem class, I am not fully sure about my assessment.
理论论述
I did check the proof and didn’t find any flaw.
实验设计与分析
Empirical results only show the final performance. It seems that different algorithms spent different iterations / function calls / time. Ideally, to compare the performance, the budget (iterations/f-calls/time, etc.) should be the same for all algorithms. Low performance algorithms with fast convergence may be trivially improved by incorporating a restart mechanism.
As I am not familiar with this topic itself, I am not sure whether the experiments done in this paper meet the standard of the community.
补充材料
I checked the proofs.
与现有文献的关系
The discussion about the potential real-world applications would improve the visibility of this paper.
遗漏的重要参考文献
No
其他优缺点
Strengths:
-
Improved theoretical results with rigorous and non-trivial proofs
-
Strong empirical results showing superior performance of the proposed approach over different baselines on different problems
Weaknesses:
-
The theoretical guarantee doesn't seem to match the practical performance, hence the improvement in theoretical guarantee does not necessarily imply the improvement in practice, hence its value is a bit questionable.
-
Empirical results only show the final performance. It seems that different algorithms spent different iterations / function calls / time. Ideally, to compare the performance, the budget (iterations/f-calls/time, etc.) should be the same for all algorithms. Low performance algorithms with fast convergence may be trivially improved by incorporating a restart mechanism.
其他意见或建议
The sizes of parentheses could be reconsidered for better presentation.
We sincerely appreciate your positive feedback and encouraging validation of our work's strengths. Please find our detailed responses below.
The discussion about the potential real-world applications would improve the visibility of this paper.
The discussion about the potential real-world applications would enhance the impact and relevance of this paper. The problem studied in this paper is highly relevant to domains where cost-aware resource allocation and the diminishing returns property of objective functions play a key role. It has applications in diverse areas, e.g., combinatorial optimization, computer networks, data mining, and machine learning. Beyond the maximum coverage and influence maximization examined in this paper, potential applications include sensor placement, balancing information gain with installation costs (Krause et al., 2006); recommendation systems, promoting products within advertising budgets while respecting user preferences (Ashkan et al., 2015); data summarization, maximizing information retention under computational resource constraints (Lin & Bilmes, 2011); active learning, selecting maximally informative data samples under limited annotation budgets (Golovin & Krause, 2011); and human-assisted learning, optimizing machine learning models with limited expert resources (De et al., 2020, 2021). We appreciate your suggestion and will incorporate these discussions into the paper. Thank you very much for your suggestion.
Reference:
Krause, A., et al. "Near-optimal sensor placements: Maximizing information while minimizing communication cost." Proc. IPSN, 2006.
Ashkan, A., et al. "Optimal Greedy Diversity for Recommendation." Proc. IJCAI, 2015.
Lin, H., and Bilmes, J. "A class of submodular functions for document summarization." Proc. ACL, 2011.
Golovin, D., and Krause, A. "Adaptive submodularity: Theory and applications in active learning and stochastic optimization." J. AI Research, 2011.
De, A., et al. "Regression under human assistance." Proc. AAAI, 2020.
De, A., et al. "Classification under human assistance." Proc. AAAI, 2021.
Empirical results only show the final performance. It seems that different algorithms spent different iterations /function calls / time. Ideally, to compare the performance, the budget (iterations/f-calls/time, etc.) should be the same for all algorithms. Low performance algorithms with fast convergence may be trivially improved by incorporating a restart mechanism.
Thank you for your comment. We fully agree, and actually have tried to make the comparison fair. The algorithms in our study fall into two categories: fixed-time algorithms such as GGA, Greedy, and 1-guess Greedy, with runtime complexities of , , and , respectively, and anytime algorithms such as POMC, EAMC, FPOMC, and EVO-SMC, whose performance improves with increased runtime. To ensure a fair comparison, we allocated the same number of objective function evaluations () to all anytime algorithms. EPOL decomposes the original problem into subproblems and solves them in parallel using POMC, allocating a budget of evaluations to each subproblem. Furthermore, we also compared EPOL with a variant of POMC, called P-POMC, as described in Appendix B.4. P-POMC directly runs the original problem on parallel processors, where each processor is also allocated a budget of evaluations. Our experimental settings align with prior work in the literature [Bian et al., 2020, Roostapour et al., 2022, Zhu et al., 2024]. It is worth noting that fixed-time algorithms (GGA, Greedy, and 1-guess Greedy) cannot benefit from restart mechanisms, as they always return the same solution for a given input. We will revise to make them clear. Thank you very much.
The theoretical guarantee doesn't seem to match the practical performance, hence the improvement in theoretical guarantee does not necessarily imply the improvement in practice, hence its value is a bit questionable.
Good point! Yes, the theoretical guarantee of an algorithm may not match its practical performance. This is because the theoretical guarantee implies the approximation performance in the worst case, which may differ from the tested real-world cases. Thus, it is possible that two algorithms with the same theoretical guarantee may demonstrate different performances in practice. However, a good theoretical guarantee is still important, which guarantees the worst-case performance of an algorithm and can characterize the robustness of the algorithm. Thus, a good algorithm should have both good theoretical guarantee and empirical performance. Our proposed algorithm EPOL has this property, which not only achieves the best-known practical theoretical guarantee, but also demonstrates empirical advantages across different settings. We will revise to add more explanation to make them clear. Thank you very much for your suggestion.
This paper studies the problem of subset selection with a linear cost constraint. The authors first improved the approximation guarantee (from (1-1/e)/2 to 1/2) of an existing evolutionary algorithm, so-called POMC, with the best empirical performance. Then, they proposed a new evolutionary algorithm EPOL with a better approximation guarantee of 0.6174. They also performed experiments on two applications (maximum coverage and influence maximization), showing that EPOL can achieve significantly better performance in most cases.
给作者的问题
-
In the introduction part, you mentioned “While there are greedy algorithms that obtain the theoretical optimal approximation of 1 − 1/e, they are generally impractical due to high computational costs” I’d like to see more explanations. Why are these algorithms impractical?
-
The proposed EPOL algorithm enumerates all the residual problems by excluding each single item. But in the experiments, only a subset of residual problems corresponding to the items with large values are enumerated. The authors did not give any explanation. I can understand this setting is sufficient to show the superiority of the proposed EPOL as the implemented version is weaker. But I still would like to see the performance of the full version of EPOL. Can it bring further improvement?
-
Why is the objective evaluation noisy for the application of influence maximization?
论据与证据
Yes, the claims made in the submission are supported by clear and convincing evidence. The improved approximation guarantee of the existing algorithm POMC and the approximation guarantee of the newly proposed algorithm EPOL are proved mathematically and correctly. The best empirical performance of EPOL has been clearly shown by extensive experiments. The authors compared EPOL with a large number of previous methods on the studied subset selection problem.
方法与评估标准
The studied problem of subset selection with a linear cost constraint, where the objective function is monotone submodular, has wide applications and has been studied extensively. This work improves the approximation guarantee of a previous algorithm (with the best observed performance in experiments) and proposes a new better algorithm both theoretically and empirically. The theoretical analysis is rigorous, and the experiments use common benchmark data sets and compare all state-of-the-art works. I think the proposed methods and/or evaluation criteria make sense.
理论论述
I have checked all the proofs (including the proof of a lemma in the appendix) carefully. The improved approximation guarantee of POMC mainly lies in a tighter analysis of the objective improvement by adding a single item. The proof is mainly accomplished by iteratively adding proper single items. The approximation bound of EPOL is proved by considering the residual problem corresponding to the item having the largest objective value in the optimal solution, which must be generated by enumerating all residual problems.
I almost did not find any mistakes. The proofs are presented clearly. I have only some minor questions or suggestions.
-
page 5, first column, line 248, I suggest using , to be consistent with the presentation in cases (2) and (3).
-
page 5, second column, “This implies that has already exceeded ” I suggest adding more explanations. Why the formulas in cases (1)-(3) satisfy Equation (6) with ?
-
page 5, second column, “After including … ” Did you consider , which will contradict with ?
实验设计与分析
Yes, I checked the experimental parts. The proposed algorithm is compared with a series of existing algorithms for the studied subset selection problem, including greedy and evolutionary algorithms. Two applications of subset selection with various parameter settings are tested. The authors performed a significance test, showing that the proposed EPOL algorithm is significantly better in most cases, and never significantly worse. They also conducted some ablation studies, e.g., comparing with multiple runs of the POMC algorithm on the original problem.
补充材料
Yes, but I just have a glance at the code provided in the supplementary material.
与现有文献的关系
The studied problem, subset selection, has wide applications across various areas such as machine learning, data mining, and computer networks. This paper proposes a new evolutionary algorithm with better theoretical guarantees and empirical performance. Another contribution of this work is improving the approximation bound of the existing evolutionary algorithm (which performed the best empirically for the subset selection problem). As far as I can tell, this work will bring some influence to the topic of both subset selection and evolutionary computation.
遗漏的重要参考文献
The related works are properly cited.
其他优缺点
Overall, this is a solid work with both theoretical analysis and empirical evaluation. It makes a good contribution to solving the significant problem of subset selection.
其他意见或建议
The paper is well written. Some minor issues:
-
page 1, second column, line 38-40, “further optimizes this approach” -> “further improves this approach”
-
page 2, first column, line 107, “objective” -> “objective evaluation”
-
page 4, second column, line 183, “Lemma 3.2” -> “The proof of Lemma 3.2”
-
page 4, the function in Theorem 3.1 is defined in Lemma 3.4 after its first appearance. I suggest giving the definition first.
-
page 7, first column, ->
-
page 7, first column, line 374, budget -> the budget
-
page 8, first column, line 417-419, “search” -> “searches” “maintain” -> “maintains”
-
page 8, second column, line 400, “, additional” -> “. Additional”
-
page 8, second column, line 407, “performs best” -> “performs the best” Please check throughout the paper, e.g., the same issue in page 12.
-
page 8, second column, line 413, “, shows” -> “, which shows”
-
page 9, first column, line 449, “EA POMC” -> “EA, POMC”
-
page 11, line 573, “Combing” -> “Combining”
-
page 11, line 587, “,and” -> “, and”
-
page 12, “conclude those” ?
-
caption of Table 10, “the probability of each 0.05” ?
-
page 16, line 831, “the optimal objective value” -> “the best objective value”
Thank you for your thorough review and valuable suggestions. We sincerely appreciate your feedback, which will help us improve our manuscript. Below are our detailed responses to your queries.
... more explanations. Why the formulas in cases (1)-(3) satisfy Equation (6) with ?
Thank you for your suggestion. We will include a more detailed explanation after analyzing cases (1)–(3) as follows:
By combining the inequalities derived from cases (1)–(3), we can conclude that the new solution satisfies:
.
Additionally, the cost of solution satisfies . Consequently, the solution must be included in ; otherwise, must be dominated by one solution in (line 6 of Algorithm 1). This implies that has already exceeded , contradicting the assumption that . Hence, after including , we have .
... Did you consider , which will contradict with ?
Yes, we did consider this case. On page 5, second column, starting from line 251, we addressed the situation where the inclusion of the new solution causes to exceed , i.e., . In this situation, we can deduce that is a -approximation solution. We will further clarify this point in the revised manuscript. Thank you.
“... greedy algorithms that obtain the theoretical optimal approximation of 1 − 1/e” ... Why are these algorithms impractical?
[Sviridenko, 2004] developed a -approximation algorithm by selecting three optimal elements and using a greedy approach, but it has a high time complexity of . Later, [Badanidiyuru & Vondrák, 2014] and [Ene & Nguyen, 2019] proposed greedy-based algorithms that achieve a -approximation. However, their computational costs remain prohibitively high, with time complexities of and , respectively. These complexities render the algorithms impractical for real-world applications. These details will be clarified in the introduction. Thank you.
...only a subset of residual problems corresponding to the items with large values are enumerated. ... But I still would like to see the performance of the full version of EPOL. Can it bring further improvement?
In our experiments, we enumerated a subset of residual problems to balance computational efficiency and performance. We agree that evaluating the full version of EPOL, is both interesting and necessary. Thanks to your suggestion, we conducted additional experiments comparing the original EPOL (as in the manuscript) and the full version (EPOL-full), which enumerates all residual problems. For , the objective values (avg std) on maximum coverage for three datasets are summarized in the tables. For each budget , the larger value is bolded, and ‘’ indicates that EPOL-full significantly outperforms EPOL (Wilcoxon signed-rank test, confidence level 0.05). The results highlight EPOL-full’s potential to improve performance by addressing all residual problems, consistently outperforming EPOL with significant advantages in several cases.
We will revise the manuscript to include these results and discussions. Thank you for your valuable feedback.
| Budget | |||||
|---|---|---|---|---|---|
| frb-30-15-1: | |||||
| EPOL | 301.1 1.0 | 329.7 0.5 | 354.4 0.9 | 375.2 1.5 | 394.1 1.8 |
| EPOL-full | 302.1 0.8 | 330.9 0.3 | 354.8 0.4 | 377.3 1.3 | 395.2 1.1 |
| frb-35-17-1: | |||||
| EPOL | 319.1 0.8 | 356.6 0.9 | 389.8 0.6 | 419.0 0.6 | 445.7 0.6 |
| EPOL-full | 319.8 0.4 | 357.9 0.3 | 390.6 0.5 | 419.4 0.3 | 446.5 0.5 |
| congress: | |||||
| EPOL | 332.8 0.4 | 358.0 0.6 | 381.2 0.6 | 399.9 0.7 | 415.5 0.5 |
| EPOL-full | 333.4 0.5 | 359.0 0.6 | 381.4 0.7 | 400.2 1.0 | 416.1 0.5 |
Why is the objective evaluation noisy for the application of influence maximization?
The objective evaluation for influence maximization, i.e., , is noisy because the propagation process is randomized, and we use the average of multiple Monte Carlo simulations to estimate the expectation. Specifically, starting from a solution , we simulate the propagation 500 times and use the average as the estimated objective value. We will revise to make it clear. Thank you.
Thanks for the detailed response!
My concerns have been addressed well. I will keep my score, and recommend accepting this paper.
This paper addresses monotone submodular maximization under a linear cost constraint using evolutionary algorithms. It reanalyzes the Pareto Optimization Algorithm for Monotone Cost functions (POMC), providing an improved approximation ratio of . Additionally, the authors propose a novel multi-objective evolutionary algorithm, EPOL, which achieves the best-known practical approximation ratio of . Empirical evaluations on maximum coverage and influence maximization tasks demonstrate the superior performance of EPOL compared to existing methods.
给作者的问题
No
论据与证据
The claims made in the submission are supported by clear and convincing evidence
方法与评估标准
Proposed methods and evaluation criteria make sense for the problem or application at hand.
理论论述
No apparent issue found.
实验设计与分析
Empirical evaluation is conducted on maximum coverage and influence maximization with six datasets. The results are sound.
补充材料
I have reviewed the supplementary material.
与现有文献的关系
The paper studies a fundamental combinatorial optimization problem: monotone submodular maximization under knapsack constraints. It improves the state-of-the-art approximation ratio achieved by evolutionary algorithms (EAs) from to , marking a significant advancement in the field.
遗漏的重要参考文献
No
其他优缺点
In this work, the authors re-analyze POMC and demonstrate that it achieves a approximation ratio. By repeatedly running POMC with a 1-guess technique, this approximation ratio is further improved to . However, the technical contribution of the paper appears to be somewhat limited, as the primary advancements rely on modifications and repeated applications of existing methods rather than introducing fundamentally new algorithmic ideas.
Moreover, the algorithm sections consist primarily of lengthy proofs without sufficient discussion of the high-level ideas. This makes it challenging for readers to grasp the intuition behind the proposed methods. The presentation of the paper could be significantly improved by including more intuitive explanations and balancing technical details with conceptual insights.
其他意见或建议
-
A linear cost constraint is typically referred to as a knapsack constraint in the literature.
-
The paper does not provide a clear definition of what a multi-objective EA is in the introduction or preliminary.
-
The core idea behind improving the approximation ratio of POMC should be introduced at a high level at the beginning of Section 3 to provide readers with a clear understanding of the approach before delving into technical details.
Thank you for reviewing our paper. We greatly appreciate your time and thoughtful feedback. Below are our detailed responses.
The core idea behind improving the approximation ratio of POMC should be introduced at a high level at the beginning of Section 3 ...
Moreover, the algorithm sections consist primarily of lengthy proofs without sufficient discussion of the high-level ideas...
Thank you for your suggestion. In the original paper, we included a high-level explanation of the core idea behind improving POMC's approximation ratio following the proof of Theorem 3.1 (Page 5, left column, lines 307–316). Previous analysis of POMC used a coarse-grained manner to evaluate the lower bound for improving . This led to POMC being able to derive a solution that satisfies , which is, however, infeasible, i.e., . A connection was then established between and two feasible solutions and , demonstrating that . This weakens the tightness of the bound. In contrast, our analysis adopts a fine-grained approach to evaluate the lower bound for improving (the analysis in Lemma 3.3 and the careful design of in Lemma 3.2), showing that POMC can find a feasible solution with and . Thanks to your suggestion, we will move this discussion to the beginning of Section 3, which will provide readers with a clearer conceptual overview before delving into technical details. Thank you very much.
To ensure clarity and readability, we have included brief explanations of each theorem and lemma, along with proof sketches:
-
Page 4, left column, lines 198–203: Connection between Theorem 3.1 and Lemma 3.2; Lemma 3.2's role in proving Theorem 4.1.
-
Page 4, left column, lines 211–219: Relationships between Lemma 3.2 and Lemmas 3.3–3.4; their importance in proofs.
-
Page 4, right column, lines 182–189: Core proof idea of Lemma 3.2: Relies on Lemma 3.3 to analyze growth and expected iterations.
-
Page 5, left column, lines 279–283: Proof sketch for Theorem 3.1.
-
Page 5, left column, lines 307–316: Main idea for improving POMC’s approximation ratio.
-
Page 5, right column, lines 293–299: Proof sketch for Theorem 4.1.
We believe these explanations provide readers with the necessary intuition to better understand the technical details. Thanks to your suggestion, we will revise to add more intuitive explanations to improve the readability of our work.
However, the technical contribution of the paper appears to be somewhat limited, as the primary advancements rely on modifications and repeated applications of existing methods rather than introducing fundamentally new algorithmic ideas.
Thank you for your feedback. We would like to clarify that the proposed EPOL algorithm is not a simple repeated application of existing methods. EPOL decomposes the original problem into multiple residual problems. For each , it creates a residual problem , which is then solved in parallel using POMC. We also compared EPOL with P-POMC (i.e., a simple repetition of POMC, as detailed in Appendix B.4), where POMC directly executes the original problem in parallel, and the best solution is selected as the final output. The experimental results show that EPOL outperforms P-POMC, thereby highlighting the importance of our design over simple repetition.
In fact, EPOL is carefully designed to analyze the residual problem anchored at the key element , defined as . It leverages POMC's 1/2-approximation guarantee on the residual problem and connects this guarantee to the original problem, thereby further improving the theoretical bound. Consequently, EPOL not only advances theoretical insights but also attains a SOTA practical performance guarantee.
From a practical problem-solving perspective, the primary goal is to design an algorithm that brings measurable performance improvements. We understand and respect the high expectations for fundamentally new algorithmic ideas. While such breakthroughs are exciting, most research focuses on advancing existing methods to achieve measurable improvements. To our understanding, exploring the potential of existing algorithms and further developing them into approaches that yield improvements in both theoretical analysis and empirical performance is, in itself, a significant contribution.
We hope these clarifications address your concerns. Additionally, we will revise the paper to improve presentation, including defining multi-objective EAs in the introduction, adding intuitive proof explanations, and clarifying that a linear cost constraint refers to a knapsack constraint.
Please do not hesitate to reach out if you have further questions or suggestions. Thank you very much.
This work studies evolutionary algorithms (EAs) for monotone submodular maximization subject to a linear cost constraint. In particular, it:
- improves the analysis of the approximation ratio for the POMC EA [Qian et al., IJCAI 2017] from to ,
- proposes a new multi-objective EA called EPOL, which achieves an approximation ratio of , matching that of the best-known (combinatorial) greedy algorithm. The paper is well written, gives clean proofs for all results, and presents a compelling set of experiments.
The majority of reviewers are in favor of "accept" (scores: [2, 4, 4, 4]). The reviewer who recommended "weak reject" did not respond to the authors' rebuttal, so our final recommendation for this paper is acceptance.