PepTune: De Novo Generation of Therapeutic Peptides with Multi-Objective-Guided Discrete Diffusion
PepTune generates therapeutic peptides in discrete space using Monte Carlo Tree Guidance for multi-objective optimization.
摘要
评审与讨论
The authors propose a framework based on the Masked Discrete Language Model (MDLM) (PepTune) to generate and optimize therapeutic peptides. They claim that the main contributions of their model are: 1) using MDLM to generate peptide SMILES representations, 2) introducing NELBO and reverse-posterior, 3) using Monte Carlo tree search to guide the generation of SMILES, and 4) training a set of classifiers and regressors for multiple properties.
给作者的问题
See Weaknesses.
论据与证据
The source code is not provided. The experiments are incomplete, and there is no comparison with other baseline models, which makes the paper unconvincing.
方法与评估标准
Yes
理论论述
Yes
实验设计与分析
Yes
补充材料
Yes
与现有文献的关系
Using MDLM to generate peptide SMILES representations and introducing NELBO and reverse-posterior
遗漏的重要参考文献
Wang Y, Liu X, Huang F, et al. A multi-modal contrastive diffusion model for therapeutic peptide generation[C]//Proceedings of the AAAI Conference on Artificial Intelligence. 2024, 38(1): 3-11.
其他优缺点
Strengths:
- This paper is well presented and well-written.
- The appendix provides clear supplementary material for better understanding.
- Therapeutic peptide generation is a challenging and novel area.
Weaknesses:
- The paper lacks significant novelty. The authors claim to use NELBO as a loss function, but this is not a new approach, as many prior works have already utilized similar loss functions, such as in Bartosh et al. (2023) [1]. MDLM itself was proposed last year in Sahoo et al. (2025) [2]. Furthermore, MCTS has been used for calculating rewards in discrete models since 2017, as demonstrated by Yu et al. (2017) [3], and there are several papers that have applied MCTS to diffusion models, such as Yoon et al. (2025) [4].
- The paper lacks essential comparisons with baseline models, such as Wang et al. (2024) [5]. Additionally, the authors did not provide the necessary resource code, which undermines the credibility of the model.
- The experimental results are incomplete and do not outperform existing models. The experiments only measure two statistical metrics (validity and uniqueness), and the novelty metric is not provided. Moreover, the results do not surpass those of Data and PepMDLM, indicating that the proposed method does not offer significant improvements over such models.
其他意见或建议
None
We are grateful for your insightful review.
Essential References Not Discussed:
We appreciate the reviewer’s suggestion to consider Wang et al. (2024). Although the title may suggest a close relation to our work, the scope, modeling assumptions, and core goals differ significantly. MMCD generates peptides composed of 20 natural amino acids only using a multi-modal approach. While MMCD uses the label “therapeutic peptide generation,” its evaluation is limited to docking scores and embedding-level metrics, without any testing of therapeutic properties such as solubility, hemolysis, or non-fouling. In contrast, our work directly targets real-world therapeutic peptide design by expanding the chemical space to include non-canonical residues and cyclic structures and by incorporating classifier-guided objectives tied to experimentally actionable properties. This direction aligns closely with the goals of the Application-Driven Machine Learning Track. Given these fundamental differences in scope and design goals, a direct comparison is not appropriate.
Addressing Weaknesses:
Weakness 1: We would like to push back on the assertion that our paper lacks significant novelty. Although NELBO is used in previous works, we derive a unique case of the NELBO given our novel bond-dependent masking schedule (See Appendix D.3). While the MDLM was already introduced by Sahoo et al., we make a novel extension with a bond-dependent masking schedule and derive a reverse posterior that differs from the general case in Sahoo et al. (See Appendix D.1-D.2). In addition, we introduce a novel invalidity loss that penalizes invalid peptide structure during training.
Finally, despite the use of MCTS in prior discrete models, our paper introduces the first application of MCTS to diffusion models, with our method being published in arXiv before Yu et al. (2025). In addition, our submission date to ICML was 9 Jan 2025, which was before the publication of Yu et al. (2025) on 11 Feb 2025.
Weakness 2: The baseline mentioned (Wang et al.) is designed to generate peptides consisting only of the 20 natural amino acids, while the core innovation of PepTune is to expand the search space to peptides with modified amino acids and cyclicizations which has not been achieved by prior models, which has been shown to enhance various therapeutic properties. These are two fundamentally different problems that are difficult to compare. Furthermore, the motivation behind the key components of our architecture, including our bond-dependent masking schedule and invalid penalty loss, stems from our unique goal of generating valid peptides from SMILES tokens despite the increased granularity.
Our source code is included in the arXiv version of the manuscript; however, given the anonymity of the review process, please refer to the anonymous repo link: https://anonymous.4open.science/r/peptune-86CB/README.md.
Weakness 3: We believe you may have misinterpreted the comparisons in Table 1. PepMDLM is our model trained with our bond-dependent masking schedule only without MCTS guidance. We make this comparison to demonstrate that our MCTS-guidance strategy does not significantly lower the diversity of sequences and increases validity through our iterative MCTS-guidance algorithm that rewards valid expansion steps.
Given that most of your concerns are regarding clarity and not the methodological and experimental soundness of our paper, and we have addressed your doubts on the novelty of our approach, we hope you consider raising your score in response to our answers. We will plan to introduce clarifications into the main text for the camera-ready version.
Thank you for the authors' response. They have clearly addressed most of my concerns:
-
their work was conducted earlier than that of Yu et al. (2025).
-
the source code has been provided, and the difference between their work and the baseline (Wang et al 2024) are clearly articulated.
-
the question regarding the "Novelty" metric (i.e., the proportion of generated novel samples that do not appear in the training set) remains unaddressed.
Anyway, as the authors have adequately resolved the first two issues, I am willing to raise my score.
We greatly appreciate your time in reviewing our paper.
Regarding the "Novelty" metric: given that there are 11 million peptides in our dataset, comparing against all sequences in the dataset is not feasible, so we report the similarity to nearest neighbor (SNN), which takes the maximum Tanimoto similarity score for each generated sequence against ~100K sequences in the training dataset (=1 for two identical sequences). We found an SNN of 0.513 for PepMDLM and 0.486 for PepTune (compared to 0.975 for HelmGPT), indicating that our generated sequences are novel and not represented in the dataset.
Thank you for your support!
This paper presents PepTune, a Masked Discrete Language Model (MDLM)-based generative framework for generating new peptides. Unlike existing approaches that struggle with multi-objective optimization or rely on continuous approximations, PepTune operates natively in a discrete sequence space while optimizing for properties such as binding affinity, membrane permeability, solubility, hemolysis, and non-fouling.
The key contributions of the paper are:
-
Introduction of Bond-Dependent Masking Schedule during training and inference.
-
Introduction of Invalid Loss Function: A penalty term that discourages the generation of invalid SMILES sequences.
-
Multi-Objective MCTS Guidance: Classifier-based rewards steer sequence generation to ensure Pareto-optimal solutions across conflicting therapeutic properties.
The paper uses several real case studies to demonstrate the efficacy of the proposed method.
给作者的问题
-
In equation 3, what is the meaning of ? The left multiplier is transposed, so it must at least be a vector, but what is ? Similary, what is the meaning of ?
-
What is in ?
-
Why is the validity percentage for PepTune much lower than HELM-GPT?
论据与证据
- Token-dependent masking and unmasking help in the generation of peptides.
There is not direct evidence supporting this claim in isolation. Idealy, PepMDLM should have been compared with vanialla MDLM for this.
- The proposed MCTS-based decoding procedure improves pareto optimality w.r.t multiple objectives.
This claim is supported using the results in Table 1 PepMDLM vs PepTune. However, it would be good to have comparisons to simple existing baselines for guided discrete diffusion for single objective guidance like Gruver et. al. (2023) and Nisonoff et. al. (2024), and them compare with them even for multi-objective guidance by making appropriate extensions.
方法与评估标准
Currently, the paper only presents results for two versions of their own method: PepMDLM and PepTune, where both methods only differ in the inference procedure. Since the paper uses multiple modifications for training as well as inference over vanilla MDLM, it is important to show ablations for each one of them.
- Comparing PepTune just with PepMDLM is not sufficient. Before going to the case studies, it would be good to have comparisons to simple existing baselines for guided discrete diffusion for single objective guidance like Gruver et. al. (2023) and Nisonoff et. al. (2024).
理论论述
I could not check the correctness of the proof for equation (3) because I had some trouble understanding the notation. Please see the questions and the comments section below.
实验设计与分析
Please see the "Claims and Evidence" section above.
补充材料
I reviewed the results (tables and figures) that are placed in the appendix and are referenced in the main paper.
与现有文献的关系
The paper adds a promising approach for scalable inference time guidance for discrete diffusion models using MCTS. To the best of my knowledge, the use of MCTS-like algorithm for sampling from a discrete diffusion model is novel and has not been explored before.
遗漏的重要参考文献
Please discuss how the state-dependent training objective used in this paper is different (if it is different) from Shi et. al. (2024).
其他优缺点
One of the major strengths of the paper are the real case studies. However, some sections of the paper are not as clear (see comments and questions). Moreover, the paper can be improved by including more baselines (see Claims & Evidence and Methods & Evaluation sections above).
其他意见或建议
-
Line 058, column 2: is used without specifying its meaning, and at this point in the paper it is unclear what is meant by "takes the vector encoding of the token ". Isn't bold-faced a sequence? There are a couple of missed words around the equation, which makes it an awkward read: "where represents the with ones..". Also, what is ? I'm assuming it is a fixed scalar greater than , but this should be mentioned.
-
In equation 3, what is the meaning of ? The left multiplier is transposed, so it must at least be a vector, but what is ? Similary, what is the meaning of ? While I could still understand equation 1, in spite of undefined symbols. I can't follow equation 3 at all.
-
The colors and fonts used in the main figure of the paper (Figure 1) make it extremely difficult to read. I'm unable to read the small fonts even after zooming in quite a bit.
-
I think there should be a on the left-hand side of equation 11.
Thank you for your detailed suggestions.
Claims and Evidence:
- We conduct an ablation study as suggested. From each model, we sampled 100 sequences of token length 100 and checked validity with our SMILES2PEPTIDE decoder. We show that both the bond-dependent masking and invalid loss improve the validity of peptides.
| Model | Fraction of Valid Peptides |
|---|---|
| PepMDLM | 0.40 |
| PepMDLM + No Bond Dependent Masking | 0.16 |
| PepMDLM + No Invalidity Loss | 0.21 |
- Due to the significant compute time needed to train additional baseline models on our 11 million peptide dataset, we decided it was not feasible to do so in this rebuttal period. Moreover, we note that Gruver et al. evaluate at most two objectives and require performing guidance updates in the latent embedding space based which can result in invalid conversion back to a discrete sequence, and Nisonoff et al. is limited to single-objective guidance. To the best of our knowledge, PepTune is the first framework for classifier-based multi-objective guidance for discrete diffusion, overcoming previous limitations via reward-based tree expansion and Pareto-aware backpropagation.
Essential References Not Discussed:
Shi et al. leverage a per-token parameter , which is trained in parallel with the discrete diffusion model. In contrast, we fix for tokens that exist within a peptide bond to ensure the model learns to reconstruct them earlier. This construction presents a practical extension of state-dependent masking to preserve peptide structure, which could be further explored for preserving various other structured data types. In addition, we rederive the state-dependent NELBO loss in a fundamentally different manner than Shi et al. (See Appendix D.3).
Other Comments or Suggestions:
We thank you for the thoughtful suggestions. We have made the notations and figures clearer upon your request in the manuscript for the camera-ready version.
- We notice the inconsistencies in notation that you refer to, but in Equations 1-4, we simplify the notation so that refers to the one-hot vector of the true token at an arbitrary position in the sequence. We have added clarification and removed the superscript in the other equations to maintain consistency.
- Yes, is a scalar greater than one. Please refer to the last sentence of Section 2.1, where the exact value is specified: “Empirically, we found that increased peptide validity while maintaining diversity across generated samples.”
- The sentence is fixed to “where is the vector with ones at indices of peptide bond tokens and zeroes in remaining indices.”
- We have increased the font size for clarity and fixed Equation 11 to include the .
Answers to Questions:
- In equation 3, the first is a one-hot vector is used to extract the probability associated with the unmasking the true token by taking an inner product. The second indicates that if the current token is a mask token, it has a probability of being unmasked to and a probability of remaining masked. For clarity, we have changed the transpose to an inner product:
The paper proposes a new model for novo generation and optimization of peptide. The model operates on SMILES and performs multi-objective-guided discrete diffusion; it can handle non-canonical amino acids and cyclic modifications. PepTune employs a bond-dependent masking schedule. To guide the diffusion, PepTune uses Monte Carlo Tree Search. The model is evaluated on different targets, e.g, generation of peptide binders for the GLP-1 receptor and bispecific binders for TfR and GLAST; PepTune produces peptides with better predicted properties compared to existing therapeutics.
给作者的问题
- if model trained on valid peptides, why would it sample invalid peptides ?
- which size of peptides can PepTune handle ?
- can PepTune be adapted to other types of molecules, e.g. proteins ?
- can the authors clarify what they mean with de novo ? it looks like the model would still need to have the targets to be in distribution for the affinity oracles to work; as such the model might not perform correctly under a more challenging train/test split with OOD test targets which is usually qualified as "de novo generation"; as such the current terminology can be misleading.
论据与证据
yes
方法与评估标准
yes
理论论述
n/a
实验设计与分析
yes
补充材料
quick pass over the supplementary; it is well detailed but a bit long for a standard conference paper.
与现有文献的关系
The paper introduces a sequence-based approach for peptide generation based on smiles and using a masking schedule that depends on the bonding to ensure that e.g., peptide bonds, side chains are preserved during the diffusion process; the model generates sequences that are more likely to be chemically valid and optimize multiple properties. As such few existing models can model non canonical amino acids and generate cyclic peptides.
遗漏的重要参考文献
the authors do not detail much the related work; it would be helpful to have a separate related work section that is more detailed than the current introduction paragraph. for example, the authors could discuss more how multi property optimization is done for other molecular tasks e.g. for proteins; Accelerating Bayesian Optimization for Biological Sequence Design with Denoising Autoencoders - Stanton et al.
其他优缺点
- The paper presents a new combination of masked diffusion with MCTS for multi-objective optimization in discrete sequence spaces. This addresses the challenges with guiding discrete diffusion processes without gradient estimation.
- The bond masking schedule is well suited to the peptide SMILES representation as it helps preserve structural elements of peptides.
- PepTune successfully optimizes multiple properties simultaneously.
- The paper includes extensive experiments and case studies, such as the generation of GLP-1R binders and bispecific peptides.
Weakness
- it seems like the generation process is not conditioned on the target, which if I understood correctly is only involved through the affinity oracles. as such the generation process can be wasteful as the generated molecules are not conditioned to the target of interest. Moreover, trained affinity oracles are known to generalize poorly out-of-distribution. As such, although target conditioned generation is possible, there might be several caveats with it.
其他意见或建议
see questions
We are very grateful for the reviewer’s thoughtful review.
Essential References Not Discussed:
We request that the reviewer refer to Appendix C, which provides an extensive analysis of prior discrete diffusion and classifier-based and classifier-free guidance methods. We also discuss at the end of Appendix C.3 the limitations of prior discrete guidance methods and challenges in multi-objective guidance.
Addressing Weaknessses:
To ensure that our model can scale to an arbitrary number of objectives, we leverage a classifier-based guidance approach which doesn’t take the protein sequence explicitly as input to the diffusion model but instead injects the target protein information at each denoising step by computing multi-objective rewards that are back-propagated through the MCTS tree. This means that unmasking steps that produce high-scoring peptides are explored further, limiting the “wasted” peptides. We also note that this exploration method enables us to generate a batch of Pareto-optimal sequences in a single run of the algorithm, which is much faster than generating only one sequence per run like prior models.
In addition, we show that our binding affinity and property classifiers generalize well on the validation data (Table 6 and Figure 14). We have to accept that only limited protein-peptide binding quantification data is available, and downstream in silico validation like VINA docking can be used to narrow down candidate peptides to be synthesized for experiments.
Answers to Questions:
-
PepMDLM is trained on SMILES tokens that decompose amino acids into smaller tokens that can be pieced together into valid amino acids during generation. While this enables us to represent a greater diversity of non-natural amino acids, even a single token generated in the wrong position can result in an invalid peptide sequence, resulting in a lower validity rate, which we calculate by decoding the SMILES sequences with regular expression matching. However, we note that our MCTS guidance strategy increases the validity rate to 100% due to its iterative unmasking process that is rewarded on high-scoring and valid peptides.
-
For the experiments, we generate peptides of token length 50 (7 amino acids) up to 200 (30 amino acids). This aligns with existing therapeutic peptides in clinical use, where the average length is about 20 amino acids or less [1].
-
Yes, our MCTS-guidance algorithm is a modular multi-objective guidance framework that can be adapted for any type of discrete data (e.g. protein sequences, promoter and enhancer DNA, RNA, small molecule drugs, natural language, etc.) and any pre-trained classifier in a plug-and-play fashion. Furthermore, our bond-dependent masking schedule can be extended to preserve structural components of other data types (e.g. functional motifs in proteins, punctuation and grammar in natural language, etc.).
-
In our work, we use “de novo” to refer to the generation of novel peptide sequences from scratch—i.e., not based on templates or fragments from known binders—and guided by pre-trained classifiers that predict binding affinity, solubility, and immunogenicity of these peptides. The model does not rely on known peptide backbones or motifs but instead generates sequences purely from noise through discrete diffusion sampling.
Furthermore, the target protein is not restricted to the training set of the classifier. Both the target protein and de novo peptide binder sequence are converted into feature-rich embeddings and passed through a classifier trained to understand binding patterns for accurate prediction. Our classifiers generalize well across unseen sequences, as shown by the high validation Spearman correlation coefficients (Table 6 and Figure 14).
We appreciate the reviewer’s thoughtful feedback and hope the clarifications provided demonstrate the uniqueness and adaptability of our approach. Our response highlights how PepTune addresses key limitations in prior work through a scalable, modular guidance framework, efficient multi-objective generation via MCTS, and strong generalization to arbitrary discrete data types and guidance objectives. Given our detailed answers, we hope you can consider raising your score to demonstrate continued support for our work.
[1] Dougherty, Patrick G et al. “Understanding Cell Penetration of Cyclic Peptides.” Chemical reviews vol. 119,17 (2019): 10241-10287. doi:10.1021/acs.chemrev.9b00008
The paper proposes PepTune, a discrete diffusion model that enables multi-objective guidance for generating and optimizing therapeutic peptide SMILES. PepTune adapts a bond-dependent masking schedule and global sequence invalid loss to improve discrete diffusion performance on peptide, and uses an MCTS-based strategy to guide the masked discrete diffusion generation and demonstrates good results for several therapeutic properties.
给作者的问题
- In Table 1, for PepTune, what are the rewards/properties optimized by the MCTS, and how are they related to the values reported in Table 1? They seem to be some general metrics that are irrelevant to the reward and thus cannot measure how well the rewards are optimized by the model.
论据与证据
Most claims in the submission are supported by convincing evidence. However, there are several gaps in the proposed method and experiments to support the novelty and performance of this paper. Please refer to details in "Methods and Evaluation Criteria" and "Experimental Designs or Analyses" sections.
方法与评估标准
There are two major components of the proposed method: 1. PepMDLM that modifies the MDLM method with prior knowledge and inductive bias of the peptide data distribution; 2. MCTS based guided sampling for multi-objective design. The datasets and evaluation criteria for PepMDLM look fine, while for guided sampling, only case studies are applied without a general rigorous evaluation. Also, commonly used criteria for Pareto optimization, eg, the hypervolume metric, should be calculated.
理论论述
There are no theoretical claims in this paper.
实验设计与分析
- For PepMDLM, an ablation study is missing for the importance of each component (i.e., bond-dependent masking schedule, bond-dependent diffusion loss, invalid peptide loss, etc.).
- In Table 5, the validity of HELM-GPT is much higher than PepMDLM. The author claims that validity is assessed differently. Could the author clarify this more and how this affects the validity comparison with HELM-GPT?
- The effectiveness of the guided sampling part is majorly demonstrated using case studies. How are these case study settings chosen? It is helpful if the authors could provide some evidence to support that these case studies offer a general and reliable evaluation for the proposed method that is applicable to a broad range of cases without cherry picking.
- For guided sampling, there is no baseline model other than the pretrained PepMDLM that is compared, for example the simplest best-of-N approach with N setting as a value comparable to the computational cost in the MCTS sampling, as well as guided sampling approaches, eg. [1].
- For each case study, the paper only shows the distribution of each objective individually. It would be helpful to draw the Pareto frontier of the generated peptides to better demonstrate how the model output balances different objectives and explore the Pareto frontier compared to other baselines.
- An analysis of the sampling cost of MCTS with respect to different hyperparameter choices and how the model performance scale with the sampling cost is helpful
[1] Paretoflow: Guided flows in multi-objective optimization. ICLR 2025.
补充材料
I review all parts of the supplementary material.
与现有文献的关系
The paper is related to: machine-learning wise, discrete diffusion models, guided sampling; and biological application wise, de novo peptide design.
遗漏的重要参考文献
-
Multi-objective guided flow models:
[1] Paretoflow: Guided flows in multi-objective optimization. ICLR 2025.
其他优缺点
The paper has good domain-specific designs in the method for peptide and evaluates the method in many real-world use cases, which have good potential to be applied in real-world peptide design problems. My main concern in this paper is the lack of baselines, benchmarks, and ablation studies in the evaluation of the method apart from the case studies.
其他意见或建议
NA
We thank the reviewer for the constructive feedback.
First, we would like to respectfully disagree with the assertion that the guidance strategy is inadequately evaluated. Given that the paper is on Application-Driven ML Track, we chose case studies that provide robust evaluation to prove efficacy in peptide-based therapeutic design. This is far more impactful and applicable to society and medicine than classical benchmarks. We ask that the reviewer reconsider their review of the paper with this in mind.
Addressing Concerns Experimental Designs Or Analyses:
- We conduct an ablation study as suggested. From each model, we sampled 100 sequences of token length 100 and checked validity with our SMILES2PEPTIDE decoder. We show that both the bond-dependent masking and invalid loss improve the validity of peptides.
| Model | Fraction of Valid Peptides |
|---|---|
| PepMDLM | 0.40 |
| PepMDLM + No Bond Dependent Masking | 0.16 |
| PepMDLM + No Invalidity Loss | 0.21 |
-
In HELM-GPT, the peptides are represented in HELM notation [1], where each token is a valid, synthesizable amino acid. In contrast, PepMDLM is trained on SMILES tokens that decompose amino acids into smaller tokens that can be pieced together into valid amino acids during generation. While this enables us to represent a greater diversity of non-natural amino acids, even a single token generated in the wrong position can result in an invalid peptide sequence, resulting in a lower validity rate. However, we note that our MCTS guidance strategy increases the validity rate to 100% due to its iterative unmasking process that is rewarded on high-scoring and valid peptides.
-
We selected three target classes to demonstrate PepTune’s robustness: (1) proteins with known binders to benchmark against known binders, (2) proteins without known binders, and (3) dual-target cases. This is a rigorous evaluation for application in therapeutic peptide design.
While previous guidance methods for discrete diffusion are limited to optimizing 1-2 objectives, we evaluate guidance across 4-5 non-trivial objectives that are highly relevant to therapeutic peptide design.
-
We appreciate the reviewer’s suggestion to conduct best-of-N comparison. However, we believe that our comparisons between PepTune and PepMDLM are sufficient in demonstrating the efficacy of our guidance strategy for the following reasons:
- In Pareto optimization, scalarizing or ranking sequences across conflicting objectives without access to the Pareto frontier leads to suboptimal trade-offs, making it difficult to rank best-of-N.
- Despite exploring more sequences, PepTune achieves comparable runtime by reusing high-reward unmasking paths within the MCTS tree. Rather than starting from scratch, PepTune expands only unexplored nodes, reducing redundant computation. Empirically, we find similar runtimes for generating 100 valid sequences using PepMDLM (403 sec) and PepTune (347 sec for top 100 sequences with and ), while PepTune yields substantially higher property scores (Figures 4–9).
-
We also appreciate the suggestion to conduct a benchmark against ParetoFlow. However, ParetoFlow is designed for continuous flow-matching models, whereas our approach is designed specifically for discrete diffusion. Due to the fundamental differences in model class and data type, a direct comparison would not be meaningful; to our knowledge, PepTune is the first to apply Pareto-aware guidance for discrete diffusion, and we hope this novelty is considered.
-
Given that our method supports robust guidance beyond 2-3 objectives, it is impossible to visualize the high-dimensional Pareto frontier. Instead, our graphs demonstrate that our approach simultaneously raises the scores across 5 objectives (see Figure 9) to a plateau, indicating that no further optimization is possible without sacrificing another objective.
-
Each MCTS run explores sequences. The sampling cost scales linearly with respect to both of these parameters. We compared and found similar guidance curves and set constant.
Responses to Questions
- In this experiment, we optimized 5 properties: binding affinity to GFAP, membrane permeability, solubility, hemolysis, and non-fouling, all of which are not related to the values in Table 1. In Table 1, we aim to show that our MCTS-guidance algorithm does not significantly lower the diversity of sequences and increases validity through our iterative MCTS-guidance algorithm that rewards valid expansion steps.
We thank you for your suggestions and hope you reconsider raising your score in light of these responses and the broader goals of the Application-Driven ML Track.
[1] Zhang et al. Helm: A hierarchical notation language for complex biomolecule structure representation. http://dx.doi.org/10.1021/ci3001925
The authors address most of my concerns, and I will increase my score.
Regarding Best-of-N, I understand the difficulty in ranking samples, but I suggest the authors try an alternative by dropping samples that are Pareto inferior to some other sample and report the number of samples left and their performance on different properties. This will give a much better idea of how a naive approach would work for the multi-objective optimization than only comparing with PepMDLM, which has no access to the target property.
Besides, in addition to the running time, I would suggest adding a table of the NFE (number of functions evaluated) for PepMDLM, PepTune, and the naive Best-of-N like baseline mentioned above.
We deeply appreciate your detailed suggestions, and we will try our best to include the best-of-N and NFE table in our camera-ready version. Thank you for your support!
Every reviewer agrees on the acceptance due to its methodological and application contribution. I also recommend its acceptance.
This is a suggestion, but it appears that this work did not cite several highly related works, such as guidance in discrete diffusion models or methods [a] that can easily be adapted to a discrete diffusion model [b]. While the focus there is not on multi-objective optimization, it proposed a searchish based method in discrete diffusion models. I believe it should be cited in the final version.
[a] LI, Xiner, et al. Derivative-free guidance in continuous and discrete diffusion models with soft value-based decoding. arXiv preprint arXiv:2408.08252, 2024.
[b] WU, Luhuan, et al. Practical and asymptotically exact conditional sampling in diffusion models. Advances in Neural Information Processing Systems, 2023, 36: 31372-31403.