PaperHub
7.0
/10
Poster3 位审稿人
最低6最高8标准差0.8
6
7
8
4.0
置信度
正确性3.7
贡献度3.3
表达3.3
NeurIPS 2024

Designing Cell-Type-Specific Promoter Sequences Using Conservative Model-Based Optimization

OpenReviewPDF
提交: 2024-05-15更新: 2025-01-16
TL;DR

We propose a workflow to design cell-type-specific promoters while accounting for various practical considerations and demonstrate its efficacy in a difficult setting.

摘要

关键词
ML applicationscomputational genomicscomputational biologymodel-based optimization

评审与讨论

审稿意见
6

Building on existing models, the authors used MBO to design promoters in a data-efficient manner, with a particular focus on discovering promoters for similar cell types. This approach was tested on three relatively similar blood cancer cell lines, demonstrating that the method successfully identified numerous new cell type-specific promoters after experimentally validating the designed sequences.

优点

A design process is proposed, building on existing models, which uses MBO to design promoters in a data-efficient manner. This approach has some impact on promoter optimization.

缺点

The authors did not focus much on improving the design of the model itself, instead relying on existing models. The innovation in the model is limited, and the experimental evaluation is not comprehensive, as it did not assess a wider range of biological sequence design algorithms.

问题

  1. The model lacks innovation and primarily relies on the application of existing models.
  2. The design process proposed by the author is roughly similar to the existing biological sequence design algorithm process, and the screening based on model uncertainty is also mentioned in GflowNets. GflowNets also has strategies to increase diversity and optimize properties. The author's contribution is more of a simple combination of current method processes. What is the author's innovation?
  3. If the author's contribution is to propose a workflow, then should they use more extensive sequence design models in the workflow, rather than just using COMs, to prove the effectiveness of the design algorithm?
  4. In terms of experimental evaluation, the authors should include more comparative models. Given the many existing biological sequence design algorithms, the authors need to demonstrate the advantages of their algorithm.

局限性

yes

评论

References:

  • Reddy, Aniketh Janardhan, et al. "Strategies for effectively modelling promoter-driven gene expression using transfer learning." bioRxiv (2024).
  • Trabucco, Brandon, et al. "Conservative objective models for effective offline model-based optimization." International Conference on Machine Learning. PMLR, 2021.
  • Jain, Moksh, et al. "Biological sequence design with gflownets." International Conference on Machine Learning. PMLR, 2022.
  • Linder, Johannes, et al. "A generative neural network for maximizing fitness and diversity of synthetic DNA and protein sequences." Cell systems 11.1 (2020): 49-62.
作者回复

We thank the reviewer for their comments and address their concerns below. First, we clarify that the novel contribution of our work is our cell-type-specific promoter design workflow. Although we use existing modeling strategies and MBO algorithms, we combine them in a novel way to tackle a difficult but important problem – an accepted axis of novelty at NeurIPS as per the guidelines. Unlike existing approaches, our elaborate workflow accounts for many practical considerations such as data-efficiency, avoiding adversarial sequences, and obtaining diverse sequences (PDF attached to general rebuttal illustrates it). And unlike most other offline MBO works that rely on computational evaluations, which are often unrepresentative of real-world experiments, we validate our workflow using expensive and time-consuming wet lab experiments. Next, we also explain the benefits of our workflow over GFlowNets, which are generally harder to train and tune, do not account for adversarial examples, and have not been validated using real-world experiments. Finally, we justify the limited number of benchmarks, as the cost of running wet lab experiments is extremely high and our budget did not allow us to benchmark more methods. We hope that the reviewer can reconsider their assessment based on our responses.

  1. Our contributions: We would like to clarify that the specific models used in our work are not the main contribution. Instead, the focus of this paper is to propose a novel workflow for discovering cell-type-specific promoters using offline MBO that accounts for various practical considerations. Our workflow has the following advantages over existing promoter design workflows (detailed in the Related Work section):

    • Compared to traditional workflows that use heuristics and manual curation, our workflow is automated and generalizable and can be used to design promoters for even relatively understudied cell types.
    • We improve on existing MBO-based workflows in three main ways. First, we use the transfer learning strategies identified by Reddy et al. (2024) to train models in a data-efficient manner - existing workflows do not employ transfer learning and rely on large training datasets that are expensive to collect. Second, we use Trabucco et al. (2021)'s COMs framework to mitigate adversarial designs - existing workflows use simple optimization techniques (e.g. vanilla gradient ascent) that are more prone to adversarial designs. Finally, as experiments are expensive and a limited number of sequences can be tested, we propose a final sequence selection step to choose a small subset of diverse yet desirable sequences for testing from a large pool of candidate designs – this step is also missing in existing workflows.

    Although we use the modeling strategies identified by Reddy et al. (2024) and the COMs framework, we combine them to solve an important problem in a novel way – an accepted axis of novelty at NeurIPS as per the guidelines. We also add a novel sequence selection step to balance diversity, optimality, and uncertainty. This novel workflow has the potential to improve gene therapies and should be of interest to many ML researchers working on synthetic biology problems such as promoter design. Moreover, we validate the utility of our workflow by performing real-world wet lab experiments - most MBO methods are never validated using wet lab experiments. We show that our workflow is effective and outperforms both a traditional design method (motif tiling) and an offline-MBO-based method (deep exploration networks - DENs) that had previously been validated using wet lab experiments. These points support the novelty and significance of our work.

  2. GFlowNets are harder to train, do not account for adversarial examples and have not been validated using real-world wet lab experiments: Our workflow improves on existing promoter design workflows in many ways, especially by accounting for practical considerations such as data-efficiency, adversarial designs, diversity, and uncertainty. The reviewer mentions that GFlowNets also try to optimize a given property while retaining diversity. However, since GFlowNets are generative models, they require specialized architectures that are significantly harder to tune and train vs. training a discriminative model as in our workflow. They also do not account for adversarial designs. Most importantly, GFlowNets have not been validated using real-world wet lab experiments to the best of our knowledge (Jain et al., 2022). DENs are also generative models that are aimed at producing diverse desirable designs and have been validated using real-world wet lab experiments (Linder et al., 2020), which is why we chose DENs as a comparator to benchmark our approach. Figure 3 shows that our workflow outperforms DENs.

  3. High experimental costs limit our ability to benchmark more methods: Most existing MBO methods for biological sequence design do not validate their designs using wet lab experiments – instead, their evaluations are based on computational oracles whose predictions are not always reflective of real measurements. In our wet lab experiments, we show that our workflow is indeed capable of designing cell-type-specific promoters. While we benchmark against DENs and motif tiling, the reason for not benchmarking more design methods is the extremely high cost of performing these experiments. They are also time consuming, and took us ~6 months to complete. Since our budget limited the total number of sequences that could be tested, and since we require a significant number of designs from every method to thoroughly validate it, we chose to evaluate a representative set of methods. In particular, we chose motif tiling as representative of widely-used traditional methods, and we chose DENs as the only MBO algorithm that has been evaluated for biological sequence design using wet lab experiments. Our workflow outperforms both of these methods.

评论

Thank you for addressing my concerns and providing detailed explanations in your rebuttal. As a result, I have raised my score.

评论

Thank you for increasing your score! We’re glad to hear that we've successfully addressed your concerns.

审稿意见
7

The paper outlines a workflow to designing promoter sequences that are specific to cell types, especially closely related cell types. The proposed approach consists of five steps:

  1. Pretraining a model on existing massively parallel reporter assay (MPRA) datasets, which are large but restricted to a few well-studies cell types.
  2. Fine-tuning the pretrained model on experimental data with the aid of a conservative regularizer.
  3. A gradient ascent-based approach to design sequences that have high predicted differential expression (DE).
  4. Selection of smaller subset that is both optimal and diverse.
  5. Experimental verification of designed promoter sequences. The authors compare the performance of the proposed workflow to two existing approaches – motif tiling and deep exploration networks (DENs) on experimental data on both optimality and diversity metrics. Their results show instances where their workflow either outperforms existing methods in optimality or yields more diverse promoter sequences than DENs.

优点

The paper is very clear with an intuitive flow. It begins with a description of the desiderata for any workflow for designing promoter sequences, followed by a summary of the related work and a series of preliminaries to acquaint the reader with the objectives of the problem being considered. I also appreciate the references to related works and motivations for the design decisions in Section 4. This delineates the differences in the proposed workflow compared to existing works. The authors are also explicit with the details of their model architectures, training and experiments. This is readily visible from the great level of detail in the relevant sections of the appendix, as well as a clear statement of their objective function and its components (see Eq. 3 and 4).

Besides the writing, the inclusion of a wet lab experimental evaluation of the designed promoter sequences lends a lot of credibility to their work as it exhibits the practical utility of the proposed workflow. Suitable baselines have also been chosen for comparison of performance, and consequently make substantive claims to the utility of the proposed workflow.

缺点

The paper has no glaring weaknesses. However, there are some places where the authors have made statements that are confusing or unsubstantiated.

  1. The claim of data efficiency over small PE datasets is not substantiated. It appears that the authors accomplish this via fine-tuning over the smaller dataset. However, they take their cue for this from existing offline MBO algorithms, so this likely cannot be claimed as a key benefit to their proposed workflow.
  2. The outlined workflow uses an ensemble of fine-tuned models (without the conservative regularizer) to compute a pessimistic estimate of the DE of a given sequence and cell pair. Why should we expect such an ensemble to estimate a DE that is consistently lower than the true PE? Additionally, this ensemble comprises models with slightly different architectures, but no further specifics are given as to what these architectures are and how they are chosen?
  3. The selection algorithm outlined in Section 4.3 appears to be greedily optimizing for an objective that balances optimality and diversity. However, the authors then state on line 339 that they set the diversity coefficient (β\beta) to 0 as the workflow naturally designed diverse promoter sequences. This renders most of the discussion between lines 274 and 294 moot to their workflow.
  4. In the conservative regularizer, the authors penalize high predictions for both unseen and potentially undesirable sequences, summarized as μ(x)\mu(\mathbf{x}). However, it is not clear how the formulation in Eq. 2 handles unseen sequences as the second term ExD[f_θ(x)]\mathbb{E}_{\mathbf{x}\sim D} \left[ f\_{\theta} (\mathbf{x}) \right] is restricted to sequences seen in the dataset(s).

问题

Questions:

  1. What is the “linear probing” in line 179 referring to?
  2. Doesn’t the conservative regularizer limit the method’s ability to “discover” new promoter sequences? Is it possible that there are promoter sequences that are unlikely to directly evolve from the sequences in the fine-tuning dataset?
  3. Why are 6-mer frequency vectors used to compute the Euclidean distance K\mathcal{K}?
  4. Why should the base pair entropy of a diverse sequence set be 2? Based on the definition used in the paper, should this not be ln(4)ln(4)?
  5. Why are different metrics used to compare the DEs of the proposed workflow against motif tiling (Figure 2) and DENs (Figure 3)?

Suggestions:

  1. On line 142, it is not immediately clear if the “some objective function” being referred to is different from the true objective function. While this becomes clearer later in the paper, a rewording of this line may improve clarity.
  2. Points 1 and 2 in Section 4.1 could be merged as the primary barrier to using Enformer is the cost of retraining it to suit the user’s setting.
  3. The titles in Figure 2 are difficult to read. Moving them to the caption of each sub-figure may improve readability.
  4. In Figure 3, there is a visible improvement in DE only for Jurkat cells. Are there any other reasons for why the proposed workflow is preferable to DENs for the other cell types (computational savings perhaps because of the use of a discriminative model in the workflow vs generative models in the DEN)? A restatement of these reasons will strengthen the case for the proposed workflow.
  5. The authors could be more concise. For instance, the information in the section from line 98 to line 106. The information here is largely similar to that contained in the outline of the workflow at the end of the introduction (Section 1). Similarly, the discussion on the diversity-based selection (starting on line 274) serves no role in their final algorithm as β\beta is set to zero. Being more concise may enable the authors to include more relevant details and improve the clarity of their writing.

局限性

The authors have made a good faith effort to acknowledge the limitations of their work in Section 6. I believe that their method is also limited in discovering novel promoter sequences due to the use of the conservative regularizer. Since their design method starts from sequences in the finetuning data to evolve the design sequences, there is likely a limit on how divergent they are from the seen data. This is somewhat also borne out by the poorer performance of their workflow on THP1 cells, which are scarcer in their training data.

评论

Regarding the reviewer's suggestions

We appreciate the reviewer’s suggestions for improving the presentation and readability of the paper, and we will incorporate them in the final version of the manuscript.

Regarding figure 3, although it shows a visible improvement in mean DE only for Jurkat cells when using gradient ascent vs. DENs, the difference in DE is significant also for K562 cells. Moreover, we see that gradient ascent produces many more sequences with high DE when compared to DENs, while also producing generally more diverse sequences (Table 1). Finally, as the reviewer noted, DENs are more computationally intensive and more difficult to tune vs. gradient ascent, since they require training separate generative models in addition to training the design models. We will add a more detailed discussion of these reasons for choosing gradient ascent over DENs in the final version.

References:

  • Reddy, Aniketh Janardhan, et al. "Strategies for effectively modelling promoter-driven gene expression using transfer learning." bioRxiv (2024).
  • Trabucco, Brandon, et al. "Conservative objective models for effective offline model-based optimization." International Conference on Machine Learning. PMLR, 2021.
  • Zheng, An, et al. "Deep neural networks identify context-specific determinants of transcription factor binding affinity." BioRxiv (2020): 2020-02.
  • Hutchinson, Gordon B. "The prediction of vertebrate promoter regions using differential hexamer frequency analysis." Bioinformatics 12.5 (1996): 391-398.
评论

Answering the reviewer's questions:

  1. Linear probing: As an alternative to fine-tuning a pretrained network, which involves updating all of the weights in the network, one can extract embeddings or outputs from the pretrained network and fit a simple linear model to make predictions for a downstream task without updating the full pretrained network. This approach is called linear probing. We will clarify this point in the final version.

  2. Certain promoters could be undiscoverable but our analysis shows that designs are representative of diverse sequence spaces that are also quite distinct from those containing the training set: The conservative regularizer mitigates adversarial designs, but as the reviewer noted, this regularization could prevent the discovery of some desirable promoters that are very different from the promoters in the training set. Discovering these promoters using MBO requires a model to be accurate in the sequence space around them. However, given a limited training set, it is difficult for a design model to generalize to such a sequence space (distribution shift problem). Therefore, it is difficult to discover such promoters even in the absence of the conservative regularizer. Since we have a limited experimental budget, the focus of our workflow is to make the best use of the available training data to design sequences while avoiding adversarial sequences.

    As the reviewer noted, it is also possible that certain promoters are harder to directly evolve from the sequences in the fine-tuning dataset; however, this problem is most pronounced when the fine-tuning dataset is not very diverse. The fine-tuning dataset we use in our experiments is diverse, and from table 1 we see that the resulting designs are also very diverse (with any pair of sequences differing by close to 180 bp on average), indicating that many parts of the sequence space are being represented in the designs. Moreover, table 1 in the document attached to the general rebuttal shows that on average, the designs differ from the most similar training set sequences by 125-140 bp, indicating that our workflow produces sequences that are quite different from the training set.

    We will mention these possible limitations in the final version and open this as an avenue for future work.

  3. We use 6-mer frequency vectors to compute the Euclidean distance K\mathcal{K} so as to capture the frequencies of potential transcription factor (TF) binding motifs that are around 6-12 bp long: TFs are proteins that bind to promoters to facilitate gene expression, each binding to specific DNA substrings called motifs. Most motifs are around 6-12 bp long (Zheng et al., 2020). By computing 6-mer frequencies, we can approximate the frequencies of different motifs in a given sequence. Then, choosing promoters with distinct frequency vectors (separated by a long Euclidean distance) yields a diverse set of promoters with unique regulatory landscapes, enhancing the likelihood of identifying cell-type-specific promoters. While higher k-mer frequencies (e.g., 7-mer, 8-mer) could be used, they would result in more distinct k-mers and sparser frequency vectors, making it harder to meaningfully differentiate between promoters. Thus, we opt for 6-mer frequency vectors to balance the need to capture TF-binding motifs and maintain sufficiently dense frequency vectors. Notably, hexamer statistics have historically been used to identify promoter regions (Hutchinson, 1996).

  4. Base pair entropy is computed with base 2 logarithms: Therefore, a set of purely random sequences will have a base pair entropy of log24=2\log_2 4 = 2. We failed to mention the base in our manuscript but will clarify this in the final version.

  5. Since DENs are generative models, the metrics in figure 2 cannot be computed for them, leading us to use the metrics in figure 3: The metrics shown in figure 2 compare each designed sequence to the starting sequence from the fine-tuning dataset that was used to generate it. These results illustrate that our workflow significantly improves most starting sequences from the fine-tuning dataset, while motif tiling mostly reduces the DE of starting sequences. Since DENs are generative models, they do not take in a starting sequence – instead, they directly generate a pool of design sequences using random noise vectors as inputs. Thus, we cannot use the same metrics as in figure 2 to analyze the performance of DENs vs. gradient ascent. In figure 3, therefore, we directly compare the DE of sequences produced using DENs to those produced using gradient ascent.

作者回复

We thank the reviewer for their detailed comments and for recognizing the significance of our work. In the responses that follow, we clarify that our data-efficiency claim is due to the pretraining step which is missing in previous promoter design workflows – we will make this clear in the final version of our manuscript. Next, we justify why we expect the pessimistic estimate of DE to underestimate the true DE due to its definition as the lower confidence bound (LCB) of the ensemble’s constituent models’ predictions. We point to more details about the ensemble’s architecture and explain our architectural choices – our explanation will be added to the final manuscript. Then, we explain that we retain the diversity component of our final selection algorithm to allow a practitioner to boost diversity if necessary. Using an additional analysis and results from Table 1, we show that our designs represent diverse regions of the sequence space that are also distant from the space of training sequences, but we acknowledge that certain desirable promoters will be undiscoverable – we will discuss this more in the final manuscript. We also address the other weaknesses and questions mentioned by the reviewer. We hope that our responses make a stronger case for the acceptance of our paper.

Addressing weaknesses identified by the reviewer:

  1. Our workflow’s data-efficiency comes from the pretraining step that is absent in existing promoter design workflows: The main reason we claim data-efficiency is indeed due to the pretraining step (lines 71-74 and Section 4.1). Our modeling strategies are derived from Reddy et al. (2024), who showed that pretraining on related genomic datasets can boost downstream performance when the pretrained model is fine-tuned using a small PE dataset. As we point out in the related work section, existing MBO-based promoter design workflows are reliant on large MPRA datasets and train their design models exclusively on these large datasets. By adopting the transfer learning strategies identified by Reddy et al. (2024) in our workflow, we can obtain more accurate design models using the same amount of data from target cells when compared to existing workflows that do not leverage transfer learning, thereby supporting our data-efficiency claim. We will make this point clearer in the final version of our manuscript.

  2. Since the pessimistic estimate of DE is computed as the lower confidence bound (LCB) of the ensemble’s constituent models’ predictions (lines 269-270), we expect it to underestimate the true DE if the ensemble is accurate: If we have reasonably accurate models of DE, the mean prediction from the ensemble should be an accurate estimate of the true DE. Since the LCB is computed as the mean prediction minus the standard deviation across predictions, this pessimistic estimate should underestimate the true DE, provided the ensemble is reasonably accurate.

    Ensemble model architectures: Details about the ensemble’s constituent models are in Section D.2 of the appendix. Since it is expensive to pretrain many different architectures, we use the same pretrained backbone in all constituent models while using different output layer architectures during fine-tuning. In particular, we vary the depth and number of hidden units of the output layers. We also vary the activation functions, as using different activation functions in the constituent models’ output layers allows us to approximate the PE function using different piecewise functions, thereby improving the ensemble’s overall robustness.

  3. Diversity component of final selection algorithm is retained to have the flexibility to boost diversity if necessary: While we set the diversity coefficient to zero for sequences designed using gradient ascent, we set it to 10 when using DENs (line 351), as the top designed sequences produced by DENs were not as diverse in our preliminary experiments as those produced using gradient ascent. We retained this ability to modulate diversity in our workflow for the sake of flexibility – if a practitioner observes that the top designed sequences are very similar to one another, the diversity coefficient can be increased to avoid selecting highly similar sequences in the final sequence set.

  4. The conservative regularizer only penalizes potentially undesirable (i.e. potentially adversarial) sequences among those not seen in the training set – it does not penalize all unseen sequences: We will improve the clarity of this description in line 217. The second term in Eqn 2 prevents uniform underestimation of the function being modeled across all sequences by maximizing the expected value of the function over the training set (see Trabucco et al. (2021) that presents the COMs framework for more details).

评论

Thank you for the detailed response! I believe you have addressed most of my concerns and I have revised my score to reflect this.

For point 4 (on unseen sequences), I suggest you edit the text in the paper to clarify the same point.

评论

Thank you for increasing your score! We are happy to know that we have addressed your concerns and we'll be sure to clarify the point on unseen sequences in the final manuscript.

审稿意见
8

The paper presents a comprehensive guide for designing cell-type-specific promoter sequences using a conservative model-based optimization (MBO) approach. The primary goal is to develop promoters that drive gene expression specifically in target cells while minimizing off-target effects in closely related cell types. The authors propose a detailed workflow that incorporates data efficiency, sequence diversity, and model uncertainty. This method is validated through empirical experiments on blood cancer cell lines, demonstrating its effectiveness compared to traditional and simpler optimization methods.

优点

The manuscript is well-written, and it is easy to follow. The problem of promotor design is important and it has driven increasing interest. The authors here present a comprehensive work and a practical guide for effectively designing the cell type specific promoter. The wet lab experiment further strengthens this paper.

缺点

  1. I believe many generative-based models are specifically designed for DNA-sequence design, e.g., the genetic algorithm, simulated annealing, and other RL-based models. Why do the authors only focus on "gradient ascent"? It may introduce more adversarial effects compared with the other methods.
  2. Lack of sufficient figures to better visualize the proposed workflow and practical guide.

问题

  1. How do we balance diversity and conservative regularization? They seem to be contradictory.
  2. Could the authors add decision-tree-like guideline figures at the end of the paper? It will help biologists better understand the workflow.

局限性

The authors have addressed the limitations.

作者回复

We thank the reviewer for their comments and for recognizing the significance of our work. In our responses below, we first justify the use of gradient ascent vs. other optimizers such as genetic algorithms and simulated annealing, due to its computational efficiency. Then, we clarify that our workflow can be used to produce diverse designs, even when conservative regularization is used, due to our use of multiple design models that are trained using different regularization strengths, as well as our final sequence selection algorithm that balances optimality and diversity to select the final set of designs for wet lab validation. Finally, to better illustrate our workflow in the final version of the paper, we will be adding a decision-tree-like figure as suggested by the reviewer - a draft of this figure is in the document attached to the general rebuttal.

  1. Gradient ascent is used instead of other optimizers due to its computational efficiency, and the conservative objective mitigates adversarial effects: While discrete optimizers such as genetic algorithms and simulated annealing are suited for making discrete mutations to a DNA sequence during optimization, it can be computationally expensive (requiring many samples and filtering) to use them in a large search space such as ours (250 bp sequence = 42504^{250} possible sequences), since these algorithms do not use the function’s derivative to take a step towards an optimum. Similarly, we did not experiment with RL-based models for DNA-sequence design since it is difficult to integrate them with the conservative objective models framework, which requires a fast optimizer during training to mine for potentially adversarial sequences. On the other hand, gradient ascent is computationally efficient and allows us to quickly discover an optimum. Furthermore, in preliminary experiments, we experimented with adding discrete mutations during training and sequence optimization in addition to performing gradient ascent, but we did not observe a significant improvement in sequence quality vs. just using gradient ascent (as measured by design model-predicted differential expression). Overall, simplicity and efficiency considerations push us to use gradient ascent instead of other discrete optimizers.

    As the reviewer points out, it is true that gradient ascent algorithms can produce adversarial effects. However, we utilize a conservative objective to mitigate them, as described in Section 4.2 in the paper. The other optimizers are also not immune to producing adversarial effects – if run for a sufficient number of steps, any of the aforementioned algorithms can produce adversarial sequences since the designed sequences can be arbitrarily different from the training set.

  2. Collecting sequences from design models trained using different conservative regularization levels, and using our final sequence selection algorithm produces a diverse yet desirable set of final designs: As the reviewer notes, strong levels of conservative regularization could hinder the diversity of designed sequences. In order to maintain a high level of diversity, we obtain a large set of candidate sequences using many different levels of regularization as mentioned in lines 240-242 (controlled by the α\alpha parameter in Eqn 3). Then, we use the algorithm in Section 4.3 to choose a subset of sequences for wet lab validation that are diverse yet desirable. This combination of multiple regularization levels and the final sequence selection algorithm yields a diverse final sequence set as illustrated in Table 1, overcoming potential issues of an over-regularized model producing a narrow distribution of candidate sequences. That said, we agree that fundamentally, conservatism (to stay close to the data) and diversity (going beyond the data in several ways) are at odds with each other more broadly from an algorithmic perspective, and studying these topics is a good avenue for future work, even for general ML, outside of biological problems, as we will discuss in the final version of paper.

  3. We will add additional figures that illustrate our guidelines to the final version of the paper. A draft decision-tree-like figure is in the document attached to the general rebuttal. We will refine this figure for the final manuscript.

评论

The authors have addressed all my concerns. I will keep my score unchanged.

评论

Thank you! We're happy to know that we've addressed your concerns!

作者回复

In the attached document, we provide an additional table that shows the average Hamming distance to the closest training set sequence for designs from every method. This table illustrates that designs from our workflow are quite distinct from the training set sequences, differing by 125-140 bp on average. We also provide a draft figure that details the various steps of our workflow.

最终决定

The tools and guides developed in this submission were unanimously viewed as a timely, relevant, and well executed contribution to promoter sequence design. The clarifications provided in the rebuttal further helped clarify the contribution and how it compares to existing work. As such, I agree with all three reviewers in that it should be published in NeurIPS and encourage the authors to include their main arguments from the rebuttal into the camera ready version of the paper.