PaperHub
6.1
/10
Poster4 位审稿人
最低3最高4标准差0.4
3
3
4
3
ICML 2025

Stochastic Encodings for Active Feature Acquisition

OpenReviewPDF
提交: 2025-01-24更新: 2025-08-07
TL;DR

We present a new method using stochastic encoders for Active Feature Acquisition, the test time task of iteratively observing features to improve the current prediction.

摘要

关键词
Active Feature AcquisitionDynamic Feature Selection

评审与讨论

审稿意见
3

Considering the training challenges of reinforcement-learning approaches and the myopic shortcomings of conditional mutual information (CMI) strategies, this paper proposed a method called Stochastic Encodings for Active Feature Acquisition (SEFA). By encoding features into a regularized, stochastic latent space, it can better predict future observations; it then scores each unobserved feature using a gradient-based measure of how it would disambiguate among likely classes. This paper tested their method on standard tabular/image data, and cancer gene-expression classification.

给作者的问题

Figures like Figure 4 are really hard to read.

论据与证据

Yes.

方法与评估标准

The paper’s benchmark datasets and evaluation setup from synthetic datasets designed to showcase the pitfalls of myopic conditional mutual information to tabular, image, and real-world medical tasks to demonstrate both practical applicability.

理论论述

These formal statements (Propositions 4.1 and 4.2) are demonstrated on that indicator toy example. Seems that the authors use it as a focused scenario to prove where purely myopic strategies fall short and how considering possible unobserved values can address that limitation. They do not provide general proofs.

实验设计与分析

The experimental settings, including datasets, baselines and evaluation metrics are standard.

补充材料

I reviewed Appendix C-H.

与现有文献的关系

SEFA draws inspiration from several established concepts and methods, like RL, CMI. SEFA goes beyond these existing methods by incorporating a stochastic, latent-space perspective that is regularized via an information bottleneck. It introduces a gradient-based sensitivity strategy in which each feature owns a block of the latent representation.

遗漏的重要参考文献

N/A

其他优缺点

Strengths: The method has been tested across synthetic data, real tabular, image, and cancer genomics datasets.

Weaknesses: The method’s factorized latent encoder implicitly assumes feature independence in the unobserved blocks, however, considering some strong feature interdependencies situations, this method might not work well.

其他意见或建议

N/A

作者回复

Thank you for the positive review, we are grateful for the feedback, we answer your questions below. All changes below will be added.

Theory

We agree that the theory provided is specific to the indicator and does not provide general statements about bounds on SEFA's performance. The purpose was not to provide fully general proofs, but insight into why CMI maximization can fail (it does not take into account possible future feature observations).

Purpose of Proposition 4.2: Proposition 4.1 gives a necessary condition for optimality, proposition 4.2 shows that it can be sufficient, we include 4.2 because only necessity without showing sufficiency is a weaker result. The connection to SEFA is to use the idea of considering possible future observations in the model design (the expectation in the latent space), since this meets a necessary condition for optimality. We verify this empirically in our ablations (Table 1 and Table 3 Deterministic Encoder). We do not use the objective from 4.2 because it is intractable and does not fit into our latent set up.

Generalizing Beyond the Indicator: Proposition 4.1 and 4.2 can be generalized beyond the indicator: CMI will be suboptimal where there are features that are jointly informative but individually non-informative. By definition, the CMI of the jointly informative features is zero, so any other features that are individually informative (even if they are very noisy) will be selected first by CMI. In contrast, the objective in 4.2 can capture the effect of jointly informative features, since it considers the possible values of the unobserved features.

Feature Independence

There are two reasons why we believe our conclusion overstates this as a limitation:

  1. Whilst the encoders do not model the feature interdependence, this can be accounted for by the predictor network p_ϕ(yz)p\_{\phi}(y|\mathbf{z}). The complex interdependencies are captured by the layers after the latent space. Existing work uses this principle, for example, β\beta-VAE [Higgins et al. 2017] has disentangled latent spaces where the complex relationships are handled by the decoder. The key point is that we are not modelling distributions of the features themselves, but of their representation - their effects on the predictor network. The predictor can learn to act differently based on other latent values, for example, using samples of feature 2's latent components differently depending on if latent component 1 is positive or not. Whilst the distribution of feature 2's latent components cannot be affected by observing feature 1, how the latent samples are used by the predictor can be affected. As a result the gradients are affected, and therefore the acquisition scores. This is another reason to run the acquisition in the latent space. The predictor network can learn the complex relationships (provided it trains on enough latent samples), allowing the encoder to have a structure that does not learn the complex feature distributions. Whereas if we ran the acquisition in feature space, the predictor would only be trained on real feature values, and so the generative model would need to be able to learn the complex multivariate feature distributions.

  2. Empirically we see this. The Feature Space Ablation (Table 1 and Table 3), uses a generative model for p(x_Ux_O)p(\mathbf{x}\_{U}|\mathbf{x}\_{O}), to run the acquisition in feature space and does not perform as well as SEFA that runs in the latent space, despite being able to model the dependencies. Additionally, our datasets have high interdependencies (image and genomics datasets), and SEFA still performs well on these.

We appreciate this is still a modelling restriction, however, we believe that due to the predictor being able to model the complex interdependencies and the ablation results, we have overstated this as a limitation. We will adjust our conclusion to accurately reflect this.

Figure 4

When we were making Figure 4, we initially had just the feature numbers on the y-axis. However, the trade-off we made was to put the feature names, providing more immediate information, but making the text smaller. The figure is a vector image so can be zoomed in. We shall add that the figure is best viewed zoomed in to the caption.

As suggested by Reviewer V2wP, without the supporting text it is not immediately clear which features are relevant for certain tumor locations. Therefore, we will edit Figure 4, such that class-wise selections with high frequencies (darker rectangles on the heat map) are highlighted using green bounding boxes, as was done with Syn 1-3. And notable low frequency selections, (where the selection frequency is low, but high for other classes) will be highlighted with black bounding boxes. This way the instance-wise differences are more noticeable, the selections are easy to identify when reading about them in the text, and the figure should be clearer by guiding attention to relevant parts of the figure.

审稿意见
3

This paper addresses Active Feature Acquisition (AFA), the task of sequentially selecting which features to measure for a specific test instance to improve prediction accuracy while minimizing the number of features acquired. The authors identify limitations in existing approaches: Reinforcement Learning (RL) methods face training difficulties, while Conditional Mutual Information (CMI) maximization makes myopic decisions that don't consider future acquisitions.

The method was evaluated on synthetic datasets with known optimal feature orderings and on real-world datasets including image classification and cancer classification tasks. The authors demonstrate that SEFA consistently outperforms baseline methods, including both RL-based and CMI-based approaches. For the cancer classification task, they also validate that the features selected by SEFA align with biomarkers identified in scientific literature.

给作者的问题

  1. Computational Overhead Analysis: How does the inference time of SEFA compare to the baseline methods across different datasets? Since SEFA requires multiple latent samples and gradient calculations during acquisition, it would be helpful to understand the practical trade-off between performance gains and computational costs. A response showing reasonable computational requirements would strengthen the practical applicability of the method.

  2. Adaptation to Varying Acquisition Costs: Many real-world scenarios have different costs associated with acquiring different features (e.g., some medical tests are more expensive or invasive than others). How would SEFA be adapted to account for variable acquisition costs? A clear adaptation strategy would significantly enhance the method's practical utility in cost-sensitive domains.

  3. Extension to Regression Tasks: Given that the current probability weighting mechanism is specific to classification, what modifications would be required to extend SEFA to regression tasks? Understanding this extension would clarify the method's generalizability beyond classification problems, which is currently listed as a limitation.

  4. Robustness to Noisy Features: How does SEFA perform when features contain noise or measurement errors? Since the method relies on gradient information to score features, it's important to understand its sensitivity to noise. Evidence of robustness would strengthen confidence in SEFA's real-world applicability.

  5. Selection of Latent Dimensionality: How should practitioners determine the number of latent components per feature for new datasets? This hyperparameter seems crucial for balancing model capacity with regularization, and guidance beyond empirical tuning would make the method more accessible to new users.

论据与证据

Overall, the paper's claims are generally well-supported by evidence, though with some limitations:

Well-supported claims:

  1. Theoretical limitations of CMI maximization: The authors provide both theoretical arguments and concrete examples showing why greedy CMI maximization can be sub-optimal for AFA (Section 4).

  2. Performance of SEFA vs. baselines: The extensive empirical evaluation across multiple datasets consistently shows that SEFA outperforms baseline methods. The authors report means with standard errors across 5 runs, providing statistical confidence in their results.

  3. Ablation studies: The impact of each SEFA component is demonstrated through comprehensive ablations on synthetic and real datasets, showing that each design element contributes to performance.

  4. Feature relevance on medical datasets: For TCGA data, the authors cite scientific literature supporting that their model selects biologically relevant features for different cancer types.

Claims with more limited evidence:

  1. Overcoming RL difficulties: While the authors cite RL challenges as motivation for SEFA, they don't directly demonstrate how these specific challenges affect performance on their tasks. SEFA does outperform the RL baseline, but the exact connection to the cited RL difficulties isn't experimentally verified.

  2. Non-greedy acquisitions: The synthetic experiments show SEFA makes better decisions than greedy methods, but it's not entirely clear if this is due to the specific "non-greedy" design or other aspects of the method. The indicator example helps, but more direct experimental validation of this specific claim would strengthen it.

  3. Scalability claims: While the paper discusses computational complexity in Table 4, there's no empirical evaluation of actual runtime comparisons between methods, which would help verify the practical implications of the theoretical complexities.

  4. Generalization beyond classification: The method is currently limited to classification tasks, as noted in the limitations section. Claims about the general superiority of the approach should be considered in this context.

The paper's claims are largely substantiated by the evidence presented, with the experimental methodology being thorough and the conclusions reasonably drawn from the results.

方法与评估标准

Proposed Methods

The proposed SEFA method makes sense for the Active Feature Acquisition (AFA) problem and addresses known limitations of existing approaches:

  1. Latent space reasoning: Using a regularized latent space to simplify decision-making is appropriate for complex feature relationships, particularly since the Information Bottleneck approach helps focus on label-relevant information.

  2. Stochastic encoding: The approach of using multiple latent samples to consider diverse feature realizations addresses the key issue of myopic decision-making in CMI-based methods.

  3. Probability weighting: Weighting by predicted class probabilities is a sensible way to focus on distinguishing between likely classes rather than ruling out unlikely ones.

  4. Supervised training: Avoiding reinforcement learning complexities by using supervised training with a predictive loss is a reasonable design choice given the known difficulties with RL.

The factorization of the latent distribution (each feature responsible for specific latent components) is a pragmatic simplification, though as the authors acknowledge, it limits modeling conditional dependencies between features.

Evaluation Criteria

The evaluation approach is comprehensive and appropriate:

  1. Dataset selection: The mix of synthetic datasets (with known optimal strategies), image datasets, tabular datasets, and medical datasets provides good coverage of different application scenarios.

  2. Synthetic datasets: Using synthetic data with known optimal feature ordering provides clear validation of the method's ability to learn non-myopic acquisition strategies.

  3. Metrics: Using AUROC for binary classification and accuracy for multi-class classification are standard and appropriate choices. The "average evaluation metric during acquisition" appropriately captures performance across the entire acquisition process.

  4. Baselines: The comparison against diverse baselines (RL-based, CMI-based, and fixed ordering) provides a thorough evaluation landscape.

  5. Ablation studies: The detailed ablation experiments effectively isolate the contribution of each component of the method.

  6. Qualitative analysis: For the TCGA dataset, connecting selected features to literature-supported biomarkers adds biological plausibility to the results.

One potential limitation is the lack of evaluation under different acquisition budget constraints, which would be relevant for resource-constrained applications. Also, while the authors pre-select features for high-dimensional datasets (MNIST, TCGA, etc.) using STG for computational efficiency, this pre-selection might impact the true feature acquisition performance in those domains.

Overall, the methods and evaluation criteria are well-aligned with the AFA problem and provide convincing evidence for the effectiveness of SEFA.

理论论述

  1. The proof of Proposition 4.2 doesn't fully explain why considering possible unobserved feature values in the acquisition objective enables optimal feature acquisition more generally beyond the indicator example.

  2. The paper doesn't provide formal guarantees about SEFA's optimality, only demonstrating that considering unobserved feature values is necessary for optimality in the indicator example.

  3. The connection between the theoretical insights from Propositions 4.1 and 4.2 and the design of SEFA isn't made fully explicit—particularly how the stochastic latent space approach relates to the integral in Proposition 4.2.

实验设计与分析

The synthetic dataset experiments (Section 6.1) are well-designed to test whether models can learn optimal feature orderings. The use of known optimal paths, multiple runs with error reporting, and clear visualizations provides strong validation of acquisition strategies. This controlled setting effectively demonstrates SEFA's ability to prioritize features optimally.

The real dataset experiments (Section 6.2) appropriately evaluate performance across varied acquisition steps. The diverse dataset selection, consistent protocols, and multiple runs strengthen validity. One limitation is the pre-selection of features for high-dimensional datasets using STG, which might introduce bias if different acquisition methods would naturally prefer different feature subsets.

The cancer classification experiments (Section 6.3) effectively combine quantitative performance with qualitative analysis. The supporting literature citations for selected features and analysis across cancer types validate that the model makes biologically plausible decisions. While not exhaustive, this literature validation adds credibility to the feature selections.

The ablation studies are comprehensive, systematically removing each proposed component while maintaining consistent evaluation protocols across synthetic and real datasets. The accompanying visualizations of acquisition behavior clearly demonstrate each component's impact, strongly supporting the claim that all components are necessary for optimal performance.

The sensitivity analyses and hyperparameter selection procedures follow sound practices, with appropriate exploration of parameter ranges and validation metrics. The reporting of final configurations enhances reproducibility, though more explicit justification for initial hyperparameter ranges would strengthen methodological transparency.

Overall, the experimental designs employ appropriate controls, metrics, and validation procedures. The analyses thoroughly support the paper's claims about SEFA's effectiveness for active feature acquisition.

补充材料

briefly looked at the additional results

与现有文献的关系

The paper's critique of CMI-based acquisition relates to broader discussions about greedy vs. non-greedy information acquisition. Chen et al. (2015a) established theoretical bounds for when greedy information maximization is near-optimal, but SEFA demonstrates scenarios where this breaks down. SEFA's factorized latent space approach conceptually relates to Ma et al.'s EDDI (2019), but fundamentally differs by focusing on gradients in latent space rather than estimating CMI.

SEFA addresses the exploration-exploitation tradeoff in active learning without using reinforcement learning. This connects to broader challenges in RL, offering an alternative supervised approach.

SEFA's probability weighting mechanism addresses a known issue in information-theoretic approaches where minimizing entropy can focus on low-probability events. This relates to research on decision-theoretic active learning and optimal experimental design, where the goal is to directly minimize expected error rather than maximize information gain. SEFA's approach of encoding features individually before integration relates to research on partial variable-wise encoders in missing data contexts. While prior works like ACFlow (Li et al., 2020) used normalizing flows for this purpose, SEFA provides a simpler yet effective alternative specifically designed for the acquisition task.

遗漏的重要参考文献

While the paper provides a thorough literature review, there are a few notable omissions that would help contextualize their contributions:

  • Monte Carlo Tree Search (MCTS): The non-greedy acquisition problem shares similarities with planning problems where MCTS has been successful (Lim et al., 2012). MCTS explicitly balances immediate rewards with long-term payoffs through rollouts, conceptually similar to how SEFA samples multiple latent realizations to evaluate acquisition decisions.

  • Feature attribution methods: SEFA's gradient-based approach for calculating feature importance shares similarities with feature attribution methods like Integrated Gradients (Sundararajan et al., 2017) and LIME (Ribeiro et al., 2016). These connections could strengthen the theoretical foundation of the gradient-based scoring function.

  • Partial observability in sequential decision making: The paper lacks references to Partially Observable Markov Decision Processes (POMDPs) literature

其他优缺点

Strengths

Novel Integration of Techniques: The paper creatively combines stochastic encoders, latent space regularization, and gradient-based feature scoring into a coherent framework. While each individual component builds on existing ideas, their integration into a unified acquisition approach is original and effective.

Principled Design Choices: Each aspect of SEFA addresses specific limitations of prior methods. The stochastic latent space addresses myopic decision-making, probability weighting focuses on distinguishing between likely classes, and the factorized architecture enables tractable feature scoring. These design choices are well-motivated by theoretical insights.

Strong Empirical Results: The consistent outperformance across diverse datasets is impressive. Particularly noteworthy is the performance on medical datasets where domain knowledge validates that selected features align with biological mechanisms.

Pragmatic Balance: The method strikes a good balance between theoretical principles and practical implementation. By using supervised training instead of reinforcement learning, the authors provide a more accessible approach to AFA that performs better and is easier to implement.

Clear Writing and Visualizations: The paper is generally well-written with excellent visualizations. The heat maps showing acquisition trajectories (Figures 2, 4) are particularly effective at conveying complex patterns in feature selection across different scenarios.

Weaknesses

Limited Theoretical Guarantees: While the paper provides examples where SEFA works well, it lacks formal guarantees about its optimality or convergence properties. The theoretical foundation primarily motivates the approach rather than providing rigorous performance bounds.

Parameter Sensitivity: The method introduces several hyperparameters (β, number of latent components per feature, number of samples). While ablations show their importance, the paper doesn't provide strong guidance on how to select these parameters for new datasets beyond empirical tuning.

Computational Overhead: At inference time, SEFA requires sampling from latent distributions and computing gradients, which may be computationally demanding compared to simpler methods. The paper acknowledges this limitation but doesn't provide empirical runtime comparisons.

Restricted to Classification: As noted in the limitations, the method is currently designed for classification tasks. This restricts its applicability to regression problems, which are common in many domains where feature acquisition is relevant (e.g., predictive maintenance, climate modeling).

Feature Independence Assumption: The factorized latent space, while computationally convenient, assumes independent encoding of features. This may limit the model's ability to capture complex interactions between features, particularly in datasets with strong feature correlations.

Limited Discussion of Scalability: While the paper demonstrates effectiveness on datasets with moderate dimensionality, it's unclear how the approach would scale to problems with thousands or millions of features, which appear in genomics and other high-dimensional domains.

Overall, the paper presents a significant contribution to the AFA literature with strong empirical validation. Its originality lies more in the creative combination of techniques and their application to an important problem than in fundamental theoretical breakthroughs.

其他意见或建议

  1. When describing the latent space regularization (Section 5.3), the paper refers to using a variational upper bound but doesn't explicitly state the form of this bound. For clarity, it would be helpful to provide the specific form.
作者回复

Thank you for the positive review, we are grateful for the feedback, we answer your questions below. All changes below will be added.

Scalability

We have addressed this shared point in our response to Reviewer ABGZ.

Generalization to Regression

SEFA can be modified for regression so that the predictor network has two heads, one that predicts the label directly and one that predicts the label as a classification problem (after it has been discretized into equally sized bins). We can use the regression head to predict the label, and the classification head to carry out the feature acquisitions.

Feature Costs and Budget

Feature Costs: There are two common ways to include cost: (1) additively, e.g. minusing the cost from the reward in RL [Li and Oliva 2021], and (2) to divide by the cost either in the reward [Kachuee et al. 2019] or in the CMI estimate [Gadgil et al. 2024]. We believe dividing by the cost is better. Intuitively, the adjusted scores represent the advantage of acquiring a feature per unit cost. And it enforces the desired property that features with zero cost are acquired. Additionally, like CMI, the scores and training of SEFA are independent of cost, scores are adjusted based on cost after calculation. So if feature costs vary over time, SEFA would not need to be retrained like RL methods that learn a policy.

Acquisition Budget: Figure 3 shows the evaluation metrics throughout the acquisition - how the models would perform if the acquisition was ended after nn features. SEFA consistently performs best throughout the acquisition.

Theoretical Claims

We have addressed this shared point in our response to Reviewer N1oA.

Hyperparameter Ranges

β\beta: The value of β\beta typically needs to be small, similar to a weight decay parameter. This is because it limits how much information from the features can be found in the latent space, papers such as Alemi et al. 2017 typically use values smaller than 0.01. In the sensitivity analysis in Figure 10 we see that if β\beta is too large, SEFA fails because it cannot predict effectively, and if β\beta is too low, then the latent space is not regularized enough. Based on this, we recommend 0.0001 - 0.01 as an effective range.

Number of Samples: SEFA improves with the number of acquisition and training samples, shown by the sensitivity analyses in Figures 11 and 12. This is because during training it encourages a diverse latent space, and during acquisition the diversity can be fully sampled. We recommend 50-200 as a good tradeoff between acquisition performance and computation.

Number of Latent Components: We use more than one latent component per feature so that there is a large capacity in the latent space for a rich representation of each feature. However, using too many increases the possibility of overfitting. Since each feature is only one value we recommend a value between 5-10. We tested 4, 6 and 8 as possible values and these were effective.

Other Hyperparameters: The initial ranges were determined by manually finding good values on the synthetic datasets and using ranges around those. We started with broadly accepted values, e.g. 0.001 as the default PyTorch Adam learning rate.

References

We agree with all suggestions and will add them. We will also cite Baehrens et al. 2009 and Simonyan et al. 2014 as gradient based feature attribution methods.

Feature Independence

We have addressed this shared point in our response to Reviewer N1oA.

Variational Upper Bound

The regularization term to minimize is I(Z;X)I(Z; X). This requires access to the marginal p(z)=p(zx)p(x)dxp(z)=\int p(z|x)p(x)dx, which is intractable. Instead we use the standard normal as a variational approximation since it produces an analytical upper bound. The derivation can be found in equations 13 and 14 of Alemi et al. 2017.

Noisy Features

We have run Syn 1 (with two additional baselines) with added Gaussian noise with three standard deviations: 0.1, 0.2 and 0.4. The number of acquisitions to acquire the correct features are:

ModelNo Noise0.10.20.4
DIME4.079±0.0574.079\pm0.0574.688±0.2114.688\pm0.2114.703±0.2764.703\pm0.2765.368±0.2675.368\pm0.267
EDDI9.183±0.1879.183\pm0.1878.988±0.2858.988\pm0.2859.216±0.1369.216\pm0.1369.382±0.2459.382\pm0.245
Fixed MLP6.009±0.0006.009\pm0.0006.312±0.2026.312\pm0.2027.321±0.3787.321\pm0.3786.009±0.0006.009\pm0.000
GDFS4.568±0.1954.568\pm0.1954.566±0.1874.566\pm0.1874.543±0.1644.543\pm0.1645.583±0.2835.583\pm0.283
Opportunistic RL4.203±0.0344.203\pm0.0344.347±0.0914.347\pm0.0914.720±0.1424.720\pm0.1425.488±0.0845.488\pm0.084
VAE6.593±0.0856.593\pm0.0856.866±0.0416.866\pm0.0417.005±0.0867.005\pm0.0867.095±0.0787.095\pm0.078
GSMRL5.570±0.1275.570\pm0.1275.495±0.1265.495\pm0.1265.778±0.1205.778\pm0.1206.858±0.2696.858\pm0.269
Random9.484±0.0069.484\pm0.0069.495±0.0109.495\pm0.0109.495±0.0109.495\pm0.0109.495±0.0109.495\pm0.010
SEFA (ours)4.017±0.0034.017\pm0.0034.100±0.0044.100\pm0.0044.207±0.0084.207\pm0.0084.406±0.0044.406\pm0.004

The models get worse with more noise. SEFA remains the best model at all levels of noise. This is likely due to the latent space regularization removing noise between the feature space and latent space.

审稿意见
4

This paper considers active feature acquisition, which is a dynamic test-time selection (usually sequential) of features to make predictions on each test instances. The selected features are chosen independently for each considered test instance. The authors argue (with some theory) how prior methods based on conditional mutual information can fail in pathological cases, and how they might fail more generally. Instead the authors produce an information bottleneck inspired approach that utilizes a latent space to encode surrogate behavior for how unobserved variables might couple with selected features, in predicting the label. This relies on a decoupling of latent variables, one grouping for each candidate, along with a gradient-based acquisition function. There are extensive synthetic and real-world experiments to support the use of this method in practice.

给作者的问题

Please see the comments above

update after rebuttal

Thank you for your response. I maintain my score at accept

论据与证据

The empirical results are strong, with sufficient comparisons against baselines and ablations, and results that seem significant over these baselines. I also really liked the explanation of why conditional mutual information might fail in proposition 4.1 and the surrounding discussion. This was well done.

方法与评估标准

Yes

理论论述

I did not look carefully at the proofs

实验设计与分析

Yes. There are a nice range of analyzes including both number of features needed to reach the optimal set (synthetic data) as well as accuracy vs number of acquired feature curves, and qualitative visualizations showing the "trajectories" of selected features.

补充材料

Briefly examined appendix

与现有文献的关系

While I appreciate the discussion around Shannon entropy not necessarily encouraging a maximal probability most likely class, I think more discussion is needed here around alternate metrics. Sometimes Renyi entropies are used instead of Shannon entropy to encourage different properties. For instance the min-entropy exactly encourages maximum probability on a single class. Extrinsic Jensen-Shannon divergence is another example, as an alternative to mutual information that better distinguishes between likely hypotheses (see e.g., Naghshvar & Javidi 2012).

Naghshvar, M., & Javidi, T. (2012, July). Extrinsic Jensen-Shannon divergence with application in active hypothesis testing. In 2012 IEEE International Symposium on Information Theory Proceedings (pp. 2191-2195). IEEE.

遗漏的重要参考文献

See above. These alternate metrics instead of Shannon entropy are not in the critical path of the paper, but would strengthen the discussion

其他优缺点

  • one additional baseline that I'm wondering about is, why didn't you compare against the baseline of selecting a random order of features, for each dataset?

I think Section 5 needs more exposition. It just gets right into the details of the design choices made, but this paper would be strengthened with additional discussion around each of these choices. In particular:

  • What is the core intuition for z? Does it essentially serve as a surrogate for a marginalization over p(xU | xO), which would otherwise be too computationally expensive to consider?
  • As discussed in the limitations, the factorization of the latent encoding is a huge assumption. This entirely decouples any relationship between the selected features. It is not clear to me at all why this is a good idea, intuitively. In fact, I'm a bit surprised that the empirical results are as good as they are, since the features do not seem to be coupled in any way. Am I missing something?
  • The discussion around Shannon entropy set things up really nicely to argue how a different scoring metric is needed to better encourage maximal class probabilities on a single class. Yet, to me the choice of the gradient-based r(c,z,i) is still very much a heuristic. I think more discussion is needed around this scoring function. Why is this a good idea? Sensitivity of p(Y|z) is listed as the reason, but how does this in any way fulfill the property that Shannon entropy is lacking?

其他意见或建议

I like the visualizations showing the "trajectories" of the selected features. However, without the context of the scientific literature discussion in Section 6.3, I don't know how much Figure 4 is really adding. For someone without a medical background, the trajectories look random. This figure could be strengthened by annotating which features are actually scientifically important for each tumor location, so we can see that it is actually grabbing the important features, early on.

作者回复

Thank you for the positive review, we are grateful for the feedback, we answer your questions below. All changes below will be added.

Renyi Entropy

Thank you for bringing this to our attention, we agree that Renyi entropies are relevant. In particular how the min-entropy relates to Shannon Entropy focussing on making low likelihoods lower (Section 4). We shall add these citations.

That being said, Renyi entropies would still suffer from making myopic decisions. They would still fail on the Indicator for the same reasons as Shannon Entropy due to Proposition 4.1. Additionally it is highly non-trivial to incorporate these metrics into the SEFA architecture.

Random Selections

We originally tested random orderings to determine whether datasets would benefit from a non-random ordering. Random orderings performed poorly so we did not present the results to save space. We have provided the original results in our response to Reviewer ABGZ in the Additional Baselines sections. The random ordering always performs worse than the Fixed Global MLP Ordering - the datasets benefit from non-random orders.

Using the Latent Space

Marginalization over p(x_Ux_O)p(\mathbf{x}\_{U} | \mathbf{x}\_{O}) is possible, we do it in our Feature Space ablations (Table 1 and Table 3). Instead the choice to use the latent space was made to improve performance, due to the following reasons:

  • Acquisitions are not made using gradients calculated with feature level noise. By using latent space regularization we make acquisitions with a feature representation with noise removed containing only label relevant information.
  • We avoid the need to train a generative model, avoiding the complex feature distributions (e.g. multi-modal) with complex interdependencies. In latent space we can model the distribution as independent Gaussians and the predictor learns to predict using these, and handles the interdependencies. See more in the Feature Independence response for a further explanation.
  • Tabular data may contain features that are categorical (gradients are not meaningful), or features at different scales (the gradients are different scales), so gradients with respect to the raw features is not a reliable way to score them. We are able to train the encoders so that the latent components are all approximately the same scale (by regularizing the latent space) and real numbers, so that the gradients give reliable feature scores.

In summary, by calculating scores using representations of the features, we avoid both the need to train a generative model, and the associated pitfalls (feature level noise, complex distributions, complex interdependencies, less meaningful gradients).

Feature Independence

As suggested by the ICML Program Chairs we have addressed this shared point in our response to Reviewer N1oA. Briefly, the complex interdependencies are able to be accounted for by the predictor network. The encoders do not need to learn the couplings between the features, because the predictor is able to treat latent components differently based on observed features, and therefore the gradients change based on this.

Scoring Metric

Overcoming Shannon Entropy Drawbacks: The two main properties that Shannon Entropy lacks are (1) considering unobserved values in the current decision and (2) focussing on identifying the most likely class. SEFA addresses the first property by taking an expectation of r(c,z,i)r(c, \mathbf{z}, i) over z\mathbf{z} considering the possible latent realizations of the unobserved features. SEFA addresses the second property by using probability weighting (the outer sum over p_θ,ϕ(yx_O)p\_{\theta, \phi}(y| \mathbf{x}\_{O}) in Equation 1).

Intuition behind r(c,z,i)r(c, \mathbf{z}, i): As pointed out by Reviewer Pdip, a good perspective on r(c,z,i)r(c, \mathbf{z}, i) is via Feature Attribution. Gradients are a common technique for explaining which features are most important for predicting a particular instance [Baehrens et al. 2009, Simonyan et al. 2014]. And LIME [Ribeiro et al. 2016] suggests attributions using locally linear approximations of the predicted probability for each class separately. So we take many samples from the latent space, representing possible future latent realizations. For each of these, r(c,z,i)r(c, \mathbf{z}, i) uses gradients as local linear approximations, to calculate feature attributions via the latent components based on how important they are for predicting class cc. After the mean over the latent samples is taken, we have feature attributions for each class, we take a sum over classes, weighted by the current predicted probability, p_θ,ϕ(yx_O)p\_{\theta, \phi}(y| \mathbf{x}\_{O}), to give more focus to features that are relevant to more likely classes.

Figure 4

Thank you for the suggestion, we agree that the figure can be improved to highlight the class relevant features. As suggested by the ICML Program Chairs, we have addressed this shared point in more detail in our response to Reviewer N1oA.

审稿意见
3

The paper proposes a Stochastic Encodings for Feature Acquisition (SEFA) framework addressing the limitations of CMI and RL methods. SEFA uses an encoder-predictor architecture with intermediate stochastic latent variables. The architecture makes predictions and calculates the acquisition objective.

给作者的问题

Please check the weaknesses section.

论据与证据

Yes.

方法与评估标准

Yes.

理论论述

Yes.

实验设计与分析

Yes.

补充材料

Yes.

与现有文献的关系

The contributions are novel addressing the limitations of CMI and RL methods.

遗漏的重要参考文献

NA

其他优缺点

Strengths -

The theoretical proofs to establish the limitations of CMI and RL and the proposed SEFA framework, look solid. The performance look pretty good.

Weaknesses -

There's a high probability of encountering scalability issues with the proposed method and may turn out to be difficult to extend this to various other applications.

Some recent RL baselines can be used as baselines.

其他意见或建议

NA

作者回复

Thank you for the positive review, we are grateful for the feedback, we answer your questions below. All changes below will be added.

Scalability

We provide the computational complexities for a single training step and single acquisition step for each model in Table 4 (Appendix H.3). The main takeaway is that SEFA scales better than half the methods at training time, better than the other half during acquisition and never the worst.

As suggested by Reviewer Pdip, to add to this, we have timed each model to determine the wall-clock times for inference and training. The average times in seconds for a single acquisition are given below:

ModelSyn 1Syn 2Syn 3
DIME0.017±0.0000.017\pm0.0000.018±0.0000.018\pm0.0000.024±0.0000.024\pm0.000
EDDI2.871±0.0592.871\pm0.0592.613±0.0022.613\pm0.0022.854±0.0562.854\pm0.056
Fixed MLP0.014±0.0000.014\pm0.0000.011±0.0000.011\pm0.0000.017±0.0000.017\pm0.000
GDFS0.018±0.0000.018\pm0.0000.015±0.0010.015\pm0.0010.022±0.0000.022\pm0.000
Opportunistic RL0.029±0.0010.029\pm0.0010.021±0.0010.021\pm0.0010.032±0.0000.032\pm0.000
VAE0.153±0.0000.153\pm0.0000.154±0.0000.154\pm0.0000.267±0.0050.267\pm0.005
GSMRL0.047±0.0000.047\pm0.0000.040±0.0000.040\pm0.0000.055±0.0000.055\pm0.000
Random0.014±0.0000.014\pm0.0000.011±0.0000.011\pm0.0000.018±0.0010.018\pm0.001
SEFA (ours)0.246±0.0010.246\pm0.0010.165±0.0010.165\pm0.0010.240±0.0020.240\pm0.002
ModelCubeBank MarketingCalifornia HousingMiniBooNE
DIME0.021±0.0000.021\pm0.0000.016±0.0000.016\pm0.0000.010±0.0010.010\pm0.0010.021±0.0000.021\pm0.000
EDDI8.872±0.2138.872\pm0.2131.456±0.0021.456\pm0.0020.583±0.0060.583\pm0.0069.039±0.2459.039\pm0.245
Fixed MLP0.020±0.0000.020\pm0.0000.012±0.0000.012\pm0.0000.006±0.0000.006\pm0.0000.019±0.0000.019\pm0.000
GDFS0.021±0.0000.021\pm0.0000.017±0.0020.017\pm0.0020.008±0.0000.008\pm0.0000.020±0.0000.020\pm0.000
Opportunistic RL0.031±0.0000.031\pm0.0000.022±0.0010.022\pm0.0010.010±0.0000.010\pm0.0000.034±0.0010.034\pm0.001
VAE0.278±0.0020.278\pm0.0020.136±0.0000.136\pm0.0000.048±0.0010.048\pm0.0010.291±0.0010.291\pm0.001
GSMRL0.061±0.0000.061\pm0.0000.035±0.0010.035\pm0.0010.021±0.0000.021\pm0.0000.045±0.0000.045\pm0.000
Random0.019±0.0000.019\pm0.0000.012±0.0000.012\pm0.0000.006±0.0000.006\pm0.0000.017±0.0000.017\pm0.000
SEFA (ours)0.861±0.0040.861\pm0.0040.130±0.0010.130\pm0.0010.077±0.0000.077\pm0.0000.314±0.0020.314\pm0.002
ModelMNISTFashion MNISTMETABRICTCGA
DIME0.024±0.0000.024\pm0.0000.023±0.0000.023\pm0.0000.003±0.0000.003\pm0.0000.010±0.0000.010\pm0.000
EDDI12.360±0.02912.360\pm0.0298.543±0.2008.543\pm0.2000.108±0.0010.108\pm0.0010.765±0.0110.765\pm0.011
Fixed MLP0.020±0.0000.020\pm0.0000.020±0.0000.020\pm0.0000.002±0.0000.002\pm0.0000.008±0.0000.008\pm0.000
GDFS0.024±0.0000.024\pm0.0000.025±0.0000.025\pm0.0000.003±0.0000.003\pm0.0000.009±0.0000.009\pm0.000
Opportunistic RL0.030±0.0000.030\pm0.0000.031±0.0000.031\pm0.0000.003±0.0000.003\pm0.0000.012±0.0000.012\pm0.000
VAE0.452±0.0070.452\pm0.0070.452±0.0070.452\pm0.0070.009±0.0000.009\pm0.0000.031±0.0010.031\pm0.001
GSMRL0.097±0.0010.097\pm0.0010.069±0.0010.069\pm0.0010.016±0.0000.016\pm0.0000.021±0.0000.021\pm0.000
Random0.019±0.0000.019\pm0.0000.019±0.0000.019\pm0.0000.002±0.0000.002\pm0.0000.008±0.0000.008\pm0.000
SEFA (ours)1.498±0.0021.498\pm0.0021.497±0.0031.497\pm0.0030.019±0.0000.019\pm0.0000.098±0.0000.098\pm0.000

As the number of classes increases (MiniBooNe to Cube and Cube to MNIST/Fashion MNIST), SEFA's acquisition time increases, despite the number of features staying the same, this matches Table 4. The time for Generative Models increases significantly with the number of features in line with Table 4. As expected, models with policy networks are fastest, and scale well with the number of features, but SEFA is never the slowest (faster than EDDI, often by a significant amount).

The average time in milliseconds for a single training step (one gradient descent step) on Syn 1 is given below:

ModelIteration Time (mS)
DIME35.156±0.08135.156\pm0.081
EDDI7.290±0.0287.290\pm0.028
Fixed MLP5.152±0.0275.152\pm0.027
GDFS21.661±0.05621.661\pm0.056
Opportunistic RL89.565±0.37089.565\pm0.370
VAE6.761±0.0756.761\pm0.075
GSMRL90.383±0.28590.383\pm0.285
SEFA (ours)15.843±0.09115.843\pm0.091

Models that train by simulating acquisition are slowest to train despite being faster at inference time. Supervised models, including SEFA, are faster because they do not need to generate rollouts.

Additional Baselines

Thank you for the suggestion, we have run the experiments on the following new baselines:

  • GSMRL [Li and Oliva 2021], a more recent and popular RL baseline for AFA.
  • A random ordering with an MLP predictor as requested by Reviewer V2wP (to show feature ordering is necessary).

The results are given below:

ModelSynthetic 1Synthetic 2Synthetic 3
GSMRL5.570±0.1275.570\pm0.1276.227±0.1856.227\pm0.1858.199±0.0678.199\pm0.067
Random9.484±0.0069.484\pm0.0069.499±0.0059.499\pm0.0059.987±0.0089.987\pm0.008
SEFA (ours)4.017±0.0034.017\pm0.0034.098±0.0074.098\pm0.0075.081±0.0215.081\pm0.021
ModelCubeBank MarketingCalifornia HousingMiniBooNE
GSMRL0.891±0.0010.891\pm0.0010.879±0.0060.879\pm0.0060.638±0.0030.638\pm0.0030.946±0.0010.946\pm0.001
Random0.699±0.0010.699\pm0.0010.816±0.0030.816\pm0.0030.569±0.0030.569\pm0.0030.912±0.0010.912\pm0.001
SEFA (ours)0.904±0.0000.904\pm0.0000.919±0.0010.919\pm0.0010.676±0.0050.676\pm0.0050.957±0.0000.957\pm0.000
ModelMNISTFashion MNISTMETABRICTCGA
GSMRL0.701±0.0020.701\pm0.0020.683±0.0010.683\pm0.0010.665±0.0020.665\pm0.0020.781±0.0030.781\pm0.003
Random0.661±0.0010.661\pm0.0010.648±0.0010.648\pm0.0010.647±0.0050.647\pm0.0050.753±0.0030.753\pm0.003
SEFA (ours)0.761±0.0010.761\pm0.0010.721±0.0000.721\pm0.0000.708±0.0020.708\pm0.0020.845±0.0020.845\pm0.002

SEFA is better than both new baselines. Random performs poorly in all cases.

最终决定

The reviewers thinks that this is a solid paper: it highlights the limitation of conditional mutual information-based methods, which motivated their novel approach of using latent space. Extensive synthetic and real-world experiments supported the use of this method in practice. Additional baselines, noisy feature experiments, as well as scalability evaluation in the rebuttal stage helped confirm the utility of the proposed methods.