PaperHub
4.4
/10
withdrawn5 位审稿人
最低3最高6标准差1.2
3
5
3
6
5
3.2
置信度
ICLR 2024

Learned Mixing Weights for Transferable Tabular Data Augmentation

OpenReviewPDF
提交: 2023-09-22更新: 2024-03-26

摘要

We present an architecture-agnostic method for tabular data augmentation, which mixes pairs of samples from the training set. The mixing procedure is based on a set of per-feature weights that are assigned by a learned network $g$ that is separate from the primary classification model $f$. The features are selected between the two samples at random, and the sum of the weights that $g$ assigns the features that are selected from each sample determines the mixing of the target label. $g$ itself is trained based on two loss terms, one that encourages variability in the assigned weights between the features and one that encourages, for every training sample, the model $f$ to be agnostic to the features for which $g$ assigns low weights. Our experiments show that this learned data augmentation method improves multiple neural architectures designed for tabular data. Even more notable is that the network $g$ that was trained on an MLP produces mixed samples that improve non-differentiable methods, including classical methods and gradient-boosting decision tree methods. This is done without any further tuning and with the default parameters of the classifiers. The outcome achieved this way, using the cutting-edge CatBoost method now represents the state of the art.
关键词
Tabuler datamix augmentationdata augmentation

评审与讨论

审稿意见
3

The authors propose a method to perform mixup on tabular data. Instead of mixing the features themselves, some features are chosen directly from one sample, some from the other sample. An MLP is used to determine the weighting of the two samples in the loss.

优点

The method is simple and widely applicable across architectures. It appears to provide consistent benefits on the majority of datasets used.

缺点

  • The paper provides little motivation for many of the aspects of the approach. The writing is not particularly clear, the clarity of the exposition can be much improved still.
  • While there are many classification models the method is applied on in the experiments, there is only single data augmentation baseline that is compared against. The difference in performance does not seem significant. The paper also does not contain any comparisons to numbers in the literature, which makes it difficult to judge how good the results are in the first place. Since no per-dataset numbers are provided, it is unclear how large the variance between multiple repeats is. I would especially like to see a comparison against Chen et al. (2023), since they propose a very similar idea where features are swapped between samples (their Feat-Mix variant). The authors would benefit from elaborating on the similarities and differences further than the small paragraph in the current version of the paper, considering the similarity in terms of approach. Chen et al don't evaluate their Feat-Mix with other architectures, but in principle I believe it should apply just the same as the proposed approach here.
  • It would be good to have a statistical analysis of whether the differences between models are statistically significant, for example using the Friedman and Nemenyi tests, see [1]

[1] https://www.jmlr.org/papers/volume7/demsar06a/demsar06a.pdf

问题

Can you specify a and b more precisely in equation 1? Are the per-sample weights predicted by the MLP consistent between different training runs of the MLP?

审稿意见
5

This paper proposes a structure-agnostic data mixing scheme tailored towards unstructured tabular datasets. Two data points are mixed randomly at a 1:7 ratio of features. A feature selection model gg is used to assign per-feature weights for each example. The weight of an example for the mixed up vector is the sum of weight of the features used from that example in the mix. These weights are used to construct a loss function during training, with added regularization terms to avoid uniform feature weight assignment. The weights are also employed during evaluation by linearly mixing the prediction with the output from a weighted version of the feature. The work proposes a transfer learning scheme where the feature importance model gg is first pre-trained alongside a differentiable MLP (as the main predictor). The importance model is then used on black-box classifiers such as gradient boosted decision trees. The empirical results show that the proposed method applied on top of current SOTA improves the results on standard benchmark datasets.

优点

Originality and significance: The use of data mixing with a learned feature importance model for unstructured data (to the best of my knowledge) is novel to this work. The results show improvements as the method is used with different choices of the predictor model (except for logistic regression). A limited study shows that for continuous domains, the proposed method matches the performance of SnapMix (a feature mixing method that uses CAM scores from the main classifier to calculate mixing weights).

Quality and clarity: The paper is mostly easy to follow and the building blocks of the methods are explained clearly.

缺点

Originality and significance: It can be hard to assess the significance of the work since there is no comparison with other baseline methods suited for data mixing with unstructured data. I would have liked to see an ablation study where gg is a constant (uniform feature weights, mixing based on the count of features used from each example) and instead the ratio of the mix is varied.

Quality and clarity: The paper has an unusual flow. The introduction section includes many details of the experiments, while the motivations, intuitions and discussions are scattered across other sections. I had to go back and forth several times to find relevant details in various sections. IMO a restructuring of the text can help with readability.

问题

Notes and questions:

  • Mixing at evaluation time is not well motivated in the paper. Please elaborate why you would expect to see a performance boost by mixing with a weighted example prediction.
  • If we average g(.)g(.) across a dataset, would it result in average feature importance weights that could be used for feature selection, or are these weights mainly useful for example mixing?
  • Figure 1 is missing a legend. You can also rearrange and resize to a much smaller footprint without losing clarity.
  • SnapMix needs to be explained in more detail, maybe with a short description of CAM.
审稿意见
3

This paper proposes a learnable mixup-style data augmentation technique for tabular classification problems. At a high level, the proposed method learns a weighting function to appropriately weight the contribution of different features in virtual samples generated by stitching together different features of different input samples. Empirically, this method improves the performance of many popular learning algorithms.

优点

  1. The concept of learning a mixing function for mixup-style data augmentation is interesting and novel.
  2. Empirically, the proposed method improves classification accuracy for many common algorithms.

缺点

  1. The paper opens up with high-level overview of the proposed method and is missing a lot of context. More concretely, it's missing (1) a clear definition the problem of interest, (2) why the problem is of interest to the community, (3) what methods have prior works proposed and why these methods don't address the problem of interest, and (4) How the proposed method improves upon prior these prior works. Here is my current understanding of the context:
  • The problem of interest: training neural models with unstructured tabular data.
  • Why the problem is interesting (motivation): unclear.
  • Prior works: discussed in related work, but should also be discussed in the intro.
  • Core novelty: this is the first work to learn a mixing augmentation function for tabular problems. Please let me know if these short descriptions are generally correct. Further detail on these four points would help me better understand the better and also provide more feedback during the review period.
  1. I found the method description to be a bit haphazard and hard to follow. A figure illustrating the flow of information in training gg and ff as well as the method's novelties would be very helpful. The algorithmic details are clear, but the motivation behind these details is unclear to me. Please see the Question section my question related to the algorithm.
  2. Section 3 also introduce many new variables which makes it difficult to follow. If it's possible to consolidate notation, that would help immensely with readability. For instance, in Eq 1, the authors may consider replacing mium_i^u with g(xi)g(x_i) and then stating that gradients do not backpropagate through gg (analogous for mjum_j^u).
  3. It seems the empirical analysis does not compare the proposed mixup method with existing mixup-style methods. This method should at least be compared to a naive, unweighted mixup strategy.

Overall, I vote to reject. The proposed mixup method does seem to accomplish what the paper claims it should accomplish (as seen by Figure 1 and 3), but the paper is quite difficult to follow, much of the algorithm is missing justification, and lacks relevant baselines.

Other minor comments:

  1. aa and bb are not defined in Eq. 1.
  2. In Table 1, its difficult to identify the top performing methods. I suggest bolding/highlighting the best methods.
  3. I prefer the term "augmented sample" over "virtual sample" as is emphasizes that the data was generated using a data augmentation function.
  4. Notation for the Hadamard product is defined After Eq 2 in Section 3 but is used throughout the intro.

问题

  1. Could the authors explain the decision to use 7/8 of the features from the first sample? Ablations on this value would shed some light on this decision.
  2. I don't quite understand Eq. 3. Why this loss? At a high-level, what is the goal of the learned mixing model? What sorts of mixtures does it aim to generate? Is this loss applied only to virtual samples (if so, please use xˉ\bar{x} instead of xx here).
  3. Is the "weighted sample dg(x)xdg(x) \circ x" the sample thing as a virtual sample xˉ\bar{x}? If so, I suggest rewriting as "virtual sample xˉ=dg(x)x\bar{x} = dg(x) \circ x
  4. Why do we want to prevent uniformity?
  5. The loss in Eq. 4 can be minimized by updated gg such that just a single entry of mm large and the rest are uniform. Is this the intended effect?
  6. The first contribution mentions that the proposed method operates in the input domain. Could the authors explain why this detail is important?
  7. The paper mentions that the proposed method is suitable for unstructured data, but it is not clear to me what is meant by "unstructured." Could the authors please explain?
审稿意见
6

This paper presents a novel data augmentation strategy for tabular data classification. It utilizes an auxiliary network g to learn feature importance for each individual input example. Given a pair of examples, data mixing is realized by replacing 1/8 of one example’s features with the other example’s features. The coefficient of the loss objectives regarding the two examples is determined by the total importance of the mixed features. With the techniques of surrogate networks or gradient-free optimization, the proposed method can be applicable to any classification function. Authors conduct experiments on the AMLB benchmark. Results show that the proposed data mixing strategy outperforms baselines including CatBoost, XGBoost, MLP, TabTransformer, K-NN and Logistic Regression.

优点

  1. Mixup is a very popular data augmentation scheme. This paper conducts an interesting exploration to apply mixup-like augmentation approaches to tabular data.
  2. The proposed idea sounds reasonable. It is model-agnostic.
  3. Clear improvements are achieved on most datasets.

缺点

  1. As a core component of this paper, feature importance calculation achieved by g sounds a common idea in feature selection. Even if this paper is the first to apply feature selection techniques on tabular data augmentation, the relevant studies should be mentioned.
  2. In the experiments, no existing tabular data augmentation methods are evaluated and compared with.
  3. It is not clear why feature importance is necessary in the whole method. How about using uniform importance for all features?
  4. The improvement on Logistic regression is marginal.

问题

See weaknesses.

审稿意见
5

This work proposed an architecture-agnostic method for tabular data augmentation, which features close cooperation between a weight-assigning network and a classifier in both the training and inference phases. Experiments demonstrate the effectiveness of the proposed feature mixing technique on tabular data.

优点

  • the idea of co-train a weight assignment network gg together with a classifier ff is interesting. Such a paradigm of close cooperation between two networks may be used in other tasks.
  • the proposed technique is effective in improving multiple tabular data classifiers.

缺点

  • This paper violates the double-blind policy. The first paragraph of Sec.1 reads, "... we develop a data mixing (Guo et al., 2019; Huang et al., 2021) technique ...".
  • The proposed technique is not subject to theoretical analysis, nor is an intuitive picture explained.
    • it is hard to judge the applicability of the proposed technique, learning a network gg for data mixing, in general scenarios. It is possible that gg overfits and degrades the resulting ff.
    • the authors claim their technique is "architecture-agnostic." However, for more complex ff, simple MLP-based gg may not be able to capture its behavior (recall that gg is trained to replicate the logits of ff). It is interesting to ask how the architecture of gg will affect the performance.
    • why should we expect that the feature weight is beneficial during inference?
  • The presentation of ideas could be better organized and clearer.
    • there lacks a description of how to co-train the two networks, ff and gg,
    • In Sec. 3, it is unclear what the authors mean by stating, "... we sum the weights assigned by network g that are associated with the relevant coordinates of each sample". A mathematical expression would be helpful.
      • as features chosen from the two samples are non-overlapping, do the authors mean that we add the corresponding weights of two samples regardless of whether the feature is selected for constructing the virtual sample?
    • Eq. 1, a summation ranges from u=au=a to bb. However, there is no definition for aa and bb.
    • Eq. 3, the symbol dgdg could be confusing; it may be better to write it as dgd*g

问题

  • why should we expect that the feature weight is beneficial during inference?
  • the experiments mainly consider the tabular data. It is interesting to ask whether the proposed technique also shows competing performance on different types of data, e.g., images, compared to other feature mixing methods.
  • how would the performance change when adopting different architectures for gg
  • how are the two networks, ff and gg, co-trained in detail?
  • it seems one can also use a differentiable loss for Lm\mathcal{L}_m, e.g., the entropy, which encourages the weights to concentrate on one particular element of mm. If adopting, say, the entropy as Lm\mathcal{L}_m, how would the resulting network gg be affected?