PaperHub
5.8
/10
Poster4 位审稿人
最低5最高7标准差0.8
6
5
7
5
4.0
置信度
正确性2.8
贡献度3.0
表达3.3
NeurIPS 2024

TabEBM: A Tabular Data Augmentation Method with Distinct Class-Specific Energy-Based Models

OpenReviewPDF
提交: 2024-05-15更新: 2024-11-06
TL;DR

We introduce a new data augmentation method for tabular data, which trains class-specific generators.

摘要

Data collection is often difficult in critical fields such as medicine, physics, and chemistry, yielding typically only small tabular datasets. However, classification methods tend to struggle with these small datasets, leading to poor predictive performance. Increasing the training set with additional synthetic data, similar to data augmentation in images, is commonly believed to improve downstream tabular classification performance. However, current tabular generative methods that learn either the joint distribution $ p(\mathbf{x}, y) $ or the class-conditional distribution $ p(\mathbf{x} \mid y) $ often overfit on small datasets, resulting in poor-quality synthetic data, usually worsening classification performance compared to using real data alone. To solve these challenges, we introduce TabEBM, a novel class-conditional generative method using Energy-Based Models (EBMs). Unlike existing tabular methods that use a shared model to approximate all class-conditional densities, our key innovation is to create distinct EBM generative models for each class, each modelling its class-specific data distribution individually. This approach creates robust energy landscapes, even in ambiguous class distributions. Our experiments show that TabEBM generates synthetic data with higher quality and better statistical fidelity than existing methods. When used for data augmentation, our synthetic data consistently leads to improved classification performance across diverse datasets of various sizes, especially small ones. Code is available at https://github.com/andreimargeloiu/TabEBM.
关键词
tabular datadata augmentationsynthetic data generationenergy based model

评审与讨论

审稿意见
6

The authors address the problem of data augmentation for classification tasks by proposing a novel approach that utilizes separate energy-based models (EBMs) to generate synthetic data for each class. These EBMs are derived directly from the logits of a binary classifier whose goal is to distinguish real data from negative data. These logits (their logsumexp over positive and negative classes) are reinterpreted them as an energy function for each class. In order to avoid training these classifiers on the data itself, the authors take advantage of a pre-trained Prior-Data Fitted Network (PFN).

The authors perform a comprehensive evaluation of their method comparing it to competing methods for data augmentation on a number of small real world datasets. They show that their method is the only one that offers a consistent improvement in classification from the addition of synthetic data.

优点

  • The paper is well-structured and clearly written, effectively communicating the key ideas and concepts.
  • The authors introduce a novel approach by leveraging a Prior-Data Fitted Network (PFN) to obtain energy-based models (EBMs), which, to the best of my knowledge, has not been previously explored.
  • The experimental results section is thorough and comprehensive. They compare their approach to a wide range of relevant and competitive methods for synthetic data generation.
  • The appendices are well-organized and provide valuable supplementary information. The paper strikes a good balance with the most relevant information presented in main text and additional details provided in the appendix.

缺点

  • Interpreting classifier logits as energy functions may not necessarily yield good density models. Given a classifier with logits f(x)[y]f(x)[y], one can add an arbitrary function g(x)g(x) that is constant over yy, f(x)[y]f(x)[y]+g(x)f(x)[y] \rightarrow f(x)[y] + g(x), without affecting the conditional probabilities p(yx)p(y\vert x) due to the invariance of the softmax function to the addition of a constant. However, this changes the energy function for the density model to E(x)E(x)g(x)E(x) \rightarrow E(x) - g(x). Consequently, there exists a functional degree of freedom that can arbitrarily alter the energy function without impacting the discriminative loss optimized by a classifier. As a result, a classifier trained solely to distinguish between positive and negative samples is not inherently encouraged to learn a useful energy function for the density over xx. My interpretation of the results of [1], from which the authors seem to draw inspiration is that reinterpreting the logits of a classification model as an energy based model by itself accomplishes nothing if the training process of the original classifier is not changed to encourage a good density model to be learned and encoded into the free degree of freedom. [1] achieves this by optimizing the joint likelihood, Ex,ypθ(y,x)E_{x,y} p_\theta(y,x), instead of the conditional likelihood, Ex,ypθ(yx)E_{x,y} p_\theta(y\vert x), used in training classifiers. This significantly complicates training by requiring sampling from the EBM during training. Therefore, it is unclear why the authors assume that the pre-trained PFN has learned useful energy functions for density modeling when it is derived from a simple classifier and this is not discussed in the paper.

  • Another weakness of the paper is in the experimental setup where only default hyperparameters are used for all competing methods. For many of these methods, tuning hyperparameters can substantially improve their results and the default hyperparameters may not be optimal. Personally, I would find the presented results more convincing if the paper compared with fewer competing methods but made a better effort of getting good results out of those as it is impossible to know how much the authors tuned the degrees of freedom available during experimentation with their own proposed method.

[1] Will Grathwohl, Kuan-Chieh Wang, Jörn-Henrik Jacobsen, David Duvenaud, Mohammad Norouzi, Kevin Swersky. Your Classifier is Secretly an Energy Based Model and You Should Treat it Like One, 2022 (https://arxiv.org/abs/1912.03263)

问题

  • Can you provide a convincing explanation for why the logits from discriminative classifiers derived from TabPFN should be useful when reinterpreted as a class-conditional energy function in light of the first point above? Without a clear justification, it is difficult to understand the basis for expecting this method to work effectively.

  • Since the authors marginalize over yy by taking the LogSumExp over the y=0y=0 and y=1y=1 logits corresponding to negative and empirical class data, respectively, one would expect the energy function to represent a mixture distribution between these two densities. Why aren't the logits for y=1y=1 used directly as an energy function instead? In other words, what is the motivation for marginalizing over yy when it is simply an indicator of whether the data is real or negative?

  • The details of how the negative examples are generated are not clear. The text simply states: "These surrogate negative samples Xcneg\mathcal{X}^{neg}_c are constructed to be far from the distribution of real data, ensuring that the classifier can easily distinguish them. This placement prevents class ambiguity and facilitates a robust energy landscape". The appendix does provide some additional details, but it is still ambiguous how this data is generated. Given the potential importance of this process in understanding why the method works, could you provide more clarity on how the negative data is generated and discuss how changing the distribution of negative data affects the results?

  • The authors helpfully point out all datasets that were also used as test sets for TabPFN's evaluation since these could potentially be tainted if good practices were not followed on the original TabPFN paper [1]. I am more concerned however with datasets that were part of the meta-validation set that was used to tune the hyperparameter priors of TabPFN. It seems that one of the datasets used for evaluation of TabEBM (stock) is part of this meta-validation set (Tables 7 and 8 in the appendix of [1]). Even though this is only one of 150 datasets that comprise this validation set and it is not clear in the original paper the extent of this finetuning, I would still consider it good practice to avoid using this dataset for evaluating downstream methods for fear of data leakage. I also would have liked the paper to prioritize evaluating on more datasets that weren't used at all in [1]. I understand if [1] might have somewhat exhausted the pool of small datasets in OpenML but evaluating on larger datasets by first subsampling them would be an option. This would also allow for the use of real data as an ideal baseline to measure TabEBM's improvement against.

  • For TabEBM and the other methods that learn class conditional densities (CTGAN, TabDDPM, ARF, GOGGLE, TabPFGen) what distribution of classes (i.e., p(y)p(y)) was used for the synthetic data? This is not specified in the paper so I am assuming it was an empirical estimate of the p(y) from the training data otherwise this would be somewhat unfair when comparing to methods that learn a joint distribution. Furthermore it would be potentially problematic when evaluating a balanced accuracy in a test set if the p(y) of generated data is not the same across methods (for conditional methods).

  • How are statistical fidelity results computed? The metrics reported seem to be univariate ones so how are they being used to compare multivariate distributions? Are only marginal distributions being compared? Is it an average over univariate metrics computed for each marginal separately? How is the KS test used on categorical variables? This section of the experiments is sorely lacking in details and the appendices seem to provide no further information. If it is indeed the case that only marginals are being compared I find this evaluation lackluster since it is easy to achieve similar marginal distributions without learning the relationships between different input variables which is the main difficulty in learning multivariate density models.

[1] Noah Hollmann, Samuel Müller, Katharina Eggensperger, Frank Hutter. TabPFN: A Transformer That Solves Small Tabular Classification Problems in a Second, 2022 (https://arxiv.org/abs/2207.01848)

局限性

The authors aknowledge that due to the reliance of the method on a PFN it is currently constrained for use in small datasets.

作者回复

Thank you for the detailed review and constructive feedback! We address all your questions and comments below, and provide a rebuttal PDF in the general rebuttal. Due to space limits, we summarise the new results. We will update our manuscript with additional clarifications and results.

>Q3a: Clarify the generation of the negative samples

The surrogate classification task is to determine if a sample belongs to class cc by comparing Xc\mathcal{X}_c against a set of "negative samples" Xcneg\mathcal{X}_c^{\text{neg}}. We label the true samples Xc\mathcal{X}_c as 1 and the negative samples Xcneg\mathcal{X}_c^{\text{neg}} as 0, and train TabPFN on this datasets, resulting in a class-specific binary classifier used to define the Ec(x)E_c(x) (using Equation 3)

We generate the negative samples at the corners of a hypercube in RDR^D. For each dimension dd, the coordinates of a negative sample are either ασd\alpha\sigma_d or ασd-\alpha\sigma_d, where α\alpha is a fixed constant and σd\sigma_d is the standard deviation of dimension dd. For example, in R3R^3, a negative sample might have coordinates [ασ1,ασ2,ασ3][\alpha\sigma_1, \alpha\sigma_2, -\alpha\sigma_3]. In the paper we use four negative samples with α=5\alpha = 5, placing them far from the real data.

>Q3b: How does the distribution of negative data affect the results?

We found it essential to use negative samples far from the real data, to allow the binary classifier to easily differentiate between Xc\mathcal{X}_c and Xcneg\mathcal{X}_c^{\text{neg}}.

Figure F1 from the PDF shows a new experiment varying the distribution of the negative samples. TabEBM infers an accurate energy surface with distant negative samples, and the energy surface becomes inaccurate when negative samples resemble real samples. This occurs because TabPFN is uncertain, affecting its logits magnitude and making them unsuitable for density estimation (we further investigate the logits in our next answer to Q1/W1).

TabEBM is robust to the distribution of negative samples if they are easily distinguishable from real data. We ran two new experiments, keeping the negative samples at the corners of a hypercube and varying their distance and number. When varying the per-dimension distance of those samples α[0.1,5]\alpha \in [0.1, 5], TabEBM provides consistent improvements between 2.85-3.18%. Additionally, our results in Table R2 (in our answer to Q1 from reviewer Cc3y) demonstrate that TabEBM performs similarly regardless of the number of negative samples in the surrogate binary tasks.

>Q1/W1: Why can TabPFN's logits be useful for energy estimation?

Indeed, we agree with your observation! As the TabPFN classifier is trained on different class-specific surrogate tasks, it learns different logits for each task. We found it essential to place the negative samples far from the real samples, because TabPFN's confidence depends on the distance to the training data [1], as it was pre-trained to approximate Bayesian inference.

We present new experiments in Figure F2 from the PDF. The top row shows that as the distance to the real data increases, the logit f(x)[1]f(x)[1] for the real data smoothly decreases until the two logits become similar. Thus, the classifier predicts is uncertain in these far-away regions. As maximum logits decrease, TabEBM's inferred density drops significantly (as shown on the bottom row) because pc(x)(exp(f(x)[0])+exp(f(x)[1]))p_c(x) \propto (\exp(f(x)[0]) + \exp(f(x)[1])). One could possibly fine-tune TabPFN on the surrogate tasks to further improve the logits for density estimation.

>Q2: Why fit p(x) rather than p(x|y=1) in the surrogate binary tasks?

Both approaches lead to virtually identical results due to the design of our surrogate task. The new results in Figure F2 (bottom row) from the PDF shows that the TabEBM's inferred energy using p(x)p(x) (in red) is essentially identical to the energy of p(xy=1)p(x|y=1) (in grey), especially near the real data. As the SGLD sampling starts near the original points (line 676), it will visit the neighbourhood of the real samples, where the energy surfaces are virtually identical, leading to similar results.

>Q4: Evaluating more datasets

We run new experiments on six leakage-free datasets from UCI, with 1,000-9105 samples of 7-42 features. Table R2 (in our response to Q2 for tdoQ) shows that TabEBM consistently outperforms the baseline and all other benchmark methods.

>Q5: What's the class distribution of the synthetic data?

The synthetic data has the same class distribution as the real training data.

>Q6: Is statistical fidelity computed over the univariate marginals?

Yes, these metrics are univariate, and we computed them using the open-source library Synthcity. We acknowledge the reviewer's point about the imperfection of univariate metrics, although evaluating generators' ability to capture the joint feature relationships remains underexplored [2].

>W2: How does tuning the generators/predictors affect the ranking?

We run new experiments tuning three generators on three datasets using the average validation accuracy of six downstream classifiers. The ranges are: SMOTE (k{3,5,10}k \in \{3, 5, 10\}), TVAE (lr{5e4,1e3,5e3}\text{lr} \in \{5e-4, 1e-3, 5e-3\}) and CTGAN (lr{1e3,5e3,1e2}\text{lr} \in \{1e-3, 5e-3, 1e-2\}). TabEBM remains the most competitive data augmentation method, improving accuracy by +2.25, followed by SMOTE (+1.81) and CTGAN (-5.71). We can provide the full results in the discussion.

We also tuned the downstream predictors, and the new results in Table R7 from the PDF show that TabEBM remains the best-performing method, providing the largest improvements for data augmentation, even after tuning the predictors.

Thank you again for your constructive review! We would appreciate it if you would consider raising your score in light of our response.

References:

评论

I thank the authors for the clarifications and for taking the time to run these additional experiments.

I must confess that the reason this approach works is still somewhat mysterious to me and I would like to have a better theoretical justification for it. However, the linked blogpost and the author's additional experiments do shed some light on the matter and seem to suggest that this may be due to some property of TabPFN. The authors have, at least, successfully convinced me that their surrogate task procedure works and is somewhat reliable and I believe these are valuable empirical results.

I believe the clarification on the generation of negative data should be in the paper.

Since I consider my concerns mostly addressed, I am happy to raise my score to a 6.

审稿意见
5

The paper proposes TabEBM, a novel data augmentation method designed for low-sample-size tabular classification tasks. TabEBM generates synthetic tabular data using class-specific Energy-Based Models (EBMs) to learn the marginal distribution for each class. Experimental results on various real-world datasets demonstrate that TabEBM improves downstream performance via data augmentation, generates high-fidelity synthetic data, and strikes a competitive balance between accuracy and privacy in data sharing.

优点

  • Novel approach using class-specific EBMs for tabular data generation.
  • Comprehensive evaluation across multiple datasets, metrics, and downstream tasks.
  • Strong performance, especially in low data regimes.
  • Thorough analysis of statistical fidelity and privacy preservation.
  • Open-source implementation provided.

缺点

  • Scalability Issues: The reliance on TabPFN, which struggles with large sample sizes, limits TabEBM's scalability.
  • Implementation Complexity: The need for class-specific surrogate tasks and the iterative nature of the sampling process may complicate implementation and increase computational overhead.

问题

  • How does TabEBM perform on datasets with hundreds of features?
  • Could the approach be extended to use other pre-trained models besides TabPFN?
  • How does the method handle highly imbalanced datasets where some classes have significantly fewer samples than others?
  • What is the computational complexity of generating samples compared to other methods?
  • How sensitive is the performance to the choice of hyperparameters in the SGLD sampling process?
  • Clarify the meaning of the word "fit" which is used in line 122 of the paper.

局限性

  • Performance may degrade for datasets much larger than those tested.
  • Inherits limitations of the underlying TabPFN model.
  • May not be suitable for datasets with extremely high dimensionality.
作者回复

Thank you for your thoughtful review! We address all your questions and comments below. Due to limited space, we summarised the new results. We will update the manuscript to include the complete new experiments and clarifications.

>Q3: How does TabEBM perform on highly imbalanced datasets?

We conducted new experiments, adjusting the class imbalance on two binary datasets (with Nreal=100N_\text{real}=100). We keep the setup from Section 3 and report the balanced accuracy averaged over six downstream predictors. Table R4 shows that under high class imbalanced, TabEBM outperforms the Baseline and SMOTE, which is a method specifically designed to handle imbalanced datasets.

Table R4. Test balanced classification accuracy (%) varying the class imbalance.

Datasets(#Minority:#Majority)BaselineSMOTETabEBM(Ours)
steel(50:50)88.7977.7493.77
(10:90)60.7460.7673.02
(5:95)52.2956.1165.95
stock(50:50)88.6490.2190.16
(10:90)76.0583.8684.92
(5:95)65.9875.2477.68

>Q5: How sensitive is TabEBM to the hyperparameters of the SGLD sampling?

We run new experiments varying two key hyperparameters of SGLD on the “biodeg” dataset with Nreal=100N_{\text{real}}=100. Specifically, we vary the step size αstep\alpha_{\text{step}} and the noise scale αnoise\alpha_{\text{noise}}, reporting the accuracy averaged over six downstream predictors.

Table R5 shows that TabEBM is stable to these hyperparameters, as the difference between the highest and lowest accuracy is less than 1.5%. Note that increasing αnoise\alpha_{\text{noise}} (which is added at each SGLD step) is expected to degrade performance because we standardized the data to have unit standard deviation.

Table R5. Classification accuracy (%) of TabEBM with different SGLD settings

αstep\alpha_{\text{step}}\rightarrow0.10.30.51.0
αnoise\alpha_{\text{noise}}\downarrow
0.0176.4577.0977.0476.58
0.0276.8676.9676.7776.26
0.0575.9375.8975.9475.70

>Q6: Clarify the meaning of the word "fit" on line 122

We use the term “fit” to mean “train” the in-context model TabPFN. For TabPFN, “training” is similar to the K Nearest Neighbour algorithm, where "training" means defining the training dataset used for inference. Thus, we simply update TabPFN's training datasets and the model in inference mode, thus never updating its parameters.

>Q4: How does TabEBM's computational complexity compare to other methods?

Figure 4 from the paper illustrates the trade-off between accuracy and the time required for training and generating 500 synthetic samples. The results show that TabEBM is the fastest method for generating data besides SMOTE. Other methods are 3-30 times slower than TabEBM. Also, TabEBM surpasses all other methods in improving downstream accuracy through data augmentation.

>W2: Class-specific surrogate tasks and iterative sampling can complicate implementation and increase computational overhead

On implementation complexity: TabEBM simplifies implementation by eliminating the need for dedicated training, pipelines, protocols, and resources—making it ready to use for generating data. We provide an open-source implementation of TabEBM, allowing users to generate data with just two lines of code for new datasets. We believe this ease of use will encourage the widespread application of tabular data augmentation on small datasets.

On the computational overhead: As mentioned in our answer to Q4 and Figure 4, TabEBM is computationally efficient, being the fastest method for generating data besides SMOTE —while outperforming all competing methods.

>Q2: Can TabEBM use other pre-trained models besides TabPFN?

Yes, TabEBM can be used with other pre-trained models besides TabPFN. In the paper, we use TabPFN because it is the only tabular model of this type with open-sourced weights. However, TabEBM is a general method for transforming classifiers into class-specific generators and can use any gradient-based in-context classifier that computes logits (through Equation 3). As new tabular foundational models are developed [1], they can be readily integrated into TabEBM.

>Q1/W1/L1/L2/L3: On the scalability and limitations of TabEBM

TabEBM is a method that uses an underlying binary classifier, thus its capabilities and limitations are directly influenced by the in-context model it employs (TabPFN in our case), as discussed in the paper (Lines 317-323). Since TabPFN handles only up to 100 features, this limitation is also inherited by TabEBM.

Our new experiment (Table R6) shows that TabEBM can be applied to larger datasets beyond TabPFN's limitation of 1000 samples. On larger datasets TabEBM still outperforms other generators (not shown due to space limits), but training on real data alone appears sufficient. This highlights TabEBM's usefulness in fields with limited training samples. Note that despite TabPFN's 10-class limit, TabEBM can handle unlimited classes by using TabPFN only for surrogate binary tasks.

As foundational models for tabular data evolve [1], new models that can accommodate more features are anticipated. Integrating these models into the TabEBM framework will enable it to handle high-dimensional datasets, thus increasing its versatility and utility.

Table R6. Classification accuracy (%) comparing data augmentation with increased real data availability on the “texture” dataset

NrealN_{\text{real}}BaselineImprovement by TabEBM (%)
5072.40+6.50
10082.42+3.59
100096.37-0.07
200097.76+0.07
300098.20+0.15
400098.51+0.04

Thank you again for your thoughtful review! We would appreciate it if you would consider raising your score in light of our response.

References:

评论

Thank you to the authors for clarifying and conducting additional experiments. These efforts have significantly enhanced the quality of the research. However, I believe my initial evaluation remains valid, so my score remains unchanged.

评论

Dear Reviewer iojh,

Thank you for acknowledging the additional experiments and clarifications we provided. We are glad that you found that our efforts have significantly enhanced the quality of our research. We are also pleased that, in your original review, you highlighted several strengths of our work, including its novelty, comprehensive evaluation, strong performance, and open-source implementation.

We conducted additional experiments to further support our argument, which we believe will interest the reviewer. Specifically, we evaluated four additional generative models—TVAE, CTGAN, TabDDPM, and TabPFGen—addressing your questions on handling imbalanced datasets and the effectiveness of data augmentation on large datasets.

  • Table R5-extended demonstrates that TabEBM outperforms these methods, especially on highly imbalanced datasets. The results indicate that TabEBM is particularly effective for class balancing, and we will highlight this application in the revised manuscript.
  • Table R6-extended suggests that data augmentation may not be necessary when ample real data is available for model training. As the availability of real data increases, training downstream predictors solely on this real data can guarantee optimal performance. Existing methods, such as TVAE, CTGAN and TabDDPM, tend to decrease performance on smaller datasets, where performance enhancement is most needed, while TabEBM consistently delivers the largest improvements. This suggests that data augmentation is beneficial primarily in data-scarce scenarios, where TabEBM excels. We will include these results in the updated manuscript.

Given the new results and your positive feedback, we kindly ask if you might reconsider your initial score in light of these results. If there are any remaining concerns, please let us know, as we would be more than happy to discuss them further while the discussion period is open. We would greatly appreciate your feedback in order to further improve our paper.

Thank you again for your time and consideration,

Authors

Table R5-extended. Test balanced classification accuracy (%) varying the class imbalance. We aggregated the performance over six downstream predictors. TabEBM is effective for datasets with high class imbalance.

Datasets (#Minority:#Majority)BaselineSMOTETVAECTGANTabDDPMTabPFGenTabEBM (Ours)
steel
(50:50)88.7977.7474.2177.6183.7889.6793.77
(20:80)74.3366.5668.1061.5269.0683.3884.65
(10:90)60.7460.7654.8553.5058.4670.3673.02
(5:95)52.2956.1151.1251.3051.1764.8465.95
stock
(50:50)88.6490.2188.6785.5289.2388.1690.16
(20:80)85.3488.9585.8785.3787.1688.7289.38
(10:90)76.0583.8674.6076.9378.8183.5284.92
(5:95)65.9875.2462.6661.1168.6975.5677.68

Table R6-extended. Test balanced classification accuracy (%) aggregated over six downstream predictors, comparing data augmentation with increased real data availability of the “texture” dataset. TabPFGen is not applicable because it supports only datasets with up to 10 classes.

N_realBaselineSMOTETVAECTGANTabDDPMTabPFGenTabEBM (Ours)Improvement vs Baseline by TabEBM (%)
5072.4076.4055.33254.8062.94N/A78.90+6.50
10082.4284.3566.0069.4976.34N/A86.01+3.59
20087.5489.2978.3782.4482.53N/A89.77+2.23
50092.9693.6990.0991.4891.24N/A93.76+0.80
100096.3796.2193.6195.3694.56N/A96.30-0.07
200097.7696.8496.6297.1097.13N/A97.83+0.07
300098.2098.2897.6097.6097.73N/A98.35+0.15
400098.5198.5998.1198.0098.46N/A98.55+0.04
审稿意见
7

The authors present a new method of tabular data augmentation, called TabEBM. The unique feature of TabEBM is that it creates distinct generative models for each class in a classification problem setting. With extensive and thorough evaluations, the authors prove that TabEBM sets the new state of the art.

优点

  1. The manuscript is well organized. The text is clear. The quality of the figures is high.
  2. The authors address a very common problem for the broad scientific community. Application of machine learning algorithms on small tabular datasets is limited by design. Introducing TabEBM as an effective data augmentation method, this work can have an extremely broad impact in the future.
  3. The experiments are rich and rigorous. The method is compared against many existing methods on a variety of datasets. The evidence proving its effectiveness is convincing.
  4. The authors claim TabEBM will be available as an open-source library.

缺点

Some important details on study design, model training, and practical considerations of TabEBM application are missing. See questions.

问题

  1. The definition of a surrogate binary classification task (lines 111-113) might result in class imbalance. How is this potential problem addressed?
  2. Can the authors provide more details on training the models for the surrogate binary classification task? This could eliminate some other questions.
  3. Have the authors specifically investigated the propensity of TabEBM to generate outliers? How does TabEBM compare to other methods in this regard?
  4. In section 3.2, the authors describe many statistical evaluations performed. However, the multiple hypothesis testing aspect is not described in sufficient detail. How were the p-values adjusted? How strong is the impact of a particular correction method on the final results (tables of C3)?
  5. Can the authors share more insights about the open-source library? This is positioned as one of the key contributions, but the code is not provided and the practical considerations of using the library are not clear from the submission.

局限性

Briefly discussed in section 4.

作者回复

Thank you for your thoughtful review! We address all of your questions and comments below. Due to space constraints, we summarised the new results. We will update the manuscript to include the complete new experiments and clarifications.

>Q2: Provide details on training models for the proposed surrogate binary classification tasks

In TabEBM, for each class cc, we train a class-specific EBM, Ec(x)E_c(x), using exclusively the data from that class, denoted as Xc\mathcal{X}_c. The energy function Ec(x)E_c(x) is derived by adapting the logits from a binary classifier, as outlined in Equation 3. To train this classifier with only class-specific data, we create surrogate binary classification tasks.

Our surrogate classification tasks determine if a sample belongs to class cc by comparing Xc\mathcal{X}_c against a set of "negative samples," Xcneg\mathcal{X}_c^{\text{neg}}. We intentionally create these negative samples at a distance of five standard deviations from the original data to ensure the binary classifier can easily differentiate between the real class data and these artificially distant samples.

We label the true class samples Xc\mathcal{X}_c as 1 and the negative samples Xcneg\mathcal{X}_c^{\text{neg}} as 0, forming a combined dataset Dc=(XcXcneg,{1}Xc{0}Xcneg)\mathcal{D}_c = (\mathcal{X}_c \cup \mathcal{X}_c^{\text{neg}}, \{1\}^{|\mathcal{X}_c|} \cup \{0\}^{|\mathcal{X}_c^{\text{neg}}|}). Following this, we train the TabPFN classifier on Dc\mathcal{D}_c, resulting in a binary classifier specific to class cc. This classifier is then used to define the energy function Ec(x)E_c(x) using Equation 3, from which we can generate data using SGLD.

>Q1: Does the surrogate binary classification task result in class imbalance issues?

No, the imbalance between the "negative samples" and the real samples used in the surrogate binary classification task has a negligible impact on TabEBM's performance.

We conducted a new experiment to evaluate the effects of varying the ratio Xcneg:Xc\|\mathcal{X}^{\text{neg}}_c\|:\|\mathcal{X}_c\|. To ensure reliable outcomes, we maintained a consistent ratio across all classes, keeping the same proportion of negative samples for each class. We varied the number of negative samples Xcneg\|\mathcal{X}^{\text{neg}}_c\| to represent various ratios of Xc\|\mathcal{X}_c\|, thus simulating both balanced and highly imbalanced scenarios.

The results, presented in Table R3, computed across six datasets with Nreal=100N_{\text{real}} = 100 real samples, demonstrate that TabEBM is robust to potential imbalances in the surrogate binary tasks.

Table R3. Classification accuracy (%) on data augmentation, showing the impact of the number of negative samples Xcneg\|\mathcal{X}^{\text{neg}}_c\| in TabEBM across six datasets. TabEBM's performance is robust regarding the number of negative samples.

RatioXcneg:Xc\|\mathcal{X}^{\text{neg}}_c\|:\|\mathcal{X}_c\|10.50.20.1FixedXcneg=4\|\mathcal{X}^{\text{neg}}_c\|=4
Average accuracy improvments+2.91+2.87+2.88+2.92+2.90

>Q3: TabEBM's propensity to generate outliers

We provide fair and thorough comparisons by (i) including a wide range of evaluation metrics and (ii) utilising a large synthetic set (Lines 183-184). Given our extensive test scope (i.e., eight datasets ×\times five sample sizes), it is computationally infeasible to employ complex and specialised methods to detect outliers in synthetic data. However, the downstream accuracy and statistical fidelity metrics are highly indicative of the distribution shifts between real and synthetic data. Notably, Figure 3 illustrates that TabEBM consistently outperforms the baseline across various sample sizes and class numbers. We believe these two metrics adequately demonstrate that TabEBM is more stable to generate in-distribution samples than benchmark methods.

>Q4: Details on the multiple hypothesis testing in Section 3.2

We aim to provide a fair and coherent comparison between TabEBM and existing methods, and thus, we follow the widely-adopted evaluation process in prior studies. Specifically, we compute the statistical fidelity metrics with the open-source implementations from the well-established open-source library Synthcity. However, the previous studies often operate under the assumption that the issues associated with multiple comparisons are less pronounced in generating low-dimensional tabular data, hence correction methods for multiple hypothesis testing are seldom employed. We will further clarify this in the revised manuscript.

>Q5: The code for the open-source library is not provided

We included the TabEBM library as a zip file in the Supplementary material. The library has two core functionalities:

  1. Generate synthetic data: The library can generate data that can be used as additional training material for data augmentation. We have included a demo notebook, TabEBM_generate_data.ipynb, in the attached codebase. This notebook demonstrates how to use the generated data for data augmentation and shows a toy example comparing the quality and position of the synthetic data with the real data.
from TabEBM import TabEBM

tabebm = TabEBM()
augmented_data = tabebm.generate(X_train, y_train, num_samples=100)
# augmented_data[class_id] = generated data for a specific `class_id`
  1. Compute & visualise the energy function: We allow users to compute TabEBM's energy function and the unnormalised data density. The demo notebook, TabEBM_approximated_density.ipynb, shows TabEBM's energy under conditions of data noise and class imbalance.

We will make the TabEBM library freely available after publication. It is easy to use, domain-agnostic, and requires no training, making it suitable for data augmentation, especially for small-sized datasets.

Thank you again for your thoughtful review! We would appreciate it if you would consider raising your score in light of our response.

评论

Thanks to the authors for the clarifications and additional experiments! I believe my initial evaluation was fair, so my score remains the same.

审稿意见
5

The paper introduces TabEBM, a class-conditional generative method for tabular data augmentation using distinct EBMs for each class. By modeling each class's data distribution individually, TabEBM generates high-quality synthetic data. Experiments demonstrate that using TabEBM for data augmentation improves classification performance across various datasets, particularly small ones.

优点

The paper introduces distinct class-specific EBMs for tabular data augmentation. Moreover, it presents thorough experiments demonstrating effectiveness across various datasets, especially small ones. Additionally, the writing is well-structured and clear.

缺点

  1. On the technical level, TabPFN, EBMs, and SGLD have all been introduced in TabPFGen [48]. The main difference in the proposed TabEBM method is the use of class-specific EBMs. However, this idea is straightforward and lacks novelty.
  2. The most significant difference between tabular tasks and common tasks like images is that they include both continuous and categorical features. However, TabEBM treats all features as continuous, lacking targeted consideration for categorical features.
  3. In TabEBM, although the choice of surrogate binary classifier is arbitrary, not using training-free models like TabPFN will result in training costs being positively correlated with the number of classes, significantly increasing the training overhead.
  4. The datasets used in the experiments have a maximum dimensionality of 77, lacking experimental results on high-dimensional data (e.g., hundreds of dimensions).

问题

  1. Do the datasets used in the experiments only contain continuous features and no categorical features?
  2. The TabEBM method is similar to TabPFGen [48]. Why didn't you use the 18 datasets mentioned in the TabPFGen paper for experiments, but instead chose 6 different datasets (with a smaller test scope) for evaluation?

局限性

N/A

作者回复

Thank you for your thoughtful review! We address all of your questions and comments below. Due to space constraints, we summarised the new results. We will update the manuscript to include the complete new experiments and clarifications.

>Q1/W2: Do the datasets contain categorical features? How does TabEBM handle them?

Yes, we evaluated TabEBM on datasets containing both continuous and categorical features. Specifically, Table R1 shows the two considered datasets with mixed feature types. In lines 639-640, we also clarify that TabEBM encodes the categorical features with leave-one-out target statistics. As the reviewer mentioned, the evaluation results (Table 1) show that TabEBM is effective across various datasets, including those containing categorical features.

Table R1. Number of categorical features in the datasets considered.

Dataset#Categorical
Main text
protein4
energy1
New datasets in our response to Q2
support27
mushroom22
abalone1
statlog13

>Q2: Why not evaluate the 18 datasets in the TabPFGen paper?

Firstly, we wanted to avoid data leakage and ensure a robust comparison by also evaluating datasets not in TabPFN's meta-test (Appendix F.3 in its original paper). As the 18 datasets in the TabPFGen paper were all used for constructing and evaluating TabPFN, we also use different, leakage-free datasets. Secondly, our test scope is not smaller than that of the TabPFGen paper. We evaluate a wide range of test scopes, considering different datasets and varying the availability of real data. We also include datasets with categorical data, unlike TabPFGen, which investigates only numerical features. Our setup (i.e., eight datasets × five sample sizes) leads up to 33 different test cases. For reference, ARF has the most test cases among prior studies, with 20 test cases, while TabPFGen follows with 18 test cases. Thirdly, we further provide new results on six more leakage-free datasets for UCI with 7-42 features, some including categorical features. We set Nreal=100N_\text{real}=100 and provide results for the Top-5 benchmark methods. Table R2 shows that TabEBM continues to outperform all other benchmark methods.

Table R2. Classification accuracy (%) comparing data augmentation on six new leakage-free datasets.

DatasetsBaselineSMOTETVAECTGANTabDDPMTabPFGenTabEBM (Ours)
clinical68.6371.0761.8065.2154.0369.6671.20
support264.2365.6060.7059.1458.3164.3465.28
mushroom95.5195.8493.7593.2679.8797.0596.82
auction51.9057.3553.0952.3551.1456.8257.97
abalone11.59N/A8.497.729.95N/A13.56
statlog56.2257.3053.1255.5553.0757.6557.85

>W1: The core ideas “have all been introduced in TabPFGen” … "this idea is straightforward and lacks novelty”

TabEBM makes a novel contribution to the field of tabular data augmentation by being the first to introduce class-specific generators, as we discussed in our related work, as well as acknowledged by all other reviewers.

Specifically, TabEBM is the first method to create distinct class-specific EBMs. It consists of individual models—one for each class—designed to learn distinct marginal distributions for the inputs associated with each class. TabEBM's class-specific generation is not available to TabPFGen, and it is only possible due to our proposed surrogate binary tasks, which enable the creation of a binary classification task for each class and the obtaining of the class-specific EBM via Equation 3. The results demonstrate that training class-specific EBMs significantly improves the inferred energy over TabPFGen under high levels of noise (Figure 7) and high data imbalance (Figure 8). Unlike TabPFGen, which supports only 10 classes, TabEBM can handle an unlimited number of classes. As you rightly point out, our extensive experiments demonstrate that our method, with its class-specific generation, outperforms all other methods in accuracy, statistical fidelity, and privacy.

In addition to the method, our paper makes two additional key contributions. First, we conduct the first extensive analysis of tabular data augmentation across various dataset sizes. Our benchmark reveals that existing tabular augmentation methods may improve performance only on small datasets, a previously uninvestigated issue. Second, we will release our method as an open-source library for tabular data augmentation, allowing users to generate data immediately without additional training.

>W3: “In TabEBM … not using training-free models like TabPFN will result in … significantly increasing the training overhead.”

Because we tackle classification problems on small-size datasets, training models from scratch is not practical, often leading to suboptimal performance due to overfitting. Instead, using pre-trained models, such as TabPFN, is integral to our approach. From a practical standpoint, using pre-trained models makes TabEBM immediately usable. A significant limitation of existing tabular generators is their need for training, which can be time-consuming, error-prone, and typically requires GPUs. In contrast, our TabEBM library (included with the submission) allows users to start generating data immediately without additional training or tuning.

>W4: Lacking experimental results on high-dimensional data

As noted in Lines 317-323, scaling up in-context tabular models for high-dimensional data remains an unresolved challenge and is not the primary focus of our study. Instead, TabEBM is designed to demonstrate the effectiveness of using pretrained in-context tabular classifiers for generating tabular data. As acknowledged by the reviewer, and further shown in our response to Q2, our "thorough experiments" demonstrate TabEBM's effectiveness.

Thank you again for your thoughtful review! We would appreciate it if you would consider raising your score in light of our response.

评论

Dear Reviewer tdoQ,

Thank you once more for taking the time to provide your feedback on our work. In light of our response and the new experiments, and with the discussion window closing soon, please let us know if you have any further questions or if there is anything else we can clarify. If not, we would appreciate it if you would consider updating your review based on our rebuttal.

Thank you,

Authors

评论

I appreciate the additional experiments conducted by the author. These evaluations addressed my concerns regarding the experiment scale and performance for datasets with many categorical features. I have increased my overall rating for this paper from Reject to Borderline Accept.

作者回复

We thank the reviewers for the feedback!

(1) Summary of positive things

  • Novel method
    • Cc3y: “a new method of tabular data augmentation”; “the unique feature is that it creates distinct generative models for each class”
    • iojh: “Novel approach using class-specific EBMs for tabular data generation”
    • XdB4: “a novel approach that utilizes separate energy-based models (EBMs) to generate synthetic data for each class”
  • Comprehensive evaluation showing TabEBM's effectiveness
    • tdoQ: “thorough experiments demonstrating effectiveness”
    • Cc3y: “extensive and thorough evaluations”; ”TabEBM sets the new state of the art”
    • iojh: ”Comprehensive evaluation across multiple datasets, metrics, and downstream tasks”; “TabEBM improves downstream performance … generates high-fidelity synthetic data”
    • XdB4: “The experimental results section is thorough and comprehensive”
  • Clear presentation
    • tdoQ: “well-structured and clear”
    • Cc3y: “The manuscript is well organized”; “The text is clear”; “The quality of the figures is high”
    • XdB4: “The paper is well-structured and clearly written, effectively communicating the key ideas and concepts.”
  • Impactful method
    • Cc3y: “The authors address a very common problem for the broad scientific community”; “this work can have an extremely broad impact in the future”
  • Reproducibility
    • iojh: “Open-source implementation provided”

(2) Summary of our responses and new experiments

We replied to all questions and concerns raised by the reviewers:

Table R0. Number of our responses to the reviewers’ comments (#Raised/#Replied).

Reviewer# Questions# Weaknesses# Limitations# New experiments
tdoQ2/24/4N/A1
Cc3y5/5N/AN/A1
iojh6/62/23/33
XdB46/62/2N/A5

We provide 10 new experiments and attach a rebuttal PDF. The new results are consistent with the main text and further support TabEBM’s effectiveness. We will add them to the revised manuscript. Below, we detail the new experiments for each reviewer.

  1. [tdoQ, XdB4] Evaluation on six new, leakage-free datasets with mixed feature types.
  2. [Cc3y] Ablation studies on the surrogate binary classification task w.r.t. the imbalance between the negative samples and the real samples.
  3. [iojh] Evaluation on highly imbalanced datasets to show TabEBM’s robustness.
  4. [iojh] Ablation studies on TabEBM’s sensitivity to the hyperparameters of the SGLD sampling process.
  5. [iojh] Evaluation on larger datasets to show TabEBM’s scalability.
  6. [XdB4] Ablation studies on the distance of the negative samples and the real samples.
  7. [XdB4] Evaluating the impact of the distribution of the negative samples.
  8. [XdB4] Investigation into TabPFN's logit values on the surrogate binary tasks.
  9. [XdB4] Ablation studies on tuning the hyper-parameters of generative benchmarks.
  10. [XdB4] Ablation studies on tuning the downstream predictors.
最终决定

The authors propose a generative method for tabular data augmentation, utilizing a distinct generative process for each class in a classification setting.

  • Reviews: The paper received mostly positive scores, though some were near and one below the threshold. The weaknesses noted were in the areas of novelty, clarity, and missing technical details.
  • Rebuttal: The authors made significant efforts to address the reviewers' concerns, particularly by providing additional experiments. As a result, two reviewers raised their scores. It appears that the authors successfully addressed most of the points that were raised.
  • Post-decision Revision: As noted by one reviewer,
    • the new dataset collins seems to have been used to tune TabPFN and hence is not meaningful to derive conclusions from;
    • the statistical fidelity metrics should be better explained, as it is 1) not clear how they are aggregated, 2) the KS metric only considers numerical features, and 3) the KL is potentially based only on categorical features.