PaperHub
6.8
/10
Poster4 位审稿人
最低3最高5标准差0.8
5
4
5
3
3.8
置信度
创新性3.0
质量3.0
清晰度2.5
重要性2.8
NeurIPS 2025

Inverse Methods for Missing Data Imputation

OpenReviewPDF
提交: 2025-05-12更新: 2025-10-29

摘要

关键词
missing data imputationmissing data completionkernelridge regressionnon-missing feature

评审与讨论

审稿意见
5

This paper presents Kernel Point Imputation (KPI), a novel method that addresses key weaknesses in iterative imputation. It uses a bi-level optimization framework to tackle model misspecification by adapting model forms within an RKHS and leverages fully observed oracle features for supervision.

优缺点分析

Strengths: The method's key strengths are its innovative bi-level optimization framework, strong empirical results outperforming established baselines, and a principled method for leveraging oracle features as supervision, which existing methods often neglect. The result is fair. The algorithm is supported by theoretical justification

Weaknesses: The primary weaknesses are potential scalability issues due to computationally expensive kernel matrix inversion and the use of a heuristic kernel ensemble strategy.
experiments details lacked

问题

  • The word oracle feature is not properly defined. Are they just observed features? Then the word "oracle" seems wordy. Or is it defined by the features with minimal missing values, as in line 94? Also, in line 94, the phrase "minimal" seems unclear.
  • How does the computational cost of KPI scale compared to baselines, given the expensive kernel matrix inversion? No running time is reported. The dataset's details and relevant data citations are not given. So, it's not certain that the method can be scaled to big datasets.
  • The theoretical results should be placed in the main text, while the proofs can be given in the appendix.
  • the missing rate explored are between 0.1-0.4, lacks of analysis for high missing percentages.
  • regarding noise in datasets as in line 305, it has been studied that missing data imputation can have a denoising property, at least for time series, as in Missing data imputation for noisy time-series data and applications in healthcare LP Le, XHN Thi, T Nguyen, MA Riegler, P Halvorsen, BT Nguyen arXiv preprint arXiv:2412.11164
  • the related works should be improved

局限性

NA. the author says that "There is no societal impact of the work performed."

最终评判理由

The authors have addressed my concern very thoroughly with promising newly updated results. So, I recommend accepting the paper.

格式问题

no concern

作者回复

Thank you very much for your encouraging support and appreciation of our empirical results and theoretical justification. Below are our responses to the specific query raised.

We observed that some comments in the weaknesses and questions sections convey similar points. We will selectively merge them to make the response concise and easy to follow.


[Q1] The word oracle feature is not properly defined. Are they just observed features? Then the word "oracle" seems wordy. Or is it defined by the features with minimal missing values, as in line 94? Also, in line 94, the phrase "minimal" seems unclear.

Response. Thank you very much for your concrete and meticulous comment.

  • Your understanding is absolutely correct: in our work, "oracle features" are defined as features that are fully observed, i.e., with no missing entries across all samples. We agree that the word "oracle" is wordy and will replace oracle features with fully observed features in the revised manuscript.
  • Moving forward, we understand that despite ambiguity of fully observed features, there are some scenarios without fully observed features. In this case, the scope of "oracle features" may be relaxed to include those features with very few (minimal) missing entries, which can be solved through suitable imputation prior to KPI process. We will discern this operational flexibility from the formal definition in the revised manuscript.

[Q2,W1] How does the computational cost of KPI scale compared to baselines, given the expensive kernel matrix inversion? No running time is reported. The dataset's details and relevant data citations are not given. So, it's not certain that the method can be scaled to big datasets.

Response. We express our gratitude for your meticulous and actionable feedback. There are two concerns in this query and we address them as follows:

  • Dataset details and citation. The datasets we use are subsets of varying scales from the UCI repository. We will provide detailed descriptions and proper citations for these datasets [1] in the revised manuscript.
  • Computational cost and scalability concern.
    • KPI updates imputation values by minimizing the loss function (Eq. 8) in a batch-wise manner. Therefore, the complexity is determined by the batch size (which defines the kernel matrix size), rather than by the total number of samples, which enables scalability to large-scale datasets.
    • On the basis, we add experiments to investigate the running time per iteration given different batch size. The results in the table below indicate that the running time remains under 3.3.295 ms for a batch size of 1024, demonstrating KPI’s efficiency. Meanwhile, the table demonstrates that the complexity is less relevant to the feature number, since altering the number of features does not affect the kernel matrix size.
    • Collectively, KPI's computational cost per iteration is primarily related to batch size and not to dataset size or feature dimensionality, making it scalable to large-scale datasets.
Batch size32642565121024
Gaussian kernel2.1212.0612.1142.6233.295
Polynomial kernel1.3781.3011.3811.8593.041
Feature number163264128256
Gaussian kernel2.0642.0622.0642.0682.223
Polynomial kernel1.3071.3111.3291.3111.302

[Q3] The theoretical results should be placed in the main text, while the proofs can be given in the appendix.

Response. Thank you very much for your thoughtful suggestion. We fully agree that presenting the main theoretical results in the main text will enhance the clarity and impact of our manuscript. To address this, we will explicitly include the key theorems at the end of Section 3.4, ensuring that readers can readily access the central theoretical contributions without referring to the appendix. The detailed proofs will be provided in the appendix for completeness. We will also revise the "Theoretical justification" paragraph (lines 192-194) to reflect these changes and to clearly highlight the implications of our theoretical findings. We appreciate your guidance and are committed to making these improvements in the revised manuscript.

[Q4] The missing rate explored are between 0.1-0.4, lacks of analysis for high missing percentages.

Response. We express our gratitude for your constructive and actionable comment. We have added the results for a high missing ratio of 0.6 in the table below. Overall, KPI continues to outperform baseline methods even under this more challenging condition, highlighting its robustness to high missingness. We sincerely appreciate your valuable feedback and will update the manuscript accordingly to provide a more comprehensive evaluation.

DatasetsBTCCCBVISPSQBWQW
MAEWASSMAEWASSMAEWASSMAEWASSMAEWASSMAEWASSMAEWASS
CSDI-T0.7461.6470.9313.0450.8434.1110.77215.4190.72110.4290.55821.5280.7544.54
Miss.D0.7831.7461.0063.320.8774.3990.84116.9610.74311.2640.59122.6040.7694.858
MICE0.7440.2311.0090.9830.9871.7850.7911.0910.5313.2240.69810.5440.9551.684
GAIN0.7270.721.0162.2331.2945.0831.2624.4060.7997.7581.18853.391.4237.177
Sinkhorn0.6550.6960.7661.3170.7421.7820.5877.5770.5394.6010.4358.5630.6831.848
TDM0.7360.7660.811.4260.8082.2210.7149.3940.6926.3130.56210.2340.752.199
MIWAE0.7120.6310.8281.2440.8612.1660.6919.8330.5954.630.52112.0820.7121.615
MISS.F0.6490.5050.8321.4670.8372.0540.75811.1390.583.8860.4918.3650.7682.063
Remasker0.5030.5031.8984.3030.7322.4271.32819.882.76126.8294.26582.5970.5371.455
NewImp0.750.8070.8261.4910.8182.3170.7599.9040.756.6240.611.450.7652.264
KPI0.5980.4350.5570.8410.7321.5850.5176.8350.4653.0970.3947.8030.7081.533

[Q5] Regarding noise in datasets as in line 305, it has been studied that missing data imputation can have a denoising property, at least for time series, as in Missing data imputation for noisy time-series data and applications in healthcare.

Response. The listed reference is inspiring to investigate the MDI problem in the presence of noise. Thank you for highlighting this important work. We will add the related works in the revised manuscript.

  • Currently, despite that KPI does not explicitly design for noisy data, it would be helpful to investigate KPI’s robustness to noise. To this end, we add controlled experiments by introducing additive noise (±20% perturbation to 10% of observed entries) at a fixed missing rate of 0.1.
  • As shown in the table below, KPI consistently outperforms all baselines across diverse datasets under these noisy conditions. These results empirically substantiate that, although KPI is not explicitly designed for denoising, it remains applicable to noisy case. We will incorporate a discussion of these findings and the relevant literature in the revised manuscript to provide a comprehensive and academically grounded perspective.
Dataset:BTCCCBVISPKQBWQW
CSDI-T0.68960.90240.85010.75800.70010.55790.7834
GAIN0.78010.82630.79570.62420.52580.47360.7403
MICE0.87431.04031.05210.91350.74020.85411.0602
MIRACLE0.55870.65910.77353.71383.86310.64230.5978
MISS.F0.48860.73590.72670.58020.59190.43460.6718
MIWAE0.57820.73290.86040.62420.61070.53080.6985
Sinkhorn0.52600.49900.55490.50430.4782Nan0.6262
TDM0.63940.75650.76370.61910.60640.50830.7354
NewImp0.78020.84650.81450.74790.75580.60460.7649
KPI0.44460.36720.39820.41230.34010.23960.5112

[Q6] the related works should be improved

Response. We appreciate the reviewer’s suggestion to enhance the related works section. Currently we group the related MDI methods into two main groups: discriminative and generative approaches. Within the discriminative category, we further differentiate between iterative methods (which KPI aims to improve) and non-iterative methods (e.g., TDM). Building on this taxonomy, we promise to update the related works section in the revised manuscript to incorporate recent advances and ensure a more comprehensive and up-to-date coverage of the field.

Reference

[1] Asuncion, Arthur, and David Newman. "UCI machine learning repository." Nov. 2007.

评论

[25] Boris Muzellec, Julie Josse, Claire Boyer, and Marco Cuturi. Missing data imputation using optimal transport. In Proc. Int. Conf. Mach. Learn., volume 119, pages 7130–7140, 2020.

[26] Danilo Jimenez Rezende, Shakir Mohamed, and Daan Wierstra. Stochastic backpropagation and approxi- mate inference in deep generative models. In Proc. Int. Conf. Mach. Learn., volume 32, pages 1278–1286, 2014.

[27] Patrick Royston and Ian R White. Multiple imputation by chained equations (mice): implementation in stata. J. Statist. Softw., 45:1–20, 2011.

[28] Juntong Shi, Minkai Xu, Harper Hua, Hengrui Zhang, Stefano Ermon, and Jure Leskovec. Tabdiff: a mixed-type diffusion model for tabular data generation. In Proc. Int. Conf. Learn. Represent., 2024.

[29] Daniel J Stekhoven and Peter Bühlmann. Missforest—non-parametric missing value imputation for mixed-type data. Bioinformatics, 28(1):112–118, 2012.

[30] Ziyue Sun, Haozhe Li, Wenhai Wang, Jiaqi Liu, and Xinggao Liu. DTIN: dual transformer-based imputation nets for multivariate time series emitter missing data. Knowl. Based Syst., 284:111270, 2024. [31] Yusuke Tashiro, Jiaming Song, Yang Song, and Stefano Ermon. Csdi: Conditional score-based diffusion models for probabilistic time series imputation. Proc. Adv. Neural Inf. Process. Syst., 34:24804–24816, 2021.

[32] Daeho Um, Ji Won Yoon, Seong Jin Ahn, and Yunha Yeo. Gene-gene relationship modeling based on genetic evidence for single-cell rna-seq data imputation. In Proc. Adv. Neural Inf. Process. Syst., volume 37, pages 18882–18909, 2024.

[33] Hao Wang, Tai-Wei Chang, Tianqiao Liu, Jianmin Huang, Zhichao Chen, Chao Yu, Ruopeng Li, and Wei Chu. ESCM2: entire space counterfactual multi-task model for post-click conversion rate estimation. In Proc. Annu. Int. ACM SIGIR Conf. Res. Develop. Inf. Retrieval, pages 363–372, 2022.

[34] Jinsung Yoon, James Jordon, and Mihaela van der Schaar. GAIN: missing data imputation using generative adversarial nets. In Proc. Int. Conf. Mach. Learn., volume 80, pages 5675–5684, 2018.

[35] Aoqian Zhang, Shaoxu Song, Yu Sun, and Jianmin Wang. Learning individual models for imputation. In Proc. IEEE Int. Conf. Data Eng., pages 160–171. 2019.

[36] Hengrui Zhang, Liancheng Fang, Qitian Wu, and Philip S Yu. Diffputer: Empowering diffusion models for missing data imputation. In Proc. Int. Conf. Learn. Represent., 2025.

[37] Hengrui Zhang, Liancheng Fang, Qitian Wu, and Philip S Yu. Tabnat: A continuous-discrete joint generative framework for tabular data. In Proc. Int. Conf. Mach. Learn., 2025.

[38] Yimei Zhang, Xiangjie Kong, Wenfeng Zhou, Jin Liu, Yanjie Fu, and Guojiang Shen. A comprehensive survey on traffic missing data imputation. IEEE Trans. Intell. Transp. Syst., 2024.

[39] He Zhao, Ke Sun, Amir Dezfouli, and Edwin V. Bonilla. Transformed distribution matching for missing value imputation. In Proc. Int. Conf. Mach. Learn., volume 202, pages 42159–42186, 2023.

[40] Yuxuan Zhao, Eric Landgrebe, Eliot Shekhtman, and Madeleine Udell. Online missing value imputation and change point detection with the gaussian copula. In Proc. AAAI Conf. Artif. Intell., volume 36, pages 9199–9207, 2022.

[41] Yuxuan Zhao, Alex Townsend, and Madeleine Udell. Probabilistic missing value imputation for mixed categorical and ordered data. In Proc. Adv. Neural Inf. Process. Syst., volume 35, pages 22064–22077, 2022.

[42] Yuxuan Zhao and Madeleine Udell. Missing value imputation for mixed data via gaussian copula. In Proc. ACM SIGKDD Int. Conf. Knowl. Discovery Data Mining

[43] Lien P Le, Xuan-Hien Nguyen Tai, Thu Nguyen, et al. Missing data imputation for noisy time-series data and applications in healthcare[J]. arXiv preprint arXiv:2412.11164, 2024.

评论

References

[1] Hanyang Chen, Yang Jiang, Shengnan Guo, Xiaowei Mao, Youfang Lin, and Huaiyu Wan. Difflight: a partial rewards conditioned diffusion model for traffic signal control with missing data. In Proc. Adv. Neural Inf. Process. Syst., volume 37, pages 123353–123378, 2024.

[2] Zhichao Chen, Haoxuan Li, Fangyikang Wang, Haotian Zhang, Hu Xu, Xiaoyu Jiang, Zhihuan Song, and Hao Wang. Rethinking the diffusion models for missing data imputation: A gradient flow perspective. In Proc. Adv. Neural Inf. Process. Syst., 2024.

[3] Tianyu Du, Luca Melis Melis, and Ting Wang. Remasker: Imputing tabular data with masked autoencoding. In Proc. Int. Conf. Learn. Represent., 2024.

[4] Ye Du, Chen Yang, Nanxi Yu, Wanyu Lin, Qian Zhao, and Shujun Wang. Latent imputation before prediction: A new computational paradigm for de novo peptide sequencing. In Proc. Int. Conf. Mach. Learn., 2025.

[5] Xinxin Feng, Haitao Zhang, Can Wang, and Haifeng Zheng. Traffic data recovery from corrupted and incomplete observations via spatial-temporal trpca. IEEE Trans. Intell. Transp. Syst., 23(10):17835–17848, 2022.

[6] Liyang Hu, Yuheng Jia, Weijie Chen, Longhui Wen, and Zhirui Ye. A flexible and robust tensor completion approach for traffic data recovery with low-rankness. IEEE Trans. Intell. Transp. Syst., 25(3):2558–2572, 2023.

[7] Yue Hu and Daniel B Work. Robust tensor recovery with fiber outliers for traffic events. ACM Trans. Knowl. Discov. D., 15(1):1–27, 2020.

[8] Daniel Jarrett, Bogdan Cebere, Tennison Liu, Alicia Curth, and Mihaela van der Schaar. Hyperimpute: Generalized iterative imputation with automatic model selection. In Proc. Int. Conf. Mach. Learn., volume 162, pages 9916–9937, 2022.

[9] Jungkyu Kim, Kibok Lee, and Taeyoung Park. To predict or not to predict? proportionally masked autoencoders for tabular data imputation. In Proc. AAAI Conf. Artif. Intell., volume 39, pages 17886– 17894, 2025.

[10] Trent Kyono, Yao Zhang, Alexis Bellot, and Mihaela van der Schaar. MIRACLE: causally-aware imputation via learning missing data mechanisms. In Proc. Adv. Neural Inf. Process. Syst., pages 23806–23817, 2021.

[11] Marine Le Morvan, Julie Josse, Erwan Scornet, and Gaël Varoquaux. What’sa good imputation to predict with missing values? In Proc. Adv. Neural Inf. Process. Syst., volume 34, pages 11530–11540, 2021.

[12] Daniel Lee and H Sebastian Seung. Algorithms for non-negative matrix factorization. Proc. Adv. Neural Inf. Process. Syst., 13, 2000.

[13] Haoxuan Li, Yanghao Xiao, Chunyuan Zheng, Peng Wu, and Peng Cui. Propensity matters: Measuring and enhancing balancing for recommendation. In Proc. Int. Conf. Mach. Learn., volume 202, pages 20182–20194. PMLR, 2023.

[14] Haoxuan Li, Chunyuan Zheng, Wenjie Wang, Hao Wang, Fuli Feng, and Xiao-Hua Zhou. Debiased recommendation with noisy feedback. In Proc. ACM SIGKDD Int. Conf. Knowl. Discovery Data Mining, page 1576–1586, 2024.

[15] Haozhe Li, Yilin Liao, Zijian Tian, Zhaoran Liu, Jiaqi Liu, and Xinggao Liu. Bidirectional stackable recurrent generative adversarial imputation network for specific emitter missing data imputation. IEEE Trans. Inf. Forensics Security, 19:2967–2980, 2024.

[16] Haozhe Li, Yilin Liao, Zijian Tian, Zhaoran Liu, Jiaqi Liu, and Xinggao Liu. Bidirectional stackable recurrent generative adversarial imputation network for specific emitter missing data imputation. IEEE Trans. Inf. Forensics Security, 19:2967–2980, 2024.

[17] Jingchen Liu, Andrew Gelman, Jennifer Hill, Yu-Sung Su, and Jonathan Kropko. On the stationary distribution of iterative imputations. Biometrika, 101(1):155–173, 2014.

[18] Yang Liu and Anthony Constantinou. Improving the imputation of missing data with markov blanket discovery. In Proc. Int. Conf. Learn. Represent., 2022.

[19] Chao Ma and Cheng Zhang. Identifiable generative models for missing not at random data imputation. In Proc. Adv. Neural Inf. Process. Syst., volume 34, pages 27645–27658, 2021.

[20] R Malarvizhi and Antony Selvadoss Thanamani. K-nearest neighbor in missing data imputation. Int. J. Eng. Res. Dev, 5(1):5–7, 2012.

[21] Pierre-Alexandre Mattei and Jes Frellsen. Leveraging the exact likelihood of deep latent variable models. In Proc. Adv. Neural Inf. Process. Syst., pages 3859–3870, 2018.

[22] Pierre-Alexandre Mattei and Jes Frellsen. MIWAE: deep generative modelling and imputation of incomplete data sets. In Proc. Int. Conf. Mach. Learn., volume 97, pages 4413–4423, 2019.

[23] Rahul Mazumder, Trevor Hastie, and Robert Tibshirani. Spectral regularization algorithms for learning large incomplete matrices. J. Mach. Learn. Res., 11:2287–2322, 2010.

[24] Xiaoye Miao, Yangyang Wu, Lu Chen, Yunjun Gao, and Jianwei Yin. An experimental survey of missing data imputation algorithms. IEEE Trans. Knowl. Data Eng., 35(7):6630–6650, 2022.

评论

Dear reviewer is18,

Thank you for your valuable suggestion to improve the related work section. We have taken your feedback to action and have significantly expanded our literature review, with a special focus on the top venues (e.g., ICML, NeurIPS, ICLR, SIGKDD, and AAAI) over the last three years. The updated review now incorporates 42 references, the majority of which are recent publications, to provide a more comprehensive and current overview of the field. The revised literature review is presented below.

The pervasive presence of missing data undermines the integrity of collected datasets and the reliability of data-driven applications, underscoring the necessity for effective missing data imputation (MDI) in broad fields such as traffic [1, 38], signal processing [15], healthcare [43] and bioinformatics [4, 32]. To achieve accurate MDI, existing approaches can be broadly categorized into two paradigms: discriminative and generative, each with distinct advantages and limitations [8, 24].

The iterative method [35, 17, 11] is one of the most popular methods in discriminative imputation, initiated from imputation by chained equations (ICE) [27], which employs specific models to estimate missing values for each feature based on the remaining observable features. On the basis of ICE, a line of work advocates for employing modern parametric models, such as neural networks [22, 10], Bayesian models [27] and random forest [29], which enhances the capacity of imputation models and thereby accommodating complex missing patterns. In a different line of work, various training techniques are investigated within the paradigm, such as multiple imputation [27], ensemble learning [29], variable selection [18], and multitask learning [22], which enhances the utility to accommodate diverse contexts. While this paradigm offers enhanced flexibility and accuracy, it fails to utilize the oracle features effectively and risks model misspecification, which can lead to suboptimal imputation results. Our research advances this methodology by handling the two limitations.

Apart from the iterative methods, there are other notable approaches in the discriminative paradigm. The simple direct paradigm employs elementary statistical measures like mean, median, and mode to replace missing values, offering quick and straightforward solutions. However, this approach lacks the capacity to accommodate complex relationships [20, 23], often producing trivial and inadequate imputation results that fail to meet the expectation in practice. Another notable approach is matrix and tensor factorization, which decomposes the dataset into two low-rank matrices, capturing the latent structure of the data for imputation [12, 6]. This method has robustness to noise [5, 7] and is particularly effective in collaborative filtering and recommendation systems [33, 13, 14]. Gaussian corpula models are easy to implement and excels in imputing mixed data type [41, 42, 40]. Recent advancements have introduced a novel methodology through distribution discrepancy minimization [39, 25]. This method is based on the reasonable assumption that if datasets follow the independent and identically distributed (i.i.d.) principle, then any two batches should mirror the same distribution, leading to minimal discrepancy.

The generative paradigm restates imputation as a conditional generation problem, using advanced neural architectures and generative training strategies, such as autoencoders [9, 3], generative ad- versarial networks [34, 30, 16] and diffusions [31, 1, 2, 36], to approximate data distributions and perform imputation. This strategy incorporates the strengths of generative models, capturing and utilizing complex relationships, which potentially enhances the imputation quality when ample data is available. Recent studies further discover its potential to jointly impute continuous and discrete features [37, 28] and handling complex missing mechanisms [19]. However, it also bears the defects with generative models, such as the instability associated with adversarial training and the operational complexity of diffusions [21, 26], hampering their use in practice.


We sincerely appreciate your initial supports and meticulous comments. Your suggestions on refining literature review were especially helpful, which makes this work much more comprehensive as a result. If our response and additional experiments have properly resolved your raised issues, we politely invite you to further reconsider your overall rating.

If you have any further question and concern, we are happy to discuss.

Thanks for your time,

Submission28751 Authors

审稿意见
4

This paper addresses two critical defects in existing iterative imputation methods for handling missing data: model misspecification (i.e., applying a uniform model form to heterogeneous data) and underutilization of "oracle features" (fully observed features) It proposes a bi-level optimization framework called Kernel Point Imputation (KPI) to solve these problems.

优缺点分析

Strengths:

  1. Proposes a bi-level optimization framework that cleverly addresses the two core issues in iterative imputation: model misspecification and underutilization of "oracle features."

  2. By adaptively selecting function forms in RKHS, KPI can better capture complex heterogeneous dependencies among different features.

Weaknesses:

  1. The baselines compared in this paper are too outdated; at least some more competitive baselines published in 2024 should be added. Such as:

[1]Jolicoeur-Martineau, Alexia, Kilian Fatras, and Tal Kachman. "Generating and imputing tabular data via diffusion and flow-based gradient-boosted trees." In International Conference on Artificial Intelligence and Statistics, pp. 1288-1296. PMLR, 2024. [2]Wei, T. R., Wang, Y., Inoue, Y., Wu, H. T., & Fang, Y. (2024). Table Transformers for imputing textual attributes. Pattern Recognition Letters, 186, 258-264. [3]Liu, Yixin, Thalaiyasingam Ajanthan, Hisham Husain, and Vu Nguyen. "Self-supervision improves diffusion models for tabular data imputation." In Proceedings of the 33rd ACM International Conference on Information and Knowledge Management, pp. 1513-1522. 2024.

  1. Computational Cost: The bi-level optimization framework and the use of RKHS might increase computational complexity, especially when dealing with large-scale datasets, potentially making it less efficient than some simpler imputation methods.

  2. Despite the adaptive kernel ensemble, the choice and parameterization of kernel functions might still impact performance and require careful tuning.

  3. Definition and Availability of "Oracle" Features: The assumption of "oracle features" might be difficult to meet or clearly define in some real-world scenarios, which could limit the method's universality.

问题

  1. Does the paper provide more detailed guidelines for identifying and defining "oracle" features? In practical applications, how would an inappropriate selection of "oracle" features impact imputation performance?

  2. What are the detailed implementation specifics of the adaptive kernel ensemble strategy? How does it ensure dynamic selection of the optimal kernel during training while avoiding overfitting?

局限性

The applicability of the model to missing imputation in non-tabular data (e.g., time series data) is not yet clear.

最终评判理由

Thank you for the authors’ response. After carefully reviewing the reply, I have decided to retain my current score.

格式问题

None

作者回复

Thank you very much for your encouraging support and appreciation of our framework and addressed research problems. Below are our responses to the specific query raised.


[W1] The baselines compared in this paper are too outdated; at least some more competitive baselines published in 2024 should be added.

Response. We appreciate your insightful comment regarding the inclusion of more recent and competitive baselines. We agree that it is essential to ensure up-to-date comparison.

  • We would like to clarify that our current manuscript already incorporates NewImp (NeurIPS 2024) and Remasker (ICLR 2024), both of which represent state-of-the-art methods published in 2024. Their inclusion represents the recent advances in the field and could contemporarily alleviate the concern regarding baseline selection.

  • In response to the suggestion, we are actively working to implement them in our experimental framework. Given the technical complexity of diffusion models that we are not familiar with, we respectfully request additional time to implement them. We will report the results of these new baselines in the revised manuscript or, if the review permits this, in a subsequent discussion phase.

[W2] Computational Cost: The bi-level optimization framework and the use of RKHS might increase computational complexity, especially when dealing with large-scale datasets, potentially making it less efficient than some simpler imputation methods.

Yes, we acknowledge that computational cost is a key consideration when evaluating the proposed method. We address this concern with a rigorous and evidence-based discussion as follows:

  • Firstly, although KPI targets to solve a bi-level optimization problem, the final solution reduces to a standard quadratic loss function (see Eq. 8) that is directly computable. Similar to standard deep learning methods, KPI updates the loss function in a batch-wise manner without requiring bi-level optimization in practice, and the batch-wise update makes it scalable to large-scale datasets.
  • Secondly, we report below the computational cost of KPI under various settings. As complexity is primarily determined by the kernel matrix size—which is equal to the batch size—we focus on the running time of computing the learning objective (Eq. 8) per batch across different batch sizes. Results show that while larger batch sizes increase computational cost, the running time remains limited (<3.295<3.295 ms given batch size 1024). We also showcase the complexity is less relevant to the feature dimension, since it does not change the kernel matrix size.
Batch size32642565121024
Gaussian kernel2.1212.0612.1142.6233.295
Polynomial kernel1.3781.3011.3811.8593.041
Feature number163264128256
Gaussian kernel2.0642.0622.0642.0682.223
Polynomial kernel1.3071.3111.3291.3111.302

[W3] Despite the adaptive kernel ensemble, the choice and parameterization of kernel functions might still impact performance and require careful tuning.

Response. Thank you for highlighting the important issue of kernel selection and parameterization. To address this concern, we demonstrate that Gaussian kernel is a good default choice prior to selection. Our response contains theoretical justification and empirical evidence.

  • Theoretically, the Gaussian kernel is known for its universality, allowing it to approximate any continuous function in RKHS. This property is essential for our method, as it ensures sufficient expressiveness to represen ff^* with kernel matrices (Lemma 3.3). Thus, the universality property makes the Gaussian kernel a strong default choice.
  • Empirically, we compare the Gaussian kernel to other commonly used kernels, such as linear, polynomial, and Laplacian. As shown in the table below, the Gaussian kernel consistently provides competitive or superior performance.
  • In conclusion, Gaussian kernel serves as a theoretically and empirically advantageous choice, reducing the need for kernel selection. Its sole hyperparameter, the kernel width σ, can be efficiently addressed using a multi-kernel ensemble strategy that adaptively selects and weights kernels with different σ values.
KernelMSEWASSMAE
CC
Linear0.0990.0510.203
Poly0.0650.0390.091
Laplacian0.0680.0420.093
Gaussian0.0760.0350.082
CBV
Linear0.0890.0870.219
Poly0.0870.0880.214
Laplacian0.0830.0800.211
Gaussian0.0870.0660.205
BT
Linear0.3260.1010.359
Poly0.3050.0900.342
Laplacian0.3160.0910.346
Gaussian0.3020.0890.338

[W4] Definition and Availability of "Oracle" Features: The assumption of "oracle features" might be difficult to meet or clearly define in some real-world scenarios, which could limit the method's universality.

Response. Once again, thank you very much for your constructive comment. In our work, "oracle features" are defined as features that are fully observed, i.e., with no missing entries across all samples.

  • Oracle features are ambiguous in real-world datasets: For example, in electronic health records, demographic variables such as age, gender, and height are typically fully observed. Similarly, in industrial settings, sensor readings from robust or easily accessible equipment (e.g., ambient temperature, pressure at main control points) are often fully recorded, whereas measurements from sensors in harsh environments (e.g., temperature at the base of a high-pressure tower) are more prone to missingness due to equipment failure or maintenance constraints.

  • A few oracle features improve performance. We recognize that the number of fully observed features may be limited in some domains. However, our empirical results (Section 4.3) show that even a small number of oracle features can suffice for effective application of KPI.

  • Relaxed definition. In cases where no feature is entirely observed, the definition of "oracle features" may be relaxed to include those with very minimal missingness, after suitable imputation prior to KPI process. We will clarify this operational flexibility and its implications in the revised manuscript.

评论

Dear Reviewer 8EZS,

Thank you for your constructive feedback on expanding the baseline comparisons.** We have acted on your suggestion and added experiments on the suggested baselines.**

We would like to note that reference [2] is specifically designed for text-based features and is not directly applicable to the numerical data central to our work. Therefore, we have focused the new experiments on ForestVP [1] and SimpDM [3]. To provide a more thorough comparison against the state-of-the-art, we have also included two strong, concurrent generative methods: NewImp and Remasker.

MethodBTCCCBVISPKQBWQW
MAEWASSMAEWASSMAEWASSMAEWASSMAEWASSMAEWASSMAEWASS
Remasker0.4390.1310.7670.7500.5280.5220.5993.5840.4471.2680.4012.8110.5460.636
NewImp0.4650.1770.4120.2920.4050.4010.4312.4950.3200.8570.3322.9920.4970.692
SimpDM0.5040.1480.4680.2850.4300.4250.4722.6120.3520.9060.4062.8090.5640.732
ForestDiffusion0.4720.1380.4530.3240.4300.4360.4452.5010.3430.8890.3422.9170.5610.754
KPI (Ours)0.3970.1210.3470.2840.4020.3940.4002.3870.3190.7470.2642.1310.4910.685

The results of additional comparisons are presented above. While the new baselines, particularly NewImp and Remasker, demonstrate competitive performance, our proposed KPI approach consistently achieves the best overall results across the benchmarks. We will integrate these new results and all corresponding clarifications into the final revised manuscript. We hope this added evaluation would address your concerns - thank you!


[1]Jolicoeur-Martineau, Alexia, Kilian Fatras, and Tal Kachman. "Generating and imputing tabular data via diffusion and flow-based gradient-boosted trees." In International Conference on Artificial Intelligence and Statistics, pp. 1288-1296. PMLR, 2024.

[2]Wei, T. R., Wang, Y., Inoue, Y., Wu, H. T., & Fang, Y. (2024). Table Transformers for imputing textual attributes. Pattern Recognition Letters, 186, 258-264.

[3]Liu, Yixin, Thalaiyasingam Ajanthan, Hisham Husain, and Vu Nguyen. "Self-supervision improves diffusion models for tabular data imputation." In Proceedings of the 33rd ACM International Conference on Information and Knowledge Management, pp. 1513-1522. 2024.

审稿意见
5

The paper presents an iterative imputation strategy based on bi-level optimization, proposed as an alternative to the existing MDI method. The approach consists of three steps: (1) randomly select a feature, (2) find a function in a Reproducing Kernel Hilbert Space (RKHS) that accurately predicts this feature using the others, and (3) update the imputation of other features to minimize the prediction risk associated with the selected feature. The method leverages oracle features—those with no missing data—and is empirically validated on a large set of datasets with artificially introduced missing values.

优缺点分析

Strengths

  • The method is built upon a well-presented heuristic and offers an interesting alternative for practical applications.
  • The imputation literature is well-situated and contextualized.
  • The use of RKHS is clearly introduced, and the derivation of its solution is well explained.
  • A wide variety of datasets are used for empirical evaluation, which supports the general relevance of the method—although the study remains incomplete in some aspects (see weaknesses).

Weaknesses

  • Some of the theoretical results (in main) are incorrect or imprecise. In particular, key assumptions are missing or improperly stated—for instance, the Gaussian kernel is not convex in general, yet convexity is assumed without justification. There are overstatements in the theoretical justifications (e.g., convergence and rates are claimed without sufficient assumptions or guarantees). Moreover, the algorithm uses an approximate minimizer, not the true one, and the true minimization problem is non-convex.
  • The paper does not discuss the mechanism of missing data. The proposed method appears to be valid only under the MCAR assumption, which is the most favorable case. Experiments are conducted under this assumption, limiting the scope of comparison with methods that perform well under MAR or MNAR .
  • The method seems to be more of a completion strategy than a true imputation method. It is unclear how the method could be used to impute missing values on new data points without retraining the model, unlike other methods such as MICE or MissForest. The experimental section should include a scenario with a held-out test dataset to evaluate generalization.
  • Notation is sometimes imprecise or not rigorous. For example, the function ff^\star is applied to inputs that may contain Na values, which is not clearly defined or explained.

Typos (no response needed):

  • XimpX^{imp} instead of X(imp)X^{(\mathrm{imp})} (177)
  • Nan (Not a number )-> Na (not attribute)

问题

You could address the points raised under 'Weaknesses' to improve your rating.

Overall, the theoretical overstatements and the omission of the missing data mechanism are the main reasons for my current recommendation to bordeline reject the paper . However, these two issues are relatively easy to address, and given the novelty and practical interest of the method, my evaluation could be revised to "accept" if they are properly corrected.

局限性

yes

最终评判理由

After rebuttal period the two main concerns are well addressed by the authors: rate: 3->5

格式问题

na

作者回复

Thank you very much for your constructive comments and appreciation of our heuristic and empirical studies. We are grateful that you have highlighted the key issues and expressed openness to rating adjustment. Below are our responses to the specific concerns and queries.


[W1] Some of the theoretical results are incorrect or imprecise. In particular, key assumptions are missing or improperly stated—for instance, the Gaussian kernel is not convex in general. There are overstatements in the theoretical justifications (e.g., convergence and rates are claimed without sufficient assumptions or guarantees). The algorithm uses an approximate minimizer, not the true one, and the true minimization problem is non-convex.

Response. Thank you for your insightful comments on the theoretical analysis. We acknowledge the raised points and will revise our analysis accordingly.

  • On the convexity of kernel function. In the main paper, we assumed kernel convexity. For Gaussian kernel we demonstrated that it has a convex region given dimension D=1\mathrm{D}=1, but it is not convex in general cases with D>1\mathrm{D}>1. We acknowledge this imprecision and will now restrict our discussion to the linear kernel, which is convex.
  • On the overstatements like convergence assumptions. Given linear kernel, we immediately have the seperate convexity of the loss function to XtX_t (since it is a standard quadratic function). Under these conditions, we can demonstrate convergence for the minimization process by updating XtX_t.

Considering the learning objective in (8) with linear kernel, define g(Xt)=(YtXtw)2g(\mathbf{X}^t) = (\mathbf{Y}^t - \mathbf{X}^t \mathbf{w})^2 where w=Xs(XsXs+λI)1Ys\mathbf{w} = \mathbf{X}^{s\top}(\mathbf{X}^s\mathbf{X}^{s\top} + \lambda\mathbf{I})^{-1}\mathbf{Y}^s. Suppose Xkt\mathbf{X}^t_k is the kk-th gradient descent iteration to minimize g(Xt)g(\mathbf{X}^t), Xt\mathbf{X}^t_* is the minimizer of g(Xt)g(\mathbf{X}^t). We make assumtions as follows: <!- > - (1) gg is LL-smooth in Xt\mathbf{X}^t, i.e., for all X,Z\mathbf{X}, \mathbf{Z} we have g(X)g(Z)FLXZF\|\nabla g(\mathbf{X}) - \nabla g(\mathbf{Z})\|_F \leq L \|\mathbf{X} - \mathbf{Z}\|_F;->

  • (1) gg is LL-strongly convex in Xt\mathbf{X}^t, i.e., for all X,Z\mathbf{X}, \mathbf{Z} we have g(Z)g(X)+g(X),ZX+L2ZXF2g(\mathbf{Z}) \geq g(\mathbf{X}) + \langle \nabla g(\mathbf{X}), \mathbf{Z} - \mathbf{X} \rangle + \frac{L}{2}\|\mathbf{Z} - \mathbf{X}\|_F^2
  • (2) the update step size is small, i.e., η2L\eta \leq \frac{2}{L}.
  • (3) the initial value X0t\mathbf{X}^t_0 is proximal to Xt\mathbf{X}^t_*, i.e., X0tXt22<ϵ2\|\mathbf{X}^t_0-\mathbf{X}^t_*\|_2^2 < \epsilon^2.

Then, for any K0\mathrm{K} \geq 0, we haveg(XKt)g(Xt)12ηKϵ2.g(\mathbf{X}^t_\mathrm{K})-g(\mathbf{X}^t_*)\leq \frac{1}{2\eta\mathrm{K}}\epsilon^2.

Proof. We first consider the iterative update using gradient descent:Xk+1t=XktηXtg(Xkt).\mathbf{X}^t_{k+1} = \mathbf{X}^t_k - \eta \nabla_{\mathbf{X}^t} g(\mathbf{X}^t_k). For a LL-strongly convex function gg, we derive the core equation: g(Xk+1t)g(Xkt)+Xttg(Xkt)(Xk+1tXkt)+L2Xk+1tXkt2=g(Xkt)ηXttg(Xkt)2+Lη22Xttg(Xkt)2g(Xkt)η2Xttg(Xkt)2g(Xt)+Xttg(Xkt)(XktXt)η2Xttg(Xkt)2=g(Xt)+12ηXktXt212ηXktXt2+Xttg(Xkt)(XktXt)η2Xttg(Xkt)2=g(Xt)+12ηXktXt212ηXktXtηXttg(Xkt)2=g(Xt)+12η(XktXt2Xk+1tXt2),g(\mathbf{X}^t_{k+1})\leq g(\mathbf{X}^t_k)+\nabla_{\mathbf{X}^{t}}^t g(\mathbf{X}^t_k)(\mathbf{X}^t_{k+1}-\mathbf{X}^t_{k})+\frac{L}{2}\|\mathbf{X}^t_{k+1}-\mathbf{X}^t_{k}\|^2\\ =g(\mathbf{X}^t_k)-\eta\|\nabla_{\mathbf{X}^{t}}^tg(\mathbf{X}^t_k)\|^2+\frac{L\eta^2}{2}\|\nabla_{\mathbf{X}^{t}}^tg(\mathbf{X}^t_k)\|^2\\ \leq g(\mathbf{X}^t_k)-\frac{\eta}{2}\|\nabla_{\mathbf{X}^{t}}^tg(\mathbf{X}^t_k)\|^2 \\ \leq g(\mathbf{X}^t_*)+\nabla_{\mathbf{X}^{t}}^tg(\mathbf{X}^t_k)(\mathbf{X}^t_k-\mathbf{X}^t_*)-\frac{\eta}{2}\|\nabla_{\mathbf{X}^{t}}^tg(\mathbf{X}^t_k)\|^2 \\ = g(\mathbf{X}^t_*)+\frac{1}{2\eta}\|\mathbf{X}^t_k-\mathbf{X}^t_*\|^2 - \frac{1}{2\eta}\|\mathbf{X}^t_k-\mathbf{X}^t_*\|^2 +\nabla_{\mathbf{X}^{t}}^tg(\mathbf{X}^t_k)(\mathbf{X}^t_k-\mathbf{X}^t_*)-\frac{\eta}{2}\|\nabla_{\mathbf{X}^{t}}^tg(\mathbf{X}^t_k)\|^2\\ = g(\mathbf{X}^t_*)+\frac{1}{2\eta}\|\mathbf{X}^t_k-\mathbf{X}^t_*\|^2 - \frac{1}{2\eta}\|\mathbf{X}^t_k-\mathbf{X}^t_*-\eta\nabla_{\mathbf{X}^{t}}^tg(\mathbf{X}^t_k)\|^2\\ = g(\mathbf{X}^t_*)+\frac{1}{2\eta}(\|\mathbf{X}^t_k-\mathbf{X}^t_*\|^2 - \|\mathbf{X}^t_{k+1}-\mathbf{X}^t_*\|^2), where the first equation uses assumption 1; the second equation uses the update rules of the gradient descent; the fourth step holds as we set η1L\eta \leq \frac{1}{L} which makes η+Lη22η2- \eta + \frac{L \eta^2}{2} \leq -\frac{\eta}{2}.

Summing up the inequality above with k=0,1,...,Kk=0,1,...,\mathrm{K}, we have: k=1Kg(Xkt)Kg(Xt)12η(X0tXt)2XKtXt2)12ηX0tXt2. \sum_{k=1}^\mathrm{K} g(\mathbf{X}^t_{k}) - \mathrm{K}g(\mathbf{X}^t_*) \leq \frac{1}{2\eta}(\|\mathbf{X}^t_0-\mathbf{X}^t_*)\|^2-\|\mathbf{X}^t_\mathrm{K}-\mathbf{X}^t_*\|^2)\leq \frac{1}{2\eta}\|\mathbf{X}^t_0-\mathbf{X}^t_*\|^2. According to the 4-th step of the core equation, we have g(Xk+1t)g(Xkt)...g(X0t)g(\mathbf{X}^t_{k+1})\leq g(\mathbf{X}^t_{k})\leq...\leq g(\mathbf{X}^t_{0}). Therefore, Kg(XKt)k=1Kg(Xkt)\mathrm{K}g(\mathbf{X}^t_{\mathrm{K}})\leq\sum_{k=1}^\mathrm{K}g(\mathbf{X}^t_{k}) holds, which immediately follows by Kg(XKt)Kg(Xt)k=1Kg(Xkt)Kg(Xt)12ηX0tXt212ηϵ2. \mathrm{K}g(\mathbf{X}^t_\mathrm{K})-\mathrm{K}g(\mathbf{X}^t_*)\leq\sum_{k=1}^\mathrm{K} g(\mathbf{X}^t_{k}) - \mathrm{K}g(\mathbf{X}^t_*)\leq \frac{1}{2\eta}\|\mathbf{X}^t_0-\mathbf{X}^t_*\|^2\leq\frac{1}{2\eta}\epsilon^2. Therefore, we have: g(XKt)g(Xt)12ηKϵ2. g(\mathbf{X}^t_\mathrm{K})-g(\mathbf{X}^t_*)\leq \frac{1}{2\eta\mathrm{K}}\epsilon^2.

  • On the approximate minimizer. We agree that the convergence justification relied on stringent assumptions (e.g., Gaussian kernel and separate convexity), and in practice only targets the approximate minimizer, which largely weakens its significance. In our revised manuscript, we will remove discussions on algorithmic convergence. Instead, we will limit the discussion in the appendix to the separate convexity of the objective under the linear kernel, thereby ensuring precision and avoiding overstatement.

[W2] The paper does not discuss the mechanism of missing data, limiting the scope of comparison with methods that perform well under MAR or MNAR.

Response. Thank you very much for your meticulous and actionable suggestion. In submission, relevant content was included in a seperate supplementary appendix, which may not have been sufficiently visible. To address this, we have now explicitly referenced the relevant discussions and results:

  • The MAR and MNAR mechanisms are discussed in Appendix B.1.
  • The MAR and MNAR results are discussed in Appendix C.3 with missing ratio fixed as 0.1.

[W3] The method seems to be more of a completion strategy than a true imputation method. It is unclear how the method could be used to impute missing values on new data points without retraining the model, unlike other methods such as MICE or Miss.F. The experimental section should include a scenario with a held-out test dataset to evaluate generalization.

Response. Thank you for your insightful comment. Please see our concise response below.

  • Firstly, We agree that our method is best characterized as a completion strategy, similar to prior works [1,2]. To clarify this, we will revise the terminology throughout the manuscript to completion where appropriate. In particular, we will add the following statement to the preliminary note (line 58):

As a preliminary note, there are two primary paradigms for handling missing data: (1) completion, where the model is optimized to fill in missing entries within a given incomplete dataset and does not generalize to new missing indices; and (2) imputation, where the model is trained to impute missing values for previously unseen indices or samples. The completion setting is practical in contexts where the entire dataset is collected prior to analysis. Consistent with prior work such as Sinkhorn and TDM [1,2], this study adopts the completion paradigm.

  • Secondly, as more of a completion approach, the current method does not naturally generalize to imputing missing values in new data points without training.
  • Finally, we are pleased that it is feasible to add the requested experiment using a held-out test dataset to evaluate how well KPI generalizes to new samples. Specifically, for each new sample, KPI minimizes the learning objective (Eq. 8) via gradient descent, updating only the values on missing indices in the test sample while keeping those in the training data fixed. The results in the table below demonstrate that KPI retains effective in the held-out test setting.
DatasetCCIS
MAEWASSMAEWASS
MICE0.2580.0450.5440.737
MISS.F0.2220.0400.1940.085
KPI0.1070.0540.1070.069

Reference

[1] Missing data imputation using optimal transport. ICML.

[2] Transformed distribution matching for missing value imputation. ICML.

[W4] Notation is sometimes imprecise or not rigorous. For example, the function is applied to inputs that may contain Na values, which is not clearly defined or explained.

Response. Thank you for highlighting the notational clarity. In Eq. 4, we formulated f(X(miss)_,d,X(obs)_,d)f^* (\mathbf{X}^\mathrm{(miss)}\_{\cdot,-d},\mathbf{X}^\mathrm{(obs)}\_{\cdot,-d}). As you noted, it is ambiguous to apply ff ^* to inputs with NaN values. Nevertheless, in practice, as you could have noticed, the missing entries are initialized with real values prior to optimization, so ff only receives real-valued inputs during the KPI process. To avoid confusion, we will replace X(miss)\mathbf{X}^\mathrm{(miss)} with X(comp)\mathbf{X}^\mathrm{(comp)} —explicitly denoting completed values—in this and related expressions. Other notations will similarly be standarized to ensure rigor and consistency.

评论

To build on my earlier suggestion in light of your (satisfying) response: For W1, in the context of your paper, I believe the theoretical result is not the main contribution. Rather, the introduction of your heuristic and the use of RKHS are the real strengths. Therefore, avoiding any overstatement—even at the cost of a weaker theoretical result—seems the right approach to preserve the integrity of this clever work.

For W2, I think it would be useful to clarify that, in theory, non-MCAR settings are not handled by your approach. At the same time, you can argue that this potential limitation is mitigated by your empirical results.

This paper is a refreshing and inspiring contribution to the imputation literature. It combines originality with significance—so there’s no need to overstate your claims. It’s highly interesting even without any formal theoretical guarantees.

I’m truly sorry about the code link issue. ML conferences can be unforgiving on such details. Regardless of this outcome, I have no doubt that the paper will succeed in future resubmissions.

评论
MethodBTCCCBVISPKQBWQW
MAEWASSMAEWASSMAEWASSMAEWASSMAEWASSMAEWASSMAEWASS
MICE0.5891.3620.7330.6950.9451.6370.7712.7980.5992.0830.6974.5630.9170.962
MISS.F0.8250.8150.7390.5170.8760.6590.6731.3380.9031.6040.6852.8020.8140.745
CSDI_T1.1875.6321.0023.2750.9344.4120.81216.1431.35519.1780.89124.2310.9215.084
MissDiff1.1252.9820.9952.3170.9476.0630.79814.3121.32723.8450.86935.0420.9126.953
GAIN1.1751.3420.8910.7050.7930.5870.5630.7340.8161.5260.6451.8790.8370.805
MIRACLE0.8040.6530.4620.2360.8170.6453.94220.9154.62524.7910.6981.7810.6020.472
MIWAE0.8520.9210.8290.6040.8950.7320.5680.8970.8641.7590.7553.0210.7350.662
Sinkhorn1.2131.4531.0650.8431.0170.8910.9351.4871.3643.5211.0113.1851.0121.028
TDM1.1081.3960.9620.7290.9180.7680.7861.1231.3023.3780.8472.8140.9010.891
ReMasker0.5260.5240.6020.4350.4910.3760.6282.2390.6121.5970.6073.8130.5190.591
NewImp0.5270.3820.3570.2570.3890.3300.4651.5640.4461.2130.4513.5090.4680.556
KPI (ours)0.3940.1600.1530.1290.2880.2380.3641.3960.4371.4200.2953.1150.4230.476
MethodBTCCCBVISPKQBWQW
MAEWASSMAEWASSMAEWASSMAEWASSMAEWASSMAEWASSMAEWASS
MICE0.6960.4110.6600.3840.8320.7340.7081.8310.4990.9800.5892.3770.8070.724
MISS.F0.7310.5790.6970.4830.7480.6310.5871.9440.7261.5130.5002.2910.6670.567
CSDI-T0.8853.1050.8852.9230.8383.9220.75916.8331.01614.1730.68320.3300.7954.275
MissDiff0.9122.4730.8541.1930.8423.7040.75213.3910.99119.3230.66223.2920.7715.953
GAIN0.8460.5950.7820.5450.6980.5700.5291.5110.5711.3050.4741.9430.7300.684
MIRACLE0.6550.3190.3710.1630.8420.8023.72527.0934.19625.0520.5762.2490.5180.437
MIWAE0.6580.4240.7350.5080.8080.7310.5591.5350.6281.4000.5613.3180.6410.577
Sinkhorn0.9670.7520.9400.6980.9250.9110.8542.1211.0492.9060.8443.7550.8800.859
TDM0.8580.7320.8490.6200.8210.7560.7241.6400.9692.7130.6613.2020.7720.744
Remasker0.4810.2971.0280.8860.4870.2960.6601.7010.5861.0590.5162.1280.5190.434
NewImp0.6450.4610.5850.5930.5620.8370.4423.9450.4342.3280.4417.1610.6011.102
KPI (ours)0.4550.1240.1580.0930.3570.3520.3911.3440.4230.7690.2861.9670.4580.563

We sincerely appreciate your encouraging words that Overall, the theoretical overstatements and the omission of the missing data mechanism are the main reasons for my current recommendation to bordeline reject the paper. However, these two issues are relatively easy to address, and given the novelty and practical interest of the method, my evaluation could be revised to "accept" if they are properly corrected. In light of this, if our response and additional experiments have properly resolved your raised issues, we politely invite you to further reconsider your overall score.

I personally really enjoy the discussion with you. I am particularly grateful for your thoughtful approach—not only highlighting crucial points for improvement but also offering encouragement, a practice we also value in the review process. Your comments on toning down overstatements were especially helpful, which makes work much more rigorous as a result.

If you have any further question and concern, we are happy to discuss.

Thanks for your time,

Submission28751 Authors

评论

Based on my initial review, my two main concerns have been addressed by the authors. Therefore, I am updating my recommendation to 'accept.' I appreciate the authors' responses and the quality of their work.

评论

Dear Reviewer tbrx,

We noticed the two concerns highlighted in the review report. In this thread, we would like to offer supplementary responses and results to further clarify these points.

[Additional response to W1, convergence].

Response. We acknowledged the overstatement and would like to add experiments to demonstrate convergence.

  • Revising Claims of Convergence. We concur with the reviewer that our algorithm utilizes an approximate minimizer for a non-convex optimization problem when general kernels are used. To rectify our previous overstatement, we would remove the theoretical claims regarding algorithmic convergence. The discussion in the appendix will be confined to the separate convexity of the loss function under the linear kernel, which avoids this overstatement.
  • Adding Empirical Demonstration of Convergence. While a theoretical proof of convergence remains challenging, we provide empirical results that demonstrate the convergence of our algorithm in practice. Specifically, in the tables below, we report the training loss, the validation MAE, and Wasserstein discrepancy at different epochs on CC dataset. Three key observations are noted.
    • The training loss decreases as the number of epochs increases, which indicates that the SGD-based optimizer is effective at minimizing the loss function.
    • The decrease in the training loss corresponds to a reduction in the validation set error, confirming that minimizing the loss function is effective to improve imputation accuracy.
    • The training loss and validation metrics stabilize after a certain number of epochs, indicating that the optimization process has converged.

MCAR results with missing ratio 0.1

CC
LossMAEWASSMSE
00.5820.8450.5721.027
200.4610.1200.1520.218
400.4350.0980.0440.072
600.4220.0900.0380.064

MAR results with missing ratio 0.1

CC
LossMAEWASSMSE
00.5210.8870.6271.031
200.2820.0730.0450.062
400.2710.0680.0380.059
600.2690.0650.0360.058

MNAR results with missing ratio 0.1

CC
LossMAEWASSMSE
00.4970.8290.6271.031
200.4450.3250.2280.376
400.4110.1800.1150.195
600.4030.1580.0930.163

[Additional response to W2, MNAR and MAR results].

Response. We would like to address this concern by (1) discussing the MAR and MNAR mechanisms, and (2) adding experiment results.

  • MAR and MNAR merchanism. In this paper, to simulate MAR data, features are randomly divided into two categories: observable and missing. Missingness is determined by a logistic model influenced by observable features, parameterized by randomly selected weights and a bias adjusted to achieve the desired missing ratio. To simulate MNAR data, the non-missing features from the MAR simulation are performed with a secondary MCAR masking. This additional step creates dependencies between missing values and certain observable features, thus replicating the MNAR condition.
  • MAR and MNAR results. We investigated the performance of baselines and KPI given MAR and MNAR scenarios. The results can be found in the supplementary Appendix file, for a missing ratio of 0.1. Please kindly see Table 9 for MAR results and Tabel 10 for MNAR results.
  • Additional MAR and MNAR results. We understand that the reviewer may advocate for more comprehensive results in MAR and MNAR scenarios. Therefore, we have added experiments with an increased missing ratio of 0.3 for both MAR and MNAR conditions. The results are presented in the tables below. Similar to the MCAR results, KPI exhibits competitive performance across most datasets, showing its robustness to different missing data mechanisms. These results will be collectively presented in the supplementary Appendix.
评论

Thank you so much for updating your recommendation to 'accept' and recognizing our work quality. We highly appreciate your encouraging words and we are truly grateful for your dedicated time and constructive comments!

评论

Dear reviewer tbrx,

We really appreciate your recognition of our work and your kind words, for considering "the introduction of our heuristic and the use of RKHS are the real strengths", "this paper is a refreshing and inspiring contribution to the imputation literature", "it combines originality with significance", and "it’s highly interesting even without any formal theoretical guarantees" -- thank you for your engagement with our work.

In addition, thank you for your constructive suggestions for toning down claims and incorporating the additional MNAR/MAR experiments are excellent points, and we will be integrating them in our final version.

We are also truly sorry about the accidental problem with the initial link, and during rebuttal have taken immediate action to correct this. The full code is now accessible, especially includes a detailed collection of comments (for easy understanding) and a Dockerfile (for easy replication). Meanwhile, we are reminded of the conference guidelines that leave room for addressing the link issue in the rebuttal phase as follows:

5.Open Access to Data and Code: If you ran experiments, did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? Please see the NeurIPS code and data submission guidelines for more details. While we encourage release of code and data, we understand that this might not be possible, so no is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark).

Therefore, with the link corrected, we trust the evaluation and discussion can be grounded on our technical contributions (for which you have highly recognized and we are truly encouraged).

Thanks for your time,

Submission28751 Authors

审稿意见
3

This paper proposes a kernel point imputation (KPI) framework for missing data imputation, aiming to address the issues of model misspecification and insufficient utilisation of oracle features in existing iterative imputation methods. The framework is implemented through a bi-level optimization : the inner layer adaptively selects the optimal model form for each feature in the reproducing kernel Hilbert space (RKHS) to mitigate model misspecification; while the outer layer uses oracle features as supervisory signals to optimise the imputation values. Experiments demonstrate that KPI outperforms existing methods on multiple real-world datasets and effectively leverages oracle features to enhance imputation accuracy.

优缺点分析

Advantages

  1. Reformulates iterative imputation as a bi-level optimization problem, with the inner layer optimising the model form in the RKHS and the outer layer using oracle features to supervise the imputation. This provides a new paradigm for missing data imputation.

  2. The method is simple and easy to understand.

  3. Experimental results appear promising.

Weaknesses and Questions

  1. The KPI framework proposed in the paper does not account for potential noise in the dataset, yet noise is prevalent in industrial scenarios and other real-world applications, which may limit KPI's applicability in such contexts.

  2. The paper defines oracle features as fully observed features. How can we determine which features can serve as oracle features in practical applications?

  3. The KPI method relies on oracle features. Can the KPI method perform effectively when there are few oracle features? In the experimental results shown in Figure 3, the KPI method is only compared with the Miss.F method. However, based on the experimental results in Table 1, the Miss.F method is not the suboptimal method.

  4. In the experimental results in Table 1, only the average results under different missing rates are shown, making it impossible to assess the effectiveness of the KPI method under high missing rates. It is recommended to present the results under different missing rate conditions separately to better validate the effectiveness of the KPI method.

  5. In the provided anonymous code link, there is only one file, and its content is ‘okasddsa’.

Summary

This paper proposes a new paradigm by reformulating iterative imputation as a bi-level optimization problem. However, the KPI method heavily relies on oracle features, and the experiments conducted on different proportions of oracle features are not sufficiently comprehensive. Therefore, I recommend a borderline reject.

问题

See Weaknesses and Questions.

局限性

n/a

最终评判理由

Please see my original comments. I will stand my rating.

格式问题

n/a

作者回复

[W1] The KPI framework proposed in the paper does not account for potential noise in the dataset, yet noise is prevalent in industrial scenarios and other real-world applications, which may limit KPI's applicability in such contexts.

Response. We appreciate the reviewer’s insightful observation regarding the role of noise in real-world datasets and its implications to KPI. We address this concern with a rigorous and evidence-based discussion as follows:

  • Firstly, as acknowledged in our limitations, KPI assumes that observed data are accurate and free from significant noise. Notably, that this assumption is not unique to KPI but is a foundational premise shared by most MDI methods in the literature. We restate and emphasize this general assumption in the limitations section.
  • Secondly, the absence of explicit design for noisy data does not necessarily imply poor robustness to noise. To assess KPI under noisy conditions, we conducted additional experiments:
    • Experimental protocol. We fixed the missing data ratio at 0.1 and introduced additive noise by randomly perturbing 10% of the observed entries, with each perturbed value altered by up to ±20% of its original magnitude.
    • Experimental results. The results, summarized in the table below, demonstrate that KPI consistently achieves the best performance across noise-involved datasets. The results demonstrate that, although KPI is not specifically designed for noisy data, its superior performance is retained in the noisy scenario.
Dataset:BTCCCBVISPKQBWQW
CSDIT0.68960.90240.85010.75800.70010.55790.7834
GAIN0.78010.82630.79570.62420.52580.47360.7403
MICE0.87431.04031.05210.91350.74020.85411.0602
MIRACLE0.55870.65910.77353.71383.86310.64230.5978
MISS.F0.48860.73590.72670.58020.59190.43460.6718
MIWAE0.57820.73290.86040.62420.61070.53080.6985
Sinkhorn0.52600.49900.55490.50430.4782nan0.6262
TDM0.63940.75650.76370.61910.60640.50830.7354
NewImp0.78020.84650.81450.74790.75580.60460.7649
KPI0.44460.36720.39820.41230.34010.23960.5112

[W2] The paper defines oracle features as fully observed features. How can we determine which features can serve as oracle features in practical applications?

Response. Thank you for raising this important point regarding the practical identification of oracle features. In our framework, "oracle features" are operationally defined as those features that are fully observed—i.e., features for which no entries are missing across all samples in the dataset. This definition is directly verifiable in real-world data.

  • In practical applications, the selection of oracle features is typically straightforward: one examines the data matrix and identifies features (columns) with complete observations. For instance, in electronic health records, demographic variables such as age, gender, and height are routinely and reliably collected for all patients, and thus can be designated as oracle features. In industrial monitoring scenarios, sensor readings from robust or easily accessible equipment (e.g., ambient temperature, pressure at main control points) are often fully recorded, whereas measurements from sensors in harsh or inaccessible environments (e.g., temperature at the base of a high-pressure tower) are more prone to missingness due to equipment failure or maintenance constraints. In such cases, the consistently observed features naturally serve as oracle features.

  • We recognize that the number of fully observed features may be limited in some domains. However, our empirical results (Section 4.3) show that even a small number of oracle features can suffice for effective application of KPI. Even worse, in cases where no feature is entirely observed, the definition of "oracle features" may be relaxed to include those with very few missingness, which can be solved through suitable imputation prior to KPI process. We will clarify this operational flexibility and its implications in the revised manuscript.

[W3] The KPI method relies on oracle features. Can the KPI method perform effectively when there are few oracle features? In the experimental results shown in Figure 3, the KPI method is only compared with the Miss.F method. However, based on the experimental results in Table 1, the Miss.F method is not the suboptimal method.

Response. Thank you for your valuable and actionable comment. We agree that it is important to evaluate our method in the scenario where only a few oracle features are available and to include additional baselines.

Initially, we selected Miss.F for comparison because it is a strong representative within iterative imputation methods, despite not being the overall best performer in Table 1. To address your concern, we have added experiments using NewImp as a new baseline and investigated the case where only a small number of oracle features are accessible.

  • Experimental protocol. We simulate this setting using CC and IS datasets, varying the number of oracle features (1, 2, 3, 4), where oracles are specified in order from left to right. Missingness is then introduced to the other features, with a fixed missing ratio of 0.1.
  • Experimental results. The table below summarizes the results for each method under varying numbers of oracle features. As the number of oracle features increases, the performance of KPI improves noticeably. These findings demonstrate that KPI is able to leverage the information from oracle features effectively, leading to substantial performance gains as more fully observed features are available.
Number of oracle features1234
MAEWASSMAEWASSMAEWASSMAEWASS
CC dataset
MISS.F0.8200.2310.6480.1490.6460.1490.6440.147
NewImp0.6200.1300.5300.1150.5600.1200.5400.112
KPI0.3870.0830.3700.0790.3690.0740.2060.022
IS dataset
MISS.F0.4190.4590.3280.1910.3300.2330.3380.228
NewImp0.4150.4450.3400.1750.3350.2200.3250.200
KPI0.3630.4260.2680.1140.2680.1020.1550.061

[W4] In Table 1, only the average results under different missing rates are shown, making it impossible to assess the effectiveness of the KPI method under high missing rates. It is recommended to present the results under different missing rates separately to better validate the effectiveness of the KPI method.

Response. Thank you very much for your meticulous and actionable suggestion. In submission, relevant content was included in a seperate supplementary appendix, which may not have been sufficiently visible. To address this, we have now explicitly referenced the relevant discussions and results:

  • The results under different missing ratios are seperately reported in Appendix C.2 (Table 5-8 for missing ratio 0.1, 0.2, 0.3, 0.4).
  • We further provide the results given higher missing ratio (0.6) as follow.
DatasetsBTCCCBVISPSQBWQW
MAEWASSMAEWASSMAEWASSMAEWASSMAEWASSMAEWASSMAEWASS
CSDI-T0.7461.6470.9313.0450.8434.1110.77215.4190.72110.4290.55821.5280.7544.54
Miss.D0.7831.7461.0063.320.8774.3990.84116.9610.74311.2640.59122.6040.7694.858
MICE0.7440.2311.0090.9830.9871.7850.7911.0910.5313.2240.69810.5440.9551.684
GAIN0.7270.721.0162.2331.2945.0831.2624.4060.7997.7581.18853.391.4237.177
Sinkhorn0.6550.6960.7661.3170.7421.7820.5877.5770.5394.6010.4358.5630.6831.848
TDM0.7360.7660.811.4260.8082.2210.7149.3940.6926.3130.56210.2340.752.199
MIWAE0.7120.6310.8281.2440.8612.1660.6919.8330.5954.630.52112.0820.7121.615
MISS.F0.6490.5050.8321.4670.8372.0540.75811.1390.583.8860.4918.3650.7682.063
Remasker0.5030.5031.8984.3030.7322.4271.32819.882.76126.8294.26582.5970.5371.455
NewImp0.750.8070.8261.4910.8182.3170.7599.9040.756.6240.611.450.7652.264
KPI0.5980.4350.5570.8410.7321.5850.5176.8350.4653.0970.3947.8030.7081.533

[W5] In the provided anonymous code link, there is only one file, and its content is ‘okasddsa’.

Response. We sincerely apologize for the previous oversight regarding the anonymous code repository. The issue arose from a miscommunication within our team, which has now been thoroughly addressed. We have since updated the repository to include the complete, fully reproducible implementation of the KPI method, including all necessary scripts for data preprocessing, model training, and evaluation. Additionally, we have included a comprehensive README file that provides detailed instructions on the installation of dependencies and execution of experiments. Once again, we apologize for inconvenience this may have caused.

评论

Dear Reviewer dCSM,

Thank you once again for your constructive and comprehensive feedback! In this window, we would like to add results to further address the specific issue [W3] which we find helpful to strengthen this paper.


[W3] The KPI method relies on oracle features. Can the KPI method perform effectively when there are few oracle features? In the experimental results shown in Figure 3, the KPI method is only compared with the Miss.F method. However, based on the experimental results in Table 1, the Miss.F method is not the suboptimal method.

Response. We would like to further address this concern with additional experiments. The experiment results showcase the effectiveness of KPI given a small amount of oracle features. Moreover, KPI outperforms NewImp across varying settings, a strong generative method as the additional baseline for comparison.

  • Firstly, we demonstrate the efficacy of KPI given a few oracle features on different datasets: CC, IS, BT, and WQW. Overall, KPI outperforms Miss.F and the newly added NewImp in different settings. Meanwhile, KPI can effectively exploit oracle features for improving performance: as the number of oracle features increases, the performance of KPI improves.
Number of oracle features1234
MAEWASSMAEWASSMAEWASSMAEWASS
CC dataset
MISS.F0.8200.2310.6480.1490.6460.1490.6440.147
NewImp0.6200.1300.5300.1150.5600.1200.5400.112
KPI0.3870.0830.3700.0790.3690.0740.2060.022
IS dataset
MISS.F0.4190.4590.3280.1910.3300.2330.3380.228
NewImp0.4150.4450.3400.1750.3350.2200.3250.200
KPI0.3630.4260.2680.1140.2680.1020.1550.061
BT dataset
MAEWASSMAEWASSMAEWASSMAEWASS
MISS.F0.5350.1540.5130.1450.4950.2340.4620.125
NewImp0.5020.1480.5210.1490.4570.1220.4910.144
KPI (Ours)0.4650.1040.4100.0910.3590.0760.3290.041
WQW dataset
MISS.F0.8090.1600.7690.1500.7410.1370.7580.147
NewImp0.8050.1650.7560.1420.7320.1400.7670.155
KPI (Ours)0.8020.0550.5790.0520.4970.0460.4720.044
  • Secondly, we compare the efficacy of KPI given a few oracle features on different missing ratios: 0.1, 0.2, and 0.3. Similarly, on the CC dataset, KPI outperforms Miss.F and NewImp, and the performance improves as the number of oracle features increases. Therefore, KPI exhibits competitive performance given different missig ratios and oracle feature amounts, and can effectively utilize oracle feature for performance improvement.
Missing ratio: 0.11234
MAEWASSMAEWASSMAEWASSMAEWASS
MISS.F0.8200.2310.6480.1490.6460.1490.6440.147
NewImp0.6200.1300.5300.1150.5600.1200.5400.112
KPI (Ours)0.3870.0830.3700.0790.3690.0740.2060.022
Missing ratio: 0.2
MISS.F0.6520.1530.6660.1740.6410.1520.6390.156
NewImp0.6480.1490.5920.1620.6340.1550.5870.148
KPI (Ours)0.5780.1650.3290.0610.2720.0360.1860.028
Missing ratio: 0.3
MISS.F0.7200.2100.7040.2010.6910.1860.6850.180
NewImp0.6770.1980.6720.1950.6750.1990.6590.188
KPI (Ours)0.6130.1380.3560.0690.3300.0640.2150.037

We hope that the additional experiments in our rebuttal and discussion phases could resolve the raised issues. Your comment on the oracle features was especially insightful, and the resulting discussion was instrumental in improving our work. We are very grateful for that valuable guidance.

Should you find our responses and revisions satisfactory, we would be very appreciative if you would generously adjust your score. If you have any further question and concern, we are happy to discuss.

Many thanks,

Submission28751 Authors

评论

I thank the authors for their detailed and thoughtful rebuttal. The additional experiments and clarifications are appreciated and help address several of my initial concerns. However, I maintain my original assessment and rating for the following reasons:

While the authors conducted experiments varying the number of oracle features and demonstrated performance improvements, I still find it difficult to fully distinguish the roles of oracle vs. non-oracle features in the model.

The method fundamentally relies on the presence of fully observed oracle features.

Although the authors have now corrected the code link, the initial failure to provide a working repository raises concerns about reproducibility and the review process timeline. Namely, I have no time to check the correctness of the implementation. I acknowledge the fix but consider this a non-trivial lapse, especially when reproducibility is increasingly emphasized in ML conferences.

评论

Dear reviewer dCSM,

Since the discussion period will end in a few hours, we will be online waiting for your feedback on our rebuttal, which we believe has fully addressed your concerns.

We would highly appreciate it if you could take into account our response when updating the final rating and having discussions with AC and other reviewers.

Thank you so much for your time and efforts. Sorry for our repetitive messages, but we're eager to ensure everything is addressed.

28751 Authors

评论

Thank you for your detailed feedback and we are delighted that we have cleared your previous concerns!


The method fundamentally relies on the presence of fully observed oracle features.

Thank you for the comment. We would like to clarify that KPI does not necessarily rely on the presence of oracle features.

  • Firstly, we agree that KPI excels at exploiting oracle features for improving imputation performance. It is because KPI incorporates oracle features as supervision signal (when they are available), which provides high-quality signals and therefore beneficial to imputation performance.

  • However, we emphasize that exploitation \neq reliance: Indeed, KPI does not necessarily rely on the presence of oracle features. In implementation, we iteratively treat each feature as the target feature Y\mathbf{Y} to provide supervision, and the remaining as the input features (see line 180). This process is repeated for all features, and does not rely on oracle feature. When there exists oracle features, KPI naturally involves them as part of the supervision signal. When there exists no oracle features, KPI still operates effectively.

  • Experiments further demonstrate KPI does not rely on oracle features. In Section 4.2: overall performance, no oracle feature exists in the generated datasets, but KPI still achieves the best performance. Indeed, the performance given oracle feature presence is separately investigated in Section 4.3, where we deliberately controlled the simulation to ensure the existence of oracle features.

While the authors conducted experiments varying the number of oracle features and demonstrated performance improvements, I still find it difficult to fully distinguish the roles of oracle vs. non-oracle features in the model.

Thank you for raising this concern. Please kindly note that KPI does not require distinguishing oracle vs. non-oracle features in the model.

  • In loss function calculation (see line 180), each feature (oracle or non-oracle) is treated as the target feature Y\mathbf{Y} in a round-robin manner. Even if there exists no oracle features, KPI still works.

  • Despite the above, KPI has a special advantage in exploiting oracle features in the cases where oracle features are available. The rationale is: oracle feature, as a fruitful information source, explicitly acts as supervision signal (target) to update other feature imputations. This claim is empirically validated in Section 4.3: oracle feature presence and the added experiments in rebuttal, where increasing the number of oracle features effectively boosts the performance of KPI.

The initial failure to provide a working repository raises concerns about reproducibility and the review process timeline. Namely, I have no time to check the correctness of the implementation. I acknowledge the fix but consider this a non-trivial lapse, especially when reproducibility is increasingly emphasized in ML conferences.

We are also truly sorry about the accidental problem with the initial link, and during rebuttal have taken immediate action to correct this. The full code is now accessible. To facilitate your checking, we (i) newly added detailed comments in the code, where we explicitly label the corresponding equations in the paper, and (ii) newly included a Dockerfile to automate environment setup and facilitate reproduction.

In light of the increasing importance of reproducibility in the machine learning community, we commit to releasing the full code for our method and baselines to benefit future research. We hope these actions may address your remaining concerns.

Please also kindly note that as par in NeurIPS 2025 reviewing checklist, it says:

5.Open Access to Data and Code: If you ran experiments, did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? Please see the NeurIPS code and data submission guidelines for more details. While we encourage release of code and data, we understand that this might not be possible, so no is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark).

Therefore, with the link corrected, we believe further evaluation and discussion can be grounded on our technical contributions.


Please let us know if we have resolved your concerns – thank you!

Sincerely,

Submission28751 Authors

评论

Dear Reviewer dCSM,

As the discussion deadline approaches, we are wondering whether our responses have properly addressed your concerns? Your feedback would be extremely helpful to us. If you have further comments or questions, we hope for the opportunity to respond to them.

Many thanks,

28751 Authors

评论

Reviewers, please engage in the discussion with the authors. The discussion period will end in a few days.

最终决定

This paper presents a new method that uses an RKHS and oracle features (features with no or few missing attributes) to perform imputation. The experiments are thorough and demonstrate the method's superiority in imputation tasks compared to other methods. The method is also well-principled and has some supporting theory.

Three of the four reviewers recommend or lean towards accept. The fourth reviewer (dCSM) leans toward reject. This reviewer seems to have two lingering concerns. The first is a concern is that the use of oracle features is potentially limiting. Initially, several other reviewers shared this concern. However, the authors subsequently demonstrated robustness of their method to this aspect. Their response ultimately satisfied the other reviewers as well as me.

The second concern was a lack of initial availability of code. However, NeurIPS policy makes this clear that a paper cannot be rejected solely for this, except where the code distribution is central to the paper, which it is not in this case.

Given this, I recommend acceptance of the paper despite the fourth reviewer's concerns.