PaperHub
5.3
/10
Poster4 位审稿人
最低3最高6标准差1.3
3
6
6
6
3.8
置信度
正确性2.8
贡献度2.8
表达2.8
ICLR 2025

Going Beyond Feature Similarity: Effective Dataset distillation based on Class-aware Conditional Mutual Information

OpenReviewPDF
提交: 2024-09-23更新: 2025-03-07
TL;DR

introducing conditional mutual information to enhance the performance and the efficiency of dataset distillation

摘要

关键词
dataset distillationconditional mutual information

评审与讨论

审稿意见
3

This paper proposes a plug-and-play method, termed CMI, designed to enhance existing DD techniques by minimizing conditional mutual information. By applying CMI, the distilled data is concentrated more effectively around the center of each class, thereby improving generalizability.

优点

  • The logic is clear.
  • The experiments are comprehensive.
  • The review of related works is thorough.
  • The proposed method is theoretically sound.

缺点

  • The core ideas, methodology, and formulations in this paper draw substantially from the approach proposed in [1].
  • If \hat{Z} contains excessive confusing or uninformative information related to S , then H(S | \hat{Z}, Y) will not be reduced; rather, it could remain the same or even increase. This is because conditional entropy reflects the remaining uncertainty in S after observing both \hat{Z} and Y . When \hat{Z} is noisy or irrelevant for predicting S , it does not help in reducing this uncertainty.
  • Line 213 states that “minimizing the class-aware CMI reduces the uncertainty brought to \hat{Z} conditioned on S ,” which should be “minimizing the class-aware CMI reduces the uncertainty brought to S conditioned on \hat{Z}”.
  • The authors’ derivation of Equation 6 lacks an explicit explanation, making it challenging to fully understand the transition from previous formulations.
  • Works like [2] and [3], which also target improvements in dataset distillation, are not adequately considered.
  • Equation 3 requires summing over all synthetic instances within class y , how the authors adapt this approach to instance-based distillation methods like SRe2L.

[1] Bayes conditional distribution estimation for knowledge distillation based on conditional mutual information

[2]TOWARDS LOSSLESS DATASET DISTILLATION VIA DIFFICULTY-ALIGNED TRAJECTORY MATCHING

[3]Prioritize Alignment in Dataset Distillation

问题

see weakness.

评论

Weakness 5: Comparison with recent SOTA.

Thanks for your suggestion. In contrast to [6] and [7], which are exclusively designed for trajectory matching, (i.e., selecting specific range of trajectories), both our proposed method and the majority baseline methods function as a versatile plugin (e.g., DREAM [8] serves as a universal solution applicable to various distillation methods). Thus, we consider applying the CMI constraint to [6] as suggested by R#kma8. As demonstrated in table below, our proposed approach consistently delivers substantial performance enhancements.

MethodCIFAR10CIFAR100TinyImageNet
105010501050
MTT65.3 ± 0.471.6 ± 0.239.7 ± 0.447.7 ± 0.223.2 ± 0.228.0 ± 0.3
MTT+CMI66.7 ± 0.372.4 ± 0.341.9 ± 0.448.8 ± 0.224.1 ± 0.328.8 ± 0.3
DATM66.8 ± 0.276.1 ± 0.347.2 ± 0.454.1 ± 0.231.1 ± 0.339.7 ± 0.3
DATM+CMI67.4 ± 0.476.6 ± 0.147.6 ± 0.355.1 ± 0.331.9 ± 0.440.6 ± 0.2

Weakness 6: Details about the implementation of SRe2^2L.

Thank you for pointing out the confused expression. Eq. 3 in our paper represents an abstraction of optimization-based dataset distillation methods, wherein current methods indirectly optimize Eq. 2 by aligning specific information (e.g., feature distribution). As discussed in [9] and [10]. instance-based methods (i.e., SRe2^2L) still falls under this category.

SRe2^2L optimizes the synthetic dataset by minimizing the loss function L=l(ϕθT(s),y)L = l(\phi_{\theta_{\mathcal{T}}}(\mathbf{s}), y), where y[C]y \in [C] and ll represents the cross entropy loss. Due to the instance-independent characteristic of its optimization process, we transit the optimization from the inner loop to the outer loop, enabling simultaneous optimization of the entire synthetic dataset. During the optimization process, synthetic dataset is classified based on predefined labels which are used to compute the loss function, enabling the deployment of CMI constraint.

The change of pseudocode is shown below. For a fair comparison, we present the impact of altering the loop order in table below. We attribute the performance improvement to the more precise class boundaries achieved by applying CMI constraint for datasets generated by SRe2^2L, a critical aspect that has also been emphasized in recent studies [10] [11] [12] [13], and our method can be more easily applied on recent SOTA [12] and [13] with their class-wise supervision.

Method1050100
SRe2^2L21.3 ± 0.646.8 ± 0.252.8 ± 0.3
SRe2^2L^\dagger21.5 ± 0.346.3 ± 0.453.4 ± 0.1
SRe2^2L + CMI24.2 ± 0.349.1 ± 0.154.6 ± 0.2

Original Pseudocode

for i from 1 to IPC:  # Outer loop optimizes synthetic dataset from 1 to IPC
    targets = np.arange(1000) # initial hard label
    for j from 1 to T:  # Inner loop optimizes images from 1 to epoch T
        Optimize S[i:, ] with L

Modified Pseudocode

targets = np.tile(np.arange(1000), (ipc, 1) )# initial hard label
for i from 1 to T:  # Outer loop optimizes synthetic dataset from 1 to epoch T
    for j from 1 to IPC:  # Inner loop optimizes images from 1 to IPC
        Optimize S[j, :] with L
    for k from 1 to C:  # apply CMI constraint for each class
        Optimize S[:, k] with CMI constraint

[6] Towards Lossless Dataset Distillation via Difficulty-Aligned Trajectory Matching. In ICLR, 2024.

[7] Prioritize Alignment in Dataset Distillation. In arXiv, 2024.

[8] Dream: Efficient Dataset Distillation by Representative Matching. In ICCV, 2023.

[9] Generalized Large-Scale Data Condensation via Various Backbone and Statistical Matching. In CVPR, 2024.

[10] Elucidating the Design Space of Dataset Condensation. In NeurIPS, 2024.

[11] Breaking Class Barriers: Efficient Dataset Distillation via Inter-Class Feature Compensator. Openreview Submission, 2024.

[12] Dataset Distillation via the Wasserstein Metric. In arXiv, 2023.

[13] Are Large-scale Soft Labels Necessary for Large-scale Dataset Distillation? In NeurIPS, 2024.

评论

We would like to express our gratitude for your valuable suggestions and insightful feedback of our paper. Our point-by-point responses are provided below.

Weakness 1: Contribution of our work.

Thank you for acknowledging the previous work [1] that utilizes conditional mutual information (CMI). Here, we present the distinctions between our proposed method and [1] as follows:

Different Optimization Objectives. [1] lies in the field of knowledge distillation, with the goal of enabling the teacher model to be a better estimated Bayesian classifier. While our proposed method is applied to evaluate the non-linear information between S\mathcal{S} and Z^\hat{Z} and to make the synthetic dataset a more learnable dataset.

Different Optimization Intentions. [1] maximize the CMI value to help the teacher model better capture the contextual information during pre-training phase. In contrast, our proposed method seeks to minimize the CMI value of the synthetic dataset to constrain its complexity during distillation process.

Different Optimization Methods. [1] optimizes by calculating CMI in the probability space, whereas our approach reconstructs the Markov chain (i.e., from YXY^Y \rightarrow X \rightarrow \hat{Y} to YSZ^Y \rightarrow S \rightarrow \hat{Z}) in feature space to compute CMI and validate the theoretical correctness as discussed in Sec. 4.1. A stronger constraint effect is achieved as shown in Table 7 provided in Sec 5.5.

Beyond that, we emphasize that the derivation process of CMI stems from a deterministic calculation in information theory, that is to say specific calculation is deterministic since the variable is determined. This is why our formulation is similar to [1] (Despite significant differences discussed above). Thus, the focus when applying CMI as a regularization should be why and how to determine the three corresponding random variables.

On the other hand, similar to the way of our work does, a series of studies employing methodologies utilizing information theory to enhance performance across various tasks have been widely applied, including in dataset distillation [2] [3] [4]. In contrast to the various estimation framework necessitated by the nondeterministic calculation of other information metric (e.g., mutual information), we reclaim that the major advantage of CMI is its certainty, i.e., the only estimation derived from the distribution sampling.

In summary, our study is the first to identify the intrinsic class-aware complexity challenge in dataset distillation and to introduce the CMI as a solution, elucidating its theoretical and practical significance. Furthermore, We offer a novel perspective on how information theory can be effectively applied to this field, thereby highlighting the originality and advancement of our research.

[1] Bayes conditional distribution estimation for knowledge distillation based on conditional mutual information. In ICLR, 2024.

[2] MIM4DD: Mutual Information Maximization for Dataset Distillation. In NeurIPS, 2023.

[3] One Training Fits All: Generalized Data Condensation via Mixture-of-Information Bottleneck Guidance. OpenReview Submission, 2024.

[4] GIFT: Unlocking Full Potential of Labels in Distilled Dataset at Near-zero Cost. Openreview Submission, 2024.

评论

Weakness 2: Confusion about the computation of conditional entropy H(SZ^,Y)H(S|\hat{Z}, Y)

We truly appreciate your detailed concerns about CMI for enhancing the rigor of the paper. The objective of minimizing the CMI is to mainly reduce the difference between conditional entropy H(SY)H(S \mid Y) and H(SZ^,Y)H(S \mid \hat{Z}, Y), rather than simply minimizing H(SZ^,Y)H(S \mid \hat{Z}, Y). Moreover, introducing additional conditions (i.e., the random variable Z^\hat{Z}) will either keep the entropy constant or decrease it, and it cannot lead to an increase as stated. This is a fundamental fact of information theory.

In fact, Z^\hat{Z} is a function of the random variable SS, where SS serves as the input of the deterministic fθ()f_{\theta^*}(\cdot) and Z^\hat{Z} is the output. At this point, if the goal is to minimize H(SZ^,Y)H(S \mid \hat{Z}, Y), fθf_{\theta^*} should not change any input, leading to H(SZ^,Y)=H(SS,Y)=0H(S \mid \hat{Z}, Y) = H(S \mid S, Y) = 0.

On the other hand, our research lies in the field of dataset distillation, where the optimization objective does not lead to the synthetic dataset becoming entirely noise, which has also been discussed in recent work [5]. To certify this assumption, we used different pre-trained classifier to evaluate the synthetic datasets distilled by different distillation methods as shown in Table below. For a more comprehensive understanding, we also present the performance of CMI constraint using different proxy models in Table.

MethodConvNetAlexNetResNet18VGG11
1050105010501050
MTT98.0100.097.099.098.0100.0100.0100.0
MTT + CMI100.0100.0100.0100.0100.0100.0100.0100.0
DSA72.082.679.083.871.072.275.078.4
DSA + CMI100.0100.099.0100.0100.0100.0100.0100.0
DM47.087.659.085.856.079.660.078.8
DM + CMI99.0100.098.0100.0100.0100.0100.0100.0

Weakness 3: Ambiguous Expression

Thank you for pointing out the issue of unclear expression in this section. Here, the point we aim to convey is ''Minimizing the class-aware CMI value constraints the uncertainty brought to Z^\hat{Z} with S\mathcal{S} as the input of fθ()f_{\theta^{*}}(\cdot)'', instead of H(Z^S)H(\hat{Z} \mid \mathcal{S}). The statement will be revised in the final version of the manuscript to ensure greater precision, and we sincerely appreciate your feedback in pointing this out.

Weakness 4: Lack of explicit explanation of CMI computation

Thanks for your suggestion, here, we provide a more detailed mathematical derivation. We first define the input synthetic data S\mathcal{S} is conditionally distributed according to PSY(y)P_{S \mid Y}(\cdot \mid y) and mapped into PSP([M])P_{S} \in \mathcal{P}([M]). Then we can formulate PZ^YP_{\hat{Z} \mid Y} as the average of PSP_{S} with respect to PSY(y)P_{S \mid Y}(\cdot \mid y), which is shown in Eq. 5:

PZ^y=E[PSY=y].P_{\hat{Z} \mid y } = \mathbb{E}[ P_S \mid Y =y ].

We get the conditional mutual information I(S;Z^Y=y)I(S ; \hat{Z} \mid Y=y) between SS and Z^\hat{Z} given YY:

I(S;Z^Y)=y[C]PY(y)I(S;Z^y)=E[D(PSPZ^Y)].I(S ; \hat{Z} \mid Y) = \sum_{y \in[C]} P_{Y}(y) I(S ; \hat{Z} \mid y) = \mathbb{E}\left[D\left(P_S \| P_{\hat{Z} \mid Y}\right) \right].

Given Y=yY = y, the Kullback-Leibler (KL) divergence D()D(\cdot \| \cdot) between PSP_{S} and PZ^yP_{\hat{Z} \mid y} is equal to Eq. 6:

E[D(PSPZ^y)Y=y]=E[(i=1MPS(i)lnPS(i)PZ^y(Z^=iY=y))Y=y]\mathbb{E}\left[D\left(P_{S} \| P_{\hat{Z} \mid y}\right) \mid Y=y\right] = \mathbb{E}\left[\left(\sum_{i=1}^{M} P_{S}(i) \ln \frac{P_{S}(i)}{P_{\hat{Z} \mid y}(\hat{Z}=i \mid Y=y)}\right)\mid Y=y\right] =SPSY(sy)[i=1MP(Z^=is)lnP(Z^=is)PZ^y(Z^=iY=y)].= \sum_{S} P_{S \mid Y}(\mathbf{s} \mid y)\left[\sum_{i=1}^{M} P(\hat{Z}=i \mid \mathbf{s}) \ln \frac{P(\hat{Z}=i \mid \mathbf{s})}{P_{\hat{Z} \mid y}(\hat{Z}=i \mid Y=y)}\right].

These details will be incorporated into the final version of the manuscript to enhance its rigor, and we sincerely appreciate your constructive feedback.

[5] What is Dataset Distillation Learning? In ICLR, 2024.

评论

Dear Reviewer Zh5i

Thank you once again for dedicating your valuable time to reviewing our paper and providing constructive comments! Considering the limited time available and to save the reviewer's time, we summarize our responses here.

  • 1. Contribution of our work:

    • We present a clear distinction between our method and existing approaches utilizing CMI.
    • We highlight that one major advantage of CMI stems from its computational determinism.
    • we elaborate on the originality of our method in applying information-theoretic principles and outlined its contributions.
  • 2. Confusion about the computation of conditional entropy H(SZ^,Y)H(S \mid \hat{Z}, Y):

    • We point out that incorporating additional observations will only decrease or maintain entropy.
    • We argue that Z^\hat{Z} would not be irrelevant information in the field of dataset distillation.
  • 3. Ambiguous Expression:

    • We clarify the ambiguous statements present in the paper.
  • 4. Lack of explicit explanation of CMI computation:

    • We re-derive the mathematical computation process from the definition of CMI to its computable expansion.
  • 5. Comparison with recent SOTA:

    • We provide additional experimental results by applying CMI on DATM across various datasets.
  • 6. Details about the implementation of SRe2^2L:

    • We demonstrate how to losslessly modify SRe2^2L code to implement CMI constraints.

As the end of the discussion period approaches, we kindly ask if our responses have satisfactorily addressed your concerns. Your feedback would be greatly appreciated, and we would be delighted to engage in further discussions if needed.

Sincerely,

The Authors

评论

Thank you for the explaination.

The difference you highlighted pertains to the distinction between KD and DD, rather than the core differentiator between the proposed methods. Both the methods utilize class-aware CMI for optimization, and the only modification in Eq. 7 is the substitution of XSX \to S and Y^Z^\hat{Y} \to \hat{Z}.

Furthermore, based on the provided pseudocode, the proposed CMI introduces an additional class-wise optimization step for the generated images, which incurs extra computational cost. Considering the very marginal performance gain, this additional complexity significantly diminishes the overall contribution of the proposed method, especially when compared to other advancements on SRe2L.

So I will maintain my score.

评论

We thank the reviewer for the valuable feedback and provide further clarifications below:

1. Utilization of CMI:

Conditional Mutual Information (CMI), derived from information theory, serves as a information measure of the relationship between two random variables and is not domain-specific. Unlike Mutual Information (MI), one significant advantage of CMI lies in its computational determinism. We believe that appropriately leveraging CMI according to task-specific characteristics across different fields has significantly facilitated the mutual verification of theory and experimentation, which has also been acknowledged and accepted by R#oMQb and R#kma8.

2. Confusion between KD and DD:

Here, we reemphasize that the differences between knowledge distillation (KD) and dataset distillation (DD) can lead to substantial impacts. In contrast to previous work [1], our method introduces clear differences (e.g., optimization direction). We notice that in the feedback, you seem to consider that two random variables Y^\hat{Y} and Z^\hat{Z} can be straightforwardly substituted when reconstructing Markov chains, while we have made a detailed discussion in Sec 4.1. And we have made a clear clarification that the focus when applying CMI as a regularization should be why and how to determine the three corresponding random variables and appropriate optimization intention.

3. Application to SRe2^2L:

Our work aims to constrain the excessive complexity in existing dataset distillation tasks by utilizing CMI as a general plug-and-play regularization which has been acknowledged by all the reviewers including R#Zh5i. Simply conflating our method with methods solely aimed at improving accuracy on SRe2^2L could be inappropriate. Based on your suggestion, we conduct comparisons between our method and PAD [2] recommended by the reviewer on SRe2^2L. Results on CIFAR-100 are presented in the table below.

Method10100
SRe2^2L28.257.2
SRe2^2L + PAD (FIEX)29.357.9
SRe2^2L + CMI30.159.1

More effective modification to SRe2^2L Code:

We appreciate your valuable advice and have made improvements to our method based on your suggestion and insight from existing SOTA method [3]. By simply swapping class loop with IPC loop, we successfully eliminate additional computational overhead. The revised pseudocode is provided below, and additional experimental results are shown in the table below. Under class-wise supervision without incurring extra computational cost, the performance improvements achieved with our method go beyond what could be described as "very marginal performance gain", even compared with recent SOTA plug-and-play method [4] solely designed for instance-based distillation method, further validating the generalization and robustness of our proposed method.

Method1050100
SRe2^2L21.3 ± 0.646.8 ± 0.252.8 ± 0.3
SRe2^2L^\dagger22.7 ± 0.148.4 ± 0.254.3 ± 0.3
SRe2^2L + CMI24.1 ± 0.3 (2.8\uparrow)50.3 ± 0.4(3.5\uparrow)56.0 ± 0.3(3.2\uparrow)

Original Pseudocode

for i from 1 to IPC:  # Outer loop optimizes synthetic dataset from 1 to IPC
    targets = np.arange(C) # initial hard label for all class
    for j from 1 to T:  # Inner loop optimizes images from 1 to epoch T
        Optimize S[i:, ] with L

Modified Pseudocode

for i from 1 to C:  # Outer loop optimizes synthetic dataset from 1 to class C
    targets = np.arange(IPC) # initial hard label for one class
    for j from 1 to T:  # Inner loop optimizes images from 1 to epoch T
        Optimize S[i, :] with L and CMI constraint

[1] Bayes conditional distribution estimation for knowledge distillation based on conditional mutual information

[2] Prioritize Alignment in Dataset Distillation. Openreview Submission, 2024.

[3] Are Large-scale Soft Labels Necessary for Large-scale Dataset Distillation? In NeurIPS, 2024.

[4] GIFT: Unlocking Full Potential of Labels in Distilled Dataset at Near-zero Cost. Openreview Submission, 2024.

评论

Dear Reviewer,

In response to your valuable feedback, we have provided clearer definitions, conducted additional experiments, and presented new results. If you still have any questions, we are eager to hear details. We encourage you to tell us any questions you still have, and we promise to address them in detail. Thank you.

Sincerely,

The Authors

评论

Dear Reviewer Zh5i

We appreciate your response and the contribution your feedback has made to improving our work! As the end of the discussion period approaches, if all your queries have been addressed, we kindly ask you to consider raising your rating. If you still have any doubts or reservations about our work, we are more than willing to engage in further discussion with you.

Sincerely,

The Authors

审稿意见
6

This paper introduces a novel regularization method for dataset distillation (DD) by minimizing both the distillation loss and Conditional Mutual Information (CMI) of synthetic datasets. It uses an efficient CMI estimation method to measure class-aware properties and combines CMI with existing DD techniques. Experiments show that the proposed CMI-enhanced loss significantly outperforms state-of-the-art methods, improving performance by up to 5.5%. The method can be used as a plug-and-play module for all existing DD methods with diverse optimization objectives.

优点

The strengths of this paper lie in its comprehensive experimentation across diverse datasets and network architectures, which effectively demonstrates the versatility and robustness of the proposed method. Furthermore, the method's ability to be integrated as a plug-and-play module into existing dataset distillation techniques, regardless of their optimization objectives, showcases its innovation and flexibility, making it a significant contribution to the field.

缺点

The paper lacks a clear discussion of the limitations of the proposed method. Furthermore, the authors should consider using more intuitive explanations, visual aids, and pseudocode to help readers better understand the technical details of the method.

问题

Can you discuss any potential limitations of your proposed method and suggest directions for future work to address these limitations?

评论

Thank you for dedicating your time and effort to reviewing our paper. We are deeply encouraged by your positive comments and recognition of our work. Below, we provide detailed responses to address your concerns.

Question 1: Limitation and future work.

Thank you for your positive comments and valuable suggestions. Here, we outline some limitations of our proposed method, we will add the corresponding issues in the final version of the manuscript and include them in the future work section.

Proxy Model. Computing CMI requires an proxy model to provide Z^\hat{Z} when S\mathcal{S} is used as the input, and selecting an architecture often relies on heuristic methods.

IPC=1. Minimizing CMI is equal to reduce the Kullback-Leibler (KL) divergence D()D(\cdot \|\| \cdot) between PSP_{S} and PZ^yP_{\hat{Z} \mid y}, hence, when the IPC is 1, the CMI value is 0.

Here, we suggest a future direction to better utilize the proxy model; mixing the proxy model discussed in [1] could potentially provide a more informative CMI constraint. While for the IPC=1 setting, we suggest using multi-formation introduced in [2] and [3], or other parameterization techniques. Corresponding experimental results are shown in Table 2 provided in Sec 5.2.

[1] Four eyes see more than two: Dataset Distillation with Mixture-of-Experts. OpenReview Submission, 2024.

[2] Dataset Condensation via Efficient Synthetic-Data Parameterization. In ICML, 2022.

[3] Improved Distribution Matching for Dataset Condensation. In CVPR, 2023.

评论

Dear Reviewer i6mf

Thank you once again for dedicating your valuable time to reviewing our paper and providing constructive comments! Considering the limited time available and to save the reviewer's time, we summarize our responses here.

  • 1. Limitation and future work:
    • We identify the limitations in the current deployment of CMI constraint.
    • We propose the direction of future work to address corresponding issues.

As the end of the discussion period approaches, we kindly ask if our responses have satisfactorily addressed your concerns. Your feedback would be greatly appreciated, and we would be delighted to engage in further discussions if needed.

Sincerely,

The Authors

评论

Dear Reviewer i6mf

We appreciate your response and the contribution your feedback has made to improving our work! As the end of the discussion period approaches, if all your queries have been addressed, we kindly ask you to consider raising your rating. If you still have any doubts or reservations about our work, we are more than willing to engage in further discussion with you.

Sincerely,

The Authors

审稿意见
6

This paper proposes a Conditional Mutual Information (CMI) method as a plug-and-play loss function to enhance the performance of dataset distillation methods. Experiments conducted on multiple baseline methods demonstrate the effectiveness of the CMI loss.

优点

  1. The proposed CMI method is a relatively simple yet effective approach that is plug-and-play in nature. It has demonstrated its effectiveness across multiple baseline methods.
  2. The motivation behind the method proposed in the paper is solid and is supported by a certain theoretical foundation.
  3. The experiments in the paper are comprehensive, conducted across various scales of datasets.

缺点

  1. There are now newer and more powerful methods available, such as "Towards Lossless Dataset Distillation via Difficulty-Aligned Trajectory Matching" (ICLR 2024). The authors could consider experimenting with their proposed method on these methods.
  2. The description of the method in the paper could be clearer, particularly regarding the explanation of the formula symbols, to better emphasize the key points of the approach. Currently, it appears somewhat ambiguous.
  3. In my view, using mutual information or KL divergence is not a particularly novel approach, as it has been employed in many works across various fields.

问题

Please kindly refer to the above weaknesses.

评论

We sincerely thank you for the precious time and effort in providing a wealth of suggestions to enhance the quality of our paper. We have carefully read all the comments and provide detailed point-by-point responses as follows. Hopefully, we can adequately address your concerns.

Question 1: Additional experiment on more powerful distillation method.

Thanks for the valuable suggestions on improving and completing the experimental aspects. We conduct an additional experiment by applying the CMI constraint on [1]. The experimental results show that our proposed method, as a general-purpose plugin, can achieve performance improvements across various distillation methods.

MethodCIFAR10CIFAR100TinyImageNet
105010501050
MTT65.3 ± 0.471.6 ± 0.239.7 ± 0.447.7 ± 0.223.2 ± 0.228.0 ± 0.3
MTT+CMI66.7 ± 0.372.4 ± 0.341.9 ± 0.448.8 ± 0.224.1 ± 0.328.8 ± 0.3
DATM66.8 ± 0.276.1 ± 0.347.2 ± 0.454.1 ± 0.231.1 ± 0.339.7 ± 0.3
DATM+CMI67.4 ± 0.476.6 ± 0.147.6 ± 0.355.1 ± 0.331.9 ± 0.440.6 ± 0.2

Question 2: Emphasis on the key point of the methodology.

Thank you for your suggestions on improving the presentation. Here, we make a briefly summary of our work as follows:

On the basis of observing excessive inter-class complexity in the synthetic dataset, i.e., more entangled Z^\hat{Z} when S\mathcal{S} is used as the input of a pre-trained classifier fθf_{\theta^{*}}. we introduce conditional mutual information (CMI) from information theory as a metric to evaluate the complexity of the synthetic dataset. By reconstructing the original CMI computation formula in the feature space and incorporating it as a plug-and-play regularization term, which is subsequently minimized, we achieve stable performance improvements across various distillation methods and datasets. A detailed notation is provided in table below, and we will refine our statements in the final version of the manuscript to ensure greater rigor.

NotationDescription
T\mathcal{T}the original training dataset
S\mathcal{S}the synthetic training dataset
SSthe random variable as the input to the network
YYthe random variable as the label
yysome certain label of the random variable YY
Z^\hat{Z}the random variable as the penultimate features of the network
CCthe number of the class
MMthe number of the dimensionality of Z^\hat{Z}
fθf_{\theta^{*}}the pre-trained network
QempyQ_{\text{emp}}^{y}the empirical center of class yy
λ\lambdathe hyper-parameter of CMI constraint

Question 3: Essential difference with other Information theoretic metrics.

Here, we demonstrate the advantages of CMI over mutual information (MI). In theory, the forward propagation of neural networks fθf_{\theta^{*}} is a deterministic process, causing I(X;Z^)I(X; \hat{Z}) to degrade into H(Z^)H(\hat{Z}), thereby losing its theoretical significance. On the other hand, the value spaces of XX and Z^\hat{Z} are extremely large, making I(X;Z^)I(X;\hat{Z}) very difficult to compute and approximate. As a result, variational methods are commonly employed for estimation, but these estimates are often inaccurate in practice. While CMI enables the computation process to follow a deterministic mathematical expansion by incorporating an additional variable YY. This results in a more accurate information metric without the need for additional estimation frameworks (e.g., the contrastive learning framework employed in [2]) and provides a universal regularization approach. The quantitative comparisons with [2] are shown in Table 1 provided in Sec 5.5.

[1] Towards Lossless Dataset Distillation via Difficulty-Aligned Trajectory Matching. In ICLR, 2024.

[2] MIM4DD: Mutual Information Maximization for Dataset Distillation. In NeurIPS, 2023.

评论

Dear Reviewer kma8

Thank you once again for dedicating your valuable time to reviewing our paper and providing constructive comments! Considering the limited time available and to save the reviewer's time, we summarize our responses here.

  • 1. Additional experiment on more powerful distillation method:

    • We provide additional experimental results by applying CMI on DATM across various datasets.
  • 2. Emphasis on the key point of the methodology:

    • We concisely summarize the insights and workflow of our proposed method.
    • We provide annotations for the symbols referenced in the paper.
  • 3. Essential difference with other Information theoretic metrics:

    • We demonstrate the superiority of CMI over MI based on the principle of the Information Bottleneck theory.

As the end of the discussion period approaches, we kindly ask if our responses have satisfactorily addressed your concerns. Your feedback would be greatly appreciated, and we would be delighted to engage in further discussions if needed.

Sincerely,

The Authors

评论

Dear Reviewer kma8

We appreciate your response and the contribution your feedback has made to improving our work! As the end of the discussion period approaches, if all your queries have been addressed, we kindly ask you to consider raising your rating. If you still have any doubts or reservations about our work, we are more than willing to engage in further discussion with you.

Sincerely,

The Authors

审稿意见
6

This paper proposes a new approach for dataset distillation by introducing a class-aware conditional mutual information (CMI) metric to address challenges in creating compact, representative synthetic datasets. Traditional dataset distillation methods often compress feature similarity without considering class-specific complexity, making it hard for models to generalize across different classes. This work leverages CMI as a regularization constraint, optimizes synthetic datasets and improves training efficiency as well as model performance.

优点

  1. The idea of using CMI in dataset distillation to address the inherent class-aware complexity issue is interesting.
  2. The experiments are conducted based on multiple datasets and various model architectures, providing solid evidence for the method's effectiveness.
  3. The proposed method CMI is a versatile, "plug-and-play" regularization component that can be applied to numerous dataset distillation methods, such as DSA, MTT, and IDC. This flexibility allows the approach to generalize across different scenarios and highlights its robustness.
  4. By controlling the complexity of the synthetic dataset, the CMI-enhanced loss achieves faster convergence and reduces the number of required training iterations, which is particularly beneficial for large-scale datasets and resource-intensive models.

缺点

  1. While the paper demonstrates the CMI constraint’s benefits clearly, this method also introduces additional computation overhead, especially when dealing with high-resolution datasets. Although the authors briefly mention several strategies for mitigating this cost (e.g., reducing CMI calculations frequency), a more thorough discussion on balancing cost and performance might strengthen the practical feasibility.
  2. Although empirical evidence is strong, the theoretical basis for CMI as a regularization term could be expanded. Specifically, further details on how CMI inherently captures class complexity or why it is preferable over alternative complexity measures would provide deeper insight.
  3. While the experiments on Tiny-ImageNet and ImageNet-1K are promising, it remains unclear how the proposed method scales with even larger datasets or more complex models, such as those used in real-world applications with hundreds of classes. Additional experiments in such contexts would further show the robustness of this method.

问题

  1. The paper is well-organized and clearly presents the methodology, results, and analyses. Figures and tables are effectively used to convey improvements and insights. However, further explanation of certain key terms, such as "empirical CMI," might enhance accessibility for readers unfamiliar with the topic.
  2. The ablation studies conducted to assess the influence of the weighting parameter on the CMI constraint are informative. Still, a broader exploration of other hyperparameters affecting CMI estimation, such as the dimensionality of feature space and network depth, could reveal potential optimizations.
  3. The potential of CMI for real-world applications, such as federated learning or privacy-preserving tasks, is not discussed. Given the emphasis on dataset distillation's applications in these areas, an exploration of how CMI might support these domains would align well with the broader goals of dataset distillation research.
评论

We sincerely express our gratitude for dedicating your valuable time to providing insightful suggestions that can enhance our paper. Your praise regarding our methodology, experiments, and contributions have greatly encouraged and motivated us. Our detailed responses to all of your concerns are presented below.

Weakness 1: Balance between training cost and performance.

Thanks for you suggestion on improving the experimental completeness, we conduct experiments on reducing the calculation frequency of CMI constraints as shown in table below. We report the performance and additional time consumption under different optimization settings, where the ratio of CMI optimization steps to distillation loss optimization steps is set to 1/10, 1/5, and 1. It can be observed that even under significantly reduced optimization step settings, the CMI constraint still achieves stable performance improvements.

CIFAR10CIFAR100
1/101/511/101/51
IDMadditional time (s)0.050.10.50.52.55
acc (%)61.9±0.362.1±0.262.2±0.346.7±0.147.0±0.247.2±0.4
IDCadditional time (s)0.40.84.43.26.432.0
acc (%)69.4±0.369.7±0.270.0±0.346.0±0.346.4±0.346.6±0.3

Weakness 2: Comparison with other complexity measure.

Compare to other complexity metric which evaluate the complexity of a single data point (e.g., GraDN Score [1]), CMI is a class-based concept, considering all the feature vectors within one class as a whole. On the other hand, CMI enables the computation process to follow a deterministic mathematical expansion, results in a more accurate information metric. Moreover, the insight of minimizing CMI stems from constraining the complexity of the overall synthetic dataset, making it a versatile plugin. Here, we list several information metrics used in previous distillation methods and make a brief summarization:

Influence Function [2]. influence function determines the importance of a sample by comparing the loss difference after removing a specific sample. However, this method requires retraining the model for each sample, leading to a significant computational overhead, making it impractical to optimize. Thus [3] only consider it as a evaluation metric.

Mutual Information (MI). MIM4DD [4] increases the MI between the synthetic and original datasets by incorporating a contrastive learning framework. Although it is not explicitly used for complexity constraints, MI is still an information metric and is thus included in our comparison.

Mean-square Error (MSE). IID [5] uses the MSE loss between each sample and its corresponding class center as a regularization, which is applicable only to distribution matching methods.

GraDN Score. SDC [6] uses the GraDN Score of the synthetic dataset corresponding to the proxy model during training as a regularization, which is applicable only to gradient matching methods.

For a more intuitive comparison, we have compared the performance of our work with [4], [5], and [6] as shown in Table 1 provided in Sec 5.2 and Table 8 provided in Sec A.1. We provide an additional comparative experiment where we remove another regularization term (i.e., MSE loss between the variance matrix of the synthetic dataset and that of the real dataset) from IID and compare it with our method in table below. It can be observed that CMI not only accelerates model training but also achieves better constraint performance across various distillation methods and datasets.

SVHNCIFAR10CIFAR100
105010501050
IID-DM^\dagger74.5±0.284.0±0.352.5±0.364.5±0.431.5±0.343.0±0.1
DM+CMI77.9±0.484.9±0.452.9±0.365.8±0.332.5±0.444.9±0.2
IID-IDM^\dagger81.8±0.184.7±0.359.4±0.268.3±0.245.8±0.451.0±0.3
IDM+CMI84.3±0.288.9±0.262.2±0.371.3±0.247.2±0.451.9±0.3

[1] Deep learning on a data diet: Finding important examples early in training. In NeurIPS, 2021.

[2] Understanding black-box predictions via influence functions. In ICML, 2017.

[3] What is Dataset Distillation Learning? In ICML, 2024.

[4] MIM4DD: Mutual Information Maximization for Dataset Distillation. In NeurIPS, 2023.

[5] Exploiting Inter-sample and Inter-feature Relations in Dataset Distillation. In CVPR, 2024.

[6] Not All Samples Should Be Utilized Equally: Towards Understanding and Improving Dataset Distillation. In arXiv, 2024.

评论

Weakness 3: Experimental results with larger datasets and more complex models.

Thank you for your valuable suggestions regarding enhancing the robustness of the method. We conduct additional experiments on a larger dataset (i.e., ImageNet-21K) and more complex model (e.g., ResNet101) by appling CMI constraint on SRe2^2L to evaluate the performance of our proposed method. The results shown in table below demonstrate that our method achieves significant performance improvements under all settings, validating the generality of the proposed method. We will include additional relevant experiments in the final version of the manuscript.

DatasetIPCResNet-18ResNet-50ResNet-101
SRe2^2LOursSRe2^2LOursSRe2^2LOurs
ImageNet-1K1021.3±0.624.2±0.328.2±0.430.8±0.530.5±0.323.0±0.1
5046.8±0.249.1±0.155.6±0.258.4±0.356.4±0.358.7±0.2
10052.8±0.354.6±0.261.1±0.163.6±0.362.5±0.264.2±0.3
ImageNet-21K1018.5±0.421.6±0.227.2±0.329.8±0.427.5±0.130.1±0.2

Question 1: Definition of empirical CMI.

Thank you for your suggestion on improving the presentation in the paper. Here, we propose the term empirical CMI to differentiate it from the CMI formula derived through explicit mathematical expressions in Eq. 8. In practical applications , empirical originates from the estimation of PZ^YP_{\hat{Z} \mid Y}, i.e., using QempYQ_{\text{emp}}^{Y} obtained by summing over all sampled points and averaging as a numerical approximation of PZ^YP_{\hat{Z} \mid Y}. For a specific synthetic dataset Sy\mathcal{S_y} = {(s1,y)\lbrace(s_1, y), (s2,y)(s_2, y), ..., (sn,y)}(s_n, y)\rbrace given Y=yY = y, Qempy=1SysiSyPsiQ_{\mathrm{emp}}^{y} = \frac{1}{|\mathcal{S_y}|} \sum_{s_i \in \mathcal{S_y}} P_{s_i}. We then use QempyQ_{\mathrm{emp}}^{y} as the numerical estimate of PZ^YP_{\hat{Z} \mid Y} and substitute it into Eq. 8 to calculate the computable CMI value.

Question 2: Ablation study of diverse model architectures.

Thank you for your suggestion on improving the ablation experiments. In Table 7 provided in Sec 5.5, we present the performance when using ConvNet and ResNet18 as proxy models for calculating CMI. Following your advice, we have conducted more comprehensive ablation experiments. Under the CIFAR10 with IPC=10, we utilized simply pre-trained ConvNet, AlexNet, ResNet18, and VGG11 as proxy models to compute CMI for both DM and DSA.

We sequentially report the following metrics: the depth (i.e., number of convolution layer), the dimensionality of the feature space, and the accuracy of the simply pre-trained proxy model; the start CMI value of the synthetic dataset, the CMI value at the end of the optimization process w.o. and w. CMI constraint (cstr) of the synthetic dataset, and the accuracy of the model trained using the synthetic dataset. As shown in table below, it can be observed that the dimensionality of the feature space has a greater impact on CMI values than network depth, and CMI effectively reflects the degree of clustering in Z^\hat{Z} under the same dimensional conditions. Nonetheless, the proposed CMI constraint effectively regulates the complexity of the synthetic dataset, demonstrating the strong generalization capability of our approach.

Ablation study of proxy model architecture on DM

ConvNetAlexNetVGG11ResNet18
Depth35817
Dimensionality20484096512512
Model Acc (%)79.880.279.479.5
Start CMI value0.10451.93380.00210.0016
End CMI value w.o. cstr0.08661.62170.00590.0066
End CMI value w. cstr0.03260.66420.00060.0004
Performance (%)51.250.852.452.9
评论

Question 2: Ablation study of diverse model architectures

The results on the DSA are shown in the table below.

Ablation study of proxy model architecture on DSA

ConvNetAlexNetVGG11ResNet18
Depth35817
Dimensionality20484096512512
Model Acc (%)79.880.279.479.5
Start CMI value0.10451.93380.00210.0016
End CMI value w.o. cstr0.08251.87550.00760.0051
End CMI value w. cstr0.04550.88750.00050.0004
Performance (%)53.253.454.854.7

Question 3: Application on downstream tasks

Thanks for the suggestion. Following your guidance, we explore the proposed method across two downstream tasks commonly used in dataset distillation: federated learning (FL) [7] and continual learning (CL) [8]. The experimental results are presented in the table below. It can be observed that since our proposed CMI constraint effectively provides clearer classification boundaries for the synthetic datasets, thereby offering more valuable guidance information in these real-world applications.

Federated learning

IPCDMDM + CMI
353.654.1
562.263.0
1069.270.4

Continual learning

Number of ClassesIDCIDC + CMI
2065.567.2
4061.262.4
6054.556.1
8050.453.6
10046.550.2

[7] FedDM: Iterative Distribution Matching for Communication-Efficient Federated Learning. In CVPR, 2023.

[8] Gdumb: A simple approach that questions our progress in continual learning. In ECCV, 2020.

评论

Dear Reviewer oMQb

Thank you once again for dedicating your valuable time to reviewing our paper and providing constructive comments! Considering the limited time available and to save the reviewer's time, we summarize our responses here.

  • 1. Balance between training cost and performance:

    • We provide an additional experiment to reduce the frequency of CMI computation.
  • 2. Comparison with other complexity measure:

    • We enumerated various information metrics and analyzed their theoretical shortcomings.
    • We provided a practical comparison of CMI constraint with other applicable information metrics.
  • 3. Experimental results with larger datasets and more complex models:

    • We provide additional experimental results on ImageNet-1/21K and ResNet-18/50/101.
  • 4. Definition of empirical CMI:

    • We present a more intuitive explanation for empirical CMI.
  • 5. Ablation study of diverse model architectures:

    • We employ various proxy models on different distillation methods to compute and minimize CMI.
    • we demonstrate the generalization and robustness of the CMI constraint through experimental results.
  • 6. Application on downstream tasks:

    • We validate the effectiveness of CMI in enhancing downstream tasks in federated learning scenarios.
    • We confirm the benefits of CMI for downstream tasks in continual learning settings.

As the end of the discussion period approaches, we kindly ask if our responses have satisfactorily addressed your concerns. Your feedback would be greatly appreciated, and we would be delighted to engage in further discussions if needed.

Sincerely,

The Authors

评论

We sincerely express our gratitude for dedicating your valuable time to providing insightful suggestions that can enhance our paper. Your praise regarding our methodology, experiments, and contributions have greatly encouraged and motivated us. Our detailed responses to all of your conerns are presented below.

Encouragingly, reviewers praise the insight of using CMI in dataset distillation ( R#oMQb , R#kma8 ), solid theoretical foundation ( R#kma8 , R#zh5i ), robustness ( R#oMQb , R#kma8 ) and comprehensive experiments ( R#oMQb , R#kma8, R#i6mf, R#zh5i ) of our work. To better address the reviewers' concerns, we provide point-by-point responses to each reviewer below. If there are any other questions, please don't hesitate to discuss them with us.

评论

Dear reviewer,

Could you review the author response and let them know if you are satisfied with it or if you have any additional questions?

Kind regards,

Your AC

评论

Dear Reviewers,

If you haven’t already, could you kindly review the author response and let the authors know if you are satisfied with it or if you have any additional questions? Your contribution is greatly appreciated.

Thank you!

Kind regards,

Your AC

评论

Thanks all the reviewers once again for dedicating your valuable time to reviewing our paper and providing constructive comments!

As the end of the discussion period approaches, we kindly ask if our responses have satisfactorily addressed your concerns. Your feedback would be greatly appreciated, and we would be delighted to engage in further discussions if needed.

Sincerely,

The Authors

评论

Dear reviewers,

Thank you if you have already reviewed the author response.

If not, could you kindly review it and let the authors know if you are satisfied with their response or if you have any further questions?

Kind regards,

Your AC

评论

Dear Reviewers,

Thanks again for providing the constructive review. We have tried our best to address your concerns point by point so that you can have a better understanding of our work.

However, since the discussion period is going to end in one day, we really want to hear from you and see if we have addressed your concerns. Please participate in the discussion because your feedback is important for us to improve our work. If all your queries have been addressed, we kindly ask you to consider raising your rating. If you still have other questions, we are more than willing to explain in detail.

Looking forward to your reply.

Sincerely,

The Authors

AC 元评审

This work focuses on the complexity of the synthetic dataset in dataset distillation. It proposes to apply conditional mutual information (CMI) to assess this complexity and uses this criterion as a regulariser of the existing dataset distillation methods. Experimental study shows that the proposed regulariser can be used as a plug-and-play component to work with various distillation losses to achieve improved performance and faster convergence.

Reviewers comment that the proposed method is interesting, versatile and effective; the experiments are comprehensive and solid; this work is innovative and makes significant contribution; the review is thorough; and it is technically sound. Meanwhile, the reviewers raise the issues related to the extra computational overhead, the theoretical basis of using CMI as a regulariser, the scalability with respect to dataset size and model complexity, experimenting with new methods, the novelty of this work, the discussion of limitation, the difference from an existing work on CMI, and the lack of explanation.

The authors provide a response to each of the raised issues by further clarification, additional experiments, and pseudo code. The rebuttal is overall detailed and effective. Reviewer oMQb explicitly replies that the rebuttal addresses most of the concerns they raised. However, Reviewer Zh5i still has concerns related to the difference of this work from an existing work on CMI and therefore the adequacy of the novelty. The final ratings are 6, 6, 6 and 3 (Reviewer Zh5i).

AC carefully reviews the submission, the comments, the rebuttals, the discussion, and the message from the authors. AC generally concurs with the reviewers' comments on this work. In addition, although this work adopts the CMI criterion from the literature and therefore does not have significant contribution in this regard, it does provide new insights on the complexity of the synthetic dataset and innovatively utilises CMI to consider this information in the optimisation of dataset distillation. Also, experimental study verifies the effectiveness and the generality of this work. In the follow-up discussion between reviewers and AC, Reviewer Zh5i indicates that they will not oppose acceptance if the majority of reviewers strongly advocate for it.

Taking all the factors into account, AC would like to recommend this work for acceptance. Meanwhile, the authors are strongly suggested to further improve the clarity of this work, provide more explanation, and explicitly list the differences from the work raised in the comments of Reviewer Zh5i.

审稿人讨论附加意见

The reviewers raise the issues related to the extra computational overhead, the theoretical basis of using CMI as a regulariser, the scalability with respect to dataset size and model complexity, experimenting with new methods, the novelty of this work, the discussion of limitation, the difference from an existing work on CMI, and the lack of explanation. The authors provide a response to each of the raised issues by further clarification, additional experiments, and pseudo code. The rebuttal is overall detailed and effective. Reviewer oMQb explicitly replies that the rebuttal addresses most of the concerns they raised. However, Reviewer Zh5i still has concerns related to the difference of this work from an existing work on CMI and therefore the adequacy of the novelty. The final ratings are 6, 6, 6 and 3 (Reviewer Zh5i).

AC carefully reviews the submission, the comments, the rebuttals, the discussion, and the message from the authors. AC generally concurs with the reviewers' comments on this work. In addition, although this work adopts the CMI criterion from the literature and therefore does not have significant contribution in this regard, it does provide new insights on the complexity of the synthetic dataset and innovatively utilises CMI to consider this information in the optimisation of dataset distillation. Also, experimental study verifies the effectiveness and the generality of this work. In the follow-up discussion between reviewers and AC, Reviewer Zh5i indicates that they will not oppose acceptance if the majority of reviewers strongly advocate for it. Taking all the factors into account, AC would like to recommend this work for acceptance.

最终决定

Accept (Poster)