PaperHub
5.7
/10
Poster3 位审稿人
最低5最高6标准差0.5
6
6
5
4.0
置信度
正确性2.7
贡献度2.3
表达3.0
NeurIPS 2024

Cross-modal Representation Flattening for Multi-modal Domain Generalization

OpenReviewPDF
提交: 2024-05-14更新: 2024-11-06

摘要

关键词
Multimodal LearningDomain GeneralizationSharpness-aware MinimizationRepresentation Flattening

评审与讨论

审稿意见
6

This paper extends the analysis of unimodal flatness to the domain of multi-modal domain generalization (MMDG) for the first time, and proposes the Cross-Modal Representation Flattening (CMRF) method. By constructing a shared representation space and cross-modal knowledge distillation, it addresses the issues of competition and inconsistency in flatness across multiple modalities. Experimental results demonstrate that CMRF significantly outperforms all baseline methods in the multi-modal multi-source domain generalization setting, effectively enhancing the model's generalization capability. In particular, it achieves an average improvement of up to 3.52% in the video-audio modality combination.

优点

  1. The new Cross-Modal Representation Flattening (CMRF) method is used for multi-modal domain generalization. This method constructs a consistent flat loss region in the representation space and utilizes cross-modal knowledge transfer to enhance the generalization ability of each modality, which is a novel approach in the multi-modal domain.

  2. The paper delves into two key issues in multi-modal domain generalization: inter-modal competition and inconsistent flatness between modalities. This analysis lays a solid foundation for proposing effective solutions.

  3. The paper mentions that applying SAM can only improve the generalization of the better modalities in multi-modal networks, while it has little to no benefit and may even be harmful for the weaker modalities. This indicates that traditional SAM-based single-modal methods, due to generalization gaps, fail to find the flat minimum value where each modality coexists, leading to inconsistent flatness and thus not fully utilizing the generalization ability of all modalities. The analysis and conclusions are correct.

缺点

  1. The paper mentions using simple moving average to build the teacher network, but how to extract knowledge from the mixed representation and how to transfer this knowledge to each modality is not clear enough.

  2. The paper points out that inter-modal competition is a key issue, but the specific performance of the competition and the detailed evaluation of the proposed mitigation strategies are not thorough. It is hoped that the authors can provide more quantitative analysis on inter-modal competition and validation of the effectiveness of the mitigation strategies.

问题

  1. In Section 3.3, the authors proposed a method for cross-modal representation flattening, but further explanation is needed on how to construct interpolation representations and how to optimize these interpolations to achieve flattening of the loss region.

  2. The authors should provide more details on how to dynamically adjust weights based on the differences in the generalization ability of each modality.

局限性

Authors have adequately addressed the limitations.

作者回复

We thank reviewer for valuable comments. We respond below to each of the concerns.

  1. The paper mentions using simple moving average to build the teacher network, but how to extract knowledge from the mixed representation and how to transfer this knowledge to each modality is not clear enough.

Response: The reason why using the mixed representation promotes multimodal domain generalization is twofold. Firstly, the mixed representations are constructed based on the moving averaged teacher network, which has flatter loss landscape compared with student network [A]. Thus, the representations from teacher network naturally leads to better generalization. Secondly, mixing the representations from one modality with representations from other modality means applying perturbations, and the applied perturbations guide the current modality to learn about other modality. Therefore, by distillation, we extract knowledge from the mixed representation and transfer them to each modality. Moreover, as the representations of each modality learn towards these interpolations, we optimize them by uni-modal classification losses, which ensures that when the losses of interpolations are large, we can also optimize them to lower losses, so that the representation loss landscape between modalities can be flattened.

  1. The paper points out that inter-modal competition is a key issue, but the specific performance of the competition and the detailed evaluation of the proposed mitigation strategies are not thorough. It is hoped that the authors can provide more quantitative analysis on inter-modal competition and validation of the effectiveness of the mitigation strategies.

Response: Modality competition refers to the mutual inhibition between modalities in joint training, which is reflected in in-domain performance straightforwardly as studied in previous literature. Below we give the uni-modal validation results (in-domain) on EPIC-kitchens with video and audio data. Modal competition is manifested in that each single modality of Base performs worse than uni-modal training, which further leads to worse out-of-domain performance for each modality, as in the results of Tabs. 1,4, where the generalization ability of each modality degrades under different modality combinations and different datasets, indicating the existance of modality competition. More results of the in-domain validation accuracy will be released later.

VideoAudio
D2,D3->D1D1,D3->D2D1,D2->D3AvgD2,D3->D1D1,D3->D2D1,D2->D3Avg
Uni-modal79.5875.5875.1976.7860.3254.2953.1655.92
Base75.7873.6072.4073.9354.5852.2349.1151.97
SAM77.0373.8173.7574.8654.9051.6049.6752.06
EoA78.9473.2075.1275.7556.8552.7652.4554.02
SimMMDG80.8674.8174.5776.7554.5853.3452.9053.60
CMRF(ours)81.2677.2175.6978.0558.7754.8954.3856.01

In contrast, our method achieves the best uni-modal in-domain performance as shown above, indicating that it optimizes modality competition effectively, which in turn improves the generalization ability to other domains as in Tab. 4. The detailed results with each domain as target and more results on HAC dataset are shown in Tab. 8-13 in appendix, demonstrating modality competition and our approach really mitigates this problem comprehensively.

  1. In Section 3.3, the authors proposed a method for cross-modal representation flattening, but further explanation is needed on how to construct interpolation representations and how to optimize these interpolations to achieve flattening of the loss region.

Response: How to construct interpolation representations can be explained by Eq. 8. The extracted representations from teacher network will sum weighted, where the weight δ\delta is sampled from distribution Beta(α\alpha, α\alpha). Here we set α<1\alpha <1, so that δ\delta is biased towards 0 or 1. Considering the semantic gap between modalities, we let interpolation closer to student modality act as its teacher signal. For example, for two representations zMz_M and zNz_N from modality M and N, one interpolation is zM,N=δzM+(1δ)zNz_{M,N}=\delta z_M + (1 - \delta ) z_N , if δ=0.8\delta = 0.8 (close to M), zM,Nz_{M,N} will be the teacher signal for modality M and if δ=0.2\delta = 0.2 (close to N), zM,Nz_{M,N} will be the teacher signal for modality N, the same as Eq. 9.

How to optimize these interpolations depends on two losses: the distillation loss and the uni-modal classification loss (L186, L187). As our response for Question 1, by distillation, we make the representations of each modality learn towards interpolations. Then by uni-modal classification losses, if the losses of interpolations are large, we can optimize them to lower losses, so that the representation loss landscape between modalities can be flattened.

  1. The authors should provide more details on how to dynamically adjust weights based on the differences in the generalization ability of each modality.

Response: The weight is adjusted under different epochs. We use the teacher network with its uni-modal classifier to get each uni-modal average validation accuracy on validation set, i.e. A^k\hat{A}_k. And if the performance gap between two modalities is greater than a certain threshold μ\mu, we put a larger weight (1.0) on the better modality to make it play a greater role in the distillation process and put a smaller weight (0.5) for the other.

[A] Devansh Arpit,et al.Ensemble of averages: Improving model selection and boosting performance in domain generalization. NIPS 2022.

评论

The authors' rebuttal dispelled some of my concerns, and I choose to maintain the score.

审稿意见
6

The paper addresses the problem of multi-modal domain generalization (MMDG). The authors identify two main challenges in MMDG: modality competition and discrepant uni-modal flatness and propose a novel method called Cross-Modal Representation Flattening (CMRF). This method optimizes the representation-space loss landscapes instead of the traditional parameter space, allowing for direct connections between modalities. The authors demonstrate the effectiveness of CMRF through extensive experiments on benchmark datasets EPIC-Kitchens and Human-Animal-Cartoon (HAC), showing superior performance in both multi-source and single-source settings.

优点

  1. The introduction of Cross-Modal Representation Flattening is novel to address the challenges of MMDG
  2. The paper provides a detailed analysis of modality competition and discrepant uni-modal flatness, which are crucial factors in multi-modal generalization
  3. Comprehensive evaluation demonstrates the effectiveness of the proposed methods
  4. The paper is very well written and easy to follow

缺点

  1. Are there any visualization or quantitative results to show that after CMRF training the loss landscape of different modalities get consistent flat region?
  2. What is the influence of α in the Beta distribution? What if it is larger than 1?

问题

Are there any visualization or quantitative results to show that after CMRF the loss landscape of different modalities get consistent flat region? What is the influence of α in the Beta distribution? What if it is larger than 1?

局限性

Currently, the paper need to test on validation set to estimate generalization of each modality for Eq. 11, which can be time-consuming with the scale increase of validation set.

作者回复

We thank reviewer for valuable comments. We respond below to each of the concerns.

  1. Are there any visualization or quantitative results to show that after CMRF training the loss landscape of different modalities get consistent flat region?

Response: Thanks a lot for the valuable comments. We can evaluate the loss flatness by applying low-frequency perturbation from the Gaussian Distribution on representations as in [A], where the Variance controls the perturbation strength. The magnitude of the performance drop indicates how flat the loss is. The results are shown Figs.1,2 in rebuttal.pdf. With the increase of Variance, the performance drop of different modalities grows quickly. Although previous SAM can obtain flatter loss landscape for the better modality (video or optical flow), its improvement to weak modality (audio) is low or even harmful. The same phenomenon also exists in some multimodal domain generalization method such as SimMMDG. In contrast, our method has the smallest performance drop on each modality, indicating that our method achieves flatter loss landscape for both modalities simultaneously, i.e., getting consistent flat region. More results on different modality combinations and different datasets will be released later.

  1. What is the influence of α in the Beta distribution? What if it is larger than 1?

Response: In this paper, we let the parameters in Beta(α\alpha, β\beta) equal, i.e., α=β\alpha=\beta, so it becomes Beta(α\alpha, α\alpha). In this way, the distribution is symmetric around the midpoint of the interval [0, 1]. When α<1\alpha <1, the Beta distribution is U-shaped, with more weight towards the boundaries (0 and 1). When α>1\alpha >1, the Beta distribution becomes symmetric and bell-shaped around 0.5. The larger the value of α\alpha, the more peaked the distribution becomes around 0.5. In this paper, we want to flatten each modality's loss landscape by distilling knowledge from the mixed multimodal representations. However, because of the knowledge heterogeneity between modalities, if the mixed representations contains much more inconsistent knowledge from other modalities, the cross-modal distillation will lead to worse performance [B]. Therefore, we want the mixed teacher information used for distillation to contain more information about the student modality, i.e., the values sampled from the beta distribution should be biased towards 0 or 1. And if α>1\alpha >1, the values sampled from the beta distribution are more likely to be 0.5, where the knowledge of two modalities is mixed equally. In this case, if the other modality contains knowledge that is inconsistent with the student modality, the performance of distillation will be degraded. This is also the example represented by Fixed Mix (mixing ratio is 0.5) in Tab. 6, which is worse than our method with α=0.1\alpha=0.1.

[A] Yixiong Zou, et al. Flatten Long-Range Loss Landscapes for Cross-Domain Few-Shot Learning. CVPR 2024.

[B] Zihui Xue, et al.The Modality Focusing Hypothesis: Towards Understanding Crossmodal Knowledge Distillation. ICLR 2023.

评论

Thank the authors for their rebuttal. Most of my concerns have been well-addressed and I thus increased my score to 6.

审稿意见
5

In this paper, the authors identify two primary limitations in multi-modal domain generalization (MMDG): modality competition and discrepant uni-modal flatness. To address these challenges, they propose a novel approach called Cross-Modal Representation Flattening (CMRF). CMRF constructs interpolations by mixing multi-modal representations from a teacher model and uses feature distillation to optimize high-loss regions between modalities. It adopts a supervised multi-modal contrastive loss to enhance the flat loss region in the shared feature space. Furthermore, CMRF employs adaptive weighting based on modal validation set accuracy to better utilize each modality. The effectiveness of the proposed method is validated through extensive experiments.

优点

  1. Flatness analysis within multi-modal domain generalization (MMDG) is innovative. Identifying modality competition and discrepant uni-modal flatness as key limitations offers a fresh perspective on the challenges in the field and has the potential to advance understanding in the broader area of multi-modal learning.
  2. Extensive experiments in various modality combinations and settings demonstrate the effectiveness of the proposed model.

缺点

  1. This paper lacks a detailed discussion on the practical implementation aspects of the proposed method, such as computational requirements and scalability.
  2. Although the authors' claims about addressing modality competition and discrepant uni-modal flatness are intuitive, the paper does not adequately address how to evaluate the loss flatness for multi-modal data. Additionally, the proposed method lacks sufficient evidence and empirical results to convincingly demonstrate its effectiveness in optimizing these specific limitations within multi-modal domain generalization.

问题

  1. What does the comparative data in Table 1 represent?

局限性

The authors have discussed the limitations in the paper.

作者回复

We thank reviewer for valuable comments. We respond below to each of the concerns.

  1. This paper lacks a detailed discussion on the practical implementation aspects of the proposed method, such as computational requirements and scalability.

Response:

1) Computational requirements: Firstly, the Projector (3-layer MLP) for each modality is very small in our method and is also used in previous method such as SimMMDG, indicating that additional computation from more parameters is negligible. Secondly, although our method requires more computation from teacher model, it only requires the forward propagation but not gradient backpropagation on it, so the additional computation is also limited. Finally, the main additional computational costs come from testing on validation set for adaptive weights Eq.11, as discussed in our Limitation section. The one epoch training time on D2,D3->D1 with video-audio data from EPIC-kitchens of different methods are shown below (in seconds). SWAD and our CMRF need more computation as they both require testing on validation set (The value in parentheses indicates the validation time). Searching for an efficient method to evaluate the generalization of each modality would be our future work, such as add low-frequency noise to stimulate domain shift for evaluation.

BaseSAMSAGMSWADEoARNA-NetSimMMDGCMRF (ours)
308.5346.4345.6367.9309.2311.3327.7369.9 (59.4)

2) Scalability: Our method has no restrictions on the type and number of modalities. Since our method focuses on representation space other than parameter space, the condition that our method can be applied is that a consistent representation space of multiple modalities is available, which is definitely satisfied in multimodal deep learning by mapping with neural networks. This also demonstrates the flexibility of our approach to different model structures. In this paper, our experiments include the three common modalities (video, audio, optical flow), as well as different numbers of modalities, to verify our claims. All these show that our method has good scalability.

  1. Although the authors' claims about addressing modality competition and discrepant uni-modal flatness are intuitive, the paper does not adequately address how to evaluate the loss flatness for multi-modal data. Additionally, the proposed method lacks sufficient evidence and empirical results to convincingly demonstrate its effectiveness in optimizing these specific limitations within multi-modal domain generalization.

Response: Thanks a lot for the valuable comments. We can evaluate the loss flatness by applying low-frequency perturbation from the Gaussian Distribution on representations as in [A], where the Variance controls the perturbation strength. The magnitude of the performance drop indicates how flat the loss is. The results are shown Figs.1,2 in rebuttal.pdf. With the increase of Variance, our method has the smallest performance drop on each modality, indicating that our method achieves flatter loss landscape for both modalities simultaneously and in turn provides flatter multi-modal loss landscape.

In order to better demonstrate the effectiveness of our approach in solving the two specific limitations, please allow us to restate the two problems. Modal competition refers to the mutual inhibition between modalities in joint training, which is reflected in in-domain performance straightforwardly as studied in previous literature. Below we give the uni-modal validation results (in-domain) on EPIC-kitchens with video and audio data. Modal competition is manifested in that each single modality of Base performs worse than uni-modal training, which further leads to worse out-of-domain performance as shown in Tab. 8 in appendix. Our method achieves the best uni-modal in-domain performance, indicating that it optimizes modal competition effectively, which in turn improves the generalization ability to other domains as in Tab. 4.

VideoAudio
D2,D3->D1D1,D3->D2D1,D2->D3AvgD2,D3->D1D1,D3->D2D1,D2->D3Avg
Uni-modal79.5875.5875.1976.7860.3254.2953.1655.92
Base75.7873.6072.4073.9354.5852.2349.1151.97
SAM77.0373.8173.7574.8654.9051.6049.6752.06
EoA78.9473.2075.1275.7556.8552.7652.4554.02
SimMMDG80.8674.8174.5776.7554.5853.3452.9053.60
CMRF(ours)81.2677.2175.6978.0558.7754.8954.3856.01

Discrepant uni-modal flatness refers to the fact that the generalization gaps between modalities make it difficult to optimize flat regions for each modality simultaneously. According to our analysis in Sec. 3.2 and results in Tabs. 1, 4 and above tabel, previous SAM-based methods contribute more to the better modality and contribute less or even are harmful to weak modalities. The Fig. 1 in rebuttal.pdf also show that SAM obtains flatter loss for better modality (video) and sharper loss for weak modality (audio). In contrast, our method achieves flatter loss landscape simultaneously for each modality, leading to consistent flatness.

The above shows that the effectiveness of our method for both modal competition and Discrepant uni-modal flatness, and the improvement of multimodal domain generalization comes from solving both problems at the same time, not just one of them.

  1. What does the comparative data in Table 1 represent?

Data in Tab. 1 are the average results by treating each domain as target domain, e.g., each data is the average of results from A,C→H H,C→A H,A→C on HAC.

[A] Yixiong Zou, et al. Flatten Long-Range Loss Landscapes for Cross-Domain Few-Shot Learning. CVPR 2024.

评论

Thank the authors for addressing my concerns. I tend to increase my score to 5.

作者回复

We thank all reviewers for their positive feedback:

  1. A fresh perspective of modality competition and discrepant uni-modal flatness for MMDG [eDBu, iu3y, 8XoV];
  2. Flatness analysis within multi-modal domain generalization (MMDG) and the proposed method for flatting representation loss landscape are innovative [eDBu, 8XoV];
  3. Good writing and easy to follow [iu3y]
  4. Extensive experiments demonstrate the effectiveness of the proposed model [eDBu, iu3y].

We respond to each reviewer's comments below. We provide some figures in the rebuttal.pdf.

最终决定

The paper identifies two main issues related to multi-modal domain generalization (MMDG) methods -- modality competition and discrepant uni-modal flatness. The authors propose a CMRF method, which mixes multi-modal representations from a teacher model and uses feature distillation to optimize high-loss regions between modalities. It uses a supervised multi-modal contrastive loss to enhance the flat loss region in the shared feature space. Furthermore, CMRF does adaptive weighting based on modal validation set accuracy to better utilize each modality. The effectiveness of the proposed method is validated through experiments.

The reviews of the paper are overall positive. The paper is recommended for acceptance as a poster paper. The authors are advised to address the remaining feedback of the reviewers, which were not handled during author feedback time.