PaperHub
5.7
/10
Poster3 位审稿人
最低5最高6标准差0.5
6
5
6
4.0
置信度
正确性3.0
贡献度2.3
表达2.3
NeurIPS 2024

Listenable Maps for Zero-Shot Audio Classifiers

OpenReviewPDF
提交: 2024-05-14更新: 2025-01-14
TL;DR

We propose a posthoc explanation method for zero shot audio classifiers.

摘要

关键词
Zero shot audio classifiersPosthoc explanations

评审与讨论

审稿意见
6

The paper introduces a post-hoc interpretation method for zero-shot audio classifiers, named LMAC-ZS (Listenable Maps for Audio Classifiers in the Zero-Shot context). It addresses the challenge of interpreting predictions from zero-shot audio classifiers that define audio classes based on textual prompts, where labels are not predefined but generated dynamically. LMAC-ZS outputs saliency maps that highlight important regions within input audio, correlating these with corresponding text prompts. The method involves a novel loss function that maintains the original audio-text similarity, enhancing interpretability. Experiments were conducted using the CLAP model on standard datasets like ESC50 and UrbanSound8K.

优点

  1. Motivation is Clearly articulated. It stresses the importance of interpretability in AI, particularly for models used in critical decision-making areas, focusing on the less-explored zero-shot audio classification.
  2. The problem of providing interpretable explanations for zero-shot audio classifiers is clearly formulated and identified as a novel challenge in the field.
  3. The paper compares LMAC-ZS with several baseline methods like GradCAM++, SmoothGrad, and Integrated Gradients, providing a thorough comparative analysis and justifying the selection based on their relevance and common use in related tasks.

缺点

  1. It is observed that in some scenarios LMAC-ZS (CT) performs better than LMAC-ZS (FULL) in metrics such as AI, AD, and AG. This is counterintuitive, as one would expect the model trained on the full CLAP dataset to perform better. An explanation for this discrepancy is needed.
  2. The paper mentions exploring the training of LMAC-ZS only on the Clotho dataset to simulate a limited computational budget scenario. It raises the question of whether training on other individual datasets or comparing the performance across different single datasets versus the full dataset would yield different insights. This aspect needs further exploration and clarification.
  3. In Table 2, the use of ESC50 contamination for ZS classification on the ESC50 dataset seems confusing. The rationale behind using the same dataset for contamination needs to be explained in detail. Similarly, in Table 3, the reason for using US8K contamination for Mel-Masking and ESC50 contamination for STFT-Masking in ZS classification on US8K requires clarification.

问题

  1. In Figure 7, the visual explanations for 'Toilet flushing' and 'Water drops' look very similar despite different similarity scores (Sim=0.69 and Sim=0.14). How does the model ensure distinguishable and meaningful explanations, and can more distinct examples be provided to demonstrate the method's effectiveness?
  2. I have a question about a set of textual prompts. What happens if there are sounds that are not in the text?

局限性

The author has addressed the limitations.

作者回复

We thank you for your constructive and positive comments. Our replies are below:

  • Regarding your comment as to the discrepancy between the L-MAC-ZS Full and L-MAC-ZS (CT) in some cases:

    This is potentially due to optimization, and choice of hyperparameters (note also that this is mainly observed for mel-masking and not STFT masking). From preliminary results on training L-MAC-ZS (Full) for few more epochs for mel-masking, and we see that the results seem to be on par.

  • Regarding your comment on training L-MAC-ZS with other subsets (subsets other than Clotho):

    In Tables 3 and 4 (the general reply above) we report the results obtained by training the decoder on different subsets. We experimented both with randomly sampled subsets of the whole CLAP dataset and with the individual datasets comprising the CLAP training data (namely AudioCaps, FSD50K, MACs, and Clotho). As mentioned above, we note that the explanation’s faithfulness is comparable when training the decoder on the full training data or a subset. We note that the training with only MACs dataset results in the lowest performance on ESC50. We note that this is likely related to the differences in data distributions. In fact, the MACs / ESC-50 Frechet Audio Distance (computed using CLAP embeddings) is the highest among all the subsets (Table 6 in the general rebuttal above). We conclude that training on a subset of the training data, the decoder can generate faithful explanations for the ZS classifier.

  • Regarding your comment:

In Table 2, the use of ESC50 contamination for ZS classification on the ESC50 dataset seems confusing.

This setting refers to the case where we create sound mixtures using two recordings with different classes from ESC50. Same is true for the US8k dataset as well. The goal is to investigate whether L-MAC-ZS is able to point out the important parts of the recording in a more complicated audio recording.

  • Regarding your comment:

In Figure 7, the visual explanations for 'Toilet flushing' and 'Water drops' look very similar despite different similarity scores (Sim=0.69 and Sim=0.14). How does the model ensure distinguishable and meaningful explanations, and can more distinct examples be provided to demonstrate the method's effectiveness?

We have provided additional qualitative examples in the .pdf document provided for the general rebuttal.

  • Regarding your question

    I have a question about a set of textual prompts. What happens if there are sounds that are not in the text?

    In this case, as demonstrated in the example in Figure 4 and more generally Figure 3, L-MAC-ZS is able to return an empty mask when the similarity between the text prompt and the input audio is low. We have provided additional evidence to this in the pdf attached to the general rebuttal.

评论

Thank you for addressing the questions! I am satisfied with the responses and have no further questions at this time.

审稿意见
5

This paper describes an extension of the LMAC method for explaining decisions made by audio classifiers. The novelty in the present article is to extend from fixed-vocabulary settings to zero-shot / open text description settings. To accomplish this, the authors propose a training objective that aims to preserve proximity of (learned) masked audio embeddings to embeddings of corresponding text descriptions. The method is evaluated on the CLAP model over two standard sound event detection datasets, and compares quite favorably to prior work along several evaluation criteria.

优点

Overall, I found this paper to be well written and easy to follow. The topic of the paper is timely and relevant, as much audio analysis research does seem to be trending toward open vocabulary settings. The empirical evaluations are appropriate and generally convincing.

The proposed method makes a lot of intuitive sense, and appears to be a natural extension of the fixed vocabulary setting (section 2.2).

缺点

The main weakness I see in this work is a somewhat shallow investigation of the proposed method itself, as opposed to high-level comparisons to prior work. The core technical contribution is described in equations 5-7, and consists of 3 terms: one to approximately preserve the embedding of audio (relative to text embedding) after masking, one to promote sparsity on the masks, and one to promote diversity of masks when conditioning on different text embeddings. As in CLAP, contrast is obtained by comparing amongst other samples within the training batch. This raises a handful of questions about the various components of the training objective which I will detail below; however, the critical issue here is that no ablation study is conducted to measure the importance of the various terms. This leads the reader at a bit of a loss to understand how these components interact with each other, whether they're all necessary, and so on.

问题

My questions primarily center around equation 6:

  • Notational quirk: Ci,jC_{i,j} is a scalar value (equation 5), right? Why are norm bars (\|) used in eq 6 when computing the loss on approximating Ci,jC_{i,j}?
  • Reasoning through the first term of eq6, a few things jump out at me. First, expanding out the definition in eq5, the summand should be equivalent to tiT(faudio(Xaudio,j)faudio(Mθ(ti,hj)Xaudio,j))\left|t_i^\mathsf{T} \left(f_\text{audio}(X_{\text{audio},j}) - f_\text{audio}(M_\theta(t_i, h_j) \odot X_{\text{audio},j}) \right)\right|. That is, the gap between the original and masked audio embedding should be small (ie the mask doesn't do much) or approximately orthogonal to tit_i, and this should hold for all tit_i in batch when the outer sum is computed.
    • What is the effect of the batch size / batch diversity on this loss? Presumably, the more (and more diverse) text embeddings we compute against, the harder it will be to achieve this orthogonality while having nontrivial masks. Similarly, a too-small batch could make it too easy to achieve this orthogonality in a high-dimensional embedding space while not producing useful masks. Is there a sweet spot or range sizes where this seems to work (on CLAP)?
  • What happens if the third term is left out of the loss? Similarly, how sensitive is this term to size and diversity of the batch?

A couple of minor questions:

  • Line 155 describes the audio reconstruction by ISTFT. It's not clear how phase is treated here, since XX appears to be magnitude spectra. Presumably phase is retained from the STFT and propagated through, but if some other method is used here (eg griffin-lim or the like) it should be made explicit.
  • In general, the spectrogram visualizations (eg figure 4) are quite difficult to read in this paper. It would be helpful to A) label the axes with appropriate units, B) use a log-scale frequency axis for display (even if the underlying method uses linear spectra, the results will be more visible), and C) use a log amplitude scaling on the magnitudes to enhance visual contrast.

局限性

The limitations and impacts are well and appropriately described by the authors.

作者回复

We thank you for your comments. Our replies are below:

  • Regarding your comment on the lack of ablation study regarding different terms of the loss function:

    We have conducted an ablation study (both quantitative and qualitative) to showcase the relevance of the unimodal diversity loss introduced in Eq. 7.

  • Regarding the notational quirk on CijC_{ij} you make reference to:

    You are correct. We will remove the norm bars in Equation 6.

  • Regarding your comment on the first term in equation 6, and whether a trivial, all-ones mask would minimize this loss function:

    It definitely would. Because of this, the second term in Equation 6 is very important, and as we indicate in line 142 of the submission, this term is needed to avoid trivial solutions (all-ones mask) : The second term in Equation 6 promotes sparsity in the generated mask to avoid trivial solutions.

  • Regarding your comment,

    What happens if the third term is left out of the loss?

    The third term in the loss function is responsible for making the decoder more sensitive to text prompts. To validate its impact on the generated explanations, we have carried out a qualitative and quantitative ablation study for this. Please see the qualitative results in the .pdf document attached for the rebuttal. The quantitative comparison indicates that by introducing the third loss term we marginally decrease the explanation’s faithfulness, while increasing the sensitivity to text prompts significantly (Fig 3 in the attached pdf).

  • Similarly, how sensitive is this term (third term) to size and diversity of the batch?

    First of all, note that the proposed loss function is not contrastive. It is rather a loss term that matches two matrices. So, the behavior is expected to be more similar to regular loss functions. To quantify the effect of batch size on the end performance of the model, we have also trained a model with BS=4 on the AudioCaps dataset and reported the results in Table 5. We observe that with BS=4 in fact the faithfulness numbers are slightly worse compared to BS=2 (potentially due to suboptimal hyperparameters).

  • Regarding your comment,

Line 155 describes the audio reconstruction by ISTFT. It's not clear how phase is treated here, since appears to be magnitude spectra. Presumably phase is retained from the STFT and propagated through, but if some other method is used here (eg griffin-lim or the like) it should be made explicit.

The phase of the original input audio is used to reconstruct the listenable audio for the explanation (same way it is done in the original L-MAC paper, equation 3.) We will clarify this in the final version of the paper.

  • Regarding your comments,

In general, the spectrogram visualizations (eg figure 4) are quite difficult to read in this paper. It would be helpful to A) label the axes with appropriate units, B) use a log-scale frequency axis for display (even if the underlying method uses linear spectra, the results will be more visible), and C) use a log amplitude scaling on the magnitudes to enhance visual contrast.

Thank you for your comment. We will use a log-amplitude scaling on the magnitudes / log-scale frequency axis for display, and label the axes.

评论

Thanks for your responses - these address all of my questions above.

评论

Thanks a lot for your positive feedback on our rebuttal! Please consider raising your score as it might help with the final decision. Thank you very much consideration.

审稿意见
6

In this paper, the authors focus on interpreting the decisions of zero-shot audio classifiers, particularly ones based on the Contrastive Language-Audio Pretraining (CLAP) model. To achieve this, the authors propose to learn a decoder that predicts a "listenable" audio saliency map (a mask on the input spectrogram) from an audio-text pair. The decoder is trained with a novel loss function that remains faithful to the characteristics of CLAP features. The authors demonstrate the effectiveness of their approach with extensive experiments across various datasets.

优点

Strengths

  1. Although there is a significant body of research focusing on interpretability of audio classifiers, interpretability of zero-shot audio classifiers (and zero-shot classifiers in general) is an under-explored area, which certainly merits more investigation.
  2. From a technical perspective, the proposed approach is sound and has no glaring weaknesses.
  3. The qualitative and quantitative results show that the method indeed works as claimed.
    • Particularly the anonymized qualitative samples provided are highly appreciated.

缺点

Weaknesses

  1. Even after reading two out of the three subsections in Section 2: Methodology, a reader learns nothing new from the paper.

    To be more specific:

    • Section 2.1 provides a high-level overview of "Contrastive Language Audio Pretraining" (CLAP) [10], while Section 2.2 gives a brief summary of the approach of the paper "Listenable Maps for Audio Classification" (L-MAC) [8]. Both these sections discuss the previous papers as is, without any new perspective.

    • It is solely Section 2.3 that pertains to the method proposed by the paper. Even in Section 2.3, the sub-subsection @line-152, "Producing Listenable Interpretations" is not something new achieved by the proposed method, but rather a feature of the previous L-MAC paper [8].

      • Notably, [8] also has an explicit sub-section with the exact same title; Section 2.2: Producing Listenable Interpretations. The authors make no effort in lines 152-156 to clarify that the "listenable" feature comes from [8]; thus this is not only redundant, but also very close to plagiarism.
    • Suggestion: I strongly suggest that Sections 2.1, 2.2, and the sub-subsection (@lines 152-156) be discussed separately as a "Background" or "Preliminaries" section, instead of Methodology. Specifically, since the proposed approach builds significantly upon L-MAC [8], it is necessary to disentangle your contributions from that of the authors of [8].

  2. The technical novelty is limited by the L-MAC paper.

    • From my understanding, the only novelty in the method is the loss function in Equation 6. And even that is mostly changing the cross entropy-based objectives in L-MAC (Equation (2) in [8]) to contrastive losses characteristic of CLAP. Overall, the same min-max objective is retained, and the same regularization term is added for mask sparsity. The same general framework from L-MAC is followed: a decoder is trained on the same dataset as the encoder to predict a mask on the input audio spectrogram.
  3. (minor weakness) Comparison against baselines.

    • The authors are comparing against the baselines of GradCAM, GradCAM++, SmoothGrad, Integrated Gradients. While all these methods were also used for comparison with the original L-MAC in [8], that was in a regular classification setting; these methods are, by nature, poorly suited to zero-shot settings.

    • For instance, for both "cat" and "glass-breaking", the CLAP model needs to "look" at the important regions of the audio to give an encoded audio feature---which can be observed in Figure 4 (for GradGAM) in the paper.

    • Suggestion: A better baseline would perhaps be from the paper "gScoreCAM: What objects is CLIP looking at?" [a]. The paper empirically establishes that for CLIP, in zero-shot settings, ScoreCAM performs better than the other forms of CAM.

    • Note: I understand that in general, there is a lack of baselines to compare to (for instance, listen-to-interpret and L-MAC are not applicable). I do not expect the authors to conduct any additional experiments.


In general, the paper has a significant amount of similarity/redundancy with the L-MAC paper [8], from the naming of sections and content (noted in Weakness 1), figures (Fig 1 in [8] and fig 2 in current paper), and even the exact set of metrics posed in the identical order in Section 3.1, which makes it difficult to discern the novelty and contributions.


[a] Chen, Peijie, et al. "gscorecam: What objects is clip looking at?." Proceedings of the Asian Conference on Computer Vision. 2022.

问题

Questions

  1. line 147 "The intuition is that the similarity between two text prompts should be reflected in the similarity of the audio embeddings from the corresponding masked spectrograms".

    Equation (7) is the only part of the loss function that is distinctly novel, as it adds a third term to the original loss in [8]. Have you tried training the overall decoder without the third term? It would be helpful to the paper if you can show that adding the third term yields a noticeable improvement over just using the first two terms; otherwise, the hypothesis remains unsupported by evidence.

  2. It is common to perform zero-shot classification with foundation models trained on a large dataset. Suppose we have one such hypothetical model, trained on a dataset with millions of samples.

    • Is it necessary to train the decoder on the same large dataset? If not, can an estimation be made as to what percentage of the training data the decoder needs to see to achieve reliable performance? (For instance, it may be the case that after seeing 20% of the data, the model achieves 80% of its full performance).
    • If it is necessary to train the decoder on the pre-training dataset, using LMAC-ZS becomes expensive in many cases. Can LMAC-ZS be jointly learned during the pre-training process of the base model (e.g. CLAP)?
      • Note that this may not be so straightforward, as the first term in the loss objective will try to match the decoder predictions to faulty entries of CC at the start of the training.

General Suggestions

  • line-114; line-122: The authors note that they omitted a part of Equation (4) for brevity. I strongly suggest against doing this; clarity is more important than brevity when posing a loss function or optimization objective.

    Equation (4) is supposed to be a min-max objective; the goal is to maximize the classification confidence of the masked-in (salient) part of the audio, while minimizing the confidence of the masked-out part. Thus the whole equation should be written together, and parts of it should not be omitted. If brevity is desired, it may be achieved by abbreviating CrossEntropy\text{CrossEntropy} to CECE or LCE\mathcal{L}_{CE}.

  • line 120: The abbreviation L-MAC is used, but it has not been defined previously. The authors should define the abbreviation on line 70.

  • line 62-63: The citation style is inconsistent. For example, in lines 59-60, it has been written "Key approaches in this category include [19, 20, 21, 22],". The same citation style can be followed in line 64: "Notable attempts in this vein include [23, 24, 25]."


[b] Shimada, Kazuki, et al. "Zero-and Few-Shot Sound Event Localization and Detection." ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2024.

局限性

Limitations

  • One limitation that I feel is not addressed is that the approach needs to be specifically trained on CLAP's data; it is not plug-and-play like Grad-CAM or similar approaches.
    • In line 216 It is mentioned that the decoder uses CNN14 layers, presumably because the audio part of CLAP is based on CNN14.
      • Now if we have another zero-shot audio classifier, LAION-CLAP, that (suppose) has the same dataset but a different transformer-like architecture, then the decoder may not be transferable. Another decoder architecture needs to be designed to suit the alternative zero-shot classifier. So for different zero-shot foundation models, it may become necessary to have different architectures, which can be a hurdle.
      • As noted earlier, for same architecture but different datasets, it is still needed to train the decoder again. Unless of course, the decoder from one pre-training dataset transfers to another dataset, as a general purpose audio method.

Apart from these scalability concerns, I believe the remaining limitations are adequately addressed.


Update: The rebuttal addresses most of my concerns, so I will raise my score to 6.

作者回复

We thank you for your comments. Our replies are below:

  • Regarding your comments on the organization of the paper, and the fact that the paper contains sections pertaining to L-MAC and CLAP:

    Our aim was to make this submission as self-contained as possible, and to provide the reader with the necessary preliminaries. We agree with your suggestion, and we will put Sections 2.1 and 2.2 under a new section called ‘Preliminaries’. We have also noted your comment on the subsection called ‘Producing Listenable Interpretations’, and we will clarify the overlaps with the original L-MAC paper.

  • Regarding your comment on limited novelty on top of L-MAC:

    This submission is an extension of the L-MAC framework to zero-shot classifiers, and hence it is called L-MAC-ZS. There are obvious similarities, like L-MAC, L-MAC-ZS is also a masking method that uses a decoder. However, beyond using a decoder network, using masking, and the fact that the masking is done in the STFT domain (to be able to produce listenable explanations) this paper addresses another, and more challenging problem compared to the original L-MAC paper.

    We would like to also state that, to the best of our knowledge, this is the first method that explores using a decoder based method to explain an audio foundation model like CLAP. Also, note that the loss function is significantly different from L-MAC. In L-MAC, the loss function tries to maximally retain logit similarity for mask-in, minimize the logit similarity for mask-out. In L-MAC-ZS however the interpreter is not trained with data with a fixed set of labels, and the loss function aims to match the original similarities between the text prompts and the audios in the batch, with the similarities obtained after masking. We also introduced a unimodal similarity term in Eq. 7, which we show is important for our results in this rebuttal (Tables 1 and 2 in the general rebuttal) .

  • Comparisons with baselines:

    We thank you for your suggestion to compare with gScoreCAM. We have added a comparison with gScoreCAM and ScoreCAM in Tables 1 and 2 of the general rebuttal. We observe that both gScoreCAM and ScoreCAM faithfulness scores are lower than L-MAC-ZS.

  • Regarding your question whether using the third term (given in equation 7) in the loss function is useful:

    We have conducted a qualitative (the rebuttal .pdf file) and quantitative ablation study (Tables 1 and 2 in the general rebuttal). We observe that by introducing the unimodal alignment loss term explanations become sensitive to the text prompt (Figure 3 in the pdf) while maintaining comparable faithfulness metrics (Tables 1 and 2). This is confirmed qualitatively as we note that explanations generated without the unimodal diversity term remain unchanged regardless of the provided text prompt.

  • Regarding your question on if we need to train the decoder on the same dataset as the foundation model to be interpreted (e.g. CLAP):

    Please note that we had partially investigated this in the submission. In Tables 1, 2, 3 of the original manuscript we have provided results with a decoder that had only been trained with Clotho (approximately ¼ of the whole training dataset). We see that results do not significantly drop. To further investigate this, we conducted an ablation study on the amount of data needed to train the decoder (Tables 3 and 4 in the general rebuttal). We experimented with randomly sampled subsets of the CLAP dataset and with the individual datasets comprising the CLAP training data (namely AudioCaps, FSD50K, MACs, and Clotho). As observed in the submission, we note that the explanation’s faithfulness is comparable when training the decoder on the full training data or a subset. The MACs-only training results in the lowest performance on ESC50. We note that this is likely related to the differences in data distributions. In fact, the MACs / ESC-50 Frechet Audio Distance (computed using CLAP embeddings) is the highest among all the subsets (Table 6). We conclude that training on a subset of the training data, the decoder can generate faithful explanations for the ZS classifier.

  • If it is necessary to train the decoder on the pre-training dataset, using LMAC-ZS becomes expensive in many cases. Can LMAC-ZS be jointly learned during the pre-training process of the base model (e.g. CLAP)?

    In this paper we have not experimented with jointly training CLAP and L-MAC-ZS. However, this is an interesting future direction. Also, please note that in response to your earlier comment, it is also possible to train LMAC-ZS on a subset of the pre-training dataset.

  • Regarding your comment,

line-114; line-122: The authors note that they omitted a part of Equation (4) for brevity. I strongly suggest against doing this; clarity is more important than brevity when posing a loss function or optimization objective.

We will include the full loss function in Equation 4.

  • Regarding your comments on citation style consistency, and undefined abbreviation in line 120:

    We will fix these issues in the final version of the paper.

  • Regarding your comment:

One limitation that I feel is not addressed is that the approach needs to be specifically trained on CLAP's data; it is not plug-and-play like Grad-CAM or similar approaches. and your related comments regarding generalizability to different foundation model architectures, and different datasets:

Generalizability to different audio encoder architectures: There’s no theoretical reason as to why the convolutional decoder would not provide faithful and understandable explanations for different foundation model architectures. However, it also remains feasible to design new decoder architectures based for a different foundation model. Regarding different training datasets: As we mentioned earlier above, we have experimented with training the decoder with subsets of the training data, and we see that the results remain comparable.

评论

Thank you, my primary concerns have been addressed. I will raise my score.

评论

Thank you for your detailed review, and your prompt response!

作者回复

We thank all the reviewers for their comments. We provide a reply to each reviewer in the corresponding rebuttal. Here, we report the additional quantitative results obtained during the rebuttal period to address the reviewer's concerns. The reviewer replies refer to the table numbers listed below.

Table 1. In-Domain Results on ESC-50, LMAC-ZS (CT) refers to LMAC-ZS trained on Clotho (with the diversity loss term), LMAC-ZS (CT) NoDiv refers to LMAC-ZS trained on Clotho without the diversity term (Eq. 7 in the paper). We only show the baselines that perform close to LMAC-ZS.

MetricAI (↑)AD (↓)AG (↑)FF (↑)Fid-In (↑)SPS (↑)COMP (↓)MM
GradCam20.3023.757.770.780.580.7211.540.14
GradCam++32.508.977.950.790.840.4112.410.35
ScoreCAM29.9712.148.820.700.750.3212.590.41
GScoreCAM29.648.566.620.790.840.3612.520.39
LMAC-ZS (CT)37.407.4311.260.780.860.5012.290.11
LMAC-ZS (CT) NoDiv37.546.3811.700.770.880.7211.590.02

Table 2. Out-of-Domain Results on ESC50 Mixtures

MetricAI (↑)AD (↓)AG (↑)FF (↑)Fid-In (↑)SPS (↑)COMP (↓)MM
GradCam23.7725.2512.240.690.490.6911.730.17
GradCam++29.5214.8410.170.700.700.3912.480.35
ScoreCAM31.397.037.050.790.870.3612.520.39
GScoreCAM28.0713.748.420.700.730.3212.590.41
LMAC-ZS (CT)35.6512.2313.040.690.740.5312.180.09
LMAC-ZS (CT) NoDiv37.359.5513.530.670.790.7511.460.02

Table 3. Ablation of the amount of training data (the model is trained only on the indicated dataset) - In-Domain Setting (ESC50)

MetricAI (↑)AD (↓)AG (↑)FF (↑)Fid-In (↑)SPS (↑)COMP (↓)MM
LMAC-ZS Clotho37.407.4311.260.780.860.5012.290.11
LMAC-ZS FSD50K34.008.3310.120.770.830.6111.830.04
LMAC-ZS AudioCaps39.005.9310.430.780.880.6811.670.07
LMAC-ZS MACs15.6122.865.320.780.610.4212.420.04
LMAC-ZS Subset (25%)41.503.487.990.790.920.6511.910.22
LMAC-ZS Subset (50%)43.703.547.860.790.910.6311.970.19
LMAC-ZS Subset (75%)40.604.747.730.790.890.6611.840.17
LMAC-ZS CLAP Data43.354.2910.570.780.900.6511.860.10

Table 4. Ablation of the amount of training data (the model is trained only on the indicated dataset) - Out-of-Domain Setting (ESC50 Mixtures)

MetricAI (↑)AD (↓)AG (↑)FF (↑)Fid-In (↑)SPS (↑)COMP (↓)MM
LMAC-ZS Clotho35.6512.2313.040.690.740.5312.180.09
LMAC-ZS AudioCaps35.9710.3511.420.680.760.7111.630.07
LMAC-ZS FSD50K26.9516.269.970.670.650.6611.590.03
LMAC-ZS MACS11.3831.544.420.680.380.4412.410.05
LMAC-ZS Subset (25%)42.655.999.810.700.840.6611.900.20
LMAC-ZS Subset (50%)39.477.529.050.710.810.6611.880.16
LMAC-ZS Subset (75%)40.427.078.840.700.830.6811.800.16
LMAC-ZS CLAP Data39.478.2811.810.690.800.6711.790.09

Table 5. Ablation on the batch size

MetricAI (↑)AD (↓)AG (↑)FF (↑)Fid-In (↑)SPS (↑)COMP (↓)MM
AudioCaps BS=4 OOD27.1518.0010.100.660.630.7011.750.03
AudioCaps BS=2 OOD35.9710.3511.420.680.760.7111.630.07
AudioCaps BS=4 ID30.0012.038.920.760.780.6411.750.06
AudioCaps BS=2 ID39.005.9310.430.780.880.6811.670.07

Table 6. Frechet Audio Distance between ESC-50 and the datasets, computed using CLAP representations

ComparisonFAD
MACs3.33
FSD50K3.04
Clotho3.09
AudioCaps3.11
CLAP Subset (25%)3.18
最终决定

All reviewers agreed that the paper has merit, with the ever increasing need of interpretability in audio tasks. There were a few concerns in the experiments, e.g., raised by reviewer vLBn and DqWc, but they were resolved during the rebuttal. The authors also provide an extensive comparison in the rebuttal to showcase the performance of their approach.

I'm sharing some of the concerns from reviewer Tsn7. The paper reads too similar to Paissan et al. (2024). I'm surprised how little contrast has been made between this paper and Paissan et al. (2024). Given how similar the approaches are, the paper should make it absolutely clear that the idea (and even the experimental design) is taken from Paissan et al. (2024) and is extended it to the zero-shot setting. This will help make the novelty clear. Given the rebuttal, I'm confident that the authors will make the best effort to improve the writing and giving more credits to Paissan et al. (2024).