PaperHub
7.0
/10
Poster3 位审稿人
最低3最高5标准差0.9
3
5
5
4.3
置信度
创新性2.7
质量2.7
清晰度2.7
重要性2.7
NeurIPS 2025

MGE-LDM: Joint Latent Diffusion for Simultaneous Music Generation and Source Extraction

OpenReviewPDF
提交: 2025-05-11更新: 2025-10-29
TL;DR

We propose unified latent diffusion framework for simultaneous music generation, source imputation, and query-driven arbitrary source extraction.

摘要

关键词
Music GenerationSource ExtractionInpainting

评审与讨论

审稿意见
3

The authors introduce MGE-LDM, a unified latent diffusion framework for simultaneous music generation, source imputation, and text-driven source separation. Unlike prior methods that depend on fixed instrument classes or assume additive mixing in the waveform domain, MGE-LDM operates in a shared latent space and models the joint distribution over mixtures, submixtures, and individual sources. By formulating both stem completion and source extraction as conditional inpainting tasks and leveraging CLAP embeddings for text conditioning, the model enables class-agnostic manipulation of arbitrary instruments.

优缺点分析

While none of the individual components of the proposed architecture are novel, the combination of diffusing stems in the latent space based on clap conditioning signals (and therefore not requiring explicit stem labels) seems novel.

Unfortunately, I’m not convinced by the results. The empirical study using VGGish FAD is poor, as multiple previous works have shown that VGGish is a suboptimal embedding model for music [1-3]. Additionally, while I appreciate the authors having trained various models (T1-4), I don’t see how it makes sense to compare these against a baseline only trained on a single subset of the data. While the results on synthetic data (slakh) seem decent, the model underperforms when listening to real music (moises, musdb). The biggest issue I have is the missing human evaluation, as the FAD numbers don’t show the full picture.

[1] Gui, Azalea, et al. "Adapting frechet audio distance for generative music evaluation." ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2024.

[2] Grötschla, Florian, et al. "Benchmarking Music Generation Models and Metrics via Human Preference Studies." ICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2025.

[3] Huang, Yichen, et al. "Aligning Text-to-Music Evaluation with Human Preferences." arXiv preprint arXiv:2503.16669 (2025).

问题

  • Why are the authors only training on 16 kHz? As far as I know the Stable Audio pipeline supports up to 44.1 kHz stereo.
  • Why are the authors using an outdated version of Demucs? The Hybrid Demucs is significantly better compared to the original.
  • Why does the demo page only have comparisons in the Slakh section but not for Moises/MusDB? Especially for the source extraction, which can be easily compared to existing methods.

局限性

The authors have a section on limitations in the appendix. I suggest they move this to the main text.

最终评判理由

Even though paper does not show substantial improvements over prior work, the approach contains novelty and should be interesting to the audience of NeurIPS.

格式问题

N/A

作者回复

We sincerely thank Reviewer cGKq for your careful reading and constructive comments. Below, we address your main concerns:

"Unfortunately, I’m not convinced by the results. The empirical study using VGGish FAD is poor, as multiple previous works have shown that VGGish is a suboptimal embedding model for music [1-3]."

Authors' Response: We appreciate your feedback and agree that VGGish-based FAD is not the ideal metric for music generation. We initially adopted it for its widespread use as in SingSong [4] and MSDM [5], but we acknowledge its limitations. We additionally calculated the FAD using CLAP-MA embeddings (LAION-CLAP checkpoint music_audioset_epoch_15_esc_90.14), which Grötschla et al. [2] have shown best correlation with human perception.

Below, Table A reports the total-generation results using FAD-CLAP-MA.


Table A: Total generation result with FAD-CLAP-MA \downarrow

ModelTrainsetSAS_ASfullS_{full}MuM_uMoM_o
MSDMSAS_A0.3290.2900.7670.744
MSG-LDSAS_A0.1910.3120.6580.697
T1T_1SAS_A0.110 (0.450)0.2850.6370.627
T2T_2SAS_A+SBS_B0.368 (0.359)0.0990.5780.621
T3T_3MuM_u+MoM_o0.731 (0.669)0.6320.2220.165
T4T_4SAS_A+SBS_B+MuM_u+MoM_o0.563 (0.712)0.4260.2350.186

It shows a similar relative ranking to Table 1 of the main paper, with lower absolute values. One notable difference is that our T1T_1 model achieves the lowest FAD-CLAP-MA score compared to the baselines, which is a clear contrast to the VGGish-based results in Table 1 of the main paper.

We also computed the FAD-CLAP-MA for partial generation task, where the results are shown below in Table B.


Table B: Partial generation result with sub-FAD-CLAP-MA \downarrow

ModelTrainsetBDGPBDBGBPDGDPGPBDGBDPBGPDGP
MSDMSAS_A0.1030.1000.0660.0960.2270.1890.2520.1350.1910.3020.2750.3570.5550.403
MSG-LDSAS_A0.0710.0700.0760.0770.1300.1350.1430.1170.1210.1570.1850.1880.2410.190
T1T_1SAS_A0.1400.1700.1530.1540.1840.1750.1810.2000.1890.1840.2210.2180.1900.214
T2T_2SAS_A+SBS_B0.2590.3290.2510.2640.4200.3020.3620.3800.3610.3370.5000.5170.4220.460
T3T_3MuM_u+MoM_o0.1780.2020.3560.2530.3070.4020.3400.3680.3200.4770.4450.4690.6320.484
T4T_4SAS_A+SBS_B+MuM_u+MoM_o0.1700.1870.3670.2510.2530.4050.3080.3500.2890.4550.4500.4360.5710.486



ModelTrainsetBrassC.P.Org.PipeReedStr.S.LeadS.Pad
T1T_1SAS_A0.2780.0860.0690.3270.4620.2010.3350.118
T2T_2SAS_A+SBS_B0.4250.0710.1290.3270.3300.2960.1230.313
T3T_3MuM_u+MoM_o0.3610.0870.0620.0450.0900.0870.0790.060
T4T_4SAS_A+SBS_B+MuM_u+MoM_o0.3120.0720.1070.0570.3400.2080.0850.063

The overall relative pattern of the Table B closely matches Table 2 of the main text, with MSG-LD achieving the best scores on most stem combinations. We believe these additional metrics offer a more reliable assessment of partial generation performance.

We also agree that FAD alone cannot capture musical coherence, so we also evaluated the COCOLA score [6] for partial generation (please refer to the table and discussion in our response to Reviewer W8on).

"Additionally, while I appreciate the authors having trained various models (T1-4), I don’t see how it makes sense to compare these against a baseline only trained on a single subset of the data."

Authors' Response: You are correct that only the baselines (MSDM and MSG-LD) and our T1T_1, each trained on the same SAS_A set, are directly comparable. We included T2T_2~T4T_4 to illustrate how training on larger and more diverse datasets affects performance, since dataset-agnostic scaling is rarely evaluated in prior work and no existing baselines cover these settings. We hope these results will serve as useful reference points for future research.

For additional comparability, we also trained Hybrid Demucs on the SAS_A set. The corresponding results are shown in Table C below to strenghten the baseline comparison.


Table C: Source extraction results with Hybrid Demucs. Column headers indicate the target stem and its corresponding set.

ModelTrainsetB (SAS_A)D (SAS_A)G (SAS_A)P (SAS_A)V (MuM_u)B (MuM_u)D (MuM_u)V (MoM_o)B (MoM_o)D (MoM_o)G (MoM_o)P (MoM_o)
DemucsMuM_u1.290.92--1.591.671.290.891.481.08--
HDemucsMuM_u1.490.90--1.501.991.530.831.711.1--
HDemucsSAS_A0.830.570.780.79-1.871.22-1.611.011.450.99
T1T_1SAS_A1.280.661.271.073.801.911.335.151.611.102.862.68

"The biggest issue I have is the missing human evaluation, as the FAD numbers don’t show the full picture."

Authors' Response: We agree that human evaluation is crucial for assessing generated music quality. Accordingly, we conducted a listening study on a total generation task, and the results are shown in Table D below.


Table D: Mean Opinion Score (MOS) for the total‐generation task, rated from 1 (bad) to 5 (excellent). Two aspects were evaluated: Quality (how much the snippet sounds like noise versus a real song) and Coherence (consistency across generated stems). A total of 24 subjects participated, each listening to 8 segments per model. Scores are reported with 95% confidence intervals.

ModelGround truthMSDMMSG-LDT1T_1T2T_2T3T_3T4T_4
Quality3.79±0.182.05±0.163.80±0.173.63±0.173.92±0.163.81±0.163.92±0.16
Coherence3.75±0.192.06±0.183.70±0.173.54±0.183.99±0.153.75±0.173.79±0.15

The results indicate that T1T_1 is slightly less preferred than MSG-LD, whereas T2T_2~T4T_4 receive higher preference ratings, demonstrating that expanding the training dataset can benefit the perceptual quality of generated music.

Q:"Why are the authors only training on 16 kHz? As far as I know the Stable Audio pipeline supports up to 44.1 kHz stereo."

Authors’ Response: We initially chose 16 kHz mono to match the baseline models and ensure a fair comparison. We agree that 16 kHz mono is not sufficient for high-fidelity music generation (as noted in our Limitations section). Having demonstrated the validity of our framework at 16 kHz, we plan to extend MGE-LDM to 44.1 kHz stereo in future work.

Q:"Why are the authors using an outdated version of Demucs? The Hybrid Demucs is significantly better compared to the original."

Authors’ Response: Thank you for pointing this out. We have attached results for Hybrid Demucs (HDemucs) in Table C above. All models were trained to convergence. In our experiments, HDemucs results have not shown significant improvements over standard Demucs, so we initially reported the original Demucs comparison in the main paper, consistent with the MSDM paper. However, we agree that including HDemucs in the main tables is more appropriate and will update the manuscript accordingly.

Q:"Why does the demo page only have comparisons in the Slakh section but not for Moises/MusDB? Especially for the source extraction, which can be easily compared to existing methods."

Authors’ Response: Thank you for your feedback regarding the demo page. On the “Slakh Full” demo page, we initially showcased only T4T_4 results, assuming that adding other references would not be informative given the diversity of instruments. However, we agree that this limits comparability. Accordingly, we have now added Demucs inference examples alongside our model under “Source Extraction (Full) / MUSDB18 & Moises". We hope this makes the demo more informative and directly comparable.

"The authors have a section on limitations in the appendix. I suggest they move this to the main text." Thank you for the suggestion. We will move the Limitations section into the main text to ensure it is clearly visible to readers.

We appreciate your valuable feedback and will incorporate these changes in the final manuscript to enhance its clarity and rigor.

References

[1] Gui et al. "Adapting frechet audio distance for generative music evaluation." ICASSP 2024

[2] Grötschla et al. "Benchmarking Music Generation Models and Metrics via Human Preference Studies." ICASSP 2025

[3] Huang et al. "Aligning Text-to-Music Evaluation with Human Preferences." arXiv:2503.16669, 2025.

[4] Donahue et al. "Singsong: Generating musical accompaniments from singing." arXiv:2301.12662, 2023.

[5] Mariani et al. "Multi-source diffusion models for simultaneous music generation and separation." ICLR 2024

[6] Ciranni et al. "Cocola: Coherence-oriented contrastive learning of musical audio representations." ICASSP 2025.

评论

I thank the authors for their effort made in responding to my concerns.

Reading the response I would like to highlight two aspects:

  1. The authors agree, that the only fair comparison between the proposed method and previous work/baselines is T1T_1. Therefore, these tables should be split as T2T_2 through T4T_4 should be treated more as an ablation on how the performance increases when more & varied data is added. For a truly fair comparison T2T_2 through T4T_4 would have to be compared with previous methods trained on the same data.
  2. Looking at the comparison of T1T_1, it outperforms previous approaches for the subset-generation (bass, drums, guitar, and piano) in FAD, it underperforms for the full generation in FAD, it slightly underperforms in the human evaluation of the total generation, it underperforms according to the partial generation results, and it underperforms against HDemucs baseline on separation results. Is my understanding in this regard correct?
评论

Thank you for your comments and your careful review of our rebuttal.

  1. As you suggested, we will present T1T_1 results alone alongside the baselines for a fair comparison, and report T2T_2-T4T_4 in a separate table. We will explicitly note in the manuscript and table captions that the T2T_2-T4T_4 results serve purely to evaluate the impact of dataset scaling.

  2. You are correct in your summary. However, if "full generation" refers to specifically to the FADCLAPFAD_{CLAP} result for SfullS_{full} in the total-generation task, T1T_1 actually outperforms the baseline. We will restate the exact performance breakdown as you summarized in the manuscript to make these trade-offs clear.

We appreciate your guidance and will incorporate these revisions into the final draft.

评论

I thank the authors for their clarifying statements. I have no further questions at this time.

评论

Thank you for your time and careful consideration. We appreciate your feedback, and please don’t hesitate to reach out if you have any further questions.

评论

The reviewer should re-consider their evaluation, especially after additional quantitative evaluation on COCOLA score, which complements the sub-FAD evaluation (and offers a more important metric for capturing coherence on the partial generation task). Also, using VGGish embeddings for sub-FAD is not really poor. Practically all the field of accompaniment generation (SingSong [1], MSDM [2], GMSDI [3], COCOLA [4], MSG-LD [5], JEN-1 Composer [6], STAGE [7]) uses VGGish embeddings for sub-FAD!. Also with the authors providing metrics with CLAP embeddings and giving a more detailed picture, the current experimental results are more than sufficient.

References

  • [1] Donahue, Chris, et al. "Singsong: Generating musical accompaniments from singing." arXiv preprint arXiv:2301.12662 (2023).
  • [2] Mariani, Giorgio, et al. "Multi-Source Diffusion Models for Simultaneous Music Generation and Separation." The Twelfth International Conference on Learning Representations.
  • [3] Postolache, Emilian, et al. "Generalized multi-source inference for text conditioned music diffusion models." ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2024.
  • [4] Ciranni, Ruben, et al. "Cocola: Coherence-oriented contrastive learning of musical audio representations." ICASSP 2025.
  • [5] Karchkhadze, Tornike, Mohammad Rasool Izadi, and Shlomo Dubnov. "Simultaneous music separation and generation using multi-track latent diffusion models." ICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2025.
  • [6] Yao, Yao, et al. "Jen-1 composer: A unified framework for high-fidelity multi-track music generation." Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 39. No. 13. 2025.
  • [7] Strano, Giorgio, et al. "STAGE: Stemmed Accompaniment Generation through Prefix-Based Conditioning." arXiv preprint arXiv:2504.05690 (2025).
评论

Dear reviewer W8on,

The provided experimental results are indeed more than sufficient, I have not said anything to the contrary. However, from how I see it, the results don't show outperformance of prior work to warrant acceptance. The results from COCOLA look promising, but the rebuttal is not a place to meticulously inspect and re-evaluate new results. For example. how can we be sure there was no inadvertent information leakage, as COCOLA trained on many of the same datasets that were also used in this work. All the provided results together give an impression of not significantly better than previous work, especially the human evaluation which is the gold-standard metric. Therefore, I politely disagree with your assessment and maintain my original score.

On the matter of VGGish. SingSong started with VGGish at a time when this drawback was not known yet, and then every other paper followed by reusing it. Does this mean it's a good embedding model? I would argue not. Especially since there is now ample evidence to support that VGGish is a poor choice for music models.

We can gladly continue this discussion during the AC-reviewer phase.

审稿意见
5

This paper follows the line of research set up by MSDM (Mariani et al. 2024), where a diffusion model is trained on stem-separated musical data (of fixed instrument classes) and is able to perform simultaneously full track generation, accompaniment generation given conditioning tracks and source separation. MSDM was generalised in GMSDI (Postolache et al. 2024), where text conditioning is introduced and by MSG-LD (Karchkhadze et al. 2025), which sets the diffusion model in a latent space, improving quality. MSG-LD though keeps the number of instrument classes fixed like MSDM and cannot generate arbitrary classes via text conditioning. Doing all tasks in the latent domain is difficult especially because of source separation, which requires linearity in the likelihood function (that's why GMSDI works in the waveform domain only). In this paper, the authors propose MGE-LDM, a method that aims to improve on both limitations, defining a latent diffusion model (via Stable Audio Open) that can be text conditioned and can solve all three tasks simultaneously. They train over triples composed by (z(m),z(u),z(s))(z^{(m)}, z^{(u)}, z^{(s)}), where z(m)z^{(m)} is a latent mixture, z(u)z^{(u)} is a latent sub-mixture (the mixture minus a source) and z(s)z^{(s)} is a source. At the same time they condition on CLAP embeddings of the form (c(m),c(u),c(s))(c^{(m)}, c^{(u)}, c^{(s)}). The three tasks at inference time:

  • Total generation: sample the triple and decode z(m)z^{(m)}.
  • Partial generation: inpaint (z(m)),z(s))(z^{(m)}), z^{(s)}) given z(u)z^{(u)} and decode z(s)z^{(s)}.
  • Source separation: inpaint (z(u)),z(s))(z^{(u)}), z^{(s)}) given z(m)z^{(m)} and decode z(s)z^{(s)}.

The authors also introduce track-individual noising in order to improve the inpainting mechanism. The authors benchmark their model against MSDM and MSG-LD reaching good results (especially on source separation), with improved generalisability of the model.

References

  • Mariani, Giorgio, et al. "Multi-Source Diffusion Models for Simultaneous Music Generation and Separation." ICLR 2024.
  • Postolache, Emilian, et al. "Generalized multi-source inference for text conditioned music diffusion models." ICASSP 2024.
  • Karchkhadze, Tornike, Mohammad Rasool Izadi, and Shlomo Dubnov. "Simultaneous music separation and generation using multi-track latent diffusion models." ICASSP 2025.

优缺点分析

Strengths

  • It is interesting to see the multi-track music generation research direction being improved, especially given the presented model it is general enough to let one train over multiple stem-separated datasets at the same time. This is a great departure from models with fixed classes (MSDM, MSG-LD).

  • Quantiative evaluation is thorough, comparing to MSDM and MSG-LD in a plethora of tasks. Would have been cool to compare to GMSDI given the commonality of text conditioning, but such paper did not release code / checkpoints so I agree it is a thorough task. (still one key metric is missing for partial generation, see Weaknesses #1)

  • The track-aware inpainting method is very interesting. Also seeing the comparison with RePaint in Table 5 of Appendix D and seeing the presented inpainting method as the best performing, gives a strong indicative that this should be the default mechanism for multi-track generation (at least in the absence of a constraining latent likelihood, see Weaknesses #2).

  • Cool to see a discussion about iterative generation in Appendix C. Having the individual separated tracks of a mixed track was one of the rationales of MSDM so it is interesting to see one is not losing that feature (it is actually generalised).

  • Overall the paper is great to read and a good contribution in the (controllable) music generation domain.

Weaknesses

  • As the author acknowledge in the limitations section, we still depend on labels in a loose way when training.

  • A limitation of the paper is that partial generation metrics are computed on sub-FAD, a metric which, while having served well as an initial foray in evaluating accompaniments models (e..g, in SingSong, in MSDM, etc.), there is empirical evidence (Ciranni et al. 2025) that it is not optimal for such a task, favouring randomly mixed stems over natural combinations. The authors should also compute metrics such as COCOLA score (Ciranni et al. 2025), which are better suited for the task. Also in that paper one finds a benchmark on a conditional model based on Stable Audio Open, released publicly, and trained on MusDB and MoisesDB, which shares the diffusion architecture of the presented paper: it would be interesting to see how the presented joint model compares to the conditional one.

  • While the authors acknowledge that the nonlinearity of the latent space makes it difficult to constrain the sources in a mixture and they opt out for their approach, I believe that the three entries of a triple are not necessarily constrained to sum well in the data (waveform) domain. This could have a negative impact in the performance of the model. Thus, finding the best way to solve compositional music generation in the latent domain it still remains an open problem, and maybe a more constrained approach would be given, for example, by learning latent-domain likelihood functions like in (Postolache et al. 2023). The authors could discuss this in the concluding remarks of the paper or in limitations.

References

  • Ciranni, Ruben, et al. "Cocola: Coherence-oriented contrastive learning of musical audio representations." ICASSP 2025.
  • Postolache, Emilian, et al. "Latent autoregressive source separation." Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 37. No. 8. 2023.

问题

  • Line 162: Do we have a different pp on each c(k)c^{(k)}?
  • Line 176: I would say: "whose score we approximate with fθf_\theta".
  • Line 270: "Results" instead of "Result"
  • Line 293: This makes sense since FAD cannot capture the distribution shift.
  • Line 297: I think the authors could improve on these results with negative conditioning of unwanted tracks (see GMSDI, Fig. 2).
  • Line 630: Before giving the expression of the velocity, the expression of ztz_t in function of ϵ\epsilon and z0z_0 should be given.
  • Line 631: The 00 subscript should be not bold.

局限性

yes

最终评判理由

I keep my full accept on this paper! This is a great improvement in the area of compositional audio generation, finally having a well-defined stem-level generative model in a latent domain. I am glad the authors introduced evaluations on COCOLA score, showcasing improved stem coherence with respect to previous baselines.

格式问题

no

作者回复

We sincerely thank Reviewer W8on for the through and insightful feedback. We appreciate your recognition of our contritbutions to multi-track music generation, and we address your comments below.

"A limitation of the paper is that partial generation metrics are computed on sub-FAD, a metric which, while having served well as an initial foray in evaluating accompaniments models (e..g, in SingSong, in MSDM, etc.), there is empirical evidence (Ciranni et al. 2025) that it is not optimal for such a task, favouring randomly mixed stems over natural combinations. The authors should also compute metrics such as COCOLA score (Ciranni et al. 2025), which are better suited for the task. Also in that paper one finds a benchmark on a conditional model based on Stable Audio Open, released publicly, and trained on MusDB and MoisesDB, which shares the diffusion architecture of the presented paper: it would be interesting to see how the presented joint model compares to the conditional one."

Authors' Response: You're correct that sub-FAD alone can bias toward random mixtures rather than musical coherence. In response to your suggestion, we computed the COCOLA score [1] on exactly the same partial-generation outputs reported in Table2 of the main paper. The results are shown below, with the highest COCOLA score per column bolded, and among models trained on the SA\mathbf{S}_A set (MSDM, MSG-LD, or our T1), the best result underlined.

ModelTrainsetBDGPBDBGBPDGDPGPBDGBDPBGPDGP
MSDMSA**S**_A57.7457.7451.8751.8757.0657.0656.1056.1045.9645.9657.6457.6454.4854.4850.6750.6749.4049.4055.7555.7543.2543.2540.4640.4652.8252.8247.0947.09
MSG‑LDSA**S**_A58.1858.1850.5850.5858.55\underline{58.55}58.3058.3046.1246.1257.6357.6356.2956.2949.1149.1147.5147.5157.9457.9443.7043.7041.3341.3354.1954.1944.9844.98
T1\mathcal{T}_1SA**S**_A60.93\underline{60.93}53.71\underline{\mathbf{53.71}}58.0958.0959.78\underline{59.78}49.77\underline{\mathbf{49.77}}60.43\underline{60.43}59.16\underline{\mathbf{59.16}}52.95\underline{\mathbf{52.95}}51.90\underline{\mathbf{51.90}}60.03\underline{60.03}46.28\underline{46.28}43.46\underline{43.46}56.77\underline{\mathbf{56.77}}49.73\underline{\mathbf{49.73}}
T2\mathcal{T}_2SA**S**_A+SB**S**_B61.29\mathbf{61.29}51.4351.4362.61\mathbf{62.61}61.35\mathbf{61.35}48.7948.7960.94\mathbf{60.94}59.1259.1250.8250.8249.3949.3960.21\mathbf{60.21}46.0646.0642.8142.8156.7356.7347.0147.01
T3\mathcal{T}_3Mu**M**_u+Mo**M**_o55.4055.4044.7644.7657.2957.2959.3459.3443.9043.9055.1855.1856.0956.0945.6845.6845.4245.4255.9155.9144.1044.1043.3543.3551.4451.4443.5343.53
T4\mathcal{T}_4SA**S**_A+SB**S**_B +Mu**M**_u+Mo**M**_o56.7356.7349.6549.6560.5860.5861.0661.0646.9246.9257.8857.8857.8257.8250.6650.6649.9349.9358.3258.3246.53\mathbf{46.53}45.39\mathbf{45.39}54.0246.86
Ground Truth60.1160.1153.5553.5556.3356.3357.1657.1650.4250.4259.7459.7458.5958.5953.1153.1152.2952.2957.8047.0647.0644.3244.3256.0756.0750.1850.18




ModelTrainsetBrs.C.P.Org.PipeReedStr.S.LeadS.Pad
T1SA\mathbf{S}_A63.2063.2056.4456.4459.5259.5260.5960.5962.3562.3561.1161.1163.7463.7458.6658.66
T2SA**S**_A+SB**S**_B64.3264.3265.21**65.21**65.46**65.46**62.3162.3164.31**64.31**62.3362.3364.5264.5261.10**61.10**
T3Mu\mathbf{M}_u+Mo\mathbf{M}_o60.2360.2354.7454.7456.7056.7060.1860.1856.5056.5060.6660.6662.0362.0361.0061.00
T4SA**S**_A+SB**S**_B+Mu\mathbf{M}_u+Mo\mathbf{M}_o65.94**65.94**64.1064.1064.4764.4763.77**63.77**59.4959.4962.39**62.39**65.02**65.02**58.5258.52
Ground Truth64.5864.5864.3464.3464.8564.8563.3863.3864.7564.7563.2063.2065.2965.2962.1662.16

The results show that our MGE-LDM variants achieve the highest coherence across nearly all stem combination compared to the baselines (MSDM and MSG-LD), often matching or even approaching the ground-truth levels.

Because COCOLA emphasizes musical coherence rather than timbral accuracy to reference stems, we believe it complements sub-FAD and provides a more balanced view of partial generation quality. We will integrate these COCOLA results either alongside sub-FAD in the main manuscript or in the Appendix, as you suggested.

(For reference, the COCOLA score (Harmonic) of stable audio + control net reported by the COCOLA paper [1] is 57.34, though it is not directly comparable to our results since it uses a different evaluation set.)

Once again, thank you for this valuable suggestion.

"While the authors acknowledge that the nonlinearity of the latent space makes it difficult to constrain the sources in a mixture and they opt out for their approach, I believe that the three entries of a triple are not necessarily constrained to sum well in the data (waveform) domain. This could have a negative impact in the performance of the model. Thus, finding the best way to solve compositional music generation in the latent domain it still remains an open problem, and maybe a more constrained approach would be given, for example, by learning latent-domain likelihood functions like in (Postolache et al. 2023). The authors could discuss this in the concluding remarks of the paper or in limitations."

Authors' Response: You rightly point out that, although we train on triplets satisfying mix=submix+sourcemix = submix + source in waveform space, our latent-diffusion model does not enforce an explicit additivity constraint for generated triplets. This issue is critical, as we believe it directly related to the “hallucination” phenomenon, in which the model extracts sources that aren’t actually present in the mixture. Postolache et al. [2] offer an elegant solution by enforcing additivity in a discrete VQ-VAE latent space: they estimate the joint likelihood of two sources by counting exact codebook occurences (bin-counting), effectively modeling p(zmixzsrc1,zsrc2)p(z_{mix}|z_{src_1}, z_{src_2}), where zz are quantized latent codes. This bin-counting approach serves as a powerful regularizer to enforce a summation constraint in the latent domain. Unfortunately, our current pipeline relies on a continuous latent space, which precludes direct application of discrete bin counts.

Nonetheless, we believe adapting this latent-domain likelihood approach to our model would be highly promising. We could either extend it to a continuous latent space by designing appropriate regularizers, or convert our encoder to a discrete representation, such as using VQ-VAE based encoder and discrete diffusion (or MaskGIT) for generation to enable bin counting in the latent space.

We thank the reviewer for this valuable suggestion, and we will discuss these directions in the Limitations and Future Work sections as promising approaches to reduce hallucinations and improve compositional consistency.

We also appreciate your helpful line-by-line suggestions and will update the manuscript accordingly. In particular, your idea of negative conditioning to supress unwanted tracks is very valuable and will be explored in our future work.

References:

[1] Ciranni, Ruben, et al. "Cocola: Coherence-oriented contrastive learning of musical audio representations." ICASSP 2025.

[2] Postolache, Emilian, et al. "Latent autoregressive source separation." AAAI 2023.

评论

Dear Reviewer W8on,

I hope you are doing well. With the author-reviewer discussion period extended until August 8, I wanted to gently follow up on our rebuttal, in which we addressed your insightful questions. We would greatly appreciate any further feedback or clarifications you might have at your convenience. Thank you again for your time and thoughtful review.

评论

I am sorry for my late answer! I deeply appreciate the authors have found time to integrate the COCOLA score in their manuscript. It definitely improves the quality of their submission, achieving such strong results on the metric. Also discussing latent consistency in the manuscript is a great addition. I keep my full accept score!

评论

We are sincerely grateful for the time you devoted to reviewing our rebuttal, for your constructive and insightful feedback, and for your full-accept recommendation. We will incorporate your suggestions into the revised manuscript.

审稿意见
5

The paper introduces MGE-LDM, a latent diffusion model that simultaneously addresses music generation, source imputation, and text-driven source separation without relying on predefined instrument classes. Unlike prior models constrained to fixed stem definitions, MGE-LDM learns a joint distribution over mixtures, submixtures, and individual sources in a compressed latent space. By framing both source extraction and stem completion as conditional inpainting tasks, the model can handle arbitrary instruments and enables language-guided manipulation using CLAP embeddings. It further introduces track-wise adaptive timestep conditioning during training to enhance inpainting fidelity, allowing the model to distinguish between observed and missing components. Empirical evaluation demonstrates competitive performance across synthetic and real-world benchmarks in total generation, partial generation, and source extraction; an accompanying demo website confirms the competitiveness of the approach.

优缺点分析

This paper presents a timely contribution to the field of music generation and source separation by proposing a latent diffusion model that, similarly to recent methods, unifies multiple tasks in a coherent framework. The technical foundation is solid, and empirically the model is clearly capable of high-fidelity generation and source manipulation, with performance that is competitive with or better than strong baselines. One of the key strengths lies in its use of submixtures as an intermediate representation, which elegantly works around the need to know the exact number or type of stems present in a mixture. This design choice improves scalability and robustness, especially when working with heterogeneous real-world datasets where source labels are often noisy or aggregated. The use of CLAP embeddings for conditioning is another good idea. While CLAP is not flawless, its integration here is conceptually compelling. Finally, I appreciated the adaptation of TD-Paint's per-region timestep conditioning to the audio domain, which is both technically interesting and practically useful in improving latent inpainting fidelity.

An aspect I find worthy of praise is the possibility to move beyond the common and somewhat restrictive four-stem modeling (bass, drums, guitar, piano/other) seen in many prior works. By demonstrating the model's ability to handle a wider range of instruments including synths, brass, pipe, and others, it takes an important step toward more realistic and musically rich applications, even if the results for these less common stems are not yet exceptional. This direction is significant because it reflects the diversity of real-world music and shows a commitment to broader applicability.

The paper is generally well-organized, though some sections (e.g. the method) can be dense and may benefit from additional high-level intuition or summarization. Nonetheless, the figures and detailed mathematical exposition help make the approach reproducible. The joint modeling of mixture, submixture, and source in the latent space, combined with conditional inpainting, represents a meaningful departure from previous work that tends to rely on fixed-class assumptions or waveform-domain additivity.

Key weaknesses include limitations in generalization and fidelity for non-canonical instruments, but these are also clearly outlined by the authors. This suggests the model struggles with finer details or less frequent classes, and points to room for improvement in generalization to rare or underrepresented sources. But I don't see this as a critical drawback. Further, as already mentioned, while the method is mathematically rigorous, its clarity suffers from heavy notation and long explanations without always grounding the reader in intuitive takeaways. The presentation could be tightened to emphasize conceptual contributions before diving into formalism.

Overall, I believe the paper meets the bar for NeurIPS, given its improvements upon prior related work published in such venues.

问题

  • Can you quantify or mitigate the limitations of CLAP-based conditioning? CLAP is conceptually appealing for language-guided extraction, but its granularity and semantic specificity are limited. Do you observe consistent issues with ambiguous or overlapping prompts (e.g., "strings" vs. "synth pad")? Have you explored alternative or complementary embeddings, or prompt disambiguation strategies?

  • What are the primary failure modes, and how often do they occur?

局限性

Yes.

最终评判理由

The authors clarified my concerns and included additional results that I explicitly asked for. Having also gone through the other reviews + rebuttal, I can see that several other experiments address all the other major concerns raised by the other reviewers. I think the authors did an excellent job for this rebuttal. Therefore, I confirm my recommendation to accept the paper for presentation at NeurIPS.

格式问题

None.

作者回复

We thank Reviewer ShCW for the thoughtful and constructive feedback and are delighted that you recognize our work as a timely contribution to the field. Below, we address each of your points:

"The paper is generally well-organized, though some sections (e.g. the method) can be dense and may benefit from additional high-level intuition or summarization. / while the method is mathematically rigorous, its clarity suffers from heavy notation and long explanations without always grounding the reader in intuitive takeaways. The presentation could be tightened to emphasize conceptual contributions before diving into formalism."

Authors' response: We agree that the Method section can feel dense. To make it more accessible, we will revise the method section to include a conscise, high-level overview outlining the key ideas in plain language before describing mathematical details.

"Q: Can you quantify or mitigate the limitations of CLAP-based conditioning? CLAP is conceptually appealing for language-guided extraction, but its granularity and semantic specificity are limited. Do you observe consistent issues with ambiguous or overlapping prompts (e.g., "strings" vs. "synth pad")? Have you explored alternative or complementary embeddings, or prompt disambiguation strategies?"

Authors' response: You’re correct that CLAP’s semantic granularity can lead to ambiguity. To illustrate, we ran a small experiment on mixtures containing strings or synth‑pad and measured Log‑Mel L1 distances between the true isolated stems and the model’s output under different text prompts:

Cond. inst. \ Ref inst.stringssynth_pad
strings2.572.12
synth_pad2.252.06

In an ideal setting, the lowest errors would lie on the diagonal, but here we see that asking for “synth pad” on a mixture with strings yields a lower L1 distance (2.25) than asking for “strings” (2.57). Listening confirms that the model often reproduces the correct pitch contour but with a synth‑like timbre, which is a hallucinatory behavior we have noted in our Limitations section. This highlights an important direction for future work in improving prompt‑based source separation (which we also discuss with the reviewer W8on).

We also intend to explore more recent embedding models such as M2D‑CLAP [1] to enhance semantic precision, investigate negative prompting strategies (e.g. “strings, not synth pad”) to explicitly suppress unwanted classes. We also can try leverage CLAP’s audio‑branch embeddings from short reference clips to anchor the model and reduce ambiguity. Although we have not yet quantified these ideas, we believe they could meaningfully improve disambiguation.

"Q: What are the primary failure modes, and how often do they occur?"

Authors' response: Empirically, we observe three primary failure modes. First, the model sometimes fails to extract a source and outputs near‑silence instead of the intended stem. Second, with ambiguous prompts the model often hallucinates sounds that aren’t present in the mixture, as discussed above. This issue seldom arises for clear prompts like “the sound of vocals,” where the model correctly outputs silence when vocals are absent. Third, the model struggles to capture the fine-grained detail of isolated single notes on piano and guitar: while chords and percussive strokes are generally well reproduced, individual notes often vanish or become blurred. We believe these issues stem from a relative scarcity of fine-detail, single-source examples in our training data. To address them, we plan to augment our dataset with additional isolated‑note samples and experiment with increased loss weighting on fine‑detail reconstruction.

Reference

[1] Niizumi, D., et al. “M2D‑CLAP: Masked modeling duo meets CLAP for learning general‑purpose audio‑language representations.” Interspeech 2024.

评论

I thank the authors for the additional clarifications and experiments. I still do recommend acceptance of this paper.

评论

We would like to express our sincere gratitude again for your constructive and insightful feedback. We will incorporate your suggestions into the revised version.

最终决定

The paper proposes MGE-LDM, a latent diffusion model for multi-track music generation, source imputation, and text-conditioned source separation without relying on fixed stem categories. The framework models mixtures, sub-mixtures, and sources jointly in a latent space and frames both extraction and completion as conditional inpainting, with adaptive timestep conditioning to improve fidelity.

All reviewers agree on the significance of the methodological contribution, which allows handling mixtures beyond the standard four-stem pipelines and offers a scalable, stem-agnostic approach. The ability to exploit heterogeneous datasets during training is also an important contribution of this work.

One reviewer remains unconvinced by the experimental evaluation, noting that the proposed method does not always outperform competitors, particularly in human evaluation. This concern was partially addressed by the introduction of the COCOLA metric.

Even if the performance is not always superior, the clear methodological advance, the ability to unify multiple tasks in a scalable latent framework, and the new experiments presented during the rebuttal make this work a clear step forward in this line of research.