PaperHub
6.1
/10
Poster4 位审稿人
最低3最高4标准差0.4
3
3
4
3
ICML 2025

M2PDE: Compositional Generative Multiphysics and Multi-component PDE Simulation

OpenReviewPDF
提交: 2025-01-22更新: 2025-07-24
TL;DR

We introduce a compositional multiphysics and multi-component simulation method by structurally composing the learned energy function of each conditional simulations

摘要

关键词
multiphysicsmulti-componentPDE simulationphysical simulationgenerative

评审与讨论

审稿意见
3

The paper proposes a compositional multiphysics and multicomponent simulation model that uses diffusion. The approach, MultiSimDiff, consists of learning conditional distributions of individual components conditional on the other physical processes. At test-time, the approach samples from the conditional distributions to build a sample from the joint distribution. The paper tests their approach on three tasks: reaction-diffusion, nuclear thermal coupling, and prismatic fuel thermal and mechanical analysis. The paper’s reported contributions are: a novel approach to multi-physics / multi-component modeling by framing in terms of a joint distribution; open-source data sets; demonstrating success for new simulation models.

给作者的问题

In addition to the above comments:

  • When should a multiphysics approach be split into conditional models, and when should you model the full joint distribution?
  • Is there any advantage to the order of the conditional distributions that are sampled from? For example, in Gibbs sampling, the order of sampling from the distributions can sometimes be varied.
  • If the approach is like EM, does that mean the overall log likelihood always increases?
  • How much compute is needed to train each individual conditional diffusion model? It looks like there is a different diffusion model for each component.
  • Line 8 of Alg. 2 runs an additional update step using the function f()f(\cdot). What is the purpose of this function, how costly is it, and why is it not needed for Alg. 1?

论据与证据

Claim 1: “To the best of our knowledge we are the first to introduce a compositional generative approach to Multiphysics and multi-component simulations”

  • Splitting up a problem into its conditionals and sampling from them sequentially has been known for a long time. This is what Gibbs sampling is. The other relevant literature not covered is simulation-based inference. A paper that seems to be tackling a somewhat similar goal in that literature is Gloecker at al. (https://arxiv.org/pdf/2404.09636v1). I would say that potentially the claim could be either grounded in the specific architecture, or in the specific applications shown in the paper.

Claim 2: “We create and open-source benchmark datasets … ”. “The code is available at the anonymous repository”.

  • I might have missed it, but I can’t see a link anywhere to the code or the benchmark datasets. Also, it would potentially be more valuable if it were the simulators as well as the data that were being released. Is that something the authors would consider releasing?

Claim 3: “MultiSimDiff, trained on single components, accurately predicts larger structures with up to 64 components.”

  • I agree with this claim.

方法与评估标准

  • The evaluation across three novel application datasets makes sense for the work.
  • One question is why the validation dataset consists of decoupled data and not coupled data when it is known that the test data is coupled data. This would seem to put the surrogate models at a disadvantage, since they are being optimized on decoupled data but tested on coupled data.

理论论述

  • Appendix E shows that there is a gap between models trained on coupled data vs. decoupled. My main question related to the theoretical claims of the paper is when is it appropriate to split up a problem into learning conditional distributions? For example, for Gibbs sampling, it may not be that easy to make this assumption. The Prismatic fuel element seems to make sense as it enables generalization. The paper seems to imply that it is taken for granted that all the problems can be split into the conditional distributions.

实验设计与分析

  • No addition to what has previously been mentioned.

补充材料

  • I could not find anonymous linked code. I read through the supplementary materials.

与现有文献的关系

  • The paper is focused at multiphysics/multi-component simulations. There has not been a large amount of work focused on these multi-component scenarios. This is in part due to a lack of available simulators/datasets.

遗漏的重要参考文献

  • As mentioned earlier, it seems like the area of simulation-based inference (SBI) is missing.

其他优缺点

Strengths:

  • Paper is well-written, and presentation is good.
  • The application domain is a strength of the paper. The three proposed applications are nice to see and it is not easy to introduce three new examples in a single paper without relying on previous benchmarks.

Weaknesses:

  • The novelty of the approach does not seem especially strong given that splitting into conditional distributions is a common procedure for generative modeling and sampling.

其他意见或建议

  • See below
作者回复

Re1: Relation to Gibbs Sampling.

Gibbs sampling splits a problem into conditional distributions and samples them sequentially. In our setting, this is essentially our baseline (named surrogate model in manuscript), where we iteratively update each physical field using a surrogate model (See Algorithm 3 in the manuscript). Specifically, in a multiphysics problem, if you randomly initialize all physical fields as: z(0)=(z1(0),z2(0),,zd(0))\mathbf{z}^{(0)} = \bigl(z_1^{(0)}, z_2^{(0)}, \ldots, z_d^{(0)}\bigr), then at each iteration kk, you loop over each physical field ziz_i and update it with the conditional probability p(zizi)p(z_i \mid z_{\neq i}). If that conditional probability is replaced by our trained surrogate model, it essentially becomes Algorithm 3.

Re2: Open-Source Dataset/Code.

We verified the anonymous repository is accessible. It contains solver input files and download links for the data, mentioned at line 294. We also provide the url here: https://anonymous.4open.science/r/MultiSimDiff-D5A3/README.md.

Re3: Why the validation dataset consists of decoupled data and not coupled data when it is known that the test data is coupled data...

We believe this is fair for both the surrogate model and the diffusion model. For multiphysics problems, the goal is to train the model on decoupled data but eventually predict the coupled solution. Consequently, the training and validation sets are composed of decoupled data, and the test set is composed of coupled data. Both models use the same training, validation, and test sets.

Re4: When is it appropriate to split up a problem into learning conditional distributions? Can all the problems be split into the conditional distributions?

Our tasks naturally decompose into conditional fields—each physical field or component is determined by the others. In multi-physics, each field typically has its own solver requiring boundary information from other fields. Similarly, in multi-component setups, a component’s solution depends on its neighbors. But there is a type of eigenvalue problem that cannot be solved by current methods, which can be seen in Appendix J.

Re5: Why simulation-based inference (SBI) is missing.

We understand SBI as inferring model parameters from given data. Gloecker et al. trained a single diffusion model for p(x,θ)p(x, \theta), thus achieving their “all in one” model to obtain both the likelihood and posterior. But for our problem, due to the difficulty in modeling the joint distribution of multiple physical fields or components, our approach is to combine individual models to achieve similar functionality. Although these two methods look very similar, their directions of operation seem to be opposite.

Re6: When should a multiphysics approach be split into conditional models, and when should you model the full joint distribution?

Modeling the full joint distribution (simultaneously solving all coupled equations) in a multiphysics system is difficult, especially in fields like nuclear engineering, where each physical process (e.g., neutron, fluid, mechanic) often has its own independently developed solver. Building a fully coupled solver to unify these modules is both development-intensive and computationally expensive. Instead, our approach trains simpler conditional distributions – each field conditioned on the others (decoupled solution) – and then composes these models during inference to approximate the fully coupled solution.

Re7: Does the Sampling Order Matter?

We have found that in multiphysics, the order of updating each field has negligible influence on final results, as shown in https://anonymous.4open.science/r/MultiSimDiff-D5A3/order.md. For multi-component simulation, we update all components simultaneously, making order irrelevant.

Re8: If the approach is like EM, does that mean the overall log likelihood always increases?

Theoretically, yes. The log-likelihood can be written as E(z)logZ-E(z) - \log Z. Ignoring constants, it is E(z)-E(z). Moving in the direction of E\nabla E decreases E(z)E(z) and thereby increases the log-likelihood.

Re9: How much compute is needed to train each individual conditional diffusion model? It looks like there is a different diffusion model for each component.

For multiphysics with d fields, we train d diffusion models. In Experiment 2 we used three (solid, neutron, fluid). For multi-component tasks, we only need one model because each component is analogous. Even in practical engineering, the number of "d" rarely exceeds four. So the training cost is not particularly high.

Re10: Purpose and Cost of f in Alg. 2?

In multi-component setups, we update all components concurrently and need function f to gather neighboring solutions for each component, costing O(n) for n components but no extra neural inference, which is very fast. In Algorithm 1 (multiphysics), each field is updated sequentially, and there is no unified function f to update their inputs.

审稿意见
3

The paper proposed a compositional diffusion model framework to handle multi-physics multicomponent surrogate model for physics systems. It leverages the existing diffusion backbone and demonstrated it effects on three multi-physics/multicomponent PDE systems. The paper claims the contribution are: 1. Introducing the compositional diffusion model for multi-physics and 2. Creating novel benchmark dataset for multi-physics/multi-components surrogate models research. The metrics comparison also indicates the effectiveness of the proposed model compared with baselines.

给作者的问题

I have briefly mentioned my questions in the above sections but I would like to summarize it here.

  1. What is the novelty for your proposed diffusion framework?

  2. Why do you think it is justified to use diffusion model to model the deterministic PDE systems. The challenges for multi physics is how to model the interface accurately while not being too expensive. And the challenges of multi-components lies in learning a scale-invariant features leveraging GNN-like NNs. None of these are related to the strength of diffusion models.

  3. Could yo compare with more baselines. For multi physics, it is better to include more advanced domain-decomposition based surrogate. And for the multicomponents, it is better to compare with existing GNN-based framework that alright shows great potential on doing the spatial extrapolation tasks. It could be related but not limited to the reference I mentioned above.

  4. What are the error distributions respective to spatial and temporal dimensions.

  5. As for the speed up, the numerical simulation cost heavily depends on the convergence criteria set in the interface. Could you report what is your convergence or residual set for numerical simulation. And the mean time, how accurate your diffusion surrogate is in the interface?

论据与证据

The claims that it introduces a effect multi-physics/multicomponents diffusion framework for physics application is supported by the result section. Moreover, some of the datasets are novel but not all of them. e.g, reaction-diffusion dataset is a standard questions researchers have used.

方法与评估标准

I don't think the proposed methods make sense for multiphysics/multicomponents surrogates. Firstly, the diffusion model itself doesn't have novelty. It just leverages the existing pipeline for a new dataset. Secondly, I am not convinced that the diffusion model is a top candidate for modeling multiphysics/multicomponents process. The underlying physics are all deterministic. And there exists extensive literatures based on domain decompositions or GNNS to build deterministic surrogates. Moreover, the success of the multicomponents model is more relied on the self-similar feature of the graphical data and the model that could capture these resolution invariants features, instead of the diffusion model itself. Moreover, the diffusion model is also not used to account for any uncertainty quantifications for the surrogate models in the current context to help explain the sample variety caused by diffusion models. Therefore, I am not convinced by the novelty contribution and the justification of using diffusion model for current applications.

理论论述

I have checked the equations in the main manuscript and they are correct with minor typos to the best of my knowledge.

实验设计与分析

I have checked the soundness and validity of experimental designs and have found issues.

For the multiphysics experiments. I think it is needed to compare with other domain-decomposition based surrogate models to justify the effectiveness of the papers. I also think it is beneficial to show spatial/temporal error plots to give better idea of how the coupling surface works more than a simple number.

For the multicomponents experiments, I think the GNN based benchmarks is not optimized for performance. I have reviewed research papers that leverage GNN to successfully spatially extrapolate to larger domains. I will list the reference below.

补充材料

I reviewed part of Appendix B,E,G,H,I.

与现有文献的关系

I think the contribution lies in trying to extend the diffusion based framework to multi-physics/multi components surrogate modeling, altought I am not convinced by the current script. Domain decomposition and multi components (train small, test big) ideas have been explored in separate works and the authors tried to show that a diffusion model could be a unified work for these tasks.

遗漏的重要参考文献

I think the several essential references about the benchmark needs to be included in the result comparison part. For example, [1] paper demonstrated that GNN itself could work very well on training on 64 by 64 domain and directly generalizes well to 1024 to 1024 domain for physics systems. However, the current paper reports that GNN-based model all performs very poor. Moreover, the author mentioned domain-decomposition based method like [2], but didn't compare with them in the result part.

[1] Fan, Shaoxun, et al. "Accelerate microstructure evolution simulation using graph neural networks with adaptive spatiotemporal resolution." Machine Learning: Science and Technology 5.2 (2024): 025027.

[2] Ranade, R., Hill, C., He, H., Maleki, A., Chang, N., andPathak, J. A composable autoencoder-based iterativealgorithm for accelerating numerical simulations. CoRR,abs/2110.03780, 2021. URL https://arxiv.org/abs/2110.03780.

其他优缺点

Strength is trying to building a unified surrogate model for multiphysics/multicomponents systesm. These applications are time consuming and there is a strong research need to get faster surrogate. Moreover, the paper writing is clear to follow the key idea. The weakness is it seems directly apply the existing pipeline on a new dataset, without fair comparison to the high performance benchmarks.

其他意见或建议

None

作者回复

Re1: What is the novelty for your proposed diffusion framework?

Our study is application-driven rather than aiming to improve diffusion models directly. Our contribution is not diffusion model itself, but the higher-level algorithm on top of diffusion models for multiphysics and multi-component simulation tasks. Specifically, our framework can learn from single-field or small-scale data and generate complete, coupled solutions or full-structure predictions at inference. This cuts development costs for coupled solvers, simplifies large-scale simulations, and extends readily to more complex scenarios.

Re2: Why use diffusion models for deterministic PDE systems?

We believe diffusion models can be used to simulate deterministic PDE systems because a deterministic structure can be considered a Gaussian distribution with a fixed mean and a very small variance—effectively negligible. We provide experimental evidence showing this variance is very small (see link and Re6 in our response to Reviewer WGGT). Moreover, when physical systems involve noise, diffusion models can also help quantify predictive uncertainty. Several studies have applied diffusion models to simulate physical systems, such as in weather prediction [1] and spatiotemporal field forecasting [2]. In our experiments, our baseline is effectively a deterministic surrogate model, which generally performs better on validation sets (decoupled data / small structures). However, our main objective is to do well on the test sets (coupled / large structures), and our experiments confirm that the diffusion model outperforms this deterministic baseline in that scenario. In summary, we believe diffusion models are a viable approach for modeling PDE systems.

[1] Mardani M, et al. Residual corrective diffusion modeling for km-scale atmospheric downscaling. Commun. Earth Environ. 2025.

[2] Li Z, et al. Learning spatiotemporal dynamics with a pretrained generative model. Nat. Mach. Intell. 2024.

Re3: Could you compare with more baselines like domain-decomposition and GNN-based network?

Domain decomposition in numerical simulation splits the computational domain rather than multiple physical fields. Therefore, machine learning approaches derived from domain decomposition are not used to solve multi-physics problems. Hence we use such methods only for multi-component baselines. For multi-component simulations, our manuscript already compares domain decomposition (surrogate), Graph Neural Networks (GIN), and Graph Transformer (SAN). Based on your suggestion, we further added MeshGraphNet for comparison and tuned certain hyperparameters (e.g., hidden layer size, number of message passing layers). The result is shown in: https://anonymous.4open.science/r/MultiSimDiff-D5A3/rebuttul/Table3.png. Indeed, MeshGraphNet is a very strong baseline. We find that on the 16-component dataset used to train MeshGraphNet, it achieves performance significantly better than any of the other models; in the 64-component dataset, our method still performs the best, demonstrating that our method’s strong multi-component generalization capability.

Re4: What are the error distributions respective to spatial and temporal dimensions.

Currently, our predictions are jointly conducted in both time and space. Therefore, theoretically, the errors in space and time should be similar. For multiphysics modeling, the results are shown in the figure at the following link: https://anonymous.4open.science/r/MultiSimDiff-D5A3/rebuttul/Fig5.png. Overall, due to the coupled physical fields, the spatial errors at the interfaces are slightly higher, while the temporal errors remain close. The flow field initially has larger errors, which decrease as the flow stabilizes. For multi-component simulations (see Fig.9 in the original submission), larger errors typically occur at the component boundaries.

Re5: Convergence criteria and accuracy at the interface.

We did not deliberately loosen the convergence criteria to slow down the numerical simulation. We used MOOSE to create our dataset, and the code repository includes the input files used to generate the data. The convergence criteria include an outer nonlinear iteration loop and an inner linear iteration loop:

  • Experiment 2: Nonlinear iteration has a relative error of 1e-8 (default 1e-8) and absolute error of 1e-8 (default 1e-50), with a maximum of 20 iterations (default 50). The linear iteration has a relative tolerance of 1e-5 (default 1e-5) and a maximum of 100 steps (default 10000).
  • Experiment 3: The maximum number of linear iterations is set to 20, and the maximum number of nonlinear iterations is 5, with other settings left as default.

Compared to the program’s default settings, we have actually tried to make the simulation faster wherever possible.

Regarding accuracy, the spatial errors at the interfaces are somewhat higher, as analyzed in Re4.

审稿人评论

Thanks, the rebuttal is reasonable and I modified my score.

作者评论

Thank you for your constructive feedback and for raising our manuscript's score. Your insights have been invaluable in enhancing the quality of our work.

审稿意见
4

This paper proposes MultiSimDiff, a novel compositional generative model for multiphysics and multi-component simulations. The core idea is to use diffusion models to learn energy functions representing the conditional probability distributions of different physical processes or components. During inference, MultiSimDiff reconstructs the joint solution by sampling from these learned distributions. This approach circumvents the need for coupled numerical solvers and allows generalization from decoupled (small structure) training data to coupled (large structure) predictions. The authors validate the proposed method in several tasks.

给作者的问题

see other parts

论据与证据

Yes

方法与评估标准

Yes.

理论论述

No theoretical claims.

实验设计与分析

Yes, the paper provides ablation studies on diffusion parameters and crucial hyper-paprameters, but a more systematic hyperparameter study (e.g., different architectures, sampling steps) would be encouraged.

补充材料

Yes, all parts.

与现有文献的关系

This work provides valuable insights and new ways of using generating models to handle real-world physical problems.

遗漏的重要参考文献

No.

其他优缺点

Strength:

  • The paper frames multiphysics and multi-component simulation as a generative modeling problem, leveraging diffusion models to compose energy-based representations. This is a pretty fresh and novel perspective on simulation-based learning.

  • The setting of experiment, especially the prismatic fuel element part, is solid and impressive, which demonstrates strong generalization from small to large structures, which is crucial for real-world engineering applications.

Weaknesses:

  • The presentation of the method still could be improved. There are some typos and inconsistent of notations, and more preliminary on the training part could help readers to better understand the work. For example:

    • the line 145, 197, right column, "z=(z1,z2,...,zn)" and "V=v1∪v2∪...∪vn", should use capital "N" instead of "n" ?

    • The statement of z_{i}^{e} in eq.9 is not clear enough. For readers who are familiar with diffusion, it's smooth to get the ituition of " estimate z_{i}^{0} when we at still at z_{i}^{s} " there. However, for readers who are not very familiar with diffusion, the mixure usage of z_{i}^{s} and z_{i}^{e} could lead to confusion.

    • In the Algorithm1 table, the inner loop (step 8-11) is set over the "i", but before that, the notation of "i" is already used in step 1-7. My understanding is the operations invloved "i" in step 1-7 should be applied over i=1,2..N. Authors should make it clear.

其他意见或建议

While the compositional approach is compelling, the choice of conditioning structure is manually designed and attached to single task. I'm curious whether this method can self-adapt to different partitioning schemes for new physical systems, and whether the learned decoupled component (represnt. by diffusion model) has the generalizability to other task.

作者回复

Re1: the line 145, 197, right column, "z=(z1,z2,...,zn)" and "V=v1∪v2∪...∪vn", should use capital "N" instead of "n" ?

Answer: Thank you for pointing this out. We wll make the correction.

Re2: The statement of z_{i}^{e} in eq.9 is not clear enough. For readers who are familiar with diffusion, it's smooth to get the intuition of " estimate z_{i}^{0} when we at still at z_{i}^{s} " there. However, for readers who are not very familiar with diffusion, the mixture usage of z_{i}^{s} and z_{i}^{e} could lead to confusion.

Answer: Thank you for your suggestion. We will add an explanation of the superscript “e” in the paper to indicate that it is an estimated value, distinguishing it from the subscripts ii and ss.

Re3: In the Algorithm1 table, the inner loop (step 8-11) is set over the "i", but before that, the notation of "i" is already used in step 1-7. My understanding is the operations involved "i" in step 1-7 should be applied over i=1,2..N. Authors should make it clear.

Answer: Thank you for pointing this out. Indeed, that is the meaning we intended to convey. We will add a statement clarifying that the operations in steps 1–7 are applied to all i=1,,Ni=1,\ldots,N. You can see in: https://anonymous.4open.science/r/MultiSimDiff-D5A3/rebuttul/ALG1.png

Re4: While the compositional approach is compelling, the choice of conditioning structure is manually designed and attached to single task. I'm curious whether this method can self-adapt to different partitioning schemes for new physical systems, and whether the learned decoupled component (represent. by diffusion model) has the generalizability to other task.

Answer: Your understanding is correct. For multi-physics problems, the choice of conditional structure does indeed depend on the specific task, because the conditional diffusion model needs the other physical fields in order to predict the current physical field. As such, the set of physical fields in the system is already fixed. If a new physical field is added or an existing one is removed, any physical field model that is coupled to it must be retrained. This is indeed a very interesting open question, and we will look into it in future work. For multi-component simulation, the current method can be extended to various combinations of components.

审稿意见
3

This paper introduces MultiSimDiff a new method for solving multi-physics/multi-compnennts simulations efficiently by learning the conditional score of each component's solution given its parameters and solutions of other components. Experiments demonstrate that MultiSimDiff outperforms largely a simpler surrogate that model each component's solution independently on a class of coupled cases.

给作者的问题

See my comments above.

论据与证据

Overall the paper is well-written and claims are supported by evidence.

方法与评估标准

Yes, the idea of building a global solutions by representing each solution as a distribution conditioned on other components' solution appears original and well motivated.

理论论述

There is not strict theoretical claims.

实验设计与分析

Experiments are sound although I would have appreciated if other alternative strategies such as the ones discussed in the related works section would have been used as additional baselines.

补充材料

No.

与现有文献的关系

In my opinion, the paper provides an insightful related work section which position this work well into the literature.

遗漏的重要参考文献

其他优缺点

Overall I very much enjoyed reading the paper and I believe the problem statement is very sound as well as the proposed solution. That said, I am not very knowledgeable in the field of compositional mulitphysics/multi-component solvers and related works using ML for that sake. For that reason and for the few remarks below, I will only weakly support the acceptance of the paper but might increase my score in light of the other reviews and response from the author. Here is a list of remarks/questions (ordered as noted when reading the manuscript):

  • The end of the abstract (3 last sentences) appears a bit narrow and verbose. I am not sure it really helps the reader position your work.
  • The intro is very clear
  • Would be nice to mention better the gain in computation time you may have and exactly when using your method will be beneficial, what are the key ingredients to train it. For instance, it is not totally clear to me as whether you need a large datasets of coupled simulations or not for the method to work well. I understand this consideration is problem dependent. For some problems it may be easy and meaningful to simply solve independently each component conditioned on a plausible state for the other components that are part of the Markov blanket but I imagine that in many cases finding these "plausible" sets for the Markov blanket really requires to have jointly solved this Markov blanket... For Multi-components I understand this may still make your method a nice to scale to a larger number of components, however for multi-physics it seems that you cannot to very much.
  • To emphasise my point above, I think you should discuss more clearly the issue of training the conditional scores and in particular that for tightly coupled systems having these sampled from pip_{\neq i} is not straightforward.
  • I am a bit confused by the baselines you used. You discussed in the relate work that you compare to GNN and Graph Transformer but I did not see these results in the tables. Similar remarks for CoAE-MLSim.
  • It seems that what 1 discusses in 3.1 regarding Markov Blanket and using diffusion model to jointly sample from dependent components is very close to your work from a technical standpoint.
  • Figure 1. is pixelized and not very clear (or at least not informative compared to the space it takes), clean...
  • By definition you algorithm will output a distribution over plausible solutions, it would interesting to discuss a bit better what exactly is modelled by that uncertainty and how you get rid of it in your numerical experiments. Also, how much uncertainty increases, decreases as a function of the exact setup considered.

其他意见或建议

作者回复

Re1: The end of the abstract (3 last sentences) appears a bit narrow and verbose.

Thank you for the suggestion! We will revise the ending of the abstract as follows:

We demonstrate the effectiveness of MultiSimDiff through two multiphysics tasks—reaction-diffusion and nuclear thermal coupling—where it achieves more accurate predictions than surrogate models in challenging scenarios. We then apply it to prismatic fuel element simulation, an exemplary multi-component problem, where MultiSimDiff successfully extrapolates from single-component training to a 64-component structure and outperforms existing domain-decomposition and graph-based approaches.

Re2: Would be nice to mention better ... sampled from pzip_{\neq z_i} is not straightforward.

I gather that this section focuses on two issues: (1) Under which scenarios does our algorithm exhibit an advantage? The key factor to train it. (2) For multiphysics problems, it is not easy to find a suitable pip_{\neq i}, or in other words, you believe that finding a suitable pzip_{\neq z_i} is almost equivalent to solving the problem itself.

For (1), our algorithm aims to solve multiphysics and multi-component simulation problems that can be numerically challenging. In the more complex experiments 2 and 3, we observed up to 29× and 41× speedups, with higher accuracy than other methods. In real engineering applications with large-scale coupling, these gains can be substantial. The key factor is to view multi physics/component simulation problems from the perspective of probability, and replace complex joint probabilities with easily obtainable conditional probability.

For (2), indeed, for multiphysics simulation, find pzip_{\neq z_i} that perfectly respects all coupling is nearly as hard as solving the entire problem. But we do not solve the coupled physical field equations, we currently use a pre-iteration technique (lines 762–772) to approximate it; however, Figure 12 shows a notable gap from the true coupled data distribution. Hence, the accuracy of predicting the coupled solutions of physical fields still needs improvement, we discuss this limitation in Section 5 and plan further improvements.

Re3: I am a bit confused by the baselines you used... Similar remarks for CoAE-MLSim.

For multi-component simulation, we use CoAE-MLSim as a baseline (implemented per the paper’s description) since it is not open-sourced. This corresponds to the “surrogate model” in Table 3. The Related Work section also discusses GNN (GIN) and Graph Transformer (SAN) approaches, which also can be seen in Table 3. We have also added MeshGraphNet for comparison, as noted in our response to Review QAdZ (Re3). However, these methods are not applicable to multiphysics simulations. The baseline for multiphysics simulations is shown in alg.3 in manuscript.

Re4: It seems that what 1 discusses in 3.1 regarding Markov Blanket and using diffusion model to jointly sample from dependent components is very close to your work from a technical standpoint.

Their method replaces the log-likelihood gradient at one time step in a long sequence with a smaller subsequence. We instead replace the gradient of a complex joint distribution (for each physical field or component) with a tractable conditional distribution. For multiphysics problems, some fields typically depend on all other fields, so their subsequence substitution is not valid. For multi-component problems, the posterior estimation step (their Algorithm 3) introduces further challenges. In their scenario, the known information is observational data; in ours, it is the entire physical system’s inputs, which would constrain the model to the same scale as the training configuration. Overall, the core difference is that we must be able to generalize to tasks beyond the training distribution, while their method focuses on the same domain for training and inference.

Re5: Figure 1. is pixelized and not very clear (or at least not informative compared to the space it takes), clean...

We have replaced Fig.1 with a clearer version at: https://anonymous.4open.science/r/MultiSimDiff-D5A3/rebuttul/schematic.pdf

Re6: By definition you algorithm will output a distribution over plausible solutions, ... , decreases as a function of the exact setup considered.

We think that you are referring to uncertainty quantification. In general, there is model uncertainty and data uncertainty. In our current experiments, we learn from simulation data, which do not contain noise. Ideally, if a physical process is deterministic, the predictive uncertainty should be close to zero. We conducted uncertainty quantification experiments in both Experiment 2 and 3, and found that the standard deviation of the model predictions is extremely small, indicating that the diffusion model can approximate a deterministic physical process. We have drawn a table of the results: https://anonymous.4open.science/r/MultiSimDiff-D5A3/rebuttul/std.md.

最终决定

This paper introduces a diffusion-based approach for multiphysics and multi-component simulations. The reviewers found this work valuable for several reasons. A lack of available datasets has so far limited research into data-driven approaches to such simulations, which previously largely required expensive numerical simulations with tight coupling. The authors introduce a set of new datasets and demonstrate that their approach which is trained on decoupled data shows strong generalization from small to large structures. The reviewers overall found the paper well-written except for some confusing notation which was addressed by the authors during the rebuttal.