PaperHub
6.3
/10
Poster4 位审稿人
最低6最高7标准差0.4
6
6
7
6
3.5
置信度
正确性3.5
贡献度3.0
表达2.5
NeurIPS 2024

PACE: Pacing Operator Learning to Accurate Optical Field Simulation for Complicated Photonic Devices

OpenReviewPDF
提交: 2024-04-24更新: 2024-11-06

摘要

关键词
AI for ScieneceOptical simulationNeural opeartorAI for PDE

评审与讨论

审稿意见
6

This paper proposes a neural operator (PACE) with improved accuracy, efficiency, and speed for electromagnetic field simulation. PACE is a network block in the surrogate machine learning model.

优点

This paper conducted extensive experimental evaluations for the proposed method. Also, the accuracy and efficiency of machine learning surrogate models for physical simulation are of great importance, therefore I think even though a little improvement is valuable.

缺点

This paper, including the writing and the experiments, is very similar to the main comparison NeurOLight, which has code available online. Therefore I think it is important to discuss if it is only an incremental work or in-depth research.

问题

  1. How does this paper reflect “complicated” photonic devices, especially when compared with NeurOLight? I found that the datasets of both papers are custom-generated with random configurations and have a similar data volume. I expect the authors to provide more detailed explanations about the differences, particularly regarding the use of the term “complicated”, to avoid misleading readers into interpreting this paper as a simple incremental study.

  2. This paper mentions that the NeurOLight method fails for “real-world” photonic devices, but I did not see any evaluation experiments of the proposed method on real-world datasets in this paper.

  3. Is the speed improvement compared against CPU algorithms (line 308)? If the proposed method uses a GPU while the baseline Angler algorithm uses only a CPU, I do not think this is a fair comparison, and the corresponding 577x time boost could be misleading.

  4. Please avoid using exaggerated language, for example: “boost the prediction fidelity to an unprecedented level” (line 12), and “extremely challenging cases” (line 57).

  5. Unexplained descriptions, for example: “inspired by human learning” (line 57). There is no explanation provided as to why this method is inspired by human learning.

  6. Some of the paper copy-paste from the NeurOLight paper, for example, equations 1 and 2. Corresponding references and explanations should be added.

  7. Figure 4 is difficult to understand, and the caption is too brief.

  8. In Table 1, the phrase “Improvement over all baselines (geo-means)” is rarely used and may not be considered as a good evaluation in research.

局限性

Limitations are discussed in this paper but some points may be missed. For example, this paper works on only simulated and customized datasets.

作者回复

Thank you so much for your time in reviewing our paper and providing valuable feedback on our work! Please see our responses below:

Q.1 Incremental work over NeurOLight with little improvement?

We would like to emphasize that our work is not simply incremental over NeurOLight but offers novel machine learning contributions and a substantial accuracy improvement over NeurOLight.

First, our work aims to tackle the complicated photonic device simulation problem where NeurOLight shows unsatisfactory accuracy.

  • Please refer to Common Response Q. 2 for a detailed explanation of real-world and complicated devices.
  • TL; DR. Our work successfully established a new strong SOTA accuracy on complicated photonic devices with >39% lower error compared to the strongest NeurOLight.

Second, we have clear and unique machine-learning contribution:

  • Novel neural operator design: We propose a parameter-efficient cross-axis factorized operator design with multiple key components to unlock superior accuracy by deeply analyzing the challenges in complicated photonic device simulation in Sec. 3.1.
  • Novel cascaded learning flow: We divide the simulation task into two progressively latent tasks, reducing learning complexity and successfully showing highly competitive accuracy on hard cases.

Q.2 Real-world dataset

Please refer to Common Response Q.2 for a detailed description of real-world.

In summary, the device we use has a real-world geometry, permittivity distribution, light source, and wavelength range. We randomly generate the device configurations to obtain sufficient data to learn the underlying rules in the photonic Maxwell PDE. This randomness also ensures thorough testing of our model’s effectiveness in predicting the optical field, not just fitting some specific examples.

Our claim is that NeurOLight fails for real-world complicated photonic devices with complex light-matter interactions, such as scattering and resonance, as reported in their original paper. On the dataset where NeurOLight fails (Etched MMI 3x3), PACE achieves 39% lower error compared to NeurOLight.

Q.3 Speed-up comparison

We focus on 2-D device-level finite difference frequency domain (FDFD) simulation. The numerical solver uses a CPU-based direct sparse linear solver (not the iterative solver), which is faster than the GPU version, as acknowledged in SPINS-B [1]. In 3D cases, the numerical solver uses an iterative solver where the GPU version is faster than the CPU version. Note that SPINS-B derives its conclusion using the slower scipy. However, in this work, we compare with the highly optimized Intel Pardiso solver on a 20-core CPU (NeurOlight only compares with slow scipy). We already provide a strong baseline and ensure a fair comparison when evaluating speedup.

We agree that 577x is somehow misleading. We will separately discuss the speedup over scipy and pardiso. We will change our wording to “In terms of runtime, our PACE model demonstrates 154-577x and 11.8-12x simulation speedup over numerical solver using scipy or highly-optimized pardiso solver, respectively”

Q.4 Unexplained human learning

Please check our common response Q.3 to see how we are inspired by human learning in cognition.

Q.5 Similar equation and notation to NeurOLight

For the Maxwell PDE description in Preliminary 2.2, as it is a common Maxwell equation, we honor the original notation. For the problem setup section, as we stated and cited in Line 104, we follow NeurOLight in formulating the optical device simulation as an operator learning problem. Therefore, we intentionally use the same notation and symbols to avoid confusion and ensure clarity.

We will adjust the wording and add more specific citations to prevent any potential misunderstanding.

Q.6 Why use Geo-means

To compare different benchmarks, we first calculate the average improvement over all baseline methods on the same benchmark, then use the geometric mean (geo-mean) to assess the effectiveness of PACE across different benchmarks. The geo-mean fairly evaluates results across benchmarks, avoiding domination by large improvements that can skew the arithmetic mean. For example, for values 0.1, 0.5, and 0.9, the geo-mean is ~0.35, while the arithmetic mean is 0.5.

We will add this justification in the revised manuscript.

Q.7 Writing and unclear Fig. 4

Thank you so much for pointing out places that can further improve our paper’s quality. Due to we cannot directly upload the revised version now, we will carefully modify the language in the final version to avoid exaggeration.

Fig. 4's caption will be updated to, “The proposed cascaded learning flow with two stages. The first stage learns an initial and rough solution, followed by the second stage to revise it further. A cross-stage distillation path is used to transfer the learned knowledge from the first stage to the second stage.” We will also modify Fig.4 to make it easier to understand.

Q.8 Limitation for customized dataset

Given that no complete and open-sourced dataset is available for the photonic Maxwell PDE problem, we follow NeurOLight to generate a dataset with real-world device geometry by randomly varying device configurations, which is a common way to generate PDE benchmarks. Our devices reflect the characteristics of real-world devices and are sufficiently complex with complex light-matter interaction inside.

Moreover, aware of the lack of an open-source dataset, we open-source the dataset to facilitate the community.

Reference

[1] SPINS-B (as we cannot provide the direct link in rebuttal, you can check the documentation of SPINS-B and read the claim in the faq section)

评论

Dear Reviewer,

Thank you very much for your thorough review and valuable comments on our paper. We have carefully considered all your feedback and have addressed your concerns and questions in our rebuttal. Especially regarding your question on whether we are doing incremental work over NeurOLight, we would like to emphasize that our work proposes a new neural operator backbone and a new learning strategy for complicated photonic device simulation problem, and we achieve > 39% lower error than NeurOLight (a significant improvement).

We would greatly appreciate it if you could take some time to review our responses. We will be online over the next few days and are happy to address any further questions or concerns you may have.

评论

Thanks a lot for your rebuttal.

审稿意见
6

This paper propose a novel cross-axis factorized PACE operator with a strong long-distance modeling capacity to connect the full-domain complex field pattern with local device structures. Additionally, the authors proposed a two-stage model learning method for extremely hard cases, with a first-stage model learning an initial solution refined by a second model. Finally, experimental results show that the proposed PACE model can achieve good performance with fewer parameters.

优点

  1. Major challenges for simulating real-world complicated optical devices is well analyzed.

  2. Their proposed PACE model significantly boost prediction fidelity to an unprecedented level with much faster running speed than traditional numerical solver on a 20-core CPU.

  3. A two-stage model learning method is used for extremely challenging cases.

  4. Experimental results show that on the complicated device benchmarks, their proposed PACE model significantly outperforms previous baseline models.

  5. A dataset on for the complicated optical device is released to facilitate AI for PDE community.

缺点

  1. I do not agree with the authors that the proposed two-stage learning method belongs to a divide-and-conquer approach. From my perspective, it is called coarse to fine.

  2. The writing should be improved as some sentences are not precise. For example, at line 19&20, “one sole PACE model …fewer parameters” (compare with which method?)

  3. It is common that deep learning based method runs much faster than traditional numerical solvers. The testing error of the traditional numerical solvers may plays as a groundtruth for the upper bound of precision. Thus, the authors should list the testing error of the traditional numerical solvers.

  4. In Table 1, except for the number of Params, you should list other metrics like Flops and testing time to show the efficiency performance of the proposed PACE model.

问题

  1. Training details in Sec. 5.1 and A.2 should be integrated.

  2. The effectiveness of pre-normalization and double skip is not well verified. In Table 6 of the appendix, you should list the results of PACE-12 layer without double skip and pre-norm.

  3. At line 211, the authors claim that non-linear activation is known to help generate high-frequency features. Therefore, some visualization of the high-frequency features is needed to support this statement.

局限性

NA

作者回复

Thank you so much for your appreciation of our contributions and for providing valuable feedback on our work! Please see our responses below:

Q.1 Use “coarse to fine” instead of “divide and conquer”

Thank you for your suggestion. We agree that our learning flow could be more precisely described as learning the PDE solution in a coarse-to-fine manner, given that we divide the challenging cases into two progressively latent tasks and generate high-fidelity PDE solutions from rough to clear. We will update our description.

Q.2 Testing errors of traditional numerical solver

Since the true optical field is not measurable, photonic designers typically rely on fine-grained numerical simulations to obtain the optical field. The simulated optical field is treated as a reliable reference (ground truth) to guide the photonic design optimization.

In this paper, we set the simulation granularity to 0.05nm, sufficient to obtain a reliable referenced optical field. We treat this fine-grained simulation as the ground truth (assuming 0 error or target) and test the error of the predicted field from ML models against it. Therefore, we don’t list the tested error of numerical solvers.

Q.3 Include run-time in Tab. 1

We report the run-time for etched MMI3x3 (same for etched MMI5x5) here for your reference, and the final Table 1 will be updated. We tested the run-time on a single A100 GPU using torch.utils.benchmark, averaging multiple runs. Models were compiled with torch.compile, except for TFNO, which is not compatible with torch.compile.

Our PACE shows excellent accuracy, superior parameter efficiency, and better run-time compared to the previous SOTA NeurOLight.

We’d like to emphasize that accuracy is the highest priority metric when using an ML surrogate model to guide the photonic device design and optimization loop. It is worthwhile to sacrifice speed for much better fidelity. As shown in the table, although several models show faster inference speed, they fail to deliver good accuracy. And model like FNO shows saturated model accuracy when further scaling to deeper models[1].

Moreover, PACE can be further accelerated with customized acceleration (e.g., grouping operation) and a better 1-D Fourier kernel. Even with the current implementation, PACE shows a much-accelerated runtime with a 12x speedup compared to the highly optimized Pardiso-based numerical solver. Therefore, PACE is highly effective in serving as an accurate and fast ML surrogate to accelerate the photonic design loop.

Parameter count is also a key metric for our problem. The presence of high-frequency features necessitates the use of high-frequency modes, which is known to greatly increase the parameter count in FNO and cause overfitting, motivating the development of factorized neural operator work (e.g., Factorized FNO, Tensorized FNO). As a novel neural operator, we utilize a physically meaningful cross-axis factorization and outperform previous factorized FNO work with a significant accuracy margin, while maintaining comparable speed with Factorized FNO.

Model#params (M)Test error (1e-2)Test time(±\pmσ) (ms)
UNet3.8863.031.79(±\pm0.02)
Dil-ResNet4.1751.345.89(±\pm0.06)
Attention-based module3.7570.0510.41(±\pm0.01)
UNO4.3834.225.34(±\pm0.02)
Latent-spectral method4.8155.073.84(±\pm0.06)
FNO-2d3.9932.514.24(±\pm0.03)
TFNO-2d2.2535.524.92(±\pm0.01)
Factorized FNO-2d4.0224.29.93(±\pm0.03)
NeurOLight2.1115.5812.98(±\pm0.14)
PACE1.719.5111.44(±\pm0.03)

Q.4 Incomplete ablation study on pre-normalization and double skip:

We include the training results of the PACE-12-layer model without double skip and pre-normalization. The training error, test error, and parameter counts are listed as follows:

ModelDouble skip & Pre-normalizationParamsTrain errorTest error
NeurOLight-16layerw/o210825815.5817.21
NeurOLight-16layerw/211030615.2615.87
PACE-12layerw/o170902610.3211.06
PACE-12layerw17105629.5110.59

As stated in the paper, double skip and pre-normalization are not the primary sources of accuracy improvement but are beneficial for stabilizing the model when scaling to deeper layers.

Q.5 Visualization of feature map before/after non-linear activation in our explicitly designed high-frequency projection path

We visualize the first 6 channels of feature maps before and after the nonlinear activation in the last PACE layer by showing them in the frequency domain. As shown in reb-Fig. 1, the nonlinear activation can ignite high-frequency features, which confirms our claim and validates our design choice of injecting an extra high-frequency projection path. Furthermore, in Sec. 5.3, we show that removing the projection path results in a considerable accuracy loss.

Q.6 Writing

Thank you so much for pointing out that some parts of our description are not precise, which is of great importance to improve our paper quality further. We will modify the wording in the final version, e.g., one sole PACE model is compared with various recent ML for PDE solvers.

We agree that the training details in Sec. 5.1 and A.2 should be integrated, and we will do it in the revised version with one more page budget.

[1] Tran, Alasdair, et al. "Factorized fourier neural operators."ICLR 2023.

评论

Dear Reviewer,

Thank you very much for your thorough review and valuable comments on our paper. We have carefully considered all your feedback and have addressed your concerns and questions in our rebuttal.

We would greatly appreciate it if you could take some time to review our responses. We will be online over the next few days and are happy to address any further questions or concerns you may have.

Thank you again for your time and consideration!

评论

Dear Reviewer 9WdN,

We sincerely appreciate your valuable comments and suggestions for our paper.

We have carefully addressed all your comments in our response above. Regarding your questions,

  • We have added an experiment for the ablation study to test the effectiveness of pre-normalization and double skip. The results confirm that pre-normalization and double skip are not the main sources of improvement. Our proposed new neural operator block with many essential model design considerations, such as cross-axis factorized operator and high-frequency projection path, is the key to unlocking the improvement.
  • We have provided extra visualization of the feature map before/after nonlinear activation in Rebuttal Figure 1. It validates our model design choice by clearly showing that the high-frequency feature is generated after the nonlinear.
  • Runtime is provided as you suggested, and the tab. 1 will be updated in the revised version. (please check our response Q3)

As the discussion period is drawing to a close in 1 day, we would be very grateful if you could take a moment to review our responses. If our replies have satisfactorily addressed your concerns, we would greatly appreciate it if you could acknowledge this and consider raising the rating.

If you have any remaining questions or concerns, please do not hesitate to reach out to us. We look forward to your feedback.

评论

Dear Reviewer 9WdN,

Thank you once again for your review. As the deadline for the author-reviewer discussion approaches, we noticed that we haven't received any comment from you.

We have addressed all your questions with additional experiments and clarifications:

  • We have added an experiment for the ablation study to test the effectiveness of pre-normalization and double skip. The results confirm that pre-normalization and double skip are not the main sources of improvement. Our proposed new neural operator block with our carefully designed model components, such as cross-axis factorized operator and high-frequency projection path, is the key to unlocking the improvement.
  • We have provided extra visualization of the feature map before/after nonlinear activation in Rebuttal Figure 1. It validates our model design choice by clearly showing that the high-frequency feature is generated after the nonlinear.
  • We provide the runtime (please check our response Q3) and the tab. 1 will be updated in the revised version.

As the discussion period is coming to an end, we would greatly appreciate any additional feedback you might have. If our responses have clarified your understanding of our paper, we sincerely hope you might consider raising the rating.

Thank you again for your effort in reviewing our paper.

Best regards, Authors of Paper 472

审稿意见
7

The authors propose PACE operators, which perform much better than existing methods on predicting optical fields for real-world complicated optical devices. The work builds up on and is compared to the NeurOLight framework which represents the SOTA in this field. The benefits of physics-agnostic PACE operators stem from their cross-axis factorized structure. This enables modeling of complicated devices in a parameter-efficient manner. The authors also propose a two-stage learning process which further improves the performance of PACE operators. Finally, the authors open-source this dataset consisting of optical field simulations for complicated devices. The authors demonstrate strong improvements over SOTA in solving the Maxwell PDE for complicated devices through novel operator design and learning workflows.

优点

Strengths:

  1. The paper is well-structured with clear background and motivation. The authors highlight the issues that make optical field prediction for complicated devices challenging for existing methods.
  2. The authors introduce a novel PACE operator that effectively utilizes the identified challenges to mitigate the difficulties in optical field prediction for complex devices.
  3. The authors demonstrate the strong performance of PACE operators compared to the existing NeurOLight framework as well as other commonly used ML PDE solvers such UNets, FNO variants and attention based methods. The proposed method achieves significantly lower errors with comparable or lower number of parameters, while also being faster than numerical methods.
  4. The authors also validate all the proposed improvements individually with well-designed ablation studies.
  5. The curation of the dataset for optical field simulations of complicated devices will also be helpful to the community for future benchmarking.

Overall the paper serves as a good example of identifying a clear problem in a specific domain,  thoroughly analyzing the shortcomings of existing methods, and leveraging domain knowledge to enhance neural operators. The results are strong and showcase the applicability of this method for optical device simulation.

缺点

  1. While demonstrating strong results on a challenging problem, my main concern for this paper is that it might be too specific of an application for a broader ML conference like NeurIPS. The paper might be better suited to a photonics journal. This can be mitigated by demonstrating results on other standard PDE benchmarks and showing improvements over existing methods. The authors allude to this in line 362 Conclusion but concrete results would greatly improve the paper.
  2. The interpretation of errors is not clear. First the normalized absolute error (relative error) of NeurOLight is repeatedly mentioned. Then the authors use the normalized mean squared error as the learning objective (which is rightly justified in A.6). However it is not clear what these errors mean. The paper would also be clearer if it is explicitly stated what error metrics are used in the various table (I assume it is N-MSE). What is the tolerance for acceptable errors for downstream tasks? What magnitude of error makes it comparable to existing numerical solvers used in practice?
  3. It is common in ML-PDE papers to show speed ups over numerical methods while the ML models have less accuracy. If you were to achieve this same lower accuracy using the numerical methods, it might greatly reduce the runtime of numerical methods as well. What is the runtime of Angler/scipy/pardiso that achieves comparable accuracy to the PACE operator?
  4. The authors mention multiple times about inspiration from "human learning". However there is no description of what this means. Section 4.2 on cascaded learning can be improved by expanding on how this related to human learning with appropriate citations.
  5. It has been mentioned that the model preserves the spectrum of the solution field. The author show the spectrum of the reference field in Figure 1d and Figure 9 (for Darcy Flow). However there is no discussion on the spectrum of the predicted solution fields. The authors should visualize the spectrum of predicted fields and compare it to the reference.

The paper would benefit from a thorough proof-reading. There are multiple typographical errors and some sentences are not clear. Examples are:

  • Line 5: space after comma ", NeurOLight,"
  • Line 162: satisfactory instead of satisfied
  • Line 188: shortcomings instead of shortness
  • Line 196: O(nlogn)
  • Line 205: "is a must" is extra/
  • Line 224: satisfactory instead of satisfying
  • Line 238: What does condensate mean here?
  • Line 252: "neural"
  • Line 255: Establish LSM acronym here
  • Line 269: I think the authors mean to say that there is a 73.85% reduction in error. The way this sentence is phrased right now implies that there is a 73.85% error.
  • Line 284: Fourier
  • Line 294: Sorely seems to not fit here. Maybe "just" or "only"?
  • Line 297: Appendix
  • Lines 314 and 315: It is a bit confusing. The sentence can be made clearer by avoiding the use of worse.
  • Line 319: Decomposition doesn't seem to fit here.
  • Line 336: Factorized
  • Line 362: Restricted might be a better alternative to dedicated.

Some figures are hard to decipher:

  • Figure 1 a,b,c should be made larger so that the readers can clearly see the issues mentioned in the Challenges section.
  • Figures 5, 6, 7 show important comparisons between different methods and should be made larger.
  • Figure 9 should have subfigures. Also 9a (Darcy flow field) should have axis labels and colorbar.
  • Figures 10 and 11 should be made bigger so that the light fields and errors are clearly visible. The absolute error is not closed with "|" in the blue boxes in Figure 10. Figure 11 is missing colorbars that show the magnitudes of the optical fields and errors.

Right now the dataset URL is behind a hyperref "URL". It might be helpful to mention the full address somewhere in the paper for accessibility.

问题

  1. The etymology of "PACE" is not clear. The all-caps spelling suggests it is an acronym but it is not expanded anywhere in the paper. If there is another contextual meaning of the word "pace" that is applicable to the model, it is not clear from the paper. Why is it called PACE?
  2. The works cited in the paper (reference 25, 12) seem to use the word "complex" to describe non-trivial optical devices. Would this be a better word to use than "complicated" used in the title and throughout the paper?
  3. In "Generalization to ood testing" (line 322), the results are only shown for interpolation, i.e., for wavelength ranges within the training domain. It is great that the model interpolates well in this range. However, for true ood generalization, the authors should also show results for extrapolation. How does the model behave for wavelengths outside the training range (outside 1.53-1.565 micrometers).
  4. Related to point of weaknesses, what is the contribution of using the different N-MSE metric as the learning objective? Does the performance of NeurOLight also increase when trained using N-MSE?
  5. In section 5.2.3 Speedup over numerical tools: FNOs are resolution invariant by design. Does PACE also exhibit this property? In the speed up section, were multiple PACE models trained for each discretization or are the results shown for one model with inference over different discretization?
  6. The authors qualitatively state in section A.9 that the prediction results show "near-black error map" (line 582). However it seems that the colorbars have different scales across different models. There are no colorbars in Figure 11. Also, since in cases like these with very low errors, it would be interesting to see the absolute errors in log space. This might shed further light into structural sources of errors. How would this look in log space?

局限性

The authors have a brief limitations section where they state that exploring operator learning on FDTD methods and general optimizations to FFT kernels on GPUs are interesting future directions.

While this is already a parameter efficient model, it is a known limitation of Neural Operators that they don't scale well with increasing dimensions. It might be helpful to address this limitation here. How would the PACE operator scale in 3D and beyond?

I don't think that there are any negative societal implications for this work.

作者回复

Thank you so much for your valuable feedback on our work! Please see our responses below:

Q.1 Narrow audience and other benchmarks

We'd like to emphasize that our paper aligns with NeurIPS’s interests and offers novel ML contributions to the AI for PDE community (check common response Q.1)

The standard PDE benchmarks typically use smoothly changing coefficients and, therefore, don’t share the same PDE characteristics with ours (e.g., our permittivity is highly discrete and contrastive), making them unsuitable for evaluating our method.

We notice that our problem shares a similar pattern with multi-scale PDE, where PDE coefficients are generated contrastively. Recent works [1][2] also show the failure of existing AI-based PDE solvers in multi-scale PDE. Unfortunately, [1][2] do not open-source their datasets. Thus, we compare SOTA [2] with PACE on our benchmark (See Sec. 5.3). PACE significantly outperforms [2], with a 10.59 N-MSE error over 17.4.

Q.2 Used errors:

We consistently use normalized mean squared error (l2 loss) in training and testing. We argue that mean squared error is a more informative/right metric for evaluating complex variables’ distance in A.6 (not a contribution to boost accuracy).

Normalized mean absolute error (l1 loss) is to show NeurOLight's indecent accuracy in complicated cases. We intentionally report the same metric in NeurOLight to avoid unmatched numbers.

Why normalized MSE, not regular MSE? We randomly sample device configuration and light combination. The resulting optical field yields distinct statistics, e.g., total energy. We need to normalize them to ensure a balance in optimization/evaluation on different optical fields.

As for acceptable error for downstream tasks, PACE achieves ~5% N-MSE error, sufficient to guide device optimization in the early device design and exploration stage (better fidelity than a larger simulation grid shown in the next question).

Q.3 Speedup over numerical methods with less accuracy:

We can’t set a targeted error in the simulator as the angler uses a direct linear solver, not an iterative solver. However, we can reduce the simulation granularity to speed up the process with sacrificed accuracy. In our experiment, we use a commonly used 0.05nm to obtain ground truth. A larger granularity (e.g., 0.075 nm) leads to a very different field (qualitatively compared in reb-Fig.3; quantitive N-MSE error 1.2).

As you suggested, we added PACE's speedup over the 0.075nm simulation in reb-Fig.5. PACE still shows a 5.1-10.6x speedup with much better fidelity.

Q.4 Unclear human learning

Please see common response Q.3.

Q.5 Spectrum of the predicted field

The predicted field spectrums of PACE and NeurOLight are in reb-Fig.3. PACE excellently aligns with the baseline spectrum compared to NeurOLight, while NeurOLight uses the same frequency modes.

Q.6 Etymology of PACE:

Sorry for not annotating PACE’s source in the title: Pacing Operator Learning to Accurate Optical Field Simulation for Complicated Photonic Devices.

It has two meanings: we pace to a new SOTA in AI for photonic device simulation; our learning flow decouples the hard task into two progressively tasks in a pacing manner.

Q.7 Ood testing

Firstly, we’d like to stress that our generalization experiment is done on a specific wavelength range (1.53-1.565 μm), which is C-band (an interested range for device design). We train with sampled wavelengths in C-band and test its ood generation on unseen frequencies. It is a vital test to prove the usefulness of PACE in helping device design within an interested wavelength range.

As you suggested, we tested the accuracy outside the C-band (See reb-Fig.6). PACE shows good accuracy on neighboring wavelengths while holding a 10-15% error at a further range. This is expected since wave propagation is sensitive to frequency. It can be mitigated by incorporating sampled wavelengths into training.

Q.8 Speed-up experiments

In speed-up evaluation on various device scales, we only test the speedup by varying device geometry sizes; therefore, no training is performed.

For resolution invariance, our cross-axis factorized neural operator (Fig.2) uses the same components, the Fourier integral operator and linear layers, with factorized FNO and FNO. Hence, we hold the same resolution-invariant property theoretically. But, as stated in [4], no model can really deliver a real zero-shot since finer scales will contain new physics interactions (re-training is preferred).

Q.9 Seemingly inconsistent colorbars A.9

In A.9, each column consistently uses the same test case to evaluate various methods using the same color bar. Across different test cases between columns, the optical field may yield distinct statistics, e.g., total energy, leading to different color bars. We put a wrong figure for Dil-ResNet in Fig.10, which may be the reason to confuse you(updated in reb-Fig.2).

Q.10 Error in log space

We show the absolute error of one device in log space in reb-fig.7, which highlights small errors to help understand the error pattern. As we can see, the error increases from left to right due to challenge 3 in Sec.3.1 (non-uniform learning complexity). Besides, error residual occurs especially when scattering or local resonance happens (challenge 1: complex light-matter interaction).

Q.11 3D cases

We leave the investigation of our operator for 3D FDFD optical simulation for future study. Collecting data in a 3D case would require much more time and computation resources. However, we believe our operator will scale well in 3D cases as a type of factorized neural operator kernel like FactFNO.

Q.12 Writing issues

Thank you so much for carefully reading and for your valuable suggestions. Based on your feedback, we will fix typos and improve the figures.

We will consistently use "complicated" for device and "complex" for light-matter interaction.

评论

Dear Reviewer,

Thank you very much for your thorough review and valuable comments on our paper. We have carefully considered all your feedback and have addressed your concerns and questions in our rebuttal.

We would greatly appreciate it if you could take some time to review our responses. We will be online over the next few days and are happy to address any further questions or concerns you may have.

评论

[1] Liu, Xinliang, et.al. "Mitigating spectral bias for the multiscale operator learning with hierarchical attention." arXiv preprint arXiv:2210.10890 (2022).

[2] Bo Xu and Lei Zhang. “Dilated convolution neural operator for multiscale partial differential equations.” arXiv preprintarXiv:2408.00775v1 (2024).

[3] Timurdogan, Erman, et al. "APSUNY process design kit (PDKv3. 0): O, C and L band silicon photonics component libraries on 300mm wafers." Optical Fiber Communication Conference. Optica Publishing Group, 2019.

[4] Github of Latent-Spectral-Models

评论

Dear Reviewer TfAG,

We sincerely appreciate your valuable comments and time. You provided a very high-quality discussion and made important suggestions to further improve our paper writing and figures.

We have carefully addressed all your comments in our response above, with extra experiments as you suggested. We would like to point out that our paper aligns with NeurIPS’s interests and offers novel ML contributions to the AI for PDE community. We built a new and stronger SOTA compared to the previous SOTA (NeurOLight @ NeurIPS’22).

As the discussion period is drawing to a close in 1 day, we would be very grateful if you could take a moment to review our responses. If our replies have satisfactorily addressed your concerns, we greatly appreciate it if you could acknowledge this in the discussion thread and consider raising the rating.

If you have any remaining questions or concerns, please do not hesitate to contact us. We look forward to your feedback.

评论

Dear Reviewer TfAG,

Thank you once again for your review. As the deadline for the author-reviewer discussion approaches, we noticed that we haven't received any comment from you.

We have addressed all your questions with additional experiments and clarifications.

  • Regarding your main concern about our scope, we would like to point out that our paper aligns with NeurIPS’s interests and offers novel ML contributions to AI for the PDE community (See global response Q.1), with a new parameter-efficient cross-axis factorized neural operator and a new learning flow for solving challenging PDE cases.
  • We would point out that no one model can rule all PDE cases. Our work shows important insights into how to enhance neural operators for challenging PDE cases. Moreover, we compare with one SOTA work for the challenging multi-scale PDE problem where previous neural operators also fail (similar to our challenge in that the PDE coefficient is highly contrasted). Because they didn't release the dataset, we compared our work with theirs on our benchmark. Our work shows significantly better accuracy than theirs, proving it can provide important insight into the multi-scale PDE problem.

As the discussion period is ending, we would greatly appreciate any additional feedback you might have. If our responses have clarified your understanding of our paper, we sincerely hope you might consider raising the rating.

Thank you again for your effort in reviewing our paper.

Best regards, Authors of Paper 472

评论

Dear authors, Thanks for the detailed point-by-point clarifications to my review. I went through the rebuttal and the attached pdf and believe that it addresses most of the points I raised.

The following questions have been answered sufficiently: Q1, Q2, Q4, Q5, Q6, Q7, Q8, Q9, Q10, Q11, Q12.

Please add the plots and explanations for Q5, Q7, Q9, Q10 in the camera-ready version.

Your explanation for relevance to NeurIPS (and energy-efficient ML hardware in general) should also be incorporated in the papers as it can serve as a strong motivation for general ML readers. Also add the origin of PACE acronym somewhere early in the paper.

I am not totally convinced by the speedup claims as the comparisons are not one-to-one wrt accuracy and computational time/hardware, but the authors made a good-faith effort to address this. Further exploration would be outside the scope of this paper at this point. For more information on assessing speed-up claims for ML-PDE models, please refer to this paper[1], specifically eq 1 on page 11.

I also agree with the other reviewer that coarse-to-fine is a better description than divide-and-conquer. You can also frame it as an iterative model. In this context, using the term human learning feels a little bit hyperbolic.

Finally, please proof-read the paper for typos that are mentioned in Q12 and also for those that might have been missed.

I am hoping that the mentioned changes are incorporated in the camera-ready version of the paper, and raising the score assuming this.

[1] McGreivy and Hakim, Weak baselines and reporting biases lead to overoptimism in machine learning for fluid-related partial differential equations. https://arxiv.org/abs/2407.07218

审稿意见
6

This paper develops a computationally efficient learning based PDE simulator for modeling optical fields within photonic devices. The proposed method builds upon NeurOLight [7] and adds a novel factorization of the integral kernel and a two-stage approach to prediction---produce a rough initial estimate that is then refined. The proposed method is noticeably more accurate than [7] and still considerably faster than conventional, non-learning based PDE solvers. An ablation study validates the importance of both the factorization adn the two-stage approach.

优点

Efficiently and accurately simulating photonic devices is an important problem with broad applications.

THe proposed method is more accurate than existing learning based solvers and 12x faster than conventional methods

缺点

The paper is written for a somewhat narrow audience and contains a number of typos. E.g., "boarder impact".

The 577x comparison number is somewhat misleading and should be dropped from the abstract. The proposed method may be much faster than scipy, but its only 12x faster than paradiso.

问题

Are the "real-world" devices actually manufactured and measured or merely simulated? I found section 5.1 unclear.

局限性

Yes

作者回复

Thank you so much for your valuable feedback on our work! Please see our responses below:

Q.1 Somewhat narrow audience

We would like to stress that our paper aligns well with NeurIPS’s interest and has essential machine learning contributions to AI for the PDE community. Please check the detailed answer in our common response Q.1.

Q.2 Confusing description in speedup

Thank you so much for pointing out the description is misleading. We will modify it to “In terms of runtime, our PACE model demonstrates 154-577x and 11.8-12x simulation speedup over numerical solver using scipy or highly-optimized pardiso solver, respectively”.

Q.3 Unclear description of “real-world” devices

The devices used are claimed to be in the real world since they have real-world geometry, permittivity distribution, input light sources, and wavelength ranges. (check detailed response in common response Q.2)

As we cannot measure the optical field after the device is manufactured, photonic designers rely on simulation tools to obtain it to aid the design loop. Therefore, in this paper, we also use the simulation tool to generate the ground-truth optical field to assess the effectiveness of our model.

Q.4 Typos

Thank you so much for your careful reading, which is crucial to improving our paper quality. Since we cannot upload the revised manuscript now, we will fix all typos and update the final manuscript.

Thank you so much for reviewing our paper and this response. We look forward to further discussions with you if you have any other questions!

评论

Dear Reviewer,

Thank you very much for your thorough review and valuable comments on our paper. We have carefully considered all your feedback and have addressed your concerns and questions in our rebuttal.

We would greatly appreciate it if you could take some time to review our responses. We will be online over the next few days and are happy to address any further questions or concerns you may have.

评论

Dear Reviewer ZkVJ,

We sincerely appreciate your valuable time in reviewing our paper.

We have carefully addressed all your comments in our response above. We would like to point out that our paper aligns with NeurIPS’s interests and offers novel ML contributions to AI for the PDE community. We built a new and stronger SOTA compared to the previous SOTA (NeurOLight @ NeurIPS’22). Moreover, we follow your suggestion to clarify the unclear description, and those parts will be merged into the revised version. Thank you again for your suggestion, which is crucial to improve our paper writing.

As the discussion period draws to a close in 1 day, we would be very grateful if you could take a moment to review our responses. If our replies have satisfactorily addressed your concerns, we would greatly appreciate it if you could acknowledge this in the discussion thread.

If you have any questions or concerns, please do not hesitate to contact us. We look forward to your feedback.

作者回复

We sincerely appreciate the constructive feedback provided by all the reviewers. We are truly inspired by their acknowledgment of our paper’s contribution and strong accuracy compared to various baselines as a new state-of-the-art in AI for optic simulation.

We appreciate the chance to address common comments in this response.

Q.1 Relevance to NeurIPS audience: (Reviewer ZkVJ & TfAG)

Our work tackles the challenging photonic device simulation problem with a novel neural operator design and a new learning flow and establishes a strong SOTA on AI for optics simulation compared to previous SOTA (NeurOLight @ NeurIPS’22).

It aligns well with the interest of the NeurIPS:

  • Photonic analog computing has gained much momentum, promising a next-generation efficient AI computing paradigm. A line of works on optics for AI [1-5] and AI for optics[6] appeared at NeurIPS or other ML conferences. Moreover, NeurIPS held a “Machine Learning with New Compute Paradigms” workshop last year and will do so this year. Our work can serve as an ultra-fast, high-fidelity surrogate model to accelerate the exploration of new customized devices and the design of scalable next-generation photonic computing systems, closing the loop of optics-AI synergy.

We have intellectual contributions to the AI for PDE community:

  • A novel parameter-efficient cross-axis factorized neural operator.
  • A new learning flow that solves hard PDE problems from coarse to fine.
  • Our discrete permittivity challenges are similar to those in multi-scale PDE [7-8]. Since they don’t open-source their dataset, we compare the most recent [8] with PACE on our benchmark(see 5.3 Discussion.4), where PACE shows a significantly lower N-MSE error (10.59% vs 17.4%)
  • We open-source the dataset and plan to include more devices and examples. This will contribute a new PDE problem to the AI for the PDE community, and the new PDE problem will have a huge impact on photonic design.

We believe our work will inspire more "AI for optics" works, and our exploration shares insights for "parameter-efficient ML PDE surrogate model" and "how to enhance neural operator for challenging PDE cases."

Q.2 Clarification of real-world and complicated photonic devices (Reviewer ZkVJ & weV6)

Let’s first recall the problem formulation: find a mapping from the observation space Ω, 𝜖, 𝑤, 𝐽 to the underlying optical field 𝑈.

We claim to be real-world since

  • The used photonic device really exists[9-10], meaning the geometry (Ω), and permittivity(𝜖) are real-world.
  • The input light source 𝐽 is real, and we test our model with various randomly generated 𝐽 combinations.
  • The wavelength (𝑤) range we focus on is C-band (1.53-1.565), a commonly interested range for device design.
  • Moreover, permittivity distribution 𝜖 is discrete and contrastive considering the real manufacturing limits. However, many PDE benchmarks generate coefficients that are slowly/smoothly transited.

Our devices are complicated since (details at Sec 3.1)

  • Complex light-matter interaction, e.g., scattering and resonance.
  • Sensitivity to local minor structures
  • Rich frequency information in the optical field

And NeurOLight fails to give decent accuracy on these complicated devices, as they reported.

Q.3: unclear explanation of human learning (Reviewer TfAG & weV6)

Thanks for pointing out that our term “human learning” is unclear. We added the explanation here and will include it in the revised version.

Existing ML for PDE solving work typically learns a model in a one-shot way by directly learning the underlying relationship from input-output pairs. Unlike AI systems, humans don’t learn new and difficult tasks in a one-shot manner; instead, they learn skills progressively, starting with easier tasks and gradually moving to harder ones. For example, instead of directly learning how to solve equations, students first learn basic operations, such as addition and multiplication, and then move on to solving complex equations.

Inspired by this human learning process, unlike previous work that directly learns a one-stage model, we propose to divide the challenging optical field prediction problem into two sequential latent tasks. The first learns to generate a rough initial solution, followed by the second to refine details.

[1] Li, Yingjie, et al. "RubikONNs: Multi-task Learning with Rubik's Diffractive Optical Neural Networks." IJCAI 2023.

[2] Gu, Jiaqi, et al. "L2ight: Enabling on-chip learning for optical neural networks via efficient in-situ subspace optimization." NeurIPS 2021.

[3] Gu, Jiaqi, et al. "Efficient on-chip learning for optical neural networks through power-aware sparse zeroth-order optimization." AAAI 2021.

[4] Ohana, Ruben, et al. "Photonic differential privacy with direct feedback alignment." NeurIPS, 2020.

[5] Gupta, Sidharth, et al. "Don't take it lightly: Phasing optical random projections with unknown operators." NeurIPS 2019.

[6] Gu, Jiaqi, et al. "Neurolight: A physics-agnostic neural operator enabling parametric photonic device simulation." NeurIPS 2022.

[7] Liu, Xinliang et.al. "Mitigating spectral bias for the multiscale operator learning with hierarchical attention." arXiv preprint arXiv:2210.10890 (2022).

[8] Bo, Xu, et al. “Dilated convolution neural operator for multiscale partial differential equations.” arXiv preprintarXiv:2408.00775v1 (2024).

[9] Mohammad, Tahersima, et.al. “Deep neural network inverse design of integrated photonic power splitters.” Sci. Rep., 2019.

[10] Sanaz Zarei et al. “Integrated photonic neural network based on silicon metalines”. Optics Express 2020.

评论

Dear Reviewer,

We sincerely appreciate your valuable time in reviewing our paper. We have carefully addressed all your comments in our response and added extra experiments/visualization as you suggested (appended in the pdf in the global rebuttal).

Our work tackles the challenging photonic Maxwell PDE simulation problem on real-world complicated devices with a novel neural operator design and a new learning flow and establishes a strong SOTA (>39% lower error) on AI for optics simulation compared to previous SOTA (NeurOLight @ NeurIPS’22).

As the discussion period draws to a close in 1 day, we would be very grateful if you could take a moment to review our responses. If our replies have satisfactorily addressed your concerns, we would greatly appreciate it if you could acknowledge this and kindly consider raising the rating.

We are always open to further suggestions or feedback that could help us improve the paper even more. If you have any questions or concerns, please do not hesitate to contact us.

Thank you again for your valuable input.

Best regards,

Authors of Paper 472

最终决定

The paper proposes a new neural operator setup for solving Maxwell equations in 2D. All reviewers recommend acceptance, so it is difficult to argue; I will add, however, that some of the claims are quite misleading. There is a speedup compared to the classical solver, but nothing is said about the speedup for the comparable accuracy; thus, this has to be correctly phrased in the final version. It is ok to beat n.o. baseline, but it is not ok to claim being faster but for a different accuracy.