PaperHub
3.7
/10
withdrawn3 位审稿人
最低1最高5标准差1.9
5
1
5
3.3
置信度
ICLR 2024

Latent Lie Group Representations

OpenReviewPDF
提交: 2023-09-24更新: 2024-03-26

摘要

关键词
deep learningsymmetrylie groups

评审与讨论

审稿意见
5

This paper introduces a method to learn symmetries in a latent representation of the data. The setting requires having pairs of inputs, one being the original and one the transformed data point. They propose a network to learn a latent space, the Lie algebra elements and the transformation parameters at the same time. They show some limited experimental results on SyMNIST, where each image is paired with an affine transformed version of the image.

优点

The paper discusses some good avenues for learning symmetries in data.

  1. For spatial data, the vector field approach taken for the Lie algebra elements Γ(x)\Gamma(x) (generators) is interesting.
  2. The overall architecture of their model and the loss have some good points, like the LZ\mathcal{L}^Z term in eq. (8)
  3. The learned generators in Figs 4 and 5 seem close to expected ground truth.
  4. The theoretical sections of the paper are fairly polished, well-structured and presented well.

缺点

This paper is definitely moving in a good direction, but it seems to be work in progress.

  1. Lack of baselines and ground truth: The experiments are incomplete. They authors should compare against other symmetry discovery methods such as Augerino (Benton 2020), LieConv (Finzi 2020), L-conv (Dehmamy 2020) and recently LieGAN (Yang 2023). Similar to your eq (5), LieConv uses a subalgebra of GL(3)GL(3). The rotated MNIST experiment in L-Conv also dealt with pairs of input, predicting tt and learning G\mathbf{G} in the process. LieConv and LieGAN also both use the exponential map.
  2. Novelty: Gabel 2023 seems to use a very similar formulation and architecture and even works with the same input data. How do your results contrast with Gabel 2023, and can their results serve as a baseline for yours?
  3. presentation of results:
    1. Fig 3 is not clear or informative. What should the reader learn from it? What is the vertical axis?
    2. Figs 4,5, qualitative and quantitative comparison with ground truth is missing (e.g. cosine similarity with GT and plotting GT)
  4. Detials missing for reproducibility: I couldn't find any appendices or supplementary material. There is also not enough information in the paper (no latent dims or hyperpaparameters), which makes the paper harder to assess.

问题

  1. The affine transformations you are learning, for the most part, act linearly on the inputs. Thus, I don't think going to a latent space is required in this case. Can you elaborate why you use latent space here?
  2. what are the latent dims? List all hyperparameters.
  3. what are the structures of neural nets TθT_\theta and autoencoder layers?
  4. Have you done ablation studies for the terms in the loss function (9)? How much do Lα\mathcal{L}_\alpha and LTZ\mathcal{L}^Z_T matter?
  5. Lα\mathcal{L}_\alpha needs more explanation. How does one go to the prime basis Di\mathbf{D}'_i? This may be doable if the zz space was also 2D, like the pixel space. But what is the exact procedure?
  6. Compared to Gabel 2023, what is new in your paper?
审稿意见
1

This paper propose to learn the symmetry of a dataset via neural networks. The symmetry is represented by a one-parameter Lie group and its generator is represented and learned in the network. The sample dependent tt parameter is also output by the network and it is used to estimate the paramter distribution. A SyMNIST dataset is built and is used to verify the method by discovering the applied affine transformations. Qualitative results are shown.

优点

This paper tackles an interesting topic of symmetry detection. Although it feels like the task is still at an early stage and only toy problems are tested, the related methods may become more influential after more developments.

缺点

  1. The general writting is hard to follow for me. Many sentences are hard to decode, and mathematical notations are not explained well in texts. For example,

    • (a) The definition of "class-generic symmetries" is unclear. What exactly are "symmetries within each class, but also common among all classes"? Any concrete examples?
    • (b) The "SyMNIST task" may be better described as "SyMNIST dataset". A figure that visualizes sample images of the dataset is helpful to readers.
    • (c) "We wish to learn non-canonical symmetries". What are the non-canonical symmetries refer to? Any examples? In section 4, only affine transformations are used in experiments. Does affine transformation belong to non-canonical symmetries?
    • (d) "A collection of multiple power series applied in succession is the neural network". What is the role of non-linear activations in this description?
    • (e) "avoid making calculations in pixel-space directly". What is the calculation in pixel-space refer to? Any examples?
    • (f) "pixel-level transformation will need to be extracted by special methods". Any examples?
    • (g) The network structures of encoder ff, decoder gg and t-network TT are not discussed.
    • (h) What is "aliasing artifacts" in the translation setting?
  2. This work focuses on the one-parameter groups. What if the underlying data contains a wider range of transformations, e.g., both rotation and translation, that cannot be described by a one-parameter group?

  3. There is a lot overlap with Gabel et al. (2023), including figures, equations and descriptions. It would be helpful if the authors could highlight the contribution and novelty compared to it.

  4. The t-test and ϵ\epsilon-test are missing in experiments.

  5. There is no quantitative evaluations, and there is no comparision to related algorithms.

  6. Fig. 4 and Fig 5 are hard to parse and understand due to the lack of description.

问题

See weakness.

审稿意见
5

The paper uses a symmetry identification technique to improve learning in neural networks. The mathematics underpinning the technique is robust. A good implementation would lead to improvements in neural network learning. It will also make the behaviour of NN's more transparent, as-- at least symbolically -- we get to learn what symmetries exists in data beforehand.

优点

The use of mathematical techniques to improve data-oriented machine learning.

缺点

It is not clear if the implemented technique worked at all! A few figures are produced on page 7 (Figures 4 and 5), and we read that a z-test has been performed. No numerical results are however presented and I am not sure how did the model do!

问题

Does the new model learn the data better, faster, with less parameters etc etc? Can you please provide a comparison table of some sort?

伦理问题详情

None