PaperHub
7.2
/10
Spotlight4 位审稿人
最低3最高5标准差0.8
4
3
3
5
ICML 2025

Not all solutions are created equal: An analytical dissociation of functional and representational similarity in deep linear neural networks

OpenReviewPDF
提交: 2025-01-23更新: 2025-08-16
TL;DR

Neural networks performing identical tasks can organize information completely differently, except for particularly robust ones that employ canonical internal patterns.

摘要

关键词
deep learningrepresentational similarityrepresentation learningdeep linear networksneuroscience

评审与讨论

审稿意见
4

This paper studies the relation between functional similarity and representational similarity, finding that there is a disassociation: i.e., functional equivalent networks may have very different representations. To be concrete, they mainly study a two-layer linear networks, concerning their weights W1,W2W_1, W_2 and hidden representations hh. They study the manifold of least-square solutions, and minimum-weight-norm and minimum-repr-norm solutions. These norms show different hidden representations and representational similarity matrices, even though they all have the same overall function. Least square solutions are the most flexible -- in fact, representations can be arbitrarily manipulated (e.g., into an elephant). For the minimum-weight-norm or minimum-repr-norm solutions, the representations still have some degree of freedom, but RSM is uniquely determined from task data, which is desirable. They finally study why there is representation alignment between biological and artificial systems -- due to their robustness to noise.

Update after rebuttal

The authors have addressed my questions. Overall I think this is a great paper, but 5 is a bit of a stretch, so I keep my score as 4.

给作者的问题

  • what's the dynamic (optimization) reason that an (artificial) network would seek robust solutions? Is it due to the implicit regularization of optimizers?
  • I'm a bit confused by the illustration in Figure 1. When a smaller circle is contained inside a larger circle, what does this mean? For example in subfigure D, the grey oval H is contained in the larger orange oval which is connected to W2.

论据与证据

Yes

方法与评估标准

Yes

理论论述

I skimmed through the theorems (although I didn't derive by hand myself) and did not spot any error.

实验设计与分析

This is mainly a theory paper. The experiments are minimal but are very informative and nicely designed.

补充材料

I briefly skimmed through the whole SM

与现有文献的关系

Although training of deep linear networks can be analytically characterized, the solution space seems trivial to characterize, this paper takes a unique point -- by constraining the norm of representations/weights, hidden representations can be more interpretable or more truthful to the underlying structure of data. This paper uses both mathematical derivations and neat illustration examples to support this idea.

遗漏的重要参考文献

The paper mentioned implicit regularizations in deep learning. It would be nice to mention the implications regularization of SGD:

  • ON THE ORIGIN OF IMPLICIT REGULARIZATION IN STOCHASTIC GRADIENT DESCENT, ICLR 2021

其他优缺点

Strengths

  • The paper is nicely written and neatly presented. I like the elephant plot to show that representations can be arbitrarily engineered given the same functionality.
  • The idea is rigorously supported by mathematical derivations.
  • The minimum-norm (weight or representation) solutions are interesting objects to identify and study. In particular, their representations reveal the truthful hidden structure of data

Weaknesses

  • This is probably too much of an ask for a theory paper, but adding more experiments would make the paper more appealing to practitioners.
  • The representation alignment between artificial vs biological systems due to robustness to noise is interesting but might be over-stated.

其他意见或建议

  • Line 152, a comma out of place
  • Figure 3c. It would be nice to point out (in the caption) that the elephant is deliberately engineered. I was a bit confused at first because there is no elephant in the 16 items.
  • Figure 5 caption did not explain what "duplicate" means.
作者回复

We thank the reviewer for their constructive and detailed feedback on the manuscript.

First, we address the reviewer's question regarding the optimization reasons behind an artificial network seeking robust solutions. Our analysis is general—we study the entire solution manifold and derive broad statements about neural computations and their representations without relying on any specific learning or optimization algorithm. An investigation into why and when an algorithm converges to a particular region of the solution manifold is beyond the scope of our study. However, as noted by the reviewer, extensive work in the machine learning literature has explored this question, albeit without focusing on their relation to neural representations. We have now expanded our Discussion to address this important aspect in more detail, including relevant references concerning implicit regularisation (e.g., Smith et al., 2021, as suggested) and neural collapse (e.g., Zhu et al., 2021).

Next, we address the reviewer's second question regarding the visualizations in Figure 2. Our aim was to follow well-established standard representations found in the literature (e.g., the Wikipedia article on “Kernel (linear algebra)”). Nevertheless, we recognize the need for additional clarity. We have revised the figure caption and expanded the accompanying explanations to explicitly clarify that: (1) the orange areas represent the general vector space in which representations reside, (2) open black circles denote the image of linear transforms, with the surrounding area corresponding to their kernel, and (3) filled black circles indicate the subspaces occupied by the input, hidden, and output representations of the training data, respectively. We welcome any further suggestions for improving the clarity of our visualizations.

We appreciate the reviewer's suggestion for additional experiments. Could the reviewer specify which experiments they believe would most effectively strengthen our paper’s message? We are open to incorporating further empirical evaluations, even if complex, to reinforce the practical implications of our theoretical findings.

Furthermore, as suggested, we have tempered our claims regarding noise as the cause of representational alignment between artificial and biological systems, instead framing them as hypotheses that open up an interesting field of investigation. In particular, it remains unclear if, why, and when brains may operate in these regimes. However, we emphasise that our results suggest that assuming they do without empirical validation could lead to misleading comparisons and conclusions about decodability and representational similarity.

We have also made the following revisions based on the reviewer’s comments:

  • Corrected the typo in line 152.
  • Clarified in the caption of Figure 3c that the elephant is deliberately engineered.
  • Expanded the caption of Figure 5 to explain that "duplicate" refers to networks extended by duplicating hidden neurons, a transformation that preserves the input-output mapping.

We once again thank the reviewer for their time and valuable insights. We remain open to further suggestions, particularly regarding simulation studies that could further illustrate the scope and significance of our contributions.

审稿意见
3

The paper presents a mathematical analysis of the nature of solutions (for a given problem) in an overparameterised two layer linear neural network. This is done through a theoretical study of the manifold of generic solutions, i.e. different choices of weights values which give the same fit of the training data, and identifying several types of solutions grouped by certain mathematical properties. This leads to a conclusion that analysis of overparametrised networks is challenging, because individual transformations/internal representation is arbitrary given the next transformation can correct for it, and thus the entirety of the mapping needs to be taken into account.

给作者的问题

Does the analysis you provide tell us anything else other than it is hard to analyse parts of internal representation in isolation from other parts of the network?

论据与证据

Yes. The paper is mathematically dense, but the conclusions seem justified. It's just that I think most would take it as a given, without the presented evidence, that overparamterised networks are non-trivial due to the fact that what individual components do is not informative, and instead one has to take into account the entirety of the mapping.

方法与评估标准

Linear network analysis is one of tools for anlysing neural networks, and while it has its limits, for obvious reason of missing the non-linearity aspects, it's still tells us something. I do find the connection made in this paper to the non-linear networks by stating that they also have multiple solutions (through symmetries) somewhat weak - not in accuracy - sure, non-linear network offer multiple solutions - but in relevance of this work to that aspect - do the categories of solutions analysed here tell us anything about how non-linear solutions work?

理论论述

I did not check the proofs carefully. I think it might be sound, since I don't disagree with the ultimate conclusion of the paper.

实验设计与分析

Experimental design and analysis seems fine. I find the arbitrary 'elephant' internal representation a "cute" illustrative example. However, though I applaud authors' attempt to give a high-level/intuitive explanation of what different solutions entail (in the paragraph after Theorem 3.7) I am struggling to understand those explanations and I don't find Figure 2, while again appreciated as an attempt at intuitive explanation, all that helpful.

补充材料

I did not go carefully through supplementary materials - it's 8 pages of dense math, which makes it feel like perhaps this type of publication is more suited for a journal rather than a conference?

与现有文献的关系

This work continues the thread of linear network analysis - an analysis of simplified problem (linear netowk) in hope of understanding a hard problem (the non-linear counterpart). I think this work might be a great starting point of an interesting analysis, just at this moment seems preliminary, since it only that analysis leads to a trivial conclusions - that understanding internal representation is a hard problem. We already know that.

遗漏的重要参考文献

Not to my knowledge.

其他优缺点

Seems like early work, analysis might be promising, but the current conclusion seems obvious and not new.

其他意见或建议

The equation in Assumption 2.2 seems inverted. If the network is not bottle-necked then it seems that Nhmin(Ni,No)N_h \ge min(N_i,N_o), not Nhmin(Ni,No)N_h \le \min(N_i,N_o). In Laurent and Brecht this condition is the "thinnest layer is either input or outut", which means NhN_h is at least the same or greater than input/ouput, not smaller or the same. Not sure if this is misunerstaning on the part of the authors, or just a typo.

作者回复

We would like to sincerely thank the reviewer for their time and detailed feedback. Below, we address the key points raised in the review.

The reviewer asks whether our analysis provides insights beyond the conclusion that internal representations cannot be interpreted in isolation. First, we emphasize that this is not widely understood, as many subfields of neuroscience, machine learning and their intersections do center on analysis of representations in isolation. For example, as we outline in great detail in Section 4, a significant body of research in computational neuroscience assumes that internal representations between systems (models and/or brains) can be meaningfully decoded and compared. Our work analytically demonstrates that this assumption does not necessarily hold—in particular, we identify in which regimes representations are arbitrary with respect to the task, and instead dependent on other parameters like parameter initialization.

However, this is only one part of our findings. We further identify specific regimes on the solution manifold where representational similarities are not arbitrary. We derive precise analytical conditions under which internal representations in fact are informative and comparable without requiring knowledge of other parts of the network; in particular, when they are task-specific. Additionally, we show that these task-specific stable representations coincide with regions of the solution manifold that are robust to noise, linking identifiable representational structure to a desirable computational property. This goes beyond the claim that “understanding internal representation is a hard problem” because we establish precisely when it is hard.

It is not clear to us why the reviewer dismisses these results as “intuitive”, “preliminary”, and “not novel”. To ensure a constructive discussion, we kindly ask the reviewer to provide references to prior work that establishes the same conclusions as we do in the manuscript.

Additionally, to clarify our contributions, we have revised the manuscript by adding an explicit contributions section that is detailed in the response to Reviewer rKc5.

Minor:

We would like to thank the reviewer for catching the typo in Assumption 2.2. Indeed, the greater-than-or-equal sign should have been a less-than-or-equal sign!

We appreciate the reviewer’s explicit feedback on Figure 2. Our visualizations follow well-established standards found in mathematical visualizations (e.g., the Wikipedia article on “Kernel (linear algebra)”). However, we are happy to improve and expand on the explanations of the image and kernel of a matrix in the main text. Further, we have revised the figure caption and expanded the accompanying explanations to explicitly clarify that: (1) the orange areas represent the general vector space in which representations reside, (2) open black circles denote the image of matrices, with the surrounding area corresponding to their kernel, and (3) filled black circles indicate the subspaces occupied by the input, hidden, and output representations of the training data, respectively. We welcome any further suggestions for improving the clarity of our visualizations.

Further, the reviewer notes that they did not carefully review the supplementary material due to its mathematical density, but still conclude that the results “might be sound” as they do not disagree with the paper’s results and conclusion. This aligns precisely with ICML guidelines, which state that reviewers are encouraged (but not required) to consult supplementary material and that key claims should be understandable from the main text. Furthermore, rigorous mathematical derivations are essential for theoretical contributions in machine learning, just as extensive simulations are standard for empirical work. Appendices that extend over many pages are common in ICML theory papers, and theory of machine learning and applications to (neuro)science are subject areas in ICML’s yearly call for papers. Throughout our 8-page supplement, we have made substantial efforts to fully provide all our assumptions, proofs and derivations to ensure our results are fully reproducible, testable, and clearly presented.

Finally, we sincerely thank the reviewer again for their time and effort. We hope that, in light of our clarifications, the significance of our contributions and the value of our thorough theoretical analysis (including the detailed appendix) become evident. We remain open to any further suggestions or points for clarification and are hopeful that the reviewer may reconsider their assessment accordingly.

审稿人评论

Thank you for the rebuttal. I am not willing to concede my point that it is not surprising or novel to find out that internal representation of individual layer is arbitrary in isolation. I can't point to specific literature, because, as I said it's on "intuitive" (somewhat obvious) level that we know this - we know it is hard to discern the internal representation of neural networks, because it's distributed across neurons and layers. I suppose I could point to several works that attempted layer-wise supervised learning that come up short when compared to end-to-end supervised training. And I don't think analysis of internal representation in isolation is driven by the belief that this is the best way to go about it, but rather is necessitated by the need to make things tractable. Just like is the case with with the analysis of deep linear networks – they are not a replacement for non-linear models, and they miss important aspects of the deep neural networks we use in practice, but we study them because it’s tractable and easier. However, in light of the rebuttal, and other reviewers' comments, I am willing to grant that I might have undervalued the mathematical aspects of this work, and that the presented rigorous mathematical treatment might be a decent step towards better understanding of the types of solutions on the solution manifold. I will therefore raise my score.

作者评论

We appreciate the reviewer's revised assessment. One clarification that we would like to make: Our claimed contribution is not to be first in noting challenges in analysing representations—our related work section covers several previous negative examples (see "Comparing the solutions of artificial and biological networks")—but rather to provide a rigorous mathematical treatment of these challenges using an appropriate surrogate model (deep linear networks) that may unify and explain these negative observations.

审稿意见
3

The authors study a two-layer linear network. They characterize the space of solutions for such networks, with emphasis on several normalization schemes. The result is that there are many zero-loss solutions that differ in how minimal they are. Specifically, whether the transformation from input to hidden or from hidden to output is minimal with respect to the training data. The authors discuss possible implications of these results to representational drift and to the comparison of biological and artificial networks.

After rebuttal

After reading all the rebuttals and discussion with all reviewers, I am keeping my score.

给作者的问题

None

论据与证据

Yes. There is degeneracy in solutions, and different regularization schemes choose different solutions. There is also a relation between noise robustness and regularization. It is not clear which of these claims is novel.

方法与评估标准

Not relevant

理论论述

I read the proofs, but did not verify their correctness in detail.

实验设计与分析

Figure 5: the definitions of scaled, nuisance, etc only appear in the appendices. This seems like something that should be in the main text.

补充材料

I read all the supplement.

与现有文献的关系

The authors mention the main relevant papers (Baldi & Hornik 1989, Laurent & Brecht 2018). It is not clear which aspects of the current paper are missing from prior work.

遗漏的重要参考文献

Not aware

其他优缺点

Strengths: This is a simple setting, in which the characterization of the solution space is possible. Qualitative insights from this setting can be useful in broader scenarios. The explanation of the different degeneracies (Figure 2) was very clear and intuitive.

Weaknesses: It is hard to understand what exactly is new relative to existing work. For instance, Theorem 5.1 seems like a textbook result on linear regression. The definition of minimum representation-norm is not very intuitive. The first term is the norm of the hidden representation. But the definition also has a sum with the readout weights. Why call this sum “representation-norm”? Figure 5E: It would be useful to discuss the scaling of the effect. Can the theory provide any insights on the actual values of noise in which the different models degrade? Corrolary 5.4: What is the intuition behind the suggested scaling? Line 92 – The way assumption 2.2 is written is quite confusing. It seems as though not having a bottleneck implies narrow hidden layer.

其他意见或建议

None

作者回复

We thank the reviewer for their thorough evaluation of our work, and for providing explicit feedback!

First, we address the reviewer’s concern about the novelty of our analysis. To clarify this, we have added a dedicated "Contributions" section to the revised manuscript and reproduce it here in detail. As stated in the manuscript, deep linear networks perform “…multistage computations that give rise to hidden-layer representations.” Their overparameterization through depth, rather than width as in linear regression, introduces nontrivial representational degeneracies not present in linear regression. Our work provides a complete analytical characterization of these degeneracies and their implications for decodability and comparability of neural representations. Further, we study the consequences of our analysis on use cases in neuroscience, where neural networks are a commonly employed model of learning, in Section 4.

In particular:

  • Definition 3.1 directly follows from Laurent and Brecht (2018).
  • Theorem 3.3 resembles least-squares linear regression, but its derivation in the context of two-layer linear networks requires extra care (e.g., does not hold if the network is bottlenecked).
  • The constraint satisfaction problem in Definition 3.4 has been studied in Saxe et al. (2019), but only under a set of strong assumptions (Σxx=I\Sigma_{xx} = I, Ni=NoN_i = N_o, and that Σyx\Sigma_{yx} has full rank) as stated in the manuscript.
  • In contrast, Theorem 3.5 imposes no additional assumptions about the task's statistics or network structure (beyond the network not being bottlenecked).
  • Definition 3.6 stems from our analysis of parameter-noise robustness (Section 5), but can be related to notions from implicit regularisation (e.g., Smith et al., 2021) and neural collapse (e.g., Zhu et al., 2021).
  • Accordingly, Theorem 3.7, which shows the parameterization of minimum representation-norm solutions, is novel to the best of our knowledge.
  • Theorems 3.8 and Corollaries 3.9–3.11 are novel and central to our analysis, analytically revealing the degrees of freedom within neural representations and defining when representational similarities are fixed and can be used for functional comparison.
  • For theorems and corollaries regarding robustness to noise in Section 5, similar results exist for linear regression (e.g., minimal norm confers noise-robustness) but again require careful treatment in the multi-layer setting (e.g., cross-correlation terms in the case of parameter noise).

In summary, to our knowledge, this is the first work to make an analytical and exact connection between different regions of the solution manifold of deep linear networks and the identifiability and comparability of neural representations, as well as their correspondence to optimality in noise robustness. If the reviewer is aware of any further prior work that has similar results, we would be delighted to include it and provide appropriate reference.

Minor:

We would like to thank the reviewer for catching the typo in Assumption 2.2. Indeed, the greater-than-or-equal sign should have been a less-than-or-equal sign!

Regarding terminology, we made an effort to be precise while avoiding overly lengthy terms like "minimum-readout-and-representation-norm" . We recognize that "minimum-representation-norm" is only partially descriptive, as it omits the fact that we are also minimizing the norm of the readout weights. While minimising the norm of the readout-weights has no influence on the hidden-layer representation, we chose the constraint satisfaction problems in Definitions 3.2, 3.4 and 3.6, such that they align with the analytical results on noise robustness in Section 5. We believe our terminology is sufficiently descriptive but would be grateful for any suggestions the reviewer may have for a more concise naming convention.

Regarding Corollary 5.4, the parameter noise is scaled by the norm of the inputs and the size of the output layer (see the first and second summands in Equation 22). By inversely scaling the variance of the noise, we ensure the equation is factored consistently, independent of these measures. Without this adjustment, the solution would be identical up to a fixed scaling factor. Thus, this scaling is simply a matter of convenience. We have added a corresponding comment to the manuscript.

Regarding the scaling in Figure 5E, the initial phase of the sigmoidal curve corresponds to test error (near-zero noise), while the ceiling aligns with random guessing (high noise). We lack analytical insights into the shape of this curve due to the non-linear setting. However, we would be happy to include additional simulations if the reviewer has a specific question in mind.

Again, we would like to thank the reviewer for their time and effort, and we remain open to any further suggestions or points for clarification.

审稿意见
5

This paper analytically studies the hidden representations of two-layer feedforward networks trained to minimize differentiable, convex loss functions. The only sets of weights in the networks studied are read-in and read-out weights, leading to simple expressions for both in terms of the input data. The paper shows that, even in this simple case, function can be dissociated from representation: two networks with different representations can achieve the same loss on the task. The paper then investigates the set of solutions obtained when optimizing the network under various regularization schemes, such as minimizing the sum of squares of the weight norms. These regularized networks, in fact, preserve similarity between one another when taking random walks through function space. Since these regularized networks are also more robust and generalize better, this suggests a mechanism for the empirically observed brain-model similarity "in the wild."

给作者的问题

N/A.

论据与证据

Yes.

方法与评估标准

Yes.

理论论述

Yes, all of them.

实验设计与分析

N/A.

补充材料

Yes, all of it.

与现有文献的关系

This paper will be of interest to many researchers working in both neuroscience and machine learning.

遗漏的重要参考文献

No.

其他优缺点

This is a very, very good paper. It is hard to find any weaknesses.

其他意见或建议

  • In the supplementary, I think it would be useful to define the "exclamation over equal sign" in the notation and preliminaries section.
作者回复

We thank the reviewer for their thorough evaluation of the manuscript and supplementary material, and sincerely appreciate their effort and positive feedback! We would like to kindly ask the reviewer, if time permits, to elaborate on the specific strengths and significance of the work to help the Area Chair better understand the basis of their positive assessment.

Regarding the =!\stackrel{!}{=} notation in the supplementary material, it was used to indicate that an equality is not derived but rather an assumed condition from which a separate conclusion is derived. However, since this notation is not widely used, we have replaced it with explicit language such as “such that”, “with the requirement that”, or “we would like to show” for clarity.

Please do not hesitate to let us know if there are any further points or suggestions to address.

最终决定

The paper presents an analytical treatment of how hidden representations in two-layer linear neural networks are related to the network's function. The authors show that even under very simple conditions (linear net, convex loss) function and representation can be dissociated. They also show that certain regularization schemes preserve similarity of networks, suggesting a mechanism for why empirical results find similarities between brain representations and those in neural network models trained on tasks.

Overall, the reviewers applaud the conceptual framework and praise the mathematical rigour of the paper. They find the paper well written and the results intuitive. There were no major criticisms raised. Hence I strongly recommend accepting the paper, which I believe will make a great contribution to ICML.