- Are similar representations in neural nets evidence of shared computation? In new theory work w/ Lukas Braun (lukasbraun.com) & @saxelab.bsky.social, we prove that representational comparisons are ill-posed in general, unless networks are efficient. @icmlconf.bsky.social @cogcompneuro.bsky.socialAug 13, 2025 11:30
- Deep networks have parameter symmetries, so we can walk through solution space, changing all weights and representations, while keeping output fixed. In the worst case, function and representation are *dissociated*. (Networks can have the same function with the same or different representation.)
- To analyse this dissociation in a tractable model of representation learning, we characterize *all* task solutions for two-layer linear networks. Within this solution manifold, we identify a solution hierarchy in terms of what implicit objectives are minimized (in addition to the task objective).
- We parametrised this solution hierarchy to find differences in handling of task-irrelevant dimensions: Some solutions compress away (creating task-specific, interpretable representations), while others preserve arbitrary structure in null spaces (creating arbitrary, uninterpretable representations).
- We demonstrate that representation analysis and comparison is ill-posed, giving both false negatives and false positives, unless we work with *task-specific representations*. These are interpretable *and* robust to noise (i.e., representational identifiability comes with computational advantages).
- Function-representation dissociations and the representation-computation link persist in deep nonlinear networks! Using function-invariant reparametrisations (@bsimsek.bsky.social), we break representational identifiability but degrade generalization (a computational consequence).
- Our theory predicts that representational alignment is consistent with *efficient* implementation of similar function. Comparing representations is ill-posed in general, but becomes well-posed under minimum-norm constraints, which we link to computational advantages (noise robustness).
- Main takeaway: Valid representational comparison relies on implicit assumptions (task-optimization *plus* efficient implementation). ⚠️ More work to do on making these assumptions explicit! 🧠 CCN poster (today): 2025.ccneuro.org/poster/?id=w... 📄 ICML paper (July): icml.cc/virtual/2025/poster/44890
- This is very cool! Congrats @eringrant.me !
- many thanks to my collaborators, @saxelab.bsky.social and especially Lukas :)