Natural Language Processing

Revisiting the Platonic Representation Hypothesis: An Aristotelian View

FFabian GrögerSShuo WenMMaria Brbić
Published
February 16, 2026
Authors
3

Abstract

The Platonic Representation Hypothesis suggests that representations from neural networks are converging to a common statistical model of reality. We show that the existing metrics used to measure representational similarity are confounded by network scale: increasing model depth or width can systematically inflate representational similarity scores. To correct these effects, we introduce a permutation-based null-calibration framework that transforms any representational similarity metric into a calibrated score with statistical guarantees. We revisit the Platonic Representation Hypothesis with our calibration framework, which reveals a nuanced picture: the apparent convergence reported by global spectral measures largely disappears after calibration, while local neighborhood similarity, but not local distances, retains significant agreement across different modalities. Based on these findings, we propose the Aristotelian Representation Hypothesis: representations in neural networks are converging to shared local neighborhood relationships.

Keywords

representational similarityneural networksspectral measuresneighborhood similaritypermutation-based null-calibration frameworkPlatonic Representation HypothesisAristotelian Representation Hypothesis

More in Natural Language Processing

View all
Revisiting the Platonic Representation Hypothesis: An Aristotelian View | Paperchime