Universality of intrinsic dimension of latent representations across models


Conference


Teresa Karen Scheidt, Lars Kai Hansen
UniReps: the First Workshop on Unifying Representations in Neural Models, 2023

Paper
Cite

Cite

APA   Click to copy
Scheidt, T. K., & Hansen, L. K. (2023). Universality of intrinsic dimension of latent representations across models.


Chicago/Turabian   Click to copy
Scheidt, Teresa Karen, and Lars Kai Hansen. “Universality of Intrinsic Dimension of Latent Representations across Models.” UniReps: the First Workshop on Unifying Representations in Neural Models, 2023.


MLA   Click to copy
Scheidt, Teresa Karen, and Lars Kai Hansen. Universality of Intrinsic Dimension of Latent Representations across Models. 2023.


BibTeX   Click to copy

@conference{teresa2023a,
  title = {Universality of intrinsic dimension of latent representations across models},
  year = {2023},
  series = {UniReps: the First Workshop on Unifying Representations in Neural Models},
  author = {Scheidt, Teresa Karen and Hansen, Lars Kai}
}

Abstract
While state-of-the-art transformer networks use several hundreds of latent variables per layer, it has been shown that these features can actually be represented by relatively low dimensional manifolds. The intrinsic dimension is a geometrical property of the manifold latent representations populate, viz., a minimal number of parameters needed to describe the representations.
In this work, we compare the intrinsic dimensions of three image transformer networks for classes of the cifar10 and cifar100 dataset. We find compelling evidence that the intrinsic dimensions differ among classes but are universal across networks. This universality persists across different pretraining strategies, fine-tuning and different model sizes. Our results strengthen the hypothesis that different models learn similar representations of data and show great potential that further investigation of intrinsic dimension could lead to more insights on the universality of latent representations. 

Share