On convex conceptual regions in deep network representations


Journal article


Lenka Tvetkova, Thea Brusch, Teresa Scheidt, Fabian Martin Mager, Rasmus Aagaard, Jonathan Foldager, Tommy Sonne Alstrom, Lars Kai Hansen
arxiv, 2023

Semantic Scholar ArXiv
Cite

Cite

APA   Click to copy
Tvetkova, L., Brusch, T., Scheidt, T., Mager, F. M., Aagaard, R., Foldager, J., … Hansen, L. K. (2023). On convex conceptual regions in deep network representations. Arxiv.


Chicago/Turabian   Click to copy
Tvetkova, Lenka, Thea Brusch, Teresa Scheidt, Fabian Martin Mager, Rasmus Aagaard, Jonathan Foldager, Tommy Sonne Alstrom, and Lars Kai Hansen. “On Convex Conceptual Regions in Deep Network Representations.” arxiv (2023).


MLA   Click to copy
Tvetkova, Lenka, et al. “On Convex Conceptual Regions in Deep Network Representations.” Arxiv, 2023.


BibTeX   Click to copy

@article{lenka2023a,
  title = {On convex conceptual regions in deep network representations},
  year = {2023},
  journal = {arxiv},
  author = {Tvetkova, Lenka and Brusch, Thea and Scheidt, Teresa and Mager, Fabian Martin and Aagaard, Rasmus and Foldager, Jonathan and Alstrom, Tommy Sonne and Hansen, Lars Kai}
}

Abstract

The current study of human-machine alignment aims at understanding the geometry of latent spaces and the correspondence to human representations. G\"ardenfors' conceptual spaces is a prominent framework for understanding human representations. Convexity of object regions in conceptual spaces is argued to promote generalizability, few-shot learning, and intersubject alignment. Based on these insights, we investigate the notion of convexity of concept regions in machine-learned latent spaces. We develop a set of tools for measuring convexity in sampled data and evaluate emergent convexity in layered representations of state-of-the-art deep networks. We show that convexity is robust to basic re-parametrization, hence, meaningful as a quality of machine-learned latent spaces. We find that approximate convexity is pervasive in neural representations in multiple application domains, including models of images, audio, human activity, text, and brain data. We measure convexity separately for labels (i.e., targets for fine-tuning) and other concepts. Generally, we observe that fine-tuning increases the convexity of label regions, while for more general concepts, it depends on the alignment of the concept with the fine-tuning objective. We find evidence that pre-training convexity of class label regions predicts subsequent fine-tuning performance.


Share