@inproceedings{alshammari2025a, author = {Alshammari, Shaden Naif and Hershey, John R. and Feldmann, Axel and Freeman, William T. and Hamilton, Mark}, title = {A Unifying Framework for Representation Learning}, booktitle = {ICLR 2025}, year = {2025}, month = {March}, abstract = {As the field of representation learning grows, there has been a proliferation of different loss functions to solve different classes of problems. We introduce a single information-theoretic equation that generalizes a large collection of modern loss functions in machine learning. In particular, we introduce a framework that shows that several broad classes of machine learning methods are precisely minimizing an integrated KL divergence between two conditional distributions: the supervisory and learned representations. This viewpoint exposes a hidden information geometry underlying clustering, spectral methods, dimensionality re- duction, contrastive learning, and supervised learning. This framework enables the development of new loss functions by combining successful techniques from across the literature. We not only present a wide array of proofs, connecting over 23 different approaches, but we also leverage these theoretical results to create state-of-the-art unsupervised image classifiers that achieve a +8% improvement over the prior state-of-the-art on unsupervised classification on ImageNet-1K. We also demonstrate that I-Con can be used to derive principled debiasing methods which improve contrastive representation learners.}, url = {http://approjects.co.za/?big=en-us/research/publication/a-unifying-framework-for-representation-learning/}, }