[edit]
Text-To-Concept (and Back) via Cross-Model Alignment
Mazda Moayeri, Keivan Rezaei, Maziar Sanjabi, Soheil FeiziProceedings of the 40th International Conference on Machine Learning, PMLR 202:25037-25060, 2023.
Abstract
We observe that the mapping between an image’s representation in one model to its representation in another can be learned surprisingly well with just a linear layer, even across diverse models. Building on this observation, we proposetext-to-concept, where features from a fixed pretrained model are aligned linearly to the CLIP space, so that text embeddings from CLIP’s text encoder become directly comparable to the aligned features. With text-to-concept, we convert fixed off-the-shelf vision encoders to surprisingly strong zero-shot classifiers for free, with accuracy at times even surpassing that of CLIP, despite being much smaller models and trained on a small fraction of the data compared to CLIP. We show other immediate use-cases of text-to-concept, like building concept bottleneck models with no concept supervision, diagnosing distribution shifts in terms of human concepts, and retrieving images satisfying a set of text-based constraints. Lastly, we demonstrate the feasibility ofconcept-to-text, where vectors in a model’s feature space are decoded by first aligning to the CLIP before being fed to a GPT-based generative model. Our work suggests existing deep models, with presumably diverse architectures and training, represent input samples relatively similarly, and a two-way communication across model representation spaces and to humans (through language) is viable.
Cite this Paper
BibTeX
@InProceedings{pmlr-v202-moayeri23a, title = {Text-To-Concept (and Back) via Cross-Model Alignment}, author = {Moayeri, Mazda and Rezaei, Keivan and Sanjabi, Maziar and Feizi, Soheil}, booktitle = {Proceedings of the 40th International Conference on Machine Learning}, pages = {25037--25060}, year = {2023}, editor = {Krause, Andreas and Brunskill, Emma and Cho, Kyunghyun and Engelhardt, Barbara and Sabato, Sivan and Scarlett, Jonathan}, volume = {202}, series = {Proceedings of Machine Learning Research}, month = {23--29 Jul}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v202/moayeri23a/moayeri23a.pdf}, url = {https://proceedings.mlr.press/v202/moayeri23a.html}, abstract = {We observe that the mapping between an image’s representation in one model to its representation in another can be learned surprisingly well with just a linear layer, even across diverse models. Building on this observation, we proposetext-to-concept, where features from a fixed pretrained model are aligned linearly to the CLIP space, so that text embeddings from CLIP’s text encoder become directly comparable to the aligned features. With text-to-concept, we convert fixed off-the-shelf vision encoders to surprisingly strong zero-shot classifiers for free, with accuracy at times even surpassing that of CLIP, despite being much smaller models and trained on a small fraction of the data compared to CLIP. We show other immediate use-cases of text-to-concept, like building concept bottleneck models with no concept supervision, diagnosing distribution shifts in terms of human concepts, and retrieving images satisfying a set of text-based constraints. Lastly, we demonstrate the feasibility ofconcept-to-text, where vectors in a model’s feature space are decoded by first aligning to the CLIP before being fed to a GPT-based generative model. Our work suggests existing deep models, with presumably diverse architectures and training, represent input samples relatively similarly, and a two-way communication across model representation spaces and to humans (through language) is viable.}}
Endnote
%0 Conference Paper%T Text-To-Concept (and Back) via Cross-Model Alignment%A Mazda Moayeri%A Keivan Rezaei%A Maziar Sanjabi%A Soheil Feizi%B Proceedings of the 40th International Conference on Machine Learning%C Proceedings of Machine Learning Research%D 2023%E Andreas Krause%E Emma Brunskill%E Kyunghyun Cho%E Barbara Engelhardt%E Sivan Sabato%E Jonathan Scarlett%F pmlr-v202-moayeri23a%I PMLR%P 25037--25060%U https://proceedings.mlr.press/v202/moayeri23a.html%V 202%X We observe that the mapping between an image’s representation in one model to its representation in another can be learned surprisingly well with just a linear layer, even across diverse models. Building on this observation, we proposetext-to-concept, where features from a fixed pretrained model are aligned linearly to the CLIP space, so that text embeddings from CLIP’s text encoder become directly comparable to the aligned features. With text-to-concept, we convert fixed off-the-shelf vision encoders to surprisingly strong zero-shot classifiers for free, with accuracy at times even surpassing that of CLIP, despite being much smaller models and trained on a small fraction of the data compared to CLIP. We show other immediate use-cases of text-to-concept, like building concept bottleneck models with no concept supervision, diagnosing distribution shifts in terms of human concepts, and retrieving images satisfying a set of text-based constraints. Lastly, we demonstrate the feasibility ofconcept-to-text, where vectors in a model’s feature space are decoded by first aligning to the CLIP before being fed to a GPT-based generative model. Our work suggests existing deep models, with presumably diverse architectures and training, represent input samples relatively similarly, and a two-way communication across model representation spaces and to humans (through language) is viable.
APA
Moayeri, M., Rezaei, K., Sanjabi, M. & Feizi, S.. (2023). Text-To-Concept (and Back) via Cross-Model Alignment.Proceedings of the 40th International Conference on Machine Learning, inProceedings of Machine Learning Research 202:25037-25060 Available from https://proceedings.mlr.press/v202/moayeri23a.html.