University of Cambridge > Talks.cam > ASIF: Coupled Data Turns Unimodal Models to Multimodal Without Training > ASIF: Coupled Data Turns Unimodal Models to Multimodal Without Training

ASIF: Coupled Data Turns Unimodal Models to Multimodal Without Training

Download to your calendar using vCal

If you have a question about this talk, please contact Pietro Barbiero .

CLIP proved that aligning visual and language spaces is key to solving many vision tasks without explicit training, but required to train image and text encoders from scratch on a huge dataset. LiT improved this by only training the text encoder and using a pre-trained vision network. In this talk, we will present the ASIF construction, showing that a common space can be created without any training at all, using single-domain encoders (trained with or without supervision) and a much smaller amount of image-text pairs.

Then, we will discuss the unique properties of ASIF . Most notably, deploying a new version with updated training samples can be done in a matter of seconds. Additionally, the representations in the common space are easily interpretable as every dimension corresponds to the similarity of the input to a unique entry in the multimodal dataset. We will look at experiments on standard zero-shot visual benchmarks that demonstrate the typical transfer ability of image-text models. Overall, ASIF represents a simple yet surprisingly strong baseline for foundation multi-modal models, raising important questions on their data efficiency and on the role of retrieval in machine learning.

This talk is part of the ASIF: Coupled Data Turns Unimodal Models to Multimodal Without Training series.

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

Š 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity