A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2022; you can also visit the original URL.
The file type is application/pdf
.
Data Determines Distributional Robustness in Contrastive Language Image Pre-training (CLIP)
[article]
2022
arXiv
pre-print
Contrastively trained image-text models such as CLIP, ALIGN, and BASIC have demonstrated unprecedented robustness to multiple challenging natural distribution shifts. Since these image-text models differ from previous training approaches in several ways, an important question is what causes the large robustness gains. We answer this question via a systematic experimental investigation. Concretely, we study five different possible causes for the robustness gains: (i) the training set size, (ii)
arXiv:2205.01397v1
fatcat:qrjrygxpwzhupcm26x4omhplbe