A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2022; you can also visit the original URL.
The file type is application/pdf
.
Cross-Probe BERT for Fast Cross-Modal Search
2022
Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval
Owing to the effectiveness of cross-modal attentions, text-vision BERT models have achieved excellent performance in text-image retrieval. Nevertheless, cross-modal attentions in text-vision BERT models require expensive computation cost when tackling textvision retrieval due to their pairwise input. Therefore, normally, it is impractical for deploying them for large-scale cross-modal retrieval in real applications. To address the inefficiency issue in exiting text-vision BERT models, in this
doi:10.1145/3477495.3531826
fatcat:ciforymk2fcnpmrwcp5rmvu2h4