File Download

There are no files associated with this item.

Supplementary

Conference Paper: Miles: Visual bert pre-training with injected language semantics for video-text retrieval

TitleMiles: Visual bert pre-training with injected language semantics for video-text retrieval
Authors
KeywordsComputer Vision
Issue Date2022
PublisherOrtra Ltd..
Citation
European Conference on Computer Vision (ECCV), Tel Aviv, Israel, October 23-27, 2022 How to Cite?
AbstractDominant pre-training work for video-text retrieval mainly adopt the 'dual-encoder' architectures to enable efficient retrieval, where two separate encoders are used to contrast global video and text representations, but ignore detailed local semantics. The recent success of image BERT pre-training with masked visual modeling that promotes the learning of local visual context, motivates a possible solution to address the above limitation. In this work, we for the first time investigate masked visual modeling in video-text pre-training with the 'dual-encoder' architecture. We perform Masked visual modeling with Injected LanguagE Semantics (MILES) by employing an extra snapshot video encoder as an evolving 'tokenizer' to produce reconstruction targets for masked video patch prediction. Given the corrupted video, the video encoder is trained to recover text-aligned features of the masked patches via reasoning with the visible regions along the spatial and temporal dimensions, which enhances the discriminativeness of local visual features and the fine-grained cross-modality alignment. Our method outperforms state-of-the-art methods for text-to-video retrieval on four datasets with both zero-shot and fine-tune evaluation protocols. Our approach also surpasses the baseline models significantly on zero-shot action recognition, which can be cast as video-to-text retrieval.
Persistent Identifierhttp://hdl.handle.net/10722/315553

 

DC FieldValueLanguage
dc.contributor.authorGe, Y-
dc.contributor.authorGe, Y-
dc.contributor.authorLiu, X-
dc.contributor.authorWang, JA-
dc.contributor.authorWu, J-
dc.contributor.authorShan, Y-
dc.contributor.authorQie, X-
dc.contributor.authorLuo, P-
dc.date.accessioned2022-08-19T09:00:02Z-
dc.date.available2022-08-19T09:00:02Z-
dc.date.issued2022-
dc.identifier.citationEuropean Conference on Computer Vision (ECCV), Tel Aviv, Israel, October 23-27, 2022-
dc.identifier.urihttp://hdl.handle.net/10722/315553-
dc.description.abstractDominant pre-training work for video-text retrieval mainly adopt the 'dual-encoder' architectures to enable efficient retrieval, where two separate encoders are used to contrast global video and text representations, but ignore detailed local semantics. The recent success of image BERT pre-training with masked visual modeling that promotes the learning of local visual context, motivates a possible solution to address the above limitation. In this work, we for the first time investigate masked visual modeling in video-text pre-training with the 'dual-encoder' architecture. We perform Masked visual modeling with Injected LanguagE Semantics (MILES) by employing an extra snapshot video encoder as an evolving 'tokenizer' to produce reconstruction targets for masked video patch prediction. Given the corrupted video, the video encoder is trained to recover text-aligned features of the masked patches via reasoning with the visible regions along the spatial and temporal dimensions, which enhances the discriminativeness of local visual features and the fine-grained cross-modality alignment. Our method outperforms state-of-the-art methods for text-to-video retrieval on four datasets with both zero-shot and fine-tune evaluation protocols. Our approach also surpasses the baseline models significantly on zero-shot action recognition, which can be cast as video-to-text retrieval.-
dc.languageeng-
dc.publisherOrtra Ltd..-
dc.subjectComputer Vision-
dc.titleMiles: Visual bert pre-training with injected language semantics for video-text retrieval-
dc.typeConference_Paper-
dc.identifier.emailGe, Y: geyixiao831@gmail.com-
dc.identifier.emailGe, Y: geyixiao831@gmail.com-
dc.identifier.emailLuo, P: pluo@hku.hk-
dc.identifier.authorityLuo, P=rp02575-
dc.identifier.hkuros335567-
dc.publisher.placeIsrael-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats