Back to Search Start Over

RoME: Role-aware Mixture-of-Expert Transformer for Text-to-Video Retrieval

Authors :
Satar, Burak
Zhu, Hongyuan
Zhang, Hanwang
Lim, Joo Hwee
Publication Year :
2022
Publisher :
arXiv, 2022.

Abstract

Seas of videos are uploaded daily with the popularity of social channels; thus, retrieving the most related video contents with user textual queries plays a more crucial role. Most methods consider only one joint embedding space between global visual and textual features without considering the local structures of each modality. Some other approaches consider multiple embedding spaces consisting of global and local features separately, ignoring rich inter-modality correlations. We propose a novel mixture-of-expert transformer RoME that disentangles the text and the video into three levels; the roles of spatial contexts, temporal contexts, and object contexts. We utilize a transformer-based attention mechanism to fully exploit visual and text embeddings at both global and local levels with mixture-of-experts for considering inter-modalities and structures' correlations. The results indicate that our method outperforms the state-of-the-art methods on the YouCook2 and MSR-VTT datasets, given the same visual backbone without pre-training. Finally, we conducted extensive ablation studies to elucidate our design choices.<br />Comment: Preprint, under review in TCSVT Journal

Details

Database :
OpenAIRE
Accession number :
edsair.doi.dedup.....492faaa075fa899ec2e94f0a775d5f75
Full Text :
https://doi.org/10.48550/arxiv.2206.12845