- Authors: Boris Meinardus, Anil Batra, Anna Rohrbach, Marcus Rohrbach
- Paper: arxiv
We introduce Mr. BLIP (Mr. as in Moment Retrieval), a multimodal, single-stage model that requires no expensive video-language pretraining, no additional input signal (e.g., no transcript or audio) and has a simpler and more versatile design than prior state-of-the-art methods. We achieve a new state-of-the-art in moment retrieval on the widely used benchmarks Charades-STA, QVHighlights, and ActivityNet Captions. Notably, we attain over 9% (absolute) higher Recall (at 0.5 and 0.7 IoU) on the challenging long-video multi-moment QVHighlights benchmark.
# data & data preprocessing
./mr_BLIP_data
# pretrained checkpoints
./mr_BLIP_checkpoints
# mr_BLIP code
./lavis/
# running scripts for mr_BLIP training and inference
./run_scripts
- (Optional) Creating conda environment
conda create -n mrBlip python=3.8
conda activate mrBlip
- build from source
pip install -r requirements.txt
We train Mr. BLIP on QVHighlights, Charades-STA, and ActivityNet Captions and provide the checkpoints. Download the checkpoints and put them under /mr_BLIP_checkpoints.
We test our model on:
Please download original MR data and preprocess them via our scripts.
We provide Mr. BLIP training and inference script examples as follows.
And please refer to dataset page to customize your data path.
You might want to update the config files for the respective runs to fit on your machine. They are currently set to run on 8 A100-80GB GPUs. You can simply reduce the batch size, reduce the number of frames, or apply a frame level embeddings aggregation (32 frame tokens -> 1 token) to fit on a smaller GPU.
sh run_scripts/mr_BLIP/train/qvh.sh
sh run_scripts/mr_BLIP/train/charades.sh
sh run_scripts/mr_BLIP/train/anet.sh
Should roughly return:
[email protected] | [email protected] | mIoU | [email protected] | [email protected] | |
---|---|---|---|---|---|
Mr. BLIP | 76.16 | 62.63 | 70.32 | 68.50 | 55.06 |
sh run_scripts/mr_BLIP/eval/qvh.sh
Should roughly return:
[email protected] | [email protected] | mIoU | |
---|---|---|---|
Mr. BLIP | 69.31 | 49.29 | 58.63 |
sh run_scripts/mr_BLIP/eval/charades.sh
Should roughly return:
[email protected] | [email protected] | mIoU | |
---|---|---|---|
Mr. BLIP | 53.79 | 35.47 | 51.52 |
sh run_scripts/mr_BLIP/eval/anet.sh
We thank the developers of LAVIS and BLIP-2 for their public code release.
Please cite our paper if you use our models in your works:
@article{meinardus2024surprisingeffectivenessmultimodallarge,
title={The Surprising Effectiveness of Multimodal Large Language Models for Video Moment Retrieval},
author={Boris Meinardus and Anil Batra and Anna Rohrbach and Marcus Rohrbach},
year={2024},
eprint={2406.18113},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2406.18113},
}