Temporally-Sensitive Pretraining of Video Encoders for Localization Tasks
TSP
TSP: Temporally-Sensitive Pretraining of Video Encoders for Localization Tasks
This repository holds the source code, pretrained models, and pre-extracted features for the TSP method.
Please cite this work if you find TSP useful for your research.
@article{alwassel2020tsp,
title={TSP: Temporally-Sensitive Pretraining of Video Encoders for Localization Tasks},
author={Alwassel, Humam and Giancola, Silvio and Ghanem, Bernard},
journal={arXiv preprint arXiv:2011.11479},
year={2020}
}
We provide pre-extracted features for ActivityNet v1.3 and THUMOS14 videos. The feature files are saved in H5 format, where we map each video-name
to a features tensor of size N x 512
, where N
is the number of features and 512
is the feature size. Use h5py
python package to read the feature files. Not familiar with H5 files or h5py
? here is a quick start guide.