Hallucinating IDT descriptors and I3D optical flow features for action recognition with CNNs
Date
Authors
Wang, Lei
Koniusz, Piotr
Huynh, D. Q.
Journal Title
Journal ISSN
Volume Title
Publisher
IEEE, Institute of Electrical and Electronics Engineers
Abstract
In this paper, we revive the use of old-fashioned handcrafted video representations for action recognition and put
new life into these techniques via a CNN-based hallucination step. Despite of the use of RGB and optical flow
frames, the I3D model (amongst others) thrives on combining its output with the Improved Dense Trajectory (IDT)
and extracted with its low-level video descriptors encoded
via Bag-of-Words (BoW) and Fisher Vectors (FV). Such a
fusion of CNNs and handcrafted representations is timeconsuming due to pre-processing, descriptor extraction, encoding and tuning parameters. Thus, we propose an endto-end trainable network with streams which learn the IDTbased BoW/FV representations at the training stage and are
simple to integrate with the I3D model. Specifically, each
stream takes I3D feature maps ahead of the last 1D conv.
layer and learns to ‘translate’ these maps to BoW/FV representations. Thus, our model can hallucinate and use such
synthesized BoW/FV representations at the testing stage.
We show that even features of the entire I3D optical flow
stream can be hallucinated thus simplifying the pipeline.
Our model saves 20–55h of computations and yields stateof-the-art results on four publicly available datasets.
Description
Keywords
Citation
Collections
Source
Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, ICCV 2019
Type
Book Title
Entity type
Access Statement
License Rights
Restricted until
2099-12-31