Skip navigation
Skip navigation

Activity Recognition in Videos with Segmented Streams

CollectionsANU Student Research
Title: Activity Recognition in Videos with Segmented Streams
Author(s): Cai, Zixian
Gould, Stephen
Keywords: activity recognition;two-stream
Date published: Jun-2019
Publisher: The Australian National University
We investigate a Convolutional Neural Networks (CNN) architecture for activity recognition in short video clips. Applications are ubiquitous, ranging from guiding unmanned vehicles to captioning video clips. While the employment of CNN architectures on large image datasets (such as ImageNet) has been successfully demonstrated in many prior works, there is still no clear answer as to how one can use adapt CNNs to video data. Several different architectures have been explored such as C3D and two-stream networks. However, they all use RGB frames of the video clips as is. In this work, we introduce segmented streams, where each stream consists of the original RGB frames segmented by motion types. We find that after training on the UCF101 dataset, we are able to improve over the original two-stream work by fusing our segmented streams.


File Description SizeFormat Image
20190614-SCNC2103-Zixian.pdfReport935.56 kBAdobe PDFThumbnail

Items in Open Research are protected by copyright, with all rights reserved, unless otherwise indicated.

Updated:  17 November 2022/ Responsible Officer:  University Librarian/ Page Contact:  Library Systems & Web Coordinator