Action recognition on continuous video

Document Type

Article

Publication Date

2-1-2021

Abstract

Video action recognition has been a challenging task over the years. The challenge herein is not only due to the complication in increasing information in videos but also the requirement of an efficient method to retain information over a longer-term where human action would take to perform. This paper proposes a novel framework, named as long-term video action recognition (LVAR) to perform generic action classification in the continuous video. The idea of LVAR is introducing a partial recurrence connection to propagate information within every layer of a spatial-temporal network, such as the well-known C3D. Empirically, we show that this addition allows the C3D network to access long-term information, and subsequently improves action recognition performance with videos of different length selected from both UCF101 and miniKinetics datasets. Further confirmation of our approach is strengthened with experiments on untrimmed video from the Thumos14 dataset.

Keywords

Deep learning, Action recognition, Propagate information

Divisions

fsktm

Funders

Fundamental Research Grant Scheme (FRGS) MoHE Grant, from the Ministry of Education Malaysia (FP021-2018A),Postgraduate Research Grant (PPP) Grant, from University of Malaya, Malaysia (PG006-2016A)

Publication Title

Neural Computing and Applications

Volume

33

Issue

4

Publisher

Springer London Ltd

Publisher Location

236 GRAYS INN RD, 6TH FLOOR, LONDON WC1X 8HL, ENGLAND

This document is currently not available here.

Share

COinS