TY - GEN
T1 - Human Action Recognition in Video via Fused Optical Flow and Moment Features- towards a Hierarchical approach to Complex Scenario Recognition
AU - Clawson, Kathy
AU - Jing, Min
AU - Scotney, Bryan
AU - Wang, Hui
AU - Liu, Jun
N1 - International Conference on Multimedia Modeling ; Conference date: 08-01-2014 Through 10-01-2014
PY - 2014/1/2
Y1 - 2014/1/2
N2 - This paper explores using motion features for human action recognition in video, as the first step towards hierarchical complex event detection for surveillance and security. We compensate for the low resolution and noise, characteristic of many CCTV modalities, by generating optical flow feature descriptors which view motion vectors as a global representation of the scene as opposed to a set of pixel-wise measurements. Specifically, we combine existing optical flow features with a set of moment-based features which not only capture the orientation of motion within each video scene, but incorporate spatial information regarding the relative locations of directed optical flow magnitudes. Our evaluation, using a benchmark dataset, considers their diagnostic capability when recognizing human actions under varying feature set parameterizations and signal-to-noise ratios. The results show that human actions can be recognized with mean accuracy across all actions of 93.3%. Furthermore, we illustrate that precision degrades less in low signal-to -noise images when our moments-based features are utilized.
AB - This paper explores using motion features for human action recognition in video, as the first step towards hierarchical complex event detection for surveillance and security. We compensate for the low resolution and noise, characteristic of many CCTV modalities, by generating optical flow feature descriptors which view motion vectors as a global representation of the scene as opposed to a set of pixel-wise measurements. Specifically, we combine existing optical flow features with a set of moment-based features which not only capture the orientation of motion within each video scene, but incorporate spatial information regarding the relative locations of directed optical flow magnitudes. Our evaluation, using a benchmark dataset, considers their diagnostic capability when recognizing human actions under varying feature set parameterizations and signal-to-noise ratios. The results show that human actions can be recognized with mean accuracy across all actions of 93.3%. Furthermore, we illustrate that precision degrades less in low signal-to -noise images when our moments-based features are utilized.
U2 - 10.1007%2F978-3-319-04117-9_10
DO - 10.1007%2F978-3-319-04117-9_10
M3 - Conference contribution
SN - 978-3-319-04116-2
T3 - Lecture Notes in Computer Science
SP - 104
EP - 115
BT - MultiMedia Modeling: 20th Anniversary International Conference, MMM 2014, Dublin, Ireland, January 6-10, 2014, Proceedings, Part II
ER -