IEEE Access | |
Cross-View Action Recognition Based on Hierarchical View-Shared Dictionary Learning | |
Jianhuang Lai1  Huicheng Zheng1  Chengkun Zhang1  | |
[1] School of Data and Computer Science, Sun Yat-sen University, Guangzhou, China; | |
关键词: Cross-view; action recognition; hierarchical transfer learning; feature space transformation; dictionary learning; | |
DOI : 10.1109/ACCESS.2018.2815611 | |
来源: DOAJ |
【 摘 要 】
Recognizing human actions across different views is challenging, since observations of the same action often vary greatly with viewpoints. To solve this problem, most existing methods explore the cross-view feature transfer relationship at video level only, ignoring the sequential composition of action segments therein. In this paper, we propose a novel hierarchical transfer framework, which is based on an action temporal-structure model that contains sequential relationship between action segments at multiple timescales. Thus, it can capture the view invariance of the sequential relationship of segment-level transfer. Additionally, we observe that the original feature distributions under different views differ greatly, leading to view-dependent representations irrelevant to the intrinsic structure of actions. Thus, at each level of the proposed framework, we transform the original feature spaces of different views to a view-shared low-dimensional feature space, and jointly learn a dictionary in this space for these views. This view-shared dictionary captures the common structure of action data across the views and can represent the action segments in a way robust to view changes. Moreover, the proposed method can be kernelized easily, and operate in both unsupervised and supervised cross-view scenarios. Extensive experimental results on the IXMAS and WVU datasets demonstrate superiority of the proposed method over state-of-the-art methods.
【 授权许可】
Unknown