期刊论文详细信息
IEEE Access
Favorite Video Classification Based on Multimodal Bidirectional LSTM
Miki Haseyama1  Takahiro Ogawa1  Keisuke Maeda1  Yuma Sasaka1 
[1] Graduate School of Information Science and Technology, Hokkaido University, Sapporo, Japan;
关键词: Multimodal fusion;    video classification;    LSTM;    EEG;   
DOI  :  10.1109/ACCESS.2018.2876710
来源: DOAJ
【 摘 要 】

Video classification based on the user's preference (information of what a user likes: WUL) is important for realizing human-centered video retrieval. A better understanding of the rationale of WUL would greatly contribute to the support for successful video retrieval. However, a few studies have shown the relationship between information of what a user watches and WUL. A new method that classifies videos on the basis of WUL using video features and electroencephalogram (EEG) signals collaboratively with a multimodal bidirectional Long Short-Term Memory (Bi-LSTM) network is presented in this paper. To the best of our knowledge, there has been no study on WUL-based video classification using video features and EEG signals collaboratively with LSTM. First, we newly apply transfer learning to the WUL-based video classification since the number of labels (liked or not liked) attached to videos by users is small, and it is difficult to classify videos based on WUL. Furthermore, we conduct a user study for showing that the representation of psychophysiological signals calculated from Bi-LSTM is effective for the WUL-based video classification. Experimental results showed that our deep neural network feature representations can distinguish WUL for each subject.

【 授权许可】

Unknown   

  文献评价指标  
  下载次数:0次 浏览次数:0次