期刊论文详细信息
NEUROCOMPUTING 卷:112
Transfer learning using a nonparametric sparse topic model
Article
Faisal, Ali1  Gillberg, Jussi1  Leen, Gayle1  Peltonen, Jaakko1 
[1] Aalto Univ, HIIT, Dept Informat & Comp Sci, FI-00076 Aalto, Finland
关键词: Transfer learning;    Latent Dirichlet allocation;    Nonparametric Bayesian inference;    Sparsity;    Small sample size;    Topic models;   
DOI  :  10.1016/j.neucom.2012.12.038
来源: Elsevier
PDF
【 摘 要 】

In many domains data items are represented by vectors of counts; count data arises, for example, in bioinformatics or analysis of text documents represented as word count vectors. However, often the amount of data available from an interesting data source is too small to model the data source well. When several data sets are available from related sources, exploiting their similarities by transfer learning can improve the resulting models compared to modeling sources independently. We introduce a Bayesian generative transfer learning model which represents similarity across document collections by sparse sharing of latent topics controlled by an Indian buffet process. Unlike a prominent previous model, hierarchical Dirichlet process (HDP) based multi-task learning, our model decouples topic sharing probability from topic strength, making sharing of low-strength topics easier. In experiments, our model outperforms the HDP approach both on synthetic data and in first of the two case studies on text collections, and achieves similar performance as the HDP approach in the second case study. (C) 2013 Elsevier B.V. All rights reserved.

【 授权许可】

Free   

【 预 览 】
附件列表
Files Size Format View
10_1016_j_neucom_2012_12_038.pdf 912KB PDF download
  文献评价指标  
  下载次数:4次 浏览次数:1次