期刊论文详细信息
EPJ Data Science
Predicting and explaining behavioral data with structured feature space decomposition
  1    1    2 
[1] 0000 0001 2156 6853, grid.42505.36, USC Information Sciences Institute, Marina del Rey, USA;0000 0001 2156 6853, grid.42505.36, USC Information Sciences Institute, Marina del Rey, USA;0000 0004 1792 6846, grid.35030.35, City University of Hong Kong, Kowloon Tong, Hong Kong;
关键词: Computational social science;    Empirical studies;    Online social networks;    Human behavior;    Feature selection;   
DOI  :  10.1140/epjds/s13688-019-0201-0
来源: publisher
PDF
【 摘 要 】

Modeling human behavioral data is challenging due to its scale, sparseness (few observations per individual), heterogeneity (differently behaving individuals), and class imbalance (few observations of the outcome of interest). An additional challenge is learning an interpretable model that not only accurately predicts outcomes, but also identifies important factors associated with a given behavior. To address these challenges, we describe a statistical approach to modeling behavioral data called the structured sum-of-squares decomposition (S3D). The algorithm, which is inspired by decision trees, selects important features that collectively explain the variation of the outcome, quantifies correlations between the features, and bins the subspace of important features into smaller, more homogeneous blocks that correspond to similarly-behaving subgroups within the population. This partitioned subspace allows us to predict and analyze the behavior of the outcome variable both statistically and visually, giving a medium to examine the effect of various features and to create explainable predictions. We apply S3D to learn models of online activity from large-scale data collected from diverse sites, such as Stack Exchange, Khan Academy, Twitter, Duolingo, and Digg. We show that S3D creates parsimonious models that can predict outcomes in the held-out data at levels comparable to state-of-the-art approaches, but in addition, produces interpretable models that provide insights into behaviors. This is important for informing strategies aimed at changing behavior, designing social systems, but also for explaining predictions, a critical step towards minimizing algorithmic bias.

【 授权许可】

CC BY   

【 预 览 】
附件列表
Files Size Format View
RO201910107349317ZK.pdf 2626KB PDF download
  文献评价指标  
  下载次数:11次 浏览次数:48次