期刊论文详细信息
Entropy
Bayesian and Quasi-Bayesian Estimators for Mutual Information from Discrete Data
Evan Archer1  Il Memming Park2 
[1] Institute for Computational Engineering Sciences, The University of Texas at Austin, Austin, TX 78712, USA;Center for Perceptual Systems, The University of Texas at Austin, Austin, TX 78712, USA
关键词: mutual information;    entropy;    Dirichlet distribution;    Bayes least squares;   
DOI  :  10.3390/e15051738
来源: mdpi
PDF
【 摘 要 】

Mutual information (MI) quantifies the statistical dependency between a pair of random variables, and plays a central role in the analysis of engineering and biological systems. Estimation of MI is difficult due to its dependence on an entire joint distribution, which is difficult to estimate from samples. Here we discuss several regularized estimators for MI that employ priors based on the Dirichlet distribution. First, we discuss three “quasi-Bayesian” estimators that result from linear combinations of Bayesian estimates for conditional and marginal entropies. We show that these estimators are not in fact Bayesian, and do not arise from a well-defined posterior distribution and may in fact be negative. Second, we show that a fully Bayesian MI estimator proposed by Hutter (2002), which relies on a fixed Dirichlet prior, exhibits strong prior dependence and has large bias for small datasets. Third, we formulate a novel Bayesian estimator using a mixture-of-Dirichlets prior, with mixing weights designed to produce an approximately flat prior over MI. We examine the performance of these estimators with a variety of simulated datasets and show that, surprisingly, quasi-Bayesian estimators generally outperform our Bayesian estimator. We discuss outstanding challenges for MI estimation and suggest promising avenues for future research.

【 授权许可】

CC BY   
© 2013 by the authors; licensee MDPI, Basel, Switzerland.

【 预 览 】
附件列表
Files Size Format View
RO202003190036389ZK.pdf 1876KB PDF download
  文献评价指标  
  下载次数:4次 浏览次数:1次