学位论文详细信息
Binary classification with training under both classes
binary hypothesis testing;countably infinite alphabet;binary classification;"Steins lemma";finite sample performance
Li, Yun ; Veeravalli ; Venugopal V.
关键词: binary hypothesis testing;    countably infinite alphabet;    binary classification;    "Steins lemma";    finite sample performance;   
Others  :  https://www.ideals.illinois.edu/bitstream/handle/2142/34344/Li_Yun.pdf?sequence=1&isAllowed=y
美国|英语
来源: The Illinois Digital Environment for Access to Learning and Scholarship
PDF
【 摘 要 】

This thesis focuses on the binary classification problem with training data under both classes. We first review binary hypothesis testing problems and present a new result on the case of countably infinite alphabet. The goal of binary hypothesis testing is to decide between the two underlying probabilistic processes. Asymptotic optimality of binary hypothesis testing can be achieved with the knowledge of only one of the processes. It is also shown that the finite sample performance could improve greatly with additional knowledge of the alternate process. Most previous work focuses on the case where the alphabet is finite. This thesis extends the existing results to the case of countably infinite alphabet. It is proved that, without knowledge of the alternate process, the worst-case performance of any test is arbitrarily bad, even when the alternate process is restricted to be ``far'' in the sense of relative entropy.Binary classification problems arise in applications where a full probabilistic model of either of the processes is absent and pre-classified samples from both of the processes are available. It is known that asymptotic optimality can be achieved with the knowledge of only one pre-classified training sequence. We propose a classification function that depends on both training sequences. Then Stein's lemma for classification is proved using this new classification function. It states that the maximal error exponent under one class is given by the relative entropy between the conditional distributions of the two classes. Our results also shed light on how the classification errors depend on the relative size of the training and test data.It is shown in the simulation results that our classification method outperforms the asymptotically optimal one when the test samples are of limited size.

【 预 览 】
附件列表
Files Size Format View
Binary classification with training under both classes 358KB PDF download
  文献评价指标  
  下载次数:6次 浏览次数:6次