Version 1
: Received: 14 November 2017 / Approved: 15 November 2017 / Online: 15 November 2017 (05:15:43 CET)
How to cite:
Tafaghodi khajavi, N.; Kuh, A. The Quality of the Covariance Selection through Detection Problem and AUC Bounds. Preprints2017, 2017110095. https://doi.org/10.20944/preprints201711.0095.v1
Tafaghodi khajavi, N.; Kuh, A. The Quality of the Covariance Selection through Detection Problem and AUC Bounds. Preprints 2017, 2017110095. https://doi.org/10.20944/preprints201711.0095.v1
Tafaghodi khajavi, N.; Kuh, A. The Quality of the Covariance Selection through Detection Problem and AUC Bounds. Preprints2017, 2017110095. https://doi.org/10.20944/preprints201711.0095.v1
APA Style
Tafaghodi khajavi, N., & Kuh, A. (2017). The Quality of the Covariance Selection through Detection Problem and AUC Bounds. Preprints. https://doi.org/10.20944/preprints201711.0095.v1
Chicago/Turabian Style
Tafaghodi khajavi, N. and Anthony Kuh. 2017 "The Quality of the Covariance Selection through Detection Problem and AUC Bounds" Preprints. https://doi.org/10.20944/preprints201711.0095.v1
Abstract
This paper considers the problem of quantifying the quality of a model selection problem for a graphical model. The model selection problem often uses a distance measure such as the Kulback-Leibler (KL) distance to quantify the quality of the approximation between the original distribution and the model distribution. We extend this work by formulating the problem as a detection problem between the original distribution and the model distribution. In particular, we focus on the covariance selection problem by Dempster, [1], and consider the cases where the distributions are Gaussian distributions. Previous work showed that if the approximation model is a tree, that the optimal tree that minimizes the KL divergence can be found by using the Chow-Liu algorithm [2]. While the algorithm minimizes the KL divergence it does not minimize other measures such as other divergences and the area under the curve (AUC). These measures all depend on the eigenvalues of the correlation approximation measure (CAM). We find expressions for KL divergence, log-likelihood ratio, and AUC as a function of the CAM. Easily computable upper and lower bounds are also found for the AUC. The paper concludes by computing these measures for real and synthetic simulation data.
Keywords
covariance selection; model approximation; detection problem; area under the curve; information divergences
Subject
Computer Science and Mathematics, Computer Science
Copyright:
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.