Estimating Mutual Information with the Deep Variational Symmetric Information Bottleneck

ORAL

Abstract

Mutual Information (MI) captures nonlinear statistical relations between two variables. MI has proved to be useful in analysis of complex systems, in methods involving clustering, feature selection, and dimensionality reduction, among others. Estimation of MI between high-dimensional variables is a challenging task, often requiring impractically large sample sizes for accurate estimation, and thus limiting wider adoption of methods. One approach to resolve the sampling issue is to reduce the dimensionality of the variables. However, such a reduction can destroy correlations between the variables. We resolve this problem using the Deep Variational Symmetric Information Bottleneck (DVSIB), which simultaneously compresses the variables X and Y into two corresponding lower dimensional latent variables ZX and ZY, while maximizing the information between the latent variables. The information between ZX and ZY produced by DVSIB can be used as a proxy for the information between X and Y. We demonstrate the effectiveness of this method by assessing its performance on synthetic and real datasets, showcasing its robustness and accuracy. We show that our method can estimate mutual information between two high-dimensional variables in many cases where standard estimators fail.

* NSF Grants Nos. 2010524 and 2014173, and by the Simons Investigator award

Presenters

  • Eslam Abdelaleem

    Emory University

Authors

  • Eslam Abdelaleem

    Emory University

  • K. Michael Martini

    Emory University

  • Ilya M Nemenman

    Emory, Emory University