Abstract
Information Noise Contrastive Estimation (InfoNCE) is a popular neural estimator of mutual information (MI). While InfoNCE has demonstrated impressive results in representation learning, the estimation can be significantly off. While the original estimator is known to underestimate the MI due to the logn upper bound, where n is the sample size, we show that some subsequent fix can cause the MI estimate to overshoot apparently without any bound. We propose a novel MI variational estimator, smoothed InfoNCE, that resolves the issues by smoothing out the contrastive estimation. Experiments on high-dimensional Gaussian data confirm that the proposed estimate can break the logn curse without overshooting. © 2023 The Franklin Institute
Original language | English |
---|---|
Pages (from-to) | 12415-12435 |
Journal | Journal of the Franklin Institute |
Volume | 360 |
Issue number | 16 |
Online published | 10 Sept 2023 |
DOIs | |
Publication status | Published - Nov 2023 |
Research Keywords
- Mutual information
- Neural estimation
- Variational representation
- KullbackLeibler divergence
Publisher's Copyright Statement
- COPYRIGHT TERMS OF DEPOSITED POSTPRINT FILE: © 2023 The Franklin Institute. This manuscript version is made available under the CC-BY-NC-ND 4.0 license https://creativecommons.org/licenses/by-nc-nd/4.0/.