Abstract
Estimating the mutual information (MI) by neural networks has achieved significant practical success, especially in representation learning. Recent results further reduced the variance in the neural estimation by training a probabilistic classifier. However, the trained classifier tends to be overly confident about some of its predictions, which results in an overestimated MI that fails to capture the desired representation. To soften the classifier, we propose a novel scheme that smooths the label adaptively according to how extreme the probability estimates are. The resulting MI estimate is unbiased under a mild assumption on the model. Experimental results on MNIST and CIFAR10 datasets confirmed that our method yields better representation and achieves higher classification test accuracy among existing approaches in self-supervised representation learning.
Original language | English |
---|---|
Title of host publication | 2021 IEEE International Symposium on Information Theory |
Subtitle of host publication | Proceedings |
Publisher | IEEE |
Pages | 1035-1040 |
ISBN (Electronic) | 978-1-5386-8209-8 |
ISBN (Print) | 978-1-5386-8210-4 |
DOIs | |
Publication status | Published - 2021 |
Event | 2021 IEEE International Symposium on Information Theory (ISIT 2021) - Virtual, Melbourne, Australia Duration: 12 Jul 2021 → 20 Jul 2021 https://2021.ieee-isit.org/TechnicalProgram.asp https://2021.ieee-isit.org/default.asp |
Publication series
Name | IEEE International Symposium on Information Theory - Proceedings |
---|---|
ISSN (Print) | 2157-8095 |
Conference
Conference | 2021 IEEE International Symposium on Information Theory (ISIT 2021) |
---|---|
Country/Territory | Australia |
City | Melbourne |
Period | 12/07/21 → 20/07/21 |
Internet address |