LSMI-Sinkhorn: Semi-supervised Mutual Information Estimation with Optimal Transport

Springer International Publishing
Publication Type:
Machine Learning and Knowledge Discovery in Databases. Research Track, 2021, 12975 LNAI, pp. 655-670
Issue Date:
Filename Description Size
Liu2021_Chapter_LSMI-SinkhornSemi-supervisedMu.pdfPublished version1.87 MB
Adobe PDF
Full metadata record
Estimating mutual information is an important statistics and machine learning problem. To estimate the mutual information from data, a common practice is preparing a set of paired samples {(xi,yi)}i=1n ∼ i. i. d. p(x, y). However, in many situations, it is difficult to obtain a large number of data pairs. To address this problem, we propose the semi-supervised Squared-loss Mutual Information (SMI) estimation method using a small number of paired samples and the available unpaired ones. We first represent SMI through the density ratio function, where the expectation is approximated by the samples from marginals and its assignment parameters. The objective is formulated using the optimal transport problem and quadratic programming. Then, we introduce the Least-Squares Mutual Information with Sinkhorn (LSMI-Sinkhorn) algorithm for efficient optimization. Through experiments, we first demonstrate that the proposed method can estimate the SMI without a large number of paired samples. Then, we show the effectiveness of the proposed LSMI-Sinkhorn algorithm on various types of machine learning problems such as image matching and photo album summarization. Code can be found at
Please use this identifier to cite or link to this item: