EntropicAffinity
- class torchdr.EntropicAffinity(perplexity: float = 30, tol: float = 0.001, max_iter: int = 1000, sparsity: bool | str = 'auto', metric: str = 'sqeuclidean', zero_diag: bool = True, device: str = 'auto', keops: bool = False, verbose: bool = False)[source]
Bases:
SparseLogAffinity
Solve the directed entropic affinity problem introduced in [H02].
The algorithm computes the optimal dual variable \(\mathbf{\varepsilon}^* \in \mathbb{R}^n_{>0}\) such that
\[\forall i, \: \mathrm{h}(\mathbf{P}^{\mathrm{e}}_{i:}) = \log (\xi) + 1 \quad \text{where} \quad \forall (i,j), \: P^{\mathrm{e}}_{ij} = \frac{\exp(- C_{ij} / \varepsilon_i^\star)}{\sum_{\ell} \exp(- C_{i\ell} / \varepsilon_i^\star)} \:.\]where :
\(\mathbf{C}\): symmetric pairwise distance matrix between the samples.
\(\xi\): perplexity parameter.
\(\mathrm{h}\): (row-wise) Shannon entropy such that \(\mathrm{h}(\mathbf{p}) = - \sum_{i} p_{i} (\log p_{i} - 1)\).
\(\mathbf{\varepsilon}^*\) is computed by performing one dimensional searches since rows of \(\mathbf{P}^{\mathrm{e}}\) are independent subproblems.
Convex problem. Corresponds to the matrix \(\mathbf{P}^{\mathrm{e}}\) in [V23], solving the convex optimization problem
\[\begin{split}\mathbf{P}^{\mathrm{e}} \in \mathop{\arg\min}_{\mathbf{P} \in \mathbb{R}_+^{n \times n}} \: &\langle \mathbf{C}, \mathbf{P} \rangle \\ \text{s.t.} \quad &\mathbf{P} \mathbf{1} = \mathbf{1} \\ &\forall i, \: \mathrm{h}(\mathbf{P}_{i:}) \geq \log (\xi) + 1 \:.\end{split}\]where \(\mathbf{1} := (1,...,1)^\top\): is the all-ones vector.
The entropic affinity matrix is akin to a soft \(k\) -NN affinity, with the perplexity parameter \(\xi\) acting as \(k\). Each point distributes a unit mass among its closest neighbors while minimizing a transport cost given by \(\mathbf{C}\).
The entropic constraint is saturated at the optimum and governs mass spread. With small \(\xi\), mass concentrates on a few neighbors; with large \(\xi\), it spreads across more neighbors thus capturing larger scales of dependencies.
Note
A symmetric version is also available at
SymmetricEntropicAffinity
. It is the affinity matrix used inSNEkhorn
/TSNEkhorn
[V23]. In TSNE [V08], the entropic affinity is simply averaged with its transpose.- Parameters:
perplexity (float, optional) – Perplexity parameter, related to the number of ‘effective’ nearest neighbors. Consider selecting a value between 2 and the number of samples.
tol (float, optional) – Precision threshold at which the root finding algorithm stops.
max_iter (int, optional) – Number of maximum iterations for the root finding algorithm.
sparsity (bool or 'auto', optional) – If True, keeps only the 3 * perplexity smallest element on each row of the ground cost matrix. Recommended if perplexity is not too big.
metric (str, optional) – Metric to use for computing distances (default “sqeuclidean”).
zero_diag (bool, optional) – Whether to set the diagonal of the distance matrix to 0.
device (str, optional) – Device to use for computation.
keops (bool, optional) – Whether to use KeOps for computation.
verbose (bool, optional) – Verbosity. Default is False.
References
Examples using EntropicAffinity
:
Entropic Affinities can adapt to varying noise levels
Neighbor Embedding on genomics & equivalent affinity matcher formulation