sina.ahmadi@insight-centre.org
Adsorption is a general algorithm framework for transductive learning where the learner is given a small set of labelled examples and large set of unlablled examples.
Goal: Label all the unlabelled examples and also, relabel the labelled examples.
\(G = (V, E, W)\)
\(v \in V, e = (a, b) \in V \times V\)
\(W_{ab} \in \mathbb{R}_+\)
\(n = |V|\)
\(n_l\): number of examples for which we have prior knowledge.
\(n_u\): number of examples to be labelled.
\(n = n_l + n_u\)
\(\mathcal{L} = \{1...m\}\): set of the possible labels
\(m = |\mathcal{L}|\)
Each instance \(v \in V\) is associated with \(Y_v, \hat{Y}_v \in \mathbb{R}_+^m\):
\(Y_v= \begin{bmatrix} Y_{v0} & ... & Y_{vl} & ... & Y_{vm} \end{bmatrix}\) prior knowledge
\(\hat{Y}_v= \begin{bmatrix} \hat{Y}_{v0} & ... & \hat{Y}_{vl} & ... & \hat{Y}_{vm} \end{bmatrix}\) output of the algorithm
\(Y, \hat{Y} \in \mathbb{R}_+^{n \times m}\): the matrices whose rows are \(Y_v, \hat{Y}_v\) respectively.
\(0_d\): all-zeros row vector of dimension \(d\).
The adsorption algorithm can be viewed as a controlled random walk (RW) over \(G\) formalized via three actions:
\(p_v^{inj}, p_v^{cont}, p_v^{abnd} \geq 0\) per vertex \(v \in V\). \(p_v^{inj} + p_v^{cont} + p_v^{abnd} = 1\)
To label any vertex \(v \in V\), we initiate a random-walk starting at \(v\) facing three options:
For unlabelled examples, \(p_v^{inj} = 0\).
By definition, \(W_{v'v} = 0\) if \((v, v') \notin E\).
\(\hat{Y}_v\) for node \(v \in V\) is given by:
Encoding ignorance about the correct label by using a dummy variable \(\mathtt{v} \notin \mathcal{L}\), so \(Y_v, \hat{Y}_v \in \mathbb{R}_+^{m + 1}\)
If the RW is abandoned, then the corresponding labelling vector is zero for all true labels in \(\mathcal{L}\), and an arbitrary value of unit for \(\mathtt{v}\)
$$p_v^{cont} \propto c_v$$ $$p_v^{inj} \propto d_v$$
\(p_v^{cont} = \frac{c_v}{z_v}\) ; \(p_v^{inj} = \frac{d_v}{z_v}\) ; \(p_v^{abnd} = 1 - p_v^{cont} - p_v^{inj}\)

The adsorption algorithm minimizes some objective function Q. Talukdar et al. [1] prove that:
There is no Q such that its local optimal would be the output of the adsorption algorithm.There does not exist a function Q with continuous second partial derivatives such that the Adsorption algorithm converges when gradient of Q is equal to 1.
An objective requiring:

Talukdar, Partha Pratim, and Koby Crammer. "New regularized algorithms for transductive learning." Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, Berlin, Heidelberg, 2009.