In dual decomposition a problem is broken into smaller subproblems and a solution to the relaxed problem is found. This method can be employed for MRF optimization.[1] Dual decomposition is applied to markov logic programs as an inference technique.[2]
Discrete MRF Optimization (inference) is very important in Machine Learning and Computer vision, which is realized on CUDA graphical processing units.[3] Consider a graph
G=(V,E)
V
E
lp
p\inV
(1)
min\Sigmap\in\thetap(lp)+\Sigmapq\in\thetapq(lp)(lq)
Major MRF Optimization methods are based on Graph cuts or Message passing. They rely on the following integer linear programming formulation
(2)
minxE(\theta,x)=\theta.x=\sump\thetap.xp+\sumpq\thetapq.xpq
In many applications, the MRF-variables are -variables that satisfy:
xp(l)=1
\Leftrightarrow
l
p
xpq(l,l\prime)=1
l,l\prime
p,q
The main idea behind decomposition is surprisingly simple:
A sample problem to decompose:
minx\Sigmaifi(x)
x\inC
In this problem, separately minimizing every single
fi(x)
x
\{xi\}
min | |
\{xi\ |
,x}\Sigmaifi(xi)
xi\inC,xi=x
Now we can relax the constraints by multipliers
\{λi\}
i\}) =min | |
g(\{λ | |
\{xi\inC\ |
,x}\Sigmaifi(xi)+\Sigmaiλi.(x
i-x)=min | |
\{xi\inC\ |
,x}\Sigmai[fi(xi)+λi.x
i]-(\Sigma | |
i |
λi)x
Now we eliminate
x
x
i\})=min | |
g(\{λ | |
\{xi\inC\ |
We can set up a Lagrangian dual problem:
(3)
max | |
\{λi\ |
\inΛ}
i})=\Sigma | |
g({λ | |
i |
gi(xi),
(4)
gi(x
i)=min | |
xi |
fi(xi)+λi.xi
xi\inC
The original MRF optimization problem is NP-hard and we need to transform it into something easier.
\tau
G
T
\tau
\thetaT
xT
\theta,x
\thetaT
(5)
\sumT
T= | |
\theta | |
p |
\thetap,\sumT
T= | |
\theta | |
pq |
\thetapq.
Where
\tau(p)
\tau(pq)
\tau
p
pq
(6)
E(\theta,x)=\sumTE(\thetaT,xT)
And our constraints will be:
(7)
xT\in\chiT,
T=x | |
x | |
|T |
,\forallT\in\tau
Our original MRF problem will become:
(8)
min | |
\{xT\ |
,x}\SigmaTE(\thetaT,xT)
xT\in\chiT,\forallT\in\tau
xT\inx|T,\forallT\in\tau
And we'll have the dual problem we were seeking:
(9)
max | |
\{λT\ |
\inΛ}g(\{λT\})=\sumTgT(λT),
where each function
gT(.)
(10)
gT(λ
T)=min | |
xT |
E(\thetaT+λT,xT)
xT\in\chiT
Theorem 1. Lagrangian relaxation (9) is equivalent to the LP relaxation of (2).
min | |
\{xT\ |
,x}\{E(x,
T | |
\theta)|x | |
p,x |
\inCONVEXHULL(\chiT)\}
Theorem 2. If the sequence of multipliers
\{\alphat\}
\alphat\geq0,\limt\alphat=0,
infin | |
\sum | |
t=0 |
\alphat=infin
Theorem 3. The distance of the current solution
\{\thetaT\}
\{\bar\thetaT\}
Theorem 4. Any solution obtained by the method satisfies the WTA (weak tree agreement) condition.
Theorem 5. For binary MRFs with sub-modular energies, the method computes a globally optimal solution.