In statistical theory, a pseudolikelihood is an approximation to the joint probability distribution of a collection of random variables. The practical use of this is that it can provide an approximation to the likelihood function of a set of observed data which may either provide a computationally simpler problem for estimation, or may provide a way of obtaining explicit estimates of model parameters.
The pseudolikelihood approach was introduced by Julian Besag in the context of analysing data having spatial dependence.
Given a set of random variables
X=X1,X2,\ldots,Xn
X=x=(x1,x2,\ldots,xn)
L(\theta):=\prodiPr\theta(Xi=xi\midXj=xjforj ≠ i)=\prodiPr\theta(Xi=xi\midX-i=x-i)
in discrete case and
L(\theta):=\prodip\theta(xi\midxjforj ≠ i)=\prodip\theta(xi\midx-i)=\prodip\theta(xi\midx1,\ldots,\hatxi,\ldots,xn)
in continuous one. Here
X
x
p\theta( ⋅ \mid ⋅ )
\theta=(\theta1,\ldots,\thetap)
X=x
Xi
X
xi
x
x-i=(x1,\ldots,\hatxi,\ldots,xn)
xi
Pr\theta(X=x)
X
x
\theta
Pr\theta(X=x)
The pseudo-log-likelihood is a similar measure derived from the above expression, namely (in discrete case)
l(\theta):=logL(\theta)=\sumilogPr\theta(Xi=xi\midXj=xjforj ≠ i).
One use of the pseudolikelihood measure is as an approximation for inference about a Markov or Bayesian network, as the pseudolikelihood of an assignment to
Xi
Use of the pseudolikelihood in place of the true likelihood function in a maximum likelihood analysis can lead to good estimates, but a straightforward application of the usual likelihood techniques to derive information about estimation uncertainty, or for significance testing, would in general be incorrect.[1]