Fixed-point computation explained

Fixed-point computation refers to the process of computing an exact or approximate fixed point of a given function.[1] In its most common form, the given function

f

satisfies the condition to the Brouwer fixed-point theorem: that is,

f

is continuous and maps the unit d-cube to itself. The Brouwer fixed-point theorem guarantees that

f

has a fixed point, but the proof is not constructive. Various algorithms have been devised for computing an approximate fixed point. Such algorithms are used in economics for computing a market equilibrium, in game theory for computing a Nash equilibrium, and in dynamic system analysis.

Definitions

The unit interval is denoted by

E:=[0,1]

, and the unit d-dimensional cube is denoted by

Ed

. A continuous function

f

is defined on

Ed

(from

Ed

to itself). Often, it is assumed that

f

is not only continuous but also Lipschitz continuous, that is, for some constant

L

,

|f(x)-f(y)|\leqL|x-y|

for all

x,y

in

Ed

.

A fixed point of

f

is a point

x

in

Ed

such that

f(x)=x

. By the Brouwer fixed-point theorem, any continuous function from

Ed

to itself has a fixed point. But for general functions, it is impossible to compute a fixed point precisely, since it can be an arbitrary real number. Fixed-point computation algorithms look for approximate fixed points. There are several criteria for an approximate fixed point. Several common criteria are:[2]

\varepsilon>0

, An -residual fixed-point of

f

is a point

x

in

Ed

' such that

|f(x)-x|\leq\varepsilon

, where here

||

denotes the maximum norm. That is, all

d

coordinates of the difference

f(x)-x

should be at most .

\delta>0

, A δ-absolute fixed-point of

f

is a point

x

in

Ed

such that

|x-x0|\leq\delta

, where

x0

is any fixed-point of

f

.

\delta>0

, A δ-relative fixed-point of

f

is a point x in

Ed

such that

|x-x0|/|x0|\leq\delta

, where

x0

is any fixed-point of

f

.

For Lipschitz-continuous functions, the absolute criterion is stronger than the residual criterion: If

f

is Lipschitz-continuous with constant

L

, then

|x-x0|\leq\delta

implies

|f(x)-f(x0)|\leqL\delta

. Since

x0

is a fixed-point of

f

, this implies

|f(x)-x0|\leqL\delta

, so

|f(x)-x|\leq(1+L)\delta

. Therefore, a δ-absolute fixed-point is also an -residual fixed-point with

\varepsilon=(1+L)\delta

.

The most basic step of a fixed-point computation algorithm is a value query: given any

x

in

Ed

, the algorithm is provided with an oracle

\tilde{f}

to

f

that returns the value

f(x)

. The accuracy of the approximate fixed-point depends upon the error in the oracle

\tilde{f}(x)

.

The function

f

is accessible via evaluation queries: for any

x

, the algorithm can evaluate

f(x)

. The run-time complexity of an algorithm is usually given by the number of required evaluations.

Contractive functions

A Lipschitz-continuous function with constant

L

is called contractive if

L<1

; it is called weakly-contractive if

L\le1

. Every contractive function satisfying Brouwer's conditions has a unique fixed point. Moreover, fixed-point computation for contractive functions is easier than for general functions. The first algorithm for fixed-point computation was the fixed-point iteration algorithm of Banach. Banach's fixed-point theorem implies that, when fixed-point iteration is applied to a contraction mapping, the error after

t

iterations is in

O(Lt)

. Therefore, the number of evaluations required for a

\delta

-relative fixed-point is approximately

logL(\delta)=log(\delta)/log(L)=log(1/\delta)/log(1/L)

. Sikorski and Wozniakowski[3] showed that Banach's algorithm is optimal when the dimension is large. Specifically, when

d\geqlog(1/\delta)/log(1/L)

, the number of required evaluations of any algorithm for

\delta

-relative fixed-point is larger than 50% the number of evaluations required by the iteration algorithm. Note that when

L

approaches 1, the number of evaluations approaches infinity. No finite algorithm can compute a

\delta

-absolute fixed point for all functions with

L=1

.[4]

When

L

< 1 and d = 1, the optimal algorithm is the Fixed Point Envelope (FPE) algorithm of Sikorski and Wozniakowski. It finds a δ-relative fixed point using

O(log(1/\delta)+loglog(1/(1-L)))

queries, and a δ-absolute fixed point using

O(log(1/\delta))

queries. This is faster than the fixed-point iteration algorithm.[5]

When

d>1

but not too large, and

L\le1

, the optimal algorithm is the interior-ellipsoid algorithm (based on the ellipsoid method).[6] It finds an -residual fixed-point using

O(dlog(1/\varepsilon))

evaluations. When

L<1

, it finds a

\delta

-absolute fixed point using

O(d[log(1/\delta)+log(1/(1-L))])

evaluations.

Shellman and Sikorski[7] presented an algorithm called BEFix (Bisection Envelope Fixed-point) for computing an -residual fixed-point of a two-dimensional function with '

L\le1

, using only

2\lceillog2(1/\varepsilon)\rceil+1

queries. They later[8] presented an improvement called BEDFix (Bisection Envelope Deep-cut Fixed-point), with the same worst-case guarantee but better empirical performance. When

L<1

, BEDFix can also compute a

\delta

-absolute fixed-point using

O(log(1/\varepsilon)+log(1/(1-L)))

queries.

Shellman and Sikorski presented an algorithm called PFix for computing an -residual fixed-point of a d-dimensional function with L ≤ 1, using

O(logd(1/\varepsilon))

queries. When

L

< 1, PFix can be executed with

\varepsilon=(1-L)\delta

, and in that case, it computes a δ-absolute fixed-point, using

O(logd(1/[(1-L)\delta]))

queries. It is more efficient than the iteration algorithm when

L

is close to 1. The algorithm is recursive: it handles a d-dimensional function by recursive calls on (d-1)-dimensional functions.

Algorithms for differentiable functions

When the function

f

is differentiable, and the algorithm can evaluate its derivative (not only

f

itself), the Newton method can be used and it is much faster.[9] [10]

General functions

For functions with Lipschitz constant

L

> 1, computing a fixed-point is much harder.

One dimension

For a 1-dimensional function (d = 1), a

\delta

-absolute fixed-point can be found using

O(log(1/\delta))

queries using the bisection method: start with the interval

E:=[0,1]

; at each iteration, let

x

be the center of the current interval, and compute

f(x)

; if

f(x)>x

then recurse on the sub-interval to the right of

x

; otherwise, recurse on the interval to the left of

x

. Note that the current interval always contains a fixed point, so after

O(log(1/\delta))

queries, any point in the remaining interval is a

\delta

-absolute fixed-point of

f

Setting

\delta:=\varepsilon/(L+1)

, where

L

is the Lipschitz constant, gives an -residual fixed-point, using

O(log(L/\varepsilon)=log(L)+log(1/\varepsilon))

queries.

Two or more dimensions

For functions in two or more dimensions, the problem is much more challenging. Shellman and Sikorski proved that for any integers d ≥ 2 and

L

> 1, finding a δ-absolute fixed-point of d-dimensional

L

-Lipschitz functions might require infinitely many evaluations. The proof idea is as follows. For any integer T > 1 and any sequence of T of evaluation queries (possibly adaptive), one can construct two functions that are Lipschitz-continuous with constant

L

, and yield the same answer to all these queries, but one of them has a unique fixed-point at (x, 0) and the other has a unique fixed-point at (x, 1). Any algorithm using T evaluations cannot differentiate between these functions, so cannot find a δ-absolute fixed-point. This is true for any finite integer T.

Several algorithms based on function evaluations have been developed for finding an -residual fixed-point

f

, and deforming it towards

f

while following the fixed point. A book by Michael Todd surveys various algorithms developed until 1976.

k>1/\varepsilon

. Each vertex z corresponds to a point z/k in the unit n-cube.

f

(z/k) - z/k; note that the difference is an n-vector.

f

between these two points.In the worst case, the number of function evaluations required by all these algorithms is exponential in the binary representation of the accuracy, that is, in

\Omega(1/\varepsilon)

.

Query complexity

Hirsch, Papadimitriou and Vavasis proved that[17] any algorithm based on function evaluations, that finds an -residual fixed-point of f, requires

\Omega(L'/\varepsilon)

function evaluations, where

L'

is the Lipschitz constant of the function

f(x)-x

(note that

L-1\leqL'\leqL+1

). More precisely:

\Theta(L'/\varepsilon)

.

\Omega((L'/\varepsilon)d-2)

queries and

O((L'/\varepsilon)d)

queries.

The latter result leaves a gap in the exponent. Chen and Deng closed the gap. They proved that, for any d ≥ 2 and

1/\varepsilon>4d

and

L'/\varepsilon>192d3

, the number of queries required for computing an -residual fixed-point is in

\Theta((L'/\varepsilon)d-1)

.

Discrete fixed-point computation

A discrete function is a function defined on a subset of

Zd

(the d-dimensional integer grid). There are several discrete fixed-point theorems, stating conditions under which a discrete function has a fixed point. For example, the Iimura-Murota-Tamura theorem states that (in particular) if

f

is a function from a rectangle subset of

Zd

to itself, and

f

is
hypercubic direction-preserving, then

f

has a fixed point.

Let

f

be a direction-preserving function from the integer cube

\{1,...,n\}d

to itself. Chen and Deng prove that, for any d ≥ 2 and n > 48d, computing such a fixed point requires

\Theta(nd-1)

function evaluations.

Chen and Deng[18] define a different discrete-fixed-point problem, which they call 2D-BROUWER. It considers a discrete function

f

on

\{0,...,n\}2

such that, for every x on the grid,

f

(x) - x is either (0, 1) or (1, 0) or (-1, -1). The goal is to find a square in the grid, in which all three labels occur. The function

f

must map the square

\{0,...,n\}2

to itself, so it must map the lines x = 0 and y = 0 to either (0, 1) or (1, 0); the line x = n to either (-1, -1) or (0, 1); and the line y = n to either (-1, -1) or (1,0). The problem can be reduced to 2D-SPERNER (computing a fully-labeled triangle in a triangulation satisfying the conditions to Sperner's lemma), and therefore it is PPAD-complete. This implies that computing an approximate fixed-point is PPAD-complete even for very simple functions.

Relation between fixed-point computation and root-finding algorithms

Given a function

g

from

Ed

to R, a root of

g

is a point x in

Ed

such that

g

(x)=0. An -root of g is a point x in

Ed

such that

g(x)\leq\varepsilon

.

Fixed-point computation is a special case of root-finding: given a function

f

on

Ed

, define

g(x):=|f(x)-x|

. X is a fixed-point of

f

if and only if x is a root of

g

, and x is an -residual fixed-point of

f

if and only if x is an -root of

g

. Therefore, any root-finding algorithm (an algorithm that computes an approximate root of a function) can be used to find an approximate fixed-point.

The opposite is not true: finding an approximate root of a general function may be harder than finding an approximate fixed point. In particular, Sikorski[19] proved that finding an -root requires

\Omega(1/\varepsilond)

function evaluations. This gives an exponential lower bound even for a one-dimensional function (in contrast, an -residual fixed-point of a one-dimensional function can be found using

O(log(1/\varepsilon))

queries using the bisection method). Here is a proof sketch. Construct a function

g

that is slightly larger than everywhere in

Ed

except in some small cube around some point x0, where x0 is the unique root of

g

. If

g

is Lipschitz continuous with constant

L

, then the cube around x0 can have a side-length of

\varepsilon/L

. Any algorithm that finds an -root of

g

must check a set of cubes that covers the entire

Ed

; the number of such cubes is at least

(L/\varepsilon)d

.

However, there are classes of functions for which finding an approximate root is equivalent to finding an approximate fixed point. One example[20] is the class of functions

g

such that

g(x)+x

maps

Ed

to itself (that is:

g(x)+x

is in

Ed

for all x in

Ed

). This is because, for every such function, the function

f(x):=g(x)+x

satisfies the conditions of Brouwer's fixed-point theorem. X is a fixed-point of

f

if and only if x is a root of

g

, and x is an -residual fixed-point of

f

if and only if x is an -root of

g

. Chen and Deng show that the discrete variants of these problems are computationally equivalent: both problems require

\Theta(nd-1)

function evaluations.

Communication complexity

Roughgarden and Weinstein[21] studied the communication complexity of computing an approximate fixed-point. In their model, there are two agents: one of them knows a function

f

and the other knows a function

g

. Both functions are Lipschitz continuous and satisfy Brouwer's conditions. The goal is to compute an approximate fixed point of the composite function

g\circf

. They show that the deterministic communication complexity is in

\Omega(2d)

.

Further reading

Notes and References

  1. Book: 10.1007/978-3-642-50327-6 . The Computation of Fixed Points and Applications . Lecture Notes in Economics and Mathematical Systems . 1976 . 124 . 978-3-540-07685-8 .
  2. Shellman . Spencer . Sikorski . K. . A recursive algorithm for the infinity-norm fixed point problem . Journal of Complexity . December 2003 . 19 . 6 . 799–834 . 10.1016/j.jco.2003.06.001 . free .
  3. Sikorski . K . Woźniakowski . H . Complexity of fixed points, I . Journal of Complexity . December 1987 . 3 . 4 . 388–405 . 10.1016/0885-064X(87)90008-2 . free .
  4. Book: Sikorski . Krzysztof A. . Optimal Solution of Nonlinear Equations . 2001 . Oxford University Press . 978-0-19-510690-9 .
  5. Book: 10.1007/978-1-4615-9552-6_4 . Fast Algorithms for the Computation of Fixed Points . Robustness in Identification and Control . 1989 . Sikorski . K. . 49–58 . 978-1-4615-9554-0 .
  6. Huang . Z . Khachiyan . L . Sikorski . K . Approximating Fixed Points of Weakly Contracting Mappings . Journal of Complexity . June 1999 . 15 . 2 . 200–213 . 10.1006/jcom.1999.0504 . free .
  7. Shellman . Spencer . Sikorski . K. . A Two-Dimensional Bisection Envelope Algorithm for Fixed Points . Journal of Complexity . June 2002 . 18 . 2 . 641–659 . 10.1006/jcom.2001.0625 . free .
  8. Shellman . Spencer . Sikorski . K. . Algorithm 825: A deep-cut bisection envelope algorithm for fixed points . ACM Transactions on Mathematical Software . September 2003 . 29 . 3 . 309–325 . 10.1145/838250.838255 . 7786886 .
  9. Kellogg . R. B. . Li . T. Y. . Yorke . J. . A Constructive Proof of the Brouwer Fixed-Point Theorem and Computational Results . SIAM Journal on Numerical Analysis . September 1976 . 13 . 4 . 473–483 . 10.1137/0713041 .
  10. Smale . Steve . A convergent process of price adjustment and global newton methods . Journal of Mathematical Economics . July 1976 . 3 . 2 . 107–120 . 10.1016/0304-4068(76)90019-7 .
  11. Scarf . Herbert . The Approximation of Fixed Points of a Continuous Mapping . SIAM Journal on Applied Mathematics . September 1967 . 15 . 5 . 1328–1343 . 10.1137/0115116 .
  12. H. Scarf found the first algorithmic proof: .
  13. Kuhn . Harold W. . 1968 . Simplicial Approximation of Fixed Points . 58762 . Proceedings of the National Academy of Sciences of the United States of America . 61 . 4 . 1238–1242 . 10.1073/pnas.61.4.1238 . 16591723 . 225246 . free .
  14. Merrill . Orin Harrison . 1972 . Applications and Extensions of an Algorithm that Computes Fixed Points of Certain Upper Semi-continuous Point to Set Mappings . . 570461463 .
  15. Eaves . B. Curtis . Homotopies for computation of fixed points . Mathematical Programming . December 1972 . 3-3 . 1 . 1–22 . 10.1007/BF01584975 . 39504380 .
  16. David . Gale . 1979 . The Game of Hex and Brouwer Fixed-Point Theorem . The American Mathematical Monthly . 86 . 10 . 818–827 . 10.2307/2320146 . 2320146 .
  17. Hirsch . Michael D . Papadimitriou . Christos H . Vavasis . Stephen A . Exponential lower bounds for finding Brouwer fix points . Journal of Complexity . December 1989 . 5 . 4 . 379–416 . 10.1016/0885-064X(89)90017-4 . 1727254 .
  18. Chen . Xi . Deng . Xiaotie . On the complexity of 2D discrete fixed point problem . Theoretical Computer Science . October 2009 . 410 . 44 . 4448–4456 . 10.1016/j.tcs.2009.07.052 . 2831759 .
  19. Sikorski . K. . Optimal solution of nonlinear equations satisfying a Lipschitz condition . Numerische Mathematik . June 1984 . 43 . 2 . 225–240 . 10.1007/BF01390124 . 120937024 .
  20. Book: 10.1145/1060590.1060638 . On algorithms for discrete and approximate brouwer fixed points . Proceedings of the thirty-seventh annual ACM symposium on Theory of computing . 2005 . Chen . Xi . Deng . Xiaotie . 323–330 . 1581139608 . 16942881 .
  21. Book: 10.1109/FOCS.2016.32 . On the Communication Complexity of Approximate Fixed Points . 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS) . 2016 . Roughgarden . Tim . Weinstein . Omri . 229–238 . 978-1-5090-3933-3 . 87553 .