Remez algorithm explained

The Remez algorithm or Remez exchange algorithm, published by Evgeny Yakovlevich Remez in 1934, is an iterative algorithm used to find simple approximations to functions, specifically, approximations by functions in a Chebyshev space that are the best in the uniform norm L sense.[1] It is sometimes referred to as Remes algorithm or Reme algorithm.

A typical example of a Chebyshev space is the subspace of Chebyshev polynomials of order n in the space of real continuous functions on an interval, C[''a'', ''b'']. The polynomial of best approximation within a given subspace is defined to be the one that minimizes the maximum absolute difference between the polynomial and the function. In this case, the form of the solution is precised by the equioscillation theorem.

Procedure

The Remez algorithm starts with the function

f

to be approximated and a set

X

of

n+2

sample points

x1,x2,...,xn+2

in the approximation interval, usually the extrema of Chebyshev polynomial linearly mapped to the interval. The steps are:

b0+b1xi+...+bn

n
x
i

+(-1)iE=f(xi)

(where

i=1,2,...n+2

),

for the unknowns

b0,b1...bn

and E.

bi

as coefficients to form a polynomial

Pn

.

M

of points of local maximum error

|Pn(x)-f(x)|

.

m\inM

are of equal magnitude and alternate in sign, then

Pn

is the minimax approximation polynomial. If not, replace

X

with

M

and repeat the steps above.

The result is called the polynomial of best approximation or the minimax approximation algorithm.

A review of technicalities in implementing the Remez algorithm is given by W. Fraser.[2]

Choice of initialization

The Chebyshev nodes are a common choice for the initial approximation because of their role in the theory of polynomial interpolation. For the initialization of the optimization problem for function f by the Lagrange interpolant Ln(f), it can be shown that this initial approximation is bounded by

\lVertf-Ln(f)\rVertinfty\le(1+\lVertLn\rVertinfty)

inf
p\inPn

\lVertf-p\rVert

with the norm or Lebesgue constant of the Lagrange interpolation operator Ln of the nodes (t1, ..., tn + 1) being

\lVertLn\rVertinfty=\overline{Λ}n(T)=max-1λn(T;x),

T being the zeros of the Chebyshev polynomials, and the Lebesgue functions being

λn(T;x)=

n+1
\sum
j=1

\left|lj(x)\right|,lj(x)=\prod\stackrel{i{i\nej}}n

(x-ti)
(tj-ti)

.

Theodore A. Kilgore,[3] Carl de Boor, and Allan Pinkus[4] proved that there exists a unique ti for each Ln, although not known explicitly for (ordinary) polynomials. Similarly,

\underline{Λ}n(T)=min-1λn(T;x)

, and the optimality of a choice of nodes can be expressed as

\overline{Λ}n-\underline{Λ}n\ge0.

For Chebyshev nodes, which provides a suboptimal, but analytically explicit choice, the asymptotic behavior is known as[5]

\overline{Λ}n(T)=

2
\pi

log(n+1)+

2
\pi

\left(\gamma+log

8
\pi

\right)+\alphan

(being the Euler–Mascheroni constant) with

0<\alphan<

\pi
72n2
for

n\ge1,

and upper bound[6]

\overline{Λ}n(T)\le

2
\pi

log(n+1)+1

Lev Brutman[7] obtained the bound for

n\ge3

, and

\hat{T}

being the zeros of the expanded Chebyshev polynomials:

\overline{Λ}n(\hat{T})-\underline{Λ}n(\hat{T})<\overline{Λ}3-

1
6

\cot

\pi
8

+

\pi
64
1
\sin2(3\pi/16)

-

2
\pi

(\gamma-log\pi)0.201.

Rüdiger Günttner[8] obtained from a sharper estimate for

n\ge40

\overline{Λ}n(\hat{T})-\underline{Λ}n(\hat{T})<0.0196.

Detailed discussion

This section provides more information on the steps outlined above. In this section, the index i runs from 0 to n+1.

Step 1: Given

x0,x1,...xn+1

, solve the linear system of n+2 equations

b0+b1xi+...+bn

n
x
i

+(-1)iE=f(xi)

(where

i=0,1,...n+1

),

for the unknowns

b0,b1,...bn

and E.

It should be clear that

(-1)iE

in this equation makes sense only if the nodes

x0,...,xn+1

are ordered, either strictly increasing or strictly decreasing. Then this linear system has a unique solution. (As is well known, not every linear system has a solution.) Also, the solution can be obtained with only

O(n2)

arithmetic operations while a standard solver from the library would take

O(n3)

operations. Here is the simple proof:

Compute the standard n-th degree interpolant

p1(x)

to

f(x)

at the first n+1 nodes and also the standard n-th degree interpolant

p2(x)

to the ordinates

(-1)i

p1(xi)=f(xi),p2(xi)=(-1)i,i=0,...,n.

To this end, use each time Newton's interpolation formula with the divideddifferences of order

0,...,n

and

O(n2)

arithmetic operations.

The polynomial

p2(x)

has its i-th zero between

xi-1

and

xi,i=1,...,n

, and thus no further zeroes between

xn

and

xn+1

:

p2(xn)

and

p2(xn+1)

have the same sign

(-1)n

.

The linear combination

p(x):=p1(x)-p2(x)E

is also a polynomial of degree n and

p(xi)=p1(xi)-p2(xi)E =f(xi)-(-1)iE,    i=0,\ldots,n.

This is the same as the equation above for

i=0,...,n

and for any choice of E.The same equation for i = n+1 is

p(xn+1) =p1(xn+1)-p2(xn+1)E =f(xn+1)-(-1)n+1E

and needs special reasoning: solved for the variable E, it is the definition of E:

E:= 

p1(xn+1)-f(xn+1)
p2(xn+1)+(-1)n

.

As mentioned above, the two terms in the denominator have same sign:E and thus

p(x)\equivb0+b1x+\ldots+

n
b
nx
are always well-defined.

The error at the given n+2 ordered nodes is positive and negative in turn because

p(xi)-f(xi) = -(-1)iE,  i=0,...,n+1.

The theorem of de La Vallée Poussin states that under this condition no polynomial of degree n exists with error less than E. Indeed, if such a polynomial existed, call it

\tildep(x)

, then the difference

p(x)-\tildep(x)=(p(x)-f(x))-(\tildep(x)-f(x))

would still be positive/negative at the n+2 nodes

xi

and therefore have at least n+1 zeros which is impossible for a polynomial of degree n.Thus, this E is a lower bound for the minimum error which can be achieved with polynomials of degree n.

Step 2 changes the notation from

b0+b1x+...+

n
b
nx
to

p(x)

.

Step 3 improves upon the input nodes

x0,...,xn+1

and their errors

\pmE

as follows.

In each P-region, the current node

xi

is replaced with the local maximizer

\bar{x}i

and in each N-region

xi

is replaced with the local minimizer. (Expect

\bar{x}0

at A, the

\bar{x}i

near

xi

, and

\bar{x}n+1

at B.) No high precision is required here,the standard line search with a couple of quadratic fits should suffice. (See [9])

Let

zi:=p(\bar{x}i)-f(\bar{x}i)

. Each amplitude

|zi|

is greater than or equal to E. The Theorem of de La Vallée Poussin and its proof alsoapply to

z0,...,zn+1

with

min\{|zi|\}\geqE

as the newlower bound for the best error possible with polynomials of degree n.

Moreover,

max\{|zi|\}

comes in handy as an obvious upper bound for that best possible error.

Step 4: With

min\{|zi|\}

and

max\{|zi|\}

as lower and upper bound for the best possible approximation error, one has a reliable stopping criterion: repeat the steps until

max\{|zi|\}-min\{|zi|\}

is sufficiently small or no longer decreases. These bounds indicate the progress.

Variants

Some modifications of the algorithm are present on the literature. These include:

External links

Notes and References

  1. E. Ya. Remez, "Sur la détermination des polynômes d'approximation de degré donnée", Comm. Soc. Math. Kharkov 10, 41 (1934);
    "Sur un procédé convergent d'approximations successives pour déterminer les polynômes d'approximation, Compt. Rend. Acad. Sc. 198, 2063 (1934);
    "Sur le calcul effectiv des polynômes d'approximation des Tschebyscheff", Compt. Rend. Acade. Sc. 199, 337 (1934).
  2. 10.1145/321281.321282 . W. . Fraser . A Survey of Methods of Computing Minimax and Near-Minimax Polynomial Approximations for Functions of a Single Independent Variable . J. ACM . 12 . 295–314 . 1965 . 3 . 2736060 . free .
  3. 10.1016/0021-9045(78)90013-8 . T. A. . Kilgore . A characterization of the Lagrange interpolating projection with minimal Tchebycheff norm . J. Approx. Theory . 24 . 273–288 . 1978 . 4 .
  4. 10.1016/0021-9045(78)90014-X . C. . de Boor . A. . Pinkus . Proof of the conjectures of Bernstein and Erdös concerning the optimal nodes for polynomial interpolation . . 24 . 289–303 . 1978 . 4 . free .
  5. F. W. . Luttmann . T. J. . Rivlin . Some numerical experiments in the theory of polynomial interpolation . IBM J. Res. Dev. . 9 . 187–191 . 1965 . 3 . 10.1147/rd.93.0187.
  6. T. Rivlin, "The Lebesgue constants for polynomial interpolation", in Proceedings of the Int. Conf. on Functional Analysis and Its Application, edited by H. G. Garnier et al. (Springer-Verlag, Berlin, 1974), p. 422; The Chebyshev polynomials (Wiley-Interscience, New York, 1974).
  7. 10.1137/0715046 . L. . Brutman . On the Lebesgue Function for Polynomial Interpolation . SIAM J. Numer. Anal. . 15 . 694–704 . 1978 . 4 . 1978SJNA...15..694B .
  8. 10.1137/0717043 . R. . Günttner . Evaluation of Lebesgue Constants . SIAM J. Numer. Anal. . 17 . 512–520 . 1980 . 4 . 1980SJNA...17..512G .
  9. David G. Luenberger: Introduction to Linear and Nonlinear Programming, Addison-Wesley Publishing Company 1973.
  10. Temes, G.C.; Barcilon, V.; Marshall, F.C. (1973). "The optimization of bandlimited systems". Proceedings of the IEEE. 61 (2): 196–234. doi:10.1109/PROC.1973.9004. ISSN 0018-9219.
  11. Dunham . Charles B. . 1975 . Convergence of the Fraser-Hart algorithm for rational Chebyshev approximation . Mathematics of Computation . en . 29 . 132 . 1078–1082 . 10.1090/S0025-5718-1975-0388732-9 . 0025-5718. free .