Try   HackMD

Lecture 22: Orthogonal Polynomials, 3-term recurrence, Jacobi Coeffs.

We defined orthogonal polynomials

{pn} with respect to a weight function
μ
on an interval. We showed that they are an orthogonal basis of
L2
when the interval is finite (Weierstrass), and that they satisfy a three term recurrence
pn+1(x)=(xan)pn(x)bnpn1(x)

for some
an
and
bn>0
called Jacobi coefficients (proof: just expand
xpn(x)
as a linear combination of
p0,,pn+1
and notice that most terms vanish).

We also showed that they have real, distinct roots and that the roots of

pn interlace those of
pn+1
, by recognizing that
pn
is the characteristic polynomial of an
n×n
tridiagonal matrix
Jn
with diagonals
an
and off-diagonals
bn
. This follows by writing the three-term recurrence in matrix form, evaluating the recurrence for
p0,,pn1
at the zeros
x1,,xn
of
pn(x)
, and applying a similarity by the diagonal matrix with entries
cj:=pj
, using further that
bj=cj/cj1
.

Lecture 23: Gauss Quadrature, Weights, Favard's Theorem

Theorem. Let

x1,,xn be the zeros of
pn(x)
the
n
th OP with respect to a measure
μ
. Then there are positive weights
wi
such that
q(x)dμ(x)=iq(xi)wi

for every polynomial
f
of degree at most
2n1
.
Proof. Use the Euclidean algorithm to write
f(x)=q(x)pn(x)+r(x)
for
q,r
of degree at most
n1
. This reduces the problem to
r(x)
, which is easily solved by considering Lagrange interpolants.

We then showed that the weights in the quadrature can be computed by computing eigenvectors of the associated Jacobi matrix.

Favard's theorem is a converse to the fact that every measure generates a sequence of Jacobi coefficients. We proved the theorem for compactly supported measures.

Theorem. Suppose

an and
bn>0
are two bounded sequences. Then there is a unique compactly supported measure
μ
with them as its Jacobi coefficients.
Proof.
Existence. Form the infinite tridiagonal Jacobi matrix
J
with diagonal entries
a0
and
J(n+1,n)=J(n,n+1)=bn
for
n=0,1,
. Then
J
is a bounded operator. Let
μ
be the spectral measure of
J
for the vector
e0
, i.e.:
(e0,Jke0)=xkdμ(x)

for every
k
. Let
vn
be the (unnormalized) iterates of Gram-Schmidt run on the vectors
e0,Je0,J2e0,
. We will show inductively that
vn/vn=en
and
vn=bn
; this is trivially true for
n=0
. For the induction step:
vn+1=Jen(Jen,en)en(Jen,en1)en1=Jenanenbnen1=bnen+1

by the tridiagonal structure of the matrix, and normalizing yields
vn+1/vn+1=en+1
.

Let

ϕn be the orthonormal polynomials with respect to
μ
. We will show inductively that
ϕn(J)e0=en
. By Gram-Schmidt, for the unnormalized monic polynomials
pn
:
pn+1=xϕn(xϕn,ϕn)ϕn(xϕn,ϕn1)ϕn1.
Evaluating this identity at
J
and applying to the vector
e0
, and using that
(ϕm,xkϕn)=(ϕm(J)e0,Jkϕn(J)e0)
for every
m,n
by the spectral theorem:
pn+1(J)e0=Jϕn(J)e0(Jϕn(J)e0,ϕn(J)e0)ϕn(J)e0(Jϕn(J)e0,ϕn1(J)e0)ϕn1(J)e0

=Jen(Jen,en)en(Jen,en1)en1, by induction

Thus, the vectors

ϕn(J)e0 satisfy exactly the same recurrence as the
vn
with the same initial condition
e0
, so we must have
vn/vn=ϕn(J)e0=en
for all
n
.

Thus, the polynomials

ϕn satisfy the recurrence:
bn+1ϕn+1=(xan)ϕnbn1ϕn

which is the normalized version of the standard three term recurrence, as desired.

Uniqueness. Suppose

ν is another measure with the same Jacobi coefficients
an,bn
. Then the orthogonal polynomials
pn
with respect to
μ
are
det(xJn)
where
Jn
is the finite
n×n
tridiagonal matrix with these coefficients. This is a finite submatrix of the infinite
J
above. We then have for every
k

(e0,Jke0)=(e0,Jkke0)=(Qe0,XkQe0)

by the explicit diagonalization
Jk=QXQ
in the previous lecture. But
Qe0=(ϕ0(x1),,ϕ0(xk))T
by the explicit diagonalization in the previous lecture, where
xj
are the zeros of
pk
. Thus, by the quadrature formula:
(Qe0,XkQe0)=jnϕ0(xj)xjkwj=xkdν(x)

by Gauss quadrature. Since
ν
agrees with
μ
in all moments and both measures are compactly supported, we must have
μ=ν
.

Thus, the process of generating OP (and thereby a Jacobi operator) from a measure can be seen as the "inverse" of the spectral theorem, which generates a measure from an operator.

Lecture 24: Chebyshev Series and Projections

The goal of the next few lectures is to understand efficient methods for approximating smooth enough functions on an interval by polynomials. This is one of the tools used in discretizing differential equations so that they can be solved on a computer, and is also useful in many other areas of mathematics.

We introduced the Chebyshev polynomials

Tn(cos(θ))=cos(nθ)
which are orthogonal on
[1,1]
with respect to the weight function
w(x)=(1x2)1/2
. They form an orthogonal basis of
L2([1,1],w(x)dx)
so every
f
in this space may be expanded as
f=n=0anTn(x)

where
an=(Tn,f)/Tn2
and the convergence is in
L2
.

The partial sums of the above series are called Chebyshev projections. We showed that (A) the Chebyshev coefficients

an of a Lipschitz continuous function
f
decay at rate
1/n
(B) those of a function analytic in the Bernstein ellipse
Eρ
around
[1,1]
(i.e., the image of the annulus
ann(ρ1,ρ)
under the Joukowski map
z(z+z1)/2
) decay at rate
ρn
. These imply that the uniform (sup norm) error of the Chebyshev projections is of order
1/n
and
ρn+1
, respectively.

The proof of (A) was by interpreting the Chebyshev coefficients as Fourier coefficients of

f(cos(θ)) on the circle. The proof of (B) was by interpreting them as Laurent coefficients of
f((z+z1)/2)
on an annulus containing the unit circle and shifting the contour in the integral formula for the Laurent coefficients, using analyticity of
f
.

Lecture 25: Chebyshev Interpolation, Hermite Integral Formula

Given a set of distinct points

x0,,xn[1,1] define the
n
th polynomial interpolant of a function
f
as the unique polynomial
f^n
of degree
n
such that
f^(xj)=f(xj)
for all
j=0,,n
; explicitly we have

f^(xj)=jf(xj)j(x)

where
j
are the Lagrange interpolants. The main question is: how well (and for which
xj
) does
ff^n0
?

We showed that taking

xj to be the Chebyshev extreme points (i.e. Lobatto points) yields bounds comparable to the error bounds for Chebyshev series. The key phenomenon enabling this is aliasing, namely that for
xj=cos(πj/n)
:
Tm(xj)=Tk(xj)

whenever
m±k=0(mod 2n)
. Thus, we could easily relate
f^n
to the Chebyshev series of
f
.

We then proved the Hermite Integral Formula: if

p(x) is the interpolant of
f
through
x1,,xn
, then for every
x[1,1]{x1,,xn}
:
f(x)p(x)=12πiγ(x)(s)f(s)sxds

where
(s):=j=1n(sxj),

γ
is any contour containing the unit interval, and
f
is assumed to be analytic inside
γ
. The proof was essentially an application of the residue theorem.

Using this formula, we showed that any choice of distinct interpolation points gives exponentially decaying error in

n when
f
is analytic in the "stadium"
Sα:={z:dist(z,[1,1])α}
for
α>2
. This fails when
f
is only analytic in a smaller region, in which case care must be taken to choose the points.

Using the HIF, we found that Chebyshev interpolants have exponentially decaying error

O(ρn) when
f
is analytic in the Bernstein ellipse
Eρ
(which may be much smaller than the stadium
S2
), matching the error bound for Chebyshev projection. The proof is clearly robust to small perturbations of the points
xj
, and implies essentially the same result for the Lobatto points by interlacing of extreme points and zeros.

Lecture 26: Potential Theory, Lebesgue Constants

Potential Theory

Let's begin with a motivating example. Consider the Runge function. Then, it is observed that the interpolation error diverges exponentially for equidistributed points, and converges exponentially for the Chebyshev zeros. We now develop a framework to explain this phenomenon, which will also explain why the Chebyshev points are so good.

In the last lecture, we saw that the error of polynomial interpolation at

x1,,xn is controlled by the ratio
(x)/(s).

An interpolation scheme consists of a specification of interpolation points for every

n. Letting
un(x):=
, the above quantity may be written as
expn(un(x)un(s))
. For good approximation, we would like
un(x)un(s)<0
as
n
\rightarrow\infty$.

Let us study the limiting objects. For a sequence of measures

μn, we say
μn
converges wealky to
μ
if TODO. For example, equispaced points converge to uniform, and Chebyshev points converge to the arcsin law. (Remark: this is true for all orthogonal polynomials with reasonable weights).

We now define the limiting potential to be

u(t):=,
t[1,1]
. This definition can also be extended to
x
in the interval, though
un
is certainly not continuous there.

Let's calculate the limiting potentials for Chebyshev and equispaced points.

Chebyshev: the limiting potential is equal to

log2+logρ on
Eρ
, including
ρ=1
. So for a function analytic inside
Eρ
, we have
u(x)u(s)<0
.

Equispaced: The limiting potential is TODO.

Lebesgue Constants

We defined Lebesgue constants and showed that the Lebesgue constant for Chebyshev interpolation is

O(logn), so it's never much worse than the best approximation. The proof I presented is from Natanson, Constructive Function Theory Vol III, Chap. 2.