Section 4.1 General Orthogonal Expansion
The notion of mean square convergence makes senes in a general inner product space.
Definition 4.1.1. Inner Product Space.
A vector space
\(V\) over the reals
\(\bbR\) is called an inner product space if there is a function
\((x, y)\in V\times V\mapsto (x, y)\in \bbR\) such that
-
\((a_{1}x_{1}+a_{2}x_{2},y)=a_{1}(x_{1},y)+a_{2}(x_{2}, y)\) for any \(x_{1}, x_{2}, y\in V\) and any \(a_{1}, a_{2}\in \bbR\text{;}\)
-
\((x, y)=(y, x)\) for any \(x, y\in V\text{;}\)
-
\((x, x) \ge 0\) for any \(x\in V\) and equals \(0\) iff \(x=0\text{.}\)
Note that the first two properties imply
\begin{equation*}
(x, a_{1}y_{1}+a_{2}y_{2})= a_{1}(x, y_{1})+a_{2}(x, y_{2}) \text{ for any $x, y_{1}, y_{2}\in V$.}
\end{equation*}
Due to this and the first property, we say an inner product on a vector over the reals is bilinear.
The space of real valued continuous function on a finite interval
\([a, b]\text{,}\) \(\cC[a, b]\text{,}\) has a natural inner product:
\((f, g) :=\int_{a}^{b} f(x)g(x)\, dx\text{.}\)
Definition 4.1.2. Orthogonal Relation.
Two vectors
\(x, y\) in an inner product space
\(V\) are said to be orthogonal to each other if
\((x, y)=0\text{.}\)
Note that, since
\((y, x)=(x, y)\text{,}\) it follows that if
\((x, y)=0\text{,}\) then
\((y, x)=0\text{.}\) So the orthogonal relation is symmetric in
\(x\) and
\(y\text{.}\)
In the context of
\(\cC[a, b]\text{,}\) two functions
\(f, g\in \cC[a, b]\) are orthogonal in
\(\cC[a, b]\) if
\(\int_{a}^{b} f(x)g(x)\, dx=0\text{.}\) Note that it is important to specify the interval of integration. The orthogonality relation stated earlier says that, when
\(n\ne m\text{,}\) the functions
\(\sin \left(\frac{n \pi x}{l}\right), \sin \left(\frac{m \pi x}{l}\right)\) are orthogonal on
\([0, l]\text{.}\) But these two functions may not be orthogonal on a different interval such as
\([0, l/2]\text{.}\)
Proposition 4.1.3. Basic Properties of an Inner Product Space.
Suppose that \(V\) is an inner product space over the reals \(\bbR\text{.}\) Then the Cauchy-Schwarz inequality holds:
\begin{equation*}
|(x, y)|\le \sqrt{(x, x)} \sqrt{(y, y)} \text{ for all $x, y\in V$.}
\end{equation*}
\(\Vert x \Vert :=\sqrt{(x, x)}\) defines a norm on \(V\text{:}\)
-
\(\Vert x \Vert \ge 0\) for all \(x\in V\) and equals \(0\) iff \(x=0\) in \(V\text{;}\)
-
\(\Vert a x \Vert = |a| \Vert x \Vert\) for all \(x\in V\) and real \(a\in \bbR\text{;}\)
-
\(\Vert x +y \Vert\le \Vert x \Vert + \Vert y \Vert \) for all \(x, y\in V\) (Triangle Inequality).
When \((x, y)=0\text{,}\) namely, when \(x\) is orthogonal to \(y\) in \(V\text{,}\) we also have the Pythagorean relation:
\begin{equation*}
\Vert x+y\Vert^{2} =\Vert x\Vert^{2} + \Vert y\Vert^{2}.
\end{equation*}
Proof.
For any \(x, y\in V\text{,}\) the function \(t\mapsto (x+t y, x+t y)\) in \(t\) is a quadratic function in \(t\) when \(y\ne 0\) due to the bilinear property of the inner product, and is nonnegative. Its minimum is attained at \(t=- (x, y)/(y, y)\text{.}\) Evaluating \((x+t y, x+t y)\) at this \(t=- (x, y)/(y, y)\) gives
\begin{equation*}
- \frac{(x, y)^{2}}{(y, y)}+ (x, x), \text{ which is $\ge 0$.}
\end{equation*}
This proves the Cauchy-Schwarz inequality when \(y\ne 0\text{.}\) But the case of \(y=0\) is trivial.
The triangle inequality follows from the Cauchy-Schwarz inequality by
\begin{align*}
\Vert x +y \Vert^{2} \amp =(x+y, x+y)=(x, x)+2(x, y)+(y, y) \\
\amp \le (x, x)+2\Vert x \Vert \Vert y \Vert +(y, y)
=\left(\Vert x \Vert + \Vert y \Vert \right)^{2}.
\end{align*}
The Pythagorean relation clearly follows from the above line of proof when \((x, y)=0\text{.}\)
In the context of \(\cC[a, b]\text{,}\) the Cauchy-Schwarz inequality takes the form of
\begin{equation*}
\vert \int_{a}^{b} f(x) g(x)\, dx \vert \le \sqrt{ \int_{a}^{b} |f(x)|^{2}\, dx} \sqrt{ \int_{a}^{b} |g(x)|^{2}\, dx}
\end{equation*}
for \(f, g\in \cC[a, b]\text{,}\) and the triangle inequality takes the form of
\begin{equation*}
\sqrt{ \int_{a}^{b} |f(x)+g(x)|^{2}\, dx} \le \sqrt{ \int_{a}^{b} |f(x)|^{2}\, dx} + \sqrt{ \int_{a}^{b} |g(x)|^{2}\, dx}.
\end{equation*}
Exercise 4.1.4. A Set of Functions Orthogonal on \([0, l]\).
Verify that the set of functions \(\left\{ \sin \left(\frac{n \pi x}{l}\right): n\in\mathbb N\right\}\) are mutually orthogonal to each other on \([0, l]\) and that
\begin{equation*}
\Vert \sin \left(\frac{n \pi x}{l}\right)\Vert = \sqrt{\frac l2} \text{ for
$n\in\mathbb N$.}
\end{equation*}
Then show that
\begin{equation*}
\int_{0}^{l} \vert \sum_{n=1}^{N}b_{n} \sin \left(\frac{n \pi x}{l}\right)\vert^{2} \, dx
= \frac l2 \left( \sum_{n=1}^{N} |b_{n}|^{2} \right).
\end{equation*}
Hint.
Use the relation
\(2\sin A \sin B=\cos(A-B)-\cos(A+B)\text{.}\)
Exercise 4.1.5. A Set of Functions Orthogonal on \([-l, l]\).
Verify that the set of functions \(\left\{ 1, \cos \left(\frac{n \pi x}{l}\right), \sin \left(\frac{n \pi x}{l}\right): n\in\mathbb N\right\}\) are mutually orthogonal to each other on \([-l, l]\) and that
\begin{equation*}
\Vert 1 \Vert =\sqrt{2l}, \Vert \cos \left(\frac{n \pi x}{l}\right) \Vert =\Vert \sin \left(\frac{n \pi x}{l}\right)\Vert = \sqrt{l} \text{ for
$n\in\mathbb N$.}
\end{equation*}
Then show that
\begin{align*}
\amp \int_{0}^{2l} \vert a_{0}+\sum_{n=1}^{N}\left(a_{n} \cos \left(\frac{n \pi x}{l}\right) + b_{n} \sin \left(\frac{n \pi x}{l}\right)\right) \vert^{2} \, dx\\
= \amp l\left( 2 |a_{0}|^{2} + \sum_{n=1}^{N}\left(|a_{n}|^{2} + |b_{n}|^{2}\right) \right).
\end{align*}
Hint.
Use the relations
\(2\sin A \sin B=\cos(A-B)-\cos(A+B), 2 \cos A \cos B= \cos(A-B)+\cos(A+B),
2\sin A \cos B=\sin(A+B)-\sin(A-B)\text{.}\)
It is often necessary and productive to work with spaces of complex valued functions, which should be regarded as vector spaces over
\(\bbC\text{.}\) The notion of an inner product can be extended to a vector space over
\(\bbC\text{,}\) with some modification.
Definition 4.1.6. Hermitian Inner Product Space.
A vector space \(V\) over \(\bbC\) is called a Hermitian inner product space if there is a function \((x, y)\in V\times V\mapsto (x, y)\in \bbC\) such that
-
\((a_{1}x_{1}+a_{2}x_{2},y)=a_{1}(x_{1},y)+a_{2}(x_{2}, y)\) for any \(x_{1}, x_{2}, y\in V\) and any \(a_{1}, a_{2}\in \bbC\text{;}\)
-
\((x, y)=\overline{(y, x)}\) for any \(x, y\in V\text{;}\)
-
\((x, x) \) is a nonnegative real number for any \(x\in V\) and equals \(0\) iff \(x=0\text{.}\)
Two vectors
\(x, y\in V\) are (Hermitian) orthogonal if
\((x, y)=0\text{.}\)
Note that the first two properties imply
\begin{equation*}
(x, a_{1}y_{1}+a_{2}y_{2})=\overline{(a_{1}y_{1}+a_{2}y_{2}, x)}= \overline{a_{1}}(x, y_{1})+\overline{a_{2}}(x, y_{2}) \text{ for any $x, y_{1}, y_{2}\in V$.}
\end{equation*}
Note that a Hermitian inner product on a vector space is not bilinear in both variables; it is linear in the first variable, but complex conjugate linear in the second variable.
The Cauchy-Schwarz and triangle inequalities and the notion of norm induced by the inner product extend readily to a Hermitian inner product space.
To distinguish between a Hermitian inner product and an inner product introduced earlier on a vector space over the reals, we will refer to the latter as a Euclidean inner product.
For complex valued functions
\(f, g\) in
\(\cC[a, b]\text{,}\) a natural Hermitian inner product is
\((f, g)=\int_{a}^{b} f(x)\overline{g(x)}\,dx\text{.}\) This is consistent with the inner product introduced earlier on
\(\cC[a, b]\) when
\(f, g\) are real valued.
Exercise 4.1.7. The Orthogonal Family \(\{ e^{i\frac{n\pi x}{l}}: n\in \bbZ\}\) on \([-l, l]\).
Verify that the set of functions \(\{ e^{i\frac{n\pi x}{l}}: n\in \bbZ\}\) are orthogonal on \([-l, l]\) and that
\begin{equation*}
\Vert e^{i\frac{n\pi x}{l}} \Vert= \sqrt{2l} \text{ for all $n\in \bbZ$.}
\end{equation*}
Then show that
\begin{equation*}
\int_{-l}^{l} \vert \sum_{-N}^{N}c_{n} e^{i\frac{n\pi x}{l}}\vert^{2}\, dx
= 2l\left( \sum_{-N}^{N} |c_{n}|^{2}\right).
\end{equation*}
Exercise 4.1.9. Hermitian and Euclidean Inner Product.
Use the set up of \(G(x, y)+iS(x, y)=(x, y)\) for a Hermitian inner product. Verify that
-
\(G(x, y)=G(y, x), S(x,y)=-S(y,x )\) for all \(x, y\in V\text{.}\)
-
\(G(i x, i y)=G(x, y), S(i x, i y)=S(x, y)\) for all \(x, y\in V\text{.}\)
-
\(S(x, y)=G(x, i y)\) for all \(x, y\in V\text{.}\)
-
\(G(x, ix)=0\) for all \(x\in V\text{.}\)
Conversely, if
\(G(x, y)\) is an inner product on
\(V\) as a vector space over the reals, and satisfies
\(G(i x, i y)=G(x, y)\) for all
\(x, y\in V\text{.}\) Then define
\(S(x, y)= G(x, i y)\) and
\((x, y)=G(x, y)+ i S(x, y)\) for all
\(x, y\in V\text{.}\) Verify that this
\((x, y)\) is a Hermitian inner product on
\(V\text{.}\) In other words, a Hermitian inner product is always associated with a Euclidean inner product which is preserved by multiplication by
\(i\text{.}\)
In the following we will not distinguish between a Hermitian inner product and a Euclidean inner product, and will let the context to imply the appropriate one.
Definition 4.1.12. Orthonormal Vectors.
A set of vectors
\(\{\bv_{1}, \bv_{2},\cdots \}\) (finite or infinite) in an inner product space
\(V\) is called an orthonormal set, if any two distinct vectors in this set are orthogonal to each other and each one is a unit vector.
Definition 4.1.13. Fourier Coefficients and Fourier Series.
Let
\(\{\bv_{1}, \bv_{2},\cdots \}\) be a set of orthonormal vectors in an inner product space
\(V\text{.}\) For any vector
\(\bv\in V\text{,}\) define
\(c_{k}=(\bv, \bv_{k})\text{.}\) Then
\(\{c_{k}\}\) are called the Fourier coefficients of
\(\bv\) with respect to this set of orthonormal vectors, and the series
\(\sum_{k} c_{k} \bv_{k}\) is called the Fourier series of
\(\bv\) with respect to this set of orthonormal vectors.
In the above definition the convergence of
\(\sum_{k} c_{k} \bv_{k}\) in the case of an infinite set of orthonormal vectors is not directly addressed; one either needs to show that the series converges or simply assumes it as a formal series at this point. As will be seen soon, it is also appropriate to call this sum the orthogonal projection of
\(\bv\) in the span of this set of orthonormal vectors. Note that in a setting of a set of infinite vectors, we take the span of such a set to mean the
completion of the space of
finite linear combination of vectors from this set, which allows us to make sense of an infinite series of such vectors.
Exercise 4.1.15. Find Fourier Coefficients.
Find the Fourier coefficients of the functions
\(f(x)=1\) and
\(g(x)=\cos x\) with respect to the set of orthogonal functions
\(\left\{ \sin (nx): n\in\mathbb N\right\}\) on
\([0, \pi]\text{.}\)
Exercise 4.1.16. Integral Representation of Fourier Partial Sums.
Let
\begin{equation*}
S_{N}[g](x):=a_0 +\sum_{n=1}^{N}\left[ a_n \cos \left(\frac{n \pi x}{l}\right) + b_n \sin \left(\frac{n \pi x}{l}\right) \right]
\end{equation*}
denote the partial sums of the Fourier series of \(g(x)\) with respect to the set of orthogonal functions \(\left\{ 1, \cos \left(\frac{n \pi x}{l}\right), \sin \left(\frac{n \pi x}{l}\right) : n\in\mathbb N\right\}\) on \([-l, l]\text{,}\) and let
\begin{equation*}
c_{n}=\frac{1}{2l} \int_{-l}^{l} g(x) e^{-\frac{i n\pi x}{l}}\, dx
\end{equation*}
denote the Fourier coefficients of \(g(x)\) with respect to the set of orthogonal functions \(\left\{ e^{\frac{i n\pi x}{l}}\right\}_{n=-N}^{N}\text{.}\) Verify that
\begin{equation*}
S_{N}[g](x)=\sum_{n=-N}^{N} c_{n} e^{\frac{i n\pi x}{l}}=
\frac{1}{2l} \int_{-l}^{l}g(t) D_{N}(x-t)\, dt
\end{equation*}
where
\begin{equation*}
D_{N}(t)= \sum_{-N}^{N} e^{-\frac{i n\pi t}{l}}=\frac{\sin \frac{(N+\frac 12)\pi t}{l}}{\sin\frac {\pi t}{2l}}.
\end{equation*}
Hint.
First need to work out
\begin{align*}
S_{N}[g](x)\amp =\frac{1}{2l} \int_{-l}^{l}g(t) \left(1 + \sum_{n=1}^{N} 2 \cos\left(\frac{n\pi (x-t)}{l}\right)\right)\, dt\\
\amp = \frac{1}{2l} \int_{-l}^{l}g(t) \left( \sum_{n=-N}^{N} e^{\frac{i n\pi (x-t)}{l}} \right)\, dt\text{,}
\end{align*}
then establish
\begin{equation*}
1 + \sum_{n=1}^{N} 2 \cos\left(\frac{n\pi s}{l} \right)= \sum_{n=-N}^{N} e^{\frac{i n\pi s}{l}}
=\frac{\sin \frac{(N+\frac 12)\pi s}{l}}{\sin\frac {\pi s}{2l}}
\end{equation*}
using either the relation \(2\sin \frac{\pi s}{2l} \cos \frac{(N+\frac 12)\pi s}{l} =\sin(\frac{(N+1)\pi s}{l})-\sin(\frac{N\pi s}{l})\) or \(\cos \left(\frac{n\pi s}{l} \right)= \text{Re}(e^{\frac{i n\pi s}{l} })\text{.}\)
Theorem 4.1.17. Best Approximation Property of Fourier Series.
Let \(\{\bv_{1}, \bv_{2},\cdots, \bv_{N} \}\) be a finite set of orthonormal vectors in an inner product space \(V\text{.}\) For any vector \(\bv\in V\text{,}\) let \(\sum_{k=1}^{N} c_{k} \bv_{k}\) be the Fourier series of \(\bv\) with respect to this set of orthonormal vectors. Then
\begin{equation*}
( \bv - \sum_{k=1}^{N} c_{k} \bv_{k}, \bv_{j})=0 \text{ for all $j=1, \cdots, N$}
\end{equation*}
and \(\bv - \sum_{k=1}^{N} c_{k} \bv_{k}\) is orthogonal to every vector in the span of this set of orthonormal vectors. Furthermore,
\begin{equation*}
\Vert \bv \Vert^{2}=\Vert \bv - \sum_{k=1}^{N} c_{k} \bv_{k}\Vert^{2}+ \Vert \sum_{k=1}^{N} c_{k} \bv_{k}\Vert^{2},
\end{equation*}
and
\begin{equation*}
\Vert \bv - \sum_{k=1}^{N} c_{k} \bv_{k}\Vert \le \Vert \bv - \sum_{k=1}^{N} a_{k} \bv_{k}\Vert
\text{ for any coefficients $\{a_{k}\}$,}
\end{equation*}
namely, \(\sum_{k=1}^{N} c_{k} \bv_{k}\) is closest to \(\bv\) among all vectors in the span of this set of orthonormal vectors.
Proof.
The first assertion follows directly from
\begin{equation*}
( \bv - \sum_{k=1}^{N} c_{k} \bv_{k}, \bv_{j})=(\bv, \bv_{j})- \sum_{k=1}^{N} c_{k} (\bv_{k}, \bv_{j})=c_{j}-c_{j}=0
\end{equation*}
using the orthonormal condition \((\bv_{k}, \bv_{j})=\delta_{kj}\text{.}\)
The second assertion follows by using the orthogonality relations above
\begin{align*}
\Vert \bv \Vert^{2} = \amp \left( (\bv - \sum_{k=1}^{N} c_{k} \bv_{k})+ \sum_{k=1}^{N} c_{k} \bv_{k},
(\bv - \sum_{k=1}^{N} c_{k} \bv_{k})+ \sum_{k=1}^{N} c_{k} \bv_{k} \right)\\
= \amp ( \bv - \sum_{k=1}^{N} c_{k} \bv_{k}, \bv - \sum_{k=1}^{N} c_{k} \bv_{k})+
2(\bv - \sum_{k=1}^{N} c_{k} \bv_{k}, \sum_{k=1}^{N} c_{k} \bv_{k})\\
\amp + (\sum_{k=1}^{N} c_{k} \bv_{k},\sum_{k=1}^{N} c_{k} \bv_{k})\\
= \amp \Vert \bv - \sum_{k=1}^{N} c_{k} \bv_{k}\Vert^{2}+ \Vert \sum_{k=1}^{N} c_{k} \bv_{k}\Vert^{2}
\end{align*}
using \((\bv - \sum_{k=1}^{N} c_{k} \bv_{k}, \bv_{k})=0\) for any \(k\)
Set \(\bw = \sum_{k=1}^{N} c_{k} \bv_{k}\text{.}\) Then \((\bv-\bw, \sum_{k=1}^{N} (c_{k}-a_{k}) \bv_{k})=0\text{,}\) so
\begin{align*}
\Vert \bv - \sum_{k=1}^{N} a_{k} \bv_{k}\Vert^{2} \amp = \left(\bv-\bw +( \sum_{k=1}^{N} (c_{k}-a_{k}) \bv_{k}),
\bv-\bw +( \sum_{k=1}^{N} (c_{k}-a_{k}) \bv_{k})\right)\\
\amp = \left(\bv-\bw, \bv-\bw\right)+ \left(\sum_{k=1}^{N} (c_{k}-a_{k}) \bv_{k},\sum_{k=1}^{N} (c_{k}-a_{k}) \bv_{k}\right)\\
\amp = \Vert \bv-\bw\Vert^{2}+\Vert \sum_{k=1}^{N} (c_{k}-a_{k}) \bv_{k}\Vert^{2}\\
\amp \ge \Vert \bv-\bw\Vert^{2},
\end{align*}
with equality iff \(c_{k}-a_{k}=0\) for all \(k, 1\le k \le N\text{.}\)
Definition 4.1.18. Orthogonal Projection.
Let
\(\{\bv_{1}, \bv_{2},\cdots, \bv_{N} \}\) be a finite set of orthonormal vectors in an inner product space
\(V\text{.}\) For any vector
\(\bv\in V\text{,}\) let
\(\sum_{k=1}^{N} c_{k} \bv_{k}\) be the Fourier series of
\(\bv\) with respect to this set of orthonormal vectors. Then
\(\sum_{k=1}^{N} c_{k} \bv_{k}\) is also called the orthogonal projection of
\(\bv\) in the span of this set of orthonormal vectors.
Exercise 4.1.19. Find Orthogonal Projections.
Find the orthogonal projections of the functions
\(f(x)=1\) and
\(g(x)=\cos x\) in the span of the set of orthogonal functions
\(\left\{ \sin (nx): 1\le n \le N\right\}\) on
\([0, \pi]\text{.}\)
Exercise 4.1.20.
Find the orthogonal projection of the function
\(f(x)=x\) in the span of the set of orthogonal functions
\(\left\{1, \cos (nx), \sin (nx): 1\le n \le N\right\}\) on
\([-\pi, \pi]\text{.}\)
Theorem 4.1.21. Besselβs Inequality.
Let \(\{\bv_{1}, \bv_{2},\cdots \}\) be a set of orthonormal vectors in an inner product space \(V\text{.}\) For any vector \(\bv\in V\text{,}\) let \(\sum_{k} c_{k} \bv_{k}\) be the Fourier series of \(\bv\) with respect to this set of orthonormal vectors. Then
\begin{equation*}
\sum_{k} |c_{k}|^{2}\le \Vert \bv\Vert^{2} \quad\text{ (Bessel's inequality)}.
\end{equation*}
Proof.
Take any finite subset \(\{\bv_{1}, \bv_{2},\cdots, \bv_{N}\}\text{.}\) Then we have already proved that
\begin{equation*}
\Vert \bv\Vert^{2}\ge \Vert \sum_{k=1}^{N } c_{k }\bv_{k}\Vert^{2}= \sum_{k=1}^{N} |c_{k}|^{2}.
\end{equation*}
Since this holds for any finite \(N\text{,}\) Besselβs inequality follows immediately.
Corollary 4.1.22. Besselβs inequality for trigonometric Fourier series.
Let
\begin{equation*}
g(x)\sim a_0 +\sum_{n=1}^{\infty}\left[ a_n \cos \left(\frac{n \pi x}{l}\right) + b_n \sin \left(\frac{n \pi x}{l}\right) \right].
\end{equation*}
be the Fourier series of \(g\in \cR[-l, l]\) with respect to the set of orthogonal functions \(\left\{ 1, \cos \left(\frac{n \pi x}{l}\right), \sin \left(\frac{n \pi x}{l}\right) : n\in\mathbb N\right\}\) on \([-l, l]\text{.}\) Then
\begin{align}
\amp \int_{-l}^{l} |g(x)|^{2}\, dx\tag{4.1.1}\\
= \amp \int_{-l}^{l} |g(x)-S_{N}[g](x)|^{2}\, dx+\int_{-l}^{l} |S_{N}[g](x)|^{2}\, dx \tag{4.1.2}\\
= \amp \int_{-l}^{l} |g(x)-S_{N}[g](x)|^{2}\, dx+ 2l |a_{0}|^{2}+l \sum_{n=1}^{N} \left[ |a_{n}|^{2}+|b_{n}|^{2}\right].\tag{4.1.3}
\end{align}
As a consequence the following Besselβs inequality holds
\begin{equation*}
\int_{-l}^{l} |g(x)|^{2}\, dx \ge 2l |a_{0}|^{2}+l \sum_{n=1}^{\infty} \left[ |a_{n}|^{2}+|b_{n}|^{2}\right].
\end{equation*}
Definition 4.1.24. A Complete (or Maximal) Orthonormal Set.
A set of orthonormal vectors in an inner product space is called complete (or maximal) if the only vector orthogonal to each of these vectors is the zero vector.