Hilbert space

From Wikipedia, the free encyclopedia

Jump to: navigation, search
Hilbert spaces can be used to study the harmonics of vibrating strings.

The mathematical concept of a Hilbert space, named after David Hilbert, generalizes the notion of Euclidean space. It extends the methods of vector algebra from the two-dimensional plane and three-dimensional space to infinite-dimensional spaces. In more formal terms, a Hilbert space is an inner product space — an abstract vector space in which distances and angles can be measured — which is "complete", meaning that if a sequence of vectors is Cauchy, then it converges to some limit in the space.

Hilbert spaces arise naturally and frequently in mathematics, physics, and engineering, typically as infinite-dimensional function spaces. They are indispensable tools in the theories of partial differential equations, quantum mechanics, and signal processing. The recognition of a common algebraic structure within these diverse fields generated a greater conceptual understanding, and the success of Hilbert space methods ushered in a very fruitful era for functional analysis.

Geometric intuition plays an important role in many aspects of Hilbert space theory. An element of a Hilbert space can be uniquely specified by its coordinates with respect to an orthonormal basis, in analogy with Cartesian coordinates in the plane. When that basis is countably infinite, this means that the Hilbert space can also usefully be thought of in terms of infinite sequences that are square-summable. Linear operators on a Hilbert space are likewise fairly concrete objects: in good cases, they are simply transformations that stretch the space by different factors in mutually perpendicular directions.

Contents

[edit] Introduction and history

David Hilbert, the mathematician whose groundbreaking work on the study of integral equations and quadratic forms led to the introduction of the Hilbert space.

Ordinary Euclidean space serves as a model for the notion of a Hilbert space. In Euclidean space, denoted R3, the distance between points and the angle between vectors can be expressed via the dot product, an operation on pairs of vectors whose values are real numbers. Problems from analytic geometry, such as determining whether two lines are orthogonal or finding a point on a given plane closest to the origin, can be expressed and then solved using the dot product.[1] Another important feature of R3 is that it possesses enough structure to support the methods of calculus, because of the existence of certain limits. Hilbert spaces are generalizations of R3 which possess an analog of the dot product (usually called an inner product) and are "complete" in the sense that the limits needed to perform calculus exist.

Prior to the development of Hilbert spaces, other generalizations of R3 were known to mathematicians and physicists. In particular, the idea of an abstract linear space had gained some traction towards the end of the 19th century:[citation needed] this is a space whose elements can be added together and multiplied by scalars (such as real or complex numbers) without necessarily identifying these elements with "geometric" vectors, such as position and momentum vectors in physical systems. Other objects studied by mathematicians at the turn of the 20th century, in particular spaces of sequences (including series) and spaces of functions,[2] can naturally be thought of as linear spaces. Functions, for instance, can be added together or multiplied by constant scalars, and these operations obey the algebraic laws satisfied by addition and scalar multiplication of spatial vectors.

In the first decade of the 20th century, parallel developments led to the introduction of Hilbert spaces. The first of these was the observation, which arose during David Hilbert and Erhard Schmidt's study of integral equations,[3] that two square-integrable real-valued functions f and g on an interval [a,b] have an inner product

\langle f,g \rangle = \int_a^b f(x)g(x)\,dx

which has many of the familiar properties of the Euclidean dot product. In particular, the idea of an orthogonal family of functions has meaning. Schmidt exploited the similarity of this inner product with the usual dot product to prove an analog of the spectral decomposition for an operator of the form

f(x) \mapsto \int_a^b K(x,y) f(y)\, dy

where K is a continuous function symmetric in x and y. The resulting eigenfunction expansion expresses the function K as a series of the form

K(x,y) = \sum_n \lambda_n\varphi_n(x)\varphi_n(y)\,

where the functions φn are orthogonal in the sense that φn,φm〉 = 0 for all nm. However, there are eigenfunction expansions which fail to converge in a suitable sense to a square-integrable function: the missing ingredient, which ensures convergence, is completeness.[4]

The second development was the Lebesgue integral, an alternative to the Riemann integral introduced by Henri Lebesgue in 1904.[5] The Lebesgue integral made it possible to integrate more than just continuous functions. In 1907, Frigyes Riesz and Ernst Sigismund Fischer independently proved that the space L2 of square Lebesgue-integrable functions is complete metric space.[6] As a consequence of the interplay between geometry and completeness, the 19th century results of Joseph Fourier, Friedrich Bessel and Marc-Antoine Parseval on trigonometric series easily carried over to these more general spaces, resulting in a geometrical and analytical apparatus now usually known as the Riesz-Fischer theorem.[7]

Further basic results were proved in the early 20th century. For example, the Riesz representation theorem was independently established by Maurice Fréchet and Frigyes Riesz in 1907.[8] John von Neumann coined the term abstract Hilbert space in his famous work on unbounded Hermitian operators.[9] Von Neumann was perhaps the mathematician who most clearly recognized their importance as a result of his seminal work on the foundations of quantum mechanics,[10] and continued in his work with Eugene Wigner. The name "Hilbert space" was soon adopted by others, for example by Hermann Weyl in his book on quantum mechanics and the theory of groups.[11]

The significance of the concept of a Hilbert space was underlined with the realization that it offers one of the best mathematical formulations of quantum mechanics.[12] In short, the states of a quantum mechanical system are vectors in a certain Hilbert space, the observables are hermitian operators on that space, the symmetries of the system are unitary operators, and measurements are orthogonal projections. The relation between quantum mechanical symmetries and unitary operators provided an impetus for the development of the unitary representation theory of groups, initiated in the 1928 work of Hermann Weyl.[13] On the other hand, in the early 1930s it became clear that certain properties of classical dynamical systems can be analyzed using Hilbert space techniques in the framework of ergodic theory.[14]

[edit] Applications

The orbitals of an electron in a hydrogen atom are eigenfunctions of the energy. The energy is given by the (time-independent) Schrödinger operator acting on a dense subspace of the Hilbert space of square-integrable functions on R3, and its spectrum determines the possible energy levels.

Many of the applications of Hilbert spaces exploit the fact that Hilbert spaces support generalizations of simple geometric concepts like projection and change of basis from their usual finite dimensional setting. In particular, the spectral theory of continuous self-adjoint linear operators on a Hilbert space generalizes the usual spectral decomposition of a matrix, and this often plays a major role in applications of the theory to other areas of mathematics and physics.

[edit] Sturm–Liouville theory

The overtones of a vibrating string. These are eigenfunctions of an associated Sturm–Liouville problem. The eigenvalues 1,1/2,1/3,… form the (musical) harmonic series.

In the theory of ordinary differential equations, spectral methods on a suitable Hilbert space are used to study the behavior of eigenvalues and eigenfunctions of differential equations. For example, the Sturm–Liouville problem arises in the study of the harmonics of waves in a violin string or a drum.[15] The problem is a differential equation of the form

 -\frac{d}{dx}\left[p(x)\frac{dy}{ dx}\right]+q(x)y=\lambda w(x)y

for an unknown function y on an interval [a,b], satisfying general homogeneous Robin boundary conditions

\begin{cases}
\alpha y(a)+\alpha' y'(a)=0\\
\beta y(b) + \beta' y'(b)=0.
\end{cases}

The functions p, q, and w are given in advance, and the problem is to find the function y and constants λ for which the equation has a solution. The problem only has solutions for certain values of λ, called eigenvalues of the system, and this is a consequence of the spectral theorem for compact operators applied to the integral operator defined by the Green's function for the system. Furthermore, another consequence of this general result is that the eigenvalues λ of the system can be arranged in an increasing sequence tending to infinity.[16]

[edit] Partial differential equations

Hilbert spaces provide one candidate for a weak formulation of a partial differential equation.[17]

For example, the Poisson equation −Δu = g with Dirichlet boundary conditions in a bounded domain Ω in R2 has weak formulation to find a function u such that, for all continuously differentiable functions v in Ω vanishing on the boundary:

\int_\Omega \nabla u\cdot\nabla v = \int_\Omega gv.

This can be recast in terms of the Hilbert space \scriptstyle{H^1_0(\Omega)} consisting of functions u such that u, along with its weak partial derivatives, are in L2(Ω), and which vanish on the boundary. The question then reduces to finding u in this space such that for all v in this space

a(u,v) = b(v)

where a is a continuous bilinear form, and b is a continuous linear functional, given respectively by

a(u,v) = \int_\Omega \nabla u\cdot\nabla v,\quad b(v)= \int_\Omega gv.

Since the Poisson equation is elliptic, it follows from Poincaré's inequality that the bilinear form a is coercive. The Lax-Milgram theorem is a geometrical result on Hilbert spaces that ensures the existence and uniqueness of solutions of this equation.

This procedure forms the rudiment of the Galerkin method (a finite element method) for numerical solution of PDEs.[18]

[edit] Ergodic theory

Applications to the field of ergodic theory include the von Neumann mean ergodic theorem.[19] If a dynamical system in a Hilbert space evolves according to unitary transformation, then the mean ergodic theorem says that the long-time average behavior of the system is stable under the process.

A physical consequence of the theorem is the following.[20] Let the function ƒ represent the value of a particular physical experiment on a phase space. Let Pτ be the true state of a mechanical system at time τ. By virtue of Liouville's theorem, the volume form on the phase space is preserved under the flows of P. In actual measurements, the value of ƒ itself is not computed, but rather its temporal means

\frac{1}{t-s}\int_s^t f(P_\tau)d\tau.

The theorem ensures that, for long enough time averages, there is a constant around which the dispersion of the time average of ƒ(P) is small.

[edit] Other

Applications include:[citation needed]

One goal of Fourier analysis is to write a given function as a (possibly infinite) linear combination of given basis functions. This problem can be studied abstractly in Hilbert spaces: every Hilbert space has an orthonormal basis, and every element of the Hilbert space can be written in a unique way as a sum of multiples of these basis elements. The Fourier transform then corresponds to a change of basis.[citation needed]

[edit] Definition and examples

A Hilbert space is a real or complex inner product space that is complete under the norm defined by the inner product \langle\cdot,\cdot\rangle by[21]

\|x\| = \sqrt{\langle x,x \rangle} .

Some authors use slightly different definitions. For example, Kolmogorov and Fomin[22] define a Hilbert space as above but restrict the definition to infinite-dimensional spaces. Silverman, the translator of the book, restricts it even further to separable spaces. A separable, infinite-dimensional Hilbert space is unique up to isomorphism; it is denoted by ℓ2(N), or simply 2, as it can be represented by the Lp space2. Older books and papers sometimes call a Hilbert space a unitary space or a linear space with an inner product, but this terminology has fallen out of use.

In the examples of Hilbert spaces given below, the underlying field of scalars is the complex numbers C, although similar definitions apply to the case in which the underlying field of scalars is the real numbers R.

[edit] Euclidean spaces

Every finite-dimensional inner product space is also a Hilbert space. For example, Cn with the inner product defined by

\langle x, y \rangle = \sum_{k=1}^n x_k\overline{y_k}

where the bar over a complex number denotes its complex conjugate.

[edit] Sequence spaces

Given a set B, the sequence space \ell^2 (commonly pronounced "little ell two") over B is defined by

 \ell^2(B) =\big\{ x : B \xrightarrow{x} \mathbb{C} \text{ and } \sum_{b \in B} \left|x \left(b\right)\right|^2 < \infty \big\}.

This space becomes a Hilbert space with the inner product

\langle x, y \rangle = \sum_{b \in B} x(b)\overline{y(b)}

for all x and y in \ell^2(B). B does not have to be a countable set in this definition, although if B is not countable, the resulting Hilbert space is not separable. Every Hilbert space is isomorphic to one of the form \ell^2(B) for a suitable set B. If B=N, the natural numbers, this space is separable and is simply called \ell^2.

[edit] Lebesgue spaces

Lebesgue spaces are function spaces associated to measure spaces (X, M, μ), where X is a set, M is a σ-algebra of subsets of X, and μ is a countably additive measure on M. For example, Let L2μ(X) be the space of those complex-valued measurable functions on X for which the Lebesgue integral of the square of the absolute value of the function is finite, and where functions are identified if and only if they differ only on a set of measure 0.

The inner product of functions f and g in L2μ(X) is then defined as

\langle f,g\rangle=\int_X f(t) \overline{g(t)} \ d \mu(t)

This integral exists, and the resulting space is complete.[23] The full Lebesgue integral is needed to ensure completeness, however, as not enough functions are Riemann integrable.[24]

[edit] Sobolev spaces

Sobolev spaces, denoted by Hs or W s, 2, are Hilbert spaces. These are a special kind of function space in which differentiation may be performed, but which (unlike other Banach spaces such as the Hölder spaces) retain all of the useful geometrical properties of Hilbert spaces. Because differentiation is permitted, Sobolev spaces are a convenient setting for the theory of partial differential equations.[25] They also form the basis of the theory of direct methods in the calculus of variations.[26]

For s a non-negative integer and Ω ⊂ Rn, the Sobolev space Hs(Ω) contains L2 functions whose weak derivatives of order up to s are also L2. The inner product in Hs(Ω) is

\langle f,g\rangle = \int_\Omega f(x)\bar{g}(x)\,dx + \int_\Omega D f\cdot D\bar{g}(x) + \dots + \int_\Omega D^s f(x)\cdot D^s \bar{g}(x)\, dx

where the dot indicates the dot product in the Euclidean space of partial derivatives of each order. Sobolev spaces can also be defined when s is not an integer.

[edit] Hardy spaces

The Hardy spaces are function spaces, arising in complex analysis and harmonic analysis, consisting entirely of holomorphic functions in a domain.[27] Let U denote the unit disc in the complex plane. Then the Hardy space H2(U) is defined to be the space of holomorphic functions f on U such that the means

M_r(f) = \frac{1}{2\pi}\int_0^{2\pi}|f(re^{i\theta})|^2\,d\theta

remain bounded for r < 1. The norm on this Hardy space is defined by

\|f\|_2 = \lim_{r\to 1} \sqrt{M_r(f)}.

Hardy spaces in the disc are related to Fourier series. A function f is in H2(U) if and only if

f(z) = \sum_{n=0}^\infty a_nz^n

where

\sum_n|a_n|^2 < \infty.

Thus H2(U) consists of those functions which are L2 on the circle, and whose negative frequency Fourier coefficients vanish.

[edit] Direct sums

Two Hilbert spaces H1 and H2 can be combined into another Hilbert space, called the (orthogonal) direct sum, and denoted

H_1\oplus H_2,

consisting of the set of all ordered pairs (x1x2) where xiHi, i = 1,2, and inner product defined by

\langle (x_1,x_2), (y_1,y_2)\rangle_{H_1\oplus H_2} = \langle x_1,y_1\rangle_{H_1} + \langle x_2,y_2\rangle_{H_2}.

More generally, if Hi is a family of Hilbert spaces indexed by iI, then the direct sum of the Hi, denoted

\bigoplus_{i\in I}H_i

consists of the set of all indexed families

x=(x_i\in H_i|i\in I) \in \prod_{i\in I}H_i

in the Cartesian product of the Hi such that

\sum_{i\in I} \|x_i\|^2 < \infty.

The inner product is defined by

\langle x, y\rangle = \sum_{i\in I} \langle x_i, y_i\rangle_{H_i}.

Each of the Hi is included as a closed subspace in the direct sum of all of the Hi. Moreover, the Hi are pairwise orthogonal. Conversely, if there is a system of closed subspaces Vi, iI, in a Hilbert space H which are pairwise orthogonal and whose union is dense in H, then H is canonically isomorphic to the direct sum of Vi. In this case, H is called the internal direct sum of the Vi. A direct sum (internal or external) is also equipped with a family of orthogonal projections Ei onto the ith direct summand Hi. These projections are bounded, self-adjoint, idempotent operators which satisfy the orthogonality condition

E_iE_j = 0,\quad i\not= j.

The spectral theorem for compact self-adjoint operators on a Hilbert space H states that H splits into an orthogonal direct sum of eigenspaces:

H = \bigoplus_{\lambda\in \sigma(A)} E_\lambda,

where the nonzero eigenvalues can be arranged to have decreasing magnitude with at most a single limit point at zero. The direct sum of Hilbert spaces also appears in quantum mechanics as the Fock space of a system containing a variable number of particles, where each Hilbert space in the direct sum corresponds to an additional degree of freedom for the quantum mechanical system. In representation theory, the Peter-Weyl theorem guarantees that any unitary representation of a compact group on a Hilbert space splits as the direct sum of finite-dimensional representations.

[edit] Tensor products

If H1 and H2 are two Hilbert spaces, one associates to every simple tensor product x_1 \otimes x_2 the rank one operator

 x^* \in H_1^* \rightarrow x^*(x_1) \, x_2

from the (continuous) dual H1 to H2. This mapping defined on simple tensors extends to a linear identification between  H_1 \otimes H_2 and the space of finite rank operators from H1 to H2. The finite rank operators are embedded in the Hilbert space HS(H1, H2) of Hilbert-Schmidt operators from H1 to H2. The inner product in HS(H1, H2) is given by

 \langle T_1, T_2 \rangle = \sum_{i \in I} \, \langle T_1 e_i^*, T_2 e_i^* \rangle,

where (e_i^*)_{i \in I} is an arbitrary orthonormal basis of H1. The induced inner product on simple tensors is given by

 \langle x_1 \otimes x_2, \, y_1 \otimes y_2 \rangle = \langle x_1, y_1 \rangle \, \langle x_2, y_2 \rangle.

This formula extends by sesquilinearity to  H_1 \otimes H_2.

The Hilbertian tensor product of H1 and H2 is the Hilbert space obtained by completing  H_1 \otimes H_2 for the metric associated to this inner product. Under the preceding identification, the Hilbertian tensor product of H1 and H2 is isometrically and linearly isomorphic to HS(H1, H2).

An interesting example is provided by the Hilbert space L2([0, 1]). The Hilbertian tensor product of two copies of L2([0, 1]) is isometrically and linearly isomorphic to the space L2([0, 1]2) of square-integrable functions on the square [0, 1]2. This isomorphism sends a simple tensor f_1 \otimes f_2 to the function

 (s, t) \rightarrow f_1(s) \, f_2(t)

on the square.

[edit] Properties

[edit] Pythagorean identity

Two vectors u and v in a Hilbert space H are orthogonal when \langle u, v\rangle = 0. The notation for this is uv. More generally, when S is a subset in H, the notation uS means that u is orthogonal to every element from S.
When u and v are orthogonal, one has

\|u + v\|^2 = \langle u + v, u + v \rangle = \langle u, u \rangle + 2 \, \mathrm{Re} \langle u, v \rangle + \langle v, v \rangle= \|u\|^2 + \|v\|^2.

By induction on n, this is extended to any family u1,...,un of n orthogonal vectors,

\|u_1 + \cdots + u_n\|^2 = \|u_1\|^2 + \cdots + \|u_n\|^2.

[edit] Bessel's inequality

Let ƒ1, …, ƒn be a finite orthonormal system in H, i.e., these vectors are orthogonal and have norm one. For an arbitrary vector x in H, let

y = \sum_{j=1}^n \langle x, f_j \rangle \, f_j.

Then <x, ƒk > = <y, ƒk > for every k = 1, …, n. It follows that xy is orthogonal to each ƒk, hence xy is orthogonal to y. Using the Pythagorean identity twice, one sees that

\|x\|^2 = \|x - y\|^2 + \|y\|^2 \ge \|y\|^2 = \sum_{j=1}^n|\langle x, f_j \rangle|^2.

Let {ƒi}, i ∈ I, be an arbitrary orthonormal system in H. Applying the preceding inequality to every finite subset J of I gives the Bessel inequality

\sum_{i \in I}|\langle x, f_i \rangle|^2 \le \|x\|^2, \quad x \in H

(according to the definition of the sum of an arbitrary family of non-negative real numbers).

[edit] Completeness

Completeness is the key to handling infinite-dimensional examples, such as function spaces, and is required, for instance, for the Riesz representation theorem to hold.[28] It is expressed using a form of the Cauchy criterion for sequences in H: a normed space H is complete if every Cauchy sequence converges with respect to this norm to an element in the space.

A series of vectors Σ uk converges in H when the partial sums converge to an element of H, that is called the sum of the series, and denoted by

\sum_{k=0}^\infty u_k = \lim_n \sum_{k=0}^n u_k \in H.

Completeness is required for the extension of the Pythagorean identity to series. A series Σ uk of orthogonal vectors converges in H  if and only if the series of squares of norms converges, and

\bigl\|\sum_{k=0}^\infty u_k \bigr\|^2 = \sum_{k=0}^\infty \|u_k\|^2.

Furthermore, the sum of a series of orthogonal vectors is permutation invariant (i.e. does not change if the terms are rearranged by permuting the set N of integers).

[edit] Parallelogram identity and polarization

Geometrically, the parallelogram identity asserts that AC2 + BD2 = 2(AB2 + AD2). In words, the sum of the squares of the diagonals is twice the sum of the squares of any two adjacent sides.

By definition, every Hilbert space is also a Banach space. Furthermore, in every Hilbert space the following parallelogram identity holds:

\|u+v\|^2+\|u-v\|^2=2(\|u\|^2+\|v\|^2).

Conversely, every Banach space in which the parallelogram identity holds is a Hilbert space, and the inner product is uniquely determined by the norm by the polarization identity.[29] For real Hilbert spaces, the polarization identity is

\langle u,v\rangle = \frac{1}{4}\left(\|u+v\|^2-\|u-v\|^2\right).

For complex Hilbert spaces, it is

\langle u,v\rangle = \frac{1}{4}\left(\|u+v\|^2-\|u-v\|^2+i\|u+iv\|^2-i\|u-iv\|^2\right).

[edit] Topology

Since an inner product space (such as a Hilbert space) is a normed vector space, it becomes a topological vector space by declaring that the open balls constitute a basis of topology.

This topology is locally convex and, more significantly, uniformly convex by the parallelogram law.[30]

[edit] Best approximation

If C is a non-empty closed convex subset of a Hilbert space H and x a point in H, there exists a unique point y ∈ C which minimizes the distance between x and points in C,

 y \in C, \ \ \ \|x - y\| = \mathrm{dist}(x, C) = \min \{ \|x - z\| : z \in C \}.

This is equivalent to saying that there is a point with minimal norm in the translated convex set D = Cx. The proof consists in showing that every minimizing sequence (dn) ⊂ D is Cauchy (using the parallelogram identity) hence converges (using completeness) to a point in D that has minimal norm. More generally, this holds in any uniformly convex Banach space.

When this result is applied to a closed vector subspace F of H, it can be shown that the point y ∈ F closest to x is characterized by

 y \in F, \ \ x - y \perp F.

This point y is the orthogonal projection of x onto F, and the mapping PF : xy is linear (see below, "Orthogonal complements and projections").
In particular, when F is not equal to H, one can find a non-zero vector v orthogonal to F (select x not in F and v = xy). A very useful criterion is obtained by applying this observation to the closed vector subspace F generated by a subset S of H.

A subset S of H spans a dense vector subspace if (and only if) the vector 0 is the sole vector v ∈ H orthogonal to S.

[edit] Reflexivity

Every Hilbert space is reflexive, i.e., every Hilbert space can be naturally identified with its double dual. In fact, more is true: one has a complete and convenient description of its dual space (the space of all continuous linear functions from the space H into the base field), which is itself a Hilbert space. Indeed, the Riesz representation theorem states that to every element φ of the dual H' there exists one and only one u in H such that

\varphi \left(x\right) = \langle x, u \rangle

for all x in H and the association φu provides an antilinear isomorphism between H and H'. If uφ denotes the vector in H  that represents a continuous linear functional φ, the inner product on the dual H' can be defined by

 \langle \varphi, \psi \rangle = \langle u_\psi, u_\varphi \rangle,

restoring linearity in φ. The norm of φ in the dual space coincides with <φ, φ>1/2 =||uφ||.
The representing vector u is obtained in the following way. When φ ≠ 0, the kernel F = ker φ is a closed vector subspace of H, not equal to H, hence there exists a non-zero vector v orthogonal to F. The vector u is a suitable scalar multiple λv of v. The requirement that φ(v) = <vu> yields

 u = \langle v, v \rangle^{-1} \, \overline{\varphi (v)} \, v.

This correspondence φu is exploited by the bra-ket notation popular in physics. It is common in physics to assume that the inner product, denoted by \langle x| y \rangle, is linear on the right,

\langle x| y \rangle = \langle y, x \rangle.

The result \langle x| y \rangle can be seen as the action of the linear functional  \langle x|  (the bra) on the vector  |y\rangle  (the ket).

[edit] Orthonormal bases

The notion of an orthonormal basis from linear algebra generalizes over to the case of Hilbert spaces.[31] In a Hilbert space H, an orthonormal basis is a family {ek}kB of elements of H satisfying the conditions:

  1. Orthogonality: Every two different elements of B are orthogonal: ek, ej〉= 0 for all k, j in B with kj.
  2. Normalization: Every element of the family has norm 1:||ek|| = 1 for all k in B.
  3. Completeness: The linear span of the family ek, kB, is dense in H.

A system of vectors satisfying the first two conditions basis is called an orthonormal system or an orthonormal set (or an orthonormal sequence if B is countable). It can be proved that such a system is always linearly independent. Completeness of an orthonormal system of vectors of a Hilbert space can be equivalently restated as:

if v, ek〉 = 0 for all kB and some vH then v = 0.

Examples of orthonormal bases include:

  • the set {(1,0,0), (0,1,0), (0,0,1)} forms an orthonormal basis of R3 with the dot product;
  • the sequence {ƒn : nZ} with ƒn(x) = exp(2πinx) forms an orthonormal basis of the complex space L2([0,1]);
  • the family eb : bB with eb(c) = 1 if b = c and 0 otherwise forms an orthonormal basis of ℓ2(B).

Note that in the infinite-dimensional case, an orthonormal basis will not be a basis in the sense of linear algebra; to distinguish the two, the latter basis is also called a Hamel basis. That the span of the basis vectors is dense implies that every vector in the space can be written as the sum of an infinite series, and the orthogonality implies that this decomposition is unique.

[edit] Hilbert dimension

Using Zorn's lemma, one can show that every Hilbert space admits an orthonormal basis; furthermore, any two orthonormal bases of the same space have the same cardinality, called the Hilbert dimension of the space.[32] In detail, if {ek}kB is an orthonormal basis of H, then every element x of H may be written as

x = \sum_{k \in B} \, \langle x, e_k \rangle \, e_k.

Even if B is uncountable, only countably many terms in this sum will be non-zero, and the expression is therefore well-defined. This sum is also called the Fourier expansion of x. If {ek}kB is an orthonormal basis of H, then the map Φ : H  ℓ2(B) defined by Φ(x) = (〈x,ek〉)kB is an isomorphism of Hilbert spaces: it is a bijective linear mapping such that

\langle \Phi \left(x\right), \Phi\left(y\right) \rangle_{\ell^2(B)} = \langle x, y \rangle_H

for all x and y in H. The cardinal number of B is the Hilbert dimension of H.

[edit] Separable spaces

A Hilbert space is separable if and only if it admits a countable orthonormal basis. All infinite-dimensional separable Hilbert spaces are isomorphic to \ell^2. In particular, since all infinite-dimensional separable Hilbert spaces are isomorphic, and since almost all Hilbert spaces used in physics are infinite-dimensional and separable, when physicists talk about "the Hilbert space" or just "Hilbert space", they mean any infinite-dimensional separable one.[citation needed]

[edit] Orthogonal complements and projections

If S is a subset of a Hilbert space H, the set of vectors orthogonal to S is defined by

S^\perp = \left\{ x \in H : \langle x, s \rangle = 0\ \forall s \in S \right\}.

S is a closed subspace of H and so forms itself a Hilbert space. If V is a closed subspace of H, then V is called the orthogonal complement of V. In fact, every x in H can then be written uniquely as x = v + w, with v in V and w in V. Therefore, H is the internal Hilbert direct sum of V and V. The linear operator PV : HH which maps x to v is called the orthogonal projection onto V.

Theorem. The orthogonal projection PV is a self-adjoint linear operator on H of norm ≤ 1 with the property PV2 = PV. Moreover, any self-adjoint linear operator E such that E2 = E is of the form PV, where V is the range of E. For every x in H, PV(x) is the unique element v of V which minimizes the distance||xv||.

This provides the geometrical interpretation of PV(x): it is the best approximation to x by elements of V.[33]

[edit] Operators on Hilbert spaces

[edit] Bounded operators

The continuous linear operators A : H1H2 from a Hilbert space H1 to a second Hilbert space H2 are bounded in the sense that they map bounded sets to bounded sets. The space of such bounded linear operators has a norm, the operator norm given by

\lVert A \rVert = \sup \left\{\,\lVert Ax \rVert : \lVert x \rVert \leq 1\,\right\}.

The sum and the composite of two continuous linear operators is again continuous and linear. For y in H2, the map that sends x ∈ H1 to <Ax, y> is linear and continuous, and according to the Riesz representation theorem can therefore be represented in the form

\langle x, A^* y \rangle = \langle Ax, y \rangle

for some vector Ay in H1. This defines another continuous linear operator A : H2 → H1, the adjoint of A. One can see that A∗∗ = A.

The set L(H) of all continuous linear operators on H, together with the addition and composition operations, the norm and the adjoint operation, is a C*-algebra.

An element A  of L(H) is called self-adjoint or Hermitian if A = A. These operators share many features of the real numbers and are sometimes seen as generalizations of them. If A  is Hermitian and Ax, x〉 ≥ 0 for every x, then A  is called positive. If A  has the form BB  for some B, then A  is positive.

An element A  of L(H) is called normal if AA = A A. In the complex case, if an operator A  on H  is normal and compact, then H  admits an orthonormal basis consisting of eigenvectors of A.

An element U  of L(H) is called unitary if U  is invertible and its inverse is given by U. This can also be expressed by requiring that U  be onto and〈Ux, Uy〉 =〈x, y for all x and y in H. The unitary operators form a group under composition, which can be viewed as the automorphism group of H.

[edit] Unbounded operators

If a linear operator has a closed graph and is defined on all of a Hilbert space, then, by the closed graph theorem in Banach space theory, it is necessarily bounded. However, unbounded operators can be obtained by defining a linear map on a proper subspace of the Hilbert space.

In quantum mechanics, several interesting unbounded operators are defined on a dense subspace of Hilbert space.[34] It is possible to define self-adjoint unbounded operators, and these play the role of the observables in the mathematical formulation of quantum mechanics.

Examples of self-adjoint unbounded operator on the Hilbert space L2(R) are:

  • A suitable extension of the differential operator
 (A f)(x) = i \frac{d}{dx} f(x), \,
where i is the imaginary unit and f is a differentiable function of compact support.
  • The multiplication-by-x operator:
 (B f) (x) = x f(x).\,

These correspond to the momentum and position observables, respectively. Note that neither A nor B is defined on all of H, since in the case of A the derivative need not exist, and in the case of B the product function need not be square integrable. In both cases, the set of possible arguments form dense subspaces of L2(R).

[edit] See also

[edit] Notes

  1. ^ See a multivariable calculus text, such as Stewart (2006, Chapter 9).
  2. ^ Bourbaki 1987
  3. ^ Schmidt 1907.
  4. ^ Titchmarsh 1946, §IX.1
  5. ^ Lebesgue 1904. Further details on the history of integration theory can be found in Bourbaki (1987) and Saks (2005).
  6. ^ Bourbaki 1987.
  7. ^ Dunford & Schwartz 1958, §IV.16
  8. ^ In Dunford & Schwartz (1958, §IV.16), the result that every linear functional on L2[0,1] is jointly attributed to Fréchet (1907) and Riesz (1907). The general result, that the dual of a Hilbert space is identified with the Hilbert space itself, can be found in Riesz (1934).
  9. ^ von Neumann 1929.
  10. ^ Hilbert, Nordheim & von Neumann 1927.
  11. ^ Weyl 1931.
  12. ^ Prugovečki 1981, pp. 1–10.
  13. ^ Weyl 1931.
  14. ^ von Neumann 1932
  15. ^ Young 1987, Chapter 9.
  16. ^ The eigenvalues of the Fredholm kernel are 1/λ, which tend to zero.
  17. ^ Bers, John & Schechter 1981.
  18. ^ More detail on finite element methods from this point of view can be found in Brenner & Scott (2005).
  19. ^ The version stated here in the general setting of Hilbert spaces is from Halmos (1982). Generalizations to certain other classes of Banach spaces can be found in Walters (1982).
  20. ^ von Neumann 1932.
  21. ^ The mathematical material in this article can be found in any good textbook on functional analysis, such as Dieudonné (1960), Hewitt & Stromberg (1965), Reed & Simon (1983) or Rudin (1980).
  22. ^ Kolmogorov & Fomin 1970.
  23. ^ Halmos 1950, Section 42.
  24. ^ Hewitt & Stromberg 1965.
  25. ^ Bers, John & Schechter 1981.
  26. ^ Giusti 2003.
  27. ^ A general reference on Hardy spaces is the book Duren (1970).
  28. ^ The topological dual of an inner product space can be identified with its completion.
  29. ^ Young 1988, p. 23.
  30. ^ Clarkson 1936.
  31. ^ Dunford & Schwartz 1958, §IV.4.
  32. ^ Dunford & Schwartz 1958, §IV.4. Many authors, Dunford and Schwartz included, refer to this just as the dimension. Unless the Hilbert space is finite dimensional, this is not the same thing as its dimension as a linear space.
  33. ^ This standard result can be found in any book on Hilbert spaces, such as Young (1988, Theorem 15.3).
  34. ^ See Reed & Simon (1975) and Folland (1989).

[edit] References

[edit] External links

Hilbert Space at Mathworld

Personal tools