Self-adjoint operator

In mathematics, a self-adjoint operator on a finite-dimensional complex vector space V with inner product is a linear map A (from V to itself) that is its own adjoint: . If V is finite-dimensional with a given orthonormal basis, this is equivalent to the condition that the matrix of A is Hermitian, i.e., equal to its conjugate transpose A. By the finite-dimensional spectral theorem, V has an orthonormal basis such that the matrix of A relative to this basis is a diagonal matrix with entries in the real numbers. In this article, we consider generalizations of this concept to operators on Hilbert spaces of arbitrary dimension.

Self-adjoint operators are used in functional analysis and quantum mechanics. In quantum mechanics their importance lies in the Dirac–von Neumann formulation of quantum mechanics, in which physical observables such as position, momentum, angular momentum and spin are represented by self-adjoint operators on a Hilbert space. Of particular significance is the Hamiltonian

which as an observable corresponds to the total energy of a particle of mass m in a real potential field V. Differential operators are an important class of unbounded operators.

The structure of self-adjoint operators on infinite-dimensional Hilbert spaces essentially resembles the finite-dimensional case. That is to say, operators are self-adjoint if and only if they are unitarily equivalent to real-valued multiplication operators. With suitable modifications, this result can be extended to possibly unbounded operators on infinite-dimensional spaces. Since an everywhere-defined self-adjoint operator is necessarily bounded, one needs be more attentive to the domain issue in the unbounded case. This is explained below in more detail.

Bounded self-adjoint operators

Suppose is a bounded linear operator from a Hilbert space H to itself. Then there is a unique bounded operator , called the adjoint of such that (in bracket notation)

for all in H.[1] We say that A is self-adjoint (physicists use the term "Hermitian") if . Equivalently, a bounded operator A is self-adjoint if

for all x and y in H.

Symmetric operators

Subtleties of the unbounded case

In many applications, we are led to consider operators that are unbounded; examples include the position, momentum, and Hamiltonian operators in quantum mechanics, as well as many differential operators. In the unbounded case, there are a number of subtle technical issues that have to dealt with. In particular, there is a crucial distinction between operators that are merely "symmetric" (defined in this section) and those that are "self-adjoint" (defined in the next section). In the case of differential operators defined on bounded domains, these technical issues have to do with making an appropriate choice of boundary conditions.

Definition of a symmetric operator

We now consider an unbounded operator A on a Hilbert space H. This means A is a linear map from a subspace of H—the "domain" of A, denoted —to H itself. We typically assume that is a dense subspace of H. Such an operator is called symmetric if, in bracket notation,

for all elements x and y in the domain of A.

If A is symmetric and , then A is necessarily bounded.[2] That is to say, an unbounded symmetric operator cannot be defined on the whole Hilbert space. Since the operators considered in quantum mechanics are unbounded, it is impossible to define them as symmetric operators on the whole Hilbert space.

In the physics literature, the term Hermitian is used in place of the term symmetric. It should be noted, however, that the physics literature generally glosses over the distinction between operators that are merely symmetric and operators that are actually self-adjoint (as defined in the next section).

Although the notion of a symmetric operator is easy to understand, it is not the "right" notion in the unbounded case. Specifically, the spectral theorem applies only to operators that are self-adjoint (defined in the next section) and not to operators that are merely symmetric. In particular, although the eigenvalues of a symmetric operator are necessarily real, a symmetric operator need not have any eigenvectors, let alone an orthonormal basis of them.

More generally, a partially defined linear operator A from a topological vector space E into its continuous dual space E is said to be symmetric if

for all elements x and y in the domain of A. This usage is fairly standard in the functional analysis literature.

A simple example

As noted above, the spectral theorem applies only to self-adjoint operators, and not in general to symmetric operators. Nevertheless, we can at this point give a simple example of a symmetric operator that has an orthonormal basis of eigenvectors. (This operator is actually "essentially self-adjoint.") The operator A below can be seen to have a compact inverse, meaning that the corresponding differential equation Af = g is solved by some integral, therefore compact, operator G. The compact symmetric operator G then has a countable family of eigenvectors which are complete in L2. The same can then be said for A.

Consider the complex Hilbert space L2[0,1] and the differential operator

with consisting of all complex-valued infinitely differentiable functions f on [0, 1] satisfying the boundary conditions

Then integration by parts shows that A is symmetric. The reader is invited to perform integration by parts twice and verify that the given boundary conditions for ensure that the boundary terms in the integration by parts vanish.

The eigenfunctions of A are the sinusoids

with the real eigenvalues n2π2; the well-known orthogonality of the sine functions follows as a consequence of the property of being symmetric.

We consider generalizations of this operator below.

Self-adjoint operators

Defintion of a self-adjoint operator

Briefly, a densely defined linear operator A on a Hilbert space is self-adjoint if it equals its adjoint. That is to say, A is self-adjoint that the domain of A coincides with the domain of the adjoint, and that A coincides with its adjoint on this common domain.

We now elaborate on the above definition. Given a densely defined linear operator A on H, its adjoint A is defined as follows:

(which is a densely defined linear map) is a continuous linear functional. By continuity and density of the domain of A, it extends to a unique continuous linear functional on all of H.
This vector z is defined to be A x. It can be shown that the dependence of z on x is linear.

Notice that it is the denseness of the domain of the operator, along with the uniqueness part of Riesz representation, that ensures the adjoint operator is well defined.

A result of Hellinger-Toeplitz type says that an operator having an everywhere-defined bounded adjoint is bounded.

The condition for a linear operator on a Hilbert space to be self-adjoint is stronger than to be symmetric. Although this distinction is technical, it is very important; the spectral theorem applies only to operators that are self-adjoint and not to operators that are merely symmetric. For an extensive discussion of the distinction, see Chapter 9 of Hall (2013).

For any densely defined operator A on Hilbert space one can define its adjoint operator A. For a symmetric operator A, the domain of the operator A contains the domain of the operator A, and the restriction of the operator A on the domain of A coincides with the operator A, i.e. AA, in other words A is extension of A. For a self-adjoint operator A the domain of A is the same as the domain of A, and A = A. See also Extensions of symmetric operators and unbounded operator.

Geometric interpretation

There is a useful geometric way of looking at the adjoint of an operator A on H as follows: we consider the graph G(A) of A defined by

Theorem. Let J be the symplectic mapping
Then the graph of A is the orthogonal complement of JG(A):

A densely defined operator A is symmetric if and only if AA, where the subset notation AA is understood to mean G(A) ⊆ G(A). An operator A is self-adjoint if and only if A = A; that is, if and only if {{{1}}}.

An example

Consider the complex Hilbert space L2(R), and the operator which multiplies a given function by x:

The domain of A is the space of all L2 functions for which for which is also square-integrable. Then A is self-adjoint.[3] On the other hand, A does not have any eigenfunctions. (More precisely, A does not have any normalizable eigenvectors, that is, eigenvectors that are actually in Hilbert space on which A is defined.)

As we will see later, self-adjoint operators have very important spectral properties; they are in fact multiplication operators on general measure spaces.

Spectral theorem

Statement of the spectral theorem

Partially defined operators A, B on Hilbert spaces H, K are unitarily equivalent if and only if there is a unitary transformation U : HK such that

A multiplication operator is defined as follows: Let (X, Σ, μ) be a countably additive measure space and f a real-valued measurable function on X. An operator T of the form

whose domain is the space of ψ for which the right-hand side above is in L2 is called a multiplication operator.

One version of the spectral theorem can be stated as follows.

Theorem. Any multiplication operator is a (densely defined) self-adjoint operator. Any self-adjoint operator is unitarily equivalent to a multiplication operator.[4]

Other version of the spectral theorem can be found in the spectral theorem article linked to above.

The spectral theorem for unbounded self-adjoint operators can be proved by reduction to the spectral theorem for unitary (hence bounded) operators.[5] This reduction uses the Cayley transform for self-adjoint operators which is defined in the next section. We might note that if T is multiplication by f, then the spectrum of T is just the essential range of f.

Borel functional calculus

Given the representation of T as a multiplication operator, it is easy to characterize the Borel functional calculus: If h is a bounded real-valued Borel function on R, then h(T) is the operator of multiplication by the composition hf. In order for this to be well-defined, we must show that it is the unique operation on bounded real-valued Borel functions satisfying a number of conditions.

Resolution of the identity

It has been customary to introduce the following notation

where is the characteristic function of the interval . The family of projection operators ET(λ) is called resolution of the identity for T. Moreover, the following Stieltjes integral representation for T can be proved:

The definition of the operator integral above can be reduced to that of a scalar valued Stieltjes integral using the weak operator topology. In more modern treatments however, this representation is usually avoided, since most technical problems can be dealt with by the functional calculus.

Formulation in the physics literature

In physics, particularly in quantum mechanics, the spectral theorem is expressed in a way which combines the spectral theorem as stated above and the Borel functional calculus using Dirac notation as follows:

If H is self-adjoint and f is a Borel function,

with

where the integral runs over the whole spectrum of H. The notation suggests that H is diagonalized by the eigenvectors ΨE. Such a notation is purely formal. One can see the similarity between Dirac's notation and the previous section. The resolution of the identity (sometimes called projection valued measures) formally resembles the rank-1 projections . In the Dirac notation, (projective) measurements are described via eigenvalues and eigenstates, both purely formal objects. As one would expect, this does not survive passage to the resolution of the identity. In the latter formulation, measurements are described using the spectral measure of , if the system is prepared in prior to the measurement. Alternatively, if one would like to preserve the notion of eigenstates and make it rigorous, rather than merely formal, one can replace the state space by a suitable rigged Hilbert space.

If f = 1, the theorem is referred to as resolution of unity:

In the case is the sum of an Hermitian H and a skew-Hermitian (see skew-Hermitian matrix) operator , one defines the biorthogonal basis set

and write the spectral theorem as:

(See Feshbach–Fano partitioning method for the context where such operators appear in scattering theory).

Extensions of symmetric operators

The following question arises in several contexts: if an operator A on the Hilbert space H is symmetric, when does it have self-adjoint extensions? An operator that has a unique self-adjoint extension is said to be essentially self-adjoint; equivalently, an operator is essentially self-adjoint if its closure (the operator whose graph is the closure of the graph of A) is self-adjoint. In general, a symmetric operators could have many self-adjoint extensions or none at all. Thus, we would like a classification of its self-adjoint extensions.

The first basic criterion for essential self-adjointness is the following:[6]

Theorem: If A is a symmetric operator on H, then A is essentially self-adjoint if and only if the range of the operators and are dense in H.

Equivalently, A is essentially self-adjoint if and only if the operators and have trivial kernels.[7] That is to say, A fails to be self-adjoint if and only if has an eigenvector with eigenvalue or .

Another way of looking at the issue is provided by the Cayley transform of a self-adjoint operator and the deficiency indices. (We should note here that it is often of technical convenience to deal with closed operators. In the symmetric case, the closedness requirement poses no obstacles, since it is known that all symmetric operators are closable.)

Theorem. Suppose A is a symmetric operator. Then there is a unique partially defined linear operator
such that

Here, ran and dom denote the image (in other words, range) and the domain, respectively. W(A) is isometric on its domain. Moreover, the range of 1  W(A) is dense in H.

Conversely, given any partially defined operator U which is isometric on its domain (which is not necessarily closed) and such that 1  U is dense, there is a (unique) operator S(U)

such that

The operator S(U) is densely defined and symmetric.

The mappings W and S are inverses of each other.

The mapping W is called the Cayley transform. It associates a partially defined isometry to any symmetric densely defined operator. Note that the mappings W and S are monotone: This means that if B is a symmetric operator that extends the densely defined symmetric operator A, then W(B) extends W(A), and similarly for S.

Theorem. A necessary and sufficient condition for A to be self-adjoint is that its Cayley transform W(A) be unitary.

This immediately gives us a necessary and sufficient condition for A to have a self-adjoint extension, as follows:

Theorem. A necessary and sufficient condition for A to have a self-adjoint extension is that W(A) have a unitary extension.

A partially defined isometric operator V on a Hilbert space H has a unique isometric extension to the norm closure of dom(V). A partially defined isometric operator with closed domain is called a partial isometry.

Given a partial isometry V, the deficiency indices of V are defined as the dimension of the orthogonal complements of the domain and range:

Theorem. A partial isometry V has a unitary extension if and only if the deficiency indices are identical. Moreover, V has a unique unitary extension if and only if the deficiency indices are both zero.

We see that there is a bijection between symmetric extensions of an operator and isometric extensions of its Cayley transform. The symmetric extension is self-adjoint if and only if the corresponding isometric extension is unitary.

A symmetric operator has a unique self-adjoint extension if and only if both its deficiency indices are zero. Such an operator is said to be essentially self-adjoint. Symmetric operators which are not essentially self-adjoint may still have a canonical self-adjoint extension. Such is the case for non-negative symmetric operators (or more generally, operators which are bounded below). These operators always have a canonically defined Friedrichs extension and for these operators we can define a canonical functional calculus. Many operators that occur in analysis are bounded below (such as the negative of the Laplacian operator), so the issue of essential adjointness for these operators is less critical.

Self-adjoint extensions in quantum mechanics

In quantum mechanics, observables correspond to self-adjoint operators. By Stone's theorem on one-parameter unitary groups, self-adjoint operators are precisely the infinitesimal generators of unitary groups of time evolution operators. However, many physical problems are formulated as a time-evolution equation involving differential operators for which the Hamiltonian is only symmetric. In such cases, either the Hamiltonian is essentially self-adjoint, in which case the physical problem has unique solutions or one attempts to find self-adjoint extensions of the Hamiltonian corresponding to different types of boundary conditions or conditions at infinity.

Example. The one-dimensional Schrödinger operator with the potential , defined initially on smooth compactly supported functions, is essentially self-adjoint (that is, has a self-adjoint closure) for 0 < α ≤ 2 but not for α > 2. See Berezin and Schubin, pages 55 and 86, or Section 9.10 in Hall.

The failure of essential self-adjointness for has a counterpart in the classical dynamics of a particle with potential : The classical particle escapes to infinity in finite time.[8]

Example. There is no self-adjoint momentum operator p for a particle moving on a half-line. Nevertheless, the Hamiltonian of a "free" particle on a half-line has several self-adjoint extensions corresponding to different types of boundary conditions. Physically, these boundary conditions are related to reflections of the particle at the origin (see Reed and Simon, vol.2).

Von Neumann's formulas

Suppose A is symmetric densely defined. Then any symmetric extension of A is a restriction of A*. Indeed, AB and B symmetric yields BA* by applying the definition of dom(A*).

Theorem. Suppose A is a densely defined symmetric operator. Let
Then
and
where the decomposition is orthogonal relative to the graph inner product of dom(A*):
.

These are referred to as von Neumann's formulas in the Akhiezer and Glazman reference.

Examples

A symmetric operator that is not essentially self-adjoint

We first consider the Hilbert space and the differential operator

defined on the space of continuously differentiable complex-valued functions on [0,1], satisfying the boundary conditions

Then D is a symmetric operator as can be shown by integration by parts. The spaces N+, N are given respectively by the distributional solutions to the equation

which are in L2[0, 1]. One can show that each one of these solution spaces is 1-dimensional, generated by the functions xe-x and xex respectively. This shows that D is not essentially self-adjoint,[9] but does have self-adjoint extensions. These self-adjoint extensions are parametrized by the space of unitary mappings N+N, which in this case happens to be the unit circle T.

In this case, the failure of essential self-adjointenss is due to an "incorrect" choice of boundary conditions in the definition of the domain of . Since is a first-order operator, only one boundary condition is needed to ensure that is symmetric. If we replaced the boundary conditions given above by the single boundary condition

,

then D would still be symmetric and would now, in fact, be essentially self-adjoint. This change of boundary conditions gives one particular essentially self-adjoint extension of D. Other essentially self-adjoint extensions come from imposing boundary conditions of the form .

This simple example illustrates a general fact about self-adjoint extensions of symmetric differential operators P on an open set M. They are determined by the unitary maps between the eigenvalue spaces

where Pdist is the distributional extension of P.

Constant-coefficient operators

We next give the example of differential operators with constant coefficients. Let

be a polynomial on Rn with real coefficients, where α ranges over a (finite) set of multi-indices. Thus

and

We also use the notation

Then the operator P(D) defined on the space of infinitely differentiable functions of compact support on Rn by

is essentially self-adjoint on L2(Rn).

Theorem. Let P a polynomial function on Rn with real coefficients, F the Fourier transform considered as a unitary map L2(Rn) → L2(Rn). Then F*P(D)F is essentially self-adjoint and its unique self-adjoint extension is the operator of multiplication by the function P.

More generally, consider linear differential operators acting on infinitely differentiable complex-valued functions of compact support. If M is an open subset of Rn

where aα are (not necessarily constant) infinitely differentiable functions. P is a linear operator

Corresponding to P there is another differential operator, the formal adjoint of P

Theorem. The adjoint P* of P is a restriction of the distributional extension of the formal adjoint to an appropriate subspace of . Specifically:

Spectral multiplicity theory

The multiplication representation of a self-adjoint operator, though extremely useful, is not a canonical representation. This suggests that it is not easy to extract from this representation a criterion to determine when self-adjoint operators A and B are unitarily equivalent. The finest grained representation which we now discuss involves spectral multiplicity. This circle of results is called the Hahn-Hellinger theory of spectral multiplicity.

Uniform multiplicity

We first define uniform multiplicity:

Definition. A self-adjoint operator A has uniform multiplicity n where n is such that 1 ≤ n ≤ ω if and only if A is unitarily equivalent to the operator Mf of multiplication by the function f(λ) = λ on

where Hn is a Hilbert space of dimension n. The domain of Mf consists of vector-valued functions ψ on R such that

Non-negative countably additive measures μ, ν are mutually singular if and only if they are supported on disjoint Borel sets.

Theorem. Let A be a self-adjoint operator on a separable Hilbert space H. Then there is an ω sequence of countably additive finite measures on R (some of which may be identically 0)
such that the measures are pairwise singular and A is unitarily equivalent to the operator of multiplication by the function f(λ) = λ on

This representation is unique in the following sense: For any two such representations of the same A, the corresponding measures are equivalent in the sense that they have the same sets of measure 0.

Direct integrals

The spectral multiplicity theorem can be reformulated using the language of direct integrals of Hilbert spaces:

Theorem.[10] Any self-adjoint operator on a separable Hilbert space is unitarily equivalent to multiplication by the function λ ↦ λ on

Unlike the multiplication-operator version of the spectral theorem, the direct-integral version is unique in the sense that the measure equivalence class of μ (or equivalently its sets of measure 0) is uniquely determined and the measurable function is determined almost everywhere with respect to μ.[11] The function is the spectral multiplicity function of the operator.

We may now state the classification result for self-adjoint operators: Two self-adjoint operators are unitarily equivalent if and only if (1) their spectra agree as sets, (2) the measures appearing in their direct-integral representations have the same sets of measure zero, and (3) their spectral multiplicity functions agree almost everywhere with respect to the measure in the direct integral.[12]

Example: structure of the Laplacian

The Laplacian on Rn is the operator

As remarked above, the Laplacian is diagonalized by the Fourier transform. Actually it is more natural to consider the negative of the Laplacian −Δ since as an operator it is non-negative; (see elliptic operator).

Theorem. If n=1, then −Δ has uniform multiplicity , otherwise −Δ has uniform multiplicity . Moreover, the measure μmult may be taken to be Lebesgue measure on [0, ∞).

Pure point spectrum

A self-adjoint operator A on H has pure point spectrum if and only if H has an orthonormal basis {ei}i ∈ I consisting of eigenvectors for A.

Example. The Hamiltonian for the harmonic oscillator has a quadratic potential V, that is

This Hamiltonian has pure point spectrum; this is typical for bound state Hamiltonians in quantum mechanics. As was pointed out in a previous example, a sufficient condition that an unbounded symmetric operator has eigenvectors which form a Hilbert space basis is that it has a compact inverse.

See also

Citations

  1. Hall 2013 Proposition A.53
  2. Hall 2013 Corollary 9.9
  3. Hall 2013 Proposition 9.30
  4. Hall 2013 Theorems 7.20 and 10.10
  5. Hall 2013 Section 10.4
  6. Hall 2013 Theorem 9.21
  7. Hall 2013 Corollary 9.22
  8. Hall 2013 Chapter 2, Exercise 4
  9. Hall 2013 Section 9.6
  10. Hall 2013 Theorems 7.19 and 10.9
  11. Hall 2013 Proposition 7.22
  12. Hall 2013 Proposition 7.24

References

This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.